diff --git "a/perf-df-gptq-1xA10.csv" "b/perf-df-gptq-1xA10.csv" --- "a/perf-df-gptq-1xA10.csv" +++ "b/perf-df-gptq-1xA10.csv" @@ -67,7 +67,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c137d-54cb0251522db05c7e9430a8;ef7086f7-a9c3-4fd4-ab68-e2686ec395db) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4597-0331ce6477d4a0554ad84993;0d81923f-379e-485a-a9db-142111c00863) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -138,7 +138,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c13f3-6df068fd6fa2a29d131ca18a;f60246a8-512b-421d-900e-0ea5ffe2a206) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d460b-46efe85357291070747b3d2b;56f72314-bc17-4100-b3e6-4351676e2cd2) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -195,7 +195,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpinc8pjz7/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpeq6s3eqx/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1243.045888,849.870848,0.0,203.423744,184.496128,s,17,0.16700380897521971,0.009823753469130572,0.00032406108228421964,0.009798175811767578,0.009877977752685547,0.010117497825622558,0.010848511009216308,"[0.011031264305114747, 0.009838335990905762, 0.009837311744689941, 0.009870592117309571, 0.009787679672241211, 0.00969315242767334, 0.009745439529418946, 0.009798175811767578, 0.00945907211303711, 0.009712384223937989, 0.009529888153076172, 0.009813471794128418, 0.009889056205749511, 0.009819840431213379, 0.009827679634094238, 0.009773152351379395, 0.009577312469482423]",tokens/s,26059.285873208773,kWh,1.1371295855728884e-07,6.228751583782483e-08,2.1097974316073842e-07,3.8698021755585207e-07,tokens/kWh,661532523.8506593,MB,1243.045888,849.870848,0.0,203.423744,193.844224,s,17,9.859052612304687,0.5799442713120404,0.009384098949348406,0.5792615356445312,0.582433935546875,0.589327099609375,0.609600576171875,"[0.6146689453125, 0.5820621337890625, 0.5795172729492187, 0.57863232421875, 0.5712827758789063, 0.5796158447265625, 0.5781971435546875, 0.5771149291992187, 0.5810087280273437, 0.5711693725585938, 0.5720592651367188, 0.5829916381835938, 0.5792615356445312, 0.5767430419921875, 0.573796875, 0.58035546875, 0.5805753173828125]",tokens/s,108.6311273624129,kWh,6.535953210618106e-06,3.5814351743626475e-06,1.157450808028629e-05,2.1691896465267043e-05,tokens/kWh,2904310.3769592154,,s,1070,10.00479437446596,0.00935027511632331,0.0012516390282944286,0.009197567939758301,0.009368793487548828,0.009731482028961182,0.019336427612304686,"[0.010523648262023925, 0.010109951972961426, 0.009951231956481933, 0.009835552215576172, 0.00968393611907959, 0.01023692798614502, 0.009938943862915038, 0.009607168197631836, 0.009637887954711915, 0.00972697639465332, 0.01003929615020752, 0.009821184158325195, 0.009737215995788574, 0.009621503829956055, 0.010085375785827636, 0.009829376220703125, 0.009803775787353516, 0.009574399948120118, 0.009894911766052245, 0.009881600379943848, 0.009778176307678223, 0.009654272079467773, 0.009691136360168457, 0.009679871559143067, 0.009843711853027343, 0.00973516845703125, 0.009602047920227052, 0.009592831611633301, 0.009702400207519531, 0.009350144386291503, 0.009743359565734864, 0.009999456405639649, 0.00969206428527832, 0.009797632217407226, 0.009861120223999023, 0.00974233627319336, 0.009634816169738769, 0.009597951889038087, 0.00927948760986328, 0.009200639724731445, 0.009762816429138184, 0.00970956802368164, 0.009640959739685059, 0.009963520050048828, 0.010007552146911621, 0.009977919578552246, 0.0097074556350708, 0.009759743690490723, 0.009867263793945312, 0.009690112113952636, 0.009538559913635255, 0.009515007972717286, 0.00961945629119873, 0.00930406379699707, 0.009401344299316406, 0.009560064315795898, 0.009651200294494629, 0.009743359565734864, 0.009855999946594238, 0.00983142375946045, 0.009650176048278808, 0.009390080451965332, 0.019545087814331053, 0.009326592445373535, 0.009299967765808105, 0.009189375877380371, 0.00922316837310791, 0.00914739227294922, 0.009115648269653321, 0.009276479721069335, 0.009284543991088866, 0.009153535842895508, 0.009074687957763672, 0.009202688217163087, 0.009183232307434081, 0.009252863883972168, 0.009224224090576171, 0.009331680297851563, 0.009191424369812011, 0.009270272254943847, 0.009176159858703613, 0.009543583869934083, 0.009376768112182618, 0.00923750400543213, 0.009227264404296874, 0.009970687866210937, 0.009368703842163087, 0.009185152053833008, 0.009369600296020507, 0.009630720138549804, 0.009368576049804688, 0.009257984161376954, 0.009250847816467286, 0.009198559761047363, 0.00923136043548584, 0.009145343780517578, 0.009137311935424804, 0.009149439811706543, 0.009200480461120606, 0.009233407974243164, 0.009220095634460449, 0.009228287696838379, 0.009299967765808105, 0.009211903572082519, 0.009084927558898925, 0.00913920021057129, 0.009142271995544434, 0.009117695808410644, 0.00911571216583252, 0.009382847785949708, 0.00935321617126465, 0.009257984161376954, 0.009208831787109375, 0.009076767921447754, 0.009084896087646484, 0.009037823677062988, 0.00918015956878662, 0.009111552238464356, 0.00919961643218994, 0.009134079933166504, 0.009264127731323242, 0.00928767967224121, 0.00912281608581543, 0.009064448356628419, 0.009074687957763672, 0.019499008178710937, 0.009242624282836913, 0.00923033618927002, 0.009169919967651367, 0.00912384033203125, 0.009058303833007812, 0.009085951805114746, 0.009171968460083007, 0.009233407974243164, 0.009173024177551269, 0.009084896087646484, 0.009076736450195312, 0.009096192359924317, 0.009189375877380371, 0.009201663970947266, 0.009220095634460449, 0.009248767852783203, 0.00922316837310791, 0.00923033618927002, 0.009216064453125, 0.00903878402709961, 0.009159680366516113, 0.00916380786895752, 0.009100255966186523, 0.009077759742736816, 0.00923033618927002, 0.009047039985656738, 0.009165823936462402, 0.00921497631072998, 0.00922111988067627, 0.00922214412689209, 0.00921292781829834, 0.009175040245056153, 0.009157631874084473, 0.009191424369812011, 0.0091146240234375, 0.009119744300842286, 0.009169919967651367, 0.009189375877380371, 0.0092293119430542, 0.009218048095703125, 0.009198592185974122, 0.00918835163116455, 0.009192447662353515, 0.00918835163116455, 0.009200639724731445, 0.009191424369812011, 0.009219072341918945, 0.009126912117004395, 0.009227264404296874, 0.009175040245056153, 0.009267200469970703, 0.009219072341918945, 0.009129983901977539, 0.009157631874084473, 0.009280511856079102, 0.009280511856079102, 0.009463808059692384, 0.009468928337097168, 0.009313280105590821, 0.009226240158081055, 0.009244671821594238, 0.009248767852783203, 0.019546112060546874, 0.009198592185974122, 0.00923033618927002, 0.009241600036621094, 0.009197567939758301, 0.009257984161376954, 0.00921395206451416, 0.00919961643218994, 0.009201663970947266, 0.009218048095703125, 0.009255935668945312, 0.009029631614685058, 0.009183232307434081, 0.009269280433654785, 0.009225184440612794, 0.009245696067810059, 0.009326720237731933, 0.009116543769836425, 0.009233407974243164, 0.009166848182678223, 0.00922111988067627, 0.00918015956878662, 0.009134079933166504, 0.009088000297546387, 0.009176063537597656, 0.009102335929870605, 0.009143296241760255, 0.009162752151489258, 0.009256959915161133, 0.00919654369354248, 0.009159680366516113, 0.009198592185974122, 0.009145343780517578, 0.009176063537597656, 0.009148415565490722, 0.00923136043548584, 0.00923750400543213, 0.009146368026733399, 0.009104384422302245, 0.009142271995544434, 0.00910540771484375, 0.00912384033203125, 0.00918015956878662, 0.009149439811706543, 0.009192511558532715, 0.009216959953308106, 0.009167872428894042, 0.009246720314025878, 0.009243647575378417, 0.009190400123596192, 0.00920576000213623, 0.009155584335327148, 0.009167872428894042, 0.009138175964355469, 0.009133055686950683, 0.009053183555603026, 0.00913920021057129, 0.009001055717468261, 0.009059231758117675, 0.009116671562194823, 0.00921292781829834, 0.009240575790405273, 0.009183232307434081, 0.01943142318725586, 0.009120767593383788, 0.009217023849487305, 0.009161727905273438, 0.009099264144897461, 0.009162752151489258, 0.00921497631072998, 0.009093119621276855, 0.009167872428894042, 0.009184255599975585, 0.009181183815002441, 0.009235456466674804, 0.009172991752624511, 0.009210880279541016, 0.00916592025756836, 0.009098143577575683, 0.009030655860900879, 0.009049087524414063, 0.009181183815002441, 0.009111552238464356, 0.009085951805114746, 0.009141247749328613, 0.009209856033325196, 0.009234432220458985, 0.009207807540893554, 0.009176063537597656, 0.009184255599975585, 0.009239551544189453, 0.009192447662353515, 0.009158656120300293, 0.009186304092407227, 0.009191424369812011, 0.009202688217163087, 0.009116671562194823, 0.009224191665649414, 0.009161727905273438, 0.009167872428894042, 0.009200639724731445, 0.009211903572082519, 0.0091627836227417, 0.009301983833312988, 0.00931942367553711, 0.009126912117004395, 0.008987648010253906, 0.008691712379455567, 0.008762368202209473, 0.008802304267883301, 0.00879923152923584, 0.008743935585021973, 0.00882585620880127, 0.008767487525939942, 0.008882176399230958, 0.008848383903503418, 0.008884223937988281, 0.008896512031555176, 0.0088340482711792, 0.008876031875610351, 0.008733695983886718, 0.008738816261291504, 0.008808575630187988, 0.008762240409851074, 0.008844287872314453, 0.008932352066040039, 0.019290111541748048, 0.009226240158081055, 0.009192543983459473, 0.009072735786437988, 0.009117504119873047, 0.009209856033325196, 0.009234432220458985, 0.009071616172790528, 0.009153535842895508, 0.009234432220458985, 0.009253888130187989, 0.009182208061218262, 0.009101311683654785, 0.009162752151489258, 0.009194496154785157, 0.009153535842895508, 0.009136128425598144, 0.009181183815002441, 0.00923750400543213, 0.009248767852783203, 0.009215999603271484, 0.009622528076171874, 0.00941977596282959, 0.009262080192565919, 0.009164799690246582, 0.009107456207275391, 0.009246720314025878, 0.009148415565490722, 0.009133055686950683, 0.009167872428894042, 0.009257984161376954, 0.009179136276245118, 0.009218048095703125, 0.009142271995544434, 0.009201663970947266, 0.009152511596679687, 0.009197567939758301, 0.009241600036621094, 0.009190400123596192, 0.00921497631072998, 0.009203712463378906, 0.009235456466674804, 0.009233407974243164, 0.009073792457580566, 0.009089920043945312, 0.009043968200683594, 0.009189375877380371, 0.00921497631072998, 0.009186304092407227, 0.009152511596679687, 0.009183232307434081, 0.00921395206451416, 0.009274368286132812, 0.009203712463378906, 0.009209856033325196, 0.009239583969116212, 0.009121760368347168, 0.009207807540893554, 0.009187328338623046, 0.00923033618927002, 0.009146368026733399, 0.00918015956878662, 0.00918835163116455, 0.019396608352661132, 0.009326592445373535, 0.009152511596679687, 0.009268223762512207, 0.009243647575378417, 0.009166848182678223, 0.008897536277770996, 0.008979455947875976, 0.009034751892089844, 0.009119744300842286, 0.009065471649169921, 0.009208831787109375, 0.009220095634460449, 0.009226240158081055, 0.00922111988067627, 0.009209856033325196, 0.00921292781829834, 0.009148447990417481, 0.009192416191101074, 0.009165823936462402, 0.009200639724731445, 0.009162752151489258, 0.009201663970947266, 0.009051136016845703, 0.009189375877380371, 0.009262080192565919, 0.009207903861999512, 0.00918825626373291, 0.009133055686950683, 0.00921497631072998, 0.009157631874084473, 0.009128959655761718, 0.009217087745666503, 0.009224127769470215, 0.009115839958190918, 0.009102144241333009, 0.009166848182678223, 0.00921395206451416, 0.009158656120300293, 0.00919974422454834, 0.009130047798156739, 0.009087807655334472, 0.009056256294250489, 0.009099264144897461, 0.009063424110412598, 0.009078783988952637, 0.009101311683654785, 0.009162752151489258, 0.009463808059692384, 0.009272319793701172, 0.009198592185974122, 0.009144319534301757, 0.009219072341918945, 0.009159680366516113, 0.009142271995544434, 0.009260031700134277, 0.009185279846191406, 0.00923136043548584, 0.009182208061218262, 0.009290752410888671, 0.009267200469970703, 0.009079808235168458, 0.009135104179382325, 0.019449855804443358, 0.009274368286132812, 0.009172991752624511, 0.009284607887268067, 0.009233407974243164, 0.009042943954467773, 0.009087072372436524, 0.00907356834411621, 0.0091146240234375, 0.009109503746032714, 0.009274368286132812, 0.009159680366516113, 0.00913920021057129, 0.009234432220458985, 0.009345184326171875, 0.009156448364257812, 0.009245696067810059, 0.009245696067810059, 0.0092293119430542, 0.009239551544189453, 0.00911673641204834, 0.009136063575744629, 0.009192447662353515, 0.009241600036621094, 0.009242624282836913, 0.009210880279541016, 0.009164799690246582, 0.009126912117004395, 0.009116671562194823, 0.00922111988067627, 0.009138175964355469, 0.009243647575378417, 0.009153535842895508, 0.009244671821594238, 0.009288703918457031, 0.009179136276245118, 0.009073663711547851, 0.009178208351135255, 0.009131936073303223, 0.009789440155029297, 0.010276864051818848, 0.009390080451965332, 0.009272319793701172, 0.00929792022705078, 0.009158656120300293, 0.009267264366149903, 0.009231295585632324, 0.009220095634460449, 0.009408512115478516, 0.00936355209350586, 0.00907049560546875, 0.008773632049560547, 0.008762368202209473, 0.00891596794128418, 0.008843263626098634, 0.008750080108642578, 0.008733695983886718, 0.008853504180908203, 0.008744959831237792, 0.008860671997070312, 0.00880947208404541, 0.008872960090637207, 0.008772607803344726, 0.019050495147705078, 0.00914739227294922, 0.009171968460083007, 0.00913920021057129, 0.009129983901977539, 0.009084927558898925, 0.009102399826049806, 0.009157567977905274, 0.009215999603271484, 0.009225215911865235, 0.009127936363220214, 0.009496576309204101, 0.00930303955078125, 0.009183232307434081, 0.00932249641418457, 0.0092293119430542, 0.00920787239074707, 0.009202688217163087, 0.009164735794067383, 0.00921497631072998, 0.009233407974243164, 0.00934502410888672, 0.009215999603271484, 0.009211903572082519, 0.009240575790405273, 0.009356287956237793, 0.009263104438781738, 0.009226240158081055, 0.009272319793701172, 0.009249792098999024, 0.009227264404296874, 0.009277440071105958, 0.009200672149658203, 0.009244640350341797, 0.009247743606567382, 0.00940339183807373, 0.00921395206451416, 0.009118720054626465, 0.009260031700134277, 0.009226240158081055, 0.00923033618927002, 0.009190400123596192, 0.009373696327209472, 0.009288703918457031, 0.009221152305603027, 0.009135071754455566, 0.009248767852783203, 0.009151679992675781, 0.00914412784576416, 0.00918015956878662, 0.009190400123596192, 0.009176063537597656, 0.009245696067810059, 0.00921292781829834, 0.009134079933166504, 0.009086976051330567, 0.009086112022399902, 0.009161567687988282, 0.009284607887268067, 0.00914739227294922, 0.009242624282836913, 0.009239551544189453, 0.009207807540893554, 0.019359743118286133, 0.009250816345214843, 0.009277440071105958, 0.009189375877380371, 0.009217023849487305, 0.009102335929870605, 0.009195520401000976, 0.009217023849487305, 0.009215999603271484, 0.009209856033325196, 0.009276415824890137, 0.009144319534301757, 0.00900710391998291, 0.008902655601501466, 0.00880025577545166, 0.008871935844421386, 0.008897536277770996, 0.008839167594909669, 0.008781824111938476, 0.008749055862426757, 0.00880947208404541, 0.008901632308959961, 0.008910847663879394, 0.00892518424987793, 0.008903679847717285, 0.008893440246582032, 0.008972288131713867, 0.00890777587890625, 0.00890675163269043, 0.008847519874572754, 0.008818528175354005, 0.008945664405822755, 0.008953856468200684, 0.008910847663879394, 0.008886272430419923, 0.008861760139465331, 0.008858559608459472, 0.00888319969177246, 0.009025535583496093, 0.00923033618927002, 0.009121888160705567, 0.009133983612060546, 0.009175040245056153, 0.00913920021057129, 0.00922316837310791, 0.009239551544189453, 0.009241600036621094, 0.009255935668945312, 0.009226240158081055, 0.009177087783813476, 0.009233407974243164, 0.009143296241760255, 0.00920473575592041, 0.00920473575592041, 0.009117695808410644, 0.009062399864196777, 0.009056415557861328, 0.009175999641418457, 0.009165727615356445, 0.009145343780517578, 0.009138175964355469, 0.009235456466674804, 0.009246784210205078, 0.01884774398803711, 0.00892204761505127, 0.008887295722961425, 0.00892416000366211, 0.008919039726257324, 0.008806400299072266, 0.008905728340148926, 0.00890880012512207, 0.008914943695068359, 0.008866815567016602, 0.008857600212097168, 0.00890880012512207, 0.008857600212097168, 0.008770560264587402, 0.008869888305664063, 0.008758272171020508, 0.008865792274475098, 0.008824831962585449, 0.009273344039916993, 0.009147520065307618, 0.009206656455993653, 0.009214015960693359, 0.00923539161682129, 0.0092293119430542, 0.009240575790405273, 0.009120767593383788, 0.009254912376403808, 0.009246720314025878, 0.009211903572082519, 0.009234432220458985, 0.00912281608581543, 0.009153535842895508, 0.009148415565490722, 0.009217023849487305, 0.009240575790405273, 0.009051136016845703, 0.009158656120300293, 0.009244671821594238, 0.009165823936462402, 0.009259103775024414, 0.00910428810119629, 0.009156607627868652, 0.009071616172790528, 0.009088000297546387, 0.009150464057922364, 0.009191424369812011, 0.009203712463378906, 0.009197567939758301, 0.009211903572082519, 0.009085951805114746, 0.009252863883972168, 0.009261055946350098, 0.009144319534301757, 0.0089169921875, 0.00882380771636963, 0.00890675163269043, 0.008846336364746094, 0.00879923152923584, 0.00942080020904541, 0.009156607627868652, 0.009150464057922364, 0.00923033618927002, 0.009256959915161133, 0.019325952529907226, 0.009145343780517578, 0.00921292781829834, 0.00920473575592041, 0.009232383728027344, 0.009175040245056153, 0.009247743606567382, 0.009246720314025878, 0.009155584335327148, 0.009269248008728028, 0.00930406379699707, 0.009315327644348144, 0.009176063537597656, 0.00922111988067627, 0.009227264404296874, 0.00912281608581543, 0.00921395206451416, 0.009234432220458985, 0.009240575790405273, 0.009179136276245118, 0.009125887870788574, 0.00921292781829834, 0.00922537612915039, 0.00909398365020752, 0.009247743606567382, 0.009356287956237793, 0.009209856033325196, 0.009286656379699706, 0.009254912376403808, 0.009262080192565919, 0.009192447662353515, 0.00913100814819336, 0.009253888130187989, 0.009264127731323242, 0.009240575790405273, 0.009261055946350098, 0.009332736015319825, 0.009190400123596192, 0.009409536361694336, 0.009316351890563965, 0.009192447662353515, 0.009083904266357423, 0.00921513557434082, 0.009099103927612305, 0.00910848045349121, 0.009153535842895508, 0.009072640419006347, 0.00912281608581543, 0.009566207885742188, 0.009355263710021973, 0.00941158390045166, 0.009475071907043458, 0.010164223670959472, 0.009293824195861817, 0.009284671783447265, 0.009354175567626953, 0.009158656120300293, 0.009213088035583496, 0.009136992454528809, 0.009243647575378417, 0.009215999603271484, 0.009174015998840332, 0.00939417552947998, 0.019589120864868165, 0.009201663970947266, 0.009107456207275391, 0.00919654369354248, 0.009144319534301757, 0.00919654369354248, 0.009233407974243164, 0.009263104438781738, 0.009289728164672852, 0.009229375839233398, 0.009210816383361816, 0.009252863883972168, 0.009246720314025878, 0.00920473575592041, 0.00913920021057129, 0.009211903572082519, 0.009214048385620116, 0.009109408378601073, 0.009092096328735352, 0.009267200469970703, 0.009323519706726074, 0.009115648269653321, 0.009198592185974122, 0.009293824195861817, 0.0091843843460083, 0.009113471984863281, 0.009152511596679687, 0.009136128425598144, 0.009265151977539063, 0.009224320411682129, 0.009140095710754394, 0.009064448356628419, 0.009190400123596192, 0.009260031700134277, 0.00921395206451416, 0.009155584335327148, 0.009053407669067383, 0.009170720100402832, 0.0092293119430542, 0.009143296241760255, 0.009264127731323242, 0.009234432220458985, 0.00914739227294922, 0.0092293119430542, 0.009291775703430176, 0.009253888130187989, 0.009151488304138184, 0.009143296241760255, 0.009112575531005859, 0.009107456207275391, 0.009168895721435547, 0.009296895980834961, 0.00920473575592041, 0.00914739227294922, 0.009156607627868652, 0.009156607627868652, 0.009170944213867188, 0.009110527992248535, 0.009191424369812011, 0.009135104179382325, 0.009087008476257324, 0.009227231979370117, 0.009193471908569336, 0.019509248733520508, 0.00922214412689209, 0.009254912376403808, 0.009165823936462402, 0.00908902359008789, 0.009220095634460449, 0.009126976013183593, 0.009373632431030274, 0.009495552062988282, 0.009178112030029297, 0.008806400299072266, 0.008794112205505371, 0.008841216087341308, 0.00890067195892334, 0.008871871948242188, 0.008905728340148926, 0.008902655601501466, 0.00890675163269043, 0.008870911598205567, 0.009075712203979493, 0.009273344039916993, 0.009257984161376954, 0.009076736450195312, 0.009226304054260254, 0.009189311981201171, 0.009143296241760255, 0.00908083152770996, 0.00911359977722168, 0.00919961643218994, 0.009136128425598144, 0.009226240158081055, 0.009202688217163087, 0.00930303955078125, 0.009268223762512207, 0.009245696067810059, 0.009154560089111329, 0.009047039985656738, 0.00915875244140625, 0.009263008117675782, 0.009268223762512207, 0.009255935668945312, 0.00932863998413086, 0.00923750400543213, 0.00943513584136963, 0.00921395206451416, 0.009183232307434081, 0.009143296241760255, 0.00923033618927002, 0.009234432220458985, 0.009183232307434081, 0.00921395206451416, 0.009129983901977539, 0.009178112030029297, 0.00922316837310791, 0.009220095634460449, 0.009103360176086426, 0.009099264144897461, 0.009206784248352052, 0.009102335929870605, 0.008966143608093263, 0.008952832221984864, 0.009235456466674804, 0.009166879653930664, 0.019551200866699217, 0.009203712463378906, 0.009207807540893554, 0.009195520401000976, 0.00920473575592041, 0.009081855773925781, 0.009138208389282226, 0.009206751823425292, 0.009069727897644044, 0.00909500789642334, 0.009075712203979493, 0.009251839637756347, 0.009143296241760255, 0.009044992446899413, 0.009095168113708496, 0.00909721565246582, 0.009273344039916993, 0.009202688217163087, 0.009220095634460449, 0.009200639724731445, 0.00919654369354248, 0.009187328338623046, 0.009194496154785157, 0.009220128059387206, 0.00885756778717041, 0.008859647750854491, 0.008879103660583497, 0.008896608352661133, 0.008827808380126954, 0.008890368461608887, 0.008872960090637207, 0.009155584335327148, 0.009268223762512207, 0.00923852825164795, 0.009135104179382325, 0.009187328338623046, 0.00908083152770996, 0.009074687957763672, 0.009110527992248535, 0.009248767852783203, 0.00918015956878662, 0.009179136276245118, 0.009178112030029297, 0.009066495895385742, 0.008925215721130371, 0.008839136123657226, 0.008696831703186036, 0.008950783729553222, 0.008897536277770996, 0.008856575965881347, 0.008797183990478515, 0.00890880012512207, 0.009142271995544434, 0.009152576446533204, 0.009179072380065918, 0.009263104438781738, 0.009179136276245118, 0.009220095634460449, 0.009257984161376954, 0.009208864212036132, 0.009212896347045899, 0.009203712463378906, 0.00923852825164795, 0.01962188720703125, 0.009197567939758301, 0.009276415824890137, 0.009210880279541016, 0.009266176223754884, 0.00922316837310791, 0.00920576000213623, 0.009263168334960938, 0.009200575828552247, 0.009222208023071289, 0.00924665641784668, 0.009208831787109375, 0.009190400123596192, 0.009347071647644043, 0.009125887870788574, 0.009094143867492676, 0.009088031768798828, 0.009106528282165528, 0.00907968044281006, 0.009120767593383788, 0.009215999603271484, 0.008967167854309082, 0.009009152412414552, 0.008977408409118653, 0.009110527992248535, 0.009375743865966797, 0.00921292781829834, 0.009167872428894042, 0.009269248008728028, 0.009095168113708496, 0.009136128425598144, 0.009124863624572753, 0.00910028839111328, 0.009101311683654785, 0.009169024467468262, 0.009172863960266113, 0.009210880279541016, 0.009251839637756347, 0.009948160171508789, 0.00923852825164795, 0.009289728164672852, 0.00926425552368164, 0.009149312019348144, 0.0092293119430542, 0.009250816345214843, 0.009380864143371583, 0.009245696067810059, 0.00917199993133545, 0.00909001636505127, 0.00918835163116455, 0.0091658878326416, 0.009192383766174317, 0.009256959915161133, 0.009270272254943847, 0.009243647575378417, 0.009161727905273438, 0.009264127731323242, 0.009207807540893554, 0.009226240158081055, 0.009208831787109375, 0.00914739227294922, 0.009169919967651367, 0.00922111988067627, 0.018906112670898437, 0.008923135757446288, 0.009217023849487305, 0.009284607887268067, 0.009157631874084473, 0.009200639724731445, 0.00923852825164795, 0.009200639724731445, 0.009148415565490722, 0.009266176223754884, 0.009198592185974122, 0.009218048095703125, 0.00910540771484375, 0.009266176223754884, 0.009264127731323242, 0.009206815719604492, 0.009223135948181152, 0.009200639724731445, 0.009989119529724122, 0.009961471557617188, 0.009261055946350098, 0.009271295547485351, 0.009243647575378417, 0.009207807540893554, 0.009073663711547851, 0.00886787223815918, 0.008962016105651855, 0.009142271995544434, 0.009162752151489258, 0.009256959915161133, 0.009211903572082519, 0.00918950366973877, 0.009131903648376466, 0.009225215911865235, 0.00922316837310791, 0.009207807540893554, 0.009172991752624511, 0.009079808235168458, 0.009226240158081055, 0.00923033618927002, 0.009140224456787109, 0.009226240158081055, 0.009263104438781738, 0.009155584335327148, 0.00922111988067627, 0.009246848106384277, 0.00925068759918213, 0.009235456466674804, 0.009288703918457031, 0.00912179183959961, 0.00919961643218994, 0.009259008407592773, 0.009135104179382325, 0.009215999603271484, 0.009168895721435547, 0.009155584335327148, 0.009294848442077636, 0.009154560089111329, 0.009138175964355469, 0.009104448318481445, 0.009225152015686034, 0.00921497631072998, 0.009239551544189453]",tokens/s,106.94872477647667,, @@ -320,6 +320,7 @@ ChildProcessError: Traceback (most recent call last): TypeError: DeciCoderAttention.forward() got an unexpected keyword argument 'cache_position' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2120.617984,5971.116032,0.0,5324.668928,5080.87552,s,10,3.160778564453125,0.3160778564453125,0.00014221386743603723,0.31608947753906247,0.31621777343749996,0.316263916015625,0.316300830078125,"[0.31631005859375, 0.31595333862304686, 0.31619436645507815, 0.3159564208984375, 0.315891845703125, 0.3161977844238281, 0.31620751953125, 0.31612481689453126, 0.3158882751464844, 0.31605413818359374]",tokens/s,809.9270315201371,kWh,3.7333295784062813e-06,2.0456940575684257e-06,1.835831329775016e-05,2.4137336933724864e-05,tokens/kWh,10605975.327887764,MB,2120.617984,5971.116032,0.0,5324.668928,5290.117632,s,10,173.24443359375002,17.324443359375003,0.0011951264280211484,17.324284179687503,17.325962304687497,17.325986035156248,17.326005019531248,"[17.32349609375, 17.326009765625, 17.3251484375, 17.325763671875, 17.32453515625, 17.323236328125, 17.32595703125, 17.324033203125, 17.32393359375, 17.3223203125]",tokens/s,3.636480474041204,kWh,0.00020452947697291774,0.00011209910411696113,0.00099598438706465,0.0013126129681545287,tokens/kWh,47995.868948769414,,s,629,175.81415676879894,0.27951376274848777,0.03754764840079795,0.27497882080078123,0.27517562255859374,0.27527659912109376,0.5910434716796875,"[0.2751549377441406, 0.27478631591796876, 0.274740234375, 0.2748467102050781, 0.2747781066894531, 0.27483648681640627, 0.27480575561523435, 0.27484774780273435, 0.27487026977539064, 0.2749368286132812, 0.2748989562988281, 0.2748334045410156, 0.27480679321289064, 0.2747279357910156, 0.27483444213867186, 0.2748395385742187, 0.2749573059082031, 0.2748907470703125, 0.2748989562988281, 0.274777099609375, 0.2748989562988281, 0.27485696411132815, 0.2748805236816406, 0.2748784790039063, 0.27539968872070314, 0.2749306945800781, 0.2748108825683594, 0.27496652221679685, 0.2749921264648437, 0.2748805236816406, 0.2750218200683594, 0.2749276123046875, 0.2749573059082031, 0.27505767822265625, 0.2749849548339844, 0.2750382080078125, 0.2751651916503906, 0.27496038818359375, 0.2750003051757812, 0.27507302856445315, 0.2750699462890625, 0.27509246826171874, 0.275346435546875, 0.275018798828125, 0.2749961853027344, 0.2750627746582031, 0.2750822448730469, 0.2749767761230469, 0.2750279541015625, 0.2750679016113281, 0.2748856201171875, 0.27508120727539065, 0.27509246826171874, 0.27501669311523436, 0.27497882080078123, 0.2750330810546875, 0.27511911010742185, 0.27523687744140624, 0.27515802001953127, 0.275162109375, 0.27502590942382815, 0.2751109008789063, 0.5910446166992187, 0.2748846435546875, 0.27493679809570315, 0.27485592651367186, 0.27479757690429685, 0.2749266052246094, 0.2747801513671875, 0.2748385314941406, 0.27494503784179686, 0.2749010009765625, 0.27492044067382815, 0.27491122436523435, 0.27478842163085937, 0.2750320129394531, 0.27486822509765624, 0.2753914794921875, 0.2748610534667969, 0.274977783203125, 0.27492352294921873, 0.27486309814453125, 0.27502899169921874, 0.27495526123046876, 0.2748989562988281, 0.2748948364257813, 0.27489279174804687, 0.27484774780273435, 0.2748805236816406, 0.2749081726074219, 0.27503411865234373, 0.27519793701171874, 0.27506591796875, 0.27506375122070315, 0.27527474975585936, 0.2748713073730469, 0.2750074768066406, 0.274904052734375, 0.2754652099609375, 0.2749624328613281, 0.2749819030761719, 0.2750412902832031, 0.27497369384765624, 0.27500851440429686, 0.27519180297851564, 0.274872314453125, 0.275061767578125, 0.2750320739746094, 0.2752962646484375, 0.27497268676757813, 0.2749819030761719, 0.2750351257324219, 0.2755778503417969, 0.27515597534179687, 0.2751201171875, 0.27508837890625, 0.2750494689941406, 0.27514776611328123, 0.27520306396484373, 0.2751528930664062, 0.2750115966796875, 0.275156005859375, 0.27518563842773436, 0.2749910888671875, 0.27510885620117187, 0.5912586059570313, 0.27490304565429685, 0.27485491943359375, 0.2748385314941406, 0.2748313598632813, 0.27479244995117186, 0.27500851440429686, 0.27482110595703124, 0.2751180725097656, 0.27487640380859374, 0.2750003051757812, 0.27502081298828124, 0.2749696044921875, 0.2748538818359375, 0.2746982421875, 0.2748958740234375, 0.2749091796875, 0.2749286499023437, 0.275240966796875, 0.27487744140625, 0.2748395385742187, 0.2748037109375, 0.2750064697265625, 0.2748907470703125, 0.27497882080078123, 0.27496652221679685, 0.2749358215332031, 0.27493991088867187, 0.274845703125, 0.274914306640625, 0.2749286499023437, 0.27517645263671875, 0.2750771179199219, 0.2749706115722656, 0.27486822509765624, 0.275156982421875, 0.27497268676757813, 0.2749921264648437, 0.2750484619140625, 0.2749644775390625, 0.2750044250488281, 0.2749798278808594, 0.2752276611328125, 0.27546112060546873, 0.275125244140625, 0.2750402526855469, 0.27501773071289065, 0.2749358215332031, 0.2750412902832031, 0.2750894165039062, 0.2750074768066406, 0.2750545959472656, 0.2751150207519531, 0.2751928405761719, 0.27498086547851563, 0.27527783203125, 0.2749931640625, 0.2750330810546875, 0.27503411865234373, 0.2751528930664062, 0.27550515747070314, 0.2750269470214844, 0.2751150207519531, 0.59104052734375, 0.2749900817871094, 0.2748436584472656, 0.2749163513183594, 0.27488357543945313, 0.274850830078125, 0.27485592651367186, 0.2748692626953125, 0.27481497192382814, 0.2749962158203125, 0.2748784790039063, 0.27504229736328123, 0.274956298828125, 0.274934814453125, 0.2749541931152344, 0.2749941711425781, 0.27521844482421876, 0.2750484619140625, 0.27506381225585935, 0.27490509033203125, 0.2747699279785156, 0.2748641357421875, 0.27487850952148435, 0.27475555419921877, 0.27495013427734377, 0.2749266052246094, 0.2748743591308594, 0.2752542724609375, 0.2749706115722656, 0.27507611083984373, 0.2751109008789063, 0.2750330810546875, 0.2754150390625, 0.274977783203125, 0.27513345336914063, 0.27508428955078124, 0.27500238037109376, 0.27498699951171873, 0.275009521484375, 0.27486822509765624, 0.27494708251953126, 0.27500851440429686, 0.27512216186523436, 0.27486822509765624, 0.2749327392578125, 0.27506072998046877, 0.2749296569824219, 0.275009521484375, 0.27506072998046877, 0.2750382080078125, 0.2750146484375, 0.27508837890625, 0.2751744689941406, 0.2754303283691406, 0.2751958923339844, 0.2751508483886719, 0.2751016845703125, 0.27512115478515625, 0.27508633422851564, 0.275019775390625, 0.27512933349609375, 0.275093505859375, 0.275167236328125, 0.5909524536132813, 0.2748692626953125, 0.2747965393066406, 0.274798583984375, 0.274872314453125, 0.274951171875, 0.2747801513671875, 0.2747484130859375, 0.2748190612792969, 0.2747678833007812, 0.27482009887695313, 0.2748661804199219, 0.2749358215332031, 0.27501055908203126, 0.27486822509765624, 0.27493585205078125, 0.27487127685546875, 0.2749347839355469, 0.27503103637695314, 0.2748323974609375, 0.2747781066894531, 0.27493991088867187, 0.2749276123046875, 0.27532186889648436, 0.27497268676757813, 0.2752706604003906, 0.2750003051757812, 0.2747832336425781, 0.2749266052246094, 0.2750044250488281, 0.27501669311523436, 0.2750402526855469, 0.27488973999023436, 0.2749440002441406, 0.2749849548339844, 0.2750136413574219, 0.2751754150390625, 0.27496856689453125, 0.2750382080078125, 0.2750597229003906, 0.2750392456054688, 0.2752174072265625, 0.27517745971679686, 0.27495321655273436, 0.275156982421875, 0.27486004638671874, 0.2751426696777344, 0.2749839477539063, 0.27503717041015624, 0.2749931640625, 0.27507916259765625, 0.2750403442382813, 0.2750545043945313, 0.2751549377441406, 0.2750699462890625, 0.27504434204101563, 0.2751047668457031, 0.2751129455566406, 0.2751170654296875, 0.2751405944824219, 0.2752563171386719, 0.2750802001953125, 0.275093505859375, 0.591404052734375, 0.27478732299804687, 0.2747463684082031, 0.2747463684082031, 0.2748743591308594, 0.27486309814453125, 0.27487640380859374, 0.2747135925292969, 0.27524197387695315, 0.27495321655273436, 0.27483648681640627, 0.27523275756835935, 0.27500238037109376, 0.2752081909179688, 0.27478427124023436, 0.2749214782714844, 0.27479962158203125, 0.2749573059082031, 0.27495834350585935, 0.27488973999023436, 0.2750689392089844, 0.2748938293457031, 0.2749347839355469, 0.274798583984375, 0.2747658386230469, 0.2748211669921875, 0.2749531860351562, 0.2749931640625, 0.275009521484375, 0.27486822509765624, 0.2750822448730469, 0.27492556762695314, 0.27491839599609375, 0.27491738891601564, 0.27533108520507815, 0.27536587524414063, 0.27496652221679685, 0.27481808471679686, 0.27493991088867187, 0.2749531860351562, 0.27497573852539064, 0.274914306640625, 0.2750218200683594, 0.27485491943359375, 0.27489178466796876, 0.2750146484375, 0.2750074768066406, 0.2750136413574219, 0.2751447143554687, 0.2749767761230469, 0.2750484619140625, 0.275135498046875, 0.27490203857421874, 0.2749685363769531, 0.2750413513183594, 0.2751610107421875, 0.27513446044921874, 0.2751436767578125, 0.27492044067382815, 0.27500238037109376, 0.2750679016113281, 0.27492352294921873, 0.2751355285644531, 0.5913036499023437, 0.27488153076171873, 0.27485797119140626, 0.27473101806640626, 0.27497882080078123, 0.27478631591796876, 0.2751488037109375, 0.27473202514648437, 0.27486822509765624, 0.2748518371582031, 0.2749644775390625, 0.2750003051757812, 0.2749214782714844, 0.2749296569824219, 0.2748794860839844, 0.2749767761230469, 0.27479962158203125, 0.2754744262695312, 0.27506585693359376, 0.2748374938964844, 0.27503103637695314, 0.27487539672851563, 0.2749153137207031, 0.27516619873046877, 0.27500338745117187, 0.2749542541503906, 0.27485491943359375, 0.2749767761230469, 0.27487335205078123, 0.27509246826171874, 0.27496139526367186, 0.27499725341796877, 0.27492044067382815, 0.27487335205078123, 0.2748037109375, 0.2749992980957031, 0.274988037109375, 0.27499725341796877, 0.2749624328613281, 0.2749593505859375, 0.275557373046875, 0.2749491271972656, 0.27513650512695315, 0.27502899169921874, 0.2751129455566406, 0.2749962158203125, 0.27495526123046876, 0.2748590087890625, 0.2750382080078125, 0.2751129455566406, 0.27525323486328124, 0.2751539306640625, 0.27510784912109376, 0.2751170654296875, 0.275314697265625, 0.27513345336914063, 0.2750320739746094, 0.2750699462890625, 0.27560345458984375, 0.2751047668457031, 0.27525531005859377, 0.2750750732421875, 0.2749819030761719, 0.5911541748046875, 0.27480172729492186, 0.2748220825195313, 0.2748968811035156, 0.2747218017578125, 0.2748538818359375, 0.27491122436523435, 0.2748108825683594, 0.2747555847167969, 0.2747412414550781, 0.2749276123046875, 0.2749245300292969, 0.274904052734375, 0.2748374938964844, 0.27484979248046876, 0.27491326904296876, 0.2747781066894531, 0.27500338745117187, 0.27496551513671874, 0.274798583984375, 0.274956298828125, 0.27488973999023436, 0.27482827758789063, 0.2748784790039063, 0.2750003051757812, 0.27475250244140625, 0.2749429626464844, 0.27521331787109377, 0.2749491271972656, 0.2751150207519531, 0.2752850036621094, 0.2750074768066406, 0.27493374633789064, 0.2749347839355469, 0.27506072998046877, 0.2751426696777344, 0.2750392456054688, 0.2748118896484375, 0.27491943359375, 0.27496856689453125, 0.2750136413574219, 0.274808837890625, 0.275030029296875, 0.27494708251953126, 0.274977783203125, 0.27509451293945314, 0.27521331787109377, 0.27501568603515625, 0.27507098388671875, 0.27510272216796877, 0.2749358215332031, 0.27550515747070314, 0.2751405944824219, 0.27502490234375, 0.2750003051757812, 0.2750330810546875, 0.27517132568359376, 0.27509451293945314, 0.2750965881347656, 0.2751242370605469, 0.2751856689453125, 0.2751109008789063, 0.2755092468261719, 0.591119384765625, 0.2749716491699219, 0.2747607116699219, 0.27478118896484377, 0.27481497192382814, 0.2748661804199219, 0.27484466552734377, 0.27487539672851563, 0.27488973999023436, 0.2748272705078125, 0.27498291015625, 0.2748958740234375, 0.27494091796875, 0.27488153076171873, 0.2748334045410156, 0.2749521789550781, 0.27494503784179686, 0.27500750732421875, 0.2749541931152344, 0.27481396484375, 0.2748385314941406, 0.2752174072265625, 0.27478936767578127, 0.27481805419921873, 0.2750044250488281, 0.2748856201171875, 0.2749296569824219, 0.2748856201171875, 0.275061767578125, 0.27486209106445314, 0.2749542541503906, 0.274956298828125, 0.27497573852539064, 0.27501873779296876, 0.2750433349609375, 0.27517849731445315, 0.2750965881347656, 0.2750126037597656, 0.27508428955078124, 0.2750269470214844, 0.2750136413574219, 0.27496038818359375, 0.2749716491699219, 0.274956298828125, 0.27498904418945314, 0.2750955505371094, 0.2750392456054688, 0.27496652221679685, 0.2749573059082031, 0.27499520874023436, 0.27500851440429686, 0.275030029296875, 0.2750873718261719, 0.2749859924316406, 0.2750771179199219, 0.275009521484375, 0.2750412902832031, 0.27533721923828125, 0.2750392456054688, 0.27516415405273437, 0.27510067749023437, 0.27538534545898435, 0.2750955505371094, 0.5911121826171875, 0.27488461303710937, 0.2748856201171875, 0.2748938293457031, 0.27470950317382814, 0.27487850952148435, 0.2748907165527344, 0.274914306640625, 0.27480062866210936, 0.27482623291015623, 0.27487026977539064, 0.2751856689453125, 0.2748098449707031, 0.274740234375, 0.27479757690429685, 0.27485592651367186, 0.2749327392578125, 0.27489996337890626, 0.27496038818359375, 0.2748272705078125, 0.27478732299804687, 0.2748323974609375, 0.274845703125, 0.27469515991210935, 0.2749358215332031, 0.2748661804199219, 0.27492556762695314, 0.2748293151855469, 0.27499520874023436, 0.2749327392578125, 0.2750412902832031, 0.2750218200683594, 0.2752102355957031, 0.2749224853515625, 0.2749010009765625, 0.275061767578125, 0.27493991088867187, 0.27485592651367186, 0.2750115966796875, 0.27513037109375, 0.2749706115722656, 0.27488153076171873, 0.27488461303710937, 0.27494708251953126, 0.27503717041015624, 0.27507098388671875, 0.27512832641601564, 0.27502590942382815, 0.2750474243164063, 0.2748876647949219, 0.2750802001953125, 0.27499725341796877, 0.2751907958984375, 0.27516110229492186, 0.27500543212890627, 0.2750382385253906, 0.2752030334472656, 0.2751068115234375, 0.27504434204101563, 0.27506381225585935, 0.27498699951171873, 0.27508428955078124, 0.2751539306640625]",tokens/s,3.577641366088369,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -349,7 +350,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664beebc-6e94d652186738ba13db140a;befe1ce8-3224-469a-af5a-829d918d2545) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3fc9-255d513207a6fc8f4d7dbf04;ffb1d206-a64f-45eb-ac4b-b1d5c9fc1608) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -438,7 +439,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c13a5-6e2eaf2201cead1c3b9f1f48;2931b9fa-4994-444e-a4ec-bd207df69d37) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45bd-5edbf1033d431def21a8d61b;02e0ac9c-f9e7-4934-82b6-4b77d2c1e757) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -498,7 +499,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee94-434786f14bfc35b73fac3cb2;e8a261c3-dc4a-41d2-9d53-50086dcec1a5) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3fa2-14397b02004191183f59eebf;260c805b-8781-44ea-b7fb-9bd5f8898890) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -730,7 +731,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpvcj5erii/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp9w6o1ere/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -827,7 +828,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bed0a-068b4ddb5e785373073f7396;d0785fc9-af8e-4f28-990f-8c5f16111304) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3e22-3a3704da2bb6f6ab602279ea;cf85f07a-9d93-48de-aa1e-9ef6814c9fdc) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -912,7 +913,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c132e-2114ed7016d8c868623ae780;b2bb365c-8b58-4c28-b3ba-208baaea45a4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d454a-076952361d17f30002ce7f07;86e85c29-23a7-4454-b317-423757f72d7a) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1028,7 +1029,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c122b-593a0e66517dde61157e86ad;f28b33dc-6083-46ba-b6a8-548397be0731) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4484-0dc251300fa45c0c39a65050;658e5595-63ca-4548-90ca-0408cb791f21) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1102,6 +1103,35 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1847.382016,3393.716224,0.0,2749.366272,2585.357312,s,10,0.823750862121582,0.0823750862121582,0.0014776796762060414,0.08214677047729493,0.0831559356689453,0.0847691680908203,0.08605975402832032,"[0.08638240051269531, 0.0818265609741211, 0.08206022644042969, 0.08117469024658203, 0.08168768310546876, 0.08057820892333985, 0.08223331451416016, 0.08279743957519531, 0.0825384979248047, 0.08247183990478515]",tokens/s,3107.7357459835416,kWh,9.516499386893379e-07,5.214546148738008e-07,4.1613655513126345e-06,5.634470104875773e-06,tokens/kWh,45434618.55951124,MB,1847.382016,3395.813376,0.0,2749.366272,2628.768768,s,10,45.41338720703124,4.541338720703125,0.018165998348059254,4.543606689453124,4.563778857421875,4.565016137695313,4.566005961914063,"[4.51978564453125, 4.5209658203125, 4.52827197265625, 4.56350390625, 4.56018798828125, 4.5468212890625, 4.56625341796875, 4.551375, 4.54039208984375, 4.515830078125]",tokens/s,13.872561346898575,kWh,5.326320814291636e-05,2.919147736644736e-05,0.00022564251273608083,0.0003080971982454445,tokens/kWh,204480.92471717735,,s,629,46.07871485900881,0.07325709834500603,0.009850363871964385,0.07204454040527344,0.07273390045166016,0.07303168182373046,0.1544689245605469,"[0.07243673706054687, 0.0722165756225586, 0.07173836517333984, 0.07148953247070312, 0.07192166137695312, 0.07207936096191406, 0.07273779296875, 0.0723599395751953, 0.07181619262695313, 0.07186022186279296, 0.07170764923095703, 0.07144550323486328, 0.07165235137939453, 0.07186022186279296, 0.07191142272949219, 0.07151513671875, 0.07160524749755859, 0.07166259002685547, 0.07287091064453124, 0.07176089477539062, 0.0714567642211914, 0.07153663635253907, 0.07149772644042969, 0.07140966033935547, 0.07148134613037109, 0.07159705352783204, 0.07194419097900391, 0.07158477020263672, 0.07170559692382812, 0.07132978820800781, 0.07170662689208984, 0.07147110748291016, 0.071552001953125, 0.07141478729248046, 0.07180697631835938, 0.07170150756835937, 0.07141990661621093, 0.07142912292480469, 0.07150080108642579, 0.07151513671875, 0.07165542602539063, 0.0717096939086914, 0.07222067260742188, 0.07158477020263672, 0.07184486389160157, 0.07146701049804688, 0.07157350158691406, 0.07190630340576172, 0.07302861022949218, 0.07173939514160156, 0.07169023895263672, 0.07154898834228515, 0.07212742614746094, 0.0715120620727539, 0.07172710418701173, 0.07148748779296875, 0.07151001739501953, 0.07134003448486329, 0.07155404663085937, 0.07149158477783203, 0.07145369720458984, 0.07155404663085937, 0.15449702453613282, 0.07144448089599609, 0.07186227416992187, 0.07158477020263672, 0.07216435241699219, 0.07135027313232421, 0.07165235137939453, 0.07149158477783203, 0.07154994964599609, 0.07159705352783204, 0.0714241943359375, 0.07140435028076172, 0.07174861145019532, 0.0725555191040039, 0.07297945404052734, 0.07174553680419922, 0.07139737701416016, 0.07143023681640626, 0.07148944091796874, 0.07140966033935547, 0.0728647689819336, 0.07160934448242187, 0.07147622680664062, 0.07147212982177735, 0.07145471954345703, 0.0713359375, 0.07237222290039062, 0.07177113342285156, 0.07253708648681641, 0.07189708709716797, 0.0723056640625, 0.07202098846435546, 0.07208038330078125, 0.07169229125976563, 0.07199641418457031, 0.07196774291992188, 0.07131443023681641, 0.0715335693359375, 0.07154994964599609, 0.0714803237915039, 0.07152639770507813, 0.07153459167480469, 0.07189094543457031, 0.07155404663085937, 0.07141171264648437, 0.07154176330566406, 0.07424921417236328, 0.07197798156738282, 0.07158377838134766, 0.07196975708007812, 0.07168409729003906, 0.07160524749755859, 0.07153049468994141, 0.07144755554199218, 0.07139532470703125, 0.0712806396484375, 0.07142400360107422, 0.07159398651123047, 0.07154688262939453, 0.07156735992431641, 0.07152333068847656, 0.07149263763427735, 0.07195951843261719, 0.15509400939941406, 0.07232307434082032, 0.07224524688720703, 0.07204454040527344, 0.07226675415039062, 0.07196057891845703, 0.07289667510986328, 0.0720239028930664, 0.07299593353271484, 0.07253084564208985, 0.07200870513916016, 0.07150694274902344, 0.07152639770507813, 0.07155404663085937, 0.07140863800048829, 0.07150182342529297, 0.0714229736328125, 0.07174569702148438, 0.0724744644165039, 0.07129190063476562, 0.07246540832519531, 0.0721612777709961, 0.07162163543701172, 0.07225856018066407, 0.07228518676757813, 0.0722165756225586, 0.07262515258789062, 0.07154278564453125, 0.07134719848632813, 0.07153254699707032, 0.07247666931152344, 0.07277875518798828, 0.07165337371826172, 0.07139942169189453, 0.07149874877929688, 0.0718071060180664, 0.0725533447265625, 0.07195648193359375, 0.07154790496826172, 0.07206604766845703, 0.07182854461669921, 0.071468994140625, 0.07151821136474609, 0.07143424224853516, 0.07143014526367188, 0.07146099090576172, 0.07209868621826172, 0.07252684783935547, 0.07136255645751953, 0.07155404663085937, 0.07135948944091797, 0.07153971099853515, 0.07153254699707032, 0.07149568176269532, 0.07168428802490234, 0.07149037170410157, 0.07146086120605469, 0.07137894439697266, 0.07127654266357422, 0.07141375732421874, 0.07227699279785156, 0.07252582550048828, 0.07163085174560548, 0.15379866027832031, 0.07165235137939453, 0.07263651275634765, 0.07239673614501953, 0.07263549041748046, 0.07231270599365235, 0.07247277069091797, 0.07208735656738281, 0.07200460815429688, 0.07240704345703125, 0.07304908752441407, 0.07273369598388672, 0.07322624206542969, 0.07284633636474609, 0.07248076629638672, 0.07288422393798828, 0.07182540893554687, 0.07239679718017578, 0.07256473541259766, 0.07270195007324219, 0.07253094482421875, 0.07234559631347656, 0.07266508483886719, 0.07261901092529296, 0.0730072021484375, 0.07291177368164062, 0.07205683135986328, 0.07246233367919921, 0.07198822021484375, 0.07254937744140624, 0.072521728515625, 0.07264256286621094, 0.07221965026855469, 0.07235788726806641, 0.07210495758056641, 0.07256781005859375, 0.07240499114990234, 0.073059326171875, 0.0725524444580078, 0.072195068359375, 0.0725196762084961, 0.07145471954345703, 0.07244390106201172, 0.07228006744384766, 0.07420928192138672, 0.072700927734375, 0.0726456298828125, 0.07191961669921874, 0.07188992309570312, 0.0720148468017578, 0.07125299072265626, 0.0723394546508789, 0.07194419097900391, 0.07213772583007813, 0.07230156707763671, 0.0723947525024414, 0.0721295394897461, 0.07237222290039062, 0.07256371307373047, 0.07262515258789062, 0.07254937744140624, 0.07255859375, 0.07262413024902344, 0.1548278045654297, 0.072353759765625, 0.0725002212524414, 0.07245209503173829, 0.07242854309082031, 0.07192063903808593, 0.07217766571044922, 0.07246953582763672, 0.07247049713134765, 0.0727531509399414, 0.0725770263671875, 0.07248998260498046, 0.07175382232666015, 0.07248995208740235, 0.07272748565673828, 0.07259442901611328, 0.07262310028076172, 0.07351808166503906, 0.07339826965332032, 0.07235794830322266, 0.07229228973388672, 0.07209574127197266, 0.07278079986572265, 0.0727224349975586, 0.07274598693847656, 0.07179264068603515, 0.07222579193115235, 0.07242342376708985, 0.07207017517089843, 0.0719615707397461, 0.07255654144287109, 0.07277670288085937, 0.07267021179199219, 0.07186227416992187, 0.07147929382324218, 0.07170457458496093, 0.07202098846435546, 0.07212032318115234, 0.07165548706054688, 0.07158470153808594, 0.07155404663085937, 0.07349555206298829, 0.07213772583007813, 0.07260262298583985, 0.07264870452880859, 0.07200153350830078, 0.07254528045654297, 0.071657470703125, 0.07168000030517578, 0.07270604705810547, 0.07262003326416015, 0.07252070617675781, 0.07246652984619141, 0.07231581115722656, 0.0721295394897461, 0.07235379028320313, 0.07264669036865234, 0.0744959716796875, 0.07266918182373047, 0.07270710754394531, 0.0722820816040039, 0.07173632049560547, 0.07150592041015626, 0.15439666748046876, 0.07259750366210938, 0.0715857925415039, 0.07145779418945312, 0.07145267486572265, 0.07161650848388672, 0.072521728515625, 0.07253606414794922, 0.0724295654296875, 0.07247872161865235, 0.07240396881103515, 0.07220531463623046, 0.07265996551513672, 0.07269376373291016, 0.07269171142578125, 0.07250125122070313, 0.07236608123779296, 0.07237222290039062, 0.07195340728759765, 0.07279923248291016, 0.07360511779785156, 0.07294566345214844, 0.07247769927978516, 0.07262515258789062, 0.07273471832275391, 0.07258316802978515, 0.07195452880859375, 0.0722000961303711, 0.07212854766845703, 0.07230051422119141, 0.07157759857177734, 0.07145779418945312, 0.0715448989868164, 0.071901123046875, 0.07262310028076172, 0.07254425811767579, 0.07270809936523437, 0.07179366302490234, 0.0733564453125, 0.07144534301757813, 0.07159603118896485, 0.0715857925415039, 0.0716390380859375, 0.07154994964599609, 0.07138201904296874, 0.07153868865966796, 0.07152639770507813, 0.07149362945556641, 0.07145577239990235, 0.0716349105834961, 0.07142707061767578, 0.07146189117431641, 0.07142707061767578, 0.07152230072021484, 0.07156735992431641, 0.07205580902099609, 0.0724869155883789, 0.07231078338623047, 0.07246233367919921, 0.07301737976074218, 0.07265891265869141, 0.07283404541015626, 0.07256678771972656, 0.15526809692382812, 0.07210291290283204, 0.07220326232910156, 0.07217664337158203, 0.07258521270751953, 0.07277362823486327, 0.07267737579345702, 0.07268045043945312, 0.07258627319335938, 0.07265583801269532, 0.07280332946777343, 0.07242342376708985, 0.0724245147705078, 0.07229945373535156, 0.07221759796142578, 0.07265382385253906, 0.07256678771972656, 0.0719617919921875, 0.07145760345458985, 0.07243059539794922, 0.07212134552001953, 0.07256781005859375, 0.07259961700439453, 0.07362655639648437, 0.07254118347167969, 0.07281970977783203, 0.07263948822021485, 0.07236198425292968, 0.0738150405883789, 0.07245311737060547, 0.07157247924804687, 0.07246131134033203, 0.0724726104736328, 0.07284937286376954, 0.07257804870605469, 0.07242444610595702, 0.07258726501464843, 0.07241436767578124, 0.07216009521484375, 0.07275212860107422, 0.07261695861816406, 0.07253094482421875, 0.07270502471923829, 0.07222271728515625, 0.07260467529296875, 0.07262310028076172, 0.07250534057617188, 0.0722001953125, 0.07243775939941406, 0.07218994903564453, 0.07228006744384766, 0.07340134429931641, 0.07345664215087891, 0.072163330078125, 0.0721233901977539, 0.07228928375244141, 0.07184178924560547, 0.07190630340576172, 0.07204761505126953, 0.07256269073486328, 0.07192269134521484, 0.07259750366210938, 0.07253606414794922, 0.155936767578125, 0.07303372955322265, 0.07266304016113281, 0.07238143920898438, 0.07211436462402344, 0.07191020965576173, 0.0720711669921875, 0.07248486328125, 0.07192988586425782, 0.07246336364746093, 0.07251964569091797, 0.07230873870849609, 0.0725791015625, 0.07243260955810547, 0.07255763244628906, 0.07227590179443359, 0.07335628509521484, 0.07273471832275391, 0.07248281860351563, 0.07255142211914062, 0.07281868743896484, 0.07529267120361328, 0.07326207733154297, 0.07283302307128907, 0.0724869155883789, 0.07157862091064453, 0.07152639770507813, 0.07212850952148438, 0.07244290924072265, 0.07218582153320313, 0.07301427459716797, 0.07195954895019531, 0.07162982177734376, 0.0713338851928711, 0.0723927001953125, 0.0724510726928711, 0.07208857727050781, 0.07149158477783203, 0.07150694274902344, 0.07147007751464844, 0.07175475311279297, 0.07147929382324218, 0.07141273498535156, 0.07240191650390625, 0.07315865325927734, 0.07248281860351563, 0.07143218994140625, 0.07166361236572266, 0.07237344360351562, 0.07268537902832031, 0.07203228759765624, 0.07159190368652343, 0.07179468536376953, 0.07157453155517578, 0.07146701049804688, 0.07154994964599609, 0.07143833923339844, 0.0720005111694336, 0.0722422103881836, 0.07253823852539062, 0.07229526519775391, 0.07265484619140625, 0.07151923370361328, 0.15542373657226563, 0.07168102264404297, 0.07164620971679687, 0.07139328002929687, 0.07188992309570312, 0.07269683074951172, 0.07254835510253907, 0.0723773422241211, 0.07155609893798828, 0.07221043395996093, 0.071583740234375, 0.07193702697753906, 0.07257292938232422, 0.07212134552001953, 0.07240396881103515, 0.07164518737792969, 0.07148851013183594, 0.07235382080078125, 0.07236399841308594, 0.07167692565917969, 0.07185100555419922, 0.0719288330078125, 0.07146803283691407, 0.07169843292236328, 0.07143730926513672, 0.07199641418457031, 0.07272351837158203, 0.07263839721679688, 0.07242447662353516, 0.07246947479248046, 0.0725002212524414, 0.0726087646484375, 0.07243673706054687, 0.0725382080078125, 0.07221135711669922, 0.07237840270996093, 0.07187760162353515, 0.07260671997070313, 0.07360921478271484, 0.07258624267578125, 0.07270604705810547, 0.07254425811767579, 0.07218892669677734, 0.07190220642089844, 0.07209779357910157, 0.07169741058349609, 0.07152333068847656, 0.07145894622802734, 0.07228096008300781, 0.07201996612548828, 0.07152435302734375, 0.07199436950683594, 0.07146086120605469, 0.07159606170654297, 0.07153250885009765, 0.07242854309082031, 0.07224524688720703, 0.07124889373779297, 0.07148457336425781, 0.07174845123291015, 0.07160323333740234, 0.07141065979003906, 0.07256371307373047, 0.15466802978515626, 0.07162790679931641, 0.07138700866699219, 0.07146701049804688, 0.0714229736328125, 0.0716605453491211, 0.07148851013183594, 0.0715888671875, 0.07149772644042969, 0.07135142517089844, 0.07130406188964844, 0.07140351867675782, 0.07143955230712891, 0.07148115539550781, 0.07158477020263672, 0.07155302429199219, 0.07171071624755859, 0.07169945526123046, 0.07139737701416016, 0.07129917144775391, 0.07149353790283203, 0.07145062255859375, 0.07140966033935547, 0.07134925079345703, 0.07143526458740235, 0.07143321228027344, 0.07270912170410156, 0.07148236846923828, 0.07134003448486329, 0.07151513671875, 0.0714803237915039, 0.07152333068847656, 0.07151923370361328, 0.07155097961425781, 0.07148339080810547, 0.07168000030517578, 0.07140659332275391, 0.07138508605957031, 0.07145267486572265, 0.07127040100097656, 0.07239167785644532, 0.07333273315429688, 0.0716072998046875, 0.07177728271484375, 0.07187251281738281, 0.07153254699707032, 0.07148544311523437, 0.07157247924804687, 0.07153667449951172, 0.07155606079101562, 0.0714772491455078, 0.07148544311523437, 0.0713534698486328, 0.07227993774414063, 0.0716072998046875, 0.07318630218505859, 0.0728053741455078, 0.07240191650390625, 0.07223500823974609, 0.07155404663085937, 0.07140351867675782, 0.07203433227539062, 0.07255753326416016]",tokens/s,13.650554316122056,, +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1469, in _autoset_attn_implementation + cls._check_and_enable_flash_attn_2( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 + raise ValueError( +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpw763jynu/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -1171,7 +1201,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c146b-1921abc36b54584b416d8bda;526c75dc-4801-4b7b-9be2-50f7e25504ff) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4680-65267c5367b98e643d8c8494;206a362f-168f-40b5-92ef-51ee8675e54a) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1228,7 +1258,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpqrzjvape/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpqeg205we/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -1270,7 +1300,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c11ff-43cfdf0305bda68722e4af53;200d4c7f-914c-427b-ba06-d70f1d3007d0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d445d-15dadcca02c9d3d93fd58b9e;96533f63-bfaa-44aa-8104-cb08bcdca3db) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1341,7 +1371,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12d2-06361aa902b8a91934ea6c0b;0ba24da1-2822-418f-b288-be325bc53183) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44fc-63793d957b62d183091577a7;27027c9b-5851-4e1c-877d-21edce4c43b4) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1373,7 +1403,111 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5020.81536,8173.125632,0.0,7526.678528,6833.613824,s,10,3.326392669677735,0.3326392669677734,0.0030324363165483534,0.33195497131347657,0.3341614532470703,0.3376804428100586,0.3404956344604492,"[0.3411994323730469, 0.33337945556640625, 0.33058718872070314, 0.3303564453125, 0.33089251708984374, 0.33060626220703127, 0.3314384460449219, 0.332591064453125, 0.332870361328125, 0.33247149658203123]",tokens/s,769.6024655585886,kWh,3.897490269607968e-06,2.1353897317655327e-06,2.0114881683290285e-05,2.6147761684663786e-05,tokens/kWh,9790512.973435482,MB,5020.81536,8173.125632,0.0,7526.678528,7110.583808,s,10,183.47133203125,18.347133203125,0.004664256892410071,18.349021484375,18.351409570312498,18.35145087890625,18.35148392578125,"[18.348591796875, 18.3397421875, 18.34241796875, 18.338681640625, 18.3502734375, 18.348642578125, 18.349400390625, 18.350689453125, 18.3514921875, 18.351400390625]",tokens/s,3.4337789616782985,kWh,0.00021652181481321653,0.0001186722658167949,0.0010773519962895067,0.0014125460769195183,tokens/kWh,44600.31501230066,,s,629,186.15873925781258,0.29595983983753976,0.03945461951219622,0.29114471435546874,0.29161677246093753,0.29189117431640627,0.6223596313476563,"[0.29115188598632813, 0.2917990417480469, 0.2910863342285156, 0.29130035400390625, 0.2909051818847656, 0.2912214050292969, 0.2909573059082031, 0.29103411865234374, 0.2910054321289062, 0.2911651916503906, 0.2917939147949219, 0.2913239135742188, 0.2910597839355469, 0.2918880615234375, 0.29148263549804687, 0.29144268798828127, 0.2912911376953125, 0.29127783203125, 0.2912788391113281, 0.29161370849609375, 0.291304443359375, 0.29117236328125, 0.2911733703613281, 0.29094400024414063, 0.29102081298828125, 0.2911559753417969, 0.2910597229003906, 0.2914805908203125, 0.2913699951171875, 0.29076275634765625, 0.29096661376953126, 0.29093161010742186, 0.2912389221191406, 0.29129931640625, 0.2909480895996094, 0.2912174072265625, 0.2912204895019531, 0.29107199096679687, 0.2910310363769531, 0.29094091796875, 0.29076480102539065, 0.29079766845703126, 0.29086407470703124, 0.29078524780273435, 0.2911088562011719, 0.29134234619140625, 0.2910627746582031, 0.29169970703125, 0.29144064331054687, 0.2908927917480469, 0.2910648193359375, 0.2910535583496094, 0.2909266052246094, 0.29100030517578124, 0.2910525512695312, 0.2911098937988281, 0.2910382080078125, 0.29060504150390626, 0.2909347839355469, 0.2907238464355469, 0.29095013427734373, 0.29088467407226565, 0.6249952392578125, 0.29085287475585936, 0.2911539306640625, 0.290904052734375, 0.2911037292480469, 0.2912143249511719, 0.2912419738769531, 0.291493896484375, 0.2911508483886719, 0.291056640625, 0.2912942504882812, 0.2909091491699219, 0.2910955505371094, 0.2910320739746094, 0.2910310363769531, 0.29128805541992187, 0.2911457214355469, 0.29099114990234376, 0.291136474609375, 0.291462158203125, 0.29125222778320314, 0.2911016845703125, 0.290882568359375, 0.290977783203125, 0.2910105590820313, 0.2909910888671875, 0.29104946899414064, 0.2909276123046875, 0.2908713073730469, 0.29091635131835936, 0.2908846130371094, 0.29086822509765625, 0.29094400024414063, 0.2912911376953125, 0.291083251953125, 0.2909296569824219, 0.2908917846679687, 0.2909429626464844, 0.29090509033203127, 0.29101669311523437, 0.2909276123046875, 0.2912174072265625, 0.29108837890625, 0.29093682861328124, 0.29119793701171875, 0.29107916259765626, 0.2934343566894531, 0.29095834350585936, 0.2907484130859375, 0.29136895751953124, 0.29149798583984377, 0.29114471435546874, 0.29111294555664063, 0.29156658935546875, 0.29093792724609374, 0.2908681640625, 0.2908518371582031, 0.2906265563964844, 0.2906163330078125, 0.2909542541503906, 0.29111505126953124, 0.2914887390136719, 0.29113345336914065, 0.6221834106445312, 0.291056640625, 0.2913434143066406, 0.29181332397460935, 0.2915881042480469, 0.2911866760253906, 0.2914805908203125, 0.29122662353515627, 0.29128805541992187, 0.29116824340820313, 0.2914232177734375, 0.2919321594238281, 0.29145498657226565, 0.2910115966796875, 0.29119488525390624, 0.2909060974121094, 0.29112115478515627, 0.29129421997070315, 0.29111294555664063, 0.29093887329101564, 0.2918696899414063, 0.2910146484375, 0.2910586853027344, 0.2929541015625, 0.29112115478515627, 0.29093887329101564, 0.29098599243164064, 0.2910157775878906, 0.2910462951660156, 0.291080322265625, 0.29077490234375, 0.2908006286621094, 0.2908897399902344, 0.29091943359375, 0.290830322265625, 0.29089688110351564, 0.2908344421386719, 0.29086004638671875, 0.29178369140625, 0.29098904418945315, 0.29101568603515626, 0.2907648315429687, 0.29071978759765627, 0.29085177612304686, 0.2908037109375, 0.2908999633789062, 0.2911928405761719, 0.29247796630859374, 0.29141607666015623, 0.2910228576660156, 0.2908846130371094, 0.2909020080566406, 0.2911375427246094, 0.290845703125, 0.2914642028808594, 0.29115493774414064, 0.2909491271972656, 0.290914306640625, 0.2911805419921875, 0.29089483642578123, 0.2907873229980469, 0.2908323974609375, 0.290597900390625, 0.6217267456054687, 0.29077197265625, 0.2910330810546875, 0.2909470825195313, 0.29107510375976564, 0.29102178955078123, 0.29123788452148436, 0.29106893920898436, 0.29189324951171874, 0.2911856689453125, 0.2912204895019531, 0.29157272338867185, 0.29196185302734373, 0.2910382080078125, 0.2912010498046875, 0.29094088745117186, 0.29163418579101563, 0.290946044921875, 0.29115493774414064, 0.2909962158203125, 0.2909952087402344, 0.2908518371582031, 0.29090310668945313, 0.29097158813476565, 0.29139556884765627, 0.2911918029785156, 0.29115493774414064, 0.2910177307128906, 0.291470458984375, 0.2910995178222656, 0.2909153137207031, 0.29085287475585936, 0.29129421997070315, 0.2910177307128906, 0.29091738891601565, 0.290850830078125, 0.2908016662597656, 0.2909204406738281, 0.290703369140625, 0.2909480895996094, 0.290808837890625, 0.2907821960449219, 0.2912767944335938, 0.29103521728515624, 0.2930513000488281, 0.2909010009765625, 0.2909573059082031, 0.2909429626464844, 0.29101364135742186, 0.2907362060546875, 0.29087738037109373, 0.29123077392578123, 0.2909122009277344, 0.2908518371582031, 0.29091021728515626, 0.29078021240234375, 0.291042236328125, 0.29075250244140627, 0.2906828918457031, 0.29115399169921874, 0.29102072143554686, 0.291009521484375, 0.29132696533203123, 0.6224281616210937, 0.29084161376953127, 0.29141094970703124, 0.29116824340820313, 0.29133004760742187, 0.2911754150390625, 0.2917089233398438, 0.2911651916503906, 0.29145394897460936, 0.29128704833984376, 0.29129010009765627, 0.29106585693359377, 0.2920130615234375, 0.2914622497558594, 0.2913115234375, 0.291346435546875, 0.291209228515625, 0.292105224609375, 0.29109759521484374, 0.2908907470703125, 0.29111090087890623, 0.2913320922851563, 0.291009521484375, 0.2909900817871094, 0.2909204406738281, 0.29106893920898436, 0.29144268798828127, 0.2911283264160156, 0.2910525512695312, 0.2916495361328125, 0.2910535583496094, 0.2909675598144531, 0.291346435546875, 0.2914672546386719, 0.2912665710449219, 0.29141607666015623, 0.29099212646484374, 0.29119384765625, 0.2911212463378906, 0.29125518798828126, 0.29113345336914065, 0.2911068115234375, 0.2909552612304688, 0.2910054321289062, 0.29141812133789063, 0.29120306396484374, 0.29097164916992185, 0.29154507446289063, 0.29104742431640623, 0.29088870239257814, 0.2910064697265625, 0.29087435913085935, 0.2910402526855469, 0.2909849548339844, 0.29111199951171873, 0.29143136596679686, 0.291462158203125, 0.2940467224121094, 0.29119488525390624, 0.2914396057128906, 0.2912563171386719, 0.2915348510742187, 0.291083251953125, 0.622940185546875, 0.2911180725097656, 0.291894287109375, 0.2918707885742188, 0.29123678588867186, 0.291156982421875, 0.2911457214355469, 0.2911201171875, 0.29126556396484377, 0.2914580078125, 0.2914068603515625, 0.2910320739746094, 0.29122354125976563, 0.2909378662109375, 0.29136383056640625, 0.29118463134765626, 0.2914672546386719, 0.2917652587890625, 0.2911856689453125, 0.29107098388671876, 0.29113446044921876, 0.29076788330078124, 0.29069720458984377, 0.29090304565429687, 0.29082623291015625, 0.29081088256835935, 0.291056640625, 0.29096347045898435, 0.29087435913085935, 0.291009521484375, 0.2907371520996094, 0.2910607299804687, 0.29096038818359377, 0.2913167419433594, 0.2914723815917969, 0.2914847717285156, 0.2912306213378906, 0.291462158203125, 0.2936002502441406, 0.2915155029296875, 0.2913965148925781, 0.29102386474609376, 0.2910607299804687, 0.2909644775390625, 0.29098904418945315, 0.29122970581054686, 0.29106997680664065, 0.29096649169921873, 0.2909337463378906, 0.2908917846679687, 0.2909982604980469, 0.2911559753417969, 0.29184307861328124, 0.2915625, 0.29157272338867185, 0.29153689575195313, 0.29130859375, 0.29116513061523436, 0.2910146484375, 0.2910955505371094, 0.2916024169921875, 0.291262451171875, 0.2909020080566406, 0.6232821655273437, 0.2913167419433594, 0.29129421997070315, 0.2912972717285156, 0.2918635559082031, 0.29147442626953124, 0.2911283264160156, 0.2916812744140625, 0.2915973205566406, 0.2912419738769531, 0.2914723815917969, 0.291281982421875, 0.29155218505859376, 0.29142630004882814, 0.29158706665039064, 0.2915564270019531, 0.2916771240234375, 0.29130136108398436, 0.2912850646972656, 0.2912767333984375, 0.2914969482421875, 0.2918369140625, 0.2913320922851563, 0.29271755981445313, 0.29146624755859374, 0.2911907958984375, 0.2909296569824219, 0.2910955505371094, 0.291083251953125, 0.2912593994140625, 0.2912706604003906, 0.29119589233398435, 0.29104638671875, 0.29097268676757815, 0.29096038818359377, 0.29111294555664063, 0.2913320922851563, 0.291304443359375, 0.29161575317382815, 0.2912389221191406, 0.29126040649414064, 0.2912554626464844, 0.2910042419433594, 0.2909931640625, 0.2908856201171875, 0.29083340454101564, 0.2908917846679687, 0.290946044921875, 0.2909757385253906, 0.29078323364257813, 0.29109762573242187, 0.29080374145507815, 0.29088555908203123, 0.29080780029296877, 0.2909491271972656, 0.2907904052734375, 0.2916864013671875, 0.2912890930175781, 0.2907801513671875, 0.29119692993164065, 0.29122457885742187, 0.29137411499023436, 0.291031005859375, 0.6241597290039063, 0.29123788452148436, 0.29148980712890626, 0.29125018310546874, 0.29131057739257815, 0.29118463134765626, 0.29127166748046873, 0.291272705078125, 0.2918543395996094, 0.2915051574707031, 0.291631103515625, 0.29105560302734373, 0.2923868103027344, 0.2912337951660156, 0.2911866760253906, 0.291061767578125, 0.292490234375, 0.2912214965820312, 0.291267578125, 0.2910013427734375, 0.29129421997070315, 0.2910822448730469, 0.29093582153320313, 0.29120001220703123, 0.29091943359375, 0.2908856201171875, 0.29145703125, 0.2914949035644531, 0.29112933349609377, 0.29123788452148436, 0.2912767944335938, 0.2912133178710937, 0.29131365966796874, 0.29124609375, 0.2912194519042969, 0.29173348999023435, 0.2911201171875, 0.2912788391113281, 0.2910044250488281, 0.2910013427734375, 0.29118771362304685, 0.2910330810546875, 0.2915604553222656, 0.29146316528320315, 0.2914334716796875, 0.29083953857421874, 0.29135769653320315, 0.2911631469726563, 0.29084979248046877, 0.2908221435546875, 0.29091122436523437, 0.29104229736328124, 0.2908825988769531, 0.29070947265625, 0.29156146240234376, 0.292242431640625, 0.29155123901367186, 0.29115493774414064, 0.29105670166015624, 0.2912132568359375, 0.29169766235351563, 0.2913536071777344, 0.291162109375, 0.6245776977539063, 0.2910738830566406, 0.29133721923828126, 0.29104638671875, 0.29137100219726564, 0.2918984375, 0.291302490234375, 0.29108514404296876, 0.2911313781738281, 0.29162203979492185, 0.2911015319824219, 0.29115188598632813, 0.29129421997070315, 0.2920244140625, 0.29145590209960937, 0.29112423706054685, 0.291620849609375, 0.29239004516601563, 0.29183676147460935, 0.29133517456054686, 0.29116928100585937, 0.2914068603515625, 0.29107916259765626, 0.2920724487304687, 0.2911928405761719, 0.29108837890625, 0.2911365051269531, 0.29100747680664063, 0.2909665832519531, 0.29107705688476565, 0.2915195007324219, 0.29198641967773437, 0.2912174072265625, 0.2914068603515625, 0.29100747680664063, 0.29123480224609377, 0.29113037109375, 0.29115188598632813, 0.2909010009765625, 0.2909767761230469, 0.29136282348632814, 0.2908538818359375, 0.29098599243164064, 0.29118463134765626, 0.2911078491210938, 0.29112628173828126, 0.29127783203125, 0.2916915283203125, 0.2914672546386719, 0.29112628173828126, 0.2914969482421875, 0.29150311279296875, 0.29141299438476564, 0.2911457214355469, 0.2911907958984375, 0.2912061462402344, 0.29095834350585936, 0.29087335205078124, 0.2910064697265625, 0.2916515808105469, 0.29124505615234375, 0.2910699462890625, 0.2909389038085938, 0.6244362182617188, 0.2915553283691406, 0.29160562133789064, 0.29136474609375, 0.2912962646484375, 0.2911641540527344, 0.29157272338867185, 0.2914303894042969, 0.29142630004882814, 0.291240966796875, 0.29158303833007815, 0.291295166015625, 0.2915502014160156, 0.291373046875, 0.2917918701171875, 0.29154507446289063, 0.2914375610351562, 0.2918666687011719, 0.291716064453125, 0.29142630004882814, 0.2913177490234375, 0.29200588989257814, 0.29093582153320313, 0.2909337463378906, 0.29117440795898436, 0.29120620727539065, 0.29325714111328127, 0.2910044250488281, 0.29144064331054687, 0.291175537109375, 0.29100634765625, 0.29097576904296873, 0.29103927612304686, 0.29099923706054687, 0.29096038818359377, 0.2911754150390625, 0.2910146484375, 0.29110272216796873, 0.2911170654296875, 0.2914334716796875, 0.29108428955078125, 0.29090304565429687, 0.2909296569824219, 0.290904052734375, 0.2912860107421875, 0.2911866760253906, 0.29128500366210935, 0.2911088562011719, 0.29134951782226565, 0.29139764404296875, 0.29150411987304686, 0.291240966796875, 0.29109146118164064, 0.29109658813476563, 0.2911150207519531, 0.2918696899414063, 0.29107916259765626, 0.29097369384765626, 0.29076480102539065, 0.2908671875, 0.2908917846679687, 0.29114776611328125, 0.2908713073730469]",tokens/s,3.3788368061995393,, +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1469, in _autoset_attn_implementation + cls._check_and_enable_flash_attn_2( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 + raise ValueError( +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpr265qkfc/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4706.717696,21498.42944,0.0,20851.982336,20236.60032,s,10,15.640249145507811,1.5640249145507812,0.00168203754809238,1.563786376953125,1.564931494140625,1.5666584350585937,1.5680399877929687,"[1.5630980224609374, 1.5644747314453125, 1.5625794677734375, 1.56453466796875, 1.5645272216796875, 1.5645477294921875, 1.5624769287109375, 1.5628692626953125, 1.5683853759765625, 1.5627557373046874]",tokens/s,163.68025702041214,kWh,1.8462728808323543e-05,1.0117648402811028e-05,9.129585081439906e-05,0.00011987622802553362,tokens/kWh,2135535.9958896274,MB,4706.717696,21498.42944,0.0,20851.982336,20339.706368,s,10,879.7075468749999,87.97075468749999,0.002247932535373388,87.97129296874999,87.973384375,87.973496875,87.973586875,"[87.9672265625, 87.968984375, 87.9679609375, 87.9719375, 87.972140625, 87.973609375, 87.973359375, 87.968734375, 87.9729453125, 87.9706484375]",tokens/s,0.7161470902892214,kWh,0.0010385013203819595,0.0005691906291757731,0.005147801034904198,0.0067554929844619305,tokens/kWh,9325.744271351337,,s,629,892.387736694337,1.4187404398956058,0.1859067280964402,1.39627001953125,1.3965248535156252,1.3967189697265625,2.9604817578125,"[1.39608984375, 1.3963448486328125, 1.3963612060546875, 1.3961072998046875, 1.3964923095703126, 1.3963653564453125, 1.396220947265625, 1.3965025634765624, 1.3960928955078125, 1.3962034912109376, 1.3961072998046875, 1.396042724609375, 1.396294677734375, 1.3963448486328125, 1.3963970947265625, 1.3963560791015626, 1.39631201171875, 1.396305908203125, 1.3963643798828125, 1.3958726806640624, 1.3961134033203124, 1.3959710693359375, 1.3963284912109375, 1.396166748046875, 1.3961839599609376, 1.3961236572265625, 1.3963602294921875, 1.3963294677734375, 1.3962445068359375, 1.396177001953125, 1.3961810302734374, 1.3960733642578125, 1.3959208984375, 1.396359130859375, 1.3958072509765624, 1.39637451171875, 1.396179931640625, 1.396137939453125, 1.39983154296875, 1.396505615234375, 1.3963970947265625, 1.3963028564453126, 1.396326416015625, 1.3964442138671875, 1.3961175537109376, 1.396173828125, 1.396177978515625, 1.3965987548828125, 1.3961728515625, 1.3964912109375, 1.3962935791015625, 1.39636328125, 1.39614208984375, 1.3962025146484376, 1.396126708984375, 1.3964083251953125, 1.3963018798828124, 1.396030517578125, 1.3959925537109374, 1.39601513671875, 1.395998779296875, 1.39626904296875, 2.96171923828125, 1.396209716796875, 1.3962506103515624, 1.3962557373046875, 1.3960858154296876, 1.3961513671875, 1.39597412109375, 1.39620654296875, 1.396126708984375, 1.396136962890625, 1.3962086181640625, 1.396041748046875, 1.39642578125, 1.396341796875, 1.3963878173828126, 1.396432861328125, 1.3963028564453126, 1.395948486328125, 1.3960714111328125, 1.3977159423828125, 1.395883056640625, 1.3964984130859375, 1.3963028564453126, 1.3964410400390626, 1.3959066162109375, 1.3964769287109375, 1.396209716796875, 1.3961768798828125, 1.39627001953125, 1.396104248046875, 1.396179931640625, 1.396157470703125, 1.39639404296875, 1.3961943359375, 1.39627001953125, 1.3962998046875, 1.39653125, 1.3978797607421876, 1.39601513671875, 1.39614111328125, 1.3961656494140624, 1.3962823486328124, 1.396073486328125, 1.3962823486328124, 1.3959659423828126, 1.3963233642578126, 1.3965946044921875, 1.3960704345703125, 1.396041748046875, 1.3960120849609374, 1.396105224609375, 1.3962178955078124, 1.39605810546875, 1.396042724609375, 1.3964912109375, 1.3962711181640626, 1.396250732421875, 1.396602783203125, 1.3962762451171875, 1.396516845703125, 1.396326416015625, 1.398640625, 1.3963304443359374, 2.96060302734375, 1.39647900390625, 1.39653125, 1.3961614990234374, 1.3960181884765626, 1.3962342529296874, 1.39601611328125, 1.3958338623046875, 1.39621484375, 1.3961512451171876, 1.3961553955078125, 1.3965394287109374, 1.39652294921875, 1.3962301025390624, 1.396216796875, 1.39635400390625, 1.3961707763671876, 1.396279296875, 1.39607861328125, 1.3962249755859375, 1.395962890625, 1.3963018798828124, 1.396126708984375, 1.3988648681640625, 1.3960888671875, 1.3964031982421874, 1.3962301025390624, 1.3964400634765626, 1.3959556884765625, 1.3960714111328125, 1.3961871337890626, 1.396095947265625, 1.3961861572265626, 1.3961195068359376, 1.3960130615234374, 1.395988525390625, 1.3964349365234374, 1.396515869140625, 1.3959833984375, 1.396220947265625, 1.3962230224609375, 1.396178955078125, 1.396315185546875, 1.3963509521484374, 1.3963929443359375, 1.3960755615234375, 1.3961195068359376, 1.3990062255859375, 1.396305908203125, 1.3959075927734375, 1.3962926025390625, 1.3961728515625, 1.3962230224609375, 1.39626806640625, 1.3965035400390624, 1.3963131103515625, 1.3960294189453124, 1.3960611572265624, 1.396111328125, 1.396221923828125, 1.39627734375, 1.3961727294921875, 1.3960447998046874, 2.961617919921875, 1.3961595458984375, 1.39620458984375, 1.39593212890625, 1.3961533203125, 1.3960263671875, 1.3961451416015624, 1.3962352294921876, 1.3962076416015625, 1.3960560302734375, 1.3962998046875, 1.3964493408203125, 1.39588720703125, 1.396454345703125, 1.3963284912109375, 1.39639501953125, 1.3963642578125, 1.3962127685546875, 1.396188232421875, 1.396221923828125, 1.396116455078125, 1.3964298095703125, 1.3964974365234375, 1.39631201171875, 1.3961666259765626, 1.396533203125, 1.3962506103515624, 1.396453369140625, 1.3960457763671874, 1.396447265625, 1.396515869140625, 1.39631005859375, 1.396305908203125, 1.3964073486328126, 1.396190185546875, 1.3960980224609374, 1.39657421875, 1.396528076171875, 1.3965926513671876, 1.4005042724609376, 1.396791259765625, 1.3964676513671874, 1.3963724365234376, 1.3963826904296874, 1.3961553955078125, 1.396410400390625, 1.39633251953125, 1.3959403076171875, 1.396326416015625, 1.3962608642578125, 1.3963551025390626, 1.3964451904296875, 1.396463623046875, 1.39631103515625, 1.39641552734375, 1.3962301025390624, 1.396401123046875, 1.3960284423828124, 1.3962137451171874, 1.3964359130859374, 1.3964512939453124, 1.3963233642578126, 1.39616357421875, 2.96169482421875, 1.39649951171875, 1.39685986328125, 1.396106201171875, 1.3962987060546874, 1.396274169921875, 1.3959752197265625, 1.3962496337890624, 1.3964339599609374, 1.39608984375, 1.3961011962890626, 1.3962977294921874, 1.3963929443359375, 1.39652197265625, 1.3963203125, 1.3962506103515624, 1.396452392578125, 1.3964031982421874, 1.3964390869140626, 1.396611083984375, 1.3960335693359376, 1.3959803466796874, 1.395914794921875, 1.3961964111328125, 1.3962803955078125, 1.3961553955078125, 1.3965179443359375, 1.39654248046875, 1.3962127685546875, 1.3964461669921875, 1.396337646484375, 1.3962884521484376, 1.3965230712890624, 1.3961533203125, 1.39621484375, 1.3961461181640624, 1.3964666748046874, 1.396431884765625, 1.3962578125, 1.396168701171875, 1.3964727783203126, 1.4002913818359375, 1.396400146484375, 1.3963284912109375, 1.3961728515625, 1.396716552734375, 1.396369384765625, 1.3964410400390626, 1.3962874755859376, 1.3961717529296875, 1.39602734375, 1.396189208984375, 1.39620458984375, 1.3962147216796874, 1.3963602294921875, 1.396046875, 1.39650048828125, 1.3963643798828125, 1.396154296875, 1.396696044921875, 1.396347900390625, 1.3963560791015626, 1.3960509033203126, 2.96171630859375, 1.3963919677734375, 1.3963355712890626, 1.39618505859375, 1.3962926025390625, 1.396115478515625, 1.3960120849609374, 1.396125732421875, 1.396279296875, 1.3963253173828125, 1.3961246337890625, 1.3961092529296875, 1.3965772705078126, 1.396653076171875, 1.3962025146484376, 1.395925048828125, 1.39633154296875, 1.396427734375, 1.3964052734375, 1.3962967529296875, 1.39624755859375, 1.396263916015625, 1.3963182373046874, 1.396494384765625, 1.3964830322265624, 1.3963775634765625, 1.396209716796875, 1.396411376953125, 1.3961912841796875, 1.39610107421875, 1.39620556640625, 1.39635205078125, 1.3962015380859376, 1.396042724609375, 1.3961400146484375, 1.396209716796875, 1.3963079833984375, 1.3961953125, 1.40067431640625, 1.3964073486328126, 1.396336669921875, 1.3963826904296874, 1.396295654296875, 1.39614208984375, 1.3964093017578125, 1.396348876953125, 1.39626806640625, 1.39626806640625, 1.3963868408203124, 1.3961544189453126, 1.3960079345703125, 1.39607861328125, 1.396130859375, 1.3962720947265626, 1.396485107421875, 1.39627001953125, 1.3966878662109374, 1.396177978515625, 1.3960714111328125, 1.39645751953125, 1.3967821044921875, 1.3963929443359375, 1.396221923828125, 2.9624052734375, 1.39640625, 1.396473876953125, 1.396378662109375, 1.3961103515625, 1.3961666259765626, 1.3961380615234376, 1.3961973876953124, 1.3964195556640624, 1.3963243408203125, 1.39621484375, 1.39620556640625, 1.3961348876953126, 1.3965609130859375, 1.3961318359375, 1.396334716796875, 1.396403076171875, 1.3967677001953125, 1.3963458251953125, 1.39630078125, 1.396005859375, 1.396190185546875, 1.396221923828125, 1.3991322021484376, 1.396274169921875, 1.3965567626953126, 1.3959332275390626, 1.39675439453125, 1.3960386962890625, 1.396147216796875, 1.3964830322265624, 1.3962813720703124, 1.3961563720703125, 1.3962669677734374, 1.3960653076171874, 1.3961922607421875, 1.3961605224609375, 1.3963775634765625, 1.396552734375, 1.39624853515625, 1.396236328125, 1.396547607421875, 1.396465576171875, 1.3962352294921876, 1.3962802734375, 1.3960560302734375, 1.3962137451171874, 1.39620556640625, 1.396074462890625, 1.396125732421875, 1.396263916015625, 1.3962332763671874, 1.396306884765625, 1.396263916015625, 1.39645849609375, 1.396369384765625, 1.3965506591796875, 1.396136962890625, 1.395937255859375, 1.3961185302734376, 1.3962281494140625, 1.3962301025390624, 1.400385498046875, 2.960169921875, 1.3960386962890625, 1.3961124267578124, 1.3962587890625, 1.3961328125, 1.3968507080078125, 1.396115478515625, 1.39623828125, 1.3961759033203125, 1.396401123046875, 1.39621484375, 1.3962935791015625, 1.39642578125, 1.3964779052734375, 1.3963294677734375, 1.3960775146484374, 1.397832763671875, 1.396600830078125, 1.39633154296875, 1.3963253173828125, 1.3960980224609374, 1.3963857421875, 1.3962301025390624, 1.39618408203125, 1.396242431640625, 1.3963233642578126, 1.3963560791015626, 1.3964493408203125, 1.39633154296875, 1.396285400390625, 1.3962772216796875, 1.3962454833984375, 1.3963140869140624, 1.396157470703125, 1.39630078125, 1.3960755615234375, 1.3961595458984375, 1.39652392578125, 1.396315185546875, 1.3964451904296875, 1.3963427734375, 1.3962332763671874, 1.39616357421875, 1.3960242919921875, 1.3964410400390626, 1.396295654296875, 1.396368408203125, 1.396148193359375, 1.3961932373046875, 1.396262939453125, 1.3962987060546874, 1.3963458251953125, 1.396189208984375, 1.3962659912109374, 1.3962874755859376, 1.3962987060546874, 1.3963837890625, 1.3963499755859374, 1.3961451416015624, 1.396453369140625, 1.3965670166015625, 1.3963551025390626, 1.3961656494140624, 2.9656298828125, 1.3966131591796875, 1.396486083984375, 1.396496337890625, 1.3961390380859375, 1.3962127685546875, 1.396168701171875, 1.39631005859375, 1.3962650146484374, 1.39608984375, 1.396369384765625, 1.396193359375, 1.3961533203125, 1.396337646484375, 1.396303955078125, 1.3962352294921876, 1.3963612060546875, 1.3983529052734376, 1.39636328125, 1.396357177734375, 1.396263916015625, 1.3963140869140624, 1.3959822998046876, 1.396474853515625, 1.396347900390625, 1.396595703125, 1.396111328125, 1.3961328125, 1.3962701416015626, 1.3961102294921874, 1.3961707763671876, 1.39612158203125, 1.3963203125, 1.3962547607421876, 1.396042724609375, 1.396252685546875, 1.3960447998046874, 1.396010986328125, 1.3960858154296876, 1.3964871826171874, 1.396463623046875, 1.396033447265625, 1.3962823486328124, 1.3962874755859376, 1.3961461181640624, 1.396464599609375, 1.3964410400390626, 1.396326416015625, 1.396304931640625, 1.396130859375, 1.3960826416015626, 1.396231201171875, 1.39595166015625, 1.3962086181640625, 1.39643505859375, 1.3962188720703126, 1.39656396484375, 1.396241455078125, 1.396109375, 1.396360107421875, 1.3962454833984375, 1.396515869140625, 1.4013460693359374, 2.9601669921875, 1.396242431640625, 1.3961563720703125, 1.3967093505859376, 1.3961451416015624, 1.3960601806640625, 1.3962076416015625, 1.396126708984375, 1.396243408203125, 1.3964246826171876, 1.3964666748046874, 1.39620654296875, 1.39607958984375, 1.39641552734375, 1.396093994140625, 1.3960858154296876, 1.3963212890625, 1.396093994140625, 1.3965445556640625, 1.39627001953125, 1.3963775634765625, 1.396485107421875, 1.3963131103515625, 1.3964759521484376, 1.3963275146484375, 1.3965045166015626, 1.39883935546875, 1.3962230224609375, 1.3962762451171875, 1.396242431640625, 1.3961400146484375, 1.3962178955078124, 1.3964820556640625, 1.396148193359375, 1.3961688232421876, 1.3961829833984376, 1.396322265625, 1.396190185546875, 1.396115478515625, 1.39614208984375, 1.3964031982421874, 1.39636328125, 1.3966151123046875, 1.3964134521484375, 1.3963355712890626, 1.396177978515625, 1.396125732421875, 1.396169677734375, 1.396275146484375, 1.3965240478515626, 1.3966817626953125, 1.3962578125, 1.396337646484375, 1.3967205810546874, 1.3967247314453124, 1.396284423828125, 1.396473876953125, 1.3963509521484374, 1.3962127685546875, 1.3961492919921874, 1.3961031494140625, 1.3963387451171876, 1.39633251953125]",tokens/s,0.7048505645427167,, +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status + response.raise_for_status() + File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status + raise HTTPError(http_error_msg, response=self) +requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error + metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata + r = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper + response = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper + hf_raise_for_status(response) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status + raise HfHubHTTPError(message, response=response) from e +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4abe-1f54ed505cd0f6a56cec5e76;fcf4516b-68aa-4605-8dde-0907dc1144ee) + +403 Forbidden: Authorization error.. +Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. +If you are trying to create or update content,make sure you have a token with the `write` role. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 399, in cached_file + resolved_file = hf_hub_download( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download + return _hf_hub_download_to_cache_dir( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir + _raise_on_head_call_error(head_call_error, force_download, local_files_only) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error + raise LocalEntryNotFoundError( +huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ + super().__init__(config) + File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ + self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) + File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config + return AutoConfig.from_pretrained(model, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict + resolved_config_file = cached_file( + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 442, in cached_file + raise EnvironmentError( +OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. +Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -1413,7 +1547,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1304-2ef9196801f4d23919af6acf;63d24f79-a284-4137-8d23-4c425810461b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4523-08ec7cb107f8bc9c377cbc5a;1f896f62-c957-4351-bac3-50b90fa0ccf7) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1475,7 +1609,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee6b-7589b0110a93a9862f5a3b2c;ac830afb-1163-413c-af6f-1671d7f2ae7e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3f7b-570bb65e3a48d0675d6f19e1;86290f93-4f7e-4e8a-9950-3b8c075dc546) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -1602,7 +1736,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1299-30e791073b1797ee17fe6c22;129e367a-f16b-4a04-924d-4c5733b50f44) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44d4-1d35f724772845b63fbc70af;bba85b6e-bcd4-438b-82bc-7ca9ae2ad034) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1675,7 +1809,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664beef8-3875a034421aa84f2a89b243;1a130932-b1e9-465f-98c9-bd9e5f54172c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d3fef-70abec277ed0b0fd0aecde6e;708390d0-eaef-4368-912d-92771a1ce2d8) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1734,7 +1868,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpjit8g2pu/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpk1s74iku/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -1776,7 +1910,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c141c-58531a3535f41784515548d2;a587fcdb-8daf-4f90-9759-ee3fb3bb4f1f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4633-44e7eda25cd46d70467e35e0;c4416f25-ed48-4a56-9f83-569e67477d5c) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1806,6 +1940,35 @@ Traceback (most recent call last): OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1469, in _autoset_attn_implementation + cls._check_and_enable_flash_attn_2( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 + raise ValueError( +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpnr7aenwu/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new + ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5946.503168,7053.246464,0.0,6406.79936,6098.60864,s,10,4.370441192626953,0.4370441192626952,0.0009281353036040856,0.43696878051757815,0.4381058013916016,0.4383167251586914,0.4384854641723633,"[0.4385276489257813, 0.4359717712402344, 0.43771954345703123, 0.43744598388671874, 0.4380589294433594, 0.4362575378417969, 0.43786093139648435, 0.4364915771484375, 0.436230712890625, 0.43587655639648437]",tokens/s,585.7532196792366,kWh,5.148596204997261e-06,2.8211948946868096e-06,2.4996191494520076e-05,3.296598259420415e-05,tokens/kWh,7765580.754902423,MB,5946.503168,7055.343616,0.0,6408.896512,6157.952,s,10,244.94424023437497,24.494424023437496,0.005272769015065994,24.493744140625,24.502037109375,24.502107421875,24.502163671875,"[24.487064453125, 24.502177734375, 24.49926953125, 24.4968515625, 24.502021484375, 24.493955078125, 24.493533203125, 24.4922109375, 24.487486328125, 24.489669921875]",tokens/s,2.5720139383444343,kWh,0.00028916327015789236,0.00015848618370445323,0.00135038218658508,0.0017980316404474258,tokens/kWh,35038.315557296286,,s,629,248.48397949218753,0.3950460723246224,0.05195367916433446,0.388748291015625,0.38915277099609374,0.38942371215820315,0.8254186791992187,"[0.38863052368164064, 0.3885834350585938, 0.38854656982421876, 0.38854452514648435, 0.3885762634277344, 0.3887677307128906, 0.3885189208984375, 0.3887728576660156, 0.3885434875488281, 0.38884658813476564, 0.38842572021484373, 0.3884328918457031, 0.3885711364746094, 0.3883417663574219, 0.3885240173339844, 0.38941387939453126, 0.3884031982421875, 0.3887196044921875, 0.38854452514648435, 0.3887032470703125, 0.3886264343261719, 0.38861721801757815, 0.3888977966308594, 0.3890206604003906, 0.388917236328125, 0.3884503173828125, 0.388843505859375, 0.3885404052734375, 0.38881793212890625, 0.38851788330078124, 0.3882977294921875, 0.3884666748046875, 0.388701171875, 0.38850970458984374, 0.3885455322265625, 0.3883397216796875, 0.3884103698730469, 0.388495361328125, 0.3894988708496094, 0.3882915954589844, 0.3888158569335938, 0.3886602783203125, 0.3886519775390625, 0.3887790222167969, 0.3884134521484375, 0.38843392944335936, 0.3885281372070313, 0.3888619384765625, 0.38980096435546874, 0.38914764404296875, 0.3887196044921875, 0.38860186767578125, 0.38848306274414063, 0.38838375854492185, 0.3885035400390625, 0.3884605407714844, 0.38845645141601565, 0.38830081176757814, 0.3884513244628906, 0.3889899597167969, 0.3887943725585937, 0.3889407958984375, 0.8262952880859376, 0.38884658813476564, 0.38921319580078123, 0.388748291015625, 0.3887923278808594, 0.38922137451171873, 0.3888977966308594, 0.3885906066894531, 0.3892326354980469, 0.38878106689453124, 0.38923776245117186, 0.38898483276367185, 0.3888721923828125, 0.38877694702148435, 0.38888754272460935, 0.3888670654296875, 0.38904830932617185, 0.3887943725585937, 0.38909030151367185, 0.38900018310546874, 0.3890636901855469, 0.3890831298828125, 0.3892695007324219, 0.38899404907226565, 0.39072256469726563, 0.3887984619140625, 0.3887677307128906, 0.3885977478027344, 0.38868377685546873, 0.38868069458007815, 0.38871551513671876, 0.3887615966796875, 0.38870834350585937, 0.38858139038085937, 0.388885498046875, 0.3888230285644531, 0.3896135559082031, 0.39009588623046876, 0.3886540832519531, 0.38893875122070315, 0.3886039123535156, 0.3890165710449219, 0.3886899108886719, 0.38870529174804686, 0.38863665771484374, 0.3886868591308594, 0.38889166259765623, 0.3890595703125, 0.38893771362304685, 0.3889479675292969, 0.388917236328125, 0.38900222778320315, 0.38909951782226565, 0.38838067626953127, 0.388495361328125, 0.3884666748046875, 0.389074951171875, 0.38862539672851565, 0.38855474853515626, 0.3885578308105469, 0.388843505859375, 0.389080078125, 0.3892039794921875, 0.8267151489257812, 0.3888455810546875, 0.389074951171875, 0.38913229370117186, 0.38876263427734375, 0.3887595520019531, 0.3888875732421875, 0.38918243408203124, 0.3892049865722656, 0.38914047241210936, 0.3887319030761719, 0.38899713134765623, 0.38865509033203127, 0.38901556396484377, 0.3886766052246094, 0.38860800170898435, 0.388864013671875, 0.38877593994140625, 0.3888302001953125, 0.388790283203125, 0.3885578308105469, 0.38857318115234374, 0.3885926513671875, 0.3890616455078125, 0.3892623291015625, 0.3890933837890625, 0.38865817260742186, 0.3889070129394531, 0.38874929809570313, 0.38892340087890626, 0.3887185974121094, 0.3888066711425781, 0.38876568603515627, 0.3885875244140625, 0.38916607666015623, 0.38873394775390624, 0.3885383605957031, 0.38869403076171877, 0.38857217407226563, 0.38951425170898435, 0.38901556396484377, 0.38918759155273436, 0.3889551391601562, 0.3887442016601563, 0.3890032653808594, 0.38889675903320314, 0.3889725341796875, 0.3889837951660156, 0.38876568603515627, 0.38967501831054685, 0.3893606262207031, 0.38946200561523436, 0.3886407775878906, 0.38881381225585937, 0.38867251586914064, 0.38852505493164063, 0.38857217407226563, 0.38876568603515627, 0.38846771240234373, 0.38870529174804686, 0.3890585632324219, 0.38904730224609374, 0.38865716552734375, 0.8268502807617187, 0.3887595520019531, 0.3887820739746094, 0.38875750732421877, 0.3886417846679687, 0.38880767822265627, 0.38890087890625, 0.38849432373046877, 0.38883224487304685, 0.3887718505859375, 0.38920294189453125, 0.3890390930175781, 0.3896995849609375, 0.38909747314453125, 0.3887442016601563, 0.388569091796875, 0.3890595703125, 0.3884718017578125, 0.38869195556640623, 0.38880563354492187, 0.38886605834960936, 0.38868377685546873, 0.3896105041503906, 0.38871450805664065, 0.3889029235839844, 0.38878311157226564, 0.3886878662109375, 0.38854757690429687, 0.38863873291015627, 0.3887370300292969, 0.38861004638671875, 0.38876263427734375, 0.3888015441894531, 0.3884984436035156, 0.38875546264648436, 0.38851174926757814, 0.388959228515625, 0.38882406616210935, 0.3891138610839844, 0.38870733642578126, 0.38867251586914064, 0.38911282348632814, 0.38889166259765623, 0.3884820556640625, 0.38871450805664065, 0.3886540832519531, 0.388600830078125, 0.389032958984375, 0.389180419921875, 0.38953677368164064, 0.38902886962890626, 0.3890667419433594, 0.389001220703125, 0.3887820739746094, 0.38870016479492187, 0.3886346130371094, 0.3887544250488281, 0.38863873291015627, 0.38883328247070315, 0.3885629577636719, 0.3887093811035156, 0.38889675903320314, 0.38913946533203125, 0.8280023193359375, 0.38870834350585937, 0.38864794921875, 0.3887124328613281, 0.388748291015625, 0.3888793640136719, 0.3885383605957031, 0.38862744140625, 0.38867352294921875, 0.3893258361816406, 0.3888793640136719, 0.38871551513671876, 0.38890802001953123, 0.38877490234375, 0.38909747314453125, 0.3890729064941406, 0.3887790222167969, 0.38876876831054685, 0.3889479675292969, 0.38905447387695313, 0.388927490234375, 0.38944769287109376, 0.3890114440917969, 0.3895459899902344, 0.3894302673339844, 0.3890831298828125, 0.38988287353515627, 0.388738037109375, 0.3887667236328125, 0.38885888671875, 0.3889950866699219, 0.3887370300292969, 0.3887564697265625, 0.38875546264648436, 0.3886991271972656, 0.389185546875, 0.38876058959960935, 0.38907086181640627, 0.38890188598632813, 0.3888619384765625, 0.38933401489257813, 0.388853759765625, 0.38896026611328127, 0.38860595703125, 0.38862744140625, 0.38865817260742186, 0.388463623046875, 0.3885875244140625, 0.3883212890625, 0.3891998596191406, 0.3895500793457031, 0.3895326843261719, 0.389032958984375, 0.3888066711425781, 0.388822021484375, 0.388811767578125, 0.38866329956054685, 0.38884146118164065, 0.38888958740234375, 0.38867147827148435, 0.38877490234375, 0.3891199951171875, 0.38873599243164064, 0.8254505004882813, 0.38886605834960936, 0.3887943725585937, 0.3888025512695312, 0.38903604125976565, 0.38877694702148435, 0.38867864990234374, 0.3889029235839844, 0.3888517150878906, 0.38892645263671877, 0.38902374267578127, 0.38975079345703123, 0.389823486328125, 0.389465087890625, 0.3888455810546875, 0.3890862121582031, 0.38899200439453124, 0.388611083984375, 0.3885342712402344, 0.3887175598144531, 0.38892340087890626, 0.38893875122070315, 0.3891138610839844, 0.38856704711914064, 0.3884707946777344, 0.3899740295410156, 0.388389892578125, 0.388780029296875, 0.3886213073730469, 0.3885793151855469, 0.3890093994140625, 0.38884454345703123, 0.3886991271972656, 0.388748291015625, 0.3889151916503906, 0.38854757690429687, 0.3891988525390625, 0.3890667419433594, 0.388822021484375, 0.38893670654296875, 0.3886960754394531, 0.38903399658203125, 0.38867251586914064, 0.38844415283203126, 0.3887595520019531, 0.3886796875, 0.3885424499511719, 0.3885875244140625, 0.3885660095214844, 0.388421630859375, 0.38844927978515625, 0.3885557861328125, 0.3888015441894531, 0.3884236755371094, 0.38832846069335936, 0.388347900390625, 0.388389892578125, 0.38839910888671875, 0.38838580322265626, 0.38835302734375, 0.3885363159179688, 0.38859982299804685, 0.38915277099609374, 0.8270069580078125, 0.388790283203125, 0.38852301025390623, 0.38860800170898435, 0.38870529174804686, 0.3886397399902344, 0.3888097229003906, 0.38888140869140625, 0.38900531005859373, 0.3891435546875, 0.38884454345703123, 0.38879129028320314, 0.38904934692382814, 0.38878924560546874, 0.38913946533203125, 0.38866329956054685, 0.3886970825195312, 0.3890196533203125, 0.38881076049804686, 0.38890087890625, 0.3889612731933594, 0.38866021728515626, 0.3888097229003906, 0.38903604125976565, 0.38920907592773435, 0.389138427734375, 0.38885888671875, 0.38871142578125, 0.38858648681640623, 0.3885793151855469, 0.38836428833007813, 0.38872164916992186, 0.3887953796386719, 0.3888721923828125, 0.389064697265625, 0.3893247985839844, 0.3885936584472656, 0.38854757690429687, 0.3885936584472656, 0.3884789733886719, 0.38885479736328127, 0.38876058959960935, 0.38854656982421876, 0.3885404052734375, 0.38906982421875, 0.38860287475585936, 0.38849945068359376, 0.38863665771484374, 0.38878106689453124, 0.3889336242675781, 0.38914559936523435, 0.38887322998046875, 0.3893893127441406, 0.3887462463378906, 0.38847796630859377, 0.38849127197265626, 0.38849740600585936, 0.3886684265136719, 0.3886510009765625, 0.38863565063476563, 0.38843597412109376, 0.3889039306640625, 0.38858853149414063, 0.8253368530273437, 0.38870220947265627, 0.38867864990234374, 0.3885823974609375, 0.3886929931640625, 0.3884881896972656, 0.38854656982421876, 0.38842059326171874, 0.38854757690429687, 0.38858648681640623, 0.388569091796875, 0.3887452087402344, 0.3891435546875, 0.38859982299804685, 0.3888302001953125, 0.38854144287109377, 0.3885936584472656, 0.38863052368164064, 0.38876980590820315, 0.3885404052734375, 0.38905447387695313, 0.38915277099609374, 0.38901556396484377, 0.38856704711914064, 0.3885404052734375, 0.38846157836914064, 0.38859469604492186, 0.38851071166992185, 0.38895001220703124, 0.38856805419921875, 0.38871450805664065, 0.38837246704101563, 0.3885987854003906, 0.38849127197265626, 0.38861618041992185, 0.38849639892578125, 0.38856192016601565, 0.38879641723632813, 0.38922341918945313, 0.38905752563476564, 0.38980096435546874, 0.3889090576171875, 0.3889070129394531, 0.3885916137695313, 0.38894183349609374, 0.3886796875, 0.3886407775878906, 0.3885383605957031, 0.38921829223632814, 0.38887322998046875, 0.38906265258789063, 0.3887585144042969, 0.38897869873046875, 0.38905548095703124, 0.38911077880859374, 0.3886458740234375, 0.3885455322265625, 0.3886878662109375, 0.3890585632324219, 0.38870220947265627, 0.38884658813476564, 0.38864486694335937, 0.3889561462402344, 0.82583349609375, 0.3893094482421875, 0.38874725341796873, 0.38864697265625, 0.3890738525390625, 0.3888230285644531, 0.3885772705078125, 0.3886264343261719, 0.38845541381835935, 0.38850457763671875, 0.38880767822265627, 0.3886039123535156, 0.3885486145019531, 0.388485107421875, 0.3883673706054688, 0.3887646789550781, 0.3885353088378906, 0.388632568359375, 0.38862539672851565, 0.3885936584472656, 0.38859674072265626, 0.3886069641113281, 0.388485107421875, 0.38892440795898436, 0.38891622924804686, 0.38909133911132815, 0.388453369140625, 0.38876058959960935, 0.3886592102050781, 0.3885567932128906, 0.3885189208984375, 0.3885557861328125, 0.388716552734375, 0.3882700805664063, 0.38850457763671875, 0.3887790222167969, 0.38870733642578126, 0.3885281372070313, 0.38864382934570313, 0.3885977478027344, 0.38863565063476563, 0.3885660095214844, 0.38846463012695315, 0.38862234497070314, 0.3884236755371094, 0.3884984436035156, 0.388347900390625, 0.38830181884765624, 0.38828134155273436, 0.3885066223144531, 0.38897152709960936, 0.38908721923828127, 0.3887431640625, 0.3887175598144531, 0.3884800109863281, 0.38863665771484374, 0.39151205444335935, 0.3889510498046875, 0.38859674072265626, 0.38852301025390623, 0.38847796630859377, 0.3888824462890625, 0.38877593994140625, 0.824754150390625, 0.38834994506835935, 0.38857217407226563, 0.38853323364257814, 0.3888844909667969, 0.3888005065917969, 0.3891640319824219, 0.3885987854003906, 0.3888486328125, 0.38872781372070314, 0.38896435546875, 0.3892623291015625, 0.3890882568359375, 0.3891937255859375, 0.38884658813476564, 0.3889827880859375, 0.38853323364257814, 0.3885260925292969, 0.38840933227539065, 0.38836428833007813, 0.388558837890625, 0.38827621459960937, 0.3885209655761719, 0.3886376953125, 0.38875341796875, 0.3887790222167969, 0.38885888671875, 0.3887298583984375, 0.38882406616210935, 0.38858035278320313, 0.388927490234375, 0.3888988037109375, 0.38889166259765623, 0.38841650390625, 0.3884298095703125, 0.3886039123535156, 0.3884247131347656, 0.3889837951660156, 0.38898175048828126, 0.3884728393554687, 0.3891671142578125, 0.3885987854003906, 0.3886120910644531, 0.388411376953125, 0.38858139038085937, 0.3884390258789063, 0.3885076599121094, 0.388558837890625, 0.388738037109375, 0.38887322998046875, 0.38831716918945314, 0.3885014953613281, 0.3886868591308594, 0.38870220947265627, 0.38861721801757815, 0.38846258544921874, 0.3886673889160156, 0.3884789733886719, 0.38866943359375, 0.388358154296875, 0.3887667236328125, 0.38856500244140624, 0.388558837890625]",tokens/s,2.5313503159658475,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -1847,7 +2010,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c10c4-1c1616491cb5c83a054f9972;8e9cbe16-fb9b-4abc-8b9f-e93cc169fc3c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4315-3f713b21242d0c1b42e4fe2f;ba8805c1-6291-4e42-adb6-c4ae0ad56bf8) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1878,6 +2041,7 @@ OSError: s is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2253.213696,3257.401344,0.0,2610.95424,2376.894976,s,10,1.2870804443359376,0.12870804443359377,0.00015157803773503835,0.12866516876220702,0.12899027557373047,0.1289954734802246,0.12899963180541993,"[0.12900067138671875, 0.12865913391113282, 0.12858761596679688, 0.12898912048339845, 0.12858274841308595, 0.1286431427001953, 0.12867120361328124, 0.128544189453125, 0.12870460510253906, 0.12869801330566405]",tokens/s,1988.997666203233,kWh,1.5193139385973286e-06,8.325113373967513e-07,7.587381426025696e-06,9.939206702019776e-06,tokens/kWh,25756582.761074632,MB,2253.213696,3257.401344,0.0,2610.95424,2525.750784,s,10,70.93060400390624,7.093060400390625,0.0012182744404161136,7.0930024414062505,7.094170507812501,7.094838916015625,7.095373642578125,"[7.09550732421875, 7.09132666015625, 7.09347216796875, 7.09222509765625, 7.09402197265625, 7.09253271484375, 7.0937412109375, 7.09241015625, 7.0938251953125, 7.09154150390625]",tokens/s,8.881920700482192,kWh,8.37330595117349e-05,4.5889142433513294e-05,0.0004168421247839744,0.0005464643267292225,tokens/kWh,115286.57392345577,,s,629,71.97587043762205,0.11442904680067104,0.015304620913384064,0.112574462890625,0.11269017791748047,0.1127872543334961,0.2413812921142578,"[0.11270246124267579, 0.11247513580322266, 0.11254988861083984, 0.1130250244140625, 0.11253555297851563, 0.1124659194946289, 0.11245568084716796, 0.11251609802246093, 0.11251097869873047, 0.11258163452148437, 0.11251097869873047, 0.11258470153808593, 0.11255500793457031, 0.11256934356689453, 0.11261746978759765, 0.11248435211181641, 0.11289702606201171, 0.11260825347900391, 0.11277926635742187, 0.11253043365478516, 0.11260928344726563, 0.11247206115722656, 0.11275059509277344, 0.11259801483154297, 0.11260108947753907, 0.11253247833251953, 0.11261849975585937, 0.11252735900878906, 0.11259494018554687, 0.11255398559570312, 0.11260928344726563, 0.11257344055175782, 0.11260723114013672, 0.11260825347900391, 0.11258777618408203, 0.11255705261230468, 0.11253350067138672, 0.11254476928710938, 0.1126266860961914, 0.11260211181640625, 0.11261746978759765, 0.11256422424316406, 0.11285606384277344, 0.11264921569824218, 0.11261030578613282, 0.11261746978759765, 0.1126277084350586, 0.11262258911132812, 0.11260415649414063, 0.11261337280273437, 0.1127710723876953, 0.1126277084350586, 0.11271577453613281, 0.11270861053466796, 0.11272191619873047, 0.11274956512451172, 0.11267686462402343, 0.11262464141845703, 0.11263385772705078, 0.11267481231689454, 0.11275059509277344, 0.1126645736694336, 0.24137318420410156, 0.1124853744506836, 0.11256320190429688, 0.11262566375732422, 0.11250688171386719, 0.11255091094970703, 0.11251097869873047, 0.11253964996337891, 0.11249152374267578, 0.11245977783203125, 0.11250790405273438, 0.11248230743408204, 0.11246387481689453, 0.1124352035522461, 0.1125406723022461, 0.1124485092163086, 0.11249254608154297, 0.1125191650390625, 0.11247923278808594, 0.11253453063964844, 0.11251097869873047, 0.11247309112548828, 0.11252838134765625, 0.1125580825805664, 0.11245875549316406, 0.11259085083007812, 0.11255091094970703, 0.11260006713867188, 0.11251200103759766, 0.11249868774414062, 0.11258367919921874, 0.11248025512695313, 0.11253759765625, 0.11275981140136719, 0.11252735900878906, 0.1126328353881836, 0.11258573150634765, 0.11260518646240235, 0.11257855987548829, 0.11252735900878906, 0.11254271697998047, 0.11252531433105468, 0.11264409637451171, 0.11267686462402343, 0.11256832122802735, 0.11244338989257813, 0.11257855987548829, 0.11266764831542969, 0.11267788696289062, 0.11254169464111329, 0.11269017791748047, 0.11268915557861328, 0.1125959701538086, 0.11256934356689453, 0.11263897705078126, 0.11264717102050781, 0.11258367919921874, 0.11256934356689453, 0.11263590240478516, 0.11262566375732422, 0.11256832122802735, 0.11255398559570312, 0.11261952209472656, 0.241438720703125, 0.11255398559570312, 0.11255091094970703, 0.11259801483154297, 0.11265331268310547, 0.11257855987548829, 0.11258163452148437, 0.11253043365478516, 0.11255296325683593, 0.112505859375, 0.11257138824462891, 0.1125406723022461, 0.11249766540527344, 0.1125181427001953, 0.11253862762451172, 0.11249459075927734, 0.11246284484863281, 0.11260723114013672, 0.11257344055175782, 0.11261235046386718, 0.11254476928710938, 0.11284480285644531, 0.11285606384277344, 0.11266355133056641, 0.1125038070678711, 0.11255193328857421, 0.11254783630371094, 0.11258163452148437, 0.11256114959716797, 0.1125437469482422, 0.11258879852294922, 0.11253453063964844, 0.112574462890625, 0.11258675384521484, 0.11257138824462891, 0.1126266860961914, 0.11258060455322266, 0.11259187316894531, 0.1126277084350586, 0.11262156677246093, 0.1126277084350586, 0.11259903717041016, 0.1125560302734375, 0.11265638732910156, 0.11254681396484376, 0.11247923278808594, 0.11261952209472656, 0.11259494018554687, 0.11258879852294922, 0.11258470153808593, 0.11267174530029297, 0.11263897705078126, 0.11259903717041016, 0.11261542510986328, 0.11264205169677734, 0.112648193359375, 0.11261542510986328, 0.11270758056640626, 0.11261849975585937, 0.11263999938964844, 0.1125928955078125, 0.11255705261230468, 0.11270041656494141, 0.241728515625, 0.11265331268310547, 0.11293695831298828, 0.11258060455322266, 0.11248435211181641, 0.1124485092163086, 0.11253043365478516, 0.11249971008300781, 0.11272191619873047, 0.11246182250976562, 0.1124853744506836, 0.11250688171386719, 0.11245772552490234, 0.11244441223144531, 0.11255705261230468, 0.11249254608154297, 0.11248025512695313, 0.11250278472900391, 0.11248844909667968, 0.11263999938964844, 0.11251097869873047, 0.11251507568359374, 0.11251200103759766, 0.11259699249267578, 0.11254579162597657, 0.11251507568359374, 0.11254579162597657, 0.11257651519775391, 0.11252428436279296, 0.11254579162597657, 0.11264102172851563, 0.11248844909667968, 0.11276595306396485, 0.11260928344726563, 0.11258367919921874, 0.11260006713867188, 0.1125212173461914, 0.1125038070678711, 0.1125580825805664, 0.11256320190429688, 0.1125928955078125, 0.1126277084350586, 0.11257548522949219, 0.11259391784667969, 0.11258470153808593, 0.11247821044921875, 0.11275161743164062, 0.11263078308105469, 0.11257855987548829, 0.11255091094970703, 0.11261542510986328, 0.1126451187133789, 0.11260108947753907, 0.11263180541992188, 0.11260825347900391, 0.11262156677246093, 0.11258879852294922, 0.11263692474365235, 0.11257344055175782, 0.11260518646240235, 0.11258675384521484, 0.11256012725830078, 0.11266764831542969, 0.24138444519042968, 0.11254169464111329, 0.11258163452148437, 0.11254886627197265, 0.11248230743408204, 0.11253759765625, 0.11253350067138672, 0.11252838134765625, 0.11254681396484376, 0.1125038070678711, 0.11258470153808593, 0.11254271697998047, 0.11252428436279296, 0.11249766540527344, 0.11261746978759765, 0.11248947143554687, 0.11247513580322266, 0.11254169464111329, 0.11253862762451172, 0.11260006713867188, 0.11256626892089844, 0.11257548522949219, 0.11262361907958984, 0.11265638732910156, 0.11253350067138672, 0.11247923278808594, 0.11264614105224609, 0.11268300628662109, 0.11261849975585937, 0.11249766540527344, 0.11260313415527344, 0.11253453063964844, 0.11256012725830078, 0.11267584228515624, 0.11257651519775391, 0.11323187255859375, 0.11262566375732422, 0.11254271697998047, 0.1125898208618164, 0.11256729888916016, 0.11257548522949219, 0.11260108947753907, 0.11257138824462891, 0.11269017791748047, 0.1125898208618164, 0.1124853744506836, 0.11267481231689454, 0.11263078308105469, 0.11253759765625, 0.11252633666992187, 0.11263488006591797, 0.1126666259765625, 0.11264102172851563, 0.11268812561035156, 0.11273830413818359, 0.11272089385986328, 0.11266252899169922, 0.11263180541992188, 0.11258265686035156, 0.11281510162353516, 0.11268812561035156, 0.11254169464111329, 0.11275571441650391, 0.24144178771972657, 0.11251712036132812, 0.1126328353881836, 0.11250994873046875, 0.11244646453857422, 0.11244953918457032, 0.11279257965087891, 0.11258060455322266, 0.11250176239013672, 0.11244953918457032, 0.11255398559570312, 0.11250176239013672, 0.11249971008300781, 0.11255193328857421, 0.11253145599365234, 0.11250790405273438, 0.11251302337646485, 0.11246797180175781, 0.11256832122802735, 0.11259494018554687, 0.11254988861083984, 0.11252428436279296, 0.11260620880126954, 0.11262464141845703, 0.112574462890625, 0.11257651519775391, 0.11250482940673828, 0.11253862762451172, 0.1125038070678711, 0.11250176239013672, 0.11260006713867188, 0.11254988861083984, 0.11248332977294923, 0.1125898208618164, 0.11260211181640625, 0.1126297607421875, 0.11252633666992187, 0.11251507568359374, 0.11256217956542969, 0.11253247833251953, 0.1125560302734375, 0.11251507568359374, 0.11263385772705078, 0.11258675384521484, 0.11258265686035156, 0.11251200103759766, 0.11261746978759765, 0.1127045135498047, 0.11258265686035156, 0.11248435211181641, 0.11258265686035156, 0.11281817626953125, 0.11271372985839843, 0.1126297607421875, 0.11258163452148437, 0.11262156677246093, 0.11254681396484376, 0.11303116607666015, 0.1126983642578125, 0.11258265686035156, 0.1126277084350586, 0.11263590240478516, 0.11267072296142579, 0.24141619873046874, 0.11255091094970703, 0.11249459075927734, 0.11251404571533204, 0.11253247833251953, 0.11248435211181641, 0.11257855987548829, 0.11252223968505859, 0.11253145599365234, 0.11259801483154297, 0.11253657531738281, 0.11254169464111329, 0.11248639678955077, 0.11248435211181641, 0.1125406723022461, 0.112500732421875, 0.11250278472900391, 0.112500732421875, 0.11255398559570312, 0.11258265686035156, 0.11255910491943359, 0.11248947143554687, 0.11267584228515624, 0.11257138824462891, 0.11249971008300781, 0.11251712036132812, 0.11254271697998047, 0.1126277084350586, 0.11255193328857421, 0.11256217956542969, 0.11284172821044922, 0.11293901062011719, 0.11256320190429688, 0.11262873840332031, 0.112574462890625, 0.11262873840332031, 0.11260518646240235, 0.11254988861083984, 0.1126328353881836, 0.1125928955078125, 0.11260415649414063, 0.11257548522949219, 0.1125898208618164, 0.11260108947753907, 0.11265740966796875, 0.11247513580322266, 0.11259903717041016, 0.112611328125, 0.11260518646240235, 0.11256832122802735, 0.11265638732910156, 0.11318374633789062, 0.11264614105224609, 0.11265126037597656, 0.112716796875, 0.11264102172851563, 0.11263692474365235, 0.11265638732910156, 0.11265433502197265, 0.11260108947753907, 0.11261030578613282, 0.11261235046386718, 0.1126983642578125, 0.24122265625, 0.11252019500732421, 0.11248844909667968, 0.11258060455322266, 0.11247821044921875, 0.1124505615234375, 0.11250688171386719, 0.11245568084716796, 0.1127936019897461, 0.11244953918457032, 0.11250790405273438, 0.11251712036132812, 0.11267072296142579, 0.11250482940673828, 0.1125560302734375, 0.11244032287597656, 0.1124485092163086, 0.11246284484863281, 0.11280691528320312, 0.11258675384521484, 0.11251609802246093, 0.11249152374267578, 0.11257036590576172, 0.11254988861083984, 0.11245670318603515, 0.11254783630371094, 0.11252019500732421, 0.11255500793457031, 0.11247001647949219, 0.11252428436279296, 0.11255398559570312, 0.11250688171386719, 0.11250994873046875, 0.11253453063964844, 0.11261952209472656, 0.11287757110595703, 0.11254681396484376, 0.11263692474365235, 0.11282841491699219, 0.11264409637451171, 0.11253759765625, 0.11254476928710938, 0.11256422424316406, 0.11258777618408203, 0.11260006713867188, 0.11250994873046875, 0.11263590240478516, 0.11265433502197265, 0.11255910491943359, 0.11248639678955077, 0.11258163452148437, 0.1128642578125, 0.1126266860961914, 0.11265945434570312, 0.11260415649414063, 0.11265740966796875, 0.1126983642578125, 0.11261235046386718, 0.11258163452148437, 0.11269734191894532, 0.1125959701538086, 0.11257241821289063, 0.1125928955078125, 0.24156057739257814, 0.11255500793457031, 0.11257344055175782, 0.11258265686035156, 0.11252735900878906, 0.11252735900878906, 0.11254783630371094, 0.11252223968505859, 0.11247821044921875, 0.11248332977294923, 0.11258163452148437, 0.11257241821289063, 0.11249459075927734, 0.11252735900878906, 0.11256934356689453, 0.11252735900878906, 0.11250994873046875, 0.11254783630371094, 0.1125437469482422, 0.11265535736083984, 0.11260006713867188, 0.1125212173461914, 0.11262054443359375, 0.11261849975585937, 0.11250994873046875, 0.11262464141845703, 0.11254783630371094, 0.11264614105224609, 0.11257548522949219, 0.11259187316894531, 0.11254783630371094, 0.11248230743408204, 0.11256729888916016, 0.11293798065185547, 0.11260415649414063, 0.11265638732910156, 0.11258060455322266, 0.11265433502197265, 0.1126451187133789, 0.11261337280273437, 0.11255705261230468, 0.11254476928710938, 0.11262566375732422, 0.11269427490234375, 0.1125928955078125, 0.11255910491943359, 0.11262258911132812, 0.11262361907958984, 0.11250892639160157, 0.11256114959716797, 0.11261030578613282, 0.11268096160888671, 0.11266150665283203, 0.11274956512451172, 0.11281715393066406, 0.11264614105224609, 0.11259085083007812, 0.11267584228515624, 0.11262054443359375, 0.11264614105224609, 0.11265126037597656, 0.11263897705078126, 0.11264921569824218, 0.24151962280273437, 0.11256320190429688, 0.11253555297851563, 0.11254476928710938, 0.1124505615234375, 0.11248947143554687, 0.1125038070678711, 0.11250790405273438, 0.11244544219970704, 0.11249356842041015, 0.11254169464111329, 0.11252941131591797, 0.11247206115722656, 0.11247513580322266, 0.11251609802246093, 0.11247309112548828, 0.1128253402709961, 0.11254476928710938, 0.11253145599365234, 0.11254579162597657, 0.1125212173461914, 0.11251302337646485, 0.11254886627197265, 0.11256217956542969, 0.11252941131591797, 0.11249459075927734, 0.11252019500732421, 0.11256114959716797, 0.11249664306640625, 0.11250688171386719, 0.11254681396484376, 0.11246284484863281, 0.11252735900878906, 0.11256729888916016, 0.112611328125, 0.11260415649414063, 0.11252941131591797, 0.11251097869873047, 0.11259699249267578, 0.11252428436279296, 0.11256626892089844, 0.11252531433105468, 0.11255500793457031, 0.11262156677246093, 0.11255705261230468, 0.11246080017089843, 0.11257241821289063, 0.11256934356689453, 0.11253145599365234, 0.1124853744506836, 0.11254169464111329, 0.11267788696289062, 0.11267481231689454, 0.11273113250732422, 0.11271372985839843, 0.11260415649414063, 0.11258060455322266, 0.11259391784667969, 0.11297996520996094, 0.11263078308105469, 0.11258163452148437, 0.11260108947753907, 0.11272294616699219]",tokens/s,8.739039850099807,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -1904,7 +2068,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpzqljfc1j/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp2kjepj_4/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1279.791104,1019.74016,0.0,373.293056,277.410816,s,10,0.3806483840942383,0.03806483840942383,0.0010374051807092283,0.03839825630187988,0.038631001281738284,0.03909522857666015,0.03946661041259766,"[0.03955945587158203, 0.03852783966064453, 0.03837411117553711, 0.038422401428222654, 0.03847126388549805, 0.038292991638183595, 0.03834467315673828, 0.03843852615356445, 0.036084991455078125, 0.03613212966918945]",tokens/s,6725.366787229584,kWh,4.3617786259962035e-07,2.3900503699099743e-07,1.1732389459702955e-06,1.8484218455609134e-06,tokens/kWh,138496523.73174343,MB,1280.086016,1019.74016,0.0,373.293056,323.047424,s,10,23.712255126953128,2.371225512695313,0.003882883813537052,2.3718930664062503,2.376249291992188,2.3762570922851562,2.3762633325195313,"[2.362728515625, 2.37624755859375, 2.373291748046875, 2.367230224609375, 2.371970947265625, 2.371815185546875, 2.376264892578125, 2.37161865234375, 2.37223681640625, 2.3688505859375]",tokens/s,26.56854004931377,kWh,2.7777078536237154e-05,1.5222735153430559e-05,6.781520513363374e-05,0.00011081501882330147,tokens/kWh,568514.9961527847,,s,629,24.01549103546141,0.03818043089898478,0.004509392475924156,0.03761459350585938,0.03786444854736328,0.03806208114624023,0.07505829895019543,"[0.03638784027099609, 0.03659775924682617, 0.036880382537841795, 0.03663974380493164, 0.03670937728881836, 0.03717529678344727, 0.036736000061035154, 0.03655987167358398, 0.03629260635375976, 0.03701657485961914, 0.03663974380493164, 0.0363612174987793, 0.03665100860595703, 0.038844417572021485, 0.03839078521728516, 0.038013950347900394, 0.03768627166748047, 0.03772825622558594, 0.037766143798828124, 0.03749990463256836, 0.037585918426513674, 0.03749683380126953, 0.03755929565429687, 0.03760537719726562, 0.037743614196777346, 0.03752140808105469, 0.0375551986694336, 0.03762995147705078, 0.037495807647705076, 0.03751935958862305, 0.03787571334838867, 0.03770880126953125, 0.0376360969543457, 0.03753472137451172, 0.037599231719970705, 0.03759718322753906, 0.03856595230102539, 0.03801900863647461, 0.037601280212402347, 0.03762073516845703, 0.0376545295715332, 0.03752243041992188, 0.03753881454467774, 0.03760435104370117, 0.03775795364379883, 0.037550079345703126, 0.03786649703979492, 0.03831398391723633, 0.03769036865234375, 0.037601280212402347, 0.03755929565429687, 0.037498878479003905, 0.03746303939819336, 0.03752345657348633, 0.03794432067871094, 0.037713920593261716, 0.03760435104370117, 0.037495807647705076, 0.037591041564941405, 0.03748454284667969, 0.03763916778564453, 0.03788083267211914, 0.07681126403808594, 0.03763711929321289, 0.03831808090209961, 0.03748966217041016, 0.03763507080078125, 0.03758489608764649, 0.03759308624267578, 0.037579776763916016, 0.03772419357299805, 0.03755926513671875, 0.03763097763061524, 0.03747020721435547, 0.03756748962402344, 0.03761663818359375, 0.037901313781738284, 0.03767398452758789, 0.03793817520141601, 0.03772415924072266, 0.037601280212402347, 0.03751321411132812, 0.03760332870483398, 0.03770470428466797, 0.03760435104370117, 0.03767705535888672, 0.03810611343383789, 0.0376995849609375, 0.03758694458007812, 0.03785728073120117, 0.03757878494262695, 0.03765244674682617, 0.03786956787109375, 0.03756032180786133, 0.037838848114013675, 0.037773311614990236, 0.037525505065917966, 0.037884929656982425, 0.03775692749023438, 0.03761151885986328, 0.0378869743347168, 0.03772518539428711, 0.037797889709472655, 0.03774566268920899, 0.037759998321533206, 0.03785113525390625, 0.038040576934814455, 0.03770470428466797, 0.0377077751159668, 0.03755929565429687, 0.037599231719970705, 0.037661697387695314, 0.037553150177001955, 0.037823486328125, 0.03759820938110352, 0.03769651031494141, 0.037648384094238284, 0.03772313690185547, 0.03791667175292969, 0.03774259185791016, 0.03760435104370117, 0.03763711929321289, 0.03762483215332031, 0.03786444854736328, 0.03789516830444336, 0.07634636688232421, 0.03765350341796875, 0.03746713638305664, 0.03770470428466797, 0.03758796691894531, 0.03751116943359375, 0.03785318374633789, 0.037889022827148434, 0.03752345657348633, 0.0374774398803711, 0.037649345397949216, 0.03765760040283203, 0.037743614196777346, 0.0377149429321289, 0.03764329528808594, 0.03746198272705078, 0.037561344146728515, 0.037763072967529294, 0.03768832015991211, 0.03759823989868164, 0.037550048828125, 0.03745382308959961, 0.03772825622558594, 0.03768012619018555, 0.03773952102661133, 0.03763097763061524, 0.03759820938110352, 0.03765657424926758, 0.0381921272277832, 0.037905406951904294, 0.03770675277709961, 0.03752345657348633, 0.03779072189331055, 0.0376627197265625, 0.03760844802856445, 0.03748761749267578, 0.03757670211791992, 0.0375654411315918, 0.03750713729858399, 0.03755820846557617, 0.0375551986694336, 0.03763302230834961, 0.037645313262939455, 0.03746099090576172, 0.03763507080078125, 0.0376258544921875, 0.03769651031494141, 0.0375203857421875, 0.037610496520996094, 0.037612545013427735, 0.03768627166748047, 0.03968307113647461, 0.03810201644897461, 0.03744870376586914, 0.03748761749267578, 0.03760844802856445, 0.037705726623535156, 0.0375654411315918, 0.037533695220947266, 0.037482494354248046, 0.037713920593261716, 0.037579776763916016, 0.037495807647705076, 0.07634534454345702, 0.03747840118408203, 0.037424129486083986, 0.03769753646850586, 0.03747020721435547, 0.03760639953613281, 0.03746815872192383, 0.03759308624267578, 0.03759001541137695, 0.037582847595214845, 0.037572608947753904, 0.037525505065917966, 0.03752959823608398, 0.03775590515136719, 0.03758899307250976, 0.037459968566894535, 0.0377446403503418, 0.037495807647705076, 0.03755929565429687, 0.037438465118408204, 0.03756851196289063, 0.037602302551269534, 0.037515262603759765, 0.03751321411132812, 0.03760639953613281, 0.037541889190673826, 0.03759206390380859, 0.03758694458007812, 0.03766681671142578, 0.037548030853271484, 0.037512191772460936, 0.037582847595214845, 0.037493759155273435, 0.037449726104736326, 0.03746815872192383, 0.03757056045532227, 0.037823486328125, 0.037482494354248046, 0.037579776763916016, 0.03767091369628906, 0.03766886520385742, 0.03751424026489258, 0.03748147201538086, 0.037495807647705076, 0.03765555191040039, 0.03773132705688476, 0.03749273681640625, 0.03760639953613281, 0.03768217468261719, 0.037533695220947266, 0.037394432067871096, 0.03765350341796875, 0.03762278366088867, 0.03748966217041016, 0.03768729782104492, 0.0376514892578125, 0.03766985702514648, 0.03745280075073242, 0.037541889190673826, 0.03762073516845703, 0.03751731109619141, 0.037572608947753904, 0.03764223861694336, 0.07663206481933593, 0.037482494354248046, 0.03758694458007812, 0.03754598236083984, 0.03749785614013672, 0.03743641662597656, 0.03762790298461914, 0.03742105484008789, 0.03755929565429687, 0.037628929138183595, 0.03765350341796875, 0.03744768142700195, 0.03746918487548828, 0.03767500686645508, 0.03757670211791992, 0.03751222229003906, 0.03749884796142578, 0.03762995147705078, 0.03771596908569336, 0.03759718322753906, 0.03778252792358398, 0.03771596908569336, 0.03758182525634766, 0.03766886520385742, 0.0377262077331543, 0.03749683380126953, 0.0374917106628418, 0.037571582794189456, 0.03759513473510742, 0.03750912094116211, 0.037544960021972655, 0.0376627197265625, 0.03753574371337891, 0.037466110229492186, 0.03755212783813477, 0.03748044967651367, 0.03741798400878906, 0.03748761749267578, 0.03759414291381836, 0.03739030456542969, 0.03777740859985351, 0.03743129730224609, 0.037466110229492186, 0.037466110229492186, 0.03757056045532227, 0.03745177459716797, 0.03782451248168945, 0.03930624008178711, 0.03780198287963867, 0.037553150177001955, 0.037572608947753904, 0.037628929138183595, 0.03765350341796875, 0.03766579055786133, 0.03758796691894531, 0.03759718322753906, 0.03784396743774414, 0.038711296081542966, 0.03820236968994141, 0.037833728790283204, 0.03772931289672852, 0.03775075149536133, 0.03758796691894531, 0.07656038665771485, 0.03744153594970703, 0.037580799102783204, 0.03758489608764649, 0.03763507080078125, 0.03759308624267578, 0.03766681671142578, 0.0375470085144043, 0.0374835205078125, 0.03772313690185547, 0.037703678131103514, 0.037541889190673826, 0.0375551986694336, 0.037652511596679684, 0.037497825622558593, 0.03801190567016602, 0.037644287109375, 0.03753881454467774, 0.03759718322753906, 0.03752345657348633, 0.037541889190673826, 0.03757363128662109, 0.0374835205078125, 0.03762176132202148, 0.03808870315551758, 0.03783782577514649, 0.03767705535888672, 0.03754086303710937, 0.03700121688842774, 0.03764223861694336, 0.03752959823608398, 0.037765121459960936, 0.03765862274169922, 0.03768012619018555, 0.03785116958618164, 0.03752854537963867, 0.03746201705932617, 0.037591041564941405, 0.03764022445678711, 0.03770057678222656, 0.03760435104370117, 0.037569534301757815, 0.03757875061035156, 0.037550079345703126, 0.03768320083618164, 0.03784806442260742, 0.037806079864501956, 0.037579776763916016, 0.03758796691894531, 0.03765248107910156, 0.03765862274169922, 0.03774566268920899, 0.03764121627807617, 0.03772313690185547, 0.03784294509887695, 0.037817344665527344, 0.03783987045288086, 0.03761971282958984, 0.03763814544677734, 0.03765862274169922, 0.0377077751159668, 0.03765862274169922, 0.03762995147705078, 0.07627571105957032, 0.03767603302001953, 0.03746713638305664, 0.03762278366088867, 0.03759001541137695, 0.037689342498779296, 0.03778559875488281, 0.03764019012451172, 0.03788390350341797, 0.03751424026489258, 0.03749273681640625, 0.03770880126953125, 0.03754291152954101, 0.037703678131103514, 0.03769036865234375, 0.037574657440185545, 0.03785420989990235, 0.03765657424926758, 0.03755929565429687, 0.037599231719970705, 0.03741593551635742, 0.03765555191040039, 0.0375838737487793, 0.038059009552001956, 0.03752243041992188, 0.03781324768066406, 0.03762790298461914, 0.03782761764526367, 0.03764118576049805, 0.03759513473510742, 0.03786444854736328, 0.037617664337158206, 0.03751731109619141, 0.037566463470458986, 0.03778252792358398, 0.037596160888671876, 0.03751321411132812, 0.03806412887573242, 0.03936972808837891, 0.03786342239379883, 0.03775897598266602, 0.037601280212402347, 0.03762483215332031, 0.03776921463012695, 0.03782860946655273, 0.037449726104736326, 0.037705726623535156, 0.0375654411315918, 0.03755929565429687, 0.03753267288208008, 0.03763302230834961, 0.03753881454467774, 0.037580799102783204, 0.03768832015991211, 0.037574657440185545, 0.037531646728515625, 0.03764940643310547, 0.03754393768310547, 0.03761151885986328, 0.037515262603759765, 0.03761459350585938, 0.03788288116455078, 0.03977830505371094, 0.07646412658691407, 0.03738828659057617, 0.03792588806152344, 0.037748737335205076, 0.03746713638305664, 0.037897216796875, 0.037553150177001955, 0.03762176132202148, 0.037343231201171875, 0.037651454925537106, 0.037405696868896485, 0.0374835205078125, 0.03782758331298828, 0.03766681671142578, 0.03759001541137695, 0.0375838737487793, 0.0378081283569336, 0.03773235321044922, 0.03765760040283203, 0.03766886520385742, 0.03765964889526367, 0.03771187210083008, 0.03750707244873047, 0.03753472137451172, 0.037531646728515625, 0.03755110549926758, 0.03769343948364258, 0.03748659133911133, 0.037596160888671876, 0.03750707244873047, 0.03753881454467774, 0.037820415496826174, 0.03772415924072266, 0.03755929565429687, 0.037553150177001955, 0.03769036865234375, 0.03775897598266602, 0.037594112396240234, 0.03757875061035156, 0.03759513473510742, 0.03752755355834961, 0.03780505752563477, 0.03753267288208008, 0.03759820938110352, 0.03762790298461914, 0.03850447845458985, 0.03777225494384766, 0.03763916778564453, 0.03750297546386719, 0.037561344146728515, 0.037645313262939455, 0.03790848159790039, 0.037582847595214845, 0.037553150177001955, 0.037664768218994144, 0.03753574371337891, 0.03770675277709961, 0.03749075317382813, 0.03756332778930664, 0.03781324768066406, 0.03759718322753906, 0.037773311614990236, 0.037571582794189456, 0.07186124420166015, 0.037190654754638675, 0.03747225570678711, 0.03754908752441406, 0.037579742431640625, 0.03760639953613281, 0.03770982360839844, 0.03739136123657227, 0.03773440170288086, 0.0376360969543457, 0.03769241714477539, 0.03769241714477539, 0.03752345657348633, 0.03768627166748047, 0.03762688064575195, 0.03752140808105469, 0.037733375549316404, 0.03770675277709961, 0.037754878997802735, 0.03753472137451172, 0.03745792007446289, 0.03804876708984375, 0.03773747253417969, 0.037572608947753904, 0.03754291152954101, 0.03757056045532227, 0.03814297485351562, 0.03765350341796875, 0.03766988754272461, 0.03759001541137695, 0.03777536010742188, 0.03753472137451172, 0.037574657440185545, 0.03759718322753906, 0.03750707244873047, 0.03766886520385742, 0.03754393768310547, 0.03745280075073242, 0.03773952102661133, 0.03783168029785156, 0.03754086303710937, 0.03762688064575195, 0.037483551025390624, 0.03754902267456055, 0.03769036865234375, 0.03759718322753906, 0.037456897735595705, 0.03748761749267578, 0.038013950347900394, 0.037768192291259765, 0.03762073516845703, 0.03767705535888672, 0.037928958892822266, 0.03753779220581055, 0.03802828979492188, 0.037585918426513674, 0.039975936889648435, 0.03809791946411133, 0.03769343948364258, 0.037564414978027344, 0.03762278366088867, 0.03766988754272461, 0.03770470428466797, 0.07192781066894531, 0.03523993682861328, 0.03559731292724609, 0.038373374938964845, 0.03784908676147461, 0.03781017684936523, 0.03761663818359375, 0.037765121459960936, 0.037564414978027344, 0.037615615844726565, 0.03755417633056641, 0.038040576934814455, 0.03775692749023438, 0.03775692749023438, 0.037582847595214845, 0.037716991424560545, 0.03759308624267578, 0.03775283050537109, 0.03753062438964844, 0.03760435104370117, 0.03752447891235351, 0.037569534301757815, 0.03765964889526367, 0.03779891204833984, 0.03758182525634766, 0.037698558807373043, 0.037585918426513674, 0.03753062438964844, 0.03767398452758789, 0.037664768218994144, 0.0376627197265625, 0.037787647247314454, 0.03765657424926758, 0.037833728790283204, 0.037781505584716796, 0.037684223175048825, 0.037719039916992186, 0.03794944000244141, 0.037571582794189456, 0.03768627166748047, 0.037579776763916016, 0.037574657440185545, 0.03738521575927734, 0.03818188858032227, 0.0376627197265625, 0.03749990463256836, 0.037795841217041014, 0.03761151885986328, 0.037803009033203126, 0.0377077751159668, 0.03760435104370117, 0.03778867340087891, 0.037615615844726565, 0.03751731109619141, 0.03779379272460937, 0.03823820877075195, 0.03774054336547852, 0.03762688064575195, 0.03750400161743164, 0.037553150177001955, 0.037956607818603515, 0.03771289443969727, 0.037722110748291016]",tokens/s,26.191427819286087,, @@ -2068,7 +2232,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1075-75d379d03cfd3823033d9cf3;c7d43157-6e30-45df-a828-50fd0f79925f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d42bc-65f7bc032e59e6c10002d7a9;0e51481d-fe05-4ca3-acf2-6a83db6fa7a6) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2138,7 +2302,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c14e3-7c25f1607d58929f21bb51b0;ea7607b8-ab06-4b09-bc0e-d197fe871fe5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46f7-232df3e435d4184b7657b9f4;0b0c7fc6-a9ba-4f99-95d6-64a0e4bb93b2) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2208,7 +2372,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1492-5b3724e732736ef060a4af9d;1f190432-f6dc-45d5-a11b-81c82531ada5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46a9-0052fe5c40b74b847994d5ee;41ac9e3e-84d3-40ec-8c5e-5f8b89239aa0) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2308,7 +2472,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1370-499ed5ee08b43ca4324dade9;256820a9-0d4b-4dcc-8279-7ddc5c20a7e6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d458a-7bbd6efc04b8591a1692da08;61638c95-b691-435a-b21d-438214e707f6) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2407,7 +2571,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c13e6-257601322bf5c2235b1c4ebc;5ad5c7c6-5b95-45f0-a889-d5e011897d58) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45fe-4cd77c8f4f4f9b135da7ca55;4b8c223f-cbef-4e99-8e64-3950ed38701b) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2616,6 +2780,35 @@ ChildProcessError: Traceback (most recent call last): return forward_call(*args, **kwargs) TypeError: DeciCoderAttention.forward() got an unexpected keyword argument 'cache_position' +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 563, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1478, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1644, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch @@ -2646,7 +2839,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664beeae-1767aeee1ff3b92f0d6c067a;311425fa-a7ee-4f24-94d7-b6b1e657cb2b) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3fbc-4bba41691649b5893a3e1d54;8d993afa-ceba-42d4-be56-f4e76d0bdfa8) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -2851,7 +3044,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1398-718a4e0215219abc526db24b;520bbf99-c25e-461b-8eee-3264aa416672) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45b0-16f653357bb53e3611fad150;955d08dc-0b51-4ca9-9752-6ff4f559c7bf) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2911,7 +3104,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee86-06276f9e34420fcd76b2aeeb;6691828a-349a-48fd-817e-54f1710b28eb) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3f95-5428121107e9feb24b12df1f;27d93148-d14d-4ccb-aa46-c5a012d2c39c) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -3436,7 +3629,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664becfd-270e193d04b155cb3823a940;d97d3cb4-ccc3-4e1d-83e0-29a09804d2ec) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3e15-7bab3ca4118594ee05eb660e;fe8d7c2f-7794-4a26-958c-5f0b86f42ea9) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -3521,7 +3714,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1320-0d2f591c7aab4c95050df1fa;546593f7-f23f-408b-86a4-8341966c7265) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d453d-2c3a6d7a6d7202542c99c6e3;51f37a64-f936-4c48-96e9-99a7ae48ab39) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3665,7 +3858,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c121d-42fc65390ce237bf2099d58e;f2511914-4870-4435-8125-aae370174e4f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4477-58b1ea017057deff65a45127;032535d3-2f91-48d4-a153-2914ec516e17) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3739,6 +3932,35 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1861.566464,3395.813376,0.0,2749.366272,2585.357312,s,10,0.732101089477539,0.0732101089477539,0.0009905213877066726,0.07305867004394531,0.07430353927612304,0.07468780326843262,0.07499521446228027,"[0.07394802856445312, 0.07507206726074218, 0.07251001739501953, 0.07249107360839843, 0.07266223907470704, 0.07207814025878906, 0.07345510101318359, 0.07379513549804688, 0.07421814727783203, 0.07187113952636719]",tokens/s,3496.7848522489344,kWh,8.519276714534592e-07,4.66812359389341e-07,4.015554771193711e-06,5.334294802036511e-06,tokens/kWh,47991348.34135247,MB,1861.566464,3395.813376,0.0,2749.366272,2628.768768,s,10,38.62193530273437,3.862193530273438,0.003503694519058953,3.8613565673828125,3.8657285400390626,3.8678885620117187,3.8696165795898434,"[3.870048583984375, 3.86524853515625, 3.86500830078125, 3.861310546875, 3.8626630859375, 3.861402587890625, 3.860045654296875, 3.859800048828125, 3.858201904296875, 3.8582060546875]",tokens/s,16.311973883799574,kWh,4.5592295868243344e-05,2.4984092023898362e-05,0.00021485331476820774,0.00028542970266034936,tokens/kWh,220719.84594738425,,s,629,39.21789340972901,0.062349592066341816,0.008754170010469397,0.06125347137451172,0.06140214309692383,0.06161048431396485,0.13420894287109375,"[0.06200627136230469, 0.062050304412841796, 0.062051326751708984, 0.062050304412841796, 0.062045185089111325, 0.06205952072143555, 0.062058494567871096, 0.06193766403198242, 0.061431808471679686, 0.06132940673828125, 0.061346817016601565, 0.06124851226806641, 0.06131001663208008, 0.06127609634399414, 0.06120550537109375, 0.06132633590698242, 0.06146662521362305, 0.061208576202392576, 0.06135193634033203, 0.06129971313476563, 0.06123110580444336, 0.06118707275390625, 0.061292545318603515, 0.06125568008422851, 0.06133350372314453, 0.061290496826171874, 0.06124652862548828, 0.061225921630859374, 0.061346912384033205, 0.061209503173828124, 0.061295616149902345, 0.06128947067260742, 0.061267967224121096, 0.06126182556152344, 0.06134374237060547, 0.06130483245849609, 0.06131302261352539, 0.061306880950927733, 0.06135603332519531, 0.06126182556152344, 0.06134783935546875, 0.06131097412109375, 0.06137241744995117, 0.06140518569946289, 0.06134783935546875, 0.0612853775024414, 0.06133657455444336, 0.06125875091552734, 0.061308097839355466, 0.06119721603393555, 0.061290401458740235, 0.061284351348876956, 0.06131100845336914, 0.061276126861572265, 0.0613171501159668, 0.061335521697998045, 0.06157120132446289, 0.06169484710693359, 0.06147891235351562, 0.06137855911254883, 0.06133555221557617, 0.06140723037719727, 0.13694259643554688, 0.06134783935546875, 0.06128639984130859, 0.061176830291748044, 0.06126182556152344, 0.06133760070800781, 0.06124851226806641, 0.06124748611450195, 0.061262847900390625, 0.06136524963378906, 0.061467647552490234, 0.0615208969116211, 0.06147686386108398, 0.06129663848876953, 0.06119014358520508, 0.06120755386352539, 0.06123212814331055, 0.06132633590698242, 0.061279232025146485, 0.06128742218017578, 0.0612935676574707, 0.06128025436401367, 0.061241344451904295, 0.06126489639282227, 0.06128332901000977, 0.061233150482177735, 0.06123519897460938, 0.061358081817626954, 0.061282302856445314, 0.06133046340942383, 0.061308895111083984, 0.06129663848876953, 0.06127206420898437, 0.06125568008422851, 0.06132428741455078, 0.06128639984130859, 0.06130483245849609, 0.06126694488525391, 0.06124031829833984, 0.06144204711914063, 0.06149631881713867, 0.06135500717163086, 0.06128646469116211, 0.06134777450561523, 0.06132428741455078, 0.06129663848876953, 0.06375430297851563, 0.06132217788696289, 0.06123519897460938, 0.06129971313476563, 0.061276161193847656, 0.061272224426269534, 0.06125347137451172, 0.06133964920043945, 0.06128339385986328, 0.06130579376220703, 0.06134483337402344, 0.06128326416015625, 0.061246463775634766, 0.061332576751708986, 0.06133852767944336, 0.061446144104003904, 0.061350910186767575, 0.13430271911621094, 0.06127001571655273, 0.061233150482177735, 0.06136012649536133, 0.06128947067260742, 0.06129459381103516, 0.06125056076049805, 0.06130793762207031, 0.06123107147216797, 0.061254657745361325, 0.0612935676574707, 0.06134067153930664, 0.06121779251098633, 0.06132940673828125, 0.06141030502319336, 0.06151679992675781, 0.06136217498779297, 0.061445121765136716, 0.06127523040771484, 0.061303775787353514, 0.06122387313842773, 0.061254657745361325, 0.06121062469482422, 0.062117889404296876, 0.06129971313476563, 0.061297664642333986, 0.061246463775634766, 0.06128742218017578, 0.06134080123901367, 0.06140198516845703, 0.061385726928710936, 0.06134988784790039, 0.06122598266601562, 0.06134067153930664, 0.06146662521362305, 0.06136422348022461, 0.06127308654785156, 0.06131097412109375, 0.06125260925292969, 0.06135398483276367, 0.061334560394287106, 0.06134473419189453, 0.06128639984130859, 0.06140927886962891, 0.06132940673828125, 0.061383678436279294, 0.06133760070800781, 0.062399486541748046, 0.06130995178222656, 0.061292545318603515, 0.061216766357421876, 0.061344768524169924, 0.061290496826171874, 0.061297664642333986, 0.061300735473632816, 0.06129971313476563, 0.06125158309936524, 0.06135193634033203, 0.06126489639282227, 0.06127001571655273, 0.061341697692871094, 0.06126489639282227, 0.06131609725952149, 0.1343057861328125, 0.06122393417358398, 0.06128332901000977, 0.06137446212768555, 0.061221889495849606, 0.061254657745361325, 0.06116352081298828, 0.06125875091552734, 0.0611512336730957, 0.06126492691040039, 0.06120751953125, 0.06121779251098633, 0.0611962890625, 0.06132633590698242, 0.061233150482177735, 0.061298686981201174, 0.06121779251098633, 0.061222911834716794, 0.06113391876220703, 0.06125968170166016, 0.06115020751953125, 0.0611778564453125, 0.061143039703369144, 0.06125875091552734, 0.06121779251098633, 0.06122905731201172, 0.06124649429321289, 0.06131824111938477, 0.061295486450195315, 0.061472766876220705, 0.061192192077636716, 0.06132326507568359, 0.06136832046508789, 0.061267967224121096, 0.06120550537109375, 0.06134067153930664, 0.06122905731201172, 0.06128742218017578, 0.061238273620605466, 0.06129971313476563, 0.06122905731201172, 0.061338623046875, 0.06120243072509766, 0.06127718353271484, 0.06223360061645508, 0.06153113555908203, 0.06150451278686524, 0.061418495178222655, 0.06119935989379883, 0.06135295867919922, 0.06118195343017578, 0.06127206420898437, 0.06126182556152344, 0.06143078231811523, 0.06123212814331055, 0.06126694488525391, 0.06118707275390625, 0.06128639984130859, 0.061200382232666016, 0.061346817016601565, 0.061312000274658204, 0.0612147216796875, 0.061295616149902345, 0.13448602294921874, 0.06163667297363281, 0.0612740478515625, 0.06121574401855469, 0.06132121658325195, 0.06135603332519531, 0.06117068862915039, 0.06136422348022461, 0.0611962890625, 0.06124544143676758, 0.061176864624023435, 0.061287391662597654, 0.06121779251098633, 0.06123110580444336, 0.06122598266601562, 0.06128947067260742, 0.06124236679077148, 0.06216089630126953, 0.06116659164428711, 0.06125875091552734, 0.06119833755493164, 0.06118297576904297, 0.06118809509277344, 0.06124860763549805, 0.06115318298339844, 0.06120243072509766, 0.061222911834716794, 0.06121881484985352, 0.06118707275390625, 0.061338623046875, 0.06123110580444336, 0.061228031158447264, 0.06126899337768555, 0.0612567024230957, 0.06119014358520508, 0.06126182556152344, 0.06132633590698242, 0.061192192077636716, 0.0611860466003418, 0.061295616149902345, 0.06120652770996094, 0.0612393913269043, 0.061224864959716796, 0.06126800155639649, 0.06119113540649414, 0.06131916809082031, 0.061198463439941404, 0.06121868896484375, 0.06128639984130859, 0.06151065444946289, 0.061228031158447264, 0.061271167755126955, 0.06120230484008789, 0.06123417663574219, 0.06121267318725586, 0.061297664642333986, 0.061246463775634766, 0.06344601440429687, 0.06122086334228516, 0.06132428741455078, 0.0612751350402832, 0.06124031829833984, 0.061357120513916015, 0.13396780395507812, 0.06137753677368164, 0.06133760070800781, 0.06117068862915039, 0.061176830291748044, 0.06124240112304687, 0.061154430389404296, 0.06114390563964844, 0.061158401489257816, 0.06123929595947265, 0.061139968872070315, 0.06118707275390625, 0.06120755386352539, 0.06121574401855469, 0.06114713668823242, 0.061330432891845706, 0.06128947067260742, 0.061262847900390625, 0.06122598266601562, 0.061257728576660155, 0.06117068862915039, 0.0612044792175293, 0.06126182556152344, 0.06124544143676758, 0.061168670654296875, 0.06120240020751953, 0.0611778564453125, 0.06120550537109375, 0.06122598266601562, 0.06124748611450195, 0.0611778564453125, 0.06125568008422851, 0.06140227127075195, 0.0613282241821289, 0.0612782096862793, 0.06129663848876953, 0.06118195343017578, 0.06134067153930664, 0.06349004745483398, 0.061344768524169924, 0.06118809509277344, 0.06125363159179688, 0.06121267318725586, 0.0612782096862793, 0.06121574401855469, 0.061282302856445314, 0.06121574401855469, 0.0612567024230957, 0.061241344451904295, 0.061473823547363284, 0.06134579086303711, 0.06122390365600586, 0.06124236679077148, 0.06125056076049805, 0.061193214416503904, 0.061243392944335937, 0.061246463775634766, 0.06125056076049805, 0.061238273620605466, 0.061230079650878906, 0.0612147216796875, 0.0612782096862793, 0.06139289474487305, 0.13551513671875, 0.061222911834716794, 0.061300865173339845, 0.06122995376586914, 0.06116864013671875, 0.061241344451904295, 0.06116147232055664, 0.06121574401855469, 0.061233150482177735, 0.06125056076049805, 0.061224960327148435, 0.061298686981201174, 0.0611860466003418, 0.061274112701416014, 0.0612782096862793, 0.06128742218017578, 0.06179743957519531, 0.061714366912841795, 0.06137036895751953, 0.06115532684326172, 0.06111948776245117, 0.06116659164428711, 0.061200382232666016, 0.06237392044067383, 0.06122288131713867, 0.06122412872314453, 0.061155136108398435, 0.061274112701416014, 0.0611778564453125, 0.061254657745361325, 0.061224960327148435, 0.06123110580444336, 0.06135500717163086, 0.06133452987670898, 0.061189121246337894, 0.061244415283203124, 0.06118611145019531, 0.06124025726318359, 0.06119935989379883, 0.061228031158447264, 0.061192192077636716, 0.061260799407958984, 0.06117068862915039, 0.06125888061523437, 0.061265792846679684, 0.061200382232666016, 0.06116150283813476, 0.06126895904541016, 0.061183998107910156, 0.06129459381103516, 0.061162494659423826, 0.061208576202392576, 0.0611962890625, 0.061230079650878906, 0.06120451354980469, 0.06117782211303711, 0.06115327835083008, 0.06122905731201172, 0.06116659164428711, 0.061230079650878906, 0.061257728576660155, 0.061224960327148435, 0.06124748611450195, 0.13553868103027344, 0.06118195343017578, 0.061238273620605466, 0.06116864013671875, 0.06116761779785156, 0.06121881484985352, 0.06114918518066406, 0.061135936737060546, 0.06112966537475586, 0.06120243072509766, 0.0611409912109375, 0.06120550537109375, 0.06121062469482422, 0.0611962890625, 0.06115327835083008, 0.061298686981201174, 0.061943809509277345, 0.06136524963378906, 0.06138470458984375, 0.06130588912963867, 0.061340641021728516, 0.06137241744995117, 0.06116966247558594, 0.06127206420898437, 0.061456382751464846, 0.06138777542114258, 0.061489280700683595, 0.061464447021484375, 0.06123110580444336, 0.061284351348876956, 0.06120652770996094, 0.061193214416503904, 0.061192192077636716, 0.06128025436401367, 0.06137036895751953, 0.06136627197265625, 0.06142771148681641, 0.061208576202392576, 0.06112768173217773, 0.06126185607910156, 0.06128329467773438, 0.06123110580444336, 0.06121574401855469, 0.06126182556152344, 0.06118707275390625, 0.061271041870117185, 0.061193214416503904, 0.06123417663574219, 0.061163551330566404, 0.06126399993896484, 0.06111932754516602, 0.061269153594970704, 0.061199199676513674, 0.06125980758666992, 0.061153247833251954, 0.061228031158447264, 0.061243392944335937, 0.06127308654785156, 0.06119833755493164, 0.061274112701416014, 0.06121062469482422, 0.06120550537109375, 0.06126182556152344, 0.13602406311035156, 0.06120140838623047, 0.061265918731689455, 0.06120550537109375, 0.06120755386352539, 0.06119731140136719, 0.06116761779785156, 0.06120345687866211, 0.0611256332397461, 0.061224960327148435, 0.06121881484985352, 0.061208576202392576, 0.06116556930541992, 0.061209632873535154, 0.06117071914672852, 0.06118393707275391, 0.061159423828125, 0.06195302581787109, 0.06117478561401367, 0.06131014251708984, 0.06121555328369141, 0.061200382232666016, 0.061189121246337894, 0.061211647033691405, 0.06113382339477539, 0.06121267318725586, 0.06122393417358398, 0.061211647033691405, 0.06117478561401367, 0.06120652770996094, 0.061213695526123046, 0.06119116973876953, 0.06121062469482422, 0.06128844833374023, 0.061402111053466796, 0.06131315231323242, 0.06120947265625, 0.061208576202392576, 0.06116556930541992, 0.061249534606933595, 0.06116147232055664, 0.06124236679077148, 0.061209632873535154, 0.06124848175048828, 0.06123734283447266, 0.061222816467285154, 0.06129059219360351, 0.06121871948242188, 0.06120243072509766, 0.06126387023925781, 0.061208576202392576, 0.06143283081054687, 0.06129459381103516, 0.061241344451904295, 0.061193214416503904, 0.06121984100341797, 0.06117171096801758, 0.061183998107910156, 0.06115737533569336, 0.061206687927246095, 0.06120636749267578, 0.06126694488525391, 0.06128947067260742, 0.1337159729003906, 0.06119833755493164, 0.061267967224121096, 0.061118465423583984, 0.06119731140136719, 0.06125363159179688, 0.06119116973876953, 0.06113689422607422, 0.06114918518066406, 0.06121779251098633, 0.061126655578613284, 0.06121068954467773, 0.0611736946105957, 0.06115020751953125, 0.061172737121582034, 0.0612147216796875, 0.06117375946044922, 0.061211647033691405, 0.06116352081298828, 0.061233150482177735, 0.06114508819580078, 0.06125875091552734, 0.06129971313476563, 0.061185054779052735, 0.061112289428710935, 0.061224960327148435, 0.06247423934936523, 0.06127414321899414, 0.06117987060546875, 0.061175807952880856, 0.06119935989379883, 0.06124448013305664, 0.061127616882324216, 0.06123929595947265, 0.06120550537109375, 0.061259777069091796, 0.06121779251098633, 0.06124236679077148, 0.061239326477050784, 0.06118502426147461, 0.06120547103881836, 0.06126489639282227, 0.061208576202392576, 0.06122905731201172, 0.061249534606933595, 0.06124236679077148, 0.061159423828125, 0.06130278396606445, 0.06115225601196289, 0.06123929595947265, 0.06117068862915039, 0.06123929595947265, 0.061241344451904295, 0.061357120513916015, 0.061324222564697266, 0.06140313720703125, 0.061230079650878906, 0.06125875091552734, 0.06115737533569336, 0.06119935989379883, 0.06123622512817383, 0.06115430450439453, 0.061200382232666016]",tokens/s,16.038597316498404,, +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1478, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1644, in _check_and_enable_sdpa + raise ValueError( +ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -3808,7 +4030,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c145e-388770dc7fff864d7214ca84;56d74fe2-a45d-4580-83df-49ea1fcaf79b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4673-7105734d35b6d0eb662aaac0;49f80694-dd7f-4175-a694-4a26c3dbe03a) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3907,7 +4129,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c11f2-37273d2868b3f90f16904b9e;85b3ffd3-0eb6-4f82-8af2-5892ad8d40a7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4451-4ac5c78a4460fce80b74f881;0a361a63-f01b-46ec-baea-c9e94c4abdeb) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4006,7 +4228,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12bd-50f1f5e75e74fb4559f1c35c;6101c4e3-632b-4d26-9071-b7e250abd972) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44ee-666b0f882e11aed700666155;bf27dec1-a110-4cac-b8b0-5b55509aa750) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4038,8 +4260,37 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5006.11072,8173.125632,0.0,7526.678528,6833.613824,s,10,3.234631896972656,0.3234631896972656,0.001290890363598202,0.32324838256835936,0.32507350158691406,0.3250783767700195,0.3250822769165039,"[0.3225212097167969, 0.32506610107421874, 0.32166519165039065, 0.3220767822265625, 0.32293771362304685, 0.3222037658691406, 0.32355905151367187, 0.3244464111328125, 0.3250724182128906, 0.325083251953125]",tokens/s,791.4347231893512,kWh,3.805942098713583e-06,2.0851093678459165e-06,2.0185346009376354e-05,2.6076397475935855e-05,tokens/kWh,9817307.020122128,MB,5006.11072,8173.125632,0.0,7526.678528,7110.583808,s,10,176.69872265624997,17.669872265624996,0.0012983881421126473,17.6700712890625,17.6713546875,17.67136875,17.67138,"[17.667875, 17.670087890625, 17.67133984375, 17.67002734375, 17.6713515625, 17.6700546875, 17.66839453125, 17.66792578125, 17.6713828125, 17.670283203125]",tokens/s,3.5653907992623313,kWh,0.00020855451250034905,0.00011430324463261854,0.0010754413360468291,0.0013982990931797968,tokens/kWh,45054.7385085798,,s,629,179.32957586669917,0.2851026643349749,0.03850587607379773,0.2804192810058594,0.28063436889648435,0.28075439453125,0.6034208325195313,"[0.2807080993652344, 0.2805329895019531, 0.28025140380859376, 0.28053607177734374, 0.28031283569335935, 0.2804653930664063, 0.2804398193359375, 0.2805975036621094, 0.2809886779785156, 0.2807132263183594, 0.2803640441894531, 0.28059136962890624, 0.28041012573242186, 0.28036813354492185, 0.2803312683105469, 0.2803927001953125, 0.2804172668457031, 0.28025445556640627, 0.2803117980957031, 0.28033331298828124, 0.2803240966796875, 0.2801163635253906, 0.280345458984375, 0.28016845703125, 0.280369140625, 0.28039883422851564, 0.2807910461425781, 0.280511474609375, 0.280563720703125, 0.28041317749023437, 0.2805340270996094, 0.2804440002441406, 0.28055950927734374, 0.2804561767578125, 0.28033535766601564, 0.28021453857421874, 0.2805248107910156, 0.2804111328125, 0.28032101440429685, 0.2803333740234375, 0.2802308349609375, 0.2801756286621094, 0.28041830444335936, 0.28040704345703127, 0.2803896179199219, 0.2802872314453125, 0.2804316101074219, 0.2803599243164063, 0.2805401611328125, 0.28073471069335937, 0.28038763427734376, 0.28052883911132814, 0.28028314208984373, 0.2804653930664063, 0.2805145568847656, 0.28037631225585935, 0.28035687255859376, 0.28009164428710936, 0.280563720703125, 0.2805452880859375, 0.2803619689941406, 0.2802472839355469, 0.606171142578125, 0.28023602294921873, 0.2805667724609375, 0.28041830444335936, 0.28034356689453127, 0.2801756286621094, 0.2803046264648438, 0.2802882690429688, 0.2804449157714844, 0.2802883605957031, 0.2804110412597656, 0.280369140625, 0.2804029541015625, 0.2803896179199219, 0.28032205200195315, 0.28044595336914063, 0.2805125122070313, 0.28050225830078124, 0.28046234130859377, 0.2804019165039062, 0.2804981689453125, 0.2805166015625, 0.28044082641601564, 0.2806159362792969, 0.28032818603515625, 0.28030157470703126, 0.2801919860839844, 0.28039065551757814, 0.2802534484863281, 0.2802708740234375, 0.28019607543945313, 0.28049099731445315, 0.2804510803222656, 0.2804951171875, 0.2803937377929687, 0.28054119873046873, 0.28021759033203125, 0.28069683837890624, 0.28063232421875, 0.280690673828125, 0.28052377319335936, 0.28034970092773437, 0.2803292236328125, 0.2804285583496094, 0.28027392578125, 0.2803363952636719, 0.2830735473632813, 0.280342529296875, 0.2805002136230469, 0.28026470947265625, 0.280332275390625, 0.28067636108398436, 0.28094464111328127, 0.28055142211914064, 0.2803056640625, 0.28043777465820313, 0.280479736328125, 0.2805698547363281, 0.28024627685546877, 0.2804009094238281, 0.2803466186523437, 0.2804971618652344, 0.2802001953125, 0.6034718627929687, 0.2802196350097656, 0.2803599243164063, 0.28019610595703126, 0.2802452392578125, 0.28016439819335937, 0.28058108520507813, 0.28036813354492185, 0.280489990234375, 0.28041217041015626, 0.2804766845703125, 0.280406005859375, 0.28043264770507814, 0.28037017822265625, 0.2804029541015625, 0.2804766845703125, 0.2804409484863281, 0.28033114624023436, 0.280384521484375, 0.28046234130859377, 0.2805831604003906, 0.2805032958984375, 0.2803189697265625, 0.28047564697265626, 0.2804776916503906, 0.2805032958984375, 0.2803240966796875, 0.2804725646972656, 0.280263671875, 0.2804572143554688, 0.2804981689453125, 0.28048895263671875, 0.28037939453125, 0.28040704345703127, 0.28037939453125, 0.28038247680664063, 0.28024627685546877, 0.28041522216796877, 0.280531982421875, 0.28053912353515625, 0.2805320739746094, 0.28040591430664064, 0.280352783203125, 0.28057293701171876, 0.2807500915527344, 0.2806947937011719, 0.2803097839355469, 0.28375958251953126, 0.2804418640136719, 0.28035174560546877, 0.2806302795410156, 0.2805452880859375, 0.2805770263671875, 0.28036505126953126, 0.2803804016113281, 0.2804449157714844, 0.28060671997070313, 0.28033535766601564, 0.2802309265136719, 0.2807715759277344, 0.2806528015136719, 0.28054833984375, 0.2802769775390625, 0.6030131225585937, 0.280258544921875, 0.28030975341796877, 0.28019097900390627, 0.28033331298828124, 0.2803947448730469, 0.2806056823730469, 0.2804992065429687, 0.280352783203125, 0.28040499877929687, 0.28040704345703127, 0.2803046264648438, 0.2813982849121094, 0.2808186950683594, 0.28054425048828124, 0.2802554931640625, 0.280458251953125, 0.28053811645507815, 0.280384521484375, 0.2803640441894531, 0.28048382568359376, 0.2803180236816406, 0.2802380065917969, 0.280384521484375, 0.280332275390625, 0.2803558349609375, 0.280263671875, 0.28042752075195315, 0.2803292236328125, 0.28048382568359376, 0.28054022216796876, 0.28073977661132815, 0.28061285400390623, 0.28051358032226564, 0.28070404052734377, 0.28075820922851563, 0.2804500427246094, 0.2803436279296875, 0.2801786193847656, 0.28026571655273436, 0.28032000732421875, 0.28032000732421875, 0.28045208740234373, 0.28043777465820313, 0.28250009155273437, 0.2803670959472656, 0.28039065551757814, 0.28047259521484375, 0.2804735717773438, 0.28055657958984376, 0.2805083923339844, 0.28037017822265625, 0.28052276611328125, 0.2804500427246094, 0.28042648315429686, 0.2804561767578125, 0.28051864624023437, 0.28045208740234373, 0.280205322265625, 0.28048281860351565, 0.280342529296875, 0.28044287109375, 0.28014181518554687, 0.6037985229492188, 0.28016024780273435, 0.2806005859375, 0.2805433349609375, 0.2804161682128906, 0.2803446044921875, 0.2804192810058594, 0.28039065551757814, 0.28035482788085936, 0.2803814392089844, 0.28049099731445315, 0.2804490356445313, 0.280406005859375, 0.28048590087890624, 0.2803670959472656, 0.28037939453125, 0.2804244384765625, 0.2816696472167969, 0.2804664306640625, 0.2806661071777344, 0.2807828369140625, 0.2805749816894531, 0.28032205200195315, 0.2804725646972656, 0.2803814392089844, 0.2803445739746094, 0.2803363952636719, 0.28047564697265626, 0.28027496337890623, 0.2807398071289062, 0.28061285400390623, 0.2807162780761719, 0.2804971618652344, 0.28034970092773437, 0.28027493286132815, 0.28037734985351564, 0.2802923583984375, 0.28038861083984373, 0.280300537109375, 0.28038861083984373, 0.2803998718261719, 0.2803978271484375, 0.2803036193847656, 0.28044390869140623, 0.2803814392089844, 0.2804592590332031, 0.28030877685546873, 0.2804520568847656, 0.2805401611328125, 0.28040093994140625, 0.28047869873046877, 0.2802821044921875, 0.28053607177734374, 0.280384521484375, 0.28043060302734374, 0.28053607177734374, 0.2805882873535156, 0.2829209594726563, 0.28022271728515624, 0.28046746826171876, 0.28047564697265626, 0.28037734985351564, 0.2803466186523437, 0.6032896118164063, 0.28028729248046874, 0.280363037109375, 0.2801775817871094, 0.28032717895507814, 0.28021353149414063, 0.28027389526367186, 0.2802452392578125, 0.2804510803222656, 0.2805073852539062, 0.28084429931640625, 0.2804039611816406, 0.28040093994140625, 0.28040298461914065, 0.2804899291992187, 0.2803292236328125, 0.28033535766601564, 0.2803394470214844, 0.280374267578125, 0.2803261413574219, 0.28048794555664064, 0.28052581787109376, 0.28022271728515624, 0.2803998718261719, 0.28032818603515625, 0.2804019165039062, 0.2803507080078125, 0.28050125122070313, 0.28041217041015626, 0.28047579956054686, 0.2804621887207031, 0.28067019653320313, 0.2804643859863281, 0.2803978271484375, 0.28047360229492185, 0.2806077575683594, 0.28034356689453127, 0.2804471740722656, 0.2830160522460938, 0.28041522216796877, 0.2804878845214844, 0.2803507080078125, 0.28064254760742186, 0.28075726318359373, 0.28046847534179686, 0.2805565490722656, 0.2804930419921875, 0.280406005859375, 0.28038247680664063, 0.28038861083984373, 0.28051763916015626, 0.28036300659179686, 0.2802964477539063, 0.2803466186523437, 0.28047052001953127, 0.280489990234375, 0.280458251953125, 0.2804019165039062, 0.28022988891601563, 0.2804500427246094, 0.280489990234375, 0.28060671997070313, 0.28033535766601564, 0.6047180786132812, 0.2801838073730469, 0.2804490356445313, 0.28032305908203126, 0.2802616271972656, 0.28019915771484377, 0.2802657775878906, 0.2804325866699219, 0.28046847534179686, 0.2803947448730469, 0.28039065551757814, 0.28033642578125, 0.28051556396484373, 0.2802339782714844, 0.28041522216796877, 0.2803292236328125, 0.28062619018554685, 0.28064254760742186, 0.280531982421875, 0.28036300659179686, 0.2804572143554688, 0.2804643859863281, 0.2804633483886719, 0.28211508178710937, 0.28030157470703126, 0.2803814392089844, 0.2802503662109375, 0.28055142211914064, 0.2803292236328125, 0.28034765625, 0.28033023071289065, 0.28057089233398436, 0.28050125122070313, 0.2804154052734375, 0.2804631958007813, 0.28067019653320313, 0.2802503662109375, 0.28074188232421876, 0.28085147094726565, 0.2804930419921875, 0.28036813354492185, 0.280416259765625, 0.2802923583984375, 0.2804019165039062, 0.2803046264648438, 0.2803619689941406, 0.2803456115722656, 0.2802769775390625, 0.280453125, 0.2803978271484375, 0.28041317749023437, 0.28027392578125, 0.28071218872070314, 0.28065689086914064, 0.2804090881347656, 0.28041217041015626, 0.28040704345703127, 0.28029541015625, 0.2801346435546875, 0.28052581787109376, 0.2804592590332031, 0.28048590087890624, 0.2802903137207031, 0.6054840087890625, 0.28021453857421874, 0.280310791015625, 0.28032205200195315, 0.28034765625, 0.28025753784179686, 0.2805084533691406, 0.2806282043457031, 0.28064358520507815, 0.28045111083984375, 0.28058212280273437, 0.2803957824707031, 0.28048794555664064, 0.28039886474609377, 0.2804111022949219, 0.2803732604980469, 0.28168295288085937, 0.28037734985351564, 0.280384521484375, 0.2804725646972656, 0.280384521484375, 0.28036505126953126, 0.28022476196289064, 0.2803456115722656, 0.2803947448730469, 0.28043365478515625, 0.2802872314453125, 0.2803343505859375, 0.28028005981445314, 0.2804500427246094, 0.28037118530273436, 0.280627197265625, 0.28052581787109376, 0.28042138671875, 0.28047052001953127, 0.2803825378417969, 0.28021856689453123, 0.2804111328125, 0.280310791015625, 0.28049612426757814, 0.28038348388671874, 0.280332275390625, 0.2804029541015625, 0.2804561767578125, 0.2804613037109375, 0.28042138671875, 0.2804080505371094, 0.280585205078125, 0.2807091064453125, 0.28042138671875, 0.28039678955078123, 0.2802759704589844, 0.28042648315429686, 0.2803343505859375, 0.2803670959472656, 0.28048590087890624, 0.28051864624023437, 0.28044512939453126, 0.28021951293945313, 0.2805595397949219, 0.28056781005859377, 0.28054425048828124, 0.28022271728515624, 0.6061547241210937, 0.2803138427734375, 0.280384521484375, 0.2803660888671875, 0.28042657470703125, 0.280196044921875, 0.2803947448730469, 0.28044699096679687, 0.28030670166015625, 0.2804285583496094, 0.28048895263671875, 0.2804090881347656, 0.2807029724121094, 0.280764404296875, 0.2804920349121094, 0.280521728515625, 0.2805084228515625, 0.28190618896484376, 0.2804491271972656, 0.28042230224609377, 0.2804757080078125, 0.28043154907226564, 0.28037118530273436, 0.28051864624023437, 0.28060467529296873, 0.2805248107910156, 0.2803466186523437, 0.28050942993164063, 0.2803037109375, 0.2804366455078125, 0.2803835144042969, 0.28081353759765626, 0.28074496459960935, 0.28037835693359375, 0.28037631225585935, 0.2805555114746094, 0.28037939453125, 0.2805688171386719, 0.2804551696777344, 0.280510498046875, 0.2805585632324219, 0.2803660888671875, 0.280416259765625, 0.28055859375, 0.28047052001953127, 0.2804244384765625, 0.2805206909179688, 0.28048593139648437, 0.2806968078613281, 0.28052993774414064, 0.2805851745605469, 0.28038247680664063, 0.28042138671875, 0.2804561767578125, 0.28033740234375, 0.2804776916503906, 0.2805770263671875, 0.2803240966796875, 0.28041522216796877, 0.280511474609375, 0.28045413208007813, 0.2804766845703125, 0.28029034423828125, 0.6060789184570312, 0.28018893432617187, 0.28060467529296873, 0.28061285400390623, 0.28044287109375, 0.2802237548828125, 0.2804111328125, 0.28039678955078123, 0.2804664306640625, 0.2803947448730469, 0.2804111328125, 0.2804019165039062, 0.28042752075195315, 0.2803138427734375, 0.28037631225585935, 0.2802904052734375, 0.28029736328125, 0.28034970092773437, 0.28041934204101565, 0.2805073852539062, 0.2807215576171875, 0.28053692626953125, 0.28042752075195315, 0.2806824951171875, 0.2805852661132813, 0.280438720703125, 0.28213861083984376, 0.280531982421875, 0.28032000732421875, 0.28051559448242186, 0.28042648315429686, 0.280711181640625, 0.2804653930664063, 0.28054425048828124, 0.2804316101074219, 0.2805401611328125, 0.2803804016113281, 0.2804695129394531, 0.28054629516601565, 0.2805770263671875, 0.28039486694335936, 0.280461181640625, 0.28036300659179686, 0.28038247680664063, 0.2803312683105469, 0.2803261413574219, 0.28028005981445314, 0.28033023071289065, 0.28046234130859377, 0.2805422058105469, 0.28042755126953123, 0.2803332824707031, 0.2804695129394531, 0.28044082641601564, 0.2805206909179688, 0.28069171142578125, 0.28068865966796874, 0.2804121398925781, 0.28018893432617187, 0.2804930419921875, 0.28053094482421875, 0.2804398193359375, 0.2802923583984375]",tokens/s,3.5075084350143877,, +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1478, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1644, in _check_and_enable_sdpa + raise ValueError( +ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4778.385408,21498.42944,0.0,20851.982336,20236.60032,s,10,15.667565917968751,1.566756591796875,0.0018490956286824453,1.5662525024414062,1.5679145385742188,1.569856866455078,1.5714107287597656,"[1.565250732421875, 1.566271728515625, 1.5669622802734375, 1.56748291015625, 1.5658883056640625, 1.565689697265625, 1.56708056640625, 1.5662332763671876, 1.5717991943359375, 1.5649072265625]",tokens/s,163.39487661347562,kWh,1.848894233504931e-05,1.0131964364536543e-05,9.316793564540516e-05,0.000121788842344991,tokens/kWh,2101998.796201949,MB,4784.119808,21498.42944,0.0,20851.982336,20339.706368,s,10,881.343109375,88.1343109375,0.002956912666877524,88.1353125,88.13739453125,88.137447265625,88.137489453125,"[88.129875, 88.130765625, 88.1368203125, 88.137265625, 88.1375, 88.1343046875, 88.1363203125, 88.13146875, 88.1373828125, 88.13140625]",tokens/s,0.7148180921806506,kWh,0.0010404256961577469,0.0005702452815513971,0.005204286163425598,0.006814957141134741,tokens/kWh,9244.37215015413,,s,629,894.0462337646492,1.4213771601981693,0.18617884218031844,1.398877197265625,1.3992523681640625,1.3995548828125002,2.96542931640625,"[1.39861083984375, 1.398836181640625, 1.39867138671875, 1.3985986328125, 1.3985372314453124, 1.3985423583984375, 1.3984840087890624, 1.3988331298828125, 1.3987451171875, 1.3984296875, 1.398603759765625, 1.39862939453125, 1.398592529296875, 1.39898779296875, 1.399804931640625, 1.3989600830078126, 1.3987215576171874, 1.3987052001953124, 1.3989468994140626, 1.39868359375, 1.398623291015625, 1.39879931640625, 1.3986170654296874, 1.398488037109375, 1.399183349609375, 1.3986939697265626, 1.39886279296875, 1.3986058349609376, 1.3987020263671874, 1.3987952880859376, 1.398541259765625, 1.3985986328125, 1.3989161376953125, 1.398920166015625, 1.3990419921875, 1.39896826171875, 1.3991865234375, 1.398741943359375, 1.398930419921875, 1.3989847412109375, 1.3989520263671875, 1.3987255859375, 1.398935546875, 1.39886083984375, 1.3990206298828125, 1.3991453857421876, 1.39919775390625, 1.3996278076171875, 1.3989007568359375, 1.3989898681640625, 1.3989866943359375, 1.3988802490234375, 1.3991854248046875, 1.399235595703125, 1.398929443359375, 1.3990225830078125, 1.3991126708984376, 1.398993896484375, 1.39955615234375, 1.3990440673828124, 1.3991434326171874, 1.39884033203125, 2.965783447265625, 1.3988966064453126, 1.398802490234375, 1.3985321044921875, 1.39858740234375, 1.3986273193359375, 1.3983150634765624, 1.3987532958984374, 1.3986478271484375, 1.3984962158203125, 1.3987645263671875, 1.3986058349609376, 1.3986898193359374, 1.39869189453125, 1.3983323974609374, 1.398645751953125, 1.3987481689453125, 1.39890478515625, 1.3990318603515626, 1.4008525390625, 1.3985054931640626, 1.3984737548828126, 1.3986109619140625, 1.3987440185546876, 1.3986488037109375, 1.3987255859375, 1.398550537109375, 1.3988331298828125, 1.3985792236328125, 1.398634521484375, 1.3985382080078126, 1.3987164306640625, 1.3986856689453124, 1.39848291015625, 1.3987113037109375, 1.3989376220703125, 1.398762451171875, 1.3987420654296876, 1.3987840576171875, 1.3990584716796874, 1.399045166015625, 1.401311279296875, 1.39856591796875, 1.3989765625, 1.398443115234375, 1.398836181640625, 1.39945166015625, 1.3988555908203124, 1.3991495361328126, 1.3991956787109374, 1.39893359375, 1.399204833984375, 1.3988291015625, 1.399236572265625, 1.399004150390625, 1.399005126953125, 1.3992908935546875, 1.3988536376953125, 1.399035888671875, 1.3992274169921874, 1.3993287353515624, 1.399700439453125, 1.3991219482421875, 2.966549560546875, 1.39862939453125, 1.39866015625, 1.399478271484375, 1.399192626953125, 1.3989632568359376, 1.398771728515625, 1.3986805419921875, 1.3987911376953126, 1.3989549560546874, 1.3987236328125, 1.3990697021484375, 1.3988494873046875, 1.3987430419921876, 1.3989058837890624, 1.39859765625, 1.3985792236328125, 1.398561767578125, 1.3986058349609376, 1.3985843505859374, 1.3987799072265625, 1.3989171142578125, 1.39850244140625, 1.40042041015625, 1.3986375732421874, 1.398804443359375, 1.398970458984375, 1.3991331787109376, 1.3986907958984376, 1.3987430419921876, 1.3988720703125, 1.399077880859375, 1.398720458984375, 1.3989119873046876, 1.398973388671875, 1.39915576171875, 1.39900927734375, 1.3989549560546874, 1.398836181640625, 1.3991864013671875, 1.39887109375, 1.39875537109375, 1.4007685546875, 1.39913525390625, 1.39900927734375, 1.3990001220703125, 1.39928369140625, 1.3991044921875, 1.3991658935546876, 1.3989150390625, 1.39884130859375, 1.3990972900390626, 1.3989857177734375, 1.3992447509765624, 1.399131103515625, 1.399405517578125, 1.3992642822265624, 1.3992550048828125, 1.3990594482421874, 1.3992591552734375, 1.3989908447265624, 1.39915771484375, 1.3989212646484375, 2.96693359375, 1.3985638427734375, 1.398752197265625, 1.398823974609375, 1.398572021484375, 1.3988055419921874, 1.398634521484375, 1.3986221923828126, 1.3986856689453124, 1.39827197265625, 1.398539306640625, 1.3988157958984375, 1.4004111328125, 1.39886279296875, 1.3987327880859375, 1.39856689453125, 1.3985545654296876, 1.39869287109375, 1.3988526611328125, 1.39871533203125, 1.39894580078125, 1.3987052001953124, 1.398551513671875, 1.3985792236328125, 1.3989222412109374, 1.399103515625, 1.398645751953125, 1.3989754638671874, 1.3987388916015624, 1.3983846435546874, 1.398825927734375, 1.3987491455078125, 1.3987706298828124, 1.398856689453125, 1.3985126953125, 1.3989344482421875, 1.3986201171875, 1.3990635986328126, 1.3985853271484374, 1.3990738525390625, 1.399287841796875, 1.39908203125, 1.3989232177734374, 1.3990615234375, 1.403336669921875, 1.399041015625, 1.398888427734375, 1.3992376708984375, 1.399426025390625, 1.3991126708984376, 1.39894677734375, 1.39915673828125, 1.3997025146484374, 1.3994208984375, 1.39913623046875, 1.398823974609375, 1.3998172607421875, 1.399552978515625, 1.3991270751953124, 1.3991854248046875, 1.399057373046875, 1.3992447509765624, 1.3991864013671875, 2.965474365234375, 1.3988095703125, 1.3988209228515625, 1.3986990966796875, 1.39848193359375, 1.3987869873046874, 1.398802490234375, 1.3984625244140625, 1.3988065185546874, 1.3984798583984375, 1.3986632080078125, 1.39890380859375, 1.3988720703125, 1.398951904296875, 1.3987686767578125, 1.398593505859375, 1.3987645263671875, 1.401100341796875, 1.398867919921875, 1.3988475341796875, 1.398614990234375, 1.398646728515625, 1.3986273193359375, 1.39869287109375, 1.3989990234375, 1.3990072021484374, 1.398619140625, 1.3986314697265625, 1.398825927734375, 1.3988597412109376, 1.3990645751953126, 1.398625244140625, 1.3988720703125, 1.3991177978515625, 1.398873046875, 1.3989488525390625, 1.3991280517578124, 1.398972412109375, 1.398724609375, 1.3990164794921875, 1.39892431640625, 1.398846435546875, 1.399014404296875, 1.398878173828125, 1.3987020263671874, 1.3988331298828125, 1.3991075439453124, 1.3988905029296874, 1.3992171630859376, 1.39901025390625, 1.3989990234375, 1.3992232666015625, 1.4028953857421875, 1.3991341552734375, 1.3992559814453125, 1.39980908203125, 1.399646240234375, 1.399251953125, 1.3990379638671875, 1.3988916015625, 1.39907373046875, 1.399131103515625, 1.398908935546875, 2.96527978515625, 1.3986693115234374, 1.3986282958984375, 1.3985587158203125, 1.39848193359375, 1.3986611328125, 1.3984716796875, 1.3986427001953126, 1.3985526123046874, 1.3987379150390624, 1.3991270751953124, 1.3989129638671876, 1.3986580810546876, 1.3988095703125, 1.39888232421875, 1.398624267578125, 1.3986683349609375, 1.3986641845703125, 1.3986754150390626, 1.398550537109375, 1.398556640625, 1.3986058349609376, 1.398283203125, 1.3985740966796876, 1.398614990234375, 1.398750244140625, 1.3989530029296875, 1.3992540283203125, 1.3987236328125, 1.39867138671875, 1.3988074951171876, 1.398635498046875, 1.398919189453125, 1.3987952880859376, 1.3987451171875, 1.3988116455078126, 1.3987747802734376, 1.398972412109375, 1.404094482421875, 1.3990953369140624, 1.39894580078125, 1.3994281005859375, 1.3991240234375, 1.399256103515625, 1.398873046875, 1.39911572265625, 1.3989263916015624, 1.399004150390625, 1.39951513671875, 1.3991126708984376, 1.3989796142578126, 1.39894677734375, 1.3989365234375, 1.399294921875, 1.3989478759765626, 1.398964111328125, 1.3990972900390626, 1.3990020751953125, 1.3989478759765626, 1.399329833984375, 1.39915673828125, 1.3990286865234376, 1.3991751708984375, 2.966761474609375, 1.3985423583984375, 1.39884130859375, 1.3985556640625, 1.3985556640625, 1.3988095703125, 1.39877783203125, 1.3987860107421874, 1.3986641845703125, 1.39853515625, 1.3985361328125, 1.398656982421875, 1.3986968994140625, 1.39901953125, 1.3986212158203124, 1.3989027099609375, 1.3986119384765625, 1.398583251953125, 1.398730712890625, 1.3989017333984375, 1.398697998046875, 1.398677490234375, 1.3984931640625, 1.401703369140625, 1.3986641845703125, 1.39890185546875, 1.3988289794921875, 1.3988167724609375, 1.398576171875, 1.3986744384765626, 1.398751220703125, 1.3987318115234375, 1.398581298828125, 1.39869384765625, 1.39877783203125, 1.3993779296875, 1.399005126953125, 1.3989232177734374, 1.3988587646484374, 1.39894677734375, 1.39911474609375, 1.399098388671875, 1.398730712890625, 1.39896826171875, 1.398935546875, 1.39915576171875, 1.39913623046875, 1.3991424560546875, 1.3990655517578126, 1.3991065673828125, 1.3990860595703125, 1.399183349609375, 1.3989744873046874, 1.39915576171875, 1.3991240234375, 1.399055419921875, 1.402797119140625, 1.3991197509765625, 1.398992919921875, 1.3993922119140625, 1.3990738525390625, 1.399083984375, 1.3990860595703125, 2.965937255859375, 1.3989212646484375, 1.3988382568359374, 1.398813720703125, 1.3986324462890625, 1.398729736328125, 1.3987880859375, 1.3985526123046874, 1.398720458984375, 1.39859765625, 1.39854443359375, 1.3988331298828125, 1.3987318115234375, 1.3988648681640625, 1.3985648193359375, 1.3985484619140625, 1.400943603515625, 1.3987593994140626, 1.398794189453125, 1.3986282958984375, 1.3985535888671874, 1.398667236328125, 1.3985853271484374, 1.3986427001953126, 1.398782958984375, 1.3989068603515624, 1.3985126953125, 1.39886083984375, 1.398930419921875, 1.39879931640625, 1.3986181640625, 1.3988238525390626, 1.3986898193359374, 1.3985914306640626, 1.3986754150390626, 1.3989530029296875, 1.3988914794921874, 1.3989222412109374, 1.3990400390625, 1.399320556640625, 1.398939697265625, 1.3989180908203125, 1.3988443603515626, 1.3991177978515625, 1.3987430419921876, 1.3990850830078125, 1.3988494873046875, 1.3990185546875, 1.399118896484375, 1.398983642578125, 1.399162841796875, 1.3992745361328125, 1.398998046875, 1.398919189453125, 1.3992181396484376, 1.399088134765625, 1.39907177734375, 1.3991177978515625, 1.399078857421875, 1.399394287109375, 1.399125, 1.3990706787109375, 1.3991875, 2.971739013671875, 1.399003173828125, 1.3988106689453126, 1.3985955810546875, 1.3986651611328125, 1.3988382568359374, 1.3985577392578126, 1.39854541015625, 1.3985228271484376, 1.398729736328125, 1.3986324462890625, 1.3985679931640624, 1.3984625244140625, 1.399056396484375, 1.398541259765625, 1.398877197265625, 1.39889453125, 1.4004193115234376, 1.3987532958984374, 1.3991875, 1.39894482421875, 1.39869384765625, 1.3984102783203125, 1.3985894775390626, 1.3989273681640626, 1.3987164306640625, 1.398676513671875, 1.3989805908203126, 1.398752197265625, 1.3987369384765624, 1.3986201171875, 1.3987164306640625, 1.3989530029296875, 1.3989693603515625, 1.3985843505859374, 1.3986744384765626, 1.3989273681640626, 1.3991177978515625, 1.398738037109375, 1.39888427734375, 1.399014404296875, 1.3989857177734375, 1.3989949951171874, 1.3990389404296875, 1.3988638916015625, 1.3990963134765626, 1.39890380859375, 1.3990758056640624, 1.3991700439453125, 1.399078857421875, 1.399034912109375, 1.3990799560546876, 1.398781982421875, 1.3989775390625, 1.3993604736328125, 1.3990277099609374, 1.3989385986328124, 1.3992847900390626, 1.4042296142578126, 1.3992314453125, 1.3990748291015624, 1.399125, 1.39915673828125, 2.9653134765625, 1.3987911376953126, 1.3988863525390625, 1.3984849853515624, 1.398365234375, 1.3986478271484375, 1.3986856689453124, 1.3987255859375, 1.398328369140625, 1.3986314697265625, 1.3986007080078124, 1.398823974609375, 1.398498291015625, 1.39887109375, 1.3987860107421874, 1.3986385498046876, 1.3992396240234375, 1.3990430908203124, 1.3988321533203125, 1.3987052001953124, 1.3982330322265626, 1.3985146484375, 1.3988536376953125, 1.3986785888671875, 1.398445068359375, 1.39894677734375, 1.4012242431640625, 1.3989908447265624, 1.3988177490234375, 1.3987337646484375, 1.39881982421875, 1.3986683349609375, 1.3987255859375, 1.3986907958984376, 1.3990216064453125, 1.3991240234375, 1.398856689453125, 1.399024658203125, 1.398556640625, 1.3989744873046874, 1.3988157958984375, 1.3987318115234375, 1.3989478759765626, 1.3992376708984375, 1.398750244140625, 1.3990389404296875, 1.39900927734375, 1.399330810546875, 1.3993922119140625, 1.39889453125, 1.399046142578125, 1.3991455078125, 1.39875537109375, 1.3991259765625, 1.39901025390625, 1.399046142578125, 1.399300048828125, 1.39921923828125, 1.399109619140625, 1.3992078857421875, 1.39896630859375, 1.399267333984375, 1.3988720703125]",tokens/s,0.7035430341800197,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4051,21 +4302,11 @@ ChildProcessError: Traceback (most recent call last): response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) -requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json +requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json The above exception was the direct cause of the following exception: Traceback (most recent call last): - File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 399, in cached_file - resolved_file = hf_hub_download( - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn - return fn(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download - return _hf_hub_download_to_cache_dir( - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir - _raise_on_head_call_error(head_call_error, force_download, local_files_only) - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error - raise head_call_error File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn @@ -4076,19 +4317,104 @@ Traceback (most recent call last): response = _request_wrapper( File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper hf_raise_for_status(response) - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status - raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12f0-35b129ad69b747d21cd32cf7;a6e59050-7dd4-4e35-8a81-efebb25f46c4) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status + raise HfHubHTTPError(message, response=response) from e +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4aad-04710e0424df0f9646b94647;b7c59044-3a5a-4a4c-82da-e7be2ec0c625) -Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. -Please make sure you specified the correct `repo_id` and `repo_type`. -If you are trying to access a private or gated repo, make sure you are authenticated. +403 Forbidden: Authorization error.. +Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. +If you are trying to create or update content,make sure you have a token with the `write` role. The above exception was the direct cause of the following exception: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target - report = worker(*worker_args) + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 399, in cached_file + resolved_file = hf_hub_download( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download + return _hf_hub_download_to_cache_dir( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir + _raise_on_head_call_error(head_call_error, force_download, local_files_only) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error + raise LocalEntryNotFoundError( +huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ + super().__init__(config) + File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ + self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) + File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config + return AutoConfig.from_pretrained(model, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict + resolved_config_file = cached_file( + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 442, in cached_file + raise EnvironmentError( +OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. +Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status + response.raise_for_status() + File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status + raise HTTPError(http_error_msg, response=self) +requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 399, in cached_file + resolved_file = hf_hub_download( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download + return _hf_hub_download_to_cache_dir( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir + _raise_on_head_call_error(head_call_error, force_download, local_files_only) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error + raise head_call_error + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error + metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata + r = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper + response = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper + hf_raise_for_status(response) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status + raise RepositoryNotFoundError(message, response) from e +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4516-6d05aa05681bc26e5924c7af;30758e2a-f522-45b2-a338-ad8938dea775) + +Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. +Please make sure you specified the correct `repo_id` and `repo_type`. +If you are trying to access a private or gated repo, make sure you are authenticated. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ @@ -4168,7 +4494,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee5e-750bb987115b381158b85187;b19d9c6b-dc4f-4997-8ae9-2d5bf7dd922e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3f6e-11328848271c4d1d5f884916;5dbf16ef-1f2a-4c05-bd54-8a976356af29) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -4351,7 +4677,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1281-03dc1bbc49df777843d67764;8c4b77ac-57a5-4d4d-a7c5-ffecaa3fab0b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44c7-677d91f04304b84e63799599;2932b3c9-4a54-42d1-a627-57a481b40149) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4480,7 +4806,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664beee5-2aa091d82d6239b56a51444c;cb871370-80ee-45dd-846e-f112244abffc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d3fe3-3fc56abd621e4d9869ae7673;4b9b9254-f999-423e-8fa1-d6498132504c) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4609,7 +4935,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c140e-582e1f9e77cafc7b7b7f2eae;69d7eeb3-9e2b-42b4-a69d-627b3a55b52e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4626-4f1e8baa0bb78ed54ac396e6;c554b746-f9a4-417e-878c-5a1bddfe1ad5) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4639,6 +4965,35 @@ Traceback (most recent call last): OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1478, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1644, in _check_and_enable_sdpa + raise ValueError( +ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5919.543296,7053.246464,0.0,6406.79936,6098.60864,s,10,4.240695465087891,0.42406954650878903,0.0008262784375631619,0.4239071960449219,0.4251891204833984,0.4252073989868164,0.42522202178955076,"[0.4233236083984375, 0.42293777465820315, 0.424222412109375, 0.42434298706054685, 0.4251079406738281, 0.42321951293945315, 0.42518505859375, 0.42359197998046877, 0.42353851318359376, 0.4252256774902344]",tokens/s,603.6745673146193,kWh,4.997010294485975e-06,2.7381218011896633e-06,2.5116837223083482e-05,3.285196931875912e-05,tokens/kWh,7792531.324867,MB,5919.543296,7055.343616,0.0,6408.896512,6157.952,s,10,234.877408203125,23.487740820312503,0.0010175713958813057,23.48770703125,23.489005078125,23.489256445312503,23.489457539062503,"[23.48759375, 23.488662109375, 23.486078125, 23.486736328125, 23.48778515625, 23.486681640625, 23.48773828125, 23.4895078125, 23.48894921875, 23.48767578125]",tokens/s,2.6822503058922034,kWh,0.0002773065141202123,0.00015198753169179781,0.0013534979323345177,0.0017827919781465277,tokens/kWh,35337.83008463931,,s,629,238.32171221923846,0.37888984454568886,0.05040710039151534,0.37281381225585936,0.372976025390625,0.3730767883300781,0.7965521459960938,"[0.3727319030761719, 0.3726284790039063, 0.3727953796386719, 0.3728015441894531, 0.3728506774902344, 0.373243896484375, 0.37265921020507814, 0.37258343505859376, 0.372822021484375, 0.37271755981445315, 0.37292340087890624, 0.37277798461914063, 0.3727738952636719, 0.37282098388671875, 0.3728598937988281, 0.3726346130371094, 0.3728527221679688, 0.3727595520019531, 0.3726663818359375, 0.3727462463378906, 0.37283123779296873, 0.3731046447753906, 0.37285580444335936, 0.37273703002929687, 0.37288037109375, 0.3727964172363281, 0.37278616333007814, 0.37287628173828125, 0.3728404541015625, 0.3728865356445312, 0.37287014770507815, 0.3727595520019531, 0.3728793640136719, 0.37274215698242186, 0.372790283203125, 0.3727513732910156, 0.3727032470703125, 0.372864013671875, 0.3729018859863281, 0.3727185974121094, 0.3727544250488281, 0.37287527465820314, 0.37272268676757814, 0.37280767822265626, 0.3727042541503906, 0.37269503784179686, 0.3726561279296875, 0.37276568603515625, 0.3727431640625, 0.37280667114257815, 0.37291009521484375, 0.37271450805664064, 0.37288858032226563, 0.372738037109375, 0.3729776611328125, 0.3729377136230469, 0.37281280517578125, 0.3728988037109375, 0.3729018859863281, 0.37289166259765627, 0.3729623107910156, 0.3727882385253906, 0.7962152709960938, 0.37294692993164064, 0.37268685913085936, 0.3732090759277344, 0.37280050659179687, 0.3727431640625, 0.37271038818359375, 0.3725608825683594, 0.3727677307128906, 0.37274420166015626, 0.37280667114257815, 0.37268069458007813, 0.3728465881347656, 0.37274215698242186, 0.37298892211914064, 0.3730821228027344, 0.37262539672851563, 0.3729490051269531, 0.37270733642578124, 0.3729428405761719, 0.37284146118164063, 0.3727237243652344, 0.3727564697265625, 0.37272064208984373, 0.372716552734375, 0.37280255126953127, 0.37300018310546873, 0.37303604125976564, 0.37281585693359376, 0.3727411193847656, 0.37318963623046875, 0.3728107604980469, 0.37279745483398435, 0.37266021728515625, 0.37270733642578124, 0.372701171875, 0.372843505859375, 0.37275341796875, 0.37277798461914063, 0.37293362426757815, 0.37299200439453123, 0.37289166259765627, 0.3728107604980469, 0.3728660583496094, 0.37275750732421875, 0.37293362426757815, 0.3727615966796875, 0.37291009521484375, 0.3727472839355469, 0.37279739379882815, 0.372780029296875, 0.37285479736328125, 0.3727738952636719, 0.3728291931152344, 0.37283941650390623, 0.3726684265136719, 0.3727298583984375, 0.37312820434570315, 0.37288345336914064, 0.37287115478515626, 0.3727216491699219, 0.37300634765625, 0.3729704895019531, 0.7973662719726563, 0.3729797058105469, 0.37260287475585935, 0.3728844909667969, 0.37274420166015626, 0.3728384094238281, 0.3728302001953125, 0.3727462463378906, 0.3728947143554687, 0.372822021484375, 0.3726960754394531, 0.37281793212890624, 0.37280767822265626, 0.37274725341796877, 0.37275851440429686, 0.37275546264648435, 0.37281381225585936, 0.3726407775878906, 0.37273599243164063, 0.37275750732421875, 0.3727872009277344, 0.3729070129394531, 0.37279232788085936, 0.3727820739746094, 0.37279745483398435, 0.37273599243164063, 0.3726346130371094, 0.37278616333007814, 0.3728598937988281, 0.3730073547363281, 0.3728271484375, 0.3726192626953125, 0.3728517150878906, 0.3727738952636719, 0.372664306640625, 0.37266943359375, 0.3727882385253906, 0.37283328247070313, 0.37273907470703127, 0.37279232788085936, 0.3728988037109375, 0.372748291015625, 0.37272064208984373, 0.37276568603515625, 0.37278411865234373, 0.372864013671875, 0.3727667236328125, 0.3727349853515625, 0.3727267761230469, 0.3727349853515625, 0.37279949951171876, 0.3727687683105469, 0.3728865356445312, 0.37275341796875, 0.37291827392578125, 0.3728506774902344, 0.3728271484375, 0.3728353271484375, 0.3728240661621094, 0.3727400817871094, 0.37283328247070313, 0.3728558349609375, 0.37283633422851564, 0.7973109741210938, 0.37268069458007813, 0.37289984130859377, 0.37285580444335936, 0.37260391235351564, 0.37292340087890624, 0.37282815551757814, 0.3726295166015625, 0.372527099609375, 0.3725752258300781, 0.3726981201171875, 0.3726827392578125, 0.3728598937988281, 0.3727032470703125, 0.37268173217773437, 0.3727912902832031, 0.37256805419921873, 0.3727329406738281, 0.37288140869140624, 0.37288140869140624, 0.372642822265625, 0.372927490234375, 0.3725854797363281, 0.37274520874023437, 0.37272882080078124, 0.3727319030761719, 0.37270016479492185, 0.372706298828125, 0.3729326171875, 0.37279031372070315, 0.3727431335449219, 0.37302783203125, 0.37290802001953127, 0.37291928100585936, 0.3728486328125, 0.37315277099609373, 0.3726929931640625, 0.3729797058105469, 0.372811767578125, 0.37263873291015626, 0.37270220947265625, 0.37277593994140623, 0.37300634765625, 0.37322650146484376, 0.3730268249511719, 0.37290496826171876, 0.37277490234375, 0.37283123779296873, 0.3728322448730469, 0.3726878662109375, 0.37291217041015623, 0.37272674560546876, 0.37288345336914064, 0.37274215698242186, 0.37278616333007814, 0.3730687866210938, 0.3727687683105469, 0.37286911010742185, 0.3726080017089844, 0.3728302001953125, 0.37273907470703127, 0.3728445434570313, 0.37293875122070314, 0.7982294921875, 0.3727319030761719, 0.3726755981445313, 0.37270016479492185, 0.3726704711914062, 0.37275851440429686, 0.37284661865234375, 0.3728486022949219, 0.3725230102539063, 0.3727196044921875, 0.37268991088867187, 0.3726438293457031, 0.37304217529296874, 0.37271755981445315, 0.3728660583496094, 0.37264794921875, 0.37272268676757814, 0.37281585693359376, 0.3728865356445312, 0.37282098388671875, 0.3728824462890625, 0.3727492980957031, 0.3727462463378906, 0.37274215698242186, 0.3726561279296875, 0.3726356506347656, 0.372885498046875, 0.37290802001953127, 0.3727718505859375, 0.3729284973144531, 0.3729981384277344, 0.3728384094238281, 0.37281484985351565, 0.3727431640625, 0.3732777099609375, 0.3729213562011719, 0.37280255126953127, 0.372664306640625, 0.3727872009277344, 0.37281689453125, 0.37281381225585936, 0.37278411865234373, 0.372864013671875, 0.3726704711914062, 0.37273599243164063, 0.3728660888671875, 0.3728148193359375, 0.3728404541015625, 0.3728240661621094, 0.37293057250976563, 0.372780029296875, 0.3727892456054688, 0.37296127319335937, 0.3728660583496094, 0.37288345336914064, 0.37285479736328125, 0.3731251220703125, 0.3729090576171875, 0.3728824462890625, 0.3731517333984375, 0.37291110229492186, 0.37286810302734374, 0.3728045959472656, 0.7964098510742188, 0.37276364135742185, 0.3725946960449219, 0.372684814453125, 0.37267251586914063, 0.3727667236328125, 0.3728271484375, 0.3728721923828125, 0.3726366577148438, 0.37268991088867187, 0.3727216491699219, 0.3726376953125, 0.37253939819335935, 0.37272268676757814, 0.37270733642578124, 0.3728650207519531, 0.3727216491699219, 0.3726919555664063, 0.37269403076171875, 0.37291009521484375, 0.372748291015625, 0.37273086547851564, 0.3730401306152344, 0.3728937072753906, 0.37287014770507815, 0.3727492980957031, 0.3726612548828125, 0.3728650207519531, 0.3728937072753906, 0.37268484497070314, 0.3728834228515625, 0.37263360595703127, 0.37287115478515626, 0.37277285766601564, 0.3727492980957031, 0.37283428955078124, 0.37281689453125, 0.3727329406738281, 0.37286911010742185, 0.37293057250976563, 0.37295309448242187, 0.37273599243164063, 0.3726714782714844, 0.372853759765625, 0.372885498046875, 0.3726960754394531, 0.37281689453125, 0.37285580444335936, 0.3727492980957031, 0.3728271484375, 0.37304421997070314, 0.37286810302734374, 0.3729715270996094, 0.37298175048828125, 0.37288345336914064, 0.37278616333007814, 0.3728947143554687, 0.372811767578125, 0.3727247314453125, 0.37280767822265626, 0.372864013671875, 0.3729776611328125, 0.37291110229492186, 0.7983513793945313, 0.3727083740234375, 0.3728393859863281, 0.37292544555664064, 0.37264178466796877, 0.37277081298828124, 0.3727738952636719, 0.3726858215332031, 0.3728240661621094, 0.372864013671875, 0.3728250732421875, 0.37289779663085937, 0.3727339477539062, 0.3727247314453125, 0.37256192016601564, 0.3728773193359375, 0.3727595520019531, 0.37272882080078124, 0.3727933349609375, 0.3726561279296875, 0.3726131286621094, 0.3728271484375, 0.37290087890625, 0.372822021484375, 0.3727329406738281, 0.37269503784179686, 0.37281381225585936, 0.3728189392089844, 0.3728660583496094, 0.3726981201171875, 0.37275033569335936, 0.37284249877929687, 0.3728947143554687, 0.37297457885742186, 0.37266738891601564, 0.3727513732910156, 0.37276364135742185, 0.3728445434570313, 0.3727933349609375, 0.3727872009277344, 0.37278616333007814, 0.37297561645507815, 0.3728322448730469, 0.3727349853515625, 0.37272882080078124, 0.37285784912109377, 0.3728322448730469, 0.37276263427734374, 0.3729284973144531, 0.3729090576171875, 0.37265716552734374, 0.37291827392578125, 0.37290802001953127, 0.37275851440429686, 0.3728988037109375, 0.37278616333007814, 0.3728107604980469, 0.372959228515625, 0.3728937072753906, 0.3731927185058594, 0.37290597534179687, 0.3732142028808594, 0.37290496826171876, 0.7966074829101563, 0.37313534545898436, 0.37330740356445313, 0.3728291931152344, 0.37276364135742185, 0.37275851440429686, 0.37281484985351565, 0.37298175048828125, 0.3726929931640625, 0.37274420166015626, 0.37266738891601564, 0.3726909484863281, 0.3729346618652344, 0.37308416748046874, 0.3726612548828125, 0.37258956909179686, 0.3727882385253906, 0.3728455810546875, 0.3727605895996094, 0.3727247314453125, 0.3727267761230469, 0.37292340087890624, 0.3728670654296875, 0.3726632995605469, 0.37259060668945315, 0.3727667236328125, 0.37290802001953127, 0.37288345336914064, 0.37280767822265626, 0.3733063659667969, 0.3727595520019531, 0.3728650207519531, 0.37298587036132813, 0.37279745483398435, 0.37283123779296873, 0.3728670654296875, 0.3728363647460938, 0.37290087890625, 0.3727032470703125, 0.3728517150878906, 0.37305035400390624, 0.37284967041015626, 0.37284146118164063, 0.37265921020507814, 0.37265716552734374, 0.3727492980957031, 0.3726510009765625, 0.37291827392578125, 0.3729428405761719, 0.37287628173828125, 0.3729438781738281, 0.3728189392089844, 0.37292440795898435, 0.3726991271972656, 0.37287832641601565, 0.372864013671875, 0.37295822143554686, 0.37290087890625, 0.3730196533203125, 0.3729121398925781, 0.3728937072753906, 0.37303091430664065, 0.37285580444335936, 0.7966361694335937, 0.37282815551757814, 0.3726612548828125, 0.37303909301757815, 0.3726960754394531, 0.3727933349609375, 0.37255474853515624, 0.37293362426757815, 0.3726919555664063, 0.37298892211914064, 0.3726919555664063, 0.3728660583496094, 0.37276467895507814, 0.3728291931152344, 0.37275341796875, 0.37284249877929687, 0.3728517150878906, 0.3732305908203125, 0.37283428955078124, 0.3730483093261719, 0.37289166259765627, 0.37274215698242186, 0.37266534423828124, 0.3727605895996094, 0.3726632995605469, 0.37278616333007814, 0.3726796875, 0.3727185974121094, 0.3727339477539062, 0.37286093139648435, 0.3729520568847656, 0.37278515625, 0.37287628173828125, 0.3727820739746094, 0.3728742370605469, 0.3729571838378906, 0.37285580444335936, 0.3728363647460938, 0.3727677307128906, 0.37293878173828127, 0.37301858520507813, 0.3728742370605469, 0.3727247314453125, 0.37286093139648435, 0.3728087158203125, 0.3728445434570313, 0.3729879150390625, 0.37288958740234374, 0.3727687683105469, 0.37274420166015626, 0.3729284973144531, 0.37284249877929687, 0.37293057250976563, 0.372738037109375, 0.37319064331054685, 0.37287527465820314, 0.3728773193359375, 0.3728742370605469, 0.3728824462890625, 0.3727544250488281, 0.37286297607421875, 0.3728527221679688, 0.3727605895996094, 0.7982438354492187, 0.3730882568359375, 0.37265921020507814, 0.37288345336914064, 0.37274420166015626, 0.37261004638671874, 0.3727042541503906, 0.37293875122070314, 0.3731312561035156, 0.3727555236816406, 0.37273593139648437, 0.37268173217773437, 0.372790283203125, 0.37274520874023437, 0.37270938110351565, 0.3725823974609375, 0.3725987854003906, 0.37262335205078123, 0.3727462463378906, 0.37304421997070314, 0.3727810668945313, 0.37286911010742185, 0.3727339477539062, 0.37286093139648435, 0.3728322448730469, 0.37264691162109376, 0.37285479736328125, 0.37281689453125, 0.3731998901367187, 0.3728137817382812, 0.37271450805664064, 0.37277593994140623, 0.37273086547851564, 0.3728875427246094, 0.37279949951171876, 0.372738037109375, 0.3729039306640625, 0.37285888671875, 0.3728506774902344, 0.37268685913085936, 0.37275546264648435, 0.37277593994140623, 0.3727124328613281, 0.37312103271484376, 0.37275341796875, 0.3727995300292969, 0.37268988037109374, 0.3729326171875, 0.3728322448730469, 0.37282611083984374, 0.3728721923828125, 0.3729438781738281, 0.37283123779296873, 0.37274520874023437, 0.3728619384765625, 0.37287628173828125, 0.37280767822265626, 0.37277081298828124, 0.3730196533203125, 0.37287628173828125, 0.37285784912109377, 0.3729295349121094, 0.3729520568847656]",tokens/s,2.639289530705313,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -4680,7 +5035,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c10b6-030f3fc9772fe08d3f536efe;9b9e3640-f033-4593-8a75-6287524bfa29) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4305-668df2e36ba1168b74ea76bf;af45b1f6-1d9a-4503-8473-3ab6d31de4b9) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4710,6 +5065,35 @@ Traceback (most recent call last): OSError: s is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 563, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1478, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1644, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch @@ -5032,7 +5416,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1068-5a5b84d3761d512c1aecc99f;f52b74e2-a014-4695-a25b-05d1b56874a0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d42ad-7330e61556192e4568262801;593af40c-5843-4590-8b74-df0a2fff171d) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5102,7 +5486,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c14d4-191134f53b2f7b7723f7a4a1;8fdaed70-e5b9-4092-83fe-b9f34d9012d0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46ea-61d76feb58130df55d641f84;7f9e84b9-71a3-44d7-85f1-712870cd7708) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5172,7 +5556,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1485-563deef42e57aca472a193a5;a1119ff8-1194-4d20-9813-aca12df8c791) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d469b-66f77aa61a43508b5c626b31;358c1803-3172-4738-9732-66dd1e6b88db) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5300,7 +5684,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1384-4fa9117209b9d5bf7a84fde1;c8bf9dca-b705-41ac-a9d2-5897c599f799) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d459d-21df1c6879405c0c5694daf4;a7e6363e-6a7b-4817-9db4-87ad0adb1d6e) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5371,7 +5755,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c13fa-6dd95a880c9b1d34669890aa;034afad5-9398-4eca-9c2b-fd28458b078d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4612-6c529ab8327c4f0f237f558e;776a7514-0aef-4b79-a89e-a3094ef92247) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5428,7 +5812,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpj5h86k2s/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpifhj2jeo/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1235.00544,849.870848,0.0,203.423744,184.496128,s,18,0.17785792160034175,0.009880995644463434,0.00028849658241711547,0.009792351722717285,0.010203827285766602,0.010547976160049438,0.010567451219558715,"[0.01054368019104004, 0.010040608406066895, 0.00977449607849121, 0.009805983543395996, 0.010572319984436035, 0.009761152267456054, 0.009838656425476074, 0.009730496406555177, 0.009720447540283204, 0.009830143928527832, 0.009709216117858887, 0.009674847602844238, 0.0093340482711792, 0.009970623970031739, 0.010058176040649415, 0.00970019245147705, 0.01001411247253418, 0.009778719902038574]",tokens/s,25908.32029598587,kWh,1.152578154553566e-07,6.315625308503158e-08,2.135530163707643e-07,3.919670849111525e-07,tokens/kWh,653116064.7278017,MB,1235.33312,849.870848,0.0,203.423744,193.844224,s,18,10.402695068359375,0.577927503797743,0.008339520037857094,0.5779442443847655,0.5829411499023438,0.5867088104248047,0.5985698382568359,"[0.6015350952148437, 0.5610368041992188, 0.577316650390625, 0.5771664428710938, 0.5780480346679687, 0.5784367065429687, 0.58244775390625, 0.5793241577148438, 0.5705869140625, 0.5801618041992187, 0.5777095947265625, 0.5614797973632812, 0.5778404541015625, 0.5821588134765625, 0.5775224609375, 0.577691650390625, 0.5781395263671875, 0.5840924072265625]",tokens/s,109.01021250244577,kWh,6.8063746160841315e-06,3.729610197448205e-06,1.1949573998074713e-05,2.248555881160705e-05,tokens/kWh,2801798.2798577095,,s,1133,10.560176177978532,0.009320543846406458,0.0012759106372352618,0.009179136276245118,0.00931942367553711,0.009467494201660154,0.01931186187744141,"[0.010136575698852538, 0.010117119789123535, 0.010518527984619141, 0.011207679748535156, 0.010182656288146973, 0.010301440238952637, 0.01010585594177246, 0.009370623588562011, 0.009408512115478516, 0.009662464141845703, 0.010227711677551269, 0.009968640327453614, 0.009417728424072265, 0.009302016258239745, 0.009689087867736817, 0.00939417552947998, 0.009448448181152343, 0.009433088302612304, 0.009310208320617675, 0.009481216430664062, 0.00941055965423584, 0.009150464057922364, 0.009300992012023926, 0.009226240158081055, 0.009285632133483887, 0.009370623588562011, 0.009431039810180664, 0.009575424194335937, 0.009665535926818849, 0.009380864143371583, 0.0092293119430542, 0.00941875171661377, 0.009263104438781738, 0.009482239723205567, 0.009341952323913574, 0.00931123161315918, 0.009253888130187989, 0.009623552322387695, 0.00926416015625, 0.009291744232177735, 0.009316351890563965, 0.009242624282836913, 0.00931942367553711, 0.009393152236938476, 0.009688128471374511, 0.009570240020751953, 0.009524224281311035, 0.009476096153259277, 0.009392127990722657, 0.009502719879150391, 0.00921497631072998, 0.009432064056396485, 0.00919961643218994, 0.009102335929870605, 0.00894159984588623, 0.00897225570678711, 0.009049087524414063, 0.009372672080993653, 0.00982323169708252, 0.009829376220703125, 0.010189824104309082, 0.009583616256713867, 0.019946495056152345, 0.009302016258239745, 0.00921395206451416, 0.009126912117004395, 0.00923136043548584, 0.009269248008728028, 0.00922316837310791, 0.00919961643218994, 0.009209856033325196, 0.008895487785339355, 0.008706048011779785, 0.008743935585021973, 0.008759296417236329, 0.008770560264587402, 0.00881766414642334, 0.008759296417236329, 0.00869375991821289, 0.00881868839263916, 0.00879308795928955, 0.008857600212097168, 0.008679424285888672, 0.008689663887023925, 0.008700927734375, 0.008813568115234375, 0.008691712379455567, 0.008765439987182617, 0.008687616348266602, 0.008707072257995606, 0.008688639640808106, 0.008725503921508788, 0.00882688045501709, 0.008848383903503418, 0.008791040420532227, 0.00880947208404541, 0.008836095809936523, 0.00879308795928955, 0.008792063713073731, 0.008837120056152344, 0.008756223678588868, 0.00872755241394043, 0.008868864059448242, 0.008816639900207519, 0.008864768028259277, 0.008865792274475098, 0.008829952239990235, 0.00869478416442871, 0.008721407890319824, 0.008704000473022461, 0.008741888046264648, 0.008798208236694336, 0.008839167594909669, 0.008837151527404784, 0.008818655967712403, 0.009099360466003417, 0.009208736419677734, 0.009070591926574707, 0.009049087524414063, 0.009135104179382325, 0.009229472160339355, 0.00910319995880127, 0.009110527992248535, 0.009154560089111329, 0.009095168113708496, 0.019355648040771483, 0.009136128425598144, 0.009048064231872559, 0.009202688217163087, 0.009135104179382325, 0.009116671562194823, 0.009179136276245118, 0.00918015956878662, 0.009082880020141602, 0.009126912117004395, 0.009077759742736816, 0.009078783988952637, 0.00909823989868164, 0.0090316801071167, 0.00920576000213623, 0.009142271995544434, 0.009226240158081055, 0.009197567939758301, 0.009185279846191406, 0.009160703659057617, 0.009063424110412598, 0.009071616172790528, 0.009044992446899413, 0.009179136276245118, 0.009272319793701172, 0.009160799980163574, 0.009067423820495605, 0.009197567939758301, 0.009176063537597656, 0.00908083152770996, 0.009227264404296874, 0.009239551544189453, 0.009094143867492676, 0.009169919967651367, 0.009203712463378906, 0.009226240158081055, 0.009219072341918945, 0.00922214412689209, 0.009215999603271484, 0.00923136043548584, 0.009049087524414063, 0.009220095634460449, 0.009171968460083007, 0.009270272254943847, 0.009129983901977539, 0.009188575744628906, 0.009053983688354493, 0.009191424369812011, 0.009149439811706543, 0.009174015998840332, 0.00919654369354248, 0.009157631874084473, 0.009161727905273438, 0.009043968200683594, 0.009170975685119629, 0.009255904197692871, 0.009160863876342773, 0.009115488052368163, 0.009215999603271484, 0.00920576000213623, 0.009184255599975585, 0.00912179183959961, 0.009195520401000976, 0.019381248474121093, 0.009170944213867188, 0.009167872428894042, 0.009063424110412598, 0.009252863883972168, 0.009083904266357423, 0.009169919967651367, 0.009103360176086426, 0.00918835163116455, 0.00912179183959961, 0.009159680366516113, 0.009136159896850586, 0.009019359588623047, 0.00902143955230713, 0.009095168113708496, 0.009164799690246582, 0.00909823989868164, 0.009163776397705077, 0.009117695808410644, 0.009158656120300293, 0.009170944213867188, 0.009185312271118164, 0.009213919639587403, 0.009047039985656738, 0.009116671562194823, 0.009211903572082519, 0.009175040245056153, 0.009174015998840332, 0.00913920021057129, 0.00919654369354248, 0.009142271995544434, 0.009171968460083007, 0.009154560089111329, 0.009160703659057617, 0.009229375839233398, 0.009232319831848145, 0.009142271995544434, 0.009174015998840332, 0.009191424369812011, 0.009164799690246582, 0.00908902359008789, 0.00909823989868164, 0.009107456207275391, 0.009192447662353515, 0.009194496154785157, 0.009175040245056153, 0.009224191665649414, 0.009218048095703125, 0.00922316837310791, 0.00923136043548584, 0.009201663970947266, 0.009085951805114746, 0.009075712203979493, 0.009074687957763672, 0.009172991752624511, 0.009033727645874023, 0.009083904266357423, 0.009138175964355469, 0.009157631874084473, 0.00914739227294922, 0.00923750400543213, 0.009316351890563965, 0.009243647575378417, 0.02026393508911133, 0.009143296241760255, 0.009061375617980956, 0.00904207992553711, 0.009099103927612305, 0.00919961643218994, 0.009197567939758301, 0.009195520401000976, 0.009160703659057617, 0.009228287696838379, 0.009159680366516113, 0.009232383728027344, 0.009218048095703125, 0.009169919967651367, 0.009163776397705077, 0.00920473575592041, 0.009119744300842286, 0.009210880279541016, 0.00923852825164795, 0.009668607711791993, 0.009360383987426758, 0.009329664230346679, 0.009226240158081055, 0.009163871765136718, 0.009235360145568848, 0.009151488304138184, 0.009088000297546387, 0.00922316837310791, 0.00910848045349121, 0.00919654369354248, 0.009192447662353515, 0.009226240158081055, 0.009247743606567382, 0.009218048095703125, 0.009179136276245118, 0.009210880279541016, 0.00918835163116455, 0.009234432220458985, 0.009203712463378906, 0.009241600036621094, 0.009250816345214843, 0.009194496154785157, 0.009051136016845703, 0.009227264404296874, 0.009227264404296874, 0.009191424369812011, 0.009276415824890137, 0.00912281608581543, 0.009060352325439454, 0.009172991752624511, 0.009186304092407227, 0.009084032058715821, 0.0090316801071167, 0.00905510425567627, 0.009248767852783203, 0.009243647575378417, 0.009118720054626465, 0.008986623764038085, 0.00892416000366211, 0.00885043239593506, 0.008862719535827637, 0.00911359977722168, 0.009030655860900879, 0.019396608352661132, 0.009150464057922364, 0.00912281608581543, 0.009197567939758301, 0.009101311683654785, 0.009109503746032714, 0.009166848182678223, 0.00908902359008789, 0.009137151718139648, 0.009185279846191406, 0.009181183815002441, 0.009253952026367187, 0.009087008476257324, 0.00908995246887207, 0.009132032394409179, 0.009161727905273438, 0.00913920021057129, 0.009211903572082519, 0.009278464317321777, 0.009169919967651367, 0.009077759742736816, 0.009086976051330567, 0.009208831787109375, 0.009174015998840332, 0.009181183815002441, 0.009260031700134277, 0.00909823989868164, 0.009146368026733399, 0.009189375877380371, 0.009160703659057617, 0.009184255599975585, 0.009249792098999024, 0.00914739227294922, 0.00920473575592041, 0.009245696067810059, 0.009291775703430176, 0.009069567680358886, 0.009195520401000976, 0.009206784248352052, 0.009132032394409179, 0.009063424110412598, 0.009133055686950683, 0.009143296241760255, 0.009078783988952637, 0.009349311828613281, 0.009133888244628905, 0.00906668758392334, 0.009192255973815918, 0.009144319534301757, 0.009219072341918945, 0.009079839706420898, 0.009255904197692871, 0.009261055946350098, 0.009198592185974122, 0.009141247749328613, 0.009293824195861817, 0.009278464317321777, 0.009257984161376954, 0.009269248008728028, 0.009266176223754884, 0.009197567939758301, 0.009148415565490722, 0.009181183815002441, 0.019737600326538086, 0.009267200469970703, 0.009135104179382325, 0.00921497631072998, 0.009219072341918945, 0.009197567939758301, 0.009191424369812011, 0.00922316837310791, 0.009259008407592773, 0.009307135581970214, 0.009315327644348144, 0.009284607887268067, 0.00931942367553711, 0.009195520401000976, 0.009155584335327148, 0.009136128425598144, 0.00922214412689209, 0.009190400123596192, 0.009202688217163087, 0.009332736015319825, 0.009200639724731445, 0.009227264404296874, 0.009165823936462402, 0.009000960350036622, 0.008986623764038085, 0.009043968200683594, 0.009018367767333984, 0.009227392196655273, 0.009224063873291015, 0.009268223762512207, 0.00930303955078125, 0.009286720275878907, 0.009290687561035156, 0.009325568199157714, 0.009293888092041016, 0.009355199813842774, 0.009337856292724609, 0.009239551544189453, 0.009166848182678223, 0.009272319793701172, 0.009359359741210938, 0.009284607887268067, 0.009266176223754884, 0.009273344039916993, 0.009251839637756347, 0.009235456466674804, 0.009286656379699706, 0.009333760261535644, 0.009286656379699706, 0.009310208320617675, 0.009302016258239745, 0.009225215911865235, 0.009176063537597656, 0.009224191665649414, 0.009250816345214843, 0.00930406379699707, 0.009120767593383788, 0.009150464057922364, 0.009233471870422363, 0.009338815689086915, 0.009276415824890137, 0.009270272254943847, 0.009217023849487305, 0.019334144592285156, 0.009140224456787109, 0.009145343780517578, 0.009179167747497559, 0.009268192291259766, 0.009254912376403808, 0.009278464317321777, 0.009160703659057617, 0.009176063537597656, 0.009175040245056153, 0.00941875171661377, 0.009135104179382325, 0.009235456466674804, 0.009289728164672852, 0.009273344039916993, 0.009245696067810059, 0.009188447952270508, 0.009137120246887207, 0.009135040283203125, 0.00913100814819336, 0.009266176223754884, 0.009107456207275391, 0.009163904190063476, 0.009179007530212403, 0.009244671821594238, 0.00909823989868164, 0.009096287727355956, 0.009116576194763184, 0.009085951805114746, 0.009220159530639649, 0.009242560386657715, 0.009169919967651367, 0.009153535842895508, 0.00923136043548584, 0.009198592185974122, 0.009219072341918945, 0.00921292781829834, 0.009152671813964845, 0.009095135688781739, 0.009108351707458497, 0.00913100814819336, 0.009149439811706543, 0.009228287696838379, 0.00913100814819336, 0.009441280364990234, 0.009235456466674804, 0.00921292781829834, 0.00928767967224121, 0.009194496154785157, 0.009140224456787109, 0.009195520401000976, 0.009189375877380371, 0.009197567939758301, 0.00923136043548584, 0.009284607887268067, 0.00919654369354248, 0.009151488304138184, 0.009153535842895508, 0.009083935737609863, 0.009085920333862305, 0.009227264404296874, 0.009158656120300293, 0.009119744300842286, 0.01919385528564453, 0.008897536277770996, 0.008768511772155761, 0.00910848045349121, 0.009161727905273438, 0.009086976051330567, 0.009133055686950683, 0.009117695808410644, 0.009067520141601563, 0.00921497631072998, 0.009129983901977539, 0.00918015956878662, 0.009044992446899413, 0.008796159744262694, 0.008827903747558593, 0.008866815567016602, 0.008691712379455567, 0.008760319709777833, 0.008839167594909669, 0.00882380771636963, 0.0088340482711792, 0.008731648445129395, 0.008802304267883301, 0.008795136451721192, 0.008844287872314453, 0.00890675163269043, 0.00881772804260254, 0.008837056159973145, 0.00872447967529297, 0.008742912292480469, 0.008830975532531739, 0.008797344207763671, 0.008948575973510743, 0.00951296043395996, 0.009158656120300293, 0.009202688217163087, 0.009095168113708496, 0.009115648269653321, 0.009158656120300293, 0.009143296241760255, 0.009332736015319825, 0.009127936363220214, 0.009206784248352052, 0.009300992012023926, 0.009198592185974122, 0.009124863624572753, 0.009181183815002441, 0.009224191665649414, 0.009182208061218262, 0.009182208061218262, 0.009177087783813476, 0.009157631874084473, 0.009219072341918945, 0.009235456466674804, 0.00921292781829834, 0.009206784248352052, 0.00922111988067627, 0.009272319793701172, 0.009179136276245118, 0.009088000297546387, 0.009075712203979493, 0.009242624282836913, 0.009273344039916993, 0.019775487899780272, 0.009277440071105958, 0.00930406379699707, 0.009160896301269532, 0.009081664085388184, 0.009125887870788574, 0.009217087745666503, 0.009294783592224121, 0.009259008407592773, 0.009174015998840332, 0.009237536430358887, 0.009176032066345215, 0.009359359741210938, 0.009254912376403808, 0.009415679931640625, 0.009275391578674316, 0.00912281608581543, 0.00922214412689209, 0.00928876781463623, 0.009192383766174317, 0.009159680366516113, 0.009207807540893554, 0.009309184074401856, 0.009120767593383788, 0.009166848182678223, 0.008999936103820801, 0.009118720054626465, 0.00923033618927002, 0.00922323226928711, 0.009146304130554199, 0.009227264404296874, 0.00930303955078125, 0.009255935668945312, 0.009110527992248535, 0.009076736450195312, 0.00912281608581543, 0.009102335929870605, 0.00913920021057129, 0.009243647575378417, 0.009208831787109375, 0.00909721565246582, 0.009232383728027344, 0.009280511856079102, 0.009257984161376954, 0.009200639724731445, 0.009266176223754884, 0.009227264404296874, 0.009320480346679687, 0.009240544319152832, 0.009115648269653321, 0.009124896049499513, 0.009084896087646484, 0.009176063537597656, 0.009248767852783203, 0.009246720314025878, 0.009260031700134277, 0.009182208061218262, 0.009068544387817384, 0.009130047798156739, 0.00921183967590332, 0.009153535842895508, 0.009088000297546387, 0.009128959655761718, 0.019221536636352538, 0.00914735984802246, 0.00881766414642334, 0.008739839553833008, 0.009056256294250489, 0.009115648269653321, 0.009055232048034668, 0.009151488304138184, 0.009211903572082519, 0.009256959915161133, 0.009225215911865235, 0.009248767852783203, 0.009208831787109375, 0.00923033618927002, 0.009128959655761718, 0.009155584335327148, 0.0091146240234375, 0.009055264472961426, 0.009093088150024414, 0.009335807800292969, 0.009111552238464356, 0.009262080192565919, 0.00919961643218994, 0.009070591926574707, 0.009192447662353515, 0.009271295547485351, 0.009233407974243164, 0.009169919967651367, 0.009148415565490722, 0.009164799690246582, 0.009200639724731445, 0.009159680366516113, 0.009225215911865235, 0.009076736450195312, 0.009156607627868652, 0.00912384033203125, 0.009118720054626465, 0.009143296241760255, 0.009150464057922364, 0.009190431594848633, 0.009420767784118652, 0.009259008407592773, 0.009239551544189453, 0.009190400123596192, 0.009144415855407715, 0.009088992118835449, 0.009233344078063964, 0.00922316837310791, 0.00910028839111328, 0.009146431922912597, 0.00923846435546875, 0.00919654369354248, 0.009192447662353515, 0.009241600036621094, 0.00913920021057129, 0.009236479759216308, 0.009218048095703125, 0.009034912109375, 0.009146207809448243, 0.00923852825164795, 0.00923033618927002, 0.009185279846191406, 0.00914739227294922, 0.019207231521606444, 0.009263039588928223, 0.009186304092407227, 0.009096320152282716, 0.009118592262268067, 0.009195520401000976, 0.009234432220458985, 0.009151488304138184, 0.009243647575378417, 0.00921395206451416, 0.009193471908569336, 0.009243647575378417, 0.00923750400543213, 0.009211903572082519, 0.009115648269653321, 0.009201663970947266, 0.009084927558898925, 0.009203712463378906, 0.009178112030029297, 0.009164799690246582, 0.009107456207275391, 0.00879923152923584, 0.0087326717376709, 0.008692735671997071, 0.008788991928100585, 0.008741888046264648, 0.008717311859130859, 0.008870911598205567, 0.008846336364746094, 0.008821760177612305, 0.008831999778747558, 0.008837120056152344, 0.008755200386047364, 0.008786944389343262, 0.008811552047729493, 0.008701919555664062, 0.008704000473022461, 0.008717311859130859, 0.008651776313781738, 0.008627200126647949, 0.008673279762268067, 0.008647744178771972, 0.008747967720031738, 0.008684543609619141, 0.008839167594909669, 0.008827903747558593, 0.008866815567016602, 0.008782848358154297, 0.00884227180480957, 0.008824799537658691, 0.008820735931396484, 0.008838144302368164, 0.008849408149719238, 0.00881766414642334, 0.008777728080749512, 0.00868556785583496, 0.008679424285888672, 0.008851455688476563, 0.008749055862426757, 0.008771583557128907, 0.008869888305664063, 0.008725503921508788, 0.008758272171020508, 0.018522111892700196, 0.008760479927062989, 0.008744799613952636, 0.008678400039672851, 0.008663040161132812, 0.008689663887023925, 0.008782848358154297, 0.008729599952697753, 0.009275391578674316, 0.010090496063232422, 0.009278464317321777, 0.009252863883972168, 0.009244671821594238, 0.00923852825164795, 0.009389056205749511, 0.009120767593383788, 0.009095168113708496, 0.00923852825164795, 0.009381888389587402, 0.009172991752624511, 0.009146368026733399, 0.009207807540893554, 0.009226240158081055, 0.009163776397705077, 0.009165823936462402, 0.009182208061218262, 0.009246720314025878, 0.009202688217163087, 0.009207807540893554, 0.009197567939758301, 0.009206784248352052, 0.009366527557373047, 0.009355263710021973, 0.00913100814819336, 0.00922214412689209, 0.009150464057922364, 0.009242624282836913, 0.009143296241760255, 0.009187328338623046, 0.00920473575592041, 0.009209856033325196, 0.009149439811706543, 0.009115648269653321, 0.009124863624572753, 0.009211903572082519, 0.009187328338623046, 0.009228287696838379, 0.009172991752624511, 0.009270272254943847, 0.009181183815002441, 0.009215999603271484, 0.009220095634460449, 0.009201663970947266, 0.009146400451660156, 0.009382880210876465, 0.009256959915161133, 0.009209856033325196, 0.009274368286132812, 0.009193471908569336, 0.009179136276245118, 0.009100319862365723, 0.009211872100830078, 0.009215999603271484, 0.019696672439575194, 0.009102304458618165, 0.009134079933166504, 0.00912384033203125, 0.009239551544189453, 0.009202688217163087, 0.009096192359924317, 0.00909004783630371, 0.00912281608581543, 0.009125887870788574, 0.009203712463378906, 0.009104384422302245, 0.00882585620880127, 0.008884223937988281, 0.008749055862426757, 0.008752127647399903, 0.008795136451721192, 0.008794112205505371, 0.008836095809936523, 0.008820735931396484, 0.008721407890319824, 0.009194496154785157, 0.009233407974243164, 0.009241600036621094, 0.009246720314025878, 0.009147583961486816, 0.009139007568359375, 0.009153535842895508, 0.009278464317321777, 0.009261055946350098, 0.009248767852783203, 0.009250816345214843, 0.009294848442077636, 0.009308159828186035, 0.009265151977539063, 0.009347071647644043, 0.009260031700134277, 0.009329664230346679, 0.009267200469970703, 0.0092293119430542, 0.009225215911865235, 0.009184255599975585, 0.0092293119430542, 0.00918835163116455, 0.009201663970947266, 0.009164799690246582, 0.009148415565490722, 0.009138208389282226, 0.009210847854614257, 0.009318400382995605, 0.009245696067810059, 0.009461759567260742, 0.010915840148925781, 0.010038271903991699, 0.010377216339111327, 0.009389056205749511, 0.009617568016052246, 0.009375583648681641, 0.009366527557373047, 0.009326592445373535, 0.009265151977539063, 0.009374719619750976, 0.009284607887268067, 0.01962700843811035, 0.009049087524414063, 0.009162752151489258, 0.009194496154785157, 0.00922316837310791, 0.009119744300842286, 0.00921497631072998, 0.009175040245056153, 0.009163776397705077, 0.00919961643218994, 0.009193471908569336, 0.009232383728027344, 0.009225215911865235, 0.009160703659057617, 0.009127936363220214, 0.00922214412689209, 0.009052255630493163, 0.009175968170166016, 0.009187328338623046, 0.009208831787109375, 0.009189408302307129, 0.009068511962890624, 0.009040896415710448, 0.009053183555603026, 0.009175040245056153, 0.009203712463378906, 0.009224191665649414, 0.009067520141601563, 0.009135104179382325, 0.009217023849487305, 0.009209856033325196, 0.009096192359924317, 0.009144319534301757, 0.009155584335327148, 0.009078783988952637, 0.009629695892333985, 0.009241600036621094, 0.009183232307434081, 0.009161727905273438, 0.009234432220458985, 0.009186304092407227, 0.009131168365478516, 0.009117535591125488, 0.009124863624572753, 0.00909721565246582, 0.009162816047668457, 0.009056192398071289, 0.009151488304138184, 0.009083904266357423, 0.009171968460083007, 0.00918835163116455, 0.00910028839111328, 0.009140224456787109, 0.009065471649169921, 0.00922316837310791, 0.009200672149658203, 0.0090982084274292, 0.00923033618927002, 0.009254912376403808, 0.009189375877380371, 0.009062399864196777, 0.009104384422302245, 0.009064448356628419, 0.019264511108398438, 0.009159744262695313, 0.00922822380065918, 0.009208831787109375, 0.00918835163116455, 0.009182208061218262, 0.00918835163116455, 0.009112575531005859, 0.009053183555603026, 0.009227295875549317, 0.009236448287963868, 0.009200639724731445, 0.009247743606567382, 0.009157631874084473, 0.009063424110412598, 0.009496576309204101, 0.009324543952941895, 0.00922316837310791, 0.009200639724731445, 0.009242624282836913, 0.00919961643218994, 0.009125887870788574, 0.00922214412689209, 0.009150464057922364, 0.009102335929870605, 0.0090316801071167, 0.009200639724731445, 0.009150464057922364, 0.009085951805114746, 0.009141247749328613, 0.009163776397705077, 0.009049087524414063, 0.009164799690246582, 0.009156607627868652, 0.00902143955230713, 0.009132032394409179, 0.009043968200683594, 0.009084927558898925, 0.009071616172790528, 0.009217023849487305, 0.009185279846191406, 0.009183232307434081, 0.009183232307434081, 0.009169919967651367, 0.00922214412689209, 0.009191424369812011, 0.009251839637756347, 0.00914739227294922, 0.009112640380859375, 0.009141183853149415, 0.009069567680358886, 0.009070591926574707, 0.009040063858032226, 0.00905401611328125, 0.009086976051330567, 0.009063424110412598, 0.009263104438781738, 0.009186304092407227, 0.009119903564453125, 0.009044832229614257, 0.009127936363220214, 0.009224191665649414, 0.009486495971679687, 0.01975075149536133, 0.009183232307434081, 0.0092293119430542, 0.009153535842895508, 0.009065471649169921, 0.009033760070800782, 0.009042911529541016, 0.009149439811706543, 0.009255935668945312, 0.009170944213867188, 0.009193471908569336, 0.00920576000213623, 0.009220095634460449, 0.009232383728027344, 0.009138175964355469, 0.009169919967651367, 0.009307135581970214, 0.009273344039916993, 0.009115648269653321, 0.009132032394409179, 0.009265151977539063, 0.009175040245056153, 0.009054207801818847, 0.009225215911865235, 0.009192447662353515, 0.009267200469970703, 0.009179136276245118, 0.009255071640014649, 0.00908784008026123, 0.009150464057922364, 0.009152640342712402, 0.009082752227783203, 0.009157631874084473, 0.00933683204650879, 0.009025535583496093, 0.008992768287658692, 0.008881343841552734, 0.008942399978637695, 0.008988672256469727, 0.00909004783630371, 0.009201663970947266, 0.009261055946350098, 0.009135104179382325, 0.009195648193359374, 0.009181056022644043, 0.00919961643218994, 0.009078783988952637, 0.00912281608581543, 0.00922214412689209, 0.009189375877380371, 0.009096192359924317, 0.009136128425598144, 0.009146368026733399, 0.009236479759216308, 0.009179136276245118, 0.009235456466674804, 0.009173088073730469, 0.00908073616027832, 0.009496576309204101, 0.009350144386291503, 0.009266176223754884, 0.00928767967224121, 0.009235456466674804, 0.01946224021911621, 0.008913824081420899, 0.009961471557617188, 0.009366527557373047, 0.009101311683654785, 0.00914739227294922, 0.009067520141601563, 0.009072768211364745, 0.012043135643005371, 0.009969663619995118, 0.009380864143371583, 0.00935321617126465, 0.00932249641418457, 0.00920473575592041, 0.01004851245880127, 0.009363455772399902, 0.009171968460083007, 0.00919961643218994, 0.009158656120300293, 0.009293824195861817, 0.009206784248352052, 0.009227264404296874, 0.009201727867126464, 0.00920569610595703, 0.009234432220458985, 0.009376768112182618, 0.009207807540893554, 0.009278464317321777, 0.009200639724731445, 0.009198592185974122, 0.009110655784606934, 0.009169792175292969, 0.009219072341918945, 0.009275391578674316, 0.009152511596679687, 0.009171968460083007, 0.009192447662353515, 0.009001983642578124, 0.009166848182678223, 0.009206784248352052, 0.009124863624572753, 0.00912179183959961, 0.009092096328735352, 0.009110527992248535, 0.009146368026733399, 0.009148415565490722, 0.00910028839111328, 0.009198592185974122, 0.009185279846191406, 0.009109503746032714, 0.009128959655761718, 0.009093119621276855, 0.00921497631072998, 0.009187328338623046, 0.00921292781829834, 0.009129983901977539, 0.009110527992248535, 0.009033727645874023, 0.009046015739440917, 0.009069567680358886, 0.009292799949645996, 0.009153535842895508, 0.009305120468139649]",tokens/s,107.28987669378871,, @@ -5553,6 +5937,7 @@ ChildProcessError: Traceback (most recent call last): TypeError: DeciCoderAttention.forward() got an unexpected keyword argument 'cache_position' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2101.9648,5971.116032,0.0,5324.668928,5080.87552,s,10,3.1623598327636717,0.31623598327636715,0.00024235336102111678,0.3162159729003906,0.3165054656982422,0.31660202178955077,0.31667926666259766,"[0.31669857788085937, 0.3162686462402344, 0.3164840087890625, 0.31633709716796876, 0.3158043518066406, 0.31612255859375, 0.3161309814453125, 0.3161632995605469, 0.31597808837890623, 0.31637222290039063]",tokens/s,809.5220453653268,kWh,3.7342462792164754e-06,2.046175760932556e-06,1.8911707837687697e-05,2.4692129877836726e-05,tokens/kWh,10367675.905908046,MB,2101.9648,5971.116032,0.0,5324.668928,5290.117632,s,10,173.26880078125,17.326880078125,0.0008046980240511191,17.3267978515625,17.327645117187497,17.32819267578125,17.32863072265625,"[17.3270703125, 17.32675, 17.326052734375, 17.327296875, 17.3275234375, 17.32606640625, 17.3265234375, 17.325931640625, 17.326845703125, 17.328740234375]",tokens/s,3.6359690674801186,kWh,0.0002045667352755037,0.00011211794793604098,0.0010186654833345132,0.0013353501665460579,tokens/kWh,47178.636419353796,,s,629,175.83934860229496,0.2795538133581795,0.03756451887205511,0.27501873779296876,0.2752235412597656,0.27533250122070313,0.591240341796875,"[0.275198974609375, 0.2747781066894531, 0.2748385314941406, 0.2747607116699219, 0.27491326904296876, 0.2749624328613281, 0.27463168334960936, 0.27481497192382814, 0.274840576171875, 0.2751907958984375, 0.27510272216796877, 0.2749347839355469, 0.27483544921875, 0.2749849548339844, 0.2753403015136719, 0.27486309814453125, 0.27506381225585935, 0.2750320739746094, 0.2749327392578125, 0.27479449462890626, 0.27507815551757814, 0.27491122436523435, 0.2749091796875, 0.2749859924316406, 0.2749819030761719, 0.2750750732421875, 0.2750402526855469, 0.27504434204101563, 0.2749245300292969, 0.2748487548828125, 0.2750873718261719, 0.27546316528320314, 0.2752255859375, 0.2750679016113281, 0.2751549377441406, 0.2749347839355469, 0.27496038818359375, 0.2750351257324219, 0.2749020080566406, 0.27519692993164063, 0.27499520874023436, 0.2750474243164063, 0.2749327392578125, 0.27501669311523436, 0.27505767822265625, 0.27501055908203126, 0.27503411865234373, 0.2750894165039062, 0.2751805419921875, 0.27512115478515625, 0.27502081298828124, 0.2750965881347656, 0.27512216186523436, 0.2753597412109375, 0.2751068115234375, 0.275093505859375, 0.27523275756835935, 0.275135498046875, 0.2751375427246094, 0.27496551513671874, 0.27523583984375, 0.27499520874023436, 0.5913477172851562, 0.274830322265625, 0.2748323974609375, 0.27500338745117187, 0.27521127319335936, 0.27528704833984374, 0.27483544921875, 0.2747678833007812, 0.27484979248046876, 0.2747781066894531, 0.274882568359375, 0.2750955505371094, 0.2749358215332031, 0.2748938293457031, 0.2748334045410156, 0.27500851440429686, 0.27480267333984376, 0.27513958740234373, 0.27495013427734377, 0.27490304565429685, 0.275267578125, 0.2749644775390625, 0.27491326904296876, 0.2748948364257813, 0.27508120727539065, 0.2749266052246094, 0.27535052490234374, 0.2749358215332031, 0.2749992980957031, 0.2749921264648437, 0.27484774780273435, 0.27500543212890627, 0.27491943359375, 0.2750003051757812, 0.2750115966796875, 0.2751180725097656, 0.27498699951171873, 0.27493170166015624, 0.2749542541503906, 0.2748968811035156, 0.2757120056152344, 0.2750218200683594, 0.2752235412597656, 0.2749921264648437, 0.2749245300292969, 0.27508428955078124, 0.27499520874023436, 0.2750064697265625, 0.2751129455566406, 0.27507403564453126, 0.2751037292480469, 0.2750556030273438, 0.27498086547851563, 0.27507098388671875, 0.27502590942382815, 0.27505255126953126, 0.27506585693359376, 0.2751016845703125, 0.2752962646484375, 0.27508837890625, 0.275515380859375, 0.2752000122070313, 0.2751385498046875, 0.5916641235351563, 0.27478631591796876, 0.27482009887695313, 0.27491226196289065, 0.2747607116699219, 0.27492556762695314, 0.2749245300292969, 0.27482110595703124, 0.274798583984375, 0.27490609741210936, 0.2748671875, 0.27493991088867187, 0.27506689453125, 0.27507302856445315, 0.2748938293457031, 0.27494708251953126, 0.2748016662597656, 0.2749941711425781, 0.27487640380859374, 0.2748692626953125, 0.27501055908203126, 0.27505050659179686, 0.2749010009765625, 0.2748334045410156, 0.27492556762695314, 0.274956298828125, 0.2749706115722656, 0.2750597229003906, 0.2748968811035156, 0.27507916259765625, 0.27507199096679685, 0.27513037109375, 0.27511911010742185, 0.2748968811035156, 0.274951171875, 0.27523992919921875, 0.27490304565429685, 0.2749982604980469, 0.2750699462890625, 0.27497369384765624, 0.27497369384765624, 0.27502490234375, 0.27506381225585935, 0.275009521484375, 0.2749306945800781, 0.27501055908203126, 0.2751016845703125, 0.27510272216796877, 0.275230712890625, 0.27520306396484373, 0.27552459716796873, 0.27511398315429686, 0.2750894165039062, 0.2750750732421875, 0.2749798278808594, 0.27514981079101564, 0.27515802001953127, 0.2751447143554687, 0.2750351257324219, 0.27518875122070313, 0.27545803833007815, 0.27511911010742185, 0.2752235412597656, 0.5914030151367188, 0.27480679321289064, 0.27478631591796876, 0.27494091796875, 0.27484262084960936, 0.27500851440429686, 0.274798583984375, 0.2749900817871094, 0.274830322265625, 0.27475662231445314, 0.2748590087890625, 0.27493374633789064, 0.2748610534667969, 0.27496038818359375, 0.2748938293457031, 0.2751068115234375, 0.27488357543945313, 0.275167236328125, 0.274872314453125, 0.274840576171875, 0.27497369384765624, 0.27494503784179686, 0.2751068115234375, 0.2749245300292969, 0.27491021728515624, 0.27501873779296876, 0.27484262084960936, 0.2748876647949219, 0.27505255126953126, 0.27506381225585935, 0.27500851440429686, 0.2750965881347656, 0.27498291015625, 0.27501669311523436, 0.27518362426757814, 0.2751795349121094, 0.2750402526855469, 0.27513446044921874, 0.2750003051757812, 0.27497268676757813, 0.27517132568359376, 0.2752593994140625, 0.27508428955078124, 0.27551028442382813, 0.2751467590332031, 0.27510784912109376, 0.2750074768066406, 0.2751068115234375, 0.27511602783203126, 0.2750750732421875, 0.2751539306640625, 0.2751119384765625, 0.2749624328613281, 0.2750822448730469, 0.2752163696289063, 0.275198974609375, 0.2751201171875, 0.27526144409179687, 0.27553280639648436, 0.2751682434082031, 0.2751150207519531, 0.2751877136230469, 0.2750679016113281, 0.5908869018554688, 0.2749245300292969, 0.27489279174804687, 0.27481292724609374, 0.2747852783203125, 0.27498699951171873, 0.2748948364257813, 0.27476171875, 0.27514776611328123, 0.2748016662597656, 0.2749440002441406, 0.27507098388671875, 0.274988037109375, 0.27523175048828125, 0.2749696044921875, 0.27488461303710937, 0.2750412902832031, 0.27491226196289065, 0.2748374938964844, 0.2750269470214844, 0.2748794860839844, 0.2749992980957031, 0.27503103637695314, 0.2748118896484375, 0.2750074768066406, 0.27487539672851563, 0.27500851440429686, 0.27484262084960936, 0.27503616333007813, 0.27514981079101564, 0.274988037109375, 0.2754109497070312, 0.275030029296875, 0.2749224853515625, 0.27513037109375, 0.2750474243164063, 0.2750494689941406, 0.27498904418945314, 0.27496038818359375, 0.27510580444335936, 0.27506689453125, 0.2750975952148437, 0.275125244140625, 0.27500338745117187, 0.27499725341796877, 0.2753341369628906, 0.27504638671875, 0.2751037292480469, 0.2751518859863281, 0.2750351257324219, 0.2752010192871094, 0.2751590270996094, 0.275209228515625, 0.2754037780761719, 0.2751528930664062, 0.27510272216796877, 0.2750330810546875, 0.27525222778320313, 0.27512832641601564, 0.27523275756835935, 0.27520306396484373, 0.27509451293945314, 0.275167236328125, 0.591182861328125, 0.2749153137207031, 0.27479244995117186, 0.2752962646484375, 0.27491738891601564, 0.2749798278808594, 0.2749900817871094, 0.2748671875, 0.27480780029296875, 0.27507302856445315, 0.27497882080078123, 0.2749624328613281, 0.2749849548339844, 0.27487335205078123, 0.27491943359375, 0.27487026977539064, 0.27495834350585935, 0.2748538818359375, 0.27482009887695313, 0.27472076416015623, 0.2749286499023437, 0.27513345336914063, 0.27491738891601564, 0.27482827758789063, 0.2753013610839844, 0.27489791870117186, 0.2749071350097656, 0.274956298828125, 0.27478427124023436, 0.2748241882324219, 0.2750679016113281, 0.2749542541503906, 0.2750382080078125, 0.2749859924316406, 0.2752041015625, 0.2751119384765625, 0.27506689453125, 0.274914306640625, 0.2751201171875, 0.2753064880371094, 0.2750136413574219, 0.27502285766601564, 0.2750218200683594, 0.2749214782714844, 0.27502899169921874, 0.2750044250488281, 0.275093505859375, 0.27505767822265625, 0.2751201171875, 0.27507815551757814, 0.27517031860351565, 0.2749706115722656, 0.2751231994628906, 0.27507403564453126, 0.2749992980957031, 0.2751016845703125, 0.2751488037109375, 0.275135498046875, 0.2754703369140625, 0.27516619873046877, 0.27518362426757814, 0.27513958740234373, 0.27514163208007814, 0.5912626953125, 0.2746531982421875, 0.2748385314941406, 0.27486309814453125, 0.27493887329101563, 0.2749706115722656, 0.2748252258300781, 0.2748938293457031, 0.2749163513183594, 0.2748385314941406, 0.2748856201171875, 0.27499725341796877, 0.2751129455566406, 0.27485592651367186, 0.27486209106445314, 0.2748907470703125, 0.2748590087890625, 0.274882568359375, 0.2749521789550781, 0.2750556030273438, 0.27492352294921873, 0.2749521789550781, 0.2749593505859375, 0.27478936767578127, 0.2749358215332031, 0.2749266052246094, 0.2748938293457031, 0.2749296569824219, 0.2749245300292969, 0.274956298828125, 0.2753966064453125, 0.2750392456054688, 0.2749747314453125, 0.2751201171875, 0.27508428955078124, 0.27516928100585936, 0.275093505859375, 0.2749767761230469, 0.27500543212890627, 0.2749276123046875, 0.27500851440429686, 0.27505050659179686, 0.27517031860351565, 0.27502386474609375, 0.27498291015625, 0.2751324157714844, 0.2752061462402344, 0.27502081298828124, 0.275557373046875, 0.27510272216796877, 0.2751098937988281, 0.27516415405273437, 0.27508120727539065, 0.2750771179199219, 0.27513446044921874, 0.27501773071289065, 0.27513650512695315, 0.27518875122070313, 0.2750699462890625, 0.2751385498046875, 0.2755860595703125, 0.27517031860351565, 0.27525531005859377, 0.5912658081054688, 0.27493170166015624, 0.27482827758789063, 0.2749491271972656, 0.27483544921875, 0.27496038818359375, 0.2748876647949219, 0.27492556762695314, 0.2748098449707031, 0.2748590087890625, 0.274946044921875, 0.27496038818359375, 0.2749358215332031, 0.2749859924316406, 0.27488153076171873, 0.2748692626953125, 0.27488461303710937, 0.2749286499023437, 0.27498291015625, 0.2748989562988281, 0.27534335327148435, 0.2750320739746094, 0.27485592651367186, 0.2748610534667969, 0.27494091796875, 0.2748876647949219, 0.2748487548828125, 0.27498086547851563, 0.27499725341796877, 0.27497369384765624, 0.27501873779296876, 0.2750269470214844, 0.2748538818359375, 0.27500851440429686, 0.2749931640625, 0.2751590270996094, 0.27493374633789064, 0.274951171875, 0.2750494689941406, 0.27503717041015624, 0.2751231994628906, 0.2754232177734375, 0.27509451293945314, 0.2750013427734375, 0.27502490234375, 0.2750064697265625, 0.2748436584472656, 0.27507916259765625, 0.2750965881347656, 0.27501055908203126, 0.27513650512695315, 0.2751047668457031, 0.27512728881835935, 0.275093505859375, 0.27511602783203126, 0.2750648193359375, 0.275061767578125, 0.27523175048828125, 0.275167236328125, 0.2750986328125, 0.2752481384277344, 0.27527679443359376, 0.2751201171875, 0.5914869995117188, 0.2747115478515625, 0.27491122436523435, 0.2747750549316406, 0.27482110595703124, 0.2748252258300781, 0.2748968811035156, 0.27479962158203125, 0.2747607116699219, 0.27496038818359375, 0.274872314453125, 0.27533004760742186, 0.27499520874023436, 0.2750648193359375, 0.2749327392578125, 0.27499520874023436, 0.275009521484375, 0.27496755981445314, 0.2748528747558594, 0.274766845703125, 0.2749644775390625, 0.2749266052246094, 0.2748805236816406, 0.2747740173339844, 0.27503411865234373, 0.2748190612792969, 0.27482827758789063, 0.27507916259765625, 0.27513650512695315, 0.27535565185546873, 0.27501873779296876, 0.2749849548339844, 0.27530853271484373, 0.2750218200683594, 0.2750545959472656, 0.27512832641601564, 0.2749819030761719, 0.27498904418945314, 0.27506381225585935, 0.27501055908203126, 0.275061767578125, 0.2750218200683594, 0.2751016845703125, 0.2750802001953125, 0.27516110229492186, 0.2751426696777344, 0.2751098937988281, 0.2752634887695313, 0.275267578125, 0.27502285766601564, 0.2752563171386719, 0.27503103637695314, 0.2752204895019531, 0.2750484619140625, 0.27531674194335937, 0.2751047668457031, 0.275135498046875, 0.27513446044921874, 0.27511602783203126, 0.27506585693359376, 0.2751242370605469, 0.2751047668457031, 0.2751129455566406, 0.5915084838867187, 0.27509451293945314, 0.2749163513183594, 0.27489279174804687, 0.27486004638671874, 0.2749900817871094, 0.27495321655273436, 0.2748989562988281, 0.2748374938964844, 0.27503717041015624, 0.27489791870117186, 0.2749859924316406, 0.2748805236816406, 0.2748518371582031, 0.27495834350585935, 0.27506072998046877, 0.27489791870117186, 0.27492044067382815, 0.2749245300292969, 0.2749839477539063, 0.2749910888671875, 0.2749491271972656, 0.2753484802246094, 0.274946044921875, 0.27497882080078123, 0.27499520874023436, 0.2748989562988281, 0.2750648193359375, 0.2751231994628906, 0.27494503784179686, 0.2751129455566406, 0.27514163208007814, 0.2752122802734375, 0.27512728881835935, 0.2750433349609375, 0.275262451171875, 0.27510272216796877, 0.27497369384765624, 0.2752850036621094, 0.2749706115722656, 0.27505255126953126, 0.27513650512695315, 0.27524197387695315, 0.27487539672851563, 0.27529727172851565, 0.2754805908203125, 0.2750269470214844, 0.27508428955078124, 0.27512933349609375, 0.27514163208007814, 0.2751528930664062, 0.27501055908203126, 0.2751600646972656, 0.27512728881835935, 0.27509963989257813, 0.27514163208007814, 0.27519488525390623, 0.27512728881835935, 0.27538534545898435, 0.2750556030273438, 0.27517745971679686, 0.2751447143554687, 0.2751426696777344]",tokens/s,3.577128811041278,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -5582,7 +5967,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664beec5-48299db16a7b734c4043f50b;7660e74a-c225-4c0b-a1d1-bb7697ae09a4) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3fcf-79be09596c8bb23725c935fd;570b5ed3-6f98-406f-9507-f461017a8736) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -5630,6 +6015,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm-20b,internlm/internlm-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,3699.970048,14401.667072,0.0,13755.219968,13102.766592,s,10,9.234486938476563,0.9234486938476563,0.00015694751309834682,0.9234230346679687,0.9236289245605469,0.9236960235595703,0.923749702758789,"[0.9237631225585937, 0.9233424072265625, 0.9232378540039062, 0.923614013671875, 0.92333740234375, 0.9234802856445312, 0.923286865234375, 0.9234804077148437, 0.9233657836914062, 0.9235787963867188]",tokens/s,277.22168183848555,kWh,1.0911467626239315e-05,5.978965125177638e-06,5.399514420618095e-05,7.08855769575979e-05,tokens/kWh,3611453.9937106417,MB,3699.970048,14401.667072,0.0,13755.219968,13549.146112,s,10,521.8675390625001,52.18675390625001,0.004832173035005018,52.18797265625,52.192680078125,52.1929630859375,52.1931894921875,"[52.1926171875, 52.17780078125, 52.18303125, 52.19024609375, 52.1896953125, 52.18303125, 52.18698046875, 52.18192578125, 52.19324609375, 52.18896484375]",tokens/s,1.2072028874065488,kWh,0.0006161677286314844,0.0003377143746004912,0.003021631094070626,0.003975513197302602,tokens/kWh,15847.010655818145,,s,629,529.3491154174801,0.8415725205365349,0.10977779395641166,0.828315673828125,0.8288815307617188,0.8291999267578125,1.7524884814453126,"[0.8283299560546875, 0.8288552856445313, 0.8289197998046876, 0.8288737182617187, 0.8291635131835937, 0.8292260131835938, 0.8291993408203125, 0.8290447387695312, 0.8284845581054687, 0.8295833740234375, 0.829169677734375, 0.8295208740234375, 0.8292321166992187, 0.8292198486328125, 0.8287918090820312, 0.8292393188476562, 0.8283453369140625, 0.8279398193359375, 0.8277166137695312, 0.8279326782226563, 0.82779443359375, 0.8279859008789062, 0.8278538208007813, 0.8286986083984375, 0.8280166625976563, 0.8292669677734374, 0.8286300048828125, 0.8284497680664062, 0.8281569213867187, 0.8281292724609375, 0.8283238525390625, 0.82832177734375, 0.828168212890625, 0.8286556396484375, 0.8279562377929688, 0.8285941772460937, 0.8276121826171875, 0.828205078125, 0.8276715698242187, 0.828705810546875, 0.82787841796875, 0.8283914184570312, 0.82791015625, 0.8280340576171875, 0.82789990234375, 0.828811279296875, 0.8277933959960937, 0.8286689453125, 0.8279623413085937, 0.8286515502929688, 0.8282398681640625, 0.8284702758789062, 0.828310546875, 0.828242919921875, 0.8281712646484375, 0.8282501220703125, 0.827809814453125, 0.8283566284179688, 0.8280350952148438, 0.8285419311523438, 0.8281610107421875, 0.828590087890625, 1.7526087646484374, 0.8282511596679687, 0.8280330200195313, 0.8287457275390625, 0.827747314453125, 0.8281599731445313, 0.8289617919921874, 0.8283146362304687, 0.8284682006835937, 0.8281384887695312, 0.8286207885742187, 0.828526611328125, 0.8286443481445313, 0.8279920654296875, 0.8281005859375, 0.828253173828125, 0.8283299560546875, 0.8285716552734375, 0.8289443969726562, 0.8280289306640625, 0.8287232055664062, 0.8279951171875, 0.8286760864257813, 0.8278538208007813, 0.8283197631835938, 0.8281917724609374, 0.827905029296875, 0.8280709228515625, 0.8279982299804688, 0.82779443359375, 0.82838427734375, 0.827863037109375, 0.828579833984375, 0.8280166625976563, 0.8283709716796875, 0.8280606689453125, 0.8286248779296875, 0.8277299194335938, 0.8278067016601562, 0.8275527954101562, 0.8280186767578125, 0.827683837890625, 0.8284047241210938, 0.8277227783203125, 0.8281292724609375, 0.8276500244140625, 0.828442626953125, 0.827905029296875, 0.82813134765625, 0.8279777221679687, 0.8279449462890625, 0.8278568725585937, 0.827884521484375, 0.828137451171875, 0.8275814208984374, 0.82881640625, 0.82893310546875, 0.8283596801757812, 0.8284989624023438, 0.8284825439453125, 0.8283627319335938, 0.8282060546875, 0.828095458984375, 1.7519022216796876, 0.8275968017578125, 0.8276592407226563, 0.8287620849609375, 0.8279121704101563, 0.8282992553710937, 0.8284395751953125, 0.8284252319335937, 0.82886962890625, 0.8285501708984375, 0.8285828857421875, 0.8282501220703125, 0.828315673828125, 0.8279388427734375, 0.82779443359375, 0.829212646484375, 0.8281763916015625, 0.8284119262695312, 0.8286464233398437, 0.8283207397460938, 0.828284912109375, 0.828379150390625, 0.8283883666992188, 0.8285040893554687, 0.8284805297851563, 0.8284548950195313, 0.8283238525390625, 0.8286340942382813, 0.828189697265625, 0.8283955078125, 0.8282286376953125, 0.8282521362304688, 0.8280985717773437, 0.8285787963867187, 0.8282685546875, 0.827926513671875, 0.8280299682617187, 0.8283094482421876, 0.8282941284179688, 0.8285736694335938, 0.82859521484375, 0.8283627319335938, 0.8281138916015625, 0.8280494384765625, 0.8279715576171875, 0.8282552490234375, 0.8279132080078125, 0.8279388427734375, 0.8278876342773438, 0.828315673828125, 0.827831298828125, 0.8283770751953125, 0.8280145874023438, 0.8283197631835938, 0.828168212890625, 0.828189697265625, 0.828669921875, 0.8288809204101563, 0.8283207397460938, 0.828379150390625, 0.8280166625976563, 0.8285234985351563, 0.8285552368164063, 1.7527921142578125, 0.828043212890625, 0.8282654418945312, 0.828569580078125, 0.8276694946289063, 0.8288983154296875, 0.8285204467773437, 0.8285911254882813, 0.8283862915039063, 0.8288511962890625, 0.8286136474609375, 0.8290160522460938, 0.82844775390625, 0.8288204956054688, 0.8282091674804688, 0.8283566284179688, 0.8277545166015625, 0.828521484375, 0.8283187255859376, 0.82813134765625, 0.8282613525390625, 0.8279757080078125, 0.8283146362304687, 0.8285399169921875, 0.8281292724609375, 0.8288614501953125, 0.8281251831054688, 0.8287293701171875, 0.8280320434570313, 0.8292351684570313, 0.8283986206054688, 0.8284313354492188, 0.8281497802734376, 0.828242919921875, 0.8283381958007813, 0.8286873779296875, 0.8278701782226563, 0.828253173828125, 0.8279766845703125, 0.8285347900390625, 0.8284252319335937, 0.8286965942382812, 0.8283033447265625, 0.828010498046875, 0.8280514526367188, 0.8280606689453125, 0.828368896484375, 0.8286996459960938, 0.8281978759765625, 0.8291727294921875, 0.8280750122070313, 0.8287785034179688, 0.8282081298828124, 0.8290396118164063, 0.8282664794921875, 0.8282244873046875, 0.8284303588867188, 0.8285634765625, 0.82805859375, 0.8282265625, 0.8286863403320313, 0.8288552856445313, 0.8286197509765625, 1.7528729248046875, 0.8281569213867187, 0.8281046752929687, 0.82874267578125, 0.8277800903320313, 0.8292249755859376, 0.8288839721679687, 0.828832763671875, 0.8286668701171875, 0.828669921875, 0.8286453857421875, 0.8288338012695312, 0.8287528686523438, 0.8280883178710937, 0.827683837890625, 0.82812109375, 0.8279807739257813, 0.8282572631835937, 0.8284447021484375, 0.82854296875, 0.8287160034179688, 0.8286515502929688, 0.8279459838867187, 0.8287528686523438, 0.8282070922851562, 0.8284672241210937, 0.8284119262695312, 0.828000244140625, 0.827989990234375, 0.8281989135742187, 0.8284241943359375, 0.8284898071289063, 0.8292003173828125, 0.8285255737304688, 0.8282030029296875, 0.8286361694335938, 0.82777392578125, 0.827915283203125, 0.8281702270507812, 0.8278272094726562, 0.82804736328125, 0.8277350463867188, 0.8282869873046875, 0.8278343505859375, 0.828442626953125, 0.8284241943359375, 0.8276316528320312, 0.8282521362304688, 0.828095458984375, 0.8286126098632812, 0.8279807739257813, 0.8282327270507812, 0.8277401733398437, 0.8281630859375, 0.8279183349609375, 0.828310546875, 0.8293826293945312, 0.8292290649414062, 0.8290283813476562, 0.8291409912109375, 0.828621826171875, 0.8294010620117187, 0.82875390625, 1.7525504150390625, 0.8278475952148437, 0.8281569213867187, 0.8283740234375, 0.8284754028320312, 0.82943896484375, 0.8290396118164063, 0.8289986572265625, 0.8283309936523438, 0.8293314819335937, 0.828231689453125, 0.8284518432617187, 0.8282132568359375, 0.82813134765625, 0.8279654541015625, 0.8282091674804688, 0.8285654907226563, 0.82838427734375, 0.82808935546875, 0.8284825439453125, 0.82813232421875, 0.8284815063476563, 0.8280330200195313, 0.8283402099609375, 0.8284876708984374, 0.8285337524414063, 0.8284190673828125, 0.8281508178710938, 0.8285429077148437, 0.8289249267578125, 0.8282756958007812, 0.828568603515625, 0.8282664794921875, 0.82851123046875, 0.8281937866210938, 0.8279603271484375, 0.8280504150390625, 0.8278507690429687, 0.8280791015625, 0.8282203979492188, 0.8286986083984375, 0.828205078125, 0.8285870361328125, 0.8280360717773437, 0.827936767578125, 0.8279459838867187, 0.8280350952148438, 0.8279039916992188, 0.8279265747070312, 0.8277093505859375, 0.8278538208007813, 0.8277719116210938, 0.828042236328125, 0.8284456787109375, 0.8285706176757812, 0.8286351318359375, 0.82832177734375, 0.828147705078125, 0.8280197143554687, 0.8282838745117187, 0.8279285888671875, 0.8281026611328125, 0.828169189453125, 1.7523292236328125, 0.8278825073242188, 0.8277340087890624, 0.8286156616210938, 0.827905029296875, 0.8286863403320313, 0.82865869140625, 0.8282357788085938, 0.8282726440429687, 0.8281620483398437, 0.8281221313476562, 0.8281395263671875, 0.8281405639648437, 0.8282992553710937, 0.8281384887695312, 0.8283699340820313, 0.8281774291992188, 0.8281876220703125, 0.8287948608398438, 0.8283566284179688, 0.828273681640625, 0.8285286254882812, 0.828663818359375, 0.8283955078125, 0.8284487915039063, 0.8293826293945312, 0.8286658325195313, 0.8287877197265625, 0.8280452880859375, 0.8279920654296875, 0.8284917602539063, 0.828095458984375, 0.82825634765625, 0.8284722290039063, 0.8283391723632813, 0.8282654418945312, 0.8283033447265625, 0.8285183715820312, 0.8284794921875, 0.8290130004882813, 0.8281815185546875, 0.82836181640625, 0.8288634033203125, 0.827905029296875, 0.8278640747070313, 0.8282040405273438, 0.8283299560546875, 0.8282562255859375, 0.8284856567382812, 0.828675048828125, 0.828295166015625, 0.8283176879882812, 0.8286771240234375, 0.82789990234375, 0.828400634765625, 0.8281354370117188, 0.8284415893554687, 0.8287262573242188, 0.8282142944335937, 0.8280452880859375, 0.8280760498046875, 0.8284774169921875, 0.8288265991210938, 1.7529027099609376, 0.827556884765625, 0.8284979248046875, 0.828601318359375, 0.8279746704101563, 0.8293468017578125, 0.8282152709960937, 0.82821630859375, 0.8281538696289062, 0.8292055053710937, 0.8284395751953125, 0.8282838745117187, 0.828495849609375, 0.8278896484375, 0.8282327270507812, 0.8279193725585937, 0.8277135620117188, 0.8278077392578125, 0.8276807861328125, 0.8278988647460938, 0.8280238037109375, 0.828084228515625, 0.8282398681640625, 0.8278763427734375, 0.8287344360351563, 0.8285307006835938, 0.8287344360351563, 0.828031005859375, 0.82808935546875, 0.8278344116210937, 0.8279295654296875, 0.8282808227539062, 0.828379150390625, 0.82868017578125, 0.8285399169921875, 0.828094482421875, 0.8284467163085938, 0.827958251953125, 0.8284508056640625, 0.8280094604492187, 0.828516357421875, 0.8281077880859375, 0.82827978515625, 0.8279234619140625, 0.8279653930664063, 0.828789794921875, 0.8284108276367188, 0.82785791015625, 0.8292556762695312, 0.8289863891601562, 0.8286259155273438, 0.8280647583007813, 0.8279388427734375, 0.8278200073242188, 0.82796337890625, 0.8279766845703125, 0.828020751953125, 0.8284313354492188, 0.8285388793945313, 0.82849072265625, 0.8285552368164063, 0.8282194213867188, 0.8287129516601562, 1.7530849609375, 0.8283176879882812, 0.8286771240234375, 0.82846923828125, 0.8278046875, 0.8284805297851563, 0.82914404296875, 0.8283381958007813, 0.8285050659179688, 0.8286607055664063, 0.8288368530273438, 0.8283832397460937, 0.8281456909179687, 0.82777392578125, 0.8281282348632812, 0.8281589965820313, 0.828094482421875, 0.828248046875, 0.8285931396484375, 0.8284467163085938, 0.8288184204101563, 0.8283504638671875, 0.8290365600585937, 0.8283986206054688, 0.8280831909179688, 0.8281907348632812, 0.82876416015625, 0.8283187255859376, 0.8284784545898437, 0.8284364624023437, 0.8289658813476563, 0.8285081787109375, 0.8288491821289062, 0.828727294921875, 0.8288767700195312, 0.828095458984375, 0.82791015625, 0.8276572265625, 0.8284559326171875, 0.828126220703125, 0.8284979248046875, 0.8282296142578125, 0.8292024536132813, 0.82812109375, 0.8281671752929688, 0.8280278930664062, 0.8283013305664062, 0.8280166625976563, 0.82812109375, 0.8280576171875, 0.8279807739257813, 0.828169189453125, 0.8279746704101563, 0.8283095092773437, 0.8285296630859375, 0.8288389282226563, 0.8287559814453125, 0.8289495239257813, 0.8288399658203125, 0.82905908203125, 0.8289454345703124, 0.82897509765625, 0.8292198486328125, 1.7533941650390625, 0.8275271606445312, 0.8281149291992187, 0.8286576538085938, 0.828031982421875, 0.8285419311523438, 0.8285399169921875, 0.8288491821289062, 0.8285101928710937, 0.8283607177734374, 0.8288778076171875, 0.8285839233398438, 0.8284600219726562, 0.8286679077148438, 0.829053955078125, 0.828205078125, 0.8287528686523438, 0.828284912109375, 0.828516357421875, 0.828653564453125, 0.828494873046875, 0.8280821533203125, 0.8286617431640625, 0.8286177368164063, 0.8287139282226562, 0.829022216796875, 0.82962841796875, 0.8281630859375, 0.8282173461914063, 0.8284231567382813, 0.8280617065429687, 0.8282470703125, 0.8282173461914063, 0.828221435546875, 0.8281138916015625, 0.8280043334960937, 0.828037109375, 0.8278251342773437, 0.8289996948242188, 0.8277637329101563, 0.828358642578125, 0.828105712890625, 0.8290048217773438, 0.8281620483398437, 0.8279541625976562, 0.8280166625976563, 0.8280391845703124, 0.8287129516601562, 0.8283720092773438, 0.828396484375, 0.8285409545898438, 0.8286597290039063, 0.8282623901367188, 0.82792041015625, 0.8279725952148438, 0.828685302734375, 0.82830859375, 0.8281056518554688, 0.8279388427734375, 0.8283289794921875, 0.827978759765625, 0.8282019653320313, 0.8284600219726562]",tokens/s,1.1882517259029104,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-2.7b,facebook/opt-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2064.416768,2720.530432,0.0,2074.083328,1826.138624,s,10,1.2873819351196287,0.12873819351196286,0.0006770426005378833,0.1284730758666992,0.12977086029052737,0.1297958610534668,0.12981586166381837,"[0.1284239044189453, 0.12802653503417968, 0.12845242309570312, 0.1283489990234375, 0.12849372863769531, 0.12782758331298827, 0.1286923828125, 0.12953021240234375, 0.1297653045654297, 0.12982086181640626]",tokens/s,1988.5318646809462,kWh,1.5090974846805032e-06,8.269070770519292e-07,7.080983161266559e-06,9.416987722998991e-06,tokens/kWh,27184913.852523606,MB,2064.416768,2720.530432,0.0,2074.083328,1932.190208,s,10,71.08341162109375,7.108341162109374,0.0011266889196988744,7.108484130859376,7.10943681640625,7.11006435546875,7.11056638671875,"[7.11069189453125, 7.1089921875, 7.107189453125, 7.10866162109375, 7.10855126953125, 7.1084169921875, 7.10929736328125, 7.10754833984375, 7.10741259765625, 7.10664990234375]",tokens/s,8.862827284629791,kWh,8.393480116996586e-05,4.6001814734764716e-05,0.0003951348108333315,0.000525071426738062,tokens/kWh,119983.67610932348,,s,629,72.1292452392579,0.1146728859129695,0.015315309775968402,0.11281510162353516,0.11292876586914063,0.11301908416748048,0.2412179412841797,"[0.11319398498535156, 0.11324313354492188, 0.11316326141357422, 0.1131151351928711, 0.11281919860839844, 0.1127874526977539, 0.11280281829833984, 0.11280691528320312, 0.1129195556640625, 0.11283968353271484, 0.11278028869628906, 0.11267993927001953, 0.11283251190185548, 0.11281613159179688, 0.11281100463867187, 0.1128089599609375, 0.1127710723876953, 0.1128632354736328, 0.11270758056640626, 0.11272191619873047, 0.11268402862548828, 0.1128642578125, 0.11276287841796875, 0.11280486297607421, 0.11278131103515625, 0.11277823638916015, 0.11274547576904297, 0.11276799774169922, 0.11276083374023438, 0.1128642578125, 0.11285606384277344, 0.112827392578125, 0.11289907073974609, 0.11281203460693359, 0.11281305694580078, 0.11274752044677734, 0.11283968353271484, 0.11323085021972656, 0.11279154968261719, 0.11292569732666016, 0.11282841491699219, 0.11280076599121094, 0.11284684753417969, 0.11273830413818359, 0.11285196685791016, 0.11286732482910156, 0.11285094451904297, 0.11282329559326172, 0.11285401916503907, 0.11279769897460938, 0.1128796157836914, 0.11282329559326172, 0.1129881591796875, 0.11291136169433594, 0.1127936019897461, 0.1128622055053711, 0.11282637023925782, 0.1133148193359375, 0.11288985443115235, 0.1127741470336914, 0.1128806381225586, 0.113006591796875, 0.2409820098876953, 0.11295231628417969, 0.11276697540283204, 0.11282022094726563, 0.11267276763916016, 0.11276902770996093, 0.11301990509033204, 0.11279462432861329, 0.11277619171142578, 0.11283865356445312, 0.11276390075683594, 0.11281305694580078, 0.1127034912109375, 0.11283865356445312, 0.11276799774169922, 0.11333427429199219, 0.11281203460693359, 0.11281100463867187, 0.11275878143310547, 0.11279666900634766, 0.11270553588867188, 0.1128058853149414, 0.11281613159179688, 0.11282125091552735, 0.11273318481445313, 0.1128058853149414, 0.11275059509277344, 0.11280281829833984, 0.11272396850585938, 0.11283865356445312, 0.11281407928466797, 0.1128058853149414, 0.1127874526977539, 0.11292774200439454, 0.11299839782714843, 0.11288678741455078, 0.11276390075683594, 0.11285708618164063, 0.11287142181396484, 0.11280691528320312, 0.11279872131347657, 0.11282950592041016, 0.11278636932373047, 0.11287245178222656, 0.11278643035888672, 0.11285708618164063, 0.11290214538574218, 0.11291340637207031, 0.11281407928466797, 0.11289087677001954, 0.11277823638916015, 0.11302194976806641, 0.11276083374023438, 0.1131304931640625, 0.11284070587158203, 0.11282637023925782, 0.11281407928466797, 0.11287551879882812, 0.11282943725585938, 0.11289702606201171, 0.11277005004882812, 0.11295641326904297, 0.11282637023925782, 0.24168550109863282, 0.11276083374023438, 0.11272294616699219, 0.11288166046142578, 0.11266252899169922, 0.11284377288818359, 0.11275981140136719, 0.11279052734375, 0.11276902770996093, 0.11283865356445312, 0.11269324493408203, 0.112827392578125, 0.11269427490234375, 0.11282943725585938, 0.1127372817993164, 0.11274034881591796, 0.11279564666748047, 0.1127741470336914, 0.11273420715332032, 0.11309977722167969, 0.11271372985839843, 0.11276390075683594, 0.11319500732421875, 0.11277005004882812, 0.11274034881591796, 0.11279052734375, 0.11273216247558594, 0.11281817626953125, 0.11279154968261719, 0.11274649810791015, 0.11277823638916015, 0.11277619171142578, 0.11278233337402344, 0.11285094451904297, 0.11275571441650391, 0.1128253402709961, 0.1128089599609375, 0.11281510162353516, 0.11284070587158203, 0.11279052734375, 0.11275571441650391, 0.11279666900634766, 0.11274240112304687, 0.11281100463867187, 0.11270655822753907, 0.11287551879882812, 0.1129512939453125, 0.11287347412109375, 0.11281510162353516, 0.1128427505493164, 0.112753662109375, 0.1128458251953125, 0.11273318481445313, 0.11288371276855469, 0.11284684753417969, 0.1128458251953125, 0.11282431793212891, 0.11285913848876954, 0.11295334625244141, 0.11295846557617187, 0.11269939422607422, 0.11288883209228516, 0.11284172821044922, 0.2413096923828125, 0.11277516937255859, 0.112753662109375, 0.1127710723876953, 0.11267584228515624, 0.11275571441650391, 0.11283251190185548, 0.11279154968261719, 0.11278336334228516, 0.11276390075683594, 0.11285196685791016, 0.1128253402709961, 0.11271577453613281, 0.11290930938720703, 0.11287449645996094, 0.1127936019897461, 0.1128980484008789, 0.1127936019897461, 0.11277721405029296, 0.11276287841796875, 0.11272601318359375, 0.11300249481201172, 0.11285913848876954, 0.11278540802001953, 0.1127741470336914, 0.1128089599609375, 0.11275775909423828, 0.11280691528320312, 0.11270655822753907, 0.11287449645996094, 0.1129717788696289, 0.11280281829833984, 0.11280793762207031, 0.11283251190185548, 0.1127936019897461, 0.11274854278564453, 0.11282125091552735, 0.11291136169433594, 0.11291545867919922, 0.11276595306396485, 0.11276902770996093, 0.11279666900634766, 0.11284786987304687, 0.11326258850097656, 0.11278131103515625, 0.112901123046875, 0.11286528015136718, 0.11288166046142578, 0.11337522888183593, 0.11286937713623046, 0.11274854278564453, 0.11287551879882812, 0.1127936019897461, 0.11287859344482422, 0.11287654113769531, 0.11280281829833984, 0.11282227325439453, 0.11281510162353516, 0.11282637023925782, 0.1128632354736328, 0.11277823638916015, 0.11286528015136718, 0.11287245178222656, 0.24149810791015625, 0.11275468444824219, 0.1127874526977539, 0.11275571441650391, 0.11290726470947265, 0.11273830413818359, 0.11281510162353516, 0.11276083374023438, 0.11283251190185548, 0.11291340637207031, 0.11277005004882812, 0.11277516937255859, 0.11271270751953125, 0.11282431793212891, 0.11281305694580078, 0.11274342346191406, 0.11277311706542968, 0.11275878143310547, 0.11281613159179688, 0.11301683044433594, 0.1127710723876953, 0.11278438568115234, 0.11309772491455078, 0.11286016082763672, 0.11279872131347657, 0.11275161743164062, 0.1127720947265625, 0.11301785278320313, 0.11272089385986328, 0.11279462432861329, 0.11287449645996094, 0.1128826904296875, 0.1128642578125, 0.1128304672241211, 0.11286528015136718, 0.11281100463867187, 0.11273932647705077, 0.1128427505493164, 0.11290726470947265, 0.11277926635742187, 0.11279257965087891, 0.11287039947509765, 0.11279974365234376, 0.11282125091552735, 0.11274547576904297, 0.1128826904296875, 0.11292364501953125, 0.11282125091552735, 0.11308953857421875, 0.11287347412109375, 0.11285094451904297, 0.11294515228271484, 0.11277516937255859, 0.11285298919677735, 0.11286937713623046, 0.1127874526977539, 0.11277926635742187, 0.1128806381225586, 0.11287039947509765, 0.1128826904296875, 0.11276697540283204, 0.112827392578125, 0.11284377288818359, 0.24074444580078125, 0.11279666900634766, 0.11273420715332032, 0.11283353424072265, 0.11276083374023438, 0.11281715393066406, 0.11274444580078125, 0.11301376342773438, 0.11277823638916015, 0.11279257965087891, 0.11269529724121094, 0.11279462432861329, 0.11273113250732422, 0.11278848266601563, 0.11278336334228516, 0.11278438568115234, 0.1127720947265625, 0.11289600372314453, 0.11275878143310547, 0.11274956512451172, 0.11271270751953125, 0.11279257965087891, 0.1129728012084961, 0.11294924926757813, 0.11274752044677734, 0.112827392578125, 0.11272908782958985, 0.11279257965087891, 0.11279974365234376, 0.11295948791503906, 0.1128642578125, 0.11307622528076172, 0.11276493072509766, 0.11287142181396484, 0.11277619171142578, 0.1128304672241211, 0.1127710723876953, 0.11279462432861329, 0.11315200042724609, 0.11281510162353516, 0.11276185607910157, 0.11284172821044922, 0.11278131103515625, 0.11314278411865235, 0.11274854278564453, 0.11283968353271484, 0.11279769897460938, 0.11283968353271484, 0.11278336334228516, 0.11286732482910156, 0.11274034881591796, 0.11290828704833984, 0.11287551879882812, 0.11286937713623046, 0.11282227325439453, 0.11284070587158203, 0.11276493072509766, 0.11285913848876954, 0.11283968353271484, 0.11293593597412109, 0.11289497375488282, 0.11285504150390625, 0.11289190673828126, 0.241623046875, 0.11288371276855469, 0.11290521240234375, 0.11291033935546875, 0.11266867065429688, 0.11280486297607421, 0.11278643035888672, 0.1128058853149414, 0.11277619171142578, 0.11309977722167969, 0.11277005004882812, 0.1127741470336914, 0.11267788696289062, 0.11282022094726563, 0.11276595306396485, 0.11291238403320313, 0.1127720947265625, 0.11290624237060547, 0.11285913848876954, 0.11280486297607421, 0.11278233337402344, 0.11274444580078125, 0.11285094451904297, 0.11273318481445313, 0.11276493072509766, 0.11286118316650391, 0.11291033935546875, 0.11290419006347656, 0.11285810852050782, 0.11283251190185548, 0.11283353424072265, 0.11278028869628906, 0.11282125091552735, 0.11288985443115235, 0.11281919860839844, 0.11280076599121094, 0.11274956512451172, 0.11288678741455078, 0.11283865356445312, 0.11277005004882812, 0.11273216247558594, 0.11282637023925782, 0.11300556945800781, 0.11299635314941406, 0.11285913848876954, 0.11291340637207031, 0.11292467498779298, 0.11296665954589843, 0.11279666900634766, 0.11282125091552735, 0.11284480285644531, 0.11303731536865234, 0.11290419006347656, 0.11302194976806641, 0.11295027160644532, 0.11285401916503907, 0.11284172821044922, 0.11287142181396484, 0.11282637023925782, 0.11286937713623046, 0.11272601318359375, 0.11284377288818359, 0.11280793762207031, 0.24259686279296874, 0.11276902770996093, 0.11276083374023438, 0.11273318481445313, 0.11276799774169922, 0.11278028869628906, 0.11281613159179688, 0.11273420715332032, 0.11287551879882812, 0.11277619171142578, 0.11276185607910157, 0.11278438568115234, 0.11272294616699219, 0.11282227325439453, 0.11279666900634766, 0.11277311706542968, 0.11297074890136719, 0.11292671966552735, 0.11278540802001953, 0.11278438568115234, 0.11276185607910157, 0.11275263977050781, 0.11281510162353516, 0.11269017791748047, 0.11285094451904297, 0.11281919860839844, 0.112827392578125, 0.11278336334228516, 0.11274547576904297, 0.11281919860839844, 0.11285504150390625, 0.11282841491699219, 0.11279872131347657, 0.11282022094726563, 0.11288985443115235, 0.11278028869628906, 0.11275571441650391, 0.11278438568115234, 0.11286630249023437, 0.11282125091552735, 0.11275468444824219, 0.11284070587158203, 0.11285913848876954, 0.1128642578125, 0.11276185607910157, 0.11285504150390625, 0.11280691528320312, 0.11285196685791016, 0.11281510162353516, 0.11279974365234376, 0.11277005004882812, 0.1128622055053711, 0.1128806381225586, 0.112932861328125, 0.1128427505493164, 0.11276493072509766, 0.11301580810546875, 0.1128632354736328, 0.11285504150390625, 0.11286937713623046, 0.11273625946044921, 0.11287654113769531, 0.11285196685791016, 0.2427852783203125, 0.11275775909423828, 0.11284992218017578, 0.11281305694580078, 0.11277516937255859, 0.11296051025390624, 0.1127936019897461, 0.11277823638916015, 0.11278336334228516, 0.11274854278564453, 0.11275161743164062, 0.11278643035888672, 0.11275878143310547, 0.11278950500488281, 0.1127372817993164, 0.11285196685791016, 0.11292262268066407, 0.11279257965087891, 0.11278028869628906, 0.11280076599121094, 0.11271270751953125, 0.11273011016845703, 0.11284377288818359, 0.11274956512451172, 0.1127741470336914, 0.11278848266601563, 0.11275468444824219, 0.1127710723876953, 0.11287859344482422, 0.11274137878417968, 0.11282227325439453, 0.11275878143310547, 0.11282943725585938, 0.11286630249023437, 0.11284786987304687, 0.11277823638916015, 0.11277619171142578, 0.11280281829833984, 0.11295027160644532, 0.11294105529785156, 0.11276697540283204, 0.11280178833007813, 0.11305677032470703, 0.11287449645996094, 0.11272601318359375, 0.1128304672241211, 0.11289395141601563, 0.11276799774169922, 0.11281715393066406, 0.11280076599121094, 0.11278028869628906, 0.11290930938720703, 0.11271577453613281, 0.11283353424072265, 0.11283558654785156, 0.11279872131347657, 0.11294310760498047, 0.11278233337402344, 0.11284684753417969, 0.1128980484008789, 0.11275263977050781, 0.11282329559326172, 0.11282431793212891, 0.24283648681640624, 0.11278438568115234, 0.11280281829833984, 0.11283455657958984, 0.1127014389038086, 0.11282637023925782, 0.11276390075683594, 0.11279872131347657, 0.11274240112304687, 0.11280793762207031, 0.11270758056640626, 0.11278233337402344, 0.11269222259521484, 0.112932861328125, 0.11276083374023438, 0.11274752044677734, 0.11294719696044922, 0.1127874526977539, 0.11278950500488281, 0.11275981140136719, 0.11270041656494141, 0.11275981140136719, 0.11282125091552735, 0.11272806549072266, 0.11271270751953125, 0.1127720947265625, 0.11274137878417968, 0.11284889221191406, 0.11273830413818359, 0.11278438568115234, 0.11283455657958984, 0.11283968353271484, 0.11276697540283204, 0.11281613159179688, 0.11281613159179688, 0.11278540802001953, 0.11278848266601563, 0.11285504150390625, 0.11283763122558593, 0.11283660888671875, 0.11276083374023438, 0.11286937713623046, 0.11273113250732422, 0.11288985443115235, 0.11271372985839843, 0.11285094451904297, 0.11282329559326172, 0.11287039947509765, 0.11279564666748047, 0.11285298919677735, 0.11290930938720703, 0.11287039947509765, 0.11271475219726562, 0.11291340637207031, 0.11277311706542968, 0.11281613159179688, 0.11278438568115234, 0.11281817626953125, 0.11285606384277344, 0.11285504150390625, 0.11276595306396485, 0.11286118316650391, 0.11281715393066406]",tokens/s,8.720457255771393,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1606.115328,7751.59808,0.0,7105.150976,6339.650048,s,10,3.398825378417969,0.3398825378417969,0.0001796787239087705,0.3398072967529297,0.34014054260253906,0.34015884857177736,0.34017349334716795,"[0.3401331481933594, 0.340136474609375, 0.3396676330566406, 0.3397496337890625, 0.339798583984375, 0.3398160095214844, 0.3398270568847656, 0.33977096557617187, 0.33974871826171876, 0.3401771545410156]",tokens/s,753.2013901789765,kWh,4.014880232237004e-06,2.1994003784311643e-06,2.0678386913066762e-05,2.6892667523734932e-05,tokens/kWh,9519323.427996106,MB,1606.115328,7751.59808,0.0,7105.150976,6492.977664,s,10,191.87401171875,19.187401171875003,0.0005743862691443143,19.187277343749997,19.188323046875,19.1883529296875,19.1883768359375,"[19.18681640625, 19.1883828125, 19.186537109375, 19.186984375, 19.187267578125, 19.1878203125, 19.187234375, 19.187365234375, 19.18831640625, 19.187287109375]",tokens/s,3.2834045338222118,kWh,0.00022652148161773333,0.00012415257496371852,0.0011781508221495336,0.0015288248787309856,tokens/kWh,41208.12061371849,,s,629,194.62795996093757,0.3094244196517289,0.04037499038373234,0.3045498962402344,0.30476737670898435,0.30489190673828126,0.6443736816406249,"[0.30475674438476563, 0.3042539367675781, 0.30429183959960937, 0.30432766723632815, 0.3043932189941406, 0.30428466796875, 0.30440243530273436, 0.30426931762695314, 0.3042979736328125, 0.30441677856445315, 0.30461236572265626, 0.3044607849121094, 0.304363525390625, 0.3047966613769531, 0.30441473388671875, 0.30442291259765625, 0.3044607849121094, 0.3044700012207031, 0.30437991333007813, 0.3044259948730469, 0.30444134521484373, 0.3044208679199219, 0.3044259948730469, 0.3045458068847656, 0.3044618225097656, 0.30449151611328124, 0.30450177001953127, 0.30444338989257813, 0.30464102172851565, 0.30468814086914064, 0.3046911926269531, 0.30451815795898435, 0.3048734741210937, 0.30446591186523436, 0.30462872314453127, 0.30454373168945315, 0.30467071533203127, 0.3045621643066406, 0.30447921752929685, 0.30443109130859375, 0.3044239501953125, 0.30453350830078124, 0.30460107421875, 0.3046000671386719, 0.30463079833984374, 0.30458367919921875, 0.30481304931640624, 0.30464102172851565, 0.30491546630859373, 0.30471578979492187, 0.3046144104003906, 0.3046277160644531, 0.30457550048828125, 0.30456832885742186, 0.30468814086914064, 0.30470245361328124, 0.3046748046875, 0.30466049194335937, 0.30459597778320313, 0.304690185546875, 0.30472601318359377, 0.3046666259765625, 0.64500634765625, 0.304763916015625, 0.3044136962890625, 0.3045724182128906, 0.30427749633789064, 0.30443008422851564, 0.3044013977050781, 0.3044116516113281, 0.30428262329101563, 0.30454373168945315, 0.30441677856445315, 0.3044894714355469, 0.3045038146972656, 0.3043870849609375, 0.3044822998046875, 0.30467892456054685, 0.3047976989746094, 0.30445977783203126, 0.30481304931640624, 0.3045252990722656, 0.30444338989257813, 0.30444442749023437, 0.30442495727539065, 0.3045580749511719, 0.3046277160644531, 0.30449972534179687, 0.3045252990722656, 0.3043420104980469, 0.3045877685546875, 0.30458367919921875, 0.30439630126953127, 0.3044566955566406, 0.3048796081542969, 0.30451608276367187, 0.30490008544921876, 0.3046451110839844, 0.3045519409179688, 0.304500732421875, 0.30461029052734373, 0.30453863525390623, 0.30451507568359376, 0.30459085083007814, 0.30464102172851565, 0.30447308349609375, 0.3045222473144531, 0.30469940185546873, 0.30460519409179687, 0.30456320190429687, 0.3047188415527344, 0.30492364501953123, 0.3045458068847656, 0.30499737548828126, 0.3044700012207031, 0.30495437622070315, 0.304606201171875, 0.30459597778320313, 0.30468402099609376, 0.3046666259765625, 0.30475775146484374, 0.30444134521484373, 0.3046533203125, 0.30475979614257814, 0.3046666259765625, 0.644221923828125, 0.30472601318359377, 0.3043768310546875, 0.30464102172851565, 0.30431845092773435, 0.3043645324707031, 0.3043645324707031, 0.30440447998046877, 0.30438296508789064, 0.30432257080078123, 0.30430209350585935, 0.3044085693359375, 0.3043870849609375, 0.3043512268066406, 0.30444134521484373, 0.30434100341796877, 0.30433279418945314, 0.3048509521484375, 0.30440756225585935, 0.30436044311523436, 0.304584716796875, 0.304505859375, 0.3044065246582031, 0.304505859375, 0.3044423828125, 0.3047055358886719, 0.3044843444824219, 0.3046451110839844, 0.30447308349609375, 0.30453350830078124, 0.3045089416503906, 0.30458880615234374, 0.3047055358886719, 0.3044894714355469, 0.3047690124511719, 0.3045877685546875, 0.3047823486328125, 0.3045222473144531, 0.3046082458496094, 0.30447412109375, 0.30453350830078124, 0.30452633666992185, 0.30465432739257814, 0.30454373168945315, 0.30453863525390623, 0.3045529479980469, 0.3045672912597656, 0.30466867065429687, 0.30456832885742186, 0.30453350830078124, 0.3046830139160156, 0.3045898132324219, 0.3047526550292969, 0.30457138061523437, 0.3045396423339844, 0.30459902954101564, 0.3050967102050781, 0.3045928955078125, 0.304869384765625, 0.30459393310546873, 0.30474957275390624, 0.30460519409179687, 0.304648193359375, 0.6444441528320313, 0.3045222473144531, 0.3046420593261719, 0.30442495727539065, 0.3043133544921875, 0.3046400146484375, 0.30432461547851564, 0.3043481750488281, 0.304405517578125, 0.3047004089355469, 0.30430413818359375, 0.30436557006835935, 0.3045478515625, 0.3045089416503906, 0.3043420104980469, 0.3045191650390625, 0.3044188232421875, 0.3044925537109375, 0.30435736083984377, 0.3046574096679687, 0.3043819580078125, 0.30489190673828126, 0.30441265869140627, 0.30437374877929685, 0.30438296508789064, 0.3044751281738281, 0.30437786865234373, 0.304458740234375, 0.3044208679199219, 0.30458367919921875, 0.30448025512695315, 0.30448538208007814, 0.3044945983886719, 0.3045591125488281, 0.3045928955078125, 0.30496563720703124, 0.30453555297851564, 0.30449868774414063, 0.3048529968261719, 0.3045877685546875, 0.3043932189941406, 0.3047137145996094, 0.30466253662109377, 0.3045191650390625, 0.3044822998046875, 0.30458367919921875, 0.3046911926269531, 0.3045621643066406, 0.3046400146484375, 0.3045928955078125, 0.3045191650390625, 0.304542724609375, 0.3050680236816406, 0.3045539855957031, 0.30495437622070315, 0.3046635437011719, 0.30473727416992186, 0.30457855224609376, 0.3047034912109375, 0.3045928955078125, 0.304648193359375, 0.304648193359375, 0.3046358947753906, 0.6444646606445312, 0.30448843383789065, 0.30429901123046876, 0.3041771545410156, 0.30473214721679687, 0.30434100341796877, 0.30440447998046877, 0.30466867065429687, 0.30458880615234374, 0.304468994140625, 0.3043092346191406, 0.30449868774414063, 0.3044474792480469, 0.30434713745117187, 0.30449151611328124, 0.3043891296386719, 0.3043973083496094, 0.3046860656738281, 0.30452120971679686, 0.3044700012207031, 0.30440243530273436, 0.3046553649902344, 0.30435226440429686, 0.3047424011230469, 0.3046697082519531, 0.3044976501464844, 0.3044771728515625, 0.3045498962402344, 0.3044638671875, 0.3046031494140625, 0.3045252990722656, 0.30449972534179687, 0.3045140380859375, 0.30445159912109376, 0.3045304260253906, 0.3045539855957031, 0.3044874267578125, 0.3049830322265625, 0.30450994873046877, 0.3045396423339844, 0.30475775146484374, 0.3045222473144531, 0.30458880615234374, 0.3045089416503906, 0.3045529479980469, 0.3044700012207031, 0.30457754516601565, 0.30464102172851565, 0.304869384765625, 0.3045724182128906, 0.30462872314453127, 0.3046973571777344, 0.304616455078125, 0.3048714294433594, 0.304637939453125, 0.30467071533203127, 0.30475775146484374, 0.3047106628417969, 0.304637939453125, 0.30467684936523437, 0.30463897705078125, 0.3046072387695313, 0.3046400146484375, 0.644495361328125, 0.30448333740234373, 0.30435736083984377, 0.3042928771972656, 0.3043502197265625, 0.30474343872070314, 0.30459185791015625, 0.304427001953125, 0.30459393310546873, 0.3044505615234375, 0.30435226440429686, 0.30443521118164063, 0.30451608276367187, 0.3044618225097656, 0.30436044311523436, 0.30443826293945314, 0.30443109130859375, 0.3045048217773437, 0.30441778564453126, 0.30472909545898436, 0.30432257080078123, 0.30451815795898435, 0.30475161743164064, 0.3044198303222656, 0.30478744506835936, 0.30471578979492187, 0.3044474792480469, 0.30443521118164063, 0.3045458068847656, 0.30457855224609376, 0.30445773315429686, 0.30441473388671875, 0.30459902954101564, 0.3044894714355469, 0.3044669494628906, 0.30454373168945315, 0.30460211181640623, 0.3046277160644531, 0.3049144287109375, 0.30460415649414063, 0.3044566955566406, 0.304785400390625, 0.304716796875, 0.30463284301757815, 0.30456524658203127, 0.30466049194335937, 0.3045693359375, 0.3046133728027344, 0.3046246337890625, 0.3046369323730469, 0.304564208984375, 0.3045898132324219, 0.3046144104003906, 0.30452838134765625, 0.30464102172851565, 0.30493798828125, 0.3047250061035156, 0.3049830322265625, 0.3047127075195312, 0.304606201171875, 0.3046133728027344, 0.3046174621582031, 0.30467379760742186, 0.6443817138671875, 0.30435531616210937, 0.30422219848632814, 0.3042764892578125, 0.3044116516113281, 0.30445773315429686, 0.30447308349609375, 0.30473626708984375, 0.30443008422851564, 0.30441265869140627, 0.30463897705078125, 0.3045314636230469, 0.30438092041015624, 0.3043932189941406, 0.30443212890625, 0.3043665771484375, 0.304400390625, 0.304837646484375, 0.3044454345703125, 0.30445977783203126, 0.30457958984375, 0.3045242919921875, 0.30442291259765625, 0.30445159912109376, 0.3047331848144531, 0.30453863525390623, 0.304869384765625, 0.3046031494140625, 0.3044607849121094, 0.3046297607421875, 0.30459085083007814, 0.3044566955566406, 0.30446795654296877, 0.30453350830078124, 0.3045222473144531, 0.3045601196289062, 0.304532470703125, 0.3045089416503906, 0.3045877685546875, 0.3045867614746094, 0.3047331848144531, 0.30458367919921875, 0.30455706787109377, 0.30482125854492187, 0.30450177001953127, 0.3045498962402344, 0.30447103881835935, 0.30464715576171875, 0.30456524658203127, 0.30449868774414063, 0.30496563720703124, 0.304679931640625, 0.3046369323730469, 0.3045458068847656, 0.3045693359375, 0.3046277160644531, 0.30489190673828126, 0.30474343872070314, 0.3046533203125, 0.3046860656738281, 0.30468402099609376, 0.30468197631835936, 0.30458367919921875, 0.6444154663085937, 0.3043891296386719, 0.3043133544921875, 0.30430514526367186, 0.30449151611328124, 0.30438296508789064, 0.30441574096679686, 0.3043512268066406, 0.3043614807128906, 0.30452633666992185, 0.3043502197265625, 0.304421875, 0.3047782287597656, 0.3045242919921875, 0.304427001953125, 0.3044331665039062, 0.3043891296386719, 0.3044638671875, 0.304400390625, 0.3045027770996094, 0.30462667846679686, 0.30457037353515626, 0.3045928955078125, 0.3045498962402344, 0.3043819580078125, 0.3048796081542969, 0.30445159912109376, 0.30474649047851565, 0.3044751281738281, 0.30451507568359376, 0.30451815795898435, 0.3044085693359375, 0.30458572387695315, 0.3044669494628906, 0.3044843444824219, 0.30457650756835936, 0.3044894714355469, 0.3045560302734375, 0.3046533203125, 0.304584716796875, 0.3045447692871094, 0.30480078125, 0.30474957275390624, 0.3044894714355469, 0.30452633666992185, 0.3049359436035156, 0.3046000671386719, 0.3044976501464844, 0.30457958984375, 0.30448333740234373, 0.30450686645507813, 0.3046092834472656, 0.3045928955078125, 0.30478643798828126, 0.3046492309570312, 0.3046574096679687, 0.3046973571777344, 0.3045980224609375, 0.30497381591796874, 0.3045119934082031, 0.30507928466796874, 0.304822265625, 0.30474649047851565, 0.64435302734375, 0.30439013671875, 0.3046666259765625, 0.30432461547851564, 0.30434100341796877, 0.304294921875, 0.304331787109375, 0.3043788757324219, 0.30436248779296876, 0.3043665771484375, 0.30476287841796873, 0.30451608276367187, 0.30432257080078123, 0.30437478637695314, 0.3048458251953125, 0.3043512268066406, 0.3044894714355469, 0.30464614868164064, 0.3044894714355469, 0.3044085693359375, 0.3044669494628906, 0.3044331665039062, 0.30443008422851564, 0.3045816345214844, 0.3044423828125, 0.3044290466308594, 0.30440756225585935, 0.3048161315917969, 0.304532470703125, 0.3049717712402344, 0.3045038146972656, 0.3045027770996094, 0.3044331665039062, 0.30450177001953127, 0.3045826416015625, 0.3046082458496094, 0.30457855224609376, 0.3047208862304687, 0.3044700012207031, 0.30459597778320313, 0.30450994873046877, 0.30466253662109377, 0.3046072387695313, 0.30487655639648437, 0.3046246337890625, 0.304611328125, 0.304711669921875, 0.30482022094726563, 0.3044925537109375, 0.3044607849121094, 0.3047280578613281, 0.30459902954101564, 0.30459698486328124, 0.3045816345214844, 0.30468914794921875, 0.30466867065429687, 0.30472705078125, 0.3048253479003906, 0.3050352783203125, 0.3046133728027344, 0.3048192138671875, 0.3045826416015625, 0.30507211303710935, 0.645001220703125, 0.30450177001953127, 0.3042662353515625, 0.30435531616210937, 0.30431130981445315, 0.3043143615722656, 0.3043717041015625, 0.30475057983398435, 0.3044505615234375, 0.3044505615234375, 0.3045580749511719, 0.3043348388671875, 0.30459085083007814, 0.3044546508789063, 0.3043645324707031, 0.3044464721679688, 0.3043932189941406, 0.3045805969238281, 0.3044566955566406, 0.30447412109375, 0.3044208679199219, 0.30441473388671875, 0.30437274169921874, 0.3045509033203125, 0.30446490478515625, 0.30469528198242185, 0.3044034423828125, 0.3043870849609375, 0.30474649047851565, 0.30448025512695315, 0.30467379760742186, 0.30461236572265626, 0.304427001953125, 0.30441061401367187, 0.3044751281738281, 0.304510986328125, 0.3045898132324219, 0.30455502319335936, 0.3045447692871094, 0.3045478515625, 0.30445773315429686, 0.3046195068359375, 0.30470449829101565, 0.3048529968261719, 0.3047669677734375, 0.304763916015625, 0.30464306640625, 0.30459393310546873, 0.30501580810546874, 0.30453863525390623, 0.3045222473144531, 0.3046563720703125, 0.3047004089355469, 0.3045509033203125, 0.30455502319335936, 0.30465228271484374, 0.3047208862304687, 0.30474853515625, 0.304763916015625, 0.30464715576171875, 0.30462362670898435, 0.304964599609375, 0.30475979614257814]",tokens/s,3.2318069825437337,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-125m,facebook/opt-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1189.961728,860.356608,0.0,213.909504,170.743808,s,10,0.1959003505706787,0.019590035057067873,0.0010233483732580177,0.01937462329864502,0.020302376747131347,0.021179235935211177,0.021880723285675047,"[0.022056095123291014, 0.018721183776855468, 0.018744319915771485, 0.019989568710327147, 0.020107519149780272, 0.018741504669189453, 0.018763519287109374, 0.018740480422973632, 0.019985727310180664, 0.020050432205200194]",tokens/s,13067.8683960618,kWh,2.301899942545679e-07,1.2613136213985853e-07,4.833423667859141e-07,8.396637231803405e-07,tokens/kWh,304883958.8190915,MB,1189.961728,860.356608,0.0,213.909504,191.396864,s,10,12.166089111328125,1.2166089111328127,0.0319861550794473,1.2338013916015624,1.2383833374023436,1.2499125549316406,1.259135928955078,"[1.2614417724609375, 1.1966971435546876, 1.181048828125, 1.234597412109375, 1.2346832275390625, 1.14687353515625, 1.2073231201171875, 1.2342279052734375, 1.2358212890625, 1.2333748779296876]",tokens/s,51.78328008574197,kWh,1.4253392052042297e-05,7.809042589724415e-06,2.786581920141153e-05,4.992825384317823e-05,tokens/kWh,1261810.601225498,,s,629,12.318024711608887,0.019583505105896484,0.0023829722213566315,0.019557376861572266,0.01972305908203125,0.02029158363342285,0.03738846145629883,"[0.020954111099243163, 0.02042470359802246, 0.020282367706298828, 0.02053222465515137, 0.020568063735961914, 0.02066227149963379, 0.020767744064331056, 0.020784128189086915, 0.02030284881591797, 0.02060697555541992, 0.020592639923095703, 0.020899839401245117, 0.020716543197631835, 0.020485120773315428, 0.020345855712890625, 0.0206878719329834, 0.02050764846801758, 0.020297727584838866, 0.020989952087402345, 0.020699136734008788, 0.020332544326782227, 0.020163583755493163, 0.020156415939331054, 0.020057088851928712, 0.020076543807983398, 0.020504575729370117, 0.01987276840209961, 0.019698688507080078, 0.019663871765136717, 0.019551231384277345, 0.01963007926940918, 0.019599359512329103, 0.019568639755249022, 0.01963212776184082, 0.019991552352905274, 0.01957683181762695, 0.01958502388000488, 0.019543039321899415, 0.019604480743408204, 0.019619840621948242, 0.019553279876708983, 0.019523584365844726, 0.01966080093383789, 0.019588096618652344, 0.019532800674438477, 0.01963315200805664, 0.019535871505737306, 0.019535871505737306, 0.019550207138061524, 0.019558399200439454, 0.01962393569946289, 0.019531776428222656, 0.019560447692871095, 0.01969152069091797, 0.019631103515625, 0.019572736740112305, 0.019572736740112305, 0.019575807571411134, 0.019619840621948242, 0.019572736740112305, 0.01948569679260254, 0.019566591262817384, 0.03738627243041992, 0.018154464721679687, 0.018157567977905274, 0.018150400161743165, 0.0182108154296875, 0.01820569610595703, 0.0182108154296875, 0.018380800247192384, 0.018157567977905274, 0.018233343124389647, 0.018192384719848635, 0.018266111373901366, 0.018149375915527344, 0.01822412872314453, 0.018326528549194337, 0.01826918411254883, 0.01822719955444336, 0.018206720352172853, 0.018197504043579102, 0.01816985511779785, 0.018214912414550782, 0.018164735794067383, 0.0182794246673584, 0.018106367111206053, 0.01821388816833496, 0.01857535934448242, 0.01816268730163574, 0.019144704818725586, 0.019530752182006835, 0.019603456497192383, 0.0196177921295166, 0.019500032424926757, 0.019572736740112305, 0.019545087814331053, 0.019552255630493166, 0.019547136306762695, 0.01966592025756836, 0.019560447692871095, 0.019570688247680663, 0.019511295318603517, 0.019537919998168944, 0.01943142318725586, 0.019607551574707033, 0.01963417625427246, 0.01958502388000488, 0.019562496185302734, 0.01963827133178711, 0.019575807571411134, 0.019573759078979493, 0.019533824920654298, 0.019612672805786133, 0.01959321594238281, 0.019735551834106444, 0.01957891273498535, 0.01953481674194336, 0.019566591262817384, 0.019574783325195313, 0.019582975387573243, 0.019554304122924804, 0.01964339256286621, 0.019540992736816407, 0.019548160552978516, 0.019606527328491212, 0.03739648056030274, 0.01824358367919922, 0.018276351928710938, 0.018159616470336915, 0.01822105598449707, 0.0182609920501709, 0.01824051284790039, 0.01823846435546875, 0.01822412872314453, 0.018264064788818358, 0.01821286392211914, 0.01829478454589844, 0.018601984024047852, 0.018198528289794923, 0.01818009567260742, 0.018249727249145507, 0.01821388816833496, 0.01824665641784668, 0.01821900749206543, 0.018145280838012694, 0.018276351928710938, 0.01815449523925781, 0.018118656158447266, 0.018185216903686522, 0.01818009567260742, 0.0182476806640625, 0.018157567977905274, 0.01822412872314453, 0.01822003173828125, 0.018184192657470705, 0.018136064529418947, 0.01823027229309082, 0.01817087936401367, 0.01837772750854492, 0.018481151580810547, 0.01847603225708008, 0.018388992309570314, 0.018203712463378905, 0.018227136611938477, 0.01820569610595703, 0.018513919830322266, 0.020970495223999023, 0.01985740852355957, 0.019615743637084963, 0.019555328369140625, 0.019587072372436523, 0.019527679443359376, 0.019525632858276368, 0.019548160552978516, 0.01964339256286621, 0.019611648559570313, 0.019594240188598632, 0.019577856063842772, 0.01964339256286621, 0.019556352615356445, 0.01961881637573242, 0.019572736740112305, 0.019711999893188475, 0.019596288681030274, 0.01946931266784668, 0.019466239929199217, 0.01961881637573242, 0.019501056671142578, 0.04007424163818359, 0.019982336044311523, 0.019549184799194336, 0.019538944244384765, 0.01963212776184082, 0.019587072372436523, 0.019568639755249022, 0.019598335266113282, 0.019677183151245118, 0.01964339256286621, 0.0196997127532959, 0.01963724708557129, 0.01960140800476074, 0.019594240188598632, 0.019583999633789064, 0.019521535873413084, 0.01963417625427246, 0.019558399200439454, 0.01946112060546875, 0.01943756866455078, 0.019561471939086913, 0.01963929557800293, 0.019555328369140625, 0.019528703689575197, 0.019534847259521485, 0.01962700843811035, 0.019612672805786133, 0.019558399200439454, 0.019587072372436523, 0.019590143203735352, 0.019615743637084963, 0.019552255630493166, 0.019547136306762695, 0.01960038375854492, 0.019570688247680663, 0.019604480743408204, 0.019610624313354492, 0.019544063568115236, 0.01947750473022461, 0.019571712493896484, 0.01958198356628418, 0.01954198455810547, 0.019543039321899415, 0.019545087814331053, 0.019611648559570313, 0.019611648559570313, 0.019506175994873046, 0.01946828842163086, 0.01965363121032715, 0.019554304122924804, 0.019501056671142578, 0.020116479873657226, 0.019562496185302734, 0.019538944244384765, 0.01959219169616699, 0.019552255630493166, 0.019556352615356445, 0.01947340774536133, 0.019664896011352538, 0.019647487640380858, 0.019602432250976562, 0.019554304122924804, 0.019566591262817384, 0.04012441635131836, 0.019507200241088866, 0.019568639755249022, 0.019520511627197267, 0.01946214485168457, 0.019583999633789064, 0.01960038375854492, 0.01966080093383789, 0.019688447952270507, 0.01967411231994629, 0.019579904556274414, 0.019631103515625, 0.01964134407043457, 0.019552255630493166, 0.01965465545654297, 0.019544063568115236, 0.019631103515625, 0.01962495994567871, 0.019570688247680663, 0.019612672805786133, 0.019560447692871095, 0.019519487380981446, 0.019546112060546874, 0.019504127502441407, 0.019583999633789064, 0.019588096618652344, 0.019529727935791014, 0.01959219169616699, 0.019548160552978516, 0.019536895751953123, 0.019577856063842772, 0.019619840621948242, 0.019551231384277345, 0.019515392303466796, 0.019559423446655275, 0.019523584365844726, 0.019549184799194336, 0.019587072372436523, 0.019784704208374023, 0.019539968490600586, 0.019999744415283204, 0.019573759078979493, 0.019604480743408204, 0.0194652156829834, 0.019555328369140625, 0.020063232421875, 0.019693567276000978, 0.019557376861572266, 0.019511295318603517, 0.019560447692871095, 0.019519487380981446, 0.019636224746704102, 0.019534847259521485, 0.019510271072387696, 0.019554304122924804, 0.019515392303466796, 0.019604480743408204, 0.019573759078979493, 0.019559423446655275, 0.01963929557800293, 0.01962700843811035, 0.019518463134765626, 0.0196997127532959, 0.037389312744140625, 0.01823232078552246, 0.018184192657470705, 0.01819443130493164, 0.018266111373901366, 0.018296831130981444, 0.01817087936401367, 0.01820364761352539, 0.018176000595092775, 0.01821286392211914, 0.01819340705871582, 0.018344959259033202, 0.01821183967590332, 0.01821183967590332, 0.018226175308227538, 0.01818009567260742, 0.01822822380065918, 0.01821183967590332, 0.018174976348876954, 0.018164735794067383, 0.01820569610595703, 0.01843814468383789, 0.018155519485473632, 0.01821900749206543, 0.018125823974609375, 0.018139135360717772, 0.018066432952880858, 0.018139135360717772, 0.018531328201293946, 0.01824358367919922, 0.018174976348876954, 0.0181494083404541, 0.018142175674438477, 0.018144256591796876, 0.018242559432983398, 0.018108415603637695, 0.018157567977905274, 0.01819647979736328, 0.018183168411254884, 0.018158592224121094, 0.01817087936401367, 0.018284543991088868, 0.018093055725097656, 0.018166784286499024, 0.018137088775634767, 0.018189311981201172, 0.018338815689086914, 0.018275327682495117, 0.018127872467041017, 0.01820467185974121, 0.018148351669311523, 0.01821900749206543, 0.01808793640136719, 0.018124799728393554, 0.018150400161743165, 0.018116607666015624, 0.018328575134277342, 0.01821900749206543, 0.01820569610595703, 0.01816985511779785, 0.018148351669311523, 0.018176000595092775, 0.01819647979736328, 0.03731254577636719, 0.018205663681030274, 0.018168832778930662, 0.018190336227416993, 0.018148351669311523, 0.018160640716552736, 0.018159616470336915, 0.01823846435546875, 0.018206720352172853, 0.018249727249145507, 0.01823641586303711, 0.018167808532714845, 0.018159616470336915, 0.018198528289794923, 0.01821388816833496, 0.01821696090698242, 0.018347007751464844, 0.01824051284790039, 0.018160640716552736, 0.018929664611816405, 0.019581951141357423, 0.019544063568115236, 0.019556352615356445, 0.01966592025756836, 0.01946931266784668, 0.01960038375854492, 0.01962495994567871, 0.019525632858276368, 0.01963212776184082, 0.019564544677734375, 0.01964134407043457, 0.019534847259521485, 0.019706880569458008, 0.019734527587890623, 0.019536895751953123, 0.01964236831665039, 0.019714048385620117, 0.0192542724609375, 0.01961881637573242, 0.019769344329833984, 0.019588096618652344, 0.019536895751953123, 0.019538944244384765, 0.019577856063842772, 0.019558399200439454, 0.019537919998168944, 0.019794944763183595, 0.019558399200439454, 0.019546112060546874, 0.019570688247680663, 0.019512319564819337, 0.019525632858276368, 0.019516416549682617, 0.01944883155822754, 0.01964031982421875, 0.019586048126220702, 0.019599359512329103, 0.01960038375854492, 0.01960038375854492, 0.019544063568115236, 0.019568639755249022, 0.019663871765136717, 0.01960960006713867, 0.0374200325012207, 0.018233343124389647, 0.01944268798828125, 0.020205568313598633, 0.01985228729248047, 0.019703807830810546, 0.020083711624145507, 0.019512319564819337, 0.01945497512817383, 0.019554304122924804, 0.019573759078979493, 0.019563520431518554, 0.019555328369140625, 0.01963417625427246, 0.01963929557800293, 0.019603456497192383, 0.01964339256286621, 0.0195020809173584, 0.01950111961364746, 0.019572736740112305, 0.01955731201171875, 0.019578880310058593, 0.01967616081237793, 0.02067251205444336, 0.01993011283874512, 0.019767295837402343, 0.019664896011352538, 0.01948467254638672, 0.019550207138061524, 0.01963724708557129, 0.019551231384277345, 0.0194652156829834, 0.019545087814331053, 0.01959219169616699, 0.019512319564819337, 0.019516416549682617, 0.01962291145324707, 0.019572736740112305, 0.019574783325195313, 0.019531776428222656, 0.019543039321899415, 0.019579904556274414, 0.01960857582092285, 0.01962393569946289, 0.019769344329833984, 0.019540992736816407, 0.019471359252929688, 0.019556352615356445, 0.01946214485168457, 0.019763200759887696, 0.019562496185302734, 0.019558399200439454, 0.019575807571411134, 0.019524608612060547, 0.019578880310058593, 0.01962294387817383, 0.01953276824951172, 0.019543039321899415, 0.019977216720581056, 0.01957683181762695, 0.019551231384277345, 0.019655679702758787, 0.01960857582092285, 0.04004556655883789, 0.019578880310058593, 0.019562496185302734, 0.019501056671142578, 0.019458047866821288, 0.019745792388916016, 0.019575807571411134, 0.019598335266113282, 0.01959321594238281, 0.019606527328491212, 0.019698688507080078, 0.01969254493713379, 0.01963315200805664, 0.01981644821166992, 0.01962598419189453, 0.019631103515625, 0.019644416809082032, 0.01970278358459473, 0.019922943115234376, 0.019612672805786133, 0.01960140800476074, 0.01960140800476074, 0.019581951141357423, 0.01960140800476074, 0.01960857582092285, 0.01958502388000488, 0.019740671157836915, 0.019596288681030274, 0.019589120864868165, 0.019586048126220702, 0.019589120864868165, 0.019544063568115236, 0.019588096618652344, 0.019516416549682617, 0.019517440795898438, 0.01962393569946289, 0.01965363121032715, 0.01965056037902832, 0.019519487380981446, 0.019560447692871095, 0.019532800674438477, 0.01961881637573242, 0.019560447692871095, 0.019529727935791014, 0.020074495315551756, 0.019489791870117186, 0.019594240188598632, 0.019582975387573243, 0.019606527328491212, 0.019512319564819337, 0.019510271072387696, 0.01957481575012207, 0.019650527954101563, 0.019710975646972655, 0.01962291145324707, 0.019509279251098632, 0.01963926315307617, 0.019560447692871095, 0.019543039321899415, 0.019517440795898438, 0.019553279876708983, 0.019552255630493166, 0.019688447952270507, 0.04007424163818359, 0.019550207138061524, 0.019545087814331053, 0.019547136306762695, 0.019542015075683594, 0.019607551574707033, 0.019512319564819337, 0.019720191955566405, 0.019548160552978516, 0.019571712493896484, 0.019580928802490235, 0.019500032424926757, 0.019478527069091797, 0.01963007926940918, 0.019588096618652344, 0.019546112060546874, 0.01968639945983887, 0.019561471939086913, 0.019581951141357423, 0.019511295318603517, 0.01948876762390137, 0.0194652156829834, 0.019535871505737306, 0.019500032424926757, 0.01964031982421875, 0.019557376861572266, 0.019516416549682617, 0.019527679443359376, 0.01962598419189453, 0.01949286460876465, 0.019511295318603517, 0.0198922233581543, 0.019628032684326172, 0.019589120864868165, 0.019573759078979493, 0.019551231384277345, 0.019553279876708983, 0.01962291145324707, 0.019566591262817384, 0.01945497512817383, 0.019610624313354492, 0.019580928802490235, 0.019589120864868165, 0.019628032684326172, 0.019550207138061524, 0.019540992736816407, 0.019580928802490235, 0.019573759078979493, 0.019679231643676756, 0.01960038375854492, 0.019539968490600586, 0.019647487640380858, 0.019408895492553712, 0.019381248474121093, 0.01969254493713379, 0.019680255889892577, 0.019569664001464843, 0.019570688247680663, 0.01955638313293457, 0.019572704315185548, 0.019602432250976562, 0.019544063568115236, 0.019591167449951173]",tokens/s,51.063381891677075,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,x,x,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch @@ -5670,7 +6056,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c13ab-0cc08b26063c8dbb3728feb9;0a8e7fb5-be22-445d-934c-40187f1d3f7e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45c4-3a39d52e6e534d4d3ffe5a00;1bb23040-a220-4686-9d50-838b4cc349cc) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5730,7 +6116,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee9a-789d9ef260f055a741ff1458;34029ec6-9980-41e1-b7e1-aa3ff9b0ec01) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3fa9-1c53d6b7333c095c534b7623;eb5e72ee-4144-4815-9d8e-87a5c3531fdb) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -5962,7 +6348,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp0im2uvrl/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp0sex1a8k/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -6059,7 +6445,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bed11-6d775b554fe494c53f27005f;b64dd5f6-be2e-4edf-b31b-2a7eee56546c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3e28-5ea2bd8b570db9e02c2acb53;f8606aa7-a488-47d5-850b-3fe042a30cc5) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -6144,7 +6530,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1334-5ba466035aca3b16162cc699;d4e4fa20-820a-4152-b6e4-dacd7de5e712) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4551-4d8595044435dbb928cfacd8;c77d3a2a-4523-4d7e-9508-9896353299d3) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6260,7 +6646,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1233-3887479126a01f833bd55f16;3c2f5b6c-b16e-44e1-8504-6854aff7ceb0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d448b-02be5df9729a28647a87a061;07482305-009c-4ced-8f54-d446146e6076) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6334,6 +6720,35 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1844.498432,3393.716224,0.0,2749.366272,2585.357312,s,10,0.817713020324707,0.08177130203247071,0.0017450673609261166,0.08117209243774415,0.08364522857666015,0.0849723907470703,0.08603412048339844,"[0.08629955291748047, 0.08335030364990234, 0.08125138854980468, 0.0808326416015625, 0.0810927963256836, 0.08075574493408202, 0.08150611114501953, 0.0821209945678711, 0.08023673248291016, 0.08026675415039063]",tokens/s,3130.6826923013223,kWh,9.541744039156959e-07,5.228433619278451e-07,4.351788875694268e-06,5.828806641537809e-06,tokens/kWh,43919796.236791916,MB,1844.498432,3395.813376,0.0,2749.366272,2628.768768,s,10,45.24670263671875,4.524670263671875,0.011893284876975272,4.52194580078125,4.53378251953125,4.544506494140625,4.553085673828124,"[4.55523046875, 4.5133427734375, 4.5271787109375, 4.52349462890625, 4.52039697265625, 4.527849609375, 4.5313994140625, 4.51648095703125, 4.51110009765625, 4.52022900390625]",tokens/s,13.923666549985024,kWh,5.321501913543884e-05,2.9164818669173503e-05,0.00023333654016370403,0.0003157163779683163,tokens/kWh,199546.18890985235,,s,629,45.90427850341802,0.0729797750451796,0.009729252653977099,0.07159503936767578,0.07251025695800782,0.07281602630615235,0.15288290954589845,"[0.07412633514404297, 0.07440076446533203, 0.0745185317993164, 0.07484521484375, 0.07359587097167969, 0.07270297241210938, 0.07295487976074219, 0.0725032958984375, 0.0718704605102539, 0.0717998046875, 0.07187251281738281, 0.07140557098388672, 0.07241011047363281, 0.0728279037475586, 0.07255449676513671, 0.07251353454589844, 0.07184178924560547, 0.07262105560302734, 0.07237939453125, 0.072342529296875, 0.07164313507080078, 0.07216537475585938, 0.07139225769042969, 0.07148646545410156, 0.07167488098144531, 0.07242137908935548, 0.07154073333740234, 0.07204863739013671, 0.07214796447753906, 0.07182643127441406, 0.07173836517333984, 0.07221862030029297, 0.0719452133178711, 0.07195136260986328, 0.07149260711669922, 0.07235382080078125, 0.07215001678466797, 0.07205680084228516, 0.07161958312988281, 0.07250943756103516, 0.07279821014404297, 0.07184793853759766, 0.07290367889404296, 0.0722012176513672, 0.07165235137939453, 0.072595458984375, 0.07175475311279297, 0.07177327728271485, 0.07174237060546874, 0.07140249633789063, 0.07150489807128907, 0.07124172973632813, 0.07160115051269532, 0.07147634887695313, 0.07142079925537109, 0.07171481323242188, 0.07252889251708984, 0.07654195404052734, 0.07267225646972657, 0.07200358581542969, 0.07152947235107422, 0.07169741058349609, 0.15563571166992188, 0.07144140625, 0.07148544311523437, 0.0714229736328125, 0.07137792205810548, 0.07154073333740234, 0.07146803283691407, 0.07168921661376954, 0.07148748779296875, 0.07125196838378907, 0.072015869140625, 0.0714434585571289, 0.07144652557373046, 0.07144755554199218, 0.07143526458740235, 0.07141580963134765, 0.07135846710205078, 0.07186329650878906, 0.07159503936767578, 0.07138301086425781, 0.07177523040771484, 0.07191756439208985, 0.07138511657714844, 0.07143011474609375, 0.0714691162109375, 0.07133177947998047, 0.07206400299072266, 0.07146598052978516, 0.07148339080810547, 0.07149465942382813, 0.07135244750976563, 0.07228915405273438, 0.07199129486083984, 0.07153971099853515, 0.07214195251464844, 0.07162457275390625, 0.07148236846923828, 0.07145574188232422, 0.07147622680664062, 0.07152435302734375, 0.07147622680664062, 0.07173017883300781, 0.07242240142822266, 0.07143628692626953, 0.07144550323486328, 0.07149062347412109, 0.07370745849609375, 0.071552001953125, 0.07143014526367188, 0.07157561492919921, 0.07137686157226562, 0.07150895690917969, 0.0715448989868164, 0.071289794921875, 0.07157247924804687, 0.07189810943603515, 0.07256575775146484, 0.0720179214477539, 0.07132978820800781, 0.07143628692626953, 0.07151615905761718, 0.071478271484375, 0.07195954895019531, 0.15412019348144532, 0.07261593627929687, 0.07204249572753907, 0.07153766632080077, 0.07150694274902344, 0.07190835571289063, 0.07203942108154297, 0.07188070678710938, 0.0718387222290039, 0.07253708648681641, 0.07149874877929688, 0.07151312255859375, 0.07189910125732422, 0.07219609832763672, 0.07241522979736328, 0.07219609832763672, 0.07200665283203125, 0.07194111633300782, 0.07144857788085937, 0.07177216339111328, 0.07156428527832032, 0.07152435302734375, 0.071404541015625, 0.07229337310791016, 0.0715982437133789, 0.07232189178466797, 0.07304499053955078, 0.07218585968017578, 0.07157247924804687, 0.07147212982177735, 0.07146905517578125, 0.0714250259399414, 0.07145369720458984, 0.0720711669921875, 0.07135027313232421, 0.07165337371826172, 0.07138304138183593, 0.07151001739501953, 0.07149158477783203, 0.0714434585571289, 0.07209881591796875, 0.07256575775146484, 0.07190835571289063, 0.07150592041015626, 0.07148441314697265, 0.07140665435791016, 0.07138195037841796, 0.07330815887451173, 0.07221965026855469, 0.07159101104736328, 0.07144233703613281, 0.0722165756225586, 0.07188787078857421, 0.07187558746337891, 0.07241216278076172, 0.07215513610839844, 0.0714250259399414, 0.07158169555664062, 0.07205580902099609, 0.07226470184326172, 0.07152249908447265, 0.07164704132080078, 0.07141785430908203, 0.15294053649902345, 0.07213772583007813, 0.07148236846923828, 0.07144550323486328, 0.07141990661621093, 0.07155097961425781, 0.07134003448486329, 0.0716882553100586, 0.07207417297363282, 0.07252582550048828, 0.07232307434082032, 0.07238553619384766, 0.07231795501708985, 0.07151103973388671, 0.07148134613037109, 0.07183052825927734, 0.07228518676757813, 0.07243673706054687, 0.07223193359375, 0.07178342437744141, 0.07154994964599609, 0.07213056182861328, 0.07134528350830079, 0.07198809814453125, 0.07157759857177734, 0.07243981170654297, 0.07139225769042969, 0.07143321228027344, 0.0725002212524414, 0.07220428466796874, 0.07141069030761718, 0.07173426818847656, 0.0720742416381836, 0.07153049468994141, 0.07139839935302734, 0.07158169555664062, 0.07130623626708985, 0.07154605102539062, 0.07140230560302735, 0.07144652557373046, 0.07150796508789062, 0.07149465942382813, 0.07262413024902344, 0.07216844940185548, 0.073059326171875, 0.07163391876220702, 0.07140966033935547, 0.07162879943847657, 0.07215309143066406, 0.07194834899902344, 0.07152428436279297, 0.07163085174560548, 0.07161036682128906, 0.07260262298583985, 0.07228108978271484, 0.07144550323486328, 0.0714434585571289, 0.07155712127685547, 0.07133695983886719, 0.07153254699707032, 0.0716072998046875, 0.07149568176269532, 0.07149158477783203, 0.15309414672851562, 0.07136255645751953, 0.07175580596923828, 0.07164310455322266, 0.07139532470703125, 0.07184896087646485, 0.07149158477783203, 0.07176089477539062, 0.0714567642211914, 0.07160524749755859, 0.07143116760253906, 0.07155097961425781, 0.07128268432617188, 0.07143628692626953, 0.0721080322265625, 0.07152742767333985, 0.07195340728759765, 0.0717998046875, 0.07170150756835937, 0.07143628692626953, 0.07166976165771484, 0.07139328002929687, 0.07143334197998047, 0.07141567993164062, 0.07132784271240235, 0.07156521606445312, 0.07134419250488282, 0.07141779327392578, 0.07136870574951172, 0.07146701049804688, 0.07180595397949219, 0.07261695861816406, 0.07249817657470703, 0.07219513702392578, 0.07214688110351562, 0.0726087646484375, 0.07276953887939454, 0.07257190704345703, 0.07155097961425781, 0.07176396942138671, 0.07192063903808593, 0.07148041534423828, 0.07149251556396484, 0.07153049468994141, 0.0720097885131836, 0.07163180541992188, 0.07188992309570312, 0.07222169494628906, 0.07132569885253906, 0.07175081634521484, 0.07155388641357421, 0.07143321228027344, 0.07141580963134765, 0.07153868865966796, 0.07144963073730469, 0.07194624328613282, 0.07234966278076171, 0.0735129623413086, 0.07201398468017578, 0.07184162902832031, 0.07150899505615234, 0.07152639770507813, 0.07161958312988281, 0.15273472595214843, 0.0721817626953125, 0.07285453033447266, 0.0722841567993164, 0.07220633697509765, 0.07234150695800781, 0.0714803237915039, 0.07147212982177735, 0.07221247863769531, 0.07245523071289063, 0.0723117446899414, 0.07214387512207031, 0.07150694274902344, 0.07175373077392579, 0.07198617553710937, 0.07157350158691406, 0.0715489273071289, 0.07148851013183594, 0.07142707061767578, 0.07142092895507812, 0.0714629135131836, 0.07176806640625, 0.07141785430908203, 0.07138406372070312, 0.07139647674560547, 0.07157030487060546, 0.07146803283691407, 0.07146495819091797, 0.07141375732421874, 0.07154790496826172, 0.07151107025146484, 0.07144342041015625, 0.07185100555419922, 0.07181517028808594, 0.07148646545410156, 0.0713328628540039, 0.07203743743896485, 0.07147821044921875, 0.07322930908203125, 0.07284429168701172, 0.07214694213867187, 0.07164927673339844, 0.07231283569335938, 0.07260262298583985, 0.07171686553955078, 0.07191961669921874, 0.07252889251708984, 0.07248281860351563, 0.07143628692626953, 0.07152742767333985, 0.07151001739501953, 0.0715335693359375, 0.07144652557373046, 0.0727040023803711, 0.07292620849609376, 0.07252684783935547, 0.07237529754638672, 0.07170252990722656, 0.07182028961181641, 0.0714229736328125, 0.07204557037353515, 0.07154176330566406, 0.07145881652832031, 0.15435366821289062, 0.07202508544921875, 0.07218994903564453, 0.07186739349365234, 0.07218073272705078, 0.07134719848632813, 0.07238559722900391, 0.07221548461914062, 0.07241216278076172, 0.07184178924560547, 0.07194214630126954, 0.07146803283691407, 0.07222271728515625, 0.07166976165771484, 0.07240601348876953, 0.0716042251586914, 0.07220451354980469, 0.07180572509765625, 0.0720343017578125, 0.07153049468994141, 0.07171891021728516, 0.07154073333740234, 0.07239577484130859, 0.07312076568603515, 0.07245823669433593, 0.07169741058349609, 0.07252790069580078, 0.07206703948974609, 0.07212345886230469, 0.07148537445068359, 0.07223910522460937, 0.0724295654296875, 0.0715489273071289, 0.07143526458740235, 0.07208550262451172, 0.0715868148803711, 0.07253298950195312, 0.07192985534667969, 0.0719288330078125, 0.0714424285888672, 0.07221759796142578, 0.07254220581054688, 0.07175885009765624, 0.07147622680664062, 0.07144960021972656, 0.07139942169189453, 0.07147622680664062, 0.07175270080566407, 0.07210086059570313, 0.07172198486328125, 0.07231078338623047, 0.07225138854980469, 0.07237836456298828, 0.07164521789550782, 0.07197487640380859, 0.07171071624755859, 0.07147110748291016, 0.07145785522460937, 0.07145772552490234, 0.07148953247070312, 0.07153049468994141, 0.07147929382324218, 0.07183257293701172, 0.15419290161132812, 0.0719319076538086, 0.07157453155517578, 0.07133900451660156, 0.07158086395263671, 0.07154566192626953, 0.07155609893798828, 0.07177932739257813, 0.07223910522460937, 0.07164825439453125, 0.07151103973388671, 0.07149874877929688, 0.07160832214355468, 0.07286483001708985, 0.07185404968261719, 0.07166358184814453, 0.07239065551757813, 0.07162777709960938, 0.07142400360107422, 0.07145779418945312, 0.07137689971923829, 0.07147929382324218, 0.07150899505615234, 0.07147929382324218, 0.07143014526367188, 0.07144550323486328, 0.07150489807128907, 0.07147929382324218, 0.07139942169189453, 0.07138111877441407, 0.07176595306396484, 0.07276025390625, 0.07230156707763671, 0.07154994964599609, 0.07149874877929688, 0.071984130859375, 0.07214704132080078, 0.07197277069091797, 0.07147622680664062, 0.07143628692626953, 0.07141990661621093, 0.07155097961425781, 0.07132876586914062, 0.07161753845214844, 0.07156735992431641, 0.07142707061767578, 0.07145574188232422, 0.07148236846923828, 0.07152947235107422, 0.07150387573242188, 0.07297023773193359, 0.07153459167480469, 0.07137391662597656, 0.07174543762207031, 0.07197593688964844, 0.07133900451660156, 0.07137586975097657, 0.07162675476074219, 0.07164524841308594, 0.07149254608154297, 0.07288114929199219, 0.07158067321777344, 0.07160832214355468, 0.1534167022705078, 0.07255654144287109, 0.07152025604248047, 0.07147110748291016, 0.07152127838134766, 0.07155718231201172, 0.07147615814208984, 0.07213878631591797, 0.0719688949584961, 0.07150473785400391, 0.07153254699707032, 0.0713677749633789, 0.07146691131591797, 0.07148236846923828, 0.07166361236572266, 0.07142195129394531, 0.07141273498535156, 0.07221247863769531, 0.07152947235107422, 0.07151411437988281, 0.07147727966308594, 0.07141168212890625, 0.07130111694335937, 0.07153561401367188, 0.07139328002929687, 0.07157868957519531, 0.07202604675292969, 0.07173222351074218, 0.07142809295654297, 0.07216230773925782, 0.07140249633789063, 0.07125708770751953, 0.07135030364990234, 0.07157039642333984, 0.07152537536621094, 0.07153663635253907, 0.07136067199707032, 0.07143817901611328, 0.07143218994140625, 0.07128883361816406, 0.07141171264648437, 0.07147212982177735, 0.07205785369873047, 0.0722841567993164, 0.07163801574707031, 0.07146086120605469, 0.07142400360107422, 0.07148851013183594, 0.0714567642211914, 0.07157453155517578, 0.07137894439697266, 0.07154994964599609, 0.07136358642578125, 0.07146806335449218, 0.07140144348144531, 0.07132978820800781, 0.07141990661621093, 0.0713338851928711, 0.07141683197021484, 0.07153254699707032, 0.072774658203125, 0.07168000030517578, 0.07149465942382813, 0.15201997375488283, 0.07148339080810547, 0.07143628692626953, 0.07144038391113282, 0.07139942169189453, 0.07197293090820313, 0.07162464141845704, 0.07138406372070312, 0.07135743713378906, 0.07148544311523437, 0.07146803283691407, 0.07145881652832031, 0.07143218994140625, 0.07152742767333985, 0.07162982177734376, 0.07121612548828125, 0.07146495819091797, 0.0720404510498047, 0.0724695053100586, 0.07181721496582032, 0.07156018829345703, 0.07254528045654297, 0.07181414031982422, 0.07141072082519531, 0.0720823974609375, 0.0720711669921875, 0.07299686431884765, 0.07165644836425782, 0.07156121826171875, 0.071299072265625, 0.07145369720458984, 0.07160329437255859, 0.07154678344726563, 0.07152537536621094, 0.07154688262939453, 0.0715868148803711, 0.0715704345703125, 0.0713881607055664, 0.07157350158691406, 0.0714260482788086, 0.07144038391113282, 0.07158988952636719, 0.07161138916015625, 0.07159295654296875, 0.07171071624755859, 0.07247052764892578, 0.07238553619384766, 0.07274291229248046, 0.07178034973144531, 0.0724695053100586, 0.07223808288574218, 0.07157257843017578, 0.07217967987060547, 0.07244691467285157, 0.07151718139648437, 0.07182745361328124, 0.07148748779296875, 0.07153971099853515, 0.07172608184814454, 0.07232614135742188, 0.07174861145019532, 0.07182540893554687, 0.07194419097900391]",tokens/s,13.702426451450826,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1469, in _autoset_attn_implementation + cls._check_and_enable_flash_attn_2( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 + raise ValueError( +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp4xqzchjj/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -6403,7 +6818,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1472-5afa93fa0dd2f3de457c044f;40bbd32e-a5d9-44bd-8248-4e6e551b075a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4686-76337dd644ea2cd31e133602;6b892cba-92b7-418f-8039-dfcbbb891db1) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6460,7 +6875,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp5pee6r1e/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp91y4_erl/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -6502,7 +6917,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1205-7787106b04b9899d7de49161;3f2f5bd8-e678-4d37-8156-7335973861bd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4463-02977a8c02be53072628123e;00b659d1-02ab-4c3c-9a47-a0bf6d27e2d4) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6573,7 +6988,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12d9-7e6eca2a60aefe033b0541de;dfca232d-0ecf-4d62-bc79-f2dfbf88a318) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4502-6d9423ec14cf6f1771a3a784;fe601118-2e84-4f42-a0d1-0f9a2dadae2e) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6605,8 +7020,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5016.45312,8173.125632,0.0,7526.678528,6833.613824,s,10,3.326199401855469,0.33261994018554686,0.003080048658705347,0.3318737487792969,0.3344139831542969,0.33784259643554687,0.34058548706054687,"[0.34127120971679686, 0.3336520690917969, 0.33123345947265626, 0.33050531005859374, 0.330927734375, 0.32976422119140625, 0.33141766357421876, 0.332329833984375, 0.33259820556640624, 0.33249969482421876]",tokens/s,769.647183079867,kWh,3.900687001298405e-06,2.137033586950565e-06,2.0206181039483795e-05,2.6243901627732764e-05,tokens/kWh,9754647.141699262,MB,5016.45312,8173.125632,0.0,7526.678528,7110.583808,s,10,183.519623046875,18.3519623046875,0.004288348522690653,18.3511787109375,18.358202734375,18.358737109375,18.359164609375,"[18.359271484375, 18.348994140625, 18.349845703125, 18.3483828125, 18.3546015625, 18.358083984375, 18.350859375, 18.351498046875, 18.353595703125, 18.344490234375]",tokens/s,3.4328754034062285,kWh,0.00021665113780019963,0.00011874283365397788,0.001081981506265517,0.0014173754777196945,tokens/kWh,44448.34907215681,,s,629,186.2062080688478,0.2960353069457038,0.03944703464753294,0.2912143249511719,0.2916943786621094,0.2919354309082031,0.62257517578125,"[0.29116619873046873, 0.2915358581542969, 0.2909962158203125, 0.29132696533203123, 0.29103924560546873, 0.29121844482421877, 0.29109759521484374, 0.29120819091796873, 0.2913249206542969, 0.2912593994140625, 0.2912757873535156, 0.29111911010742186, 0.2910730285644531, 0.2911375427246094, 0.29116928100585937, 0.29137100219726564, 0.2911713256835938, 0.29115289306640624, 0.291462158203125, 0.2916761474609375, 0.29171713256835935, 0.2913617858886719, 0.29141400146484375, 0.29209906005859376, 0.29184716796875, 0.29161471557617186, 0.29145599365234376, 0.2914078674316406, 0.29128500366210935, 0.29134030151367185, 0.2919741516113281, 0.29177548217773436, 0.29137921142578127, 0.29178778076171874, 0.2918133850097656, 0.2917447814941406, 0.29174578857421873, 0.2921359252929687, 0.29115188598632813, 0.290977783203125, 0.2910382080078125, 0.2909767761230469, 0.2910525512695312, 0.2908293151855469, 0.29078936767578123, 0.2910259094238281, 0.2909624328613281, 0.29080474853515625, 0.29093582153320313, 0.2909808959960937, 0.29161981201171877, 0.29114471435546874, 0.290914306640625, 0.29103717041015625, 0.2916168518066406, 0.29117019653320314, 0.29111090087890623, 0.29126046752929685, 0.2911712646484375, 0.2911160278320313, 0.29122662353515627, 0.2909982604980469, 0.6255001831054687, 0.29120819091796873, 0.2914027404785156, 0.2912962646484375, 0.29111090087890623, 0.29109246826171875, 0.291209228515625, 0.2909593505859375, 0.29141094970703124, 0.291378173828125, 0.2910279541015625, 0.29116824340820313, 0.29119793701171875, 0.29114163208007815, 0.2911231994628906, 0.29109146118164064, 0.2910648193359375, 0.29119488525390624, 0.29128704833984376, 0.29122457885742187, 0.2912767944335938, 0.29190347290039065, 0.29122354125976563, 0.2913259582519531, 0.2910484619140625, 0.29144677734375, 0.29113037109375, 0.2916597900390625, 0.2917017517089844, 0.2913945617675781, 0.291080322265625, 0.2912357177734375, 0.29142630004882814, 0.29292755126953124, 0.2914487609863281, 0.29130239868164065, 0.2910894775390625, 0.2912685546875, 0.2911784973144531, 0.29153689575195313, 0.2912389221191406, 0.29098809814453125, 0.2912736511230469, 0.2915430297851562, 0.29123175048828126, 0.2910330810546875, 0.29101364135742186, 0.29123992919921876, 0.29120306396484374, 0.29105767822265627, 0.29113345336914065, 0.29084671020507813, 0.2909900817871094, 0.29090304565429687, 0.290951171875, 0.29124609375, 0.2915348510742187, 0.2909532470703125, 0.2907913818359375, 0.2914089050292969, 0.2912010192871094, 0.2910945129394531, 0.29110693359375, 0.6228212280273437, 0.2909962158203125, 0.29121636962890624, 0.2909593505859375, 0.291162109375, 0.2909972534179687, 0.29106689453125, 0.2911058044433594, 0.29145703125, 0.291662841796875, 0.2912471008300781, 0.29108941650390624, 0.2911866760253906, 0.2912204895019531, 0.2912194519042969, 0.29111090087890623, 0.2910064697265625, 0.29104129028320314, 0.2909378662109375, 0.2911313781738281, 0.29149594116210936, 0.29147647094726564, 0.2911918029785156, 0.2929029541015625, 0.2915798645019531, 0.29134539794921877, 0.29141195678710935, 0.29131787109375, 0.2910135498046875, 0.291483642578125, 0.2914928588867188, 0.29125323486328125, 0.2910013427734375, 0.29123992919921876, 0.29137716674804687, 0.29116619873046873, 0.29122354125976563, 0.2912204895019531, 0.29115805053710936, 0.29145083618164064, 0.29133416748046875, 0.29131875610351565, 0.2913157348632813, 0.29126959228515625, 0.2911016845703125, 0.2911048583984375, 0.29092343139648436, 0.2929295349121094, 0.29161471557617186, 0.29102386474609376, 0.291267578125, 0.2909962158203125, 0.2914375610351562, 0.2910167236328125, 0.2913781433105469, 0.2911907958984375, 0.2911426696777344, 0.2909788513183594, 0.2909173583984375, 0.2910310363769531, 0.2912215881347656, 0.29127874755859373, 0.2908927917480469, 0.6225029296875, 0.291240966796875, 0.29127987670898436, 0.2913617858886719, 0.2913382263183594, 0.2910033874511719, 0.2911457214355469, 0.2911641540527344, 0.29143243408203123, 0.2911856689453125, 0.2912788391113281, 0.2914222106933594, 0.2920099792480469, 0.2911703186035156, 0.2913064880371094, 0.2911098937988281, 0.2911754150390625, 0.2914334716796875, 0.29155123901367186, 0.29108941650390624, 0.29109963989257814, 0.291051513671875, 0.2908641357421875, 0.2909010009765625, 0.29108941650390624, 0.29113037109375, 0.29089797973632814, 0.29110470581054687, 0.290946044921875, 0.2912204895019531, 0.2911457214355469, 0.29170281982421875, 0.2912634582519531, 0.2912041015625, 0.2915431213378906, 0.29133712768554687, 0.2911508483886719, 0.29135565185546874, 0.29162701416015624, 0.29137615966796876, 0.2912163391113281, 0.29111294555664063, 0.2913996887207031, 0.2912686157226563, 0.2932623291015625, 0.2910310363769531, 0.2911180725097656, 0.2912808837890625, 0.29141299438476564, 0.29114471435546874, 0.290977783203125, 0.29087554931640625, 0.2910574951171875, 0.29096038818359377, 0.29089483642578123, 0.2909276123046875, 0.2909982604980469, 0.2909296569824219, 0.2908979187011719, 0.29132696533203123, 0.29107199096679687, 0.2911231994628906, 0.2913863830566406, 0.622603271484375, 0.2917027893066406, 0.29134951782226565, 0.29125018310546874, 0.2917386169433594, 0.2914908142089844, 0.29173348999023435, 0.2911733703613281, 0.29134335327148436, 0.2910812072753906, 0.2911098937988281, 0.29106381225585937, 0.29109658813476563, 0.29120205688476564, 0.2910320739746094, 0.2910947265625, 0.29100115966796875, 0.292279296875, 0.2911508483886719, 0.2910453796386719, 0.29116928100585937, 0.29125735473632813, 0.29149594116210936, 0.29107611083984375, 0.2910126037597656, 0.29110272216796873, 0.291125244140625, 0.29108941650390624, 0.2909861145019531, 0.29115072631835937, 0.2912153625488281, 0.29102386474609376, 0.2911457214355469, 0.29115289306640624, 0.29117236328125, 0.29140069580078126, 0.29126553344726563, 0.29102386474609376, 0.29119589233398435, 0.2918922119140625, 0.2919342041015625, 0.29135769653320315, 0.29173965454101564, 0.29154611206054687, 0.29135565185546874, 0.2912491455078125, 0.2913812561035156, 0.29117745971679687, 0.29169253540039064, 0.29132186889648437, 0.2913034362792969, 0.29131365966796874, 0.29162701416015624, 0.2916198425292969, 0.29157684326171873, 0.29161575317382815, 0.29180108642578123, 0.2936002502441406, 0.29074740600585935, 0.2910423583984375, 0.29112109375, 0.29105889892578124, 0.290897705078125, 0.6212341918945312, 0.29106689453125, 0.2912808837890625, 0.291009521484375, 0.291135498046875, 0.2910699462890625, 0.2918184814453125, 0.2910679016113281, 0.2911058044433594, 0.2919096374511719, 0.29177752685546876, 0.29155328369140626, 0.29200897216796873, 0.29156658935546875, 0.29147137451171873, 0.29141812133789063, 0.29161895751953126, 0.29161663818359373, 0.292021240234375, 0.29205810546875, 0.2920222778320313, 0.2917530212402344, 0.2912767333984375, 0.29116110229492187, 0.29100747680664063, 0.29129010009765627, 0.2913894348144531, 0.29166693115234377, 0.29114886474609375, 0.2912081298828125, 0.29109759521484374, 0.2914437255859375, 0.2911088562011719, 0.2910955505371094, 0.29103512573242185, 0.2910525512695312, 0.2909982604980469, 0.29115191650390626, 0.293800048828125, 0.29129302978515625, 0.29165670776367186, 0.29106381225585937, 0.290946044921875, 0.29162905883789064, 0.2911651916503906, 0.29107098388671876, 0.2913679504394531, 0.29179495239257813, 0.29130035400390625, 0.2910279541015625, 0.29130853271484375, 0.2909982604980469, 0.29144580078125, 0.29087945556640626, 0.29143243408203123, 0.29117953491210935, 0.2911651916503906, 0.29131265258789063, 0.2911754150390625, 0.2913474426269531, 0.2915758056640625, 0.2920316162109375, 0.29120806884765627, 0.6234173583984375, 0.291557373046875, 0.2918380126953125, 0.29149484252929686, 0.29161676025390626, 0.29160858154296876, 0.2914078674316406, 0.29153176879882814, 0.2913587341308594, 0.2913239135742188, 0.29197210693359377, 0.29134848022460935, 0.2913802185058594, 0.29154815673828127, 0.291346435546875, 0.2913414306640625, 0.2914947814941406, 0.2916659240722656, 0.2913699951171875, 0.2909399108886719, 0.2911231994628906, 0.29095013427734373, 0.290977783203125, 0.2927216491699219, 0.29113345336914065, 0.29114981079101565, 0.29107403564453127, 0.2911918029785156, 0.2909122619628906, 0.291061767578125, 0.29106893920898436, 0.2917857360839844, 0.29131570434570314, 0.2910535583496094, 0.2911559753417969, 0.29109146118164064, 0.2910341796875, 0.29110470581054687, 0.29129931640625, 0.2910627746582031, 0.2909686279296875, 0.29092755126953124, 0.2909470825195313, 0.2911006774902344, 0.2909060974121094, 0.29136587524414065, 0.2909696044921875, 0.2912542724609375, 0.2915430297851562, 0.2912122802734375, 0.2912204895019531, 0.29156558227539064, 0.29109658813476563, 0.2909214782714844, 0.29097164916992185, 0.29119692993164065, 0.291061767578125, 0.29099417114257814, 0.29109042358398435, 0.29147442626953124, 0.29125531005859373, 0.291282958984375, 0.2911488037109375, 0.6243696899414063, 0.2917304382324219, 0.2916188049316406, 0.2910771484375, 0.29127786254882815, 0.2910637512207031, 0.2911365051269531, 0.29100747680664063, 0.2916546630859375, 0.2911180725097656, 0.2911170654296875, 0.29096652221679686, 0.2911590270996094, 0.2909399108886719, 0.29097268676757815, 0.29142630004882814, 0.29239706420898437, 0.29116110229492187, 0.2918369140625, 0.29131570434570314, 0.29151852416992186, 0.2918962707519531, 0.2915921936035156, 0.2912767944335938, 0.29122775268554685, 0.291442626953125, 0.2913709716796875, 0.2911436767578125, 0.29121023559570314, 0.2912071533203125, 0.29118975830078125, 0.2910627746582031, 0.29127783203125, 0.2911385498046875, 0.2912276611328125, 0.29141299438476564, 0.2915604553222656, 0.29117645263671876, 0.2910699462890625, 0.2913536071777344, 0.29105767822265627, 0.291409912109375, 0.29124301147460935, 0.29146929931640625, 0.2910986328125, 0.2909337463378906, 0.2911016845703125, 0.2913239135742188, 0.29193624877929686, 0.2917181396484375, 0.2913301696777344, 0.2909848327636719, 0.2914642028808594, 0.2911160278320313, 0.291262451171875, 0.2912665710449219, 0.29157888793945314, 0.2910177307128906, 0.290872314453125, 0.2911313781738281, 0.291019775390625, 0.290977783203125, 0.2908876647949219, 0.6242959594726563, 0.29109146118164064, 0.29125836181640624, 0.29093069458007814, 0.2913310852050781, 0.290766845703125, 0.29151742553710935, 0.2912389221191406, 0.29116928100585937, 0.29127996826171876, 0.2913391723632813, 0.2912143249511719, 0.29121844482421877, 0.2917212219238281, 0.29165771484375, 0.2913197937011719, 0.2915440673828125, 0.292537353515625, 0.29179800415039064, 0.29155841064453125, 0.2919966735839844, 0.291230712890625, 0.2913167419433594, 0.29126553344726563, 0.2913648681640625, 0.2911160278320313, 0.2914089050292969, 0.29123077392578123, 0.2909859313964844, 0.29102694702148435, 0.29122457885742187, 0.2911651916503906, 0.2914170837402344, 0.2914068603515625, 0.29143243408203123, 0.29179800415039064, 0.2914170837402344, 0.29126553344726563, 0.2910986328125, 0.2912225341796875, 0.2914498596191406, 0.29127783203125, 0.2909962158203125, 0.2915625, 0.2915471496582031, 0.29119692993164065, 0.29103616333007815, 0.2915000305175781, 0.29133517456054686, 0.29093582153320313, 0.29118771362304685, 0.2910597229003906, 0.2912747497558594, 0.2911539306640625, 0.29163418579101563, 0.2911805419921875, 0.29106893920898436, 0.29166387939453126, 0.2909224853515625, 0.2912563171386719, 0.29143243408203123, 0.2914846801757813, 0.2908927917480469, 0.6241136474609374, 0.29096652221679686, 0.2912112731933594, 0.29113446044921876, 0.2913536071777344, 0.2909900817871094, 0.29119488525390624, 0.2911098937988281, 0.2912286682128906, 0.29107199096679687, 0.2915706787109375, 0.2912286682128906, 0.2910863342285156, 0.2910955505371094, 0.2912276611328125, 0.2910618286132812, 0.29106475830078127, 0.29117645263671876, 0.29103411865234374, 0.29098599243164064, 0.29118881225585935, 0.29100946044921877, 0.2909153137207031, 0.29101568603515626, 0.2915973205566406, 0.2913239135742188, 0.2927995300292969, 0.2914518737792969, 0.2909532165527344, 0.29111294555664063, 0.29151028442382815, 0.29141094970703124, 0.2911375427246094, 0.29108944702148437, 0.290989013671875, 0.29102694702148435, 0.29118875122070315, 0.2913966064453125, 0.2912143249511719, 0.2912030944824219, 0.2909511413574219, 0.29100851440429687, 0.290956298828125, 0.2911313781738281, 0.29113958740234375, 0.29146829223632814, 0.29121636962890624, 0.291093505859375, 0.29149798583984377, 0.2911119689941406, 0.291536865234375, 0.29114675903320314, 0.2909624328613281, 0.2908794860839844, 0.2909542541503906, 0.29093582153320313, 0.29088870239257814, 0.29089181518554685, 0.29084463500976565, 0.29150106811523435, 0.29094400024414063, 0.29157684326171873, 0.2909276123046875]",tokens/s,3.3779754527165613,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4755.23072,21498.42944,0.0,20851.982336,20236.60032,s,10,15.643266601562498,1.5643266601562498,0.001716851893423332,1.563902587890625,1.5660734985351563,1.567352362060547,1.5683754528808596,"[1.5631807861328124, 1.5647735595703125, 1.5629490966796875, 1.56462890625, 1.563903564453125, 1.5627757568359375, 1.565789306640625, 1.563901611328125, 1.5686312255859376, 1.5627327880859374]",tokens/s,163.6486844598237,kWh,1.846329511867629e-05,1.0117909425098331e-05,9.183712902519626e-05,0.00012041833356897088,tokens/kWh,2125922.1284055826,MB,4755.23072,21498.42944,0.0,20851.982336,20339.706368,s,10,879.77465625,87.97746562500001,0.0022909340685085054,87.97753515625,87.97981953125,87.980624609375,87.981268671875,"[87.974140625, 87.978578125, 87.9789609375, 87.9772265625, 87.979640625, 87.97715625, 87.97784375, 87.97353125, 87.9814296875, 87.9761484375]",tokens/s,0.716092462455496,kWh,0.0010385792660216491,0.0005692333667929961,0.005151564232359207,0.006759376865173853,tokens/kWh,9320.38577765846,,s,629,892.4578986816405,1.4188519851854382,0.185906583948783,1.396368408203125,1.3966589599609376,1.3968528564453124,2.960364462890625,"[1.39618505859375, 1.39667041015625, 1.3964708251953124, 1.396378662109375, 1.3963243408203125, 1.3962271728515625, 1.3963038330078126, 1.3964288330078125, 1.3964349365234374, 1.396369384765625, 1.3964697265625, 1.3965823974609375, 1.3962998046875, 1.3964708251953124, 1.3964349365234374, 1.3963857421875, 1.396304931640625, 1.3962802734375, 1.3960714111328125, 1.396305908203125, 1.3961553955078125, 1.3963427734375, 1.396221923828125, 1.3963756103515625, 1.39614111328125, 1.396317138671875, 1.3963294677734375, 1.3963673095703124, 1.396279296875, 1.3963018798828124, 1.3959813232421876, 1.3964195556640624, 1.396304931640625, 1.396304931640625, 1.396275146484375, 1.3961236572265625, 1.3962794189453125, 1.396401123046875, 1.39871435546875, 1.396484130859375, 1.396558837890625, 1.396505615234375, 1.3966868896484375, 1.3962772216796875, 1.3962833251953124, 1.3960928955078125, 1.396380615234375, 1.3968824462890626, 1.3964442138671875, 1.396285400390625, 1.3964237060546876, 1.396358154296875, 1.396305908203125, 1.3966787109375, 1.3965537109375, 1.3964185791015624, 1.3962916259765625, 1.3968353271484375, 1.396358154296875, 1.396221923828125, 1.39633251953125, 1.396252685546875, 2.962231201171875, 1.396673583984375, 1.3970401611328125, 1.396295654296875, 1.3965557861328124, 1.396348876953125, 1.3965035400390624, 1.3961922607421875, 1.3968536376953125, 1.39629052734375, 1.396274169921875, 1.396199462890625, 1.3962281494140625, 1.3965833740234375, 1.396538330078125, 1.3962445068359375, 1.3963079833984375, 1.3963243408203125, 1.396515869140625, 1.3962762451171875, 1.396073486328125, 1.3978162841796875, 1.396114501953125, 1.396137939453125, 1.3962198486328126, 1.3962987060546874, 1.3961595458984375, 1.3963714599609376, 1.396452392578125, 1.3961502685546876, 1.3962720947265626, 1.3963868408203124, 1.3964185791015624, 1.3962239990234375, 1.396464599609375, 1.3964482421875, 1.396427734375, 1.396380615234375, 1.396588623046875, 1.3964810791015625, 1.3967718505859374, 1.3981634521484374, 1.3962802734375, 1.396275146484375, 1.3962301025390624, 1.396253662109375, 1.396327392578125, 1.3961728515625, 1.3966571044921876, 1.396474853515625, 1.3965128173828125, 1.3963858642578124, 1.396339599609375, 1.3961318359375, 1.3965147705078125, 1.3963387451171876, 1.396506591796875, 1.3962393798828125, 1.3964605712890625, 1.3965179443359375, 1.39639599609375, 1.3985382080078126, 1.3965301513671875, 2.9603779296875, 1.396199462890625, 1.3967340087890625, 1.3962291259765625, 1.3964554443359376, 1.396252685546875, 1.3964892578125, 1.3960601806640625, 1.39671240234375, 1.3963714599609376, 1.3963509521484374, 1.3961092529296875, 1.39653125, 1.3965814208984375, 1.3964400634765626, 1.3963172607421874, 1.39658349609375, 1.3963939208984375, 1.396515869140625, 1.3962587890625, 1.3965823974609375, 1.3966663818359375, 1.396378662109375, 1.3987901611328124, 1.3961759033203125, 1.3963653564453125, 1.3963018798828124, 1.3964093017578125, 1.39652197265625, 1.3961544189453126, 1.3962332763671874, 1.39641650390625, 1.39623828125, 1.3962393798828125, 1.3961964111328125, 1.396336669921875, 1.3964564208984376, 1.3965137939453125, 1.3963756103515625, 1.3962701416015626, 1.396137939453125, 1.396569091796875, 1.39635205078125, 1.3963131103515625, 1.396252685546875, 1.3963427734375, 1.3963970947265625, 1.3995396728515626, 1.39681591796875, 1.396443115234375, 1.396485107421875, 1.3962608642578125, 1.3962506103515624, 1.3963192138671876, 1.3964769287109375, 1.396210693359375, 1.3968311767578125, 1.396547607421875, 1.39646875, 1.39649951171875, 1.396304931640625, 1.3964041748046876, 1.3961563720703125, 2.961944580078125, 1.39614208984375, 1.396427734375, 1.39610009765625, 1.3963775634765625, 1.397116943359375, 1.3965689697265624, 1.396379638671875, 1.396430908203125, 1.3963284912109375, 1.3966080322265626, 1.3961544189453126, 1.3964501953125, 1.396547607421875, 1.396279296875, 1.3963203125, 1.3963448486328125, 1.396454345703125, 1.3965128173828125, 1.396253662109375, 1.396316162109375, 1.396046875, 1.396062255859375, 1.3964974365234375, 1.396443115234375, 1.3962864990234376, 1.3960980224609374, 1.3964810791015625, 1.3962884521484376, 1.396199462890625, 1.396241455078125, 1.3960233154296875, 1.3961175537109376, 1.396675537109375, 1.3962762451171875, 1.3964237060546876, 1.3964083251953125, 1.3962884521484376, 1.3965762939453126, 1.3964708251953124, 1.396631591796875, 1.3964022216796874, 1.396336669921875, 1.396317138671875, 1.400958984375, 1.396559814453125, 1.3964820556640625, 1.39627001953125, 1.3966888427734374, 1.3963427734375, 1.396474853515625, 1.3962342529296874, 1.3963448486328125, 1.39620654296875, 1.396552734375, 1.3967205810546874, 1.3966029052734374, 1.3963560791015626, 1.3962332763671874, 1.3963212890625, 1.3964134521484375, 1.396279296875, 1.39620654296875, 2.96119921875, 1.3962393798828125, 1.39626806640625, 1.3964124755859375, 1.3966131591796875, 1.396216796875, 1.3965548095703124, 1.3961031494140625, 1.396580322265625, 1.396380615234375, 1.3969398193359375, 1.396273193359375, 1.39626904296875, 1.39641650390625, 1.3962547607421876, 1.39647900390625, 1.396178955078125, 1.3985556640625, 1.3966253662109376, 1.3968516845703125, 1.3966131591796875, 1.3960867919921875, 1.39594140625, 1.3963447265625, 1.396115478515625, 1.396316162109375, 1.3962301025390624, 1.3962618408203125, 1.3963406982421875, 1.3959669189453126, 1.396400146484375, 1.3961195068359376, 1.3963663330078124, 1.396368408203125, 1.3962496337890624, 1.39649951171875, 1.3963284912109375, 1.3964493408203125, 1.3962188720703126, 1.396115478515625, 1.3963602294921875, 1.3963663330078124, 1.3966182861328125, 1.3964237060546876, 1.3963038330078126, 1.3963653564453125, 1.39626806640625, 1.396369384765625, 1.3965137939453125, 1.3962486572265624, 1.3964317626953124, 1.3962711181640626, 1.39637353515625, 1.3964647216796875, 1.39634375, 1.3963253173828125, 1.396611083984375, 1.4011822509765626, 1.3966837158203125, 1.3966038818359374, 1.39641650390625, 1.3965537109375, 1.3965322265625, 2.960329833984375, 1.3966304931640625, 1.396526123046875, 1.3960919189453125, 1.396494384765625, 1.39626904296875, 1.3971036376953125, 1.39640625, 1.3964461669921875, 1.3961175537109376, 1.3965537109375, 1.3965025634765624, 1.3961400146484375, 1.396515869140625, 1.396490234375, 1.396189208984375, 1.3961563720703125, 1.3964564208984376, 1.3965404052734376, 1.396285400390625, 1.3961553955078125, 1.396074462890625, 1.395947509765625, 1.3964676513671874, 1.3965732421875, 1.3963448486328125, 1.3962977294921874, 1.3964267578125, 1.3964461669921875, 1.39627001953125, 1.3965732421875, 1.396262939453125, 1.39629052734375, 1.39640625, 1.3961185302734376, 1.3963243408203125, 1.3963826904296874, 1.3964442138671875, 1.400627197265625, 1.396809814453125, 1.396582275390625, 1.396357177734375, 1.39643798828125, 1.3963756103515625, 1.3962301025390624, 1.3963294677734375, 1.39649951171875, 1.3963336181640624, 1.3967288818359376, 1.396634521484375, 1.3963294677734375, 1.3961072998046875, 1.3963233642578126, 1.3963182373046874, 1.39650146484375, 1.396305908203125, 1.3965128173828125, 1.396368408203125, 1.3962659912109374, 1.39628955078125, 1.3963355712890626, 1.39620654296875, 1.39628955078125, 2.96319189453125, 1.39616259765625, 1.3965670166015625, 1.39634375, 1.39635205078125, 1.3960509033203126, 1.39610107421875, 1.3959833984375, 1.3964052734375, 1.3961861572265626, 1.39652197265625, 1.39647998046875, 1.3962496337890624, 1.3968414306640624, 1.3966519775390625, 1.3960438232421875, 1.3962802734375, 1.3961964111328125, 1.3965936279296876, 1.3964493408203125, 1.3962025146484376, 1.396211669921875, 1.3961871337890626, 1.3990369873046875, 1.39656396484375, 1.39658447265625, 1.3963192138671876, 1.3963756103515625, 1.396262939453125, 1.3963499755859374, 1.3965291748046875, 1.3964564208984376, 1.396116455078125, 1.3963765869140625, 1.396496337890625, 1.396341796875, 1.3961666259765626, 1.396453369140625, 1.396306884765625, 1.3964974365234375, 1.396357177734375, 1.396156494140625, 1.396138916015625, 1.39642578125, 1.3961502685546876, 1.396420654296875, 1.3963642578125, 1.39631201171875, 1.396749267578125, 1.396473876953125, 1.396526123046875, 1.396474853515625, 1.3964871826171874, 1.396432861328125, 1.396547607421875, 1.3964410400390626, 1.4002022705078125, 1.3966510009765625, 1.39624658203125, 1.396135986328125, 1.3961492919921874, 1.3963140869140624, 1.39603759765625, 2.96123388671875, 1.3961964111328125, 1.3963387451171876, 1.3962198486328126, 1.39618408203125, 1.3961973876953124, 1.3963714599609376, 1.396177978515625, 1.3965966796875, 1.396106201171875, 1.396715576171875, 1.3962823486328124, 1.3963294677734375, 1.3963028564453126, 1.396380615234375, 1.3961964111328125, 1.398497314453125, 1.3963970947265625, 1.3965679931640624, 1.396317138671875, 1.3965782470703125, 1.3963919677734375, 1.3962659912109374, 1.3962720947265626, 1.396201416015625, 1.396317138671875, 1.3962393798828125, 1.396137939453125, 1.3964708251953124, 1.3964288330078125, 1.3963857421875, 1.3960755615234375, 1.3963480224609375, 1.396232177734375, 1.3961451416015624, 1.3963182373046874, 1.396641845703125, 1.3966080322265626, 1.396378662109375, 1.396273193359375, 1.3963345947265624, 1.3962322998046874, 1.3966744384765626, 1.3964073486328126, 1.3963243408203125, 1.3962373046875, 1.3963499755859374, 1.3963345947265624, 1.3965465087890625, 1.396432861328125, 1.39626904296875, 1.39639404296875, 1.3961748046875, 1.3964892578125, 1.3966878662109374, 1.396232177734375, 1.396611083984375, 1.3963765869140625, 1.39654248046875, 1.39650048828125, 1.39634375, 1.3964093017578125, 1.3962454833984375, 2.96591357421875, 1.3961614990234374, 1.396864013671875, 1.3961820068359374, 1.396379638671875, 1.3964666748046874, 1.3964503173828124, 1.3963642578125, 1.39664892578125, 1.39633154296875, 1.396452392578125, 1.3962650146484374, 1.396241455078125, 1.3963756103515625, 1.3963775634765625, 1.396111328125, 1.3964503173828124, 1.3980078125, 1.3967994384765625, 1.396642822265625, 1.3961103515625, 1.3965987548828125, 1.396348876953125, 1.3962291259765625, 1.396262939453125, 1.39631201171875, 1.3963212890625, 1.3961759033203125, 1.3965260009765625, 1.3965496826171875, 1.396384765625, 1.3963182373046874, 1.396358154296875, 1.3962158203125, 1.3962720947265626, 1.396420654296875, 1.3962281494140625, 1.3967728271484374, 1.3964512939453124, 1.39666845703125, 1.3965936279296876, 1.3961728515625, 1.39624853515625, 1.3961861572265626, 1.39656396484375, 1.3964697265625, 1.396485107421875, 1.396537353515625, 1.3965670166015625, 1.3964073486328126, 1.3962578125, 1.3964390869140626, 1.3966990966796875, 1.39630078125, 1.3968045654296875, 1.3967421875, 1.3966285400390626, 1.3964482421875, 1.396533203125, 1.396506591796875, 1.396284423828125, 1.3962977294921874, 1.400701904296875, 2.960291748046875, 1.3964564208984376, 1.396536376953125, 1.396294677734375, 1.396358154296875, 1.39620458984375, 1.3963621826171875, 1.3960714111328125, 1.396537353515625, 1.396304931640625, 1.3964451904296875, 1.396430908203125, 1.3963642578125, 1.3963837890625, 1.396828125, 1.3963817138671875, 1.39660498046875, 1.396589599609375, 1.39631201171875, 1.396283447265625, 1.396177978515625, 1.39618505859375, 1.396168701171875, 1.396116455078125, 1.39642578125, 1.396654052734375, 1.399869384765625, 1.3961328125, 1.3965076904296876, 1.3962025146484376, 1.3964339599609374, 1.3961707763671876, 1.3962506103515624, 1.396093994140625, 1.3964124755859375, 1.396494384765625, 1.3964676513671874, 1.396410400390625, 1.396167724609375, 1.3962772216796875, 1.396236328125, 1.3962117919921875, 1.39633251953125, 1.396305908203125, 1.396552734375, 1.3963140869140624, 1.39652197265625, 1.396462646484375, 1.3968414306640624, 1.39652099609375, 1.396305908203125, 1.396357177734375, 1.3965721435546874, 1.39648193359375, 1.396537353515625, 1.396272216796875, 1.396579345703125, 1.3963602294921875, 1.3965322265625, 1.396306884765625, 1.3964041748046876, 1.396275146484375, 1.3961829833984376]",tokens/s,0.7047951516023033,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6614,71 +7028,29 @@ If this is a private repository, make sure to pass a token having permission to File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch raise ChildProcessError(response[""traceback""]) ChildProcessError: Traceback (most recent call last): - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status - response.raise_for_status() - File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status - raise HTTPError(http_error_msg, response=self) -requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 399, in cached_file - resolved_file = hf_hub_download( - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn - return fn(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download - return _hf_hub_download_to_cache_dir( - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir - _raise_on_head_call_error(head_call_error, force_download, local_files_only) - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error - raise head_call_error - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error - metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn - return fn(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata - r = _request_wrapper( - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper - response = _request_wrapper( - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper - hf_raise_for_status(response) - File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status - raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c130c-63f5b06f48f7881e6fd4ef2d;1958ecfe-f2f0-4d14-86cc-7962ad480fbc) - -Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. -Please make sure you specified the correct `repo_id` and `repo_type`. -If you are trying to access a private or gated repo, make sure you are authenticated. - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ - super().__init__(config) - File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ - self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) - File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config - return AutoConfig.from_pretrained(model, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 934, in from_pretrained - config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict - config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict - resolved_config_file = cached_file( - File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 422, in cached_file - raise EnvironmentError( -OSError: a is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' -If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1469, in _autoset_attn_implementation + cls._check_and_enable_flash_attn_2( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 + raise ValueError( +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpjula_18j/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2111.176704,2846.359552,0.0,2199.912448,2075.149824,s,10,1.345602310180664,0.1345602310180664,0.0004287422965996787,0.13438082885742186,0.13517186279296875,0.13521602783203124,0.13525135986328124,"[0.1346531219482422, 0.13415478515625, 0.13441477966308593, 0.13409397888183594, 0.13418368530273436, 0.13434687805175782, 0.13423942565917968, 0.13526019287109375, 0.13516204833984374, 0.13509341430664062]",tokens/s,1902.4937610699312,kWh,1.5845098716241345e-06,8.681900968003902e-07,7.312065108906473e-06,9.764765077330997e-06,tokens/kWh,26216708.540618826,MB,2111.176704,2846.359552,0.0,2199.912448,2180.296192,s,10,72.99156640625,7.299156640625002,0.0010886665313993955,7.29881201171875,7.30088486328125,7.30103251953125,7.3011506445312495,"[7.30118017578125, 7.29998193359375, 7.298513671875, 7.29880029296875, 7.299087890625, 7.30085205078125, 7.29882373046875, 7.29793994140625, 7.29786962890625, 7.29851708984375]",tokens/s,8.631134129847297,kWh,8.616290877704268e-05,4.722355870710817e-05,0.0003964600079084879,0.0005298464753926387,tokens/kWh,118902.36686639149,,s,629,74.08624839782712,0.11778417869288894,0.016011389929587044,0.1158440933227539,0.11595162048339844,0.11602903137207031,0.25029092712402345,"[0.1164031982421875, 0.11662540435791016, 0.11636224365234375, 0.11587174224853515, 0.11590144348144531, 0.11583078765869141, 0.11578265380859375, 0.11591990661621093, 0.11577442932128906, 0.11571199798583984, 0.11582054138183594, 0.11574272155761718, 0.11578880310058594, 0.11589119720458985, 0.11576627349853516, 0.11583487701416016, 0.11584716796875, 0.11574578857421874, 0.1157570571899414, 0.11593215942382812, 0.11580108642578125, 0.11578675079345703, 0.11586764526367188, 0.11578265380859375, 0.11584819030761719, 0.11595059204101563, 0.11578163146972656, 0.11580723571777343, 0.11586252593994141, 0.11579904174804688, 0.11576627349853516, 0.11589631652832032, 0.11577753448486328, 0.11583590698242187, 0.11595366668701172, 0.1158287353515625, 0.11581644439697265, 0.1159557113647461, 0.11614822387695313, 0.11589222717285157, 0.11589631652832032, 0.11583487701416016, 0.1157734375, 0.11594956970214844, 0.11578572845458984, 0.11590758514404297, 0.1158666229248047, 0.11585126495361328, 0.1157949447631836, 0.11599155426025391, 0.11582669067382813, 0.11588813018798828, 0.1159557113647461, 0.11582975769042969, 0.11576115417480469, 0.11599155426025391, 0.11583795166015624, 0.11594751739501953, 0.11585228729248047, 0.11586048126220704, 0.11577651214599609, 0.11595164489746093, 0.250293212890625, 0.11575910186767578, 0.11586867523193359, 0.11589529418945313, 0.11572121429443359, 0.1158635482788086, 0.11580108642578125, 0.11569152069091797, 0.11583385467529297, 0.11572633361816406, 0.11583487701416016, 0.11589427185058594, 0.11576319885253906, 0.11582361602783203, 0.11587379455566406, 0.11576831817626954, 0.11571609497070312, 0.11584204864501953, 0.11576217651367188, 0.11584819030761719, 0.11591168212890625, 0.11580928039550781, 0.11572838592529297, 0.11591680145263672, 0.11588813018798828, 0.11593727874755859, 0.11585638427734375, 0.11580210876464844, 0.115704833984375, 0.1159208984375, 0.11575193786621094, 0.11585945892333985, 0.1158656005859375, 0.11578470611572265, 0.11577037048339844, 0.11588813018798828, 0.11587481689453125, 0.11586969757080078, 0.116063232421875, 0.11596800231933593, 0.11581849670410156, 0.11589017486572266, 0.1158502426147461, 0.11583078765869141, 0.11595673370361329, 0.11586048126220704, 0.11586969757080078, 0.1164031982421875, 0.11587891387939453, 0.11589836883544923, 0.11601817321777344, 0.11585945892333985, 0.11583283233642579, 0.11597618865966797, 0.11582054138183594, 0.11586969757080078, 0.115957763671875, 0.1162455062866211, 0.11581542205810547, 0.11584512329101562, 0.11583897399902343, 0.11605299377441407, 0.11594342041015625, 0.2505707550048828, 0.11568742370605468, 0.1159004135131836, 0.11583385467529297, 0.11577139282226563, 0.1158656005859375, 0.11585330963134766, 0.1156864013671875, 0.11587481689453125, 0.11572121429443359, 0.11576525115966797, 0.115884033203125, 0.1157201919555664, 0.11576729583740235, 0.11580108642578125, 0.1157673568725586, 0.1157989730834961, 0.11591577911376953, 0.1157550048828125, 0.11574886322021484, 0.11581337738037109, 0.11583795166015624, 0.11579596710205078, 0.11599052429199219, 0.1157949447631836, 0.11579702758789062, 0.11584925079345704, 0.1157795181274414, 0.11584819030761719, 0.1159393310546875, 0.1158123550415039, 0.11577145385742188, 0.1158583984375, 0.11581948852539062, 0.11586873626708985, 0.11593619537353515, 0.11581439971923828, 0.1158502426147461, 0.11589222717285157, 0.11580723571777343, 0.11586867523193359, 0.11591270446777344, 0.11583795166015624, 0.11578982543945313, 0.11590656280517578, 0.11583590698242187, 0.1159393310546875, 0.11591986846923828, 0.11583590698242187, 0.11578880310058594, 0.11588098907470704, 0.115810302734375, 0.11593007659912109, 0.11598540496826172, 0.11583078765869141, 0.11577037048339844, 0.11591474914550781, 0.1158440933227539, 0.11588301086425781, 0.11591065979003906, 0.11593523406982421, 0.11621990203857421, 0.11591577911376953, 0.25028504943847657, 0.11570687866210938, 0.11593421173095703, 0.1157949447631836, 0.1157734375, 0.11591986846923828, 0.11586867523193359, 0.11574578857421874, 0.11593727874755859, 0.1157396469116211, 0.11580723571777343, 0.11589734649658204, 0.11572223663330078, 0.11589324951171875, 0.11581337738037109, 0.11616973114013672, 0.11582566070556641, 0.115884033203125, 0.11577139282226563, 0.11591065979003906, 0.11591270446777344, 0.11577855682373046, 0.11584614562988281, 0.11589631652832032, 0.1157754898071289, 0.11585126495361328, 0.1158123550415039, 0.11571507263183593, 0.11579289245605469, 0.11591372680664062, 0.115810302734375, 0.11588198089599609, 0.11582566070556641, 0.11576934051513672, 0.11584819030761719, 0.11589939117431641, 0.11580518341064452, 0.115884033203125, 0.11592192077636719, 0.11581132507324218, 0.11585126495361328, 0.11589939117431641, 0.11581439971923828, 0.11580825805664062, 0.11589836883544923, 0.11582669067382813, 0.1158809585571289, 0.11590860748291015, 0.11583999633789062, 0.11585330963134766, 0.11591372680664062, 0.11576525115966797, 0.11583385467529297, 0.11591168212890625, 0.11582669067382813, 0.11581747436523437, 0.11591782379150391, 0.11587481689453125, 0.11585330963134766, 0.11589119720458985, 0.11584102630615234, 0.11583283233642579, 0.11592499542236329, 0.250281982421875, 0.1157201919555664, 0.11594239807128906, 0.11578470611572265, 0.11578675079345703, 0.11585536193847656, 0.11581337738037109, 0.11564441680908204, 0.11579084777832031, 0.11577961730957032, 0.11587478637695313, 0.11586969757080078, 0.1157232666015625, 0.11570175933837891, 0.11586252593994141, 0.11580928039550781, 0.11572531127929687, 0.11594445037841797, 0.11586457824707032, 0.11585228729248047, 0.11585536193847656, 0.1158123550415039, 0.11578368377685547, 0.11595161437988281, 0.11578777313232422, 0.11574169921875, 0.11583692932128906, 0.11585126495361328, 0.11589017486572266, 0.11593113708496093, 0.11582771301269532, 0.11573452758789063, 0.11586457824707032, 0.1158440933227539, 0.11572633361816406, 0.11594649505615234, 0.11585228729248047, 0.11630182647705078, 0.11595366668701172, 0.11585842895507813, 0.11582361602783203, 0.11603353881835937, 0.11589427185058594, 0.11576217651367188, 0.11597926330566406, 0.11593421173095703, 0.11581951904296875, 0.11588607788085938, 0.11583487701416016, 0.11584921264648437, 0.11600077056884765, 0.11583078765869141, 0.11583692932128906, 0.1159925765991211, 0.11582566070556641, 0.11570995330810546, 0.11591986846923828, 0.11591680145263672, 0.11578265380859375, 0.11595263671875, 0.11586156463623047, 0.11594847869873047, 0.11594035339355468, 0.2504622039794922, 0.1157201919555664, 0.11593215942382812, 0.11588198089599609, 0.11573760223388672, 0.11585126495361328, 0.1158287353515625, 0.11570381164550782, 0.11582975769042969, 0.11573554992675782, 0.11576627349853516, 0.11585740661621094, 0.1158123550415039, 0.11576422119140625, 0.11585842895507813, 0.11579392242431641, 0.11585842895507813, 0.11592192077636719, 0.11579289245605469, 0.11573248291015625, 0.11583999633789062, 0.1158287353515625, 0.11580825805664062, 0.11603353881835937, 0.11582669067382813, 0.11581132507324218, 0.11583590698242187, 0.11581449890136719, 0.11593513488769532, 0.11594342041015625, 0.11582975769042969, 0.11581747436523437, 0.11584921264648437, 0.11580313873291016, 0.1158318099975586, 0.11590553283691406, 0.1159331817626953, 0.11579698944091797, 0.1159362564086914, 0.11583795166015624, 0.11647283172607421, 0.1159331817626953, 0.1158635482788086, 0.11575193786621094, 0.11594035339355468, 0.1157949447631836, 0.11586048126220704, 0.11586969757080078, 0.11589631652832032, 0.11599155426025391, 0.11657625579833984, 0.11596492767333984, 0.11592704010009766, 0.11591986846923828, 0.1158656005859375, 0.11576525115966797, 0.11585433959960938, 0.11580416107177735, 0.11588301086425781, 0.11662028503417969, 0.11592601776123047, 0.11577855682373046, 0.11592704010009766, 0.25042636108398436, 0.11569664001464844, 0.11593113708496093, 0.11581132507324218, 0.11573145294189453, 0.11585740661621094, 0.11574886322021484, 0.11579392242431641, 0.11582672119140625, 0.11573244476318359, 0.11576217651367188, 0.11587789154052734, 0.11574886322021484, 0.11579289245605469, 0.11580723571777343, 0.11574272155761718, 0.11577855682373046, 0.11590144348144531, 0.11581747436523437, 0.11593113708496093, 0.1158502426147461, 0.1157734375, 0.115810302734375, 0.115884033203125, 0.1157734375, 0.11586457824707032, 0.11584204864501953, 0.11577037048339844, 0.11582771301269532, 0.115884033203125, 0.11576831817626954, 0.11585330963134766, 0.11584614562988281, 0.11580313873291016, 0.1159710693359375, 0.11589017486572266, 0.11580825805664062, 0.11585945892333985, 0.11588102722167969, 0.11577235412597656, 0.11583283233642579, 0.11589017486572266, 0.11583385467529297, 0.11590962982177734, 0.1159331817626953, 0.11623423767089844, 0.11592601776123047, 0.11588710021972656, 0.11581849670410156, 0.11585433959960938, 0.11590962982177734, 0.11579698944091797, 0.11589017486572266, 0.11592396545410157, 0.11611135864257813, 0.11585945892333985, 0.11591782379150391, 0.11581644439697265, 0.11584307098388671, 0.11585849761962891, 0.11583891296386718, 0.1158318099975586, 0.11591270446777344, 0.25133772277832034, 0.11565670776367187, 0.11591680145263672, 0.11575398254394531, 0.11577753448486328, 0.11584204864501953, 0.11576422119140625, 0.11593421173095703, 0.11582975769042969, 0.11589119720458985, 0.1157734375, 0.1158656005859375, 0.11576831817626954, 0.11573554992675782, 0.11584204864501953, 0.11581747436523437, 0.11577855682373046, 0.11632844543457031, 0.11574988555908203, 0.11584921264648437, 0.11586764526367188, 0.11579289245605469, 0.11578265380859375, 0.11591270446777344, 0.1157918701171875, 0.11573862457275391, 0.11580620574951171, 0.11574476623535156, 0.11570790100097657, 0.11589222717285157, 0.11574784088134765, 0.11576525115966797, 0.11582259368896484, 0.11581644439697265, 0.11572940826416016, 0.1158666229248047, 0.11577446746826171, 0.11585740661621094, 0.11585228729248047, 0.11583692932128906, 0.11584614562988281, 0.11599871826171874, 0.11582157135009766, 0.11573862457275391, 0.11592704010009766, 0.11583897399902343, 0.11578777313232422, 0.11583385467529297, 0.1158666229248047, 0.11591372680664062, 0.11602124786376954, 0.11583078765869141, 0.11584614562988281, 0.11594035339355468, 0.11581132507324218, 0.11575193786621094, 0.11592908477783204, 0.11587481689453125, 0.11580620574951171, 0.11590348815917968, 0.11584307098388671, 0.11584716796875, 0.1159557113647461, 0.2511810607910156, 0.11569152069091797, 0.11590860748291015, 0.115884033203125, 0.11576729583740235, 0.11587174224853515, 0.11579801940917969, 0.11567411041259766, 0.11580416107177735, 0.11567411041259766, 0.11572940826416016, 0.11585228729248047, 0.1157580795288086, 0.11572633361816406, 0.11580825805664062, 0.11571814727783203, 0.11586252593994141, 0.11594137573242187, 0.11578368377685547, 0.11573452758789063, 0.11582259368896484, 0.11580825805664062, 0.11578572845458984, 0.11587382507324219, 0.11579388427734374, 0.11576118469238281, 0.11580515289306641, 0.11575398254394531, 0.11578368377685547, 0.11639603424072266, 0.11579904174804688, 0.11602227020263672, 0.11582975769042969, 0.11574681854248046, 0.11587891387939453, 0.11587789154052734, 0.11582361602783203, 0.11588710021972656, 0.1158809585571289, 0.11595263671875, 0.11590656280517578, 0.11593421173095703, 0.11585228729248047, 0.115736572265625, 0.11606937408447265, 0.11581443023681641, 0.11588809967041015, 0.115884033203125, 0.11585126495361328, 0.1157550048828125, 0.11587686157226562, 0.1157734375, 0.11586969757080078, 0.11592601776123047, 0.11581747436523437, 0.11576012420654297, 0.11587583923339843, 0.11580416107177735, 0.11582566070556641, 0.11587276458740234, 0.11586764526367188, 0.11577855682373046, 0.11586457824707032, 0.25122508239746094, 0.11569459533691406, 0.1159557113647461, 0.11578777313232422, 0.11566694641113281, 0.11580416107177735, 0.11576319885253906, 0.11574988555908203, 0.11585536193847656, 0.11570893096923827, 0.11571199798583984, 0.11583590698242187, 0.11570585632324219, 0.11578880310058594, 0.11586867523193359, 0.11572838592529297, 0.11579698944091797, 0.11587993621826172, 0.1157570571899414, 0.11582669067382813, 0.1157918701171875, 0.11579289245605469, 0.11579596710205078, 0.11596390533447265, 0.11572633361816406, 0.115810302734375, 0.11584512329101562, 0.11574066925048829, 0.11578572845458984, 0.11587276458740234, 0.11574784088134765, 0.11581132507324218, 0.11581132507324218, 0.115810302734375, 0.11586252593994141, 0.11587993621826172, 0.11576831817626954, 0.11585945892333985, 0.11592294311523438, 0.11580825805664062, 0.11584921264648437, 0.1159208984375, 0.11589631652832032, 0.11581951904296875, 0.11599980926513671, 0.115837890625, 0.11596185302734376, 0.11590962982177734, 0.11582669067382813, 0.1158666229248047, 0.11588198089599609, 0.116347900390625, 0.11593830108642578, 0.11591270446777344, 0.11585945892333985, 0.11579698944091797, 0.11591782379150391, 0.11618201446533204, 0.11581644439697265, 0.11589222717285157, 0.11585330963134766, 0.1158287353515625, 0.11600077056884765]",tokens/s,8.49010462268795,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1334.587392,1531.445248,0.0,884.998144,666.95168,s,10,0.4638772163391113,0.046387721633911125,0.00324639334269881,0.04602436828613281,0.047300929641723624,0.05149969654083251,0.05485871006011963,"[0.055698463439941406, 0.04545503997802734, 0.0443658561706543, 0.04362329483032226, 0.04399552154541016, 0.046367870330810544, 0.04607097625732422, 0.045977760314941406, 0.04610259246826172, 0.04621984100341797]",tokens/s,5518.701737936934,kWh,5.258427056265466e-07,2.8813751836908824e-07,2.2516487148978766e-06,3.0656289388935114e-06,tokens/kWh,83506518.59790932,MB,1334.587392,1531.445248,0.0,884.998144,733.634048,s,10,28.599342529296877,2.859934252929688,0.040992537619671546,2.881282958984375,2.8891374511718753,2.890091064453125,2.890853955078125,"[2.868544189453125, 2.84873681640625, 2.803044921875, 2.76381982421875, 2.87860400390625, 2.88738916015625, 2.891044677734375, 2.8839619140625, 2.888925537109375, 2.885271484375]",tokens/s,22.028478429342716,kWh,3.3778575510522465e-05,1.8510113271708782e-05,0.000126169204021707,0.00017845789280393823,tokens/kWh,353024.4530524329,,s,629,28.960572414398204,0.04604224549188902,0.00549098690307536,0.04566527938842774,0.04605522003173828,0.04659916839599609,0.08962334625244144,"[0.047293441772460934, 0.04731903839111328, 0.04763340759277344, 0.047879199981689456, 0.04722787094116211, 0.04738457489013672, 0.04755558395385742, 0.04675788879394531, 0.046929920196533206, 0.04694937515258789, 0.044088321685791014, 0.045520896911621096, 0.04534067153930664, 0.045363201141357425, 0.045367294311523435, 0.04549324798583984, 0.045143039703369144, 0.04571033477783203, 0.045456382751464845, 0.04532326507568359, 0.045469696044921876, 0.04475801467895508, 0.045200382232666016, 0.04537036895751953, 0.04541747283935547, 0.045333503723144535, 0.04537961578369141, 0.04523926544189453, 0.044461055755615236, 0.044260353088378904, 0.046604286193847655, 0.046878719329833986, 0.04556902313232422, 0.045192192077636716, 0.0452229118347168, 0.045279232025146485, 0.04519424057006836, 0.04516864013671875, 0.04509798431396484, 0.045369342803955076, 0.0450334701538086, 0.04421836853027344, 0.04367257690429688, 0.04433203125, 0.04518195343017578, 0.04521574401855469, 0.04520652770996094, 0.04537241744995117, 0.0452229118347168, 0.045276161193847655, 0.0450437126159668, 0.04531302261352539, 0.04500787353515625, 0.04507033538818359, 0.04519424057006836, 0.045254657745361325, 0.04499353790283203, 0.04528844833374023, 0.044747776031494144, 0.04507648086547852, 0.045156352996826174, 0.044183551788330076, 0.09082777404785156, 0.0450621452331543, 0.045308929443359375, 0.045434879302978515, 0.04517375946044922, 0.0451328010559082, 0.04527308654785156, 0.045042686462402344, 0.04517068862915039, 0.04490956878662109, 0.0451962890625, 0.04515327835083008, 0.045297664642333986, 0.045254657745361325, 0.04452454376220703, 0.04513177490234375, 0.045290496826171874, 0.04532940673828125, 0.04523417663574219, 0.04528844833374023, 0.04501401519775391, 0.045312000274658204, 0.044951553344726565, 0.04558950424194336, 0.0451512336730957, 0.04399718475341797, 0.04349542236328125, 0.04442521667480469, 0.045143039703369144, 0.045352958679199216, 0.045178878784179685, 0.04510003280639648, 0.04530688095092773, 0.04565401458740234, 0.04506828689575195, 0.045439998626708986, 0.044837886810302735, 0.04470783996582031, 0.04506419372558594, 0.04528947067260742, 0.045042686462402344, 0.04521779251098633, 0.04625100708007813, 0.04592127990722656, 0.04533760070800781, 0.0456181755065918, 0.045358081817626954, 0.04549427032470703, 0.045290496826171874, 0.04539699172973633, 0.0452229118347168, 0.04547788619995117, 0.0452229118347168, 0.045241344451904295, 0.0459048957824707, 0.04576563262939453, 0.04546867370605469, 0.04531609725952149, 0.045254657745361325, 0.04540825653076172, 0.04514713668823242, 0.045437950134277344, 0.045328384399414064, 0.08992870330810547, 0.04531814575195312, 0.04516659164428711, 0.045379585266113284, 0.04734668731689453, 0.04551475143432617, 0.0451143684387207, 0.04497612762451172, 0.04514713668823242, 0.04520652770996094, 0.04469964981079102, 0.04357734298706055, 0.043924480438232424, 0.04542668914794922, 0.04524544143676758, 0.045948928833007815, 0.04659302520751953, 0.04534988784790039, 0.04527308654785156, 0.04506521606445312, 0.04558950424194336, 0.04588748931884765, 0.045328384399414064, 0.04538880157470703, 0.04365824127197265, 0.04369100952148437, 0.043635711669921876, 0.043466751098632815, 0.04356915283203125, 0.04354457473754883, 0.0450334701538086, 0.04534476852416992, 0.044988414764404294, 0.04504678344726563, 0.043670528411865236, 0.043428863525390625, 0.04340838241577148, 0.043568126678466795, 0.04484096145629883, 0.045059070587158204, 0.045192192077636716, 0.045009918212890625, 0.044933120727539064, 0.04521984100341797, 0.04479897689819336, 0.04347391891479492, 0.04362444686889649, 0.04348108673095703, 0.0435865592956543, 0.04369715118408203, 0.04343606567382813, 0.04346569442749024, 0.0435968017578125, 0.044453887939453124, 0.04364287948608398, 0.04352102279663086, 0.04346879959106445, 0.04338790512084961, 0.043548671722412106, 0.04335615921020508, 0.04338483047485352, 0.043425792694091796, 0.04342476654052734, 0.08736153411865234, 0.04329369735717774, 0.04355891036987305, 0.04335308837890625, 0.043437057495117185, 0.043379711151123046, 0.04334284973144531, 0.04359270477294922, 0.04504678344726563, 0.04501708984375, 0.0452782096862793, 0.04522598266601562, 0.04501913452148437, 0.04498636627197266, 0.04552294540405273, 0.0451962890625, 0.04374425506591797, 0.043600894927978515, 0.043483135223388675, 0.043622398376464845, 0.045437950134277344, 0.04535603332519531, 0.04524544143676758, 0.04540415954589844, 0.04534886550903321, 0.044862464904785154, 0.04562124633789062, 0.045243392944335936, 0.043433982849121096, 0.0446453742980957, 0.04436787033081055, 0.04350054550170898, 0.04353433609008789, 0.04348211288452149, 0.043478015899658204, 0.04355481719970703, 0.04326502227783203, 0.043483135223388675, 0.043344894409179685, 0.04354457473754883, 0.04340326309204102, 0.04332339096069336, 0.04313600158691406, 0.04352716827392578, 0.04295475387573242, 0.04282470321655273, 0.042797054290771484, 0.043466751098632815, 0.043409408569335936, 0.04335001754760742, 0.04273971176147461, 0.04292812728881836, 0.04281753540039063, 0.04318105697631836, 0.043312126159667966, 0.043453441619873044, 0.04338995361328125, 0.04338380813598633, 0.043431934356689454, 0.04356095886230469, 0.04343500900268555, 0.04333772659301758, 0.043079681396484375, 0.08883814239501953, 0.04634009552001953, 0.04591308975219727, 0.045687808990478515, 0.04552908706665039, 0.04483174514770508, 0.04478771209716797, 0.044685310363769534, 0.04570521545410156, 0.04539084625244141, 0.04580966567993164, 0.04576563262939453, 0.04599193572998047, 0.04572159957885742, 0.04599398422241211, 0.045927425384521485, 0.04571750259399414, 0.0460052490234375, 0.046080001831054686, 0.045674495697021485, 0.045886463165283206, 0.045832191467285156, 0.04559564971923828, 0.0457891845703125, 0.04580352020263672, 0.04566835021972656, 0.045620223999023435, 0.04576563262939453, 0.04555980682373047, 0.045244415283203124, 0.045591552734375, 0.046069759368896485, 0.04571955108642578, 0.04584758377075195, 0.04578096008300781, 0.045709312438964846, 0.045644798278808595, 0.04558335876464844, 0.04583321762084961, 0.04563455963134765, 0.04575539016723633, 0.04590796661376953, 0.04571340942382812, 0.045742080688476565, 0.04559769439697266, 0.04568473434448242, 0.04496895980834961, 0.04558233642578125, 0.045693950653076174, 0.04574105453491211, 0.04589056015014648, 0.045770751953125, 0.046047233581542966, 0.045753345489501954, 0.045590526580810545, 0.045486080169677735, 0.04593561553955078, 0.0459417610168457, 0.045770751953125, 0.045835262298583986, 0.04582912063598633, 0.04597043228149414, 0.04582092666625977, 0.0928174057006836, 0.0458158073425293, 0.04572467041015625, 0.045830142974853515, 0.04579020690917969, 0.04571443176269531, 0.04543385696411133, 0.04542464065551758, 0.0458260498046875, 0.0464087028503418, 0.046339073181152345, 0.045641727447509765, 0.04585881423950195, 0.04595814514160156, 0.04600012969970703, 0.046080001831054686, 0.04580147171020508, 0.04584755325317383, 0.04584447860717773, 0.045699073791503904, 0.04588748931884765, 0.0456888313293457, 0.0465530891418457, 0.045725696563720705, 0.04565913772583008, 0.04573593521118164, 0.04579020690917969, 0.04585779190063476, 0.04585062408447266, 0.045638656616210936, 0.04574924850463867, 0.04630220794677734, 0.04582195281982422, 0.045736961364746094, 0.04595302581787109, 0.04646809768676758, 0.04595609664916992, 0.04573593521118164, 0.04565708923339844, 0.04538265609741211, 0.045385726928710936, 0.04591308975219727, 0.04603801727294922, 0.04588851165771484, 0.04543078231811523, 0.045467647552490234, 0.04634726333618164, 0.046064640045166014, 0.045712383270263675, 0.0456888313293457, 0.04573491287231445, 0.045663230895996096, 0.04560076904296875, 0.04567859268188477, 0.04550656127929688, 0.04568576049804687, 0.045851646423339845, 0.04569702529907226, 0.04691763305664062, 0.04611174392700195, 0.045590526580810545, 0.045434879302978515, 0.04539084625244141, 0.09221836853027343, 0.045976577758789064, 0.04601651382446289, 0.046132225036621094, 0.04583935928344727, 0.04540825653076172, 0.04572467041015625, 0.04590796661376953, 0.04581171035766601, 0.04589875030517578, 0.04597043228149414, 0.046914558410644534, 0.046176254272460936, 0.04570521545410156, 0.045871105194091794, 0.04604313659667969, 0.04611379241943359, 0.0459950065612793, 0.04579942321777344, 0.04605440139770508, 0.04578713607788086, 0.04581171035766601, 0.0458076171875, 0.04614144134521484, 0.045982719421386715, 0.04577177429199219, 0.04592025756835937, 0.04589567947387695, 0.045770751953125, 0.04581478500366211, 0.045897727966308595, 0.04571033477783203, 0.045298686981201174, 0.045843456268310545, 0.04603084945678711, 0.04576051330566406, 0.04577283096313477, 0.045750240325927734, 0.04595199966430664, 0.04575027084350586, 0.046806015014648435, 0.04596121597290039, 0.04588748931884765, 0.04559872055053711, 0.04534175872802734, 0.04549523162841797, 0.04582912063598633, 0.04582912063598633, 0.04583321762084961, 0.04575436782836914, 0.045835262298583986, 0.04631961441040039, 0.04589875030517578, 0.04560179138183594, 0.045870079040527346, 0.0457625617980957, 0.04565401458740234, 0.04590387344360351, 0.04660326385498047, 0.046034942626953124, 0.04577280044555664, 0.04576665496826172, 0.04574003219604492, 0.09243750762939452, 0.04573183822631836, 0.04562739181518555, 0.046186496734619144, 0.045868030548095705, 0.04568985748291016, 0.045889537811279295, 0.045281280517578126, 0.04573491287231445, 0.0455997428894043, 0.045884414672851564, 0.04579328155517578, 0.04552499389648437, 0.045868030548095705, 0.046015487670898435, 0.04595711898803711, 0.046339073181152345, 0.045818878173828126, 0.0457625617980957, 0.04579020690917969, 0.04585574340820313, 0.045213695526123046, 0.04567552185058594, 0.045810688018798826, 0.04566016006469727, 0.04592127990722656, 0.045638656616210936, 0.0466165771484375, 0.046183425903320315, 0.04563455963134765, 0.04580044937133789, 0.04568473434448242, 0.045744129180908207, 0.04573798370361328, 0.04571648025512695, 0.04611993789672852, 0.04575436782836914, 0.045655040740966796, 0.045628414154052735, 0.04565094375610351, 0.04580966567993164, 0.045748222351074216, 0.04578815841674805, 0.04575743865966797, 0.04519116973876953, 0.045948928833007815, 0.04573491287231445, 0.04561510467529297, 0.045687808990478515, 0.0458260498046875, 0.04564377593994141, 0.045946880340576174, 0.04564275360107422, 0.04571955108642578, 0.04561407852172852, 0.04564582443237305, 0.04579020690917969, 0.045744129180908207, 0.04556288146972656, 0.04597862243652344, 0.04569807815551758, 0.04569392013549805, 0.0456888313293457, 0.09245184326171875, 0.04586086273193359, 0.045725696563720705, 0.04577382278442383, 0.04579225540161133, 0.045818878173828126, 0.045638656616210936, 0.04554444885253906, 0.046058494567871096, 0.04570320129394531, 0.04595503997802734, 0.04560076904296875, 0.045592575073242186, 0.04572774505615235, 0.04578406524658203, 0.045841407775878903, 0.045851646423339845, 0.04590591812133789, 0.04557209777832031, 0.04575948715209961, 0.04550041580200195, 0.045720577239990234, 0.045856769561767576, 0.045612064361572266, 0.045733856201171874, 0.04566527938842774, 0.045638656616210936, 0.04602470397949219, 0.04580147171020508, 0.04579942321777344, 0.045661182403564454, 0.04551270294189453, 0.04584960174560547, 0.04568678283691406, 0.045644798278808595, 0.04698419189453125, 0.04665753555297852, 0.04563558578491211, 0.04597452926635742, 0.04597555160522461, 0.0457625617980957, 0.045736961364746094, 0.04576153564453125, 0.04577382278442383, 0.04587724685668945, 0.046061569213867185, 0.0459048957824707, 0.045758464813232425, 0.04592844772338867, 0.04589875030517578, 0.04586700820922852, 0.04588544082641602, 0.04613119888305664, 0.045889537811279295, 0.0456519660949707, 0.046102527618408204, 0.045827072143554685, 0.04580044937133789, 0.04578303909301758, 0.04575743865966797, 0.04575436782836914, 0.04682342529296875, 0.04640153503417969, 0.09238937377929687, 0.045795326232910154, 0.045484031677246094, 0.04558950424194336, 0.04579942321777344, 0.045777919769287106, 0.0458342399597168, 0.04578406524658203, 0.046647296905517575, 0.046922752380371094, 0.04616089630126953, 0.045710369110107424, 0.04566934585571289, 0.04571852874755859, 0.04588851165771484, 0.04604108810424805, 0.04571852874755859, 0.04572774505615235, 0.04561715316772461, 0.04571340942382812, 0.04581478500366211, 0.04596736145019531, 0.045736961364746094, 0.045780990600585936, 0.04554751968383789, 0.04597964859008789, 0.046071807861328126, 0.04575027084350586, 0.04564582443237305, 0.045774848937988284, 0.045830142974853515, 0.046018558502197264, 0.04574515151977539, 0.04558540725708008, 0.04558438491821289, 0.04561305618286133, 0.045748222351074216, 0.045864959716796876, 0.045682689666748044, 0.04570009613037109, 0.045948928833007815, 0.04582912063598633, 0.04562432098388672, 0.04510515213012695, 0.045298686981201174, 0.0455997428894043, 0.04581478500366211, 0.04572159957885742, 0.04562739181518555, 0.04561612701416016, 0.04584550476074219, 0.04610047912597656, 0.045864959716796876, 0.04583833694458008, 0.0455813102722168, 0.045706241607666016, 0.04583833694458008, 0.04589977645874024, 0.04574617767333984, 0.04578508758544922, 0.04575948715209961, 0.04603289413452148, 0.04592844772338867]",tokens/s,21.71918396499936,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-2b,google/gemma-2b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4755.23072,21498.42944,0.0,20851.982336,20236.60032,s,10,15.643266601562498,1.5643266601562498,0.001716851893423332,1.563902587890625,1.5660734985351563,1.567352362060547,1.5683754528808596,"[1.5631807861328124, 1.5647735595703125, 1.5629490966796875, 1.56462890625, 1.563903564453125, 1.5627757568359375, 1.565789306640625, 1.563901611328125, 1.5686312255859376, 1.5627327880859374]",tokens/s,163.6486844598237,kWh,1.846329511867629e-05,1.0117909425098331e-05,9.183712902519626e-05,0.00012041833356897088,tokens/kWh,2125922.1284055826,MB,4755.23072,21498.42944,0.0,20851.982336,20339.706368,s,10,879.77465625,87.97746562500001,0.0022909340685085054,87.97753515625,87.97981953125,87.980624609375,87.981268671875,"[87.974140625, 87.978578125, 87.9789609375, 87.9772265625, 87.979640625, 87.97715625, 87.97784375, 87.97353125, 87.9814296875, 87.9761484375]",tokens/s,0.716092462455496,kWh,0.0010385792660216491,0.0005692333667929961,0.005151564232359207,0.006759376865173853,tokens/kWh,9320.38577765846,,s,629,892.4578986816405,1.4188519851854382,0.185906583948783,1.396368408203125,1.3966589599609376,1.3968528564453124,2.960364462890625,"[1.39618505859375, 1.39667041015625, 1.3964708251953124, 1.396378662109375, 1.3963243408203125, 1.3962271728515625, 1.3963038330078126, 1.3964288330078125, 1.3964349365234374, 1.396369384765625, 1.3964697265625, 1.3965823974609375, 1.3962998046875, 1.3964708251953124, 1.3964349365234374, 1.3963857421875, 1.396304931640625, 1.3962802734375, 1.3960714111328125, 1.396305908203125, 1.3961553955078125, 1.3963427734375, 1.396221923828125, 1.3963756103515625, 1.39614111328125, 1.396317138671875, 1.3963294677734375, 1.3963673095703124, 1.396279296875, 1.3963018798828124, 1.3959813232421876, 1.3964195556640624, 1.396304931640625, 1.396304931640625, 1.396275146484375, 1.3961236572265625, 1.3962794189453125, 1.396401123046875, 1.39871435546875, 1.396484130859375, 1.396558837890625, 1.396505615234375, 1.3966868896484375, 1.3962772216796875, 1.3962833251953124, 1.3960928955078125, 1.396380615234375, 1.3968824462890626, 1.3964442138671875, 1.396285400390625, 1.3964237060546876, 1.396358154296875, 1.396305908203125, 1.3966787109375, 1.3965537109375, 1.3964185791015624, 1.3962916259765625, 1.3968353271484375, 1.396358154296875, 1.396221923828125, 1.39633251953125, 1.396252685546875, 2.962231201171875, 1.396673583984375, 1.3970401611328125, 1.396295654296875, 1.3965557861328124, 1.396348876953125, 1.3965035400390624, 1.3961922607421875, 1.3968536376953125, 1.39629052734375, 1.396274169921875, 1.396199462890625, 1.3962281494140625, 1.3965833740234375, 1.396538330078125, 1.3962445068359375, 1.3963079833984375, 1.3963243408203125, 1.396515869140625, 1.3962762451171875, 1.396073486328125, 1.3978162841796875, 1.396114501953125, 1.396137939453125, 1.3962198486328126, 1.3962987060546874, 1.3961595458984375, 1.3963714599609376, 1.396452392578125, 1.3961502685546876, 1.3962720947265626, 1.3963868408203124, 1.3964185791015624, 1.3962239990234375, 1.396464599609375, 1.3964482421875, 1.396427734375, 1.396380615234375, 1.396588623046875, 1.3964810791015625, 1.3967718505859374, 1.3981634521484374, 1.3962802734375, 1.396275146484375, 1.3962301025390624, 1.396253662109375, 1.396327392578125, 1.3961728515625, 1.3966571044921876, 1.396474853515625, 1.3965128173828125, 1.3963858642578124, 1.396339599609375, 1.3961318359375, 1.3965147705078125, 1.3963387451171876, 1.396506591796875, 1.3962393798828125, 1.3964605712890625, 1.3965179443359375, 1.39639599609375, 1.3985382080078126, 1.3965301513671875, 2.9603779296875, 1.396199462890625, 1.3967340087890625, 1.3962291259765625, 1.3964554443359376, 1.396252685546875, 1.3964892578125, 1.3960601806640625, 1.39671240234375, 1.3963714599609376, 1.3963509521484374, 1.3961092529296875, 1.39653125, 1.3965814208984375, 1.3964400634765626, 1.3963172607421874, 1.39658349609375, 1.3963939208984375, 1.396515869140625, 1.3962587890625, 1.3965823974609375, 1.3966663818359375, 1.396378662109375, 1.3987901611328124, 1.3961759033203125, 1.3963653564453125, 1.3963018798828124, 1.3964093017578125, 1.39652197265625, 1.3961544189453126, 1.3962332763671874, 1.39641650390625, 1.39623828125, 1.3962393798828125, 1.3961964111328125, 1.396336669921875, 1.3964564208984376, 1.3965137939453125, 1.3963756103515625, 1.3962701416015626, 1.396137939453125, 1.396569091796875, 1.39635205078125, 1.3963131103515625, 1.396252685546875, 1.3963427734375, 1.3963970947265625, 1.3995396728515626, 1.39681591796875, 1.396443115234375, 1.396485107421875, 1.3962608642578125, 1.3962506103515624, 1.3963192138671876, 1.3964769287109375, 1.396210693359375, 1.3968311767578125, 1.396547607421875, 1.39646875, 1.39649951171875, 1.396304931640625, 1.3964041748046876, 1.3961563720703125, 2.961944580078125, 1.39614208984375, 1.396427734375, 1.39610009765625, 1.3963775634765625, 1.397116943359375, 1.3965689697265624, 1.396379638671875, 1.396430908203125, 1.3963284912109375, 1.3966080322265626, 1.3961544189453126, 1.3964501953125, 1.396547607421875, 1.396279296875, 1.3963203125, 1.3963448486328125, 1.396454345703125, 1.3965128173828125, 1.396253662109375, 1.396316162109375, 1.396046875, 1.396062255859375, 1.3964974365234375, 1.396443115234375, 1.3962864990234376, 1.3960980224609374, 1.3964810791015625, 1.3962884521484376, 1.396199462890625, 1.396241455078125, 1.3960233154296875, 1.3961175537109376, 1.396675537109375, 1.3962762451171875, 1.3964237060546876, 1.3964083251953125, 1.3962884521484376, 1.3965762939453126, 1.3964708251953124, 1.396631591796875, 1.3964022216796874, 1.396336669921875, 1.396317138671875, 1.400958984375, 1.396559814453125, 1.3964820556640625, 1.39627001953125, 1.3966888427734374, 1.3963427734375, 1.396474853515625, 1.3962342529296874, 1.3963448486328125, 1.39620654296875, 1.396552734375, 1.3967205810546874, 1.3966029052734374, 1.3963560791015626, 1.3962332763671874, 1.3963212890625, 1.3964134521484375, 1.396279296875, 1.39620654296875, 2.96119921875, 1.3962393798828125, 1.39626806640625, 1.3964124755859375, 1.3966131591796875, 1.396216796875, 1.3965548095703124, 1.3961031494140625, 1.396580322265625, 1.396380615234375, 1.3969398193359375, 1.396273193359375, 1.39626904296875, 1.39641650390625, 1.3962547607421876, 1.39647900390625, 1.396178955078125, 1.3985556640625, 1.3966253662109376, 1.3968516845703125, 1.3966131591796875, 1.3960867919921875, 1.39594140625, 1.3963447265625, 1.396115478515625, 1.396316162109375, 1.3962301025390624, 1.3962618408203125, 1.3963406982421875, 1.3959669189453126, 1.396400146484375, 1.3961195068359376, 1.3963663330078124, 1.396368408203125, 1.3962496337890624, 1.39649951171875, 1.3963284912109375, 1.3964493408203125, 1.3962188720703126, 1.396115478515625, 1.3963602294921875, 1.3963663330078124, 1.3966182861328125, 1.3964237060546876, 1.3963038330078126, 1.3963653564453125, 1.39626806640625, 1.396369384765625, 1.3965137939453125, 1.3962486572265624, 1.3964317626953124, 1.3962711181640626, 1.39637353515625, 1.3964647216796875, 1.39634375, 1.3963253173828125, 1.396611083984375, 1.4011822509765626, 1.3966837158203125, 1.3966038818359374, 1.39641650390625, 1.3965537109375, 1.3965322265625, 2.960329833984375, 1.3966304931640625, 1.396526123046875, 1.3960919189453125, 1.396494384765625, 1.39626904296875, 1.3971036376953125, 1.39640625, 1.3964461669921875, 1.3961175537109376, 1.3965537109375, 1.3965025634765624, 1.3961400146484375, 1.396515869140625, 1.396490234375, 1.396189208984375, 1.3961563720703125, 1.3964564208984376, 1.3965404052734376, 1.396285400390625, 1.3961553955078125, 1.396074462890625, 1.395947509765625, 1.3964676513671874, 1.3965732421875, 1.3963448486328125, 1.3962977294921874, 1.3964267578125, 1.3964461669921875, 1.39627001953125, 1.3965732421875, 1.396262939453125, 1.39629052734375, 1.39640625, 1.3961185302734376, 1.3963243408203125, 1.3963826904296874, 1.3964442138671875, 1.400627197265625, 1.396809814453125, 1.396582275390625, 1.396357177734375, 1.39643798828125, 1.3963756103515625, 1.3962301025390624, 1.3963294677734375, 1.39649951171875, 1.3963336181640624, 1.3967288818359376, 1.396634521484375, 1.3963294677734375, 1.3961072998046875, 1.3963233642578126, 1.3963182373046874, 1.39650146484375, 1.396305908203125, 1.3965128173828125, 1.396368408203125, 1.3962659912109374, 1.39628955078125, 1.3963355712890626, 1.39620654296875, 1.39628955078125, 2.96319189453125, 1.39616259765625, 1.3965670166015625, 1.39634375, 1.39635205078125, 1.3960509033203126, 1.39610107421875, 1.3959833984375, 1.3964052734375, 1.3961861572265626, 1.39652197265625, 1.39647998046875, 1.3962496337890624, 1.3968414306640624, 1.3966519775390625, 1.3960438232421875, 1.3962802734375, 1.3961964111328125, 1.3965936279296876, 1.3964493408203125, 1.3962025146484376, 1.396211669921875, 1.3961871337890626, 1.3990369873046875, 1.39656396484375, 1.39658447265625, 1.3963192138671876, 1.3963756103515625, 1.396262939453125, 1.3963499755859374, 1.3965291748046875, 1.3964564208984376, 1.396116455078125, 1.3963765869140625, 1.396496337890625, 1.396341796875, 1.3961666259765626, 1.396453369140625, 1.396306884765625, 1.3964974365234375, 1.396357177734375, 1.396156494140625, 1.396138916015625, 1.39642578125, 1.3961502685546876, 1.396420654296875, 1.3963642578125, 1.39631201171875, 1.396749267578125, 1.396473876953125, 1.396526123046875, 1.396474853515625, 1.3964871826171874, 1.396432861328125, 1.396547607421875, 1.3964410400390626, 1.4002022705078125, 1.3966510009765625, 1.39624658203125, 1.396135986328125, 1.3961492919921874, 1.3963140869140624, 1.39603759765625, 2.96123388671875, 1.3961964111328125, 1.3963387451171876, 1.3962198486328126, 1.39618408203125, 1.3961973876953124, 1.3963714599609376, 1.396177978515625, 1.3965966796875, 1.396106201171875, 1.396715576171875, 1.3962823486328124, 1.3963294677734375, 1.3963028564453126, 1.396380615234375, 1.3961964111328125, 1.398497314453125, 1.3963970947265625, 1.3965679931640624, 1.396317138671875, 1.3965782470703125, 1.3963919677734375, 1.3962659912109374, 1.3962720947265626, 1.396201416015625, 1.396317138671875, 1.3962393798828125, 1.396137939453125, 1.3964708251953124, 1.3964288330078125, 1.3963857421875, 1.3960755615234375, 1.3963480224609375, 1.396232177734375, 1.3961451416015624, 1.3963182373046874, 1.396641845703125, 1.3966080322265626, 1.396378662109375, 1.396273193359375, 1.3963345947265624, 1.3962322998046874, 1.3966744384765626, 1.3964073486328126, 1.3963243408203125, 1.3962373046875, 1.3963499755859374, 1.3963345947265624, 1.3965465087890625, 1.396432861328125, 1.39626904296875, 1.39639404296875, 1.3961748046875, 1.3964892578125, 1.3966878662109374, 1.396232177734375, 1.396611083984375, 1.3963765869140625, 1.39654248046875, 1.39650048828125, 1.39634375, 1.3964093017578125, 1.3962454833984375, 2.96591357421875, 1.3961614990234374, 1.396864013671875, 1.3961820068359374, 1.396379638671875, 1.3964666748046874, 1.3964503173828124, 1.3963642578125, 1.39664892578125, 1.39633154296875, 1.396452392578125, 1.3962650146484374, 1.396241455078125, 1.3963756103515625, 1.3963775634765625, 1.396111328125, 1.3964503173828124, 1.3980078125, 1.3967994384765625, 1.396642822265625, 1.3961103515625, 1.3965987548828125, 1.396348876953125, 1.3962291259765625, 1.396262939453125, 1.39631201171875, 1.3963212890625, 1.3961759033203125, 1.3965260009765625, 1.3965496826171875, 1.396384765625, 1.3963182373046874, 1.396358154296875, 1.3962158203125, 1.3962720947265626, 1.396420654296875, 1.3962281494140625, 1.3967728271484374, 1.3964512939453124, 1.39666845703125, 1.3965936279296876, 1.3961728515625, 1.39624853515625, 1.3961861572265626, 1.39656396484375, 1.3964697265625, 1.396485107421875, 1.396537353515625, 1.3965670166015625, 1.3964073486328126, 1.3962578125, 1.3964390869140626, 1.3966990966796875, 1.39630078125, 1.3968045654296875, 1.3967421875, 1.3966285400390626, 1.3964482421875, 1.396533203125, 1.396506591796875, 1.396284423828125, 1.3962977294921874, 1.400701904296875, 2.960291748046875, 1.3964564208984376, 1.396536376953125, 1.396294677734375, 1.396358154296875, 1.39620458984375, 1.3963621826171875, 1.3960714111328125, 1.396537353515625, 1.396304931640625, 1.3964451904296875, 1.396430908203125, 1.3963642578125, 1.3963837890625, 1.396828125, 1.3963817138671875, 1.39660498046875, 1.396589599609375, 1.39631201171875, 1.396283447265625, 1.396177978515625, 1.39618505859375, 1.396168701171875, 1.396116455078125, 1.39642578125, 1.396654052734375, 1.399869384765625, 1.3961328125, 1.3965076904296876, 1.3962025146484376, 1.3964339599609374, 1.3961707763671876, 1.3962506103515624, 1.396093994140625, 1.3964124755859375, 1.396494384765625, 1.3964676513671874, 1.396410400390625, 1.396167724609375, 1.3962772216796875, 1.396236328125, 1.3962117919921875, 1.39633251953125, 1.396305908203125, 1.396552734375, 1.3963140869140624, 1.39652197265625, 1.396462646484375, 1.3968414306640624, 1.39652099609375, 1.396305908203125, 1.396357177734375, 1.3965721435546874, 1.39648193359375, 1.396537353515625, 1.396272216796875, 1.396579345703125, 1.3963602294921875, 1.3965322265625, 1.396306884765625, 1.3964041748046876, 1.396275146484375, 1.3961829833984376]",tokens/s,0.7047951516023033,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6690,7 +7062,7 @@ ChildProcessError: Traceback (most recent call last): response.raise_for_status() File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) -requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/google/gemma-2b/resolve/main/config.json +requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json The above exception was the direct cause of the following exception: @@ -6707,10 +7079,10 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee71-77412ec3156496f90ced9eed;dd132f6c-329e-43f2-9750-718ed7180f6a) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4ac5-3c9fa7a5650a59c842e84cf5;eff27cf1-71d9-4a38-a3d4-70b70d9a9fc8) 403 Forbidden: Authorization error.. -Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. +Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. If you are trying to create or update content,make sure you have a token with the `write` role. The above exception was the direct cause of the following exception: @@ -6749,11 +7121,158 @@ Traceback (most recent call last): resolved_config_file = cached_file( File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 442, in cached_file raise EnvironmentError( -OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like google/gemma-2b is not the path to a directory containing a file named config.json. +OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,-,-,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status + response.raise_for_status() + File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status + raise HTTPError(http_error_msg, response=self) +requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/a/resolve/main/config.json + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 399, in cached_file + resolved_file = hf_hub_download( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download + return _hf_hub_download_to_cache_dir( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir + _raise_on_head_call_error(head_call_error, force_download, local_files_only) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1823, in _raise_on_head_call_error + raise head_call_error + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error + metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata + r = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper + response = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper + hf_raise_for_status(response) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status + raise RepositoryNotFoundError(message, response) from e +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d452a-2d2f4bd16e2901406b8f2937;7fb76edc-f1b3-48bd-b1a2-fc021736abd1) + +Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. +Please make sure you specified the correct `repo_id` and `repo_type`. +If you are trying to access a private or gated repo, make sure you are authenticated. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ + super().__init__(config) + File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ + self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) + File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config + return AutoConfig.from_pretrained(model, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict + resolved_config_file = cached_file( + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 422, in cached_file + raise EnvironmentError( +OSError: a is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' +If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2111.176704,2846.359552,0.0,2199.912448,2075.149824,s,10,1.345602310180664,0.1345602310180664,0.0004287422965996787,0.13438082885742186,0.13517186279296875,0.13521602783203124,0.13525135986328124,"[0.1346531219482422, 0.13415478515625, 0.13441477966308593, 0.13409397888183594, 0.13418368530273436, 0.13434687805175782, 0.13423942565917968, 0.13526019287109375, 0.13516204833984374, 0.13509341430664062]",tokens/s,1902.4937610699312,kWh,1.5845098716241345e-06,8.681900968003902e-07,7.312065108906473e-06,9.764765077330997e-06,tokens/kWh,26216708.540618826,MB,2111.176704,2846.359552,0.0,2199.912448,2180.296192,s,10,72.99156640625,7.299156640625002,0.0010886665313993955,7.29881201171875,7.30088486328125,7.30103251953125,7.3011506445312495,"[7.30118017578125, 7.29998193359375, 7.298513671875, 7.29880029296875, 7.299087890625, 7.30085205078125, 7.29882373046875, 7.29793994140625, 7.29786962890625, 7.29851708984375]",tokens/s,8.631134129847297,kWh,8.616290877704268e-05,4.722355870710817e-05,0.0003964600079084879,0.0005298464753926387,tokens/kWh,118902.36686639149,,s,629,74.08624839782712,0.11778417869288894,0.016011389929587044,0.1158440933227539,0.11595162048339844,0.11602903137207031,0.25029092712402345,"[0.1164031982421875, 0.11662540435791016, 0.11636224365234375, 0.11587174224853515, 0.11590144348144531, 0.11583078765869141, 0.11578265380859375, 0.11591990661621093, 0.11577442932128906, 0.11571199798583984, 0.11582054138183594, 0.11574272155761718, 0.11578880310058594, 0.11589119720458985, 0.11576627349853516, 0.11583487701416016, 0.11584716796875, 0.11574578857421874, 0.1157570571899414, 0.11593215942382812, 0.11580108642578125, 0.11578675079345703, 0.11586764526367188, 0.11578265380859375, 0.11584819030761719, 0.11595059204101563, 0.11578163146972656, 0.11580723571777343, 0.11586252593994141, 0.11579904174804688, 0.11576627349853516, 0.11589631652832032, 0.11577753448486328, 0.11583590698242187, 0.11595366668701172, 0.1158287353515625, 0.11581644439697265, 0.1159557113647461, 0.11614822387695313, 0.11589222717285157, 0.11589631652832032, 0.11583487701416016, 0.1157734375, 0.11594956970214844, 0.11578572845458984, 0.11590758514404297, 0.1158666229248047, 0.11585126495361328, 0.1157949447631836, 0.11599155426025391, 0.11582669067382813, 0.11588813018798828, 0.1159557113647461, 0.11582975769042969, 0.11576115417480469, 0.11599155426025391, 0.11583795166015624, 0.11594751739501953, 0.11585228729248047, 0.11586048126220704, 0.11577651214599609, 0.11595164489746093, 0.250293212890625, 0.11575910186767578, 0.11586867523193359, 0.11589529418945313, 0.11572121429443359, 0.1158635482788086, 0.11580108642578125, 0.11569152069091797, 0.11583385467529297, 0.11572633361816406, 0.11583487701416016, 0.11589427185058594, 0.11576319885253906, 0.11582361602783203, 0.11587379455566406, 0.11576831817626954, 0.11571609497070312, 0.11584204864501953, 0.11576217651367188, 0.11584819030761719, 0.11591168212890625, 0.11580928039550781, 0.11572838592529297, 0.11591680145263672, 0.11588813018798828, 0.11593727874755859, 0.11585638427734375, 0.11580210876464844, 0.115704833984375, 0.1159208984375, 0.11575193786621094, 0.11585945892333985, 0.1158656005859375, 0.11578470611572265, 0.11577037048339844, 0.11588813018798828, 0.11587481689453125, 0.11586969757080078, 0.116063232421875, 0.11596800231933593, 0.11581849670410156, 0.11589017486572266, 0.1158502426147461, 0.11583078765869141, 0.11595673370361329, 0.11586048126220704, 0.11586969757080078, 0.1164031982421875, 0.11587891387939453, 0.11589836883544923, 0.11601817321777344, 0.11585945892333985, 0.11583283233642579, 0.11597618865966797, 0.11582054138183594, 0.11586969757080078, 0.115957763671875, 0.1162455062866211, 0.11581542205810547, 0.11584512329101562, 0.11583897399902343, 0.11605299377441407, 0.11594342041015625, 0.2505707550048828, 0.11568742370605468, 0.1159004135131836, 0.11583385467529297, 0.11577139282226563, 0.1158656005859375, 0.11585330963134766, 0.1156864013671875, 0.11587481689453125, 0.11572121429443359, 0.11576525115966797, 0.115884033203125, 0.1157201919555664, 0.11576729583740235, 0.11580108642578125, 0.1157673568725586, 0.1157989730834961, 0.11591577911376953, 0.1157550048828125, 0.11574886322021484, 0.11581337738037109, 0.11583795166015624, 0.11579596710205078, 0.11599052429199219, 0.1157949447631836, 0.11579702758789062, 0.11584925079345704, 0.1157795181274414, 0.11584819030761719, 0.1159393310546875, 0.1158123550415039, 0.11577145385742188, 0.1158583984375, 0.11581948852539062, 0.11586873626708985, 0.11593619537353515, 0.11581439971923828, 0.1158502426147461, 0.11589222717285157, 0.11580723571777343, 0.11586867523193359, 0.11591270446777344, 0.11583795166015624, 0.11578982543945313, 0.11590656280517578, 0.11583590698242187, 0.1159393310546875, 0.11591986846923828, 0.11583590698242187, 0.11578880310058594, 0.11588098907470704, 0.115810302734375, 0.11593007659912109, 0.11598540496826172, 0.11583078765869141, 0.11577037048339844, 0.11591474914550781, 0.1158440933227539, 0.11588301086425781, 0.11591065979003906, 0.11593523406982421, 0.11621990203857421, 0.11591577911376953, 0.25028504943847657, 0.11570687866210938, 0.11593421173095703, 0.1157949447631836, 0.1157734375, 0.11591986846923828, 0.11586867523193359, 0.11574578857421874, 0.11593727874755859, 0.1157396469116211, 0.11580723571777343, 0.11589734649658204, 0.11572223663330078, 0.11589324951171875, 0.11581337738037109, 0.11616973114013672, 0.11582566070556641, 0.115884033203125, 0.11577139282226563, 0.11591065979003906, 0.11591270446777344, 0.11577855682373046, 0.11584614562988281, 0.11589631652832032, 0.1157754898071289, 0.11585126495361328, 0.1158123550415039, 0.11571507263183593, 0.11579289245605469, 0.11591372680664062, 0.115810302734375, 0.11588198089599609, 0.11582566070556641, 0.11576934051513672, 0.11584819030761719, 0.11589939117431641, 0.11580518341064452, 0.115884033203125, 0.11592192077636719, 0.11581132507324218, 0.11585126495361328, 0.11589939117431641, 0.11581439971923828, 0.11580825805664062, 0.11589836883544923, 0.11582669067382813, 0.1158809585571289, 0.11590860748291015, 0.11583999633789062, 0.11585330963134766, 0.11591372680664062, 0.11576525115966797, 0.11583385467529297, 0.11591168212890625, 0.11582669067382813, 0.11581747436523437, 0.11591782379150391, 0.11587481689453125, 0.11585330963134766, 0.11589119720458985, 0.11584102630615234, 0.11583283233642579, 0.11592499542236329, 0.250281982421875, 0.1157201919555664, 0.11594239807128906, 0.11578470611572265, 0.11578675079345703, 0.11585536193847656, 0.11581337738037109, 0.11564441680908204, 0.11579084777832031, 0.11577961730957032, 0.11587478637695313, 0.11586969757080078, 0.1157232666015625, 0.11570175933837891, 0.11586252593994141, 0.11580928039550781, 0.11572531127929687, 0.11594445037841797, 0.11586457824707032, 0.11585228729248047, 0.11585536193847656, 0.1158123550415039, 0.11578368377685547, 0.11595161437988281, 0.11578777313232422, 0.11574169921875, 0.11583692932128906, 0.11585126495361328, 0.11589017486572266, 0.11593113708496093, 0.11582771301269532, 0.11573452758789063, 0.11586457824707032, 0.1158440933227539, 0.11572633361816406, 0.11594649505615234, 0.11585228729248047, 0.11630182647705078, 0.11595366668701172, 0.11585842895507813, 0.11582361602783203, 0.11603353881835937, 0.11589427185058594, 0.11576217651367188, 0.11597926330566406, 0.11593421173095703, 0.11581951904296875, 0.11588607788085938, 0.11583487701416016, 0.11584921264648437, 0.11600077056884765, 0.11583078765869141, 0.11583692932128906, 0.1159925765991211, 0.11582566070556641, 0.11570995330810546, 0.11591986846923828, 0.11591680145263672, 0.11578265380859375, 0.11595263671875, 0.11586156463623047, 0.11594847869873047, 0.11594035339355468, 0.2504622039794922, 0.1157201919555664, 0.11593215942382812, 0.11588198089599609, 0.11573760223388672, 0.11585126495361328, 0.1158287353515625, 0.11570381164550782, 0.11582975769042969, 0.11573554992675782, 0.11576627349853516, 0.11585740661621094, 0.1158123550415039, 0.11576422119140625, 0.11585842895507813, 0.11579392242431641, 0.11585842895507813, 0.11592192077636719, 0.11579289245605469, 0.11573248291015625, 0.11583999633789062, 0.1158287353515625, 0.11580825805664062, 0.11603353881835937, 0.11582669067382813, 0.11581132507324218, 0.11583590698242187, 0.11581449890136719, 0.11593513488769532, 0.11594342041015625, 0.11582975769042969, 0.11581747436523437, 0.11584921264648437, 0.11580313873291016, 0.1158318099975586, 0.11590553283691406, 0.1159331817626953, 0.11579698944091797, 0.1159362564086914, 0.11583795166015624, 0.11647283172607421, 0.1159331817626953, 0.1158635482788086, 0.11575193786621094, 0.11594035339355468, 0.1157949447631836, 0.11586048126220704, 0.11586969757080078, 0.11589631652832032, 0.11599155426025391, 0.11657625579833984, 0.11596492767333984, 0.11592704010009766, 0.11591986846923828, 0.1158656005859375, 0.11576525115966797, 0.11585433959960938, 0.11580416107177735, 0.11588301086425781, 0.11662028503417969, 0.11592601776123047, 0.11577855682373046, 0.11592704010009766, 0.25042636108398436, 0.11569664001464844, 0.11593113708496093, 0.11581132507324218, 0.11573145294189453, 0.11585740661621094, 0.11574886322021484, 0.11579392242431641, 0.11582672119140625, 0.11573244476318359, 0.11576217651367188, 0.11587789154052734, 0.11574886322021484, 0.11579289245605469, 0.11580723571777343, 0.11574272155761718, 0.11577855682373046, 0.11590144348144531, 0.11581747436523437, 0.11593113708496093, 0.1158502426147461, 0.1157734375, 0.115810302734375, 0.115884033203125, 0.1157734375, 0.11586457824707032, 0.11584204864501953, 0.11577037048339844, 0.11582771301269532, 0.115884033203125, 0.11576831817626954, 0.11585330963134766, 0.11584614562988281, 0.11580313873291016, 0.1159710693359375, 0.11589017486572266, 0.11580825805664062, 0.11585945892333985, 0.11588102722167969, 0.11577235412597656, 0.11583283233642579, 0.11589017486572266, 0.11583385467529297, 0.11590962982177734, 0.1159331817626953, 0.11623423767089844, 0.11592601776123047, 0.11588710021972656, 0.11581849670410156, 0.11585433959960938, 0.11590962982177734, 0.11579698944091797, 0.11589017486572266, 0.11592396545410157, 0.11611135864257813, 0.11585945892333985, 0.11591782379150391, 0.11581644439697265, 0.11584307098388671, 0.11585849761962891, 0.11583891296386718, 0.1158318099975586, 0.11591270446777344, 0.25133772277832034, 0.11565670776367187, 0.11591680145263672, 0.11575398254394531, 0.11577753448486328, 0.11584204864501953, 0.11576422119140625, 0.11593421173095703, 0.11582975769042969, 0.11589119720458985, 0.1157734375, 0.1158656005859375, 0.11576831817626954, 0.11573554992675782, 0.11584204864501953, 0.11581747436523437, 0.11577855682373046, 0.11632844543457031, 0.11574988555908203, 0.11584921264648437, 0.11586764526367188, 0.11579289245605469, 0.11578265380859375, 0.11591270446777344, 0.1157918701171875, 0.11573862457275391, 0.11580620574951171, 0.11574476623535156, 0.11570790100097657, 0.11589222717285157, 0.11574784088134765, 0.11576525115966797, 0.11582259368896484, 0.11581644439697265, 0.11572940826416016, 0.1158666229248047, 0.11577446746826171, 0.11585740661621094, 0.11585228729248047, 0.11583692932128906, 0.11584614562988281, 0.11599871826171874, 0.11582157135009766, 0.11573862457275391, 0.11592704010009766, 0.11583897399902343, 0.11578777313232422, 0.11583385467529297, 0.1158666229248047, 0.11591372680664062, 0.11602124786376954, 0.11583078765869141, 0.11584614562988281, 0.11594035339355468, 0.11581132507324218, 0.11575193786621094, 0.11592908477783204, 0.11587481689453125, 0.11580620574951171, 0.11590348815917968, 0.11584307098388671, 0.11584716796875, 0.1159557113647461, 0.2511810607910156, 0.11569152069091797, 0.11590860748291015, 0.115884033203125, 0.11576729583740235, 0.11587174224853515, 0.11579801940917969, 0.11567411041259766, 0.11580416107177735, 0.11567411041259766, 0.11572940826416016, 0.11585228729248047, 0.1157580795288086, 0.11572633361816406, 0.11580825805664062, 0.11571814727783203, 0.11586252593994141, 0.11594137573242187, 0.11578368377685547, 0.11573452758789063, 0.11582259368896484, 0.11580825805664062, 0.11578572845458984, 0.11587382507324219, 0.11579388427734374, 0.11576118469238281, 0.11580515289306641, 0.11575398254394531, 0.11578368377685547, 0.11639603424072266, 0.11579904174804688, 0.11602227020263672, 0.11582975769042969, 0.11574681854248046, 0.11587891387939453, 0.11587789154052734, 0.11582361602783203, 0.11588710021972656, 0.1158809585571289, 0.11595263671875, 0.11590656280517578, 0.11593421173095703, 0.11585228729248047, 0.115736572265625, 0.11606937408447265, 0.11581443023681641, 0.11588809967041015, 0.115884033203125, 0.11585126495361328, 0.1157550048828125, 0.11587686157226562, 0.1157734375, 0.11586969757080078, 0.11592601776123047, 0.11581747436523437, 0.11576012420654297, 0.11587583923339843, 0.11580416107177735, 0.11582566070556641, 0.11587276458740234, 0.11586764526367188, 0.11577855682373046, 0.11586457824707032, 0.25122508239746094, 0.11569459533691406, 0.1159557113647461, 0.11578777313232422, 0.11566694641113281, 0.11580416107177735, 0.11576319885253906, 0.11574988555908203, 0.11585536193847656, 0.11570893096923827, 0.11571199798583984, 0.11583590698242187, 0.11570585632324219, 0.11578880310058594, 0.11586867523193359, 0.11572838592529297, 0.11579698944091797, 0.11587993621826172, 0.1157570571899414, 0.11582669067382813, 0.1157918701171875, 0.11579289245605469, 0.11579596710205078, 0.11596390533447265, 0.11572633361816406, 0.115810302734375, 0.11584512329101562, 0.11574066925048829, 0.11578572845458984, 0.11587276458740234, 0.11574784088134765, 0.11581132507324218, 0.11581132507324218, 0.115810302734375, 0.11586252593994141, 0.11587993621826172, 0.11576831817626954, 0.11585945892333985, 0.11592294311523438, 0.11580825805664062, 0.11584921264648437, 0.1159208984375, 0.11589631652832032, 0.11581951904296875, 0.11599980926513671, 0.115837890625, 0.11596185302734376, 0.11590962982177734, 0.11582669067382813, 0.1158666229248047, 0.11588198089599609, 0.116347900390625, 0.11593830108642578, 0.11591270446777344, 0.11585945892333985, 0.11579698944091797, 0.11591782379150391, 0.11618201446533204, 0.11581644439697265, 0.11589222717285157, 0.11585330963134766, 0.1158287353515625, 0.11600077056884765]",tokens/s,8.49010462268795,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1334.587392,1531.445248,0.0,884.998144,666.95168,s,10,0.4638772163391113,0.046387721633911125,0.00324639334269881,0.04602436828613281,0.047300929641723624,0.05149969654083251,0.05485871006011963,"[0.055698463439941406, 0.04545503997802734, 0.0443658561706543, 0.04362329483032226, 0.04399552154541016, 0.046367870330810544, 0.04607097625732422, 0.045977760314941406, 0.04610259246826172, 0.04621984100341797]",tokens/s,5518.701737936934,kWh,5.258427056265466e-07,2.8813751836908824e-07,2.2516487148978766e-06,3.0656289388935114e-06,tokens/kWh,83506518.59790932,MB,1334.587392,1531.445248,0.0,884.998144,733.634048,s,10,28.599342529296877,2.859934252929688,0.040992537619671546,2.881282958984375,2.8891374511718753,2.890091064453125,2.890853955078125,"[2.868544189453125, 2.84873681640625, 2.803044921875, 2.76381982421875, 2.87860400390625, 2.88738916015625, 2.891044677734375, 2.8839619140625, 2.888925537109375, 2.885271484375]",tokens/s,22.028478429342716,kWh,3.3778575510522465e-05,1.8510113271708782e-05,0.000126169204021707,0.00017845789280393823,tokens/kWh,353024.4530524329,,s,629,28.960572414398204,0.04604224549188902,0.00549098690307536,0.04566527938842774,0.04605522003173828,0.04659916839599609,0.08962334625244144,"[0.047293441772460934, 0.04731903839111328, 0.04763340759277344, 0.047879199981689456, 0.04722787094116211, 0.04738457489013672, 0.04755558395385742, 0.04675788879394531, 0.046929920196533206, 0.04694937515258789, 0.044088321685791014, 0.045520896911621096, 0.04534067153930664, 0.045363201141357425, 0.045367294311523435, 0.04549324798583984, 0.045143039703369144, 0.04571033477783203, 0.045456382751464845, 0.04532326507568359, 0.045469696044921876, 0.04475801467895508, 0.045200382232666016, 0.04537036895751953, 0.04541747283935547, 0.045333503723144535, 0.04537961578369141, 0.04523926544189453, 0.044461055755615236, 0.044260353088378904, 0.046604286193847655, 0.046878719329833986, 0.04556902313232422, 0.045192192077636716, 0.0452229118347168, 0.045279232025146485, 0.04519424057006836, 0.04516864013671875, 0.04509798431396484, 0.045369342803955076, 0.0450334701538086, 0.04421836853027344, 0.04367257690429688, 0.04433203125, 0.04518195343017578, 0.04521574401855469, 0.04520652770996094, 0.04537241744995117, 0.0452229118347168, 0.045276161193847655, 0.0450437126159668, 0.04531302261352539, 0.04500787353515625, 0.04507033538818359, 0.04519424057006836, 0.045254657745361325, 0.04499353790283203, 0.04528844833374023, 0.044747776031494144, 0.04507648086547852, 0.045156352996826174, 0.044183551788330076, 0.09082777404785156, 0.0450621452331543, 0.045308929443359375, 0.045434879302978515, 0.04517375946044922, 0.0451328010559082, 0.04527308654785156, 0.045042686462402344, 0.04517068862915039, 0.04490956878662109, 0.0451962890625, 0.04515327835083008, 0.045297664642333986, 0.045254657745361325, 0.04452454376220703, 0.04513177490234375, 0.045290496826171874, 0.04532940673828125, 0.04523417663574219, 0.04528844833374023, 0.04501401519775391, 0.045312000274658204, 0.044951553344726565, 0.04558950424194336, 0.0451512336730957, 0.04399718475341797, 0.04349542236328125, 0.04442521667480469, 0.045143039703369144, 0.045352958679199216, 0.045178878784179685, 0.04510003280639648, 0.04530688095092773, 0.04565401458740234, 0.04506828689575195, 0.045439998626708986, 0.044837886810302735, 0.04470783996582031, 0.04506419372558594, 0.04528947067260742, 0.045042686462402344, 0.04521779251098633, 0.04625100708007813, 0.04592127990722656, 0.04533760070800781, 0.0456181755065918, 0.045358081817626954, 0.04549427032470703, 0.045290496826171874, 0.04539699172973633, 0.0452229118347168, 0.04547788619995117, 0.0452229118347168, 0.045241344451904295, 0.0459048957824707, 0.04576563262939453, 0.04546867370605469, 0.04531609725952149, 0.045254657745361325, 0.04540825653076172, 0.04514713668823242, 0.045437950134277344, 0.045328384399414064, 0.08992870330810547, 0.04531814575195312, 0.04516659164428711, 0.045379585266113284, 0.04734668731689453, 0.04551475143432617, 0.0451143684387207, 0.04497612762451172, 0.04514713668823242, 0.04520652770996094, 0.04469964981079102, 0.04357734298706055, 0.043924480438232424, 0.04542668914794922, 0.04524544143676758, 0.045948928833007815, 0.04659302520751953, 0.04534988784790039, 0.04527308654785156, 0.04506521606445312, 0.04558950424194336, 0.04588748931884765, 0.045328384399414064, 0.04538880157470703, 0.04365824127197265, 0.04369100952148437, 0.043635711669921876, 0.043466751098632815, 0.04356915283203125, 0.04354457473754883, 0.0450334701538086, 0.04534476852416992, 0.044988414764404294, 0.04504678344726563, 0.043670528411865236, 0.043428863525390625, 0.04340838241577148, 0.043568126678466795, 0.04484096145629883, 0.045059070587158204, 0.045192192077636716, 0.045009918212890625, 0.044933120727539064, 0.04521984100341797, 0.04479897689819336, 0.04347391891479492, 0.04362444686889649, 0.04348108673095703, 0.0435865592956543, 0.04369715118408203, 0.04343606567382813, 0.04346569442749024, 0.0435968017578125, 0.044453887939453124, 0.04364287948608398, 0.04352102279663086, 0.04346879959106445, 0.04338790512084961, 0.043548671722412106, 0.04335615921020508, 0.04338483047485352, 0.043425792694091796, 0.04342476654052734, 0.08736153411865234, 0.04329369735717774, 0.04355891036987305, 0.04335308837890625, 0.043437057495117185, 0.043379711151123046, 0.04334284973144531, 0.04359270477294922, 0.04504678344726563, 0.04501708984375, 0.0452782096862793, 0.04522598266601562, 0.04501913452148437, 0.04498636627197266, 0.04552294540405273, 0.0451962890625, 0.04374425506591797, 0.043600894927978515, 0.043483135223388675, 0.043622398376464845, 0.045437950134277344, 0.04535603332519531, 0.04524544143676758, 0.04540415954589844, 0.04534886550903321, 0.044862464904785154, 0.04562124633789062, 0.045243392944335936, 0.043433982849121096, 0.0446453742980957, 0.04436787033081055, 0.04350054550170898, 0.04353433609008789, 0.04348211288452149, 0.043478015899658204, 0.04355481719970703, 0.04326502227783203, 0.043483135223388675, 0.043344894409179685, 0.04354457473754883, 0.04340326309204102, 0.04332339096069336, 0.04313600158691406, 0.04352716827392578, 0.04295475387573242, 0.04282470321655273, 0.042797054290771484, 0.043466751098632815, 0.043409408569335936, 0.04335001754760742, 0.04273971176147461, 0.04292812728881836, 0.04281753540039063, 0.04318105697631836, 0.043312126159667966, 0.043453441619873044, 0.04338995361328125, 0.04338380813598633, 0.043431934356689454, 0.04356095886230469, 0.04343500900268555, 0.04333772659301758, 0.043079681396484375, 0.08883814239501953, 0.04634009552001953, 0.04591308975219727, 0.045687808990478515, 0.04552908706665039, 0.04483174514770508, 0.04478771209716797, 0.044685310363769534, 0.04570521545410156, 0.04539084625244141, 0.04580966567993164, 0.04576563262939453, 0.04599193572998047, 0.04572159957885742, 0.04599398422241211, 0.045927425384521485, 0.04571750259399414, 0.0460052490234375, 0.046080001831054686, 0.045674495697021485, 0.045886463165283206, 0.045832191467285156, 0.04559564971923828, 0.0457891845703125, 0.04580352020263672, 0.04566835021972656, 0.045620223999023435, 0.04576563262939453, 0.04555980682373047, 0.045244415283203124, 0.045591552734375, 0.046069759368896485, 0.04571955108642578, 0.04584758377075195, 0.04578096008300781, 0.045709312438964846, 0.045644798278808595, 0.04558335876464844, 0.04583321762084961, 0.04563455963134765, 0.04575539016723633, 0.04590796661376953, 0.04571340942382812, 0.045742080688476565, 0.04559769439697266, 0.04568473434448242, 0.04496895980834961, 0.04558233642578125, 0.045693950653076174, 0.04574105453491211, 0.04589056015014648, 0.045770751953125, 0.046047233581542966, 0.045753345489501954, 0.045590526580810545, 0.045486080169677735, 0.04593561553955078, 0.0459417610168457, 0.045770751953125, 0.045835262298583986, 0.04582912063598633, 0.04597043228149414, 0.04582092666625977, 0.0928174057006836, 0.0458158073425293, 0.04572467041015625, 0.045830142974853515, 0.04579020690917969, 0.04571443176269531, 0.04543385696411133, 0.04542464065551758, 0.0458260498046875, 0.0464087028503418, 0.046339073181152345, 0.045641727447509765, 0.04585881423950195, 0.04595814514160156, 0.04600012969970703, 0.046080001831054686, 0.04580147171020508, 0.04584755325317383, 0.04584447860717773, 0.045699073791503904, 0.04588748931884765, 0.0456888313293457, 0.0465530891418457, 0.045725696563720705, 0.04565913772583008, 0.04573593521118164, 0.04579020690917969, 0.04585779190063476, 0.04585062408447266, 0.045638656616210936, 0.04574924850463867, 0.04630220794677734, 0.04582195281982422, 0.045736961364746094, 0.04595302581787109, 0.04646809768676758, 0.04595609664916992, 0.04573593521118164, 0.04565708923339844, 0.04538265609741211, 0.045385726928710936, 0.04591308975219727, 0.04603801727294922, 0.04588851165771484, 0.04543078231811523, 0.045467647552490234, 0.04634726333618164, 0.046064640045166014, 0.045712383270263675, 0.0456888313293457, 0.04573491287231445, 0.045663230895996096, 0.04560076904296875, 0.04567859268188477, 0.04550656127929688, 0.04568576049804687, 0.045851646423339845, 0.04569702529907226, 0.04691763305664062, 0.04611174392700195, 0.045590526580810545, 0.045434879302978515, 0.04539084625244141, 0.09221836853027343, 0.045976577758789064, 0.04601651382446289, 0.046132225036621094, 0.04583935928344727, 0.04540825653076172, 0.04572467041015625, 0.04590796661376953, 0.04581171035766601, 0.04589875030517578, 0.04597043228149414, 0.046914558410644534, 0.046176254272460936, 0.04570521545410156, 0.045871105194091794, 0.04604313659667969, 0.04611379241943359, 0.0459950065612793, 0.04579942321777344, 0.04605440139770508, 0.04578713607788086, 0.04581171035766601, 0.0458076171875, 0.04614144134521484, 0.045982719421386715, 0.04577177429199219, 0.04592025756835937, 0.04589567947387695, 0.045770751953125, 0.04581478500366211, 0.045897727966308595, 0.04571033477783203, 0.045298686981201174, 0.045843456268310545, 0.04603084945678711, 0.04576051330566406, 0.04577283096313477, 0.045750240325927734, 0.04595199966430664, 0.04575027084350586, 0.046806015014648435, 0.04596121597290039, 0.04588748931884765, 0.04559872055053711, 0.04534175872802734, 0.04549523162841797, 0.04582912063598633, 0.04582912063598633, 0.04583321762084961, 0.04575436782836914, 0.045835262298583986, 0.04631961441040039, 0.04589875030517578, 0.04560179138183594, 0.045870079040527346, 0.0457625617980957, 0.04565401458740234, 0.04590387344360351, 0.04660326385498047, 0.046034942626953124, 0.04577280044555664, 0.04576665496826172, 0.04574003219604492, 0.09243750762939452, 0.04573183822631836, 0.04562739181518555, 0.046186496734619144, 0.045868030548095705, 0.04568985748291016, 0.045889537811279295, 0.045281280517578126, 0.04573491287231445, 0.0455997428894043, 0.045884414672851564, 0.04579328155517578, 0.04552499389648437, 0.045868030548095705, 0.046015487670898435, 0.04595711898803711, 0.046339073181152345, 0.045818878173828126, 0.0457625617980957, 0.04579020690917969, 0.04585574340820313, 0.045213695526123046, 0.04567552185058594, 0.045810688018798826, 0.04566016006469727, 0.04592127990722656, 0.045638656616210936, 0.0466165771484375, 0.046183425903320315, 0.04563455963134765, 0.04580044937133789, 0.04568473434448242, 0.045744129180908207, 0.04573798370361328, 0.04571648025512695, 0.04611993789672852, 0.04575436782836914, 0.045655040740966796, 0.045628414154052735, 0.04565094375610351, 0.04580966567993164, 0.045748222351074216, 0.04578815841674805, 0.04575743865966797, 0.04519116973876953, 0.045948928833007815, 0.04573491287231445, 0.04561510467529297, 0.045687808990478515, 0.0458260498046875, 0.04564377593994141, 0.045946880340576174, 0.04564275360107422, 0.04571955108642578, 0.04561407852172852, 0.04564582443237305, 0.04579020690917969, 0.045744129180908207, 0.04556288146972656, 0.04597862243652344, 0.04569807815551758, 0.04569392013549805, 0.0456888313293457, 0.09245184326171875, 0.04586086273193359, 0.045725696563720705, 0.04577382278442383, 0.04579225540161133, 0.045818878173828126, 0.045638656616210936, 0.04554444885253906, 0.046058494567871096, 0.04570320129394531, 0.04595503997802734, 0.04560076904296875, 0.045592575073242186, 0.04572774505615235, 0.04578406524658203, 0.045841407775878903, 0.045851646423339845, 0.04590591812133789, 0.04557209777832031, 0.04575948715209961, 0.04550041580200195, 0.045720577239990234, 0.045856769561767576, 0.045612064361572266, 0.045733856201171874, 0.04566527938842774, 0.045638656616210936, 0.04602470397949219, 0.04580147171020508, 0.04579942321777344, 0.045661182403564454, 0.04551270294189453, 0.04584960174560547, 0.04568678283691406, 0.045644798278808595, 0.04698419189453125, 0.04665753555297852, 0.04563558578491211, 0.04597452926635742, 0.04597555160522461, 0.0457625617980957, 0.045736961364746094, 0.04576153564453125, 0.04577382278442383, 0.04587724685668945, 0.046061569213867185, 0.0459048957824707, 0.045758464813232425, 0.04592844772338867, 0.04589875030517578, 0.04586700820922852, 0.04588544082641602, 0.04613119888305664, 0.045889537811279295, 0.0456519660949707, 0.046102527618408204, 0.045827072143554685, 0.04580044937133789, 0.04578303909301758, 0.04575743865966797, 0.04575436782836914, 0.04682342529296875, 0.04640153503417969, 0.09238937377929687, 0.045795326232910154, 0.045484031677246094, 0.04558950424194336, 0.04579942321777344, 0.045777919769287106, 0.0458342399597168, 0.04578406524658203, 0.046647296905517575, 0.046922752380371094, 0.04616089630126953, 0.045710369110107424, 0.04566934585571289, 0.04571852874755859, 0.04588851165771484, 0.04604108810424805, 0.04571852874755859, 0.04572774505615235, 0.04561715316772461, 0.04571340942382812, 0.04581478500366211, 0.04596736145019531, 0.045736961364746094, 0.045780990600585936, 0.04554751968383789, 0.04597964859008789, 0.046071807861328126, 0.04575027084350586, 0.04564582443237305, 0.045774848937988284, 0.045830142974853515, 0.046018558502197264, 0.04574515151977539, 0.04558540725708008, 0.04558438491821289, 0.04561305618286133, 0.045748222351074216, 0.045864959716796876, 0.045682689666748044, 0.04570009613037109, 0.045948928833007815, 0.04582912063598633, 0.04562432098388672, 0.04510515213012695, 0.045298686981201174, 0.0455997428894043, 0.04581478500366211, 0.04572159957885742, 0.04562739181518555, 0.04561612701416016, 0.04584550476074219, 0.04610047912597656, 0.045864959716796876, 0.04583833694458008, 0.0455813102722168, 0.045706241607666016, 0.04583833694458008, 0.04589977645874024, 0.04574617767333984, 0.04578508758544922, 0.04575948715209961, 0.04603289413452148, 0.04592844772338867]",tokens/s,21.71918396499936,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/gemma-2b,google/gemma-2b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status + response.raise_for_status() + File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status + raise HTTPError(http_error_msg, response=self) +requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/google/gemma-2b/resolve/main/config.json + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error + metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata + r = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper + response = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper + hf_raise_for_status(response) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status + raise HfHubHTTPError(message, response=response) from e +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3f82-006a6b1932c13f3e1b549f37;767b9f29-47fa-4160-a3b9-602bd042a1fb) + +403 Forbidden: Authorization error.. +Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. +If you are trying to create or update content,make sure you have a token with the `write` role. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 399, in cached_file + resolved_file = hf_hub_download( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download + return _hf_hub_download_to_cache_dir( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir + _raise_on_head_call_error(head_call_error, force_download, local_files_only) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error + raise LocalEntryNotFoundError( +huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ + super().__init__(config) + File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ + self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) + File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config + return AutoConfig.from_pretrained(model, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict + resolved_config_file = cached_file( + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 442, in cached_file + raise EnvironmentError( +OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like google/gemma-2b is not the path to a directory containing a file named config.json. +Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,-,-,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6834,7 +7353,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12a1-5522c1ef5a4bc7b4711c1597;00af01c9-67d3-410b-a9de-c5591d8e9fb2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44da-08cc8c8326621e8102725929;aa9294bc-c272-4f49-9019-6dfc5ba12b48) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6907,7 +7426,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664bef02-40db68ec688b07b13a6c403e;0d995708-77e1-4ab7-b718-196a74e2f99a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d3ff6-3b18188113dd22fb50b2ff03;625ce80e-d443-4e0f-928a-a4d5b4bf9de0) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6966,7 +7485,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmplmibqqt7/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpiscdq93l/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -7008,7 +7527,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1422-7ec9e72e5fd77c390a7792f8;66bbec82-20d5-4bad-b390-6fbe2f0240c6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4639-26c9c8cc29cf5fbf6742e3c0;22642277-9e05-4615-93bb-511fda798606) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7038,6 +7557,35 @@ Traceback (most recent call last): OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1469, in _autoset_attn_implementation + cls._check_and_enable_flash_attn_2( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 + raise ValueError( +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpxlwhpvmm/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new + ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5937.815552,7053.246464,0.0,6406.79936,6098.60864,s,10,4.369107391357422,0.4369107391357422,0.0007208936334109947,0.43653796386718746,0.43786113586425784,0.4381150161743164,0.4383181204223633,"[0.438368896484375, 0.43641900634765624, 0.43653814697265625, 0.43647592163085935, 0.43767733764648437, 0.43614486694335936, 0.43780471801757814, 0.43653778076171873, 0.4368502807617187, 0.4362904357910156]",tokens/s,585.9320384442743,kWh,5.155067728913349e-06,2.8244480818023514e-06,2.4793715487132108e-05,3.2773231297847804e-05,tokens/kWh,7811252.960485814,MB,5937.815552,7055.343616,0.0,6408.896512,6157.952,s,10,245.01072265624998,24.501072265624998,0.0034318484925660374,24.5004931640625,24.5048515625,24.5070751953125,24.508854101562502,"[24.50105078125, 24.49864453125, 24.50170703125, 24.49968359375, 24.496791015625, 24.509298828125, 24.50153515625, 24.504357421875, 24.499935546875, 24.49771875]",tokens/s,2.571316035355277,kWh,0.0002892534986599056,0.000158535909402625,0.0013504422735896686,0.0017982316816521993,tokens/kWh,35034.417779869254,,s,629,248.55180075073253,0.3951538962650754,0.05191103267325261,0.38883840942382814,0.3892619018554688,0.38950277099609376,0.8256093530273438,"[0.38878106689453124, 0.38896536254882813, 0.388632568359375, 0.38873394775390624, 0.38894488525390625, 0.3886663818359375, 0.3890759582519531, 0.3891517333984375, 0.3889776611328125, 0.38876263427734375, 0.38862646484375, 0.38881890869140623, 0.3890196533203125, 0.3885875244140625, 0.38878616333007815, 0.3895060424804688, 0.38884146118164065, 0.38903399658203125, 0.389644287109375, 0.38915277099609374, 0.3891998596191406, 0.3889284973144531, 0.38931558227539065, 0.3889561462402344, 0.3889131469726563, 0.38835916137695314, 0.38861004638671875, 0.38877081298828126, 0.388674560546875, 0.3886622619628906, 0.3889428405761719, 0.38882098388671876, 0.3894353942871094, 0.3886909484863281, 0.38898175048828126, 0.38895718383789063, 0.38856500244140624, 0.3886120910644531, 0.389538818359375, 0.38871450805664065, 0.38864486694335937, 0.3886510009765625, 0.3886561279296875, 0.38851071166992185, 0.3891835021972656, 0.38876364135742186, 0.3887032470703125, 0.3888701477050781, 0.38866943359375, 0.3892899780273438, 0.3888373718261719, 0.388600830078125, 0.3887615966796875, 0.3901071472167969, 0.3886766052246094, 0.3887728576660156, 0.3887442016601563, 0.3886929931640625, 0.3890124816894531, 0.3890810852050781, 0.38887628173828126, 0.388568115234375, 0.8254780883789062, 0.3889141845703125, 0.38875750732421877, 0.38877593994140625, 0.3889254455566406, 0.3885977478027344, 0.3893872680664062, 0.38873907470703123, 0.38861721801757815, 0.3886346130371094, 0.38904217529296875, 0.3888578491210937, 0.38904013061523435, 0.38923876953125, 0.38863873291015627, 0.38872164916992186, 0.38861822509765626, 0.3884800109863281, 0.3887462463378906, 0.3886069641113281, 0.3893196716308594, 0.38932070922851564, 0.3890186157226562, 0.3887646789550781, 0.38994842529296875, 0.38882098388671876, 0.3890513916015625, 0.3891650695800781, 0.38888037109375, 0.3886991882324219, 0.3889468688964844, 0.3885434875488281, 0.3888199768066406, 0.3885844421386719, 0.388600830078125, 0.3886417846679687, 0.3889479675292969, 0.3886417846679687, 0.3886336059570312, 0.3887790222167969, 0.38892236328125, 0.38850970458984374, 0.38881793212890625, 0.3887523803710938, 0.3886868591308594, 0.38856500244140624, 0.3889899597167969, 0.38897457885742187, 0.38872164916992186, 0.3891282043457031, 0.38919577026367186, 0.3890595703125, 0.3892817993164063, 0.38935140991210937, 0.3886397399902344, 0.38855474853515626, 0.3887595520019531, 0.3887523803710938, 0.3889899597167969, 0.38894900512695313, 0.38868582153320314, 0.38889370727539063, 0.3889704895019531, 0.8256952514648438, 0.38874215698242187, 0.38912716674804687, 0.3890257873535156, 0.38883328247070315, 0.3886970825195312, 0.3889541015625, 0.38887527465820315, 0.389212158203125, 0.389137451171875, 0.38895819091796874, 0.3886970825195312, 0.3888670654296875, 0.38879641723632813, 0.38865716552734375, 0.3888598937988281, 0.3887564697265625, 0.3886991271972656, 0.38894900512695313, 0.3889837951660156, 0.38897357177734376, 0.3888793640136719, 0.38878106689453124, 0.38885479736328127, 0.3889776611328125, 0.38883840942382814, 0.38856396484375, 0.3887400817871094, 0.38860800170898435, 0.3888199768066406, 0.38868582153320314, 0.3887298583984375, 0.38889166259765623, 0.38933505249023437, 0.3892725830078125, 0.3889756164550781, 0.38903604125976565, 0.38886810302734376, 0.38863665771484374, 0.3886458740234375, 0.3889704895019531, 0.38879129028320314, 0.3888148498535156, 0.3887943725585937, 0.3887175598144531, 0.38925106811523436, 0.3889131469726563, 0.38866534423828125, 0.3887677307128906, 0.38883840942382814, 0.388853759765625, 0.388853759765625, 0.3888025512695312, 0.38896536254882813, 0.38888958740234375, 0.3887513732910156, 0.38878823852539063, 0.38919064331054687, 0.38868582153320314, 0.39011123657226565, 0.3893196716308594, 0.38929306030273436, 0.38961663818359377, 0.8260413208007813, 0.38885992431640626, 0.3889643249511719, 0.3887523803710938, 0.3887247314453125, 0.3887185974121094, 0.38895718383789063, 0.38912408447265623, 0.3887984619140625, 0.3889039306640625, 0.388822021484375, 0.3887298583984375, 0.388811767578125, 0.3887718505859375, 0.38874215698242187, 0.3888404541015625, 0.38900531005859373, 0.38877694702148435, 0.3888035888671875, 0.3886673889160156, 0.38864999389648436, 0.38919064331054687, 0.38883941650390624, 0.38895718383789063, 0.38876263427734375, 0.38873394775390624, 0.3887442016601563, 0.3888988037109375, 0.38905035400390625, 0.38884658813476564, 0.38873294067382813, 0.3889889221191406, 0.3889121398925781, 0.3887595520019531, 0.3887544250488281, 0.38894488525390625, 0.38860800170898435, 0.38863873291015627, 0.3886981201171875, 0.38873294067382813, 0.3888630065917969, 0.3888568115234375, 0.3888783264160156, 0.38862539672851565, 0.3889613037109375, 0.389233642578125, 0.38876263427734375, 0.38913946533203125, 0.3891650695800781, 0.3887237243652344, 0.38868582153320314, 0.38899713134765623, 0.3889704895019531, 0.3898193969726563, 0.3890257873535156, 0.3886673889160156, 0.38894692993164065, 0.3890390930175781, 0.3892991943359375, 0.38878515625, 0.3889059753417969, 0.3887462463378906, 0.3888424987792969, 0.826693603515625, 0.3888025512695312, 0.3887523803710938, 0.3885926513671875, 0.3887923278808594, 0.3886120910644531, 0.38872576904296874, 0.38866021728515626, 0.3889807434082031, 0.38920294189453125, 0.38865304565429687, 0.38867559814453123, 0.3889039306640625, 0.388706298828125, 0.3886264343261719, 0.3890257873535156, 0.3886796875, 0.38881893920898436, 0.388748291015625, 0.3887544250488281, 0.3898101806640625, 0.38882098388671876, 0.3889725341796875, 0.38886605834960936, 0.3887820739746094, 0.38895001220703124, 0.3888015441894531, 0.38883840942382814, 0.3887093811035156, 0.38877593994140625, 0.3885823974609375, 0.3885844421386719, 0.3888087158203125, 0.388632568359375, 0.388895751953125, 0.3889407958984375, 0.3887400817871094, 0.38869403076171877, 0.38845440673828124, 0.388706298828125, 0.3888496704101562, 0.38918142700195313, 0.3888035888671875, 0.38866534423828125, 0.3889039306640625, 0.38894488525390625, 0.3888486328125, 0.38889984130859373, 0.3886970825195312, 0.3889612731933594, 0.3889326171875, 0.38907904052734377, 0.38912612915039063, 0.388790283203125, 0.3889459228515625, 0.38889675903320314, 0.3888025512695312, 0.38870016479492187, 0.38883428955078125, 0.3888517150878906, 0.38923776245117186, 0.38900222778320315, 0.38878616333007815, 0.8254668579101563, 0.389001220703125, 0.3892408447265625, 0.38915377807617185, 0.38880767822265627, 0.38871142578125, 0.38886297607421877, 0.3891148681640625, 0.38876364135742186, 0.38882720947265625, 0.3886867980957031, 0.3885404052734375, 0.38891827392578127, 0.38905035400390625, 0.38921829223632814, 0.3889141845703125, 0.3891558532714844, 0.38889471435546874, 0.3890810852050781, 0.38909234619140626, 0.3887319030761719, 0.38957876586914064, 0.3891712036132812, 0.3887923583984375, 0.3895316162109375, 0.38924493408203126, 0.38886605834960936, 0.38888653564453124, 0.38975283813476563, 0.38893875122070315, 0.38879641723632813, 0.38891009521484377, 0.38910565185546875, 0.3892593078613281, 0.39087921142578125, 0.38871551513671876, 0.38870834350585937, 0.38860186767578125, 0.3886960754394531, 0.38911077880859374, 0.38857318115234374, 0.3887657470703125, 0.3889714660644531, 0.38878106689453124, 0.3889807434082031, 0.3888148498535156, 0.3887872009277344, 0.3890739135742188, 0.38941082763671875, 0.3891589050292969, 0.38924798583984377, 0.3888977966308594, 0.38902792358398436, 0.3890799560546875, 0.3894364013671875, 0.3890677795410156, 0.388959228515625, 0.38916094970703125, 0.3891568603515625, 0.3888988037109375, 0.38902987670898437, 0.38889370727539063, 0.38949786376953127, 0.8270551147460937, 0.38860186767578125, 0.38860595703125, 0.38868377685546873, 0.38870834350585937, 0.38866021728515626, 0.3888844909667969, 0.38888858032226564, 0.3890831298828125, 0.388947998046875, 0.38878512573242185, 0.389180419921875, 0.38898175048828126, 0.38901556396484377, 0.38881381225585937, 0.38938214111328123, 0.38876568603515627, 0.38882098388671876, 0.3885455322265625, 0.3887093811035156, 0.3890534362792969, 0.38878515625, 0.3887400817871094, 0.3889725341796875, 0.3888455810546875, 0.38887628173828126, 0.3890677795410156, 0.38898175048828126, 0.38893771362304685, 0.3888015441894531, 0.38900735473632814, 0.3888015441894531, 0.38892440795898436, 0.3890124816894531, 0.3895828552246094, 0.3893197021484375, 0.3891650390625, 0.3890083923339844, 0.38891827392578127, 0.38878311157226564, 0.3886766052246094, 0.38947021484375, 0.3891251220703125, 0.38870834350585937, 0.3887933349609375, 0.3888517150878906, 0.38911181640625, 0.38871142578125, 0.38868582153320314, 0.3887319030761719, 0.388611083984375, 0.38874215698242187, 0.388890625, 0.3886889038085938, 0.38894692993164065, 0.38870220947265627, 0.3885906066894531, 0.38906878662109373, 0.3886192626953125, 0.38860186767578125, 0.3894384765625, 0.38954290771484373, 0.3893800964355469, 0.825660400390625, 0.38914047241210936, 0.3887677307128906, 0.38940365600585936, 0.38903704833984376, 0.3888568420410156, 0.38899917602539064, 0.3896258544921875, 0.3887247314453125, 0.3890565185546875, 0.3885342712402344, 0.38850048828125, 0.38885888671875, 0.38866021728515626, 0.38881793212890625, 0.3884103698730469, 0.38903604125976565, 0.38858648681640623, 0.3889070129394531, 0.3887185974121094, 0.38888858032226564, 0.3892193298339844, 0.3894497375488281, 0.38906060791015623, 0.3886120910644531, 0.388600830078125, 0.38877490234375, 0.3888005065917969, 0.38874725341796873, 0.38922955322265623, 0.3888926696777344, 0.3885957336425781, 0.38874212646484374, 0.3889346618652344, 0.3888650207519531, 0.38876876831054685, 0.388664306640625, 0.38930841064453126, 0.3888230285644531, 0.38908517456054686, 0.38917633056640627, 0.38882611083984375, 0.3889510498046875, 0.38876058959960935, 0.3885772705078125, 0.3890186157226562, 0.3886346130371094, 0.3894353942871094, 0.389170166015625, 0.3888036193847656, 0.38960125732421874, 0.3888005065917969, 0.3895419006347656, 0.3890739135742188, 0.3897927551269531, 0.38913229370117186, 0.38885888671875, 0.3889837951660156, 0.388969482421875, 0.38935653686523436, 0.38904013061523435, 0.3888988037109375, 0.38913433837890626, 0.8261171264648437, 0.38934527587890627, 0.3890943908691406, 0.3892561950683594, 0.38904730224609374, 0.3889776611328125, 0.38878823852539063, 0.38880459594726563, 0.38874215698242187, 0.38939239501953127, 0.3890657348632813, 0.3885916137695313, 0.38864999389648436, 0.3887585144042969, 0.388790283203125, 0.3887032470703125, 0.3896012878417969, 0.3887442016601563, 0.3886336059570312, 0.38861004638671875, 0.389037109375, 0.3888260498046875, 0.388701171875, 0.38865304565429687, 0.38868069458007815, 0.38857318115234374, 0.3886213073730469, 0.3887431640625, 0.38874725341796873, 0.38863873291015627, 0.388595703125, 0.38875341796875, 0.38857522583007814, 0.3887923278808594, 0.38920806884765624, 0.3891712036132812, 0.38883123779296874, 0.3889407958984375, 0.388748291015625, 0.3887503356933594, 0.38887527465820315, 0.38860287475585936, 0.38913742065429685, 0.3887790222167969, 0.3885629577636719, 0.38873394775390624, 0.38880563354492187, 0.38863565063476563, 0.38857318115234374, 0.3885987854003906, 0.38900531005859373, 0.3888097229003906, 0.38852197265625, 0.3885834350585938, 0.38872576904296874, 0.38862234497070314, 0.3915376586914063, 0.38877694702148435, 0.3889428405761719, 0.38899917602539064, 0.38912716674804687, 0.38925726318359377, 0.3891875610351562, 0.8258375854492187, 0.388811767578125, 0.38908212280273435, 0.3887953796386719, 0.38926849365234373, 0.38856192016601565, 0.38865304565429687, 0.38886297607421877, 0.38894692993164065, 0.38873086547851565, 0.3887933349609375, 0.3884656677246094, 0.38854656982421876, 0.38860287475585936, 0.3886929931640625, 0.3886868591308594, 0.38878411865234375, 0.3888066711425781, 0.3886141357421875, 0.3886141357421875, 0.38892340087890626, 0.38862335205078125, 0.3893319702148437, 0.3891773376464844, 0.38893875122070315, 0.38863565063476563, 0.3889203186035156, 0.3887298583984375, 0.38881689453125, 0.3891302490234375, 0.3888824462890625, 0.3886489562988281, 0.3886243896484375, 0.3889407958984375, 0.38874111938476563, 0.38874111938476563, 0.38891009521484377, 0.3886141357421875, 0.38878823852539063, 0.38871142578125, 0.3888291931152344, 0.38855270385742186, 0.38883328247070315, 0.388885498046875, 0.38914764404296875, 0.38934017944335936, 0.38926025390625, 0.38955938720703126, 0.3897701416015625, 0.3890442199707031, 0.38865716552734375, 0.3886612548828125, 0.3888455810546875, 0.38867559814453123, 0.38898483276367185, 0.3890862121582031, 0.38879129028320314, 0.38848614501953127, 0.3886970825195312, 0.3888097229003906, 0.3888128356933594, 0.3887206115722656, 0.38876980590820315]",tokens/s,2.530659597315939,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -7079,7 +7627,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c10cb-5e8b10537f0ba96908ca15cf;eca885ac-eff0-4e3d-a3ff-345f531d0e77) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d431b-44881c4f3e9c0c9173b94844;a41b278e-8d55-4761-888e-78bbf389231b) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7110,6 +7658,7 @@ OSError: s is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2267.467776,3257.401344,0.0,2610.95424,2376.894976,s,10,1.2864311065673828,0.1286431106567383,0.000171776670734644,0.1285985336303711,0.1287443832397461,0.12892626876831054,0.1290717771911621,"[0.129108154296875, 0.1284718017578125, 0.1286188201904297, 0.12850921630859374, 0.12854039001464843, 0.12869644165039062, 0.12865798950195312, 0.12854608154296876, 0.12870396423339844, 0.1285782470703125]",tokens/s,1990.0016308148154,kWh,1.5196969493841515e-06,8.327151893634156e-07,7.5898600604873695e-06,9.942272199234937e-06,tokens/kWh,25748641.243165653,MB,2267.467776,3257.401344,0.0,2610.95424,2525.750784,s,10,70.94909423828125,7.094909423828125,0.001118594056454437,7.094721191406251,7.0962150878906245,7.096387573242187,7.096525561523437,"[7.094095703125, 7.09424169921875, 7.0961767578125, 7.0933984375, 7.09597705078125, 7.09488916015625, 7.09589208984375, 7.09331005859375, 7.09656005859375, 7.09455322265625]",tokens/s,8.879605959226998,kWh,8.37615760613201e-05,4.5907382613831236e-05,0.00041681028501491135,0.0005464792436900627,tokens/kWh,115283.42700556554,,s,629,71.99387335205085,0.11445766828624929,0.015289816234666156,0.11260825347900391,0.1127077865600586,0.11280117492675781,0.2412650079345703,"[0.11252735900878906, 0.1125580825805664, 0.11255398559570312, 0.11299123382568359, 0.11257651519775391, 0.11259903717041016, 0.11247718048095703, 0.11253759765625, 0.11249868774414062, 0.11256217956542969, 0.11247718048095703, 0.11243622589111328, 0.11248230743408204, 0.11250892639160157, 0.11247718048095703, 0.11255910491943359, 0.11265023803710937, 0.11251712036132812, 0.11261235046386718, 0.11255091094970703, 0.11255500793457031, 0.11261746978759765, 0.11257344055175782, 0.11254988861083984, 0.11260313415527344, 0.1126266860961914, 0.11258675384521484, 0.11254783630371094, 0.11254579162597657, 0.11261235046386718, 0.11260825347900391, 0.11254476928710938, 0.11260108947753907, 0.11259801483154297, 0.11267788696289062, 0.11262258911132812, 0.11281715393066406, 0.11266764831542969, 0.11260518646240235, 0.11264205169677734, 0.11268505859375, 0.11259699249267578, 0.11259187316894531, 0.11263999938964844, 0.11251302337646485, 0.11264614105224609, 0.11258573150634765, 0.11261746978759765, 0.11255398559570312, 0.1126297607421875, 0.11265229034423828, 0.11258879852294922, 0.11266355133056641, 0.11270758056640626, 0.11261746978759765, 0.11268608093261719, 0.11265740966796875, 0.11258060455322266, 0.11267378997802735, 0.11264102172851563, 0.11265126037597656, 0.11266867065429688, 0.24122265625, 0.11249766540527344, 0.11248230743408204, 0.1125406723022461, 0.112500732421875, 0.11253247833251953, 0.11251404571533204, 0.11253964996337891, 0.11250790405273438, 0.11243622589111328, 0.11254988861083984, 0.11251200103759766, 0.11281100463867187, 0.1125560302734375, 0.11253453063964844, 0.11255910491943359, 0.1125928955078125, 0.11246489715576172, 0.11262566375732422, 0.11268300628662109, 0.11256114959716797, 0.11256422424316406, 0.11256217956542969, 0.1125959701538086, 0.11283660888671875, 0.1125580825805664, 0.11257344055175782, 0.11258470153808593, 0.11263385772705078, 0.11261952209472656, 0.11261952209472656, 0.11258470153808593, 0.1125959701538086, 0.11259085083007812, 0.1126297607421875, 0.11261952209472656, 0.11255091094970703, 0.11260211181640625, 0.11261849975585937, 0.11261030578613282, 0.11265535736083984, 0.11266047668457031, 0.11263897705078126, 0.11258675384521484, 0.11261746978759765, 0.11266969299316407, 0.11258879852294922, 0.11264205169677734, 0.11262361907958984, 0.1125447998046875, 0.11264611053466797, 0.11268505859375, 0.11258470153808593, 0.11272089385986328, 0.11261849975585937, 0.11272499084472656, 0.11271782684326172, 0.11262156677246093, 0.11267276763916016, 0.1126277084350586, 0.11259699249267578, 0.11268096160888671, 0.11284070587158203, 0.24138444519042968, 0.11254589080810547, 0.11252623748779297, 0.11253145599365234, 0.11248127746582032, 0.11257344055175782, 0.11256832122802735, 0.11259187316894531, 0.11256114959716797, 0.1125898208618164, 0.11260006713867188, 0.11251712036132812, 0.11250688171386719, 0.11261030578613282, 0.11247513580322266, 0.11253350067138672, 0.11260620880126954, 0.11252633666992187, 0.11263078308105469, 0.11266969299316407, 0.11256114959716797, 0.11270861053466796, 0.11259391784667969, 0.11262258911132812, 0.112611328125, 0.11260415649414063, 0.11267686462402343, 0.11270963287353515, 0.11256012725830078, 0.11260928344726563, 0.11258470153808593, 0.11263180541992188, 0.11264614105224609, 0.11259494018554687, 0.11286732482910156, 0.11270655822753907, 0.11261542510986328, 0.11264205169677734, 0.11264307403564452, 0.11261440277099609, 0.11263692474365235, 0.11256320190429688, 0.11264307403564452, 0.11306905364990234, 0.11276083374023438, 0.11260006713867188, 0.11259699249267578, 0.11268710327148437, 0.11263897705078126, 0.11287859344482422, 0.11272806549072266, 0.11265945434570312, 0.11271270751953125, 0.11277926635742187, 0.11267378997802735, 0.11267174530029297, 0.11268096160888671, 0.11272089385986328, 0.11270963287353515, 0.11264409637451171, 0.1126266860961914, 0.11269017791748047, 0.11263590240478516, 0.24126873779296876, 0.11247615814208985, 0.112468994140625, 0.11252941131591797, 0.11247821044921875, 0.11256422424316406, 0.11249868774414062, 0.11254169464111329, 0.11254886627197265, 0.11241574096679688, 0.11257548522949219, 0.11246694183349609, 0.11248639678955077, 0.11256217956542969, 0.11249152374267578, 0.11253657531738281, 0.11254988861083984, 0.11252531433105468, 0.1126666259765625, 0.11258163452148437, 0.11253759765625, 0.11260313415527344, 0.11258060455322266, 0.11269017791748047, 0.1125898208618164, 0.11260825347900391, 0.112648193359375, 0.1126645736694336, 0.112648193359375, 0.11261337280273437, 0.11261746978759765, 0.11259699249267578, 0.11258470153808593, 0.11257855987548829, 0.11269529724121094, 0.11260620880126954, 0.1125580825805664, 0.11263180541992188, 0.1125181427001953, 0.11259801483154297, 0.11262464141845703, 0.11252019500732421, 0.11263488006591797, 0.11266560363769532, 0.11260825347900391, 0.1125775375366211, 0.11258879852294922, 0.11268300628662109, 0.11261440277099609, 0.11253968048095703, 0.11272086334228516, 0.11265535736083984, 0.11264921569824218, 0.11271372985839843, 0.11262566375732422, 0.11268096160888671, 0.11267378997802735, 0.11260723114013672, 0.11266764831542969, 0.11265126037597656, 0.11261440277099609, 0.112611328125, 0.11260108947753907, 0.24140492248535156, 0.11250482940673828, 0.11252531433105468, 0.11256217956542969, 0.11257344055175782, 0.11262464141845703, 0.11257241821289063, 0.11260723114013672, 0.11253350067138672, 0.11247923278808594, 0.11268608093261719, 0.11254681396484376, 0.11251302337646485, 0.11256422424316406, 0.1125437469482422, 0.11257241821289063, 0.11258675384521484, 0.11254169464111329, 0.11258777618408203, 0.11263692474365235, 0.11254169464111329, 0.11267276763916016, 0.11256320190429688, 0.11268300628662109, 0.1125928955078125, 0.11256012725830078, 0.11266252899169922, 0.11268710327148437, 0.1125181427001953, 0.11261644744873046, 0.11263385772705078, 0.11259699249267578, 0.11262361907958984, 0.1126277084350586, 0.1126666259765625, 0.11310489654541016, 0.11263590240478516, 0.11265535736083984, 0.11256729888916016, 0.11266047668457031, 0.11264717102050781, 0.11253043365478516, 0.11262258911132812, 0.11270041656494141, 0.11257855987548829, 0.1126297607421875, 0.11261440277099609, 0.113080322265625, 0.11273932647705077, 0.11254476928710938, 0.1126983642578125, 0.11264921569824218, 0.11271270751953125, 0.11278643035888672, 0.11260825347900391, 0.11276799774169922, 0.11273011016845703, 0.11268710327148437, 0.1126277084350586, 0.11264921569824218, 0.11263078308105469, 0.11264205169677734, 0.11266047668457031, 0.24146739196777345, 0.11256626892089844, 0.11253043365478516, 0.11249049377441406, 0.11251404571533204, 0.11254476928710938, 0.11256934356689453, 0.11252941131591797, 0.11252326202392578, 0.1124874267578125, 0.11258265686035156, 0.1124874267578125, 0.11245362854003907, 0.1125928955078125, 0.11243417358398437, 0.11250994873046875, 0.11249561309814453, 0.11249049377441406, 0.11260006713867188, 0.11265740966796875, 0.1125406723022461, 0.11257651519775391, 0.11254271697998047, 0.112611328125, 0.11253862762451172, 0.11254783630371094, 0.11257138824462891, 0.11258879852294922, 0.1129543685913086, 0.11272499084472656, 0.11296768188476562, 0.11258470153808593, 0.1126277084350586, 0.11254476928710938, 0.1130465316772461, 0.11263385772705078, 0.11253862762451172, 0.11317555236816407, 0.11258163452148437, 0.11261542510986328, 0.11265945434570312, 0.11256012725830078, 0.11270758056640626, 0.11262054443359375, 0.1125212173461914, 0.11261644744873046, 0.11257138824462891, 0.11265433502197265, 0.11263692474365235, 0.11256114959716797, 0.11267788696289062, 0.11265229034423828, 0.11258163452148437, 0.11269222259521484, 0.11258060455322266, 0.11271782684326172, 0.11271987152099609, 0.11268402862548828, 0.11265433502197265, 0.11262361907958984, 0.11256114959716797, 0.11266867065429688, 0.11265229034423828, 0.24139161682128907, 0.11250176239013672, 0.11253555297851563, 0.11268096160888671, 0.1125038070678711, 0.11258060455322266, 0.112505859375, 0.11255910491943359, 0.11272908782958985, 0.11249356842041015, 0.11262156677246093, 0.11254988861083984, 0.1125181427001953, 0.11256524658203125, 0.11253350067138672, 0.11256320190429688, 0.11259187316894531, 0.11249868774414062, 0.11261746978759765, 0.1126666259765625, 0.1125928955078125, 0.11259903717041016, 0.1124874267578125, 0.11262156677246093, 0.1125191650390625, 0.11256422424316406, 0.11262566375732422, 0.1126266860961914, 0.11251200103759766, 0.11266150665283203, 0.11256320190429688, 0.11260620880126954, 0.11262873840332031, 0.11253453063964844, 0.11270246124267579, 0.11262873840332031, 0.11258060455322266, 0.11270963287353515, 0.11267584228515624, 0.11267174530029297, 0.11267686462402343, 0.11259085083007812, 0.11262566375732422, 0.1126277084350586, 0.11263590240478516, 0.1127014389038086, 0.11265843200683594, 0.11269939422607422, 0.11266969299316407, 0.11256626892089844, 0.11269324493408203, 0.11265229034423828, 0.11263488006591797, 0.11290217590332031, 0.112838623046875, 0.11316223907470703, 0.11273318481445313, 0.11269939422607422, 0.11269324493408203, 0.11275981140136719, 0.11265023803710937, 0.11267378997802735, 0.11267584228515624, 0.24135270690917968, 0.11246489715576172, 0.11244646453857422, 0.11252326202392578, 0.11248844909667968, 0.1125580825805664, 0.11250482940673828, 0.11253247833251953, 0.11251097869873047, 0.11246797180175781, 0.11258879852294922, 0.11255705261230468, 0.11249971008300781, 0.11258060455322266, 0.11251609802246093, 0.11262464141845703, 0.11256422424316406, 0.11249152374267578, 0.11260518646240235, 0.11258367919921874, 0.11249356842041015, 0.11263488006591797, 0.11252019500732421, 0.11262156677246093, 0.11256422424316406, 0.11264409637451171, 0.11259903717041016, 0.11263999938964844, 0.11257241821289063, 0.11264717102050781, 0.1125560302734375, 0.11253657531738281, 0.11259085083007812, 0.11260006713867188, 0.11264614105224609, 0.11261235046386718, 0.11258777618408203, 0.11262156677246093, 0.11253555297851563, 0.11268096160888671, 0.11268812561035156, 0.11258879852294922, 0.11261542510986328, 0.11256729888916016, 0.11261542510986328, 0.11260108947753907, 0.11261746978759765, 0.11259391784667969, 0.1126328353881836, 0.11251200103759766, 0.11265229034423828, 0.11268402862548828, 0.11262258911132812, 0.1126819839477539, 0.11268608093261719, 0.11268096160888671, 0.11264717102050781, 0.1126277084350586, 0.11265945434570312, 0.11261849975585937, 0.11265229034423828, 0.11263999938964844, 0.11266560363769532, 0.24148684692382813, 0.11249254608154297, 0.1125437469482422, 0.11259187316894531, 0.11256422424316406, 0.11260108947753907, 0.1124853744506836, 0.11255910491943359, 0.11254579162597657, 0.11250688171386719, 0.11258060455322266, 0.11249152374267578, 0.11248332977294923, 0.11256934356689453, 0.11260928344726563, 0.11278233337402344, 0.11263180541992188, 0.11256524658203125, 0.11260825347900391, 0.11268505859375, 0.1125928955078125, 0.11265023803710937, 0.11258163452148437, 0.11292774200439454, 0.11261952209472656, 0.11260415649414063, 0.11264717102050781, 0.11263078308105469, 0.11259699249267578, 0.11263385772705078, 0.11253759765625, 0.11260825347900391, 0.11264307403564452, 0.11258675384521484, 0.11266252899169922, 0.11268505859375, 0.11253657531738281, 0.11328409576416015, 0.11260108947753907, 0.11265638732910156, 0.11269222259521484, 0.11260723114013672, 0.11266867065429688, 0.11265229034423828, 0.11262258911132812, 0.11260415649414063, 0.11260211181640625, 0.11268402862548828, 0.11266252899169922, 0.11253964996337891, 0.11268402862548828, 0.1127045135498047, 0.11264614105224609, 0.11278540802001953, 0.11263078308105469, 0.11262361907958984, 0.1126666259765625, 0.11310182189941406, 0.1127710723876953, 0.11265740966796875, 0.11270246124267579, 0.1127034912109375, 0.1126451187133789, 0.2412554168701172, 0.11251200103759766, 0.11250994873046875, 0.11250176239013672, 0.11246080017089843, 0.11257855987548829, 0.11251302337646485, 0.11257548522949219, 0.11251097869873047, 0.11244134521484375, 0.11257651519775391, 0.11249868774414062, 0.11251097869873047, 0.11252838134765625, 0.1125038070678711, 0.11257344055175782, 0.11258675384521484, 0.1125181427001953, 0.11258367919921874, 0.11262566375732422, 0.11253453063964844, 0.11256422424316406, 0.11253350067138672, 0.11267378997802735, 0.11257036590576172, 0.11257548522949219, 0.11260415649414063, 0.11263590240478516, 0.11255193328857421, 0.11262156677246093, 0.11258265686035156, 0.11260825347900391, 0.11259801483154297, 0.11260313415527344, 0.11262156677246093, 0.11263078308105469, 0.11259391784667969, 0.1126266860961914, 0.11296358489990234, 0.11260415649414063, 0.11274854278564453, 0.11260928344726563, 0.11267481231689454, 0.11265023803710937, 0.11275775909423828, 0.11261542510986328, 0.11258367919921874, 0.112716796875, 0.11261952209472656, 0.11258060455322266, 0.1126983642578125, 0.11266355133056641, 0.11259187316894531, 0.11265638732910156, 0.11257651519775391, 0.11263590240478516, 0.1126328353881836, 0.11285196685791016, 0.1129891815185547, 0.11267481231689454, 0.11260928344726563, 0.1127045135498047, 0.1126666259765625]",tokens/s,8.736854550444638,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -7136,7 +7685,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpqgf4stu5/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp6cy67taw/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1269.682176,1019.74016,0.0,373.293056,277.410816,s,10,0.37923129653930665,0.037923129653930666,0.0012152886645715089,0.038048974990844725,0.038519168853759764,0.0396308479309082,0.04052019119262695,"[0.04074252700805664, 0.03821686553955078, 0.03797974395751953, 0.03823471832275391, 0.03626841735839844, 0.03608185577392578, 0.03733708953857422, 0.03811382293701172, 0.03827212905883789, 0.03798412704467773]",tokens/s,6750.497713035296,kWh,4.301884120160883e-07,2.3572264180256752e-07,1.164744366137984e-06,1.83065541995664e-06,tokens/kWh,139840626.0453229,MB,1269.682176,1019.74016,0.0,373.293056,323.047424,s,10,23.493592773437502,2.3493592773437504,0.014632907695307939,2.3561588134765623,2.362567529296875,2.3635712402343754,2.3643742089843753,"[2.34352734375, 2.36090478515625, 2.362344482421875, 2.364574951171875, 2.338309326171875, 2.331607421875, 2.357506103515625, 2.3608046875, 2.3548115234375, 2.3192021484375]",tokens/s,26.81582191687153,kWh,2.642717914328431e-05,1.4482863030778391e-05,6.568572578086276e-05,0.00010659576795492543,tokens/kWh,591017.8350292469,,s,629,23.794291763305676,0.037828762739754634,0.004511848848474209,0.03739033508300781,0.03767849044799805,0.03791810684204102,0.07427948760986337,"[0.03534745788574219, 0.03654348754882813, 0.03639910507202149, 0.03678003311157227, 0.03678003311157227, 0.036736000061035154, 0.036125694274902344, 0.03654348754882813, 0.036519935607910156, 0.0366110725402832, 0.03650764846801758, 0.03670425415039062, 0.03646771240234375, 0.037282817840576174, 0.03782860946655273, 0.03862527847290039, 0.037749759674072264, 0.03757056045532227, 0.037465087890625, 0.0372408332824707, 0.03734630584716797, 0.03750092697143555, 0.03727872085571289, 0.03741388702392578, 0.03740262222290039, 0.03726131057739258, 0.037303295135498044, 0.03731353759765625, 0.037188640594482424, 0.037268447875976565, 0.03742617416381836, 0.03730124664306641, 0.03739136123657227, 0.037250049591064455, 0.03735039901733399, 0.03733913421630859, 0.03732787322998047, 0.03716505432128906, 0.03726233673095703, 0.0372490234375, 0.03732070541381836, 0.03729305648803711, 0.03743129730224609, 0.03723980712890625, 0.037269504547119144, 0.03735039901733399, 0.03727769470214844, 0.03724800109863281, 0.03734124755859375, 0.037339073181152344, 0.037269504547119144, 0.03736678314208984, 0.0371701774597168, 0.037266433715820314, 0.03737190246582031, 0.037318656921386716, 0.03726847839355469, 0.03745894241333008, 0.03732787322998047, 0.03729715347290039, 0.037501953125, 0.037501953125, 0.07589580535888672, 0.03728793716430664, 0.037407745361328126, 0.037207038879394534, 0.03731353759765625, 0.03784089660644531, 0.037353473663330077, 0.03739340972900391, 0.0374015998840332, 0.03751731109619141, 0.03856486511230469, 0.037754878997802735, 0.037369857788085936, 0.03734220886230469, 0.03731353759765625, 0.0373125114440918, 0.03733606338500976, 0.037425151824951174, 0.037303295135498044, 0.037362686157226564, 0.038662143707275394, 0.037748737335205076, 0.03751116943359375, 0.03752243041992188, 0.03749785614013672, 0.03739340972900391, 0.03728793716430664, 0.03726540756225586, 0.03750707244873047, 0.03739340972900391, 0.03738828659057617, 0.03735859298706055, 0.037395454406738284, 0.03731558227539063, 0.037868545532226565, 0.037456897735595705, 0.03736883163452148, 0.03723161697387695, 0.03758489608764649, 0.03761151885986328, 0.03742822265625, 0.037384193420410154, 0.03741900634765625, 0.03744976043701172, 0.03728688049316406, 0.03746815872192383, 0.03732275390625, 0.037525505065917966, 0.0375203857421875, 0.03739340972900391, 0.03733913421630859, 0.03781631851196289, 0.03749785614013672, 0.03746815872192383, 0.03744870376586914, 0.037324798583984374, 0.03752345657348633, 0.03728691101074219, 0.0374015998840332, 0.03738828659057617, 0.037454849243164064, 0.037454849243164064, 0.03749273681640625, 0.0757176284790039, 0.03736166381835938, 0.03737702560424805, 0.03760332870483398, 0.03744768142700195, 0.037354496002197264, 0.03724595260620117, 0.037343231201171875, 0.03800678253173828, 0.03777740859985351, 0.03753472137451172, 0.03747430419921875, 0.0374835205078125, 0.037585918426513674, 0.037338111877441404, 0.0374466552734375, 0.03750912094116211, 0.037379070281982424, 0.03765964889526367, 0.03735551834106445, 0.03742310333251953, 0.03738726425170898, 0.03740364837646484, 0.037495807647705076, 0.03738214492797851, 0.037310462951660156, 0.03733401489257813, 0.03747430419921875, 0.03743436813354492, 0.0373831672668457, 0.03754393768310547, 0.037370880126953124, 0.03760025787353516, 0.037601280212402347, 0.037510143280029294, 0.0373043212890625, 0.0373125114440918, 0.03747225570678711, 0.03762688064575195, 0.037282817840576174, 0.037294078826904296, 0.03753267288208008, 0.037607425689697264, 0.037400577545166014, 0.037495807647705076, 0.03745075225830078, 0.038043647766113284, 0.037773311614990236, 0.03741798400878906, 0.03741183853149414, 0.03771289443969727, 0.03767193603515625, 0.03745280075073242, 0.0372674560546875, 0.03750092697143555, 0.038141952514648435, 0.037969921112060545, 0.037384193420410154, 0.03739852905273437, 0.03727977752685547, 0.03753366470336914, 0.037572608947753904, 0.03738726425170898, 0.07624192047119141, 0.03899084854125977, 0.037817344665527344, 0.03795356750488281, 0.03747836685180664, 0.03777536010742188, 0.03754905700683594, 0.03737497711181641, 0.03731763076782227, 0.03754598236083984, 0.0381317138671875, 0.037395454406738284, 0.03748556900024414, 0.03738726425170898, 0.037493759155273435, 0.037318656921386716, 0.03763916778564453, 0.03733299255371094, 0.037407745361328126, 0.03729817581176758, 0.03746918487548828, 0.03743129730224609, 0.03740364837646484, 0.03732275390625, 0.03753267288208008, 0.03765350341796875, 0.03748863983154297, 0.03742617416381836, 0.03767708969116211, 0.03730940628051758, 0.03756748962402344, 0.03845017623901367, 0.03810201644897461, 0.03751321411132812, 0.03726847839355469, 0.03728998565673828, 0.03749273681640625, 0.03731148910522461, 0.03733606338500976, 0.03725107192993164, 0.03753881454467774, 0.03738726425170898, 0.037282817840576174, 0.03728179168701172, 0.0374917106628418, 0.037351425170898435, 0.03746819305419922, 0.037434337615966796, 0.037440513610839846, 0.03745587158203125, 0.03747532653808594, 0.03769651031494141, 0.03751935958862305, 0.037397502899169925, 0.03741491317749023, 0.037466110229492186, 0.03749478530883789, 0.037566463470458986, 0.03733401489257813, 0.03742822265625, 0.03756851196289063, 0.037501953125, 0.03759820938110352, 0.07184793853759766, 0.03510374450683594, 0.03521535873413086, 0.03516928100585937, 0.03511603164672852, 0.03513241577148438, 0.03520000076293945, 0.035148799896240236, 0.03514470291137695, 0.03511500930786133, 0.0367718391418457, 0.03750400161743164, 0.03728998565673828, 0.037482494354248046, 0.0374835205078125, 0.03756032180786133, 0.037303295135498044, 0.037424129486083986, 0.03768729782104492, 0.03739648056030274, 0.03758492660522461, 0.03751830291748047, 0.03748659133911133, 0.03748863983154297, 0.037410816192626956, 0.03736883163452148, 0.03754396820068359, 0.03754390335083008, 0.037591041564941405, 0.03752140808105469, 0.03739136123657227, 0.03755212783813477, 0.03739340972900391, 0.03741491317749023, 0.03769139099121094, 0.037408767700195314, 0.038091777801513675, 0.03745382308959961, 0.0376360969543457, 0.03739136123657227, 0.03732070541381836, 0.03746303939819336, 0.03743334579467773, 0.037364734649658206, 0.037526527404785154, 0.0374015998840332, 0.037416961669921874, 0.037416961669921874, 0.03750604629516602, 0.037427200317382815, 0.037367809295654295, 0.037498878479003905, 0.037397502899169925, 0.03749683380126953, 0.037622814178466794, 0.03773948669433594, 0.037416961669921874, 0.03744464111328125, 0.03751728057861328, 0.0374835205078125, 0.03757056045532227, 0.037477375030517575, 0.037720062255859374, 0.0716236801147461, 0.035062782287597655, 0.03518668746948242, 0.035019775390625, 0.03531980895996094, 0.035253246307373046, 0.03521843338012695, 0.034840576171875, 0.03494911956787109, 0.03526863861083984, 0.03520406341552734, 0.035253246307373046, 0.03508428955078125, 0.0372490234375, 0.037318656921386716, 0.03726335906982422, 0.03732787322998047, 0.03786342239379883, 0.037288959503173826, 0.0374835205078125, 0.03732992172241211, 0.03781631851196289, 0.038765567779541016, 0.03768012619018555, 0.03770982360839844, 0.037367809295654295, 0.03745894241333008, 0.03721318435668945, 0.03721420669555664, 0.03736678314208984, 0.03769343948364258, 0.03735756683349609, 0.03937484741210937, 0.03793203353881836, 0.037556224822998044, 0.03736883163452148, 0.03729510498046875, 0.03745177459716797, 0.03745792007446289, 0.037561344146728515, 0.037422080993652344, 0.037353473663330077, 0.037594112396240234, 0.03732070541381836, 0.03734732818603516, 0.037340160369873046, 0.037397502899169925, 0.037362686157226564, 0.03757875061035156, 0.03736064147949219, 0.037321727752685545, 0.03742822265625, 0.03740364837646484, 0.037318656921386716, 0.03724492645263672, 0.03730022430419922, 0.03753472137451172, 0.0374200325012207, 0.03728486251831055, 0.03722751998901367, 0.03739136123657227, 0.037362686157226564, 0.037375999450683595, 0.07522509002685547, 0.03725209426879883, 0.03725823974609375, 0.0374466552734375, 0.03733916854858398, 0.03731145477294922, 0.03730124664306641, 0.03746406555175781, 0.037375999450683595, 0.03750707244873047, 0.037292030334472655, 0.037459968566894535, 0.03742924880981445, 0.037405696868896485, 0.03738009643554688, 0.0372592658996582, 0.03754905700683594, 0.03736067199707031, 0.03739644622802735, 0.037394432067871096, 0.037394432067871096, 0.037392383575439454, 0.037389312744140625, 0.0372674560546875, 0.03746713638305664, 0.037416961669921874, 0.03733606338500976, 0.03751321411132812, 0.03757056045532227, 0.037351425170898435, 0.03736883163452148, 0.037335041046142575, 0.03735756683349609, 0.03738521575927734, 0.03734630584716797, 0.03721932983398438, 0.037397502899169925, 0.03740364837646484, 0.037465087890625, 0.03727974319458008, 0.03751116943359375, 0.03736883163452148, 0.03720294570922852, 0.03728179168701172, 0.03764019012451172, 0.037395454406738284, 0.037294078826904296, 0.037379070281982424, 0.03743231964111328, 0.03737705612182617, 0.03722953414916992, 0.03727360153198242, 0.03775692749023438, 0.03735039901733399, 0.03786652755737305, 0.03738211059570312, 0.03764019012451172, 0.03744255828857422, 0.03744153594970703, 0.03740671920776367, 0.038042625427246096, 0.03756032180786133, 0.03753267288208008, 0.07615692901611328, 0.03728486251831055, 0.03751628875732422, 0.03750092697143555, 0.03723161697387695, 0.03742105484008789, 0.03728384017944336, 0.03728793716430664, 0.03723369598388672, 0.037351390838623044, 0.0372592658996582, 0.03703603363037109, 0.03736576080322265, 0.03826483154296875, 0.03825356674194336, 0.037498878479003905, 0.03745075225830078, 0.03750604629516602, 0.03733606338500976, 0.0373493766784668, 0.037405696868896485, 0.037548030853271484, 0.037897216796875, 0.0374015998840332, 0.03749273681640625, 0.03787980651855469, 0.03769548797607422, 0.03733401489257813, 0.037424129486083986, 0.03724288177490234, 0.03732275390625, 0.03731353759765625, 0.037348350524902346, 0.0373043212890625, 0.03746099090576172, 0.03751731109619141, 0.037424129486083986, 0.037477375030517575, 0.03738623809814453, 0.037395454406738284, 0.0372828483581543, 0.0373001937866211, 0.03731455993652344, 0.03749683380126953, 0.037678081512451174, 0.03893964767456055, 0.037498878479003905, 0.03726847839355469, 0.03724595260620117, 0.037381118774414065, 0.03728998565673828, 0.03745177459716797, 0.03747635269165039, 0.037425151824951174, 0.03755724716186523, 0.03739033508300781, 0.03731763076782227, 0.03747532653808594, 0.03732070541381836, 0.03763302230834961, 0.0374917106628418, 0.03746303939819336, 0.037392383575439454, 0.07604838562011719, 0.03738726425170898, 0.03732275390625, 0.037763072967529294, 0.03760537719726562, 0.037493759155273435, 0.03744153594970703, 0.037367809295654295, 0.037353473663330077, 0.03748966217041016, 0.037321727752685545, 0.03739136123657227, 0.03747020721435547, 0.03746201705932617, 0.03758899307250976, 0.037288959503173826, 0.03742617416381836, 0.037359615325927735, 0.03741900634765625, 0.03733708953857422, 0.0372592658996582, 0.03737497711181641, 0.03758182525634766, 0.03737395095825195, 0.03727052688598633, 0.03736678314208984, 0.037477375030517575, 0.03746201705932617, 0.03737702560424805, 0.03690291213989258, 0.03730022430419922, 0.037541889190673826, 0.03752140808105469, 0.03749683380126953, 0.03740979385375977, 0.037354496002197264, 0.037292030334472655, 0.03724288177490234, 0.03722956848144531, 0.03722956848144531, 0.0375203857421875, 0.03734630584716797, 0.03717836761474609, 0.0371671028137207, 0.037372928619384765, 0.037341182708740234, 0.03730636978149414, 0.03713945770263672, 0.03719987106323242, 0.03745177459716797, 0.03732889556884766, 0.03736166381835938, 0.0373493766784668, 0.03739648056030274, 0.03730739212036133, 0.03730739212036133, 0.03730944061279297, 0.03724390411376953, 0.037353473663330077, 0.03740467071533203, 0.0375203857421875, 0.03734630584716797, 0.03746406555175781, 0.07565721893310547, 0.0372674560546875, 0.0373125114440918, 0.037479423522949216, 0.03725721740722656, 0.03723468780517578, 0.0374200325012207, 0.03733196640014649, 0.03729817581176758, 0.03751321411132812, 0.03730739212036133, 0.03743334579467773, 0.03721932983398438, 0.03751833724975586, 0.03735244750976562, 0.037408767700195314, 0.03565260696411133, 0.035312641143798826, 0.03527782440185547, 0.0352542724609375, 0.03772415924072266, 0.0377077751159668, 0.03799244689941406, 0.03948953628540039, 0.037106689453125, 0.03671449661254883, 0.03655372619628906, 0.03666534423828125, 0.03620249557495117, 0.03514470291137695, 0.03519692611694336, 0.035335166931152344, 0.03523788833618164, 0.03600076675415039, 0.03778662490844727, 0.03708927917480469, 0.03679846572875976, 0.036724769592285156, 0.03671651077270508, 0.0366192626953125, 0.036624416351318356, 0.03651375961303711, 0.03685990524291992, 0.03674524688720703, 0.03836924743652344, 0.03679436874389649, 0.03653734588623047, 0.03676671981811523, 0.03650559997558594, 0.03653529739379883, 0.036706302642822264, 0.03682099151611328, 0.036528129577636716, 0.03685887908935547, 0.03667865753173828, 0.036674560546875, 0.036618240356445314, 0.03663052749633789, 0.03661721420288086, 0.03660595321655274, 0.0365926399230957, 0.036544513702392575, 0.036792320251464845]",tokens/s,26.434911627418618,, @@ -7300,7 +7849,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c107b-7531d3ac365d20b2674e2708;14ec0e56-3359-4dd8-b313-080dced26cb3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d42c2-5eb94eb5223e4a7430b40a0c;912486c3-e58b-47aa-83ec-1a5427208b46) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7370,7 +7919,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c14ea-608295171bf29b8d22a98e6e;a9c73255-9caf-48fd-a746-916fd6c82c61) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46fe-327bfa9f17e3d2523c8de62b;57f2ca23-f104-4797-84ff-cc62037d71cb) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7440,7 +7989,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1499-06a7ed1965d01e841ad5f2f1;40e6647a-d366-438e-9269-0c9f5574c2c0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46af-77e75bfc7e5cf56323b1370d;9cb15cc2-3c31-4120-99ed-38196ca97c03) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7540,7 +8089,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1363-078676957709b2a4723f3bf1;59478935-d103-462f-b9c7-841a4267fb6b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d457d-516ba5cf2cbc07081fc6897a;4022f992-ca15-4e03-8138-1e94dc7169e7) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7611,7 +8160,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c13d9-56de52c01fc7cc5b4a378db3;a5dd7877-31ca-46bc-9d1d-7a014e339c35) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45f1-7e81833649f084f75ae3e8bb;6c56cf7b-6d1e-47da-b09d-8ff6fd214d36) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7765,6 +8314,7 @@ ChildProcessError: Traceback (most recent call last): TypeError: DeciCoderAttention.forward() got an unexpected keyword argument 'cache_position' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2123.460608,5971.116032,0.0,5324.668928,5080.876032,s,10,3.2550746459960944,0.3255074645996094,0.0008337963032000422,0.325276611328125,0.3257435028076172,0.32685052642822265,0.3277361453247071,"[0.32795755004882815, 0.32484524536132814, 0.32533059692382815, 0.3251536560058594, 0.32514990234375, 0.32549749755859375, 0.32518228149414063, 0.3252226257324219, 0.32538031005859375, 0.32535498046875]",tokens/s,786.4642991056838,kWh,3.8430350602314035e-06,2.105488517026383e-06,1.9587694881612944e-05,2.553621845887073e-05,tokens/kWh,10024976.89359605,MB,2123.460608,5971.116032,0.0,5324.668928,5290.118656,s,10,177.20532812500002,17.720532812500004,0.0052544305035837425,17.7196923828125,17.7276015625,17.7277421875,17.7278546875,"[17.71912890625, 17.71456640625, 17.720630859375, 17.7278828125, 17.71237890625, 17.727517578125, 17.716267578125, 17.719486328125, 17.7275703125, 17.7198984375]",tokens/s,3.5551978412048655,kWh,0.0002091554838338847,0.00011463415165342726,0.0010539822473069871,0.0013777718827942989,tokens/kWh,45726.00209566469,,s,629,179.85040585327167,0.2859306929304793,0.03863957620976894,0.2812538757324219,0.28160614013671875,0.2817001525878906,0.6063846411132813,"[0.28122930908203125, 0.28129177856445314, 0.2810245056152344, 0.2810921020507812, 0.2811351013183594, 0.28118630981445314, 0.28117401123046876, 0.281101318359375, 0.28111154174804687, 0.28105831909179685, 0.2810572814941406, 0.28108389282226565, 0.28105010986328127, 0.2811955261230469, 0.28121395874023436, 0.2812866516113281, 0.2815754089355469, 0.28128973388671874, 0.2816645202636719, 0.281312255859375, 0.28124160766601564, 0.28118630981445314, 0.2813624267578125, 0.28113613891601563, 0.28115045166015623, 0.28137777709960937, 0.2815621032714844, 0.28139520263671874, 0.28123544311523435, 0.281302001953125, 0.2815447082519531, 0.28124261474609374, 0.28110540771484377, 0.28135833740234373, 0.28128460693359375, 0.2810368041992187, 0.2815948791503906, 0.2811248779296875, 0.28113201904296875, 0.28118118286132815, 0.28119448852539064, 0.28112896728515624, 0.28108901977539064, 0.28126412963867187, 0.28096820068359374, 0.2810245056152344, 0.281270263671875, 0.28120883178710937, 0.28115045166015623, 0.2812313537597656, 0.2814156799316406, 0.28137677001953126, 0.281375732421875, 0.28129995727539064, 0.281628662109375, 0.2811781005859375, 0.28127743530273436, 0.28120779418945313, 0.281449462890625, 0.2811463623046875, 0.28092312622070315, 0.2812456970214844, 0.6060492553710938, 0.280764404296875, 0.2809722900390625, 0.2808719482421875, 0.281027587890625, 0.2810429382324219, 0.28127435302734377, 0.2810992736816406, 0.28099072265625, 0.2813102111816406, 0.28106854248046875, 0.28104193115234377, 0.28099172973632813, 0.28096408081054686, 0.28089447021484376, 0.281059326171875, 0.2810460205078125, 0.28098968505859373, 0.2810777587890625, 0.2809354248046875, 0.281017333984375, 0.2809866333007813, 0.2811197509765625, 0.2810849304199219, 0.2814791564941406, 0.28121600341796876, 0.28174029541015627, 0.2812938232421875, 0.28124774169921873, 0.28133477783203126, 0.2811064453125, 0.28119143676757813, 0.2810798034667969, 0.2810624084472656, 0.28098458862304687, 0.28118731689453125, 0.28110848999023436, 0.2811555786132813, 0.28107571411132815, 0.28158770751953127, 0.281459716796875, 0.2814412841796875, 0.2813644714355469, 0.2810142822265625, 0.2813040771484375, 0.28120574951171873, 0.28168295288085937, 0.28170138549804685, 0.2815539245605469, 0.2811822204589844, 0.2811094970703125, 0.28119143676757813, 0.28096307373046875, 0.2810716247558594, 0.28109722900390627, 0.2813061218261719, 0.2811678771972656, 0.2812630920410156, 0.2813839416503906, 0.2814310302734375, 0.281312255859375, 0.2810705871582031, 0.28164608764648436, 0.6065008544921875, 0.28089752197265627, 0.2808965148925781, 0.28090573120117185, 0.281059326171875, 0.28105831909179685, 0.2810101623535156, 0.28111566162109375, 0.28095489501953125, 0.2809876403808594, 0.2810449829101562, 0.2810992736816406, 0.28105831909179685, 0.28147406005859377, 0.2811781005859375, 0.28130303955078123, 0.28129791259765624, 0.2810439758300781, 0.2809374694824219, 0.2809610290527344, 0.2811955261230469, 0.28140542602539065, 0.281196533203125, 0.2814361572265625, 0.28143307495117187, 0.2812866516113281, 0.2814289855957031, 0.2812600402832031, 0.2813726806640625, 0.2812508239746094, 0.2812313537597656, 0.28151092529296873, 0.2815119323730469, 0.28151092529296873, 0.28148532104492185, 0.2814044189453125, 0.2812886962890625, 0.28121395874023436, 0.2811412353515625, 0.28146585083007813, 0.28120166015625, 0.2812876892089844, 0.28140032958984373, 0.28124261474609374, 0.2812620849609375, 0.2812119140625, 0.281275390625, 0.2814259338378906, 0.2816952209472656, 0.2817484741210938, 0.2815324096679688, 0.28149554443359376, 0.281486328125, 0.2813368225097656, 0.28129995727539064, 0.2814289855957031, 0.2814341125488281, 0.2812600402832031, 0.28163482666015627, 0.2814218139648437, 0.2814986267089844, 0.2813286437988281, 0.2812938232421875, 0.6064302368164063, 0.28105831909179685, 0.28143719482421875, 0.28149249267578125, 0.28150885009765625, 0.2818365478515625, 0.28147198486328123, 0.28145150756835935, 0.2814064636230469, 0.281459716796875, 0.28177203369140624, 0.2814310302734375, 0.281238525390625, 0.2812333984375, 0.28132659912109376, 0.28122213745117186, 0.2811781005859375, 0.28126412963867187, 0.28113409423828123, 0.2814412841796875, 0.28169830322265627, 0.2810572814941406, 0.28097537231445313, 0.28115045166015623, 0.281069580078125, 0.2811566162109375, 0.281143310546875, 0.2815662231445312, 0.2811402282714844, 0.28129995727539064, 0.28122113037109375, 0.28142385864257813, 0.2810705871582031, 0.28127435302734377, 0.2812467346191406, 0.2814709777832031, 0.28154061889648435, 0.2812989501953125, 0.281712646484375, 0.28118630981445314, 0.28149658203125, 0.2824161376953125, 0.28164913940429687, 0.28166656494140624, 0.28172903442382813, 0.28148532104492185, 0.28164813232421876, 0.2819051513671875, 0.2816860046386719, 0.28129074096679685, 0.2810828857421875, 0.28118014526367185, 0.28128564453125, 0.28141876220703127, 0.28150067138671875, 0.2815682678222656, 0.28134912109375, 0.28145562744140623, 0.2814617614746094, 0.2815119323730469, 0.28112896728515624, 0.28120065307617187, 0.2814361572265625, 0.6062673950195312, 0.2807910461425781, 0.2809088134765625, 0.2809866333007813, 0.28096511840820315, 0.2809466857910156, 0.28097024536132814, 0.2810449829101562, 0.28104702758789063, 0.2812518310546875, 0.2811535339355469, 0.2812518310546875, 0.28147406005859377, 0.28112896728515624, 0.28104806518554687, 0.2808985595703125, 0.28109722900390627, 0.2809774169921875, 0.2810019836425781, 0.2810368041992187, 0.28092007446289063, 0.28094259643554687, 0.280816650390625, 0.2810726318359375, 0.28107467651367185, 0.28106854248046875, 0.2810071105957031, 0.2813572998046875, 0.281196533203125, 0.28106341552734376, 0.28121908569335935, 0.2811473999023438, 0.28106137084960936, 0.2811832275390625, 0.28110540771484377, 0.28103372192382814, 0.2810624084472656, 0.281280517578125, 0.28099993896484377, 0.2810992736816406, 0.28094464111328127, 0.2814136352539062, 0.2813040771484375, 0.2811217956542969, 0.2810163269042969, 0.281491455078125, 0.28120574951171873, 0.28158462524414063, 0.2816952209472656, 0.28116070556640627, 0.2811617431640625, 0.28148736572265626, 0.28130712890625, 0.28105523681640626, 0.281343994140625, 0.28129278564453125, 0.28119244384765624, 0.2813644714355469, 0.2811852722167969, 0.28122213745117186, 0.2812518310546875, 0.28121600341796876, 0.2815703125, 0.6071398315429688, 0.2814771118164062, 0.28115966796875, 0.28129177856445314, 0.28113919067382814, 0.281417724609375, 0.28138290405273436, 0.28137777709960937, 0.28117913818359375, 0.28141671752929687, 0.28105831909179685, 0.2811975708007812, 0.2810941467285156, 0.28144537353515625, 0.28139620971679685, 0.28142385864257813, 0.2814464111328125, 0.2810378112792969, 0.2813409423828125, 0.2816624755859375, 0.28126412963867187, 0.2811781005859375, 0.2808565673828125, 0.2814422912597656, 0.28143820190429686, 0.28139111328125, 0.2813706359863281, 0.2814730224609375, 0.28143206787109376, 0.2816470947265625, 0.28146484375, 0.281670654296875, 0.2815037536621094, 0.2814761047363281, 0.2816030578613281, 0.2814474182128906, 0.28148019409179686, 0.2816225280761719, 0.28140338134765625, 0.28146380615234373, 0.2813716430664063, 0.28143719482421875, 0.2815867004394531, 0.28133273315429685, 0.28145150756835935, 0.28143002319335936, 0.2815672302246094, 0.2816778259277344, 0.28152935791015626, 0.28127435302734377, 0.28112896728515624, 0.28127435302734377, 0.2815170593261719, 0.281143310546875, 0.28146585083007813, 0.2816501770019531, 0.28121395874023436, 0.2810716247558594, 0.2812989501953125, 0.2811627502441406, 0.2813183898925781, 0.2815201416015625, 0.28156414794921875, 0.6065889282226562, 0.2809610290527344, 0.2809169921875, 0.28118014526367185, 0.2810624084472656, 0.28140237426757814, 0.28091595458984375, 0.2810408935546875, 0.2810378112792969, 0.28120269775390627, 0.2810142822265625, 0.2811064453125, 0.2812733459472656, 0.2811217956542969, 0.2809190368652344, 0.2811146240234375, 0.28121395874023436, 0.28103988647460937, 0.28125799560546877, 0.28095590209960936, 0.2810921020507812, 0.2812047424316406, 0.2810859375, 0.28157644653320313, 0.2811074523925781, 0.2813736877441406, 0.28132965087890627, 0.28149554443359376, 0.28099072265625, 0.2809395141601562, 0.2811760559082031, 0.2810900573730469, 0.2810798034667969, 0.28106036376953125, 0.28111770629882815, 0.2811985778808594, 0.2812600402832031, 0.28122930908203125, 0.28111358642578127, 0.2813102111816406, 0.2814115905761719, 0.28146484375, 0.28112384033203125, 0.28095281982421877, 0.28117300415039065, 0.2812620849609375, 0.28116378784179685, 0.28120065307617187, 0.28133377075195315, 0.2811207580566406, 0.28124160766601564, 0.28134808349609375, 0.2812538757324219, 0.2812600402832031, 0.281343994140625, 0.28151910400390623, 0.28132965087890627, 0.2814218139648437, 0.28138906860351565, 0.28175668334960935, 0.28138290405273436, 0.2815150146484375, 0.281712646484375, 0.6064854736328125, 0.28096517944335936, 0.280976318359375, 0.281059326171875, 0.28107366943359374, 0.28107366943359374, 0.2810449829101562, 0.28102859497070315, 0.28117196655273435, 0.2813102111816406, 0.2813982849121094, 0.28133477783203126, 0.2816522216796875, 0.28102041625976565, 0.281122802734375, 0.28104806518554687, 0.28105010986328127, 0.2810296325683594, 0.28088525390625, 0.2810900573730469, 0.28102554321289064, 0.28089547729492187, 0.28095489501953125, 0.2809866333007813, 0.28106137084960936, 0.2808616943359375, 0.281101318359375, 0.2810357666015625, 0.2813009948730469, 0.2810091552734375, 0.281481201171875, 0.2811463623046875, 0.28136038208007813, 0.28138290405273436, 0.28177407836914065, 0.2813921203613281, 0.281101318359375, 0.281275390625, 0.2812200927734375, 0.28148223876953127, 0.281270263671875, 0.28124774169921873, 0.2812549133300781, 0.281275390625, 0.28143515014648435, 0.28106341552734376, 0.2812938232421875, 0.2816450500488281, 0.28184884643554686, 0.2816512145996094, 0.2815047607421875, 0.28159896850585936, 0.2812528686523437, 0.281175048828125, 0.2812651672363281, 0.28133172607421875, 0.28128256225585935, 0.28144332885742185, 0.2814535827636719, 0.28138290405273436, 0.28144436645507814, 0.28169830322265627, 0.2818529357910156, 0.6073692016601563, 0.2816573486328125, 0.2816153564453125, 0.28132760620117186, 0.2814689331054688, 0.28154367065429686, 0.281027587890625, 0.2809866333007813, 0.28119143676757813, 0.2810900573730469, 0.280995849609375, 0.2812333984375, 0.2813255615234375, 0.28138494873046876, 0.2815784912109375, 0.28136959838867187, 0.2814412841796875, 0.28131329345703127, 0.2810378112792969, 0.2817525634765625, 0.2815733642578125, 0.28126617431640627, 0.2815447082519531, 0.2814975891113281, 0.28120883178710937, 0.2812549133300781, 0.2811269226074219, 0.28102041625976565, 0.2809354248046875, 0.2811125793457031, 0.28105523681640626, 0.28127743530273436, 0.2812518310546875, 0.28106854248046875, 0.28128256225585935, 0.2811463623046875, 0.28116378784179685, 0.2817177734375, 0.28164096069335937, 0.28217242431640627, 0.28151602172851564, 0.28213861083984376, 0.2816296997070312, 0.28145458984375, 0.28150271606445315, 0.28149249267578125, 0.28160614013671875, 0.28172491455078125, 0.2816174011230469, 0.28166656494140624, 0.28144024658203126, 0.2816296997070312, 0.2815784912109375, 0.2815047607421875, 0.28180685424804686, 0.281459716796875, 0.2812569580078125, 0.2812108764648438, 0.2812108764648438, 0.2811453552246094, 0.28124774169921873, 0.2810777587890625, 0.28122930908203125, 0.606845947265625, 0.2807613830566406, 0.28094561767578125, 0.2809241638183594, 0.28097637939453124, 0.2810091552734375, 0.2809774169921875, 0.2810460205078125, 0.28142385864257813, 0.28146585083007813, 0.2812518310546875, 0.28173208618164064, 0.28189797973632813, 0.2814361572265625, 0.28138189697265625, 0.28133273315429685, 0.2814044189453125, 0.28123544311523435, 0.28123544311523435, 0.28124774169921873, 0.2811883544921875, 0.28133377075195315, 0.2814422912597656, 0.28136651611328123, 0.28155084228515626, 0.28144436645507814, 0.28169012451171876, 0.2812231750488281, 0.28088116455078127, 0.2811125793457031, 0.2812108764648438, 0.2812108764648438, 0.28103884887695313, 0.2812651672363281, 0.28122213745117186, 0.28098968505859373, 0.2810654602050781, 0.28117300415039065, 0.28104806518554687, 0.2814894104003906, 0.281248779296875, 0.2813368225097656, 0.28116683959960936, 0.2811555786132813, 0.28160513305664064, 0.2811412353515625, 0.28160614013671875, 0.2813726806640625, 0.28156414794921875, 0.2810163269042969, 0.28116888427734377, 0.2811412353515625, 0.2812467346191406, 0.2811412353515625, 0.2812590026855469, 0.28136651611328123, 0.28124160766601564, 0.28138494873046876, 0.2813511657714844, 0.28120269775390627, 0.28121701049804687, 0.28180685424804686, 0.28130712890625]",tokens/s,3.4973510180074943,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -7794,7 +8344,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664beea1-2bbe96514c199b0f58a5c2c0;caa877d3-b680-4780-a173-7a82f7fb1790) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3faf-26e2c91b7ae530ca4b844f56;203b9e54-c410-41a3-8c78-41067b7db408) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -7883,7 +8433,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c138b-586fd2c83afb97577bc6a629;67e304b7-f6bc-46a8-891d-73a0bfabd68d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45a3-3faa5d1a03d2c2a52b761859;696f247e-6601-453e-993a-1f0155b6890e) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7943,7 +8493,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee78-37998b77361ffc3b6681e4fa;aa0c704c-f1af-4772-824c-8d0a96c98f00) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3f88-623514257272f29e662dcdd4;f94a69da-f633-49b3-ba9c-d55fd72100cf) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -8244,7 +8794,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664becf0-14be02e30af14c777bd2a3df;7917bebd-3d0c-4f80-90ed-32450a248a2b) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3e08-57b7fb7114987c10146249cb;21d9a2f7-577c-4613-af95-8816d068d73a) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -8329,7 +8879,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1313-11ab88cd407c4b84153d5de2;d1b685f2-ea8f-4a80-b843-532392577ffd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4530-65852dc614b181227f8b2339;6fc2c377-66a5-4580-bca8-551d01793649) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8445,7 +8995,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c120c-02aacefc47a1783a1090a8cf;1bb57294-63d2-49bd-b61c-ca7e4aac18b4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d446a-13d4f59e192f2bbc7821b91f;4aff3160-af8a-45e0-a69a-55f55da4afef) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8519,6 +9069,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1856.221184,3368.5504,0.0,2722.103296,2585.357312,s,10,0.7366494750976563,0.07366494750976563,0.00039249671689789645,0.0736723518371582,0.07419290008544921,0.07419320220947266,0.07419344390869141,"[0.0741935043334961, 0.07335641479492187, 0.07308262634277343, 0.07368495941162109, 0.07417286682128907, 0.07325129699707031, 0.07419283294677734, 0.07332640075683594, 0.07372882843017578, 0.07365974426269531]",tokens/s,3475.1942226805027,kWh,8.657666441394562e-07,4.743867474013832e-07,4.119346361168371e-06,5.4594997527092105e-06,tokens/kWh,46890743.03427949,MB,1856.221184,3368.5504,0.0,2722.103296,2628.769792,s,10,38.86898974609376,3.8868989746093745,0.002142874021588634,3.886506591796875,3.8892139404296877,3.890647985839844,3.891795222167969,"[3.89208203125, 3.884625, 3.88758349609375, 3.88604248046875, 3.886436767578125, 3.885375, 3.888895263671875, 3.886576416015625, 3.88693505859375, 3.884438232421875]",tokens/s,16.20829365814206,kWh,4.582713047034936e-05,2.5112934640191582e-05,0.00021524919024462234,0.0002861892553551633,tokens/kWh,220134.050531759,,s,629,39.468826690673836,0.06274853209964042,0.008805392613180215,0.061639678955078124,0.06181806030273437,0.06226268157958985,0.1355356140136719,"[0.0625162239074707, 0.06246809768676758, 0.062304256439208984, 0.06238105773925781, 0.06236467361450195, 0.06246092987060547, 0.062456832885742185, 0.06226432037353516, 0.061895679473876954, 0.06161510467529297, 0.061515777587890626, 0.06170009613037109, 0.06151987075805664, 0.06168473434448242, 0.06160793685913086, 0.061649921417236325, 0.061518848419189455, 0.061574142456054685, 0.062078079223632815, 0.06155251312255859, 0.061625343322753906, 0.061636608123779295, 0.06158028793334961, 0.061625343322753906, 0.06154547119140625, 0.06160076904296875, 0.06159564971923828, 0.06160076904296875, 0.061606910705566405, 0.06164582443237305, 0.061652992248535154, 0.061623294830322264, 0.06156185531616211, 0.06175743865966797, 0.061557758331298826, 0.061669376373291014, 0.06161100769042969, 0.061769729614257814, 0.061658111572265625, 0.06219878387451172, 0.0616703987121582, 0.061666305541992185, 0.061620223999023435, 0.0616734733581543, 0.06161510467529297, 0.06164684677124024, 0.061663230895996096, 0.06168985748291016, 0.06160179138183594, 0.06177382278442383, 0.06160179138183594, 0.06164582443237305, 0.06184550476074219, 0.06161100769042969, 0.061652992248535154, 0.06158848190307617, 0.06170828628540039, 0.06169190216064453, 0.061652992248535154, 0.061575233459472654, 0.061744064331054685, 0.06192236709594726, 0.13562771606445312, 0.06159257507324219, 0.061623294830322264, 0.06195609664916992, 0.06166732788085937, 0.06157209777832031, 0.06155878448486328, 0.06166527938842774, 0.06160793685913086, 0.06156083297729492, 0.061618209838867184, 0.06155465698242187, 0.061639678955078124, 0.06154035186767578, 0.06164787292480469, 0.061722625732421876, 0.06158950424194336, 0.06161305618286133, 0.06162435150146484, 0.06246294403076172, 0.0619683837890625, 0.0616099853515625, 0.06166220855712891, 0.06152601623535156, 0.06164889526367188, 0.06153734588623047, 0.06161401748657227, 0.06158233642578125, 0.06157004928588867, 0.06159769439697266, 0.06160179138183594, 0.06163455963134765, 0.061584384918212894, 0.061551616668701174, 0.061639678955078124, 0.06158848190307617, 0.061601951599121095, 0.06156272125244141, 0.06161612701416016, 0.06156697463989258, 0.06195609664916992, 0.06161510467529297, 0.06171657562255859, 0.061641632080078126, 0.061723648071289064, 0.061636608123779295, 0.06167552185058594, 0.06159564971923828, 0.0615997428894043, 0.061645889282226564, 0.061612991333007815, 0.06159667205810547, 0.06162227249145508, 0.061636608123779295, 0.06173286437988281, 0.061914112091064455, 0.06161305618286133, 0.06163455963134765, 0.06170214462280273, 0.06166425704956055, 0.06158643341064453, 0.06162944030761719, 0.06158540725708008, 0.13527247619628907, 0.061581279754638674, 0.06155878448486328, 0.06152499389648437, 0.06166835021972656, 0.061590526580810545, 0.061517822265625, 0.06159872055053711, 0.061587455749511716, 0.06171955108642578, 0.06164070510864258, 0.061568000793457034, 0.06157721710205078, 0.06150348663330078, 0.06161203384399414, 0.06168169784545898, 0.06163759994506836, 0.061608959197998046, 0.06161407852172852, 0.06154240036010742, 0.06161920166015625, 0.062476287841796874, 0.06166835021972656, 0.06163046264648438, 0.06195814514160156, 0.06177587127685547, 0.0623902702331543, 0.061707263946533204, 0.06162124633789062, 0.0615731201171875, 0.06163251113891602, 0.061565025329589844, 0.061674400329589846, 0.06159564971923828, 0.0616519660949707, 0.06163763046264648, 0.06163455963134765, 0.06162944030761719, 0.0616734733581543, 0.06159667205810547, 0.0616703987121582, 0.061571071624755856, 0.06169702529907226, 0.06159462356567383, 0.06166425704956055, 0.06162636947631836, 0.06164889526367188, 0.06272614288330078, 0.06170521545410156, 0.06171955108642578, 0.06169702529907226, 0.06160281753540039, 0.061656063079833984, 0.06173286437988281, 0.062260223388671876, 0.061813758850097655, 0.061811710357666014, 0.061819969177246095, 0.06171744155883789, 0.061764606475830076, 0.06167244720458984, 0.061655040740966796, 0.061584384918212894, 0.13583973693847656, 0.06158643341064453, 0.06156083297729492, 0.06164070510864258, 0.06155263900756836, 0.06152703857421875, 0.06155059051513672, 0.061587455749511716, 0.06175641632080078, 0.061603839874267576, 0.06168678283691406, 0.061625343322753906, 0.06214553451538086, 0.0615813102722168, 0.06157926559448242, 0.06165913772583008, 0.06169401550292969, 0.06158636856079101, 0.061620223999023435, 0.06160588836669922, 0.06167244720458984, 0.061709312438964846, 0.06159769439697266, 0.061652992248535154, 0.06162124633789062, 0.061636608123779295, 0.061606910705566405, 0.06158950424194336, 0.06163251113891602, 0.061591552734375, 0.06161203384399414, 0.061639678955078124, 0.06177280044555664, 0.06168988800048828, 0.06166422271728516, 0.06160588836669922, 0.06171443176269531, 0.061841407775878904, 0.06170521545410156, 0.061625343322753906, 0.06257766342163086, 0.06173388671875, 0.061682689666748045, 0.061551616668701174, 0.0617154541015625, 0.06164787292480469, 0.06163763046264648, 0.061587455749511716, 0.06166732788085937, 0.06187228775024414, 0.06169481658935547, 0.06167552185058594, 0.06157209777832031, 0.06177996826171875, 0.06175129699707031, 0.06170214462280273, 0.061555744171142575, 0.06177276611328125, 0.06169497680664063, 0.061618175506591794, 0.061625343322753906, 0.06185369491577149, 0.06161920166015625, 0.13637632751464843, 0.062094337463378904, 0.061722625732421876, 0.06158950424194336, 0.06156083297729492, 0.061636608123779295, 0.06153113555908203, 0.061567039489746095, 0.06163654327392578, 0.0615813102722168, 0.0616703987121582, 0.06159360122680664, 0.061620223999023435, 0.061590526580810545, 0.06161305618286133, 0.06163763046264648, 0.06162227249145508, 0.06240870285034179, 0.06161920166015625, 0.06155878448486328, 0.06172671890258789, 0.0615599365234375, 0.06163238525390625, 0.06171136093139649, 0.06158540725708008, 0.06151987075805664, 0.06161407852172852, 0.06159564971923828, 0.0616099853515625, 0.0615997428894043, 0.0616827507019043, 0.061606849670410156, 0.06166527938842774, 0.06172774505615235, 0.06173183822631836, 0.06160793685913086, 0.06163455963134765, 0.061603839874267576, 0.061603839874267576, 0.06159360122680664, 0.06163148880004883, 0.0615731201171875, 0.061639678955078124, 0.06152601623535156, 0.06165302276611328, 0.06166934585571289, 0.06161407852172852, 0.061620223999023435, 0.06162636947631836, 0.06263603210449219, 0.06173593521118164, 0.06169295883178711, 0.061610977172851564, 0.06161305618286133, 0.06162124633789062, 0.06166835021972656, 0.061700126647949216, 0.06224483108520508, 0.06181785583496094, 0.06161612701416016, 0.06189977645874024, 0.061739105224609375, 0.061621150970458984, 0.13536972045898438, 0.06163455963134765, 0.06161731338500977, 0.06172252655029297, 0.06154950332641602, 0.061603839874267576, 0.06166016006469727, 0.061590526580810545, 0.06159769439697266, 0.06160281753540039, 0.06174310302734375, 0.06155878448486328, 0.0615813102722168, 0.06159769439697266, 0.06164582443237305, 0.06159257507324219, 0.061636608123779295, 0.06157209777832031, 0.06174720001220703, 0.06168166351318359, 0.061584384918212894, 0.06155878448486328, 0.06153948974609375, 0.06161494445800781, 0.06175129699707031, 0.06155878448486328, 0.06163251113891602, 0.06156492614746094, 0.061671424865722656, 0.06155878448486328, 0.06164889526367188, 0.0617625617980957, 0.06167859268188477, 0.06308761596679688, 0.061780990600585936, 0.061655040740966796, 0.061669376373291014, 0.06156185531616211, 0.06179635238647461, 0.06155263900756836, 0.061687808990478515, 0.06156595230102539, 0.0617523193359375, 0.061723648071289064, 0.061685760498046874, 0.06160588836669922, 0.0616049919128418, 0.061594497680664065, 0.061568000793457034, 0.061666305541992185, 0.06168371200561523, 0.061661182403564455, 0.061633537292480466, 0.061652992248535154, 0.06165094375610351, 0.06170111846923828, 0.06174515151977539, 0.06168166351318359, 0.06168985748291016, 0.06164684677124024, 0.061617153167724606, 0.06170320129394531, 0.06162019348144531, 0.13634867858886718, 0.061669376373291014, 0.06157619094848633, 0.061677566528320314, 0.06161100769042969, 0.06153526306152344, 0.06168368148803711, 0.06178713607788086, 0.06175539016723633, 0.061590526580810545, 0.061674495697021485, 0.06164377593994141, 0.06158643341064453, 0.06159564971923828, 0.06164377593994141, 0.06164787292480469, 0.061636608123779295, 0.06153420639038086, 0.06174105453491211, 0.06166835021972656, 0.06171443176269531, 0.06157823944091797, 0.061693950653076174, 0.0633589744567871, 0.06182400131225586, 0.06165708923339844, 0.06162124633789062, 0.061742080688476565, 0.06158233642578125, 0.061774848937988285, 0.061881343841552736, 0.061818878173828126, 0.06159360122680664, 0.06161612701416016, 0.06182809448242187, 0.06172876739501953, 0.06173183822631836, 0.061664287567138674, 0.061742080688476565, 0.06162838363647461, 0.061633537292480466, 0.06159462356567383, 0.06164070510864258, 0.061852737426757814, 0.061847488403320314, 0.0616099853515625, 0.06166220855712891, 0.06165708923339844, 0.06167552185058594, 0.061625343322753906, 0.061797374725341796, 0.06176358413696289, 0.06170828628540039, 0.061644798278808595, 0.06157823944091797, 0.061695999145507815, 0.061557758331298826, 0.06290022277832032, 0.06170828628540039, 0.06167552185058594, 0.06157004928588867, 0.061685760498046874, 0.061639678955078124, 0.13560012817382813, 0.06154956817626953, 0.06159257507324219, 0.061664321899414065, 0.06175942230224609, 0.061708446502685546, 0.061510494232177734, 0.061656063079833984, 0.06162636947631836, 0.06158441543579102, 0.061638622283935546, 0.0615546875, 0.06163455963134765, 0.06155168151855469, 0.06161196899414063, 0.06161407852172852, 0.06238719940185547, 0.06159769439697266, 0.06166016006469727, 0.061658111572265625, 0.06163455963134765, 0.06155059051513672, 0.06164684677124024, 0.061532161712646485, 0.061644798278808595, 0.061692928314208986, 0.061917182922363284, 0.061932544708251956, 0.061709312438964846, 0.06181990432739258, 0.06227046585083008, 0.06178303909301758, 0.06163455963134765, 0.06159257507324219, 0.061641727447509766, 0.061695999145507815, 0.061706432342529295, 0.06160879898071289, 0.06169392013549805, 0.061598751068115236, 0.06168060684204101, 0.061641727447509766, 0.06170009613037109, 0.061723648071289064, 0.061636608123779295, 0.061666305541992185, 0.06164582443237305, 0.061813758850097655, 0.06163148880004883, 0.06166220855712891, 0.06170009613037109, 0.06167859268188477, 0.06173491287231445, 0.06168678283691406, 0.061613086700439454, 0.06166217422485352, 0.06156185531616211, 0.061712383270263675, 0.061704193115234375, 0.06167660903930664, 0.06163347244262695, 0.06173388671875, 0.06157926559448242, 0.1358909454345703, 0.061585472106933596, 0.06161913681030273, 0.06158643341064453, 0.06191007995605469, 0.06163654327392578, 0.06151379013061523, 0.06163552093505859, 0.06162739181518555, 0.06161305618286133, 0.061639678955078124, 0.061499393463134766, 0.06164582443237305, 0.06164070510864258, 0.06162944030761719, 0.061644798278808595, 0.06170009613037109, 0.06242816162109375, 0.06164787292480469, 0.06159878540039063, 0.06164985656738281, 0.06165094375610351, 0.06169702529907226, 0.06156595230102539, 0.06164889526367188, 0.06171750259399414, 0.06167244720458984, 0.06157209777832031, 0.06165200042724609, 0.06158233642578125, 0.061629409790039065, 0.06162124633789062, 0.06156288146972656, 0.061585472106933596, 0.06165599822998047, 0.06158540725708008, 0.06170111846923828, 0.06161616134643555, 0.06170723342895508, 0.06167244720458984, 0.06189977645874024, 0.06173183822631836, 0.061659168243408204, 0.06166115188598633, 0.06170431900024414, 0.06161600112915039, 0.061656063079833984, 0.061663230895996096, 0.06173081588745117, 0.06170828628540039, 0.0616734733581543, 0.06315827178955079, 0.06174924850463867, 0.061710334777832034, 0.06160486221313476, 0.06162944030761719, 0.06163558578491211, 0.06182092666625977, 0.06171955108642578, 0.061692928314208986, 0.06157823944091797, 0.061656063079833984, 0.06159769439697266, 0.13598208618164062, 0.06162636947631836, 0.06159462356567383, 0.06156492614746094, 0.06158643341064453, 0.061563934326171875, 0.06156079864501953, 0.06168678283691406, 0.06158339309692383, 0.06161507034301758, 0.06163558578491211, 0.0615618896484375, 0.06162019348144531, 0.061518848419189455, 0.06167244720458984, 0.06163046264648438, 0.06163763046264648, 0.06160588836669922, 0.06163148880004883, 0.06158540725708008, 0.06164889526367188, 0.06164892959594727, 0.061594593048095704, 0.06151168060302734, 0.06163251113891602, 0.06157926559448242, 0.0629678077697754, 0.06161305618286133, 0.06163558578491211, 0.06159360122680664, 0.06161203384399414, 0.061655040740966796, 0.061587455749511716, 0.06164275360107422, 0.06164889526367188, 0.06159872055053711, 0.061656063079833984, 0.06163455963134765, 0.061687808990478515, 0.0616099853515625, 0.061692928314208986, 0.061608959197998046, 0.06170111846923828, 0.06162739181518555, 0.06164889526367188, 0.06160179138183594, 0.06163251113891602, 0.06161407852172852, 0.061652030944824215, 0.061599681854248044, 0.06173183822631836, 0.06154444885253906, 0.06162636947631836, 0.061669376373291014, 0.06157622528076172, 0.06165910339355469, 0.06154550552368164, 0.06167855834960938, 0.06169702529907226, 0.061652992248535154, 0.061699073791503904, 0.06180044937133789, 0.06164179229736328]",tokens/s,15.936627782974552,, +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1479.114752,1841.823744,0.0,1195.37664,1022.155264,s,10,0.6871012496948242,0.06871012496948242,0.00033873627735752553,0.06860185623168946,0.06884490509033203,0.06926672515869141,0.06960418121337891,"[0.06968854522705079, 0.06850118255615234, 0.06852729797363281, 0.06862016296386719, 0.06869097900390625, 0.06858354949951172, 0.06851074981689453, 0.06875116729736327, 0.06849430084228515, 0.06873331451416016]",tokens/s,3725.797327740305,kWh,8.094942381483598e-07,4.4356150458715745e-07,3.836808624999982e-06,5.089864367735499e-06,tokens/kWh,50296035.710259095,MB,1479.114752,1841.823744,0.0,1195.37664,1091.597312,s,10,35.17510473632813,3.517510473632812,0.0019858794372678286,3.5167731933593753,3.5191181640625,3.52109130859375,3.5226698242187497,"[3.523064453125, 3.5170556640625, 3.51595458984375, 3.516420166015625, 3.516550048828125, 3.5174658203125, 3.51656103515625, 3.516367919921875, 3.5186796875, 3.5169853515625]",tokens/s,17.91039443158641,kWh,4.1521001642561397e-05,2.275550190917111e-05,0.0001961486569188012,0.00026042516047053367,tokens/kWh,241912.10974458922,,s,629,35.73639368057251,0.05681461634431242,0.008170033415171557,0.05581414413452149,0.05589258193969727,0.055985150909423825,0.12453187377929688,"[0.05623807907104492, 0.056237056732177736, 0.056357887268066405, 0.056323070526123044, 0.05624934387207031, 0.05632819366455078, 0.05631078338623047, 0.056376319885253906, 0.05601279830932617, 0.05586739349365234, 0.05577318572998047, 0.05581107330322266, 0.055946239471435545, 0.05584998321533203, 0.05613875198364258, 0.05595852661132812, 0.05590220642089844, 0.05599334335327148, 0.055934974670410156, 0.05595238494873047, 0.05628211212158203, 0.05583871841430664, 0.05575987243652344, 0.055757823944091796, 0.055818241119384764, 0.055790592193603515, 0.0557946891784668, 0.05578649520874023, 0.05578035354614258, 0.05583871841430664, 0.055836673736572265, 0.055803905487060546, 0.055874561309814455, 0.0558919677734375, 0.05581721496582031, 0.05585919952392578, 0.055828479766845705, 0.05583871841430664, 0.055795711517333986, 0.05586636734008789, 0.055826431274414064, 0.055774208068847655, 0.055801856994628904, 0.05579776000976563, 0.055820289611816405, 0.05580083084106445, 0.0557496337890625, 0.05584281539916992, 0.0558766098022461, 0.05587865447998047, 0.05584281539916992, 0.055782398223876956, 0.055809024810791016, 0.055788543701171874, 0.05576499176025391, 0.05582950210571289, 0.055911422729492184, 0.0558551025390625, 0.05586739349365234, 0.05582745742797852, 0.055890945434570315, 0.05582438278198242, 0.12451123046875, 0.05576396942138672, 0.05577523040771484, 0.05576499176025391, 0.05576806259155274, 0.05575475311279297, 0.05576499176025391, 0.0557916145324707, 0.05580799865722656, 0.05574655914306641, 0.05579673767089844, 0.05575987243652344, 0.055812095642089846, 0.055739391326904295, 0.05584384155273438, 0.0557209587097168, 0.05575372695922851, 0.055790592193603515, 0.05578342437744141, 0.05578035354614258, 0.05590323257446289, 0.055731201171875, 0.05590630340576172, 0.055772159576416014, 0.05581721496582031, 0.05579673767089844, 0.05584998321533203, 0.055812095642089846, 0.0557946891784668, 0.05591449737548828, 0.05588684844970703, 0.055831550598144535, 0.05584998321533203, 0.05613363265991211, 0.05586227035522461, 0.055790592193603515, 0.05580799865722656, 0.05578649520874023, 0.05587046432495117, 0.0557946891784668, 0.05585715103149414, 0.0557916145324707, 0.05584691238403321, 0.05583257675170898, 0.05586329650878906, 0.05585919952392578, 0.05588275146484375, 0.055853057861328125, 0.05580083084106445, 0.05595340728759766, 0.05588275146484375, 0.0557844467163086, 0.05586739349365234, 0.05587046432495117, 0.055825408935546876, 0.05580595016479492, 0.05588172912597656, 0.05586841583251953, 0.055831550598144535, 0.055839744567871094, 0.05582233428955078, 0.05582131195068359, 0.05584076690673828, 0.12454399871826172, 0.055744510650634765, 0.055806976318359375, 0.05572710418701172, 0.055826431274414064, 0.05572608184814453, 0.05575987243652344, 0.05573222351074219, 0.055831550598144535, 0.055772159576416014, 0.055736320495605465, 0.05578649520874023, 0.055790592193603515, 0.055723007202148435, 0.05581414413452149, 0.05574348831176758, 0.055772159576416014, 0.05571583938598633, 0.05583462524414062, 0.05577830505371094, 0.055772159576416014, 0.05574655914306641, 0.05588479995727539, 0.055787521362304686, 0.055787521362304686, 0.055798782348632815, 0.05585715103149414, 0.05578956985473633, 0.05578342437744141, 0.055831550598144535, 0.055812095642089846, 0.05575065612792969, 0.05581107330322266, 0.05577318572998047, 0.05585100936889648, 0.055803905487060546, 0.055812095642089846, 0.0557916145324707, 0.05586227035522461, 0.05580287933349609, 0.055769088745117185, 0.055812095642089846, 0.05584998321533203, 0.055766014099121096, 0.05582438278198242, 0.05583769607543945, 0.055818241119384764, 0.05590118408203125, 0.05585715103149414, 0.055828479766845705, 0.05595033645629883, 0.055844863891601565, 0.05581107330322266, 0.05580083084106445, 0.0558653450012207, 0.05584281539916992, 0.05595443344116211, 0.05584998321533203, 0.05580799865722656, 0.055853057861328125, 0.05577318572998047, 0.055803905487060546, 0.055818241119384764, 0.12458598327636719, 0.05576806259155274, 0.05575680160522461, 0.0557916145324707, 0.05617663955688477, 0.05577523040771484, 0.0557496337890625, 0.055771137237548826, 0.0557844467163086, 0.05573427200317383, 0.05581926345825195, 0.055929855346679686, 0.0558551025390625, 0.055806976318359375, 0.055798782348632815, 0.055728126525878906, 0.0557946891784668, 0.05577318572998047, 0.055744510650634765, 0.0557762565612793, 0.055772159576416014, 0.0557496337890625, 0.055841793060302736, 0.055755775451660154, 0.055769088745117185, 0.055782398223876956, 0.05583257675170898, 0.05581619262695312, 0.05586022567749024, 0.05580595016479492, 0.05581926345825195, 0.05580083084106445, 0.05580595016479492, 0.05582950210571289, 0.055815166473388675, 0.055844863891601565, 0.05583462524414062, 0.05582745742797852, 0.05579776000976563, 0.05574553680419922, 0.0557844467163086, 0.055785472869873044, 0.055795711517333986, 0.0557844467163086, 0.05582233428955078, 0.05586227035522461, 0.05587148666381836, 0.05582131195068359, 0.0559370231628418, 0.055825408935546876, 0.05587148666381836, 0.05581414413452149, 0.05578649520874023, 0.055856128692626954, 0.05586739349365234, 0.055809024810791016, 0.055826431274414064, 0.05580799865722656, 0.05578956985473633, 0.055820289611816405, 0.05583359909057617, 0.055810047149658204, 0.05586329650878906, 0.12467711639404297, 0.055820289611816405, 0.055785472869873044, 0.05572403335571289, 0.055782398223876956, 0.05574758529663086, 0.055757823944091796, 0.05576294326782227, 0.05584281539916992, 0.05578035354614258, 0.055801856994628904, 0.05575884628295898, 0.05578342437744141, 0.05572403335571289, 0.055812095642089846, 0.0557496337890625, 0.0562718734741211, 0.05581926345825195, 0.055861248016357425, 0.05574041748046875, 0.055828479766845705, 0.05575167846679688, 0.0558919677734375, 0.055742462158203124, 0.05580083084106445, 0.055785472869873044, 0.05584793472290039, 0.055769088745117185, 0.055836673736572265, 0.05579673767089844, 0.05579264068603516, 0.05577318572998047, 0.05582233428955078, 0.05581926345825195, 0.055810047149658204, 0.0557762565612793, 0.05582438278198242, 0.05582438278198242, 0.05586739349365234, 0.05574860763549805, 0.055815166473388675, 0.05576294326782227, 0.05587558364868164, 0.05578137588500977, 0.055825408935546876, 0.055771137237548826, 0.05580595016479492, 0.0557946891784668, 0.055907329559326174, 0.05586329650878906, 0.05581721496582031, 0.055972862243652347, 0.05586227035522461, 0.05586636734008789, 0.055795711517333986, 0.0557916145324707, 0.05586636734008789, 0.055795711517333986, 0.05584588623046875, 0.05582233428955078, 0.05583052825927735, 0.05580083084106445, 0.05589503860473633, 0.12453990173339843, 0.05574041748046875, 0.05575680160522461, 0.055809024810791016, 0.0557844467163086, 0.05577011108398437, 0.056180736541748044, 0.05583257675170898, 0.05579673767089844, 0.05576704025268555, 0.0558551025390625, 0.05578137588500977, 0.05581107330322266, 0.055779327392578126, 0.055820289611816405, 0.055755775451660154, 0.0557844467163086, 0.05576294326782227, 0.055777278900146485, 0.05581619262695312, 0.05575680160522461, 0.05578956985473633, 0.05586329650878906, 0.055803905487060546, 0.05586022567749024, 0.055777278900146485, 0.055766014099121096, 0.05580492782592773, 0.055858177185058595, 0.055806976318359375, 0.055831550598144535, 0.055782398223876956, 0.055772159576416014, 0.05580083084106445, 0.05586943817138672, 0.055809024810791016, 0.05583769607543945, 0.05579264068603516, 0.0557946891784668, 0.055731201171875, 0.05578137588500977, 0.055809024810791016, 0.05582131195068359, 0.05595238494873047, 0.05582745742797852, 0.055853057861328125, 0.05588787078857422, 0.05581619262695312, 0.055787521362304686, 0.05583257675170898, 0.05588787078857422, 0.05584691238403321, 0.055831550598144535, 0.055856128692626954, 0.05627699279785156, 0.0558766098022461, 0.05584588623046875, 0.055831550598144535, 0.05580799865722656, 0.05583052825927735, 0.055809024810791016, 0.05583564758300781, 0.05584691238403321, 0.12479897308349609, 0.05577830505371094, 0.05582438278198242, 0.05576192092895508, 0.05574655914306641, 0.05575987243652344, 0.0557916145324707, 0.05576192092895508, 0.05581619262695312, 0.05576499176025391, 0.05581107330322266, 0.05576499176025391, 0.055831550598144535, 0.055787521362304686, 0.05582950210571289, 0.05572710418701172, 0.05584076690673828, 0.05575680160522461, 0.055825408935546876, 0.055723007202148435, 0.055841793060302736, 0.05574860763549805, 0.055815166473388675, 0.055744510650634765, 0.055820289611816405, 0.05578342437744141, 0.055858177185058595, 0.055798782348632815, 0.05581107330322266, 0.055782398223876956, 0.05587251281738281, 0.0558551025390625, 0.05586431884765625, 0.05585715103149414, 0.055812095642089846, 0.05580287933349609, 0.05588275146484375, 0.05590323257446289, 0.05586431884765625, 0.05576294326782227, 0.05586022567749024, 0.055760894775390625, 0.05585203170776367, 0.05575884628295898, 0.05586431884765625, 0.055777278900146485, 0.055823360443115234, 0.05583052825927735, 0.055874561309814455, 0.05587148666381836, 0.055790592193603515, 0.05584588623046875, 0.05584588623046875, 0.05587558364868164, 0.05578342437744141, 0.055825408935546876, 0.05590220642089844, 0.05594828796386719, 0.05585919952392578, 0.055787521362304686, 0.05580595016479492, 0.05581107330322266, 0.05586943817138672, 0.12475801849365234, 0.0557496337890625, 0.05581926345825195, 0.05589811325073242, 0.05580287933349609, 0.05578137588500977, 0.055772159576416014, 0.05578137588500977, 0.05575475311279297, 0.05574758529663086, 0.05581414413452149, 0.055793663024902344, 0.05579776000976563, 0.055806976318359375, 0.05585203170776367, 0.05576704025268555, 0.05576499176025391, 0.055812095642089846, 0.05581926345825195, 0.05579776000976563, 0.05581619262695312, 0.0557844467163086, 0.05584588623046875, 0.055787521362304686, 0.05582438278198242, 0.05579980850219726, 0.055790592193603515, 0.055790592193603515, 0.05583052825927735, 0.05582438278198242, 0.05583769607543945, 0.05579980850219726, 0.05581414413452149, 0.05580595016479492, 0.055826431274414064, 0.05581619262695312, 0.05584691238403321, 0.05578035354614258, 0.05590630340576172, 0.055806976318359375, 0.05579264068603516, 0.05581107330322266, 0.055844863891601565, 0.0557946891784668, 0.05581107330322266, 0.055809024810791016, 0.05588275146484375, 0.055815166473388675, 0.055785472869873044, 0.05583462524414062, 0.0558837776184082, 0.05583871841430664, 0.055788543701171874, 0.05585203170776367, 0.055810047149658204, 0.055790592193603515, 0.05580287933349609, 0.055844863891601565, 0.05579673767089844, 0.055831550598144535, 0.055782398223876956, 0.05583052825927735, 0.05590016174316406, 0.12446822357177735, 0.055798782348632815, 0.055782398223876956, 0.05577011108398437, 0.05578035354614258, 0.05575167846679688, 0.05574655914306641, 0.05576396942138672, 0.05580287933349609, 0.05576192092895508, 0.055820289611816405, 0.05574758529663086, 0.05579673767089844, 0.055725055694580077, 0.055809024810791016, 0.055728126525878906, 0.055841793060302736, 0.055744510650634765, 0.05581619262695312, 0.05592166519165039, 0.05584998321533203, 0.05577830505371094, 0.055877632141113284, 0.05575680160522461, 0.05580492782592773, 0.05580492782592773, 0.05587865447998047, 0.055777278900146485, 0.05586431884765625, 0.05670809555053711, 0.05588684844970703, 0.05614080047607422, 0.055812095642089846, 0.05583359909057617, 0.05583769607543945, 0.05582950210571289, 0.05588582229614258, 0.05592268753051758, 0.056118270874023435, 0.05578342437744141, 0.055861248016357425, 0.056174591064453126, 0.05588172912597656, 0.055752704620361325, 0.05583564758300781, 0.05581414413452149, 0.05596979141235352, 0.05585408020019531, 0.05584384155273438, 0.05587558364868164, 0.05581721496582031, 0.05582950210571289, 0.05588479995727539, 0.05593804931640625, 0.05578956985473633, 0.0557844467163086, 0.05588684844970703, 0.05579980850219726, 0.05584691238403321, 0.05581926345825195, 0.0558131217956543, 0.055779327392578126, 0.055839744567871094, 0.12470374298095703, 0.055769088745117185, 0.05576396942138672, 0.055777278900146485, 0.05578342437744141, 0.05576806259155274, 0.05575372695922851, 0.055731201171875, 0.055752704620361325, 0.05572403335571289, 0.05583769607543945, 0.05580083084106445, 0.055795711517333986, 0.05576396942138672, 0.05582233428955078, 0.05577011108398437, 0.0557916145324707, 0.05578035354614258, 0.05583052825927735, 0.05581619262695312, 0.05581414413452149, 0.05578956985473633, 0.055841793060302736, 0.05577830505371094, 0.055798782348632815, 0.055806976318359375, 0.05582745742797852, 0.05582131195068359, 0.05592575836181641, 0.055820289611816405, 0.05586022567749024, 0.055810047149658204, 0.0557844467163086, 0.055815166473388675, 0.0558551025390625, 0.05585203170776367, 0.055874561309814455, 0.055790592193603515, 0.055785472869873044, 0.05573222351074219, 0.05581721496582031, 0.05588479995727539, 0.05584076690673828, 0.05581107330322266, 0.05585203170776367, 0.05581414413452149, 0.0559370231628418, 0.05584384155273438, 0.05582438278198242, 0.05584588623046875, 0.05587148666381836, 0.05585203170776367, 0.055803905487060546, 0.05607219314575195, 0.05584076690673828, 0.055806976318359375, 0.055844863891601565, 0.05585919952392578, 0.05580799865722656, 0.05607731246948242, 0.055861248016357425, 0.05581619262695312, 0.05583564758300781]",tokens/s,17.601104510496405,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -8588,7 +9139,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1450-4f1454266cda34d42e5b3bef;20600d0d-fd15-4b87-b990-bc2a86298c03) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4666-2943069d50fe47724939fb58;3a5d5c4e-08cc-43a1-b80e-ec60d03ab349) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8659,7 +9210,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c11e4-170bfc53025c9689685db7bf;19ef007f-1051-409a-a30d-1860d12b57c9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4444-637c9e4f1ba8eddb78086fb8;a5a8a76d-88b1-409b-a13b-5e334acbd894) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8730,7 +9281,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12aa-01a94595520bf43a54a938d3;671496e0-0f77-47d9-a5d1-ee388ab61bcf) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44e1-7119b9a743197a580a839892;903ca7d3-484e-40e5-8e43-05d058fe3026) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8762,7 +9313,139 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5028.081664,8173.125632,0.0,7526.678528,6833.744896,s,10,3.26147134399414,0.3261471343994141,0.0004409501692482163,0.3262998352050781,0.3266359405517578,0.32663985443115234,0.326642985534668,"[0.3264150085449219, 0.3254004821777344, 0.326524169921875, 0.32663507080078125, 0.32664376831054687, 0.3256611022949219, 0.32650796508789065, 0.3257033386230469, 0.3257957763671875, 0.32618466186523437]",tokens/s,784.92181288489,kWh,3.8486035547376105e-06,2.1088648073688792e-06,2.01936272659971e-05,2.615109562810359e-05,tokens/kWh,9789264.803302791,MB,5028.081664,8173.125632,0.0,7526.678528,7110.584832,s,10,177.0117890625,17.70117890625,0.0018849237304151378,17.7014580078125,17.703122265625,17.704050390625,17.704792890625,"[17.7017734375, 17.6988359375, 17.701845703125, 17.701814453125, 17.702916015625, 17.704978515625, 17.70087109375, 17.701142578125, 17.69869921875, 17.698912109375]",tokens/s,3.5590849814955385,kWh,0.00020889549900607397,0.00011449215478764955,0.0010727857193390082,0.0013961733731327316,tokens/kWh,45123.3358351769,,s,629,179.6651919555665,0.285636235223476,0.03881150448396202,0.28092312622070315,0.28114656982421876,0.28127252807617187,0.6073619189453126,"[0.2810716247558594, 0.28097842407226564, 0.2807705688476563, 0.28086477661132814, 0.28075418090820314, 0.28107284545898437, 0.28095672607421873, 0.2811484069824219, 0.2811556396484375, 0.2810254821777344, 0.2809948120117188, 0.28088116455078127, 0.28087808227539063, 0.28109619140625, 0.2810828857421875, 0.2810512084960938, 0.28083807373046876, 0.2809333801269531, 0.2807726135253906, 0.2809671630859375, 0.2808340454101563, 0.2806732788085938, 0.28127130126953126, 0.2812733459472656, 0.280880126953125, 0.28080947875976564, 0.28092312622070315, 0.280748046875, 0.2808985595703125, 0.2808719482421875, 0.28096307373046875, 0.28107571411132815, 0.280943603515625, 0.28092312622070315, 0.2808719787597656, 0.28081967163085936, 0.2808186950683594, 0.281025634765625, 0.281087890625, 0.281322509765625, 0.28115252685546877, 0.2808883056640625, 0.28102041625976565, 0.28091290283203124, 0.2809108581542969, 0.2808330383300781, 0.2808821716308594, 0.28100811767578127, 0.2808965148925781, 0.2810798034667969, 0.2807940979003906, 0.28092312622070315, 0.2808330383300781, 0.2809354248046875, 0.28095489501953125, 0.2808739929199219, 0.28079513549804686, 0.2808279113769531, 0.2809620361328125, 0.2811197509765625, 0.280922119140625, 0.28079205322265627, 0.6080081787109375, 0.28076544189453123, 0.280838134765625, 0.28076031494140624, 0.28081253051757815, 0.2807132263183594, 0.28089547729492187, 0.2808545227050781, 0.2807480773925781, 0.2809906921386719, 0.280911865234375, 0.28095590209960936, 0.28086578369140625, 0.28108901977539064, 0.2810091552734375, 0.28119961547851563, 0.2811514892578125, 0.280869873046875, 0.28105319213867186, 0.28088626098632813, 0.28103884887695313, 0.28089959716796875, 0.28075518798828125, 0.28093438720703123, 0.28087295532226564, 0.28100927734375, 0.2809219970703125, 0.28087091064453124, 0.280838134765625, 0.2807982177734375, 0.28106341552734376, 0.2812600402832031, 0.28098046875, 0.28088626098632813, 0.2808821716308594, 0.2809999694824219, 0.2808719177246094, 0.2809323425292969, 0.28095590209960936, 0.28091390991210935, 0.280911865234375, 0.2806722412109375, 0.28075726318359373, 0.28089752197265627, 0.280853515625, 0.28107672119140625, 0.2811269226074219, 0.28113018798828127, 0.28134176635742186, 0.28091290283203124, 0.28086480712890627, 0.2808719177246094, 0.2811617431640625, 0.2807859191894531, 0.2810091552734375, 0.28094259643554687, 0.28108901977539064, 0.2808760375976562, 0.28075418090820314, 0.28086477661132814, 0.2808832092285156, 0.28084735107421877, 0.2807152709960937, 0.6078546142578125, 0.28070706176757815, 0.28101324462890626, 0.2809610290527344, 0.2807173156738281, 0.2806036376953125, 0.2808186950683594, 0.2807275390625, 0.2807459716796875, 0.28081765747070314, 0.28090777587890625, 0.28082278442382813, 0.28080331420898436, 0.2808493957519531, 0.28080435180664065, 0.28077981567382815, 0.2807377624511719, 0.28094259643554687, 0.2809610290527344, 0.28080230712890625, 0.2811473999023438, 0.2808586730957031, 0.2810807800292969, 0.2825359497070313, 0.28085760498046874, 0.28076235961914064, 0.280943603515625, 0.28124774169921873, 0.28127847290039065, 0.28103988647460937, 0.28092620849609373, 0.2809927673339844, 0.28098458862304687, 0.2810777587890625, 0.28099993896484377, 0.28095693969726565, 0.28086578369140625, 0.28130712890625, 0.28111053466796876, 0.2808883056640625, 0.2807859191894531, 0.2808258666992188, 0.2808832092285156, 0.28082073974609373, 0.28086782836914065, 0.28088116455078127, 0.28091595458984375, 0.2825093078613281, 0.2810675048828125, 0.2808412170410156, 0.280953857421875, 0.2808637390136719, 0.28111770629882815, 0.280880126953125, 0.2813460388183594, 0.2811760559082031, 0.28108084106445314, 0.28080953979492185, 0.2807018737792969, 0.28099172973632813, 0.28095281982421877, 0.28108389282226565, 0.28077362060546873, 0.608311279296875, 0.28081048583984375, 0.2809405517578125, 0.2806405029296875, 0.28076031494140624, 0.28073370361328126, 0.28086578369140625, 0.2808135681152344, 0.28088116455078127, 0.280952880859375, 0.28101626586914064, 0.2809241638183594, 0.28168295288085937, 0.2808330383300781, 0.28111053466796876, 0.2809692077636719, 0.2810429382324219, 0.28080947875976564, 0.2810460205078125, 0.28084530639648436, 0.2808330383300781, 0.28081253051757815, 0.2807982177734375, 0.2809302978515625, 0.28086477661132814, 0.2809169921875, 0.2808565673828125, 0.280958984375, 0.2808299560546875, 0.28117913818359375, 0.28106341552734376, 0.28076235961914064, 0.280943603515625, 0.28079000854492187, 0.2809374694824219, 0.2807408752441406, 0.2809620361328125, 0.28110452270507813, 0.28091583251953123, 0.2808842163085937, 0.2811627502441406, 0.2810449829101562, 0.280825927734375, 0.2808196411132812, 0.28327322387695314, 0.28112588500976565, 0.2808965148925781, 0.2809743347167969, 0.2809743347167969, 0.28070501708984374, 0.280880126953125, 0.2807848815917969, 0.28094976806640626, 0.28081253051757815, 0.28092312622070315, 0.28089959716796875, 0.28097125244140625, 0.2808299560546875, 0.28092935180664064, 0.2811114807128906, 0.2811207580566406, 0.2812037048339844, 0.28112896728515624, 0.6081239013671875, 0.28064358520507815, 0.281069580078125, 0.2810849304199219, 0.2808268737792969, 0.2807080993652344, 0.2809927673339844, 0.2810705871582031, 0.2809989013671875, 0.2809968566894531, 0.28100302124023435, 0.2809395141601562, 0.280853515625, 0.2807992248535156, 0.28090777587890625, 0.280985595703125, 0.28104193115234377, 0.2819112854003906, 0.28102859497070315, 0.28089959716796875, 0.28095590209960936, 0.28086782836914065, 0.2809641418457031, 0.2810326843261719, 0.2808555603027344, 0.2808555603027344, 0.2807705688476563, 0.28092312622070315, 0.28085964965820315, 0.2809231567382813, 0.2809599609375, 0.28101223754882815, 0.28098458862304687, 0.280995849609375, 0.28102554321289064, 0.2809692077636719, 0.2812591247558594, 0.28121075439453125, 0.28093646240234377, 0.2810716247558594, 0.28083712768554686, 0.28087808227539063, 0.28104806518554687, 0.2810040283203125, 0.28094873046875, 0.2809456787109375, 0.28094680786132814, 0.28089434814453124, 0.281069580078125, 0.28089447021484376, 0.2810828857421875, 0.2811248779296875, 0.2812682189941406, 0.2808401794433594, 0.28089752197265627, 0.2809354248046875, 0.2809825134277344, 0.28154788208007814, 0.2807090148925781, 0.28145867919921874, 0.28125799560546877, 0.28101119995117185, 0.28075726318359373, 0.607119384765625, 0.28083096313476563, 0.28100607299804686, 0.28079928588867187, 0.28110943603515626, 0.2810675048828125, 0.28095489501953125, 0.2810941467285156, 0.28082891845703123, 0.2809333801269531, 0.28096307373046875, 0.28083917236328126, 0.28091494750976564, 0.28112588500976565, 0.28135220336914063, 0.2809314270019531, 0.2810141906738281, 0.28090060424804686, 0.28100506591796875, 0.28073983764648436, 0.2808986206054688, 0.28088619995117187, 0.2810705871582031, 0.28108697509765623, 0.28090264892578126, 0.28086578369140625, 0.28082891845703123, 0.2808259582519531, 0.2808493041992188, 0.2809292907714844, 0.28081048583984375, 0.28089752197265627, 0.28099172973632813, 0.2809354248046875, 0.28101229858398435, 0.2810070495605469, 0.28085861206054685, 0.28089959716796875, 0.28375860595703123, 0.28100811767578127, 0.2811023254394531, 0.2808770446777344, 0.28098458862304687, 0.2810849304199219, 0.2813665771484375, 0.28089950561523436, 0.28091802978515623, 0.280869873046875, 0.28100607299804686, 0.2809487915039062, 0.28107769775390623, 0.28094772338867186, 0.28119961547851563, 0.2809252014160156, 0.2810787658691406, 0.2808965148925781, 0.28122726440429685, 0.28111770629882815, 0.28107571411132815, 0.2812405700683594, 0.2811535339355469, 0.28105523681640626, 0.28082073974609373, 0.607952880859375, 0.28060671997070313, 0.2809661560058594, 0.2806609802246094, 0.2808463439941406, 0.2806804504394531, 0.28079205322265627, 0.2808760375976562, 0.2808392333984375, 0.28082785034179686, 0.2810624084472656, 0.28129486083984373, 0.28122213745117186, 0.2807715759277344, 0.2808842163085937, 0.28091802978515623, 0.2810572814941406, 0.2808985595703125, 0.28099172973632813, 0.2808565673828125, 0.2809518127441406, 0.28080126953125, 0.28089447021484376, 0.2825502624511719, 0.2808985595703125, 0.28085861206054685, 0.280816650390625, 0.280848388671875, 0.28089447021484376, 0.28097125244140625, 0.28085861206054685, 0.28095590209960936, 0.2809477844238281, 0.2808728942871094, 0.2808985595703125, 0.2808268737792969, 0.28089547729492187, 0.28095077514648437, 0.28085861206054685, 0.2809968566894531, 0.2809774169921875, 0.2807982177734375, 0.28089959716796875, 0.28099172973632813, 0.2807982177734375, 0.2809169921875, 0.281101318359375, 0.28097842407226564, 0.2811545715332031, 0.2810491027832031, 0.28117196655273435, 0.2810101623535156, 0.2810798034667969, 0.280890380859375, 0.2809333801269531, 0.28083917236328126, 0.28120578002929686, 0.28119549560546875, 0.2808186950683594, 0.2811064453125, 0.2810828857421875, 0.2810245056152344, 0.28082073974609373, 0.6070692749023437, 0.2808114624023437, 0.28095693969726565, 0.2807562255859375, 0.28089752197265627, 0.280748046875, 0.2808279113769531, 0.28082891845703123, 0.28091494750976564, 0.28085043334960935, 0.2809415588378906, 0.28091390991210935, 0.28088644409179686, 0.2808758850097656, 0.2809313354492188, 0.2808616943359375, 0.2821949462890625, 0.2810019836425781, 0.28163482666015627, 0.2812569580078125, 0.2809241638183594, 0.281143310546875, 0.2809333801269531, 0.2809456787109375, 0.28096511840820315, 0.2810408935546875, 0.28102041625976565, 0.28100506591796875, 0.28084530639648436, 0.28107571411132815, 0.28093438720703123, 0.28092724609375, 0.280890380859375, 0.28085043334960935, 0.28082073974609373, 0.28086782836914065, 0.280764404296875, 0.28092724609375, 0.280929443359375, 0.2809577941894531, 0.2809302978515625, 0.28088934326171877, 0.28073779296875, 0.28088626098632813, 0.2807859191894531, 0.2808637390136719, 0.28086782836914065, 0.2808883056640625, 0.28105419921875, 0.2811023254394531, 0.2812333984375, 0.28084326171875, 0.2809989013671875, 0.28085147094726565, 0.28113101196289064, 0.2810880126953125, 0.280985595703125, 0.28092620849609373, 0.28081048583984375, 0.2810357666015625, 0.28111666870117186, 0.2810572814941406, 0.2808555603027344, 0.6074562377929688, 0.280658935546875, 0.28088729858398437, 0.28080230712890625, 0.2807451171875, 0.28069366455078126, 0.280728515625, 0.28076031494140624, 0.28095077514648437, 0.2811463623046875, 0.28098458862304687, 0.28073883056640625, 0.2808392333984375, 0.2808042907714844, 0.2809539489746094, 0.28077252197265623, 0.28087808227539063, 0.28191949462890625, 0.28118118286132815, 0.2807510986328125, 0.280943603515625, 0.28087091064453124, 0.2808258666992188, 0.28086782836914065, 0.2809599914550781, 0.28128973388671874, 0.28111566162109375, 0.2807930908203125, 0.2808934326171875, 0.2809313354492188, 0.2811085510253906, 0.28094970703125, 0.280985595703125, 0.2808299560546875, 0.280943603515625, 0.2810040283203125, 0.28105624389648437, 0.28103884887695313, 0.28102859497070315, 0.2809405517578125, 0.280806396484375, 0.28071218872070314, 0.2808770446777344, 0.2809866333007813, 0.280911865234375, 0.28092825317382814, 0.2808463439941406, 0.280869873046875, 0.28091494750976564, 0.2807848815917969, 0.2809037780761719, 0.2807929992675781, 0.2809415588378906, 0.28072653198242187, 0.28112896728515624, 0.2808504638671875, 0.2809825134277344, 0.2808032836914062, 0.2806886291503906, 0.28093954467773435, 0.2810111694335938, 0.28101837158203125, 0.2809968566894531, 0.607963134765625, 0.28109722900390627, 0.28084225463867185, 0.2807828369140625, 0.28078900146484376, 0.28078900146484376, 0.28086383056640624, 0.28084625244140626, 0.2809241638183594, 0.2809098205566406, 0.28088525390625, 0.28086782836914065, 0.28103271484375, 0.28082382202148437, 0.28085964965820315, 0.28094158935546876, 0.2811105041503906, 0.28084225463867185, 0.28108901977539064, 0.2807848815917969, 0.28083917236328126, 0.28080230712890625, 0.28073779296875, 0.28079718017578126, 0.2807562255859375, 0.2808330383300781, 0.2826495971679687, 0.2808985595703125, 0.280853515625, 0.28077566528320314, 0.28073883056640625, 0.2808842163085937, 0.2811269226074219, 0.2812651672363281, 0.28103167724609374, 0.28106341552734376, 0.2807930908203125, 0.2809354248046875, 0.2808637390136719, 0.2806599731445312, 0.28091494750976564, 0.28076446533203125, 0.2810346984863281, 0.2811269226074219, 0.2809743347167969, 0.280764404296875, 0.28090167236328123, 0.280933349609375, 0.2809263000488281, 0.28092303466796875, 0.28106341552734376, 0.28090060424804686, 0.2809794616699219, 0.280806396484375, 0.2810941467285156, 0.28091290283203124, 0.2808883056640625, 0.280875, 0.28072357177734375, 0.28083596801757815, 0.2809948120117188, 0.28083712768554686, 0.28070196533203123]",tokens/s,3.5009563797731067,, +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 117, in run + _ = backend.generate(self.inputs, self.config.generate_kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 369, in generate + return self.pretrained_model.generate(**inputs, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1736, in generate + result = self._sample( + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2375, in _sample + outputs = self( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 900, in forward + transformer_outputs = self.transformer( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 797, in forward + outputs = block( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 453, in forward + attn_outputs = self.self_attention( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 291, in forward + fused_qkv = self.query_key_value(hidden_states) # [batch_size, seq_length, 3 x hidden_size] + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/auto_gptq/nn_modules/qlinear/qlinear_cuda_old.py"", line 316, in forward + weight = weight.reshape(-1, self.group_size, weight.shape[2]) +RuntimeError: shape '[-1, 128, 4672]' is invalid for input of size 21229568 + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,,,MB,4753.522688,21498.42944,0.0,20851.982336,20236.731392,s,10,15.791625,1.5791625,0.0006504668109249951,1.579173889160156,1.5798716064453124,1.5800453002929689,1.5801842553710939,"[1.57879541015625, 1.5780565185546875, 1.5798330078125, 1.5797557373046875, 1.580218994140625, 1.57864208984375, 1.5789478759765625, 1.57939990234375, 1.578477294921875, 1.5794981689453125]",tokens/s,162.11124567610997,kWh,1.8641048173109693e-05,1.0215338307862111e-05,9.709499434260027e-05,0.00012595138082357207,tokens/kWh,2032530.3170641307,MB,4754.120704,21498.42944,0.0,20851.982336,20339.707392,s,10,881.8083671875002,88.18083671874999,0.002079438758836812,88.1809140625,88.18330625,88.18352421875001,88.18369859375001,"[88.1769296875, 88.179453125, 88.1832578125, 88.1811796875, 88.18275, 88.178796875, 88.1821796875, 88.1837421875, 88.1794296875, 88.1806484375]",tokens/s,0.7144409414137963,kWh,0.001040938318769137,0.0005705258692486495,0.0053753298280382,0.0069867940160559866,tokens/kWh,9017.011215046985,,s,629,894.6216553955076,1.4222919799610616,0.18759022057657465,1.39966162109375,1.400088818359375,1.4002253417968749,2.978684599609375,"[1.3990267333984374, 1.399194580078125, 1.39926318359375, 1.399057373046875, 1.39915771484375, 1.3993861083984376, 1.3993082275390625, 1.3993277587890625, 1.3993349609375, 1.3990625, 1.3992550048828125, 1.399257080078125, 1.3994239501953125, 1.3993123779296874, 1.3992008056640626, 1.399593017578125, 1.39961962890625, 1.3998203125, 1.3995447998046875, 1.39936865234375, 1.399487548828125, 1.399593994140625, 1.3995611572265625, 1.399668701171875, 1.39974658203125, 1.3998499755859375, 1.399363525390625, 1.39985302734375, 1.399152587890625, 1.399531494140625, 1.39940771484375, 1.399680908203125, 1.39953564453125, 1.399763916015625, 1.3996451416015625, 1.3997568359375, 1.3997650146484375, 1.39991455078125, 1.3995980224609375, 1.3999871826171875, 1.399814208984375, 1.40004248046875, 1.3995816650390625, 1.4001510009765625, 1.39960009765625, 1.400014892578125, 1.4001336669921876, 1.399931884765625, 1.3994854736328124, 1.40003125, 1.3995704345703126, 1.40002197265625, 1.399784423828125, 1.3999564208984374, 1.3997066650390626, 1.400405029296875, 1.3998858642578125, 1.399877685546875, 1.399904296875, 1.3998837890625, 1.3997188720703124, 1.4000732421875, 2.978333740234375, 1.3989600830078126, 1.39924169921875, 1.3988751220703124, 1.399287841796875, 1.3991710205078125, 1.399235595703125, 1.399609375, 1.399541748046875, 1.3992969970703124, 1.39930322265625, 1.399277587890625, 1.39953564453125, 1.399251953125, 1.3993912353515625, 1.399419921875, 1.3995018310546874, 1.3994639892578125, 1.399394287109375, 1.399310302734375, 1.3996768798828125, 1.399478271484375, 1.3996656494140625, 1.3994024658203126, 1.39964208984375, 1.3995662841796874, 1.3995765380859375, 1.39959912109375, 1.399552978515625, 1.3994024658203126, 1.3995858154296874, 1.39936669921875, 1.3997916259765626, 1.3996544189453124, 1.3998448486328126, 1.3994271240234375, 1.3996441650390625, 1.39966162109375, 1.399943115234375, 1.399889892578125, 1.4001827392578126, 1.39974658203125, 1.400034423828125, 1.3994228515625, 1.399857177734375, 1.399530517578125, 1.40224609375, 1.3999288330078126, 1.4000067138671874, 1.400026123046875, 1.400090576171875, 1.3997403564453126, 1.39998828125, 1.3997435302734376, 1.4000753173828124, 1.3996533203125, 1.4001356201171875, 1.39987353515625, 1.4000650634765626, 1.399835693359375, 1.40000048828125, 1.39968408203125, 1.4000137939453126, 2.979962890625, 1.39926220703125, 1.3990328369140625, 1.399584716796875, 1.3994219970703126, 1.3992283935546874, 1.3991546630859375, 1.3993482666015624, 1.3992652587890626, 1.39936669921875, 1.399351318359375, 1.3993861083984376, 1.399299072265625, 1.3993707275390626, 1.3993062744140625, 1.3993953857421875, 1.39961962890625, 1.399415771484375, 1.399525390625, 1.39930517578125, 1.39949365234375, 1.3997568359375, 1.3993184814453126, 1.3997650146484375, 1.3997353515625, 1.399690185546875, 1.4000291748046876, 1.399877685546875, 1.399784423828125, 1.3996329345703125, 1.3998212890625, 1.3994864501953126, 1.4001602783203124, 1.3998140869140625, 1.3999124755859376, 1.3998775634765626, 1.4000035400390625, 1.3994906005859375, 1.399784423828125, 1.39968408203125, 1.3998192138671874, 1.399498779296875, 1.4001407470703124, 1.3995120849609375, 1.3999288330078126, 1.3997454833984375, 1.39981103515625, 1.4027008056640624, 1.3998243408203126, 1.39981103515625, 1.4002513427734375, 1.39980078125, 1.3999759521484374, 1.399794677734375, 1.40002001953125, 1.39987353515625, 1.3999278564453126, 1.399804931640625, 1.400158203125, 1.399793701171875, 1.4000107421875, 1.400045654296875, 1.400123291015625, 2.980041748046875, 1.39945166015625, 1.399203857421875, 1.3994495849609374, 1.3992274169921874, 1.39907177734375, 1.399314453125, 1.399120849609375, 1.39966162109375, 1.399363525390625, 1.3993165283203124, 1.399487548828125, 1.4004613037109375, 1.3995233154296876, 1.399342041015625, 1.399341064453125, 1.3994598388671875, 1.3993349609375, 1.3996513671875, 1.3996400146484376, 1.399520263671875, 1.3995191650390626, 1.3995377197265626, 1.399234619140625, 1.3996871337890624, 1.4000189208984375, 1.3999349365234375, 1.39932470703125, 1.399783447265625, 1.4000445556640626, 1.399647216796875, 1.399568359375, 1.399568359375, 1.399731201171875, 1.399889892578125, 1.39970556640625, 1.39993603515625, 1.3996287841796875, 1.3998807373046875, 1.3998387451171874, 1.399732177734375, 1.3995458984375, 1.3997874755859374, 1.3995592041015625, 1.400785888671875, 1.3997117919921875, 1.3998121337890626, 1.3999656982421875, 1.3996810302734375, 1.3996943359375, 1.39987353515625, 1.399900146484375, 1.4000506591796875, 1.3997332763671875, 1.3998817138671875, 1.3996072998046876, 1.40021044921875, 1.39983251953125, 1.400130615234375, 1.4000435791015624, 1.400005615234375, 1.3997353515625, 1.4001796875, 2.9802998046875, 1.399120849609375, 1.399341064453125, 1.399108642578125, 1.39905126953125, 1.3995653076171874, 1.39922119140625, 1.399256103515625, 1.3993154296875, 1.3994423828125, 1.3993594970703125, 1.399393310546875, 1.3994976806640624, 1.399383056640625, 1.3993502197265626, 1.399436279296875, 1.399673828125, 1.401808837890625, 1.399751708984375, 1.3995008544921874, 1.39957958984375, 1.3996103515625, 1.3996278076171875, 1.3993154296875, 1.39953564453125, 1.399587890625, 1.3997158203125, 1.3994547119140626, 1.399677978515625, 1.39930322265625, 1.3997813720703125, 1.3996236572265626, 1.3997977294921875, 1.3994639892578125, 1.399583740234375, 1.3996553955078126, 1.39995751953125, 1.3994649658203124, 1.3997015380859374, 1.3995806884765625, 1.4000537109375, 1.3995079345703125, 1.40003125, 1.3995447998046875, 1.4000128173828126, 1.3996851806640624, 1.3999759521484374, 1.3998397216796874, 1.40012646484375, 1.399877685546875, 1.400088623046875, 1.3997568359375, 1.4000291748046876, 1.3996800537109375, 1.4001787109375, 1.399815185546875, 1.40000048828125, 1.4016307373046875, 1.400141845703125, 1.4000137939453126, 1.400014892578125, 1.3999698486328125, 1.4001131591796876, 2.978821044921875, 1.399046142578125, 1.3991793212890624, 1.3989376220703125, 1.39928369140625, 1.3996728515625, 1.3994056396484376, 1.399405517578125, 1.3993594970703125, 1.3993656005859374, 1.3992847900390626, 1.3991884765625, 1.3991322021484376, 1.3994342041015626, 1.3993819580078124, 1.3996124267578125, 1.39962060546875, 1.39949365234375, 1.39947216796875, 1.3994957275390625, 1.3995653076171874, 1.399456787109375, 1.3994700927734376, 1.3995662841796874, 1.399562255859375, 1.3996544189453124, 1.399804931640625, 1.39970361328125, 1.3997813720703125, 1.39945068359375, 1.3997384033203124, 1.399419921875, 1.3995765380859375, 1.3994117431640625, 1.400079345703125, 1.3999482421875, 1.399835693359375, 1.399540771484375, 1.399688232421875, 1.399657470703125, 1.3998336181640625, 1.3995345458984374, 1.39985205078125, 1.399656494140625, 1.3999964599609376, 1.39978857421875, 1.3996011962890624, 1.39962060546875, 1.3998602294921876, 1.3998612060546876, 1.4000189208984375, 1.3999698486328125, 1.3998602294921876, 1.399774169921875, 1.40016748046875, 1.399921630859375, 1.4004654541015624, 1.39996875, 1.4002933349609374, 1.3999073486328124, 1.4000814208984376, 1.3998212890625, 1.4002340087890626, 2.97885595703125, 1.3990830078125, 1.3992325439453126, 1.39894482421875, 1.3992386474609375, 1.3991240234375, 1.399192626953125, 1.399552978515625, 1.3998336181640625, 1.3991290283203126, 1.3995643310546875, 1.3993369140625, 1.3992908935546875, 1.3994813232421874, 1.3995396728515626, 1.3994075927734375, 1.399647216796875, 1.3993912353515625, 1.399774169921875, 1.399404541015625, 1.3995263671875, 1.399309326171875, 1.39955810546875, 1.4027039794921874, 1.3995570068359375, 1.3995970458984375, 1.399773193359375, 1.3994434814453125, 1.39992578125, 1.3998408203125, 1.39961962890625, 1.3993533935546876, 1.39991455078125, 1.399616455078125, 1.3995592041015625, 1.3996400146484376, 1.3997864990234374, 1.39940966796875, 1.399609375, 1.39965234375, 1.3998223876953124, 1.39985205078125, 1.4000179443359375, 1.399384033203125, 1.40019921875, 1.3997762451171876, 1.3997117919921875, 1.399884765625, 1.4000496826171875, 1.3995694580078124, 1.4001171875, 1.399783447265625, 1.40018896484375, 1.3996298828125, 1.400130615234375, 1.3999932861328126, 1.400088623046875, 1.4003466796875, 1.400564697265625, 1.3999073486328124, 1.399972900390625, 1.399690185546875, 1.400058837890625, 2.979527587890625, 1.3993922119140625, 1.39902978515625, 1.398983642578125, 1.39919775390625, 1.398992919921875, 1.3992396240234375, 1.3993533935546876, 1.3992376708984375, 1.3990655517578126, 1.399604248046875, 1.3992969970703124, 1.3993492431640624, 1.3994608154296875, 1.39930419921875, 1.3992017822265626, 1.401822265625, 1.39940869140625, 1.3995540771484376, 1.3994659423828124, 1.39966162109375, 1.3993226318359375, 1.3994906005859375, 1.39947216796875, 1.3996871337890624, 1.399563232421875, 1.39978955078125, 1.3996298828125, 1.399773193359375, 1.3996103515625, 1.3997169189453125, 1.3991597900390624, 1.3995765380859375, 1.399605224609375, 1.3999544677734375, 1.399837646484375, 1.3997568359375, 1.3994127197265624, 1.3997230224609376, 1.3997178955078124, 1.400048583984375, 1.399815185546875, 1.4001356201171875, 1.3998172607421875, 1.4001346435546875, 1.3995457763671875, 1.4000435791015624, 1.3997373046875, 1.3998131103515625, 1.3997435302734376, 1.4027857666015624, 1.400364013671875, 1.4001920166015625, 1.39983056640625, 1.4000352783203125, 1.4002769775390624, 1.400194091796875, 1.399773193359375, 1.4002216796875, 1.39990625, 1.4000352783203125, 1.399901123046875, 1.3998612060546876, 2.978325439453125, 1.3991455078125, 1.3992120361328124, 1.3990419921875, 1.3993450927734374, 1.3994271240234375, 1.3989150390625, 1.3993011474609376, 1.3993082275390625, 1.3992078857421875, 1.39926220703125, 1.399310302734375, 1.3993021240234376, 1.399226318359375, 1.3992540283203125, 1.3992283935546874, 1.399677978515625, 1.4018785400390625, 1.399798828125, 1.3992427978515625, 1.3995345458984374, 1.3994066162109375, 1.3994649658203124, 1.39955712890625, 1.3995816650390625, 1.399498779296875, 1.4000189208984375, 1.3997578125, 1.400258544921875, 1.39966162109375, 1.399711669921875, 1.399426025390625, 1.399445556640625, 1.3996800537109375, 1.3996759033203126, 1.3994976806640624, 1.400227783203125, 1.39970458984375, 1.399803955078125, 1.3998212890625, 1.399910400390625, 1.399972900390625, 1.399868408203125, 1.399245849609375, 1.40001171875, 1.3994486083984374, 1.399951416015625, 1.3996124267578125, 1.399826416015625, 1.3997291259765625, 1.400196044921875, 1.3996390380859376, 1.400005615234375, 1.3996943359375, 1.4000947265625, 1.399920654296875, 1.3999810791015626, 1.399846923828125, 1.40007421875, 1.3998878173828124, 1.400089599609375, 1.399841796875, 1.3999400634765624, 2.979295166015625, 1.3990625, 1.399267333984375, 1.39909521484375, 1.399299072265625, 1.3992110595703124, 1.399098388671875, 1.3994393310546875, 1.3993287353515624, 1.39915576171875, 1.39934619140625, 1.3993994140625, 1.399341064453125, 1.399320556640625, 1.3994598388671875, 1.3993870849609376, 1.3995867919921876, 1.399257080078125, 1.3995181884765624, 1.399752685546875, 1.3998980712890625, 1.3995858154296874, 1.39955615234375, 1.3992734375, 1.3995673828125, 1.39959912109375, 1.4034248046875, 1.3997086181640626, 1.39966162109375, 1.3998623046875, 1.399962646484375, 1.3994066162109375, 1.3994639892578125, 1.3996851806640624, 1.399741455078125, 1.3994925537109375, 1.3996871337890624, 1.3995704345703126, 1.40006396484375, 1.39987353515625, 1.399795654296875, 1.3996441650390625, 1.3998203125, 1.3995867919921876, 1.4001162109375, 1.3995396728515626, 1.3999329833984375, 1.3997076416015626, 1.3996749267578126, 1.3995120849609375, 1.3998345947265625, 1.39978955078125, 1.4000609130859376, 1.399593017578125, 1.39999951171875, 1.3997137451171875, 1.400164306640625, 1.4000711669921875, 1.4002811279296874, 1.3999185791015625, 1.399836669921875, 1.3998961181640626, 1.3999913330078124]",tokens/s,0.7030905145280909,, +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status + response.raise_for_status() + File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status + raise HTTPError(http_error_msg, response=self) +requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error + metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata + r = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper + response = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper + hf_raise_for_status(response) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status + raise HfHubHTTPError(message, response=response) from e +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4a9f-55071c83759c874764b9062d;5fabd602-f57d-47e2-b950-8eb15ba27650) + +403 Forbidden: Authorization error.. +Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. +If you are trying to create or update content,make sure you have a token with the `write` role. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 399, in cached_file + resolved_file = hf_hub_download( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download + return _hf_hub_download_to_cache_dir( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir + _raise_on_head_call_error(head_call_error, force_download, local_files_only) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error + raise LocalEntryNotFoundError( +huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ + super().__init__(config) + File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ + self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) + File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config + return AutoConfig.from_pretrained(model, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict + resolved_config_file = cached_file( + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 442, in cached_file + raise EnvironmentError( +OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. +Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -8802,7 +9485,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12e2-54a4358e723187c36629b3b2;21920b46-22d4-4c47-a388-fe4c319a3886) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d450a-06d4d980781d77d37018ef68;a534f14d-f060-4988-bb34-45c66d982990) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8864,7 +9547,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee50-5d530d9a66b70ef6419c8a0d;a1c168f6-053d-4263-a910-53c9d9a2c9c7) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3f61-71cc121a416fcf946b72ace6;df34207f-375a-46f5-8a21-000356223800) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -8991,7 +9674,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1270-5a4762e65ab33ab504a5dec4;a0d45a80-ba9a-45dc-8a24-176d34ee0982) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44b9-043667f835d0e624568e604d;5010a1ce-2b83-453a-a3d5-62061f9f56e9) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9064,7 +9747,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664beed4-53a174062aff8e441c1e5b43;71500148-38fa-4bac-9f7a-87b5cbd3eedb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d3fd6-0968a52b675dbd201a94d1e1;564c0bf2-4777-495a-9c62-5124dfdcc282) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9137,7 +9820,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1401-04f635b47cd3a1f100594e47;3adab9ea-a60f-48aa-be2a-4d1e28191d4e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4618-278eafba595634dc08dbc87c;8ec471fc-e1df-4ae1-adc4-cc20ddc5bbb2) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9167,6 +9850,63 @@ Traceback (most recent call last): OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 117, in run + _ = backend.generate(self.inputs, self.config.generate_kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 369, in generate + return self.pretrained_model.generate(**inputs, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1736, in generate + result = self._sample( + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2375, in _sample + outputs = self( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 900, in forward + transformer_outputs = self.transformer( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 797, in forward + outputs = block( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 477, in forward + mlp_output = self.mlp(mlp_layernorm_out) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 409, in forward + x = self.act(self.dense_h_to_4h(x)) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/auto_gptq/nn_modules/qlinear/qlinear_cuda_old.py"", line 314, in forward + ).to(torch.int16 if self.bits == 8 else torch.int8) +torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 MiB. GPU + ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,6210.260992,7055.343616,0.0,6408.896512,6098.739712,s,10,4.279645782470704,0.42796457824707035,0.0005593245447652133,0.4277575378417969,0.4286687561035156,0.42885357360839843,0.4290014276123047,"[0.4275655212402344, 0.42783905029296876, 0.42729730224609375, 0.42840765380859375, 0.428627685546875, 0.427676025390625, 0.42752056884765627, 0.428287353515625, 0.42738623046875, 0.42903839111328124]",tokens/s,598.1803471880035,kWh,5.048784510129028e-06,2.7665064322491163e-06,2.4794672613500057e-05,3.26099635558782e-05,tokens/kWh,7850361.425928487,MB,6210.260992,7055.343616,0.0,6408.896512,6157.953024,s,10,235.09618749999998,23.509618749999998,0.0012595072409445369,23.509406249999998,23.5114091796875,23.51148388671875,23.51154365234375,"[23.50916796875, 23.509130859375, 23.507298828125, 23.508712890625, 23.508748046875, 23.51155859375, 23.511392578125, 23.509669921875, 23.51086328125, 23.50964453125]",tokens/s,2.679754217622096,kWh,0.00027756352315760327,0.0001521284417659126,0.0013340981089442967,0.0017637900738678128,tokens/kWh,35718.53642528296,,s,629,238.5749162597658,0.3792923946896115,0.05084076188471288,0.3731507263183594,0.37334281005859377,0.3735038024902344,0.8008048754882812,"[0.3732080688476562, 0.3730780029296875, 0.3730810852050781, 0.37292440795898435, 0.37326849365234377, 0.37307186889648436, 0.37314663696289063, 0.37291622924804685, 0.3731578979492188, 0.3729930114746094, 0.3734118347167969, 0.37305035400390624, 0.3731650695800781, 0.37292340087890624, 0.37317120361328127, 0.3731005554199219, 0.3731650695800781, 0.37304934692382813, 0.37309747314453123, 0.3731312561035156, 0.3733893127441406, 0.37327462768554687, 0.3730882568359375, 0.3729346618652344, 0.37309439086914065, 0.37311386108398437, 0.37305856323242187, 0.3731015625, 0.3731005554199219, 0.37298175048828125, 0.37313638305664065, 0.3729284973144531, 0.37317633056640626, 0.37328897094726565, 0.3731036071777344, 0.37316094970703123, 0.37328076171875, 0.3730554809570312, 0.37310565185546873, 0.3731179504394531, 0.3733811340332031, 0.3731671142578125, 0.37309133911132814, 0.3732705383300781, 0.3731312561035156, 0.3731097717285156, 0.37328076171875, 0.37323468017578126, 0.3732152404785156, 0.3732725830078125, 0.3731374206542969, 0.3732234802246094, 0.37307589721679685, 0.37328793334960936, 0.37320703125, 0.37330841064453124, 0.37322445678710936, 0.37333810424804686, 0.37332171630859373, 0.37317733764648436, 0.3731732482910156, 0.3732490234375, 0.8009871215820312, 0.37333505249023435, 0.37299404907226563, 0.3729971313476563, 0.37340570068359374, 0.37328793334960936, 0.37288858032226563, 0.37309133911132814, 0.3729848327636719, 0.37314764404296874, 0.37289166259765627, 0.3730513916015625, 0.3732152404785156, 0.37308621215820315, 0.3730554809570312, 0.37328591918945314, 0.37298480224609376, 0.3728322448730469, 0.37288858032226563, 0.37303909301757815, 0.3731015625, 0.3734876098632812, 0.3730401306152344, 0.37324594116210935, 0.373106689453125, 0.3731865539550781, 0.37314151000976564, 0.37302987670898435, 0.3732254638671875, 0.37307186889648436, 0.37301043701171877, 0.37304217529296874, 0.37334527587890626, 0.37311590576171877, 0.3730882568359375, 0.373185546875, 0.37310565185546873, 0.37387469482421876, 0.37318246459960935, 0.37328793334960936, 0.3730780029296875, 0.3730687866210938, 0.3731885986328125, 0.3731087341308594, 0.37335556030273437, 0.3732705078125, 0.3730831298828125, 0.3730677795410156, 0.3733760070800781, 0.37326641845703123, 0.37328897094726565, 0.37296844482421876, 0.3733094482421875, 0.3731671142578125, 0.37319064331054685, 0.3732940673828125, 0.3732715454101562, 0.3733114929199219, 0.3732971496582031, 0.3731578979492188, 0.37335140991210936, 0.37319064331054685, 0.37305035400390624, 0.8007680053710937, 0.3731671142578125, 0.37287115478515626, 0.3731046447753906, 0.37296435546875, 0.37311590576171877, 0.3727196044921875, 0.3730145263671875, 0.3731374206542969, 0.3730985107421875, 0.3728568420410156, 0.37312716674804686, 0.3730053100585938, 0.37314047241210935, 0.3730554809570312, 0.37310772705078127, 0.3731169128417969, 0.3729151916503906, 0.372927490234375, 0.37301657104492186, 0.37318350219726565, 0.373501953125, 0.3730124816894531, 0.3735449523925781, 0.37311590576171877, 0.373043212890625, 0.37314251708984375, 0.3731589050292969, 0.37302783203125, 0.37303909301757815, 0.37303704833984375, 0.3731619873046875, 0.3731046447753906, 0.37303091430664065, 0.37307186889648436, 0.37319985961914065, 0.3731179504394531, 0.3730350036621094, 0.37315277099609373, 0.3731312561035156, 0.37315277099609373, 0.37339340209960936, 0.3731015625, 0.37323776245117185, 0.3731599426269531, 0.3731128234863281, 0.3731937255859375, 0.37312408447265627, 0.3732305908203125, 0.37319985961914065, 0.3733421936035156, 0.3732418518066406, 0.3733534851074219, 0.373148681640625, 0.3731650695800781, 0.3731537780761719, 0.37313433837890625, 0.37315277099609373, 0.37320294189453124, 0.37319168090820315, 0.3732234191894531, 0.37321319580078127, 0.3731814270019531, 0.8016783447265625, 0.3732008972167969, 0.37310772705078127, 0.37303604125976564, 0.37309133911132814, 0.3731507263183594, 0.37292544555664064, 0.3730473022460937, 0.37297357177734375, 0.3731589050292969, 0.37297457885742186, 0.37307699584960935, 0.37298892211914064, 0.3731087341308594, 0.37291928100585936, 0.37309747314453123, 0.3734814758300781, 0.3731107788085937, 0.37312921142578126, 0.3731107788085937, 0.373064697265625, 0.3731005554199219, 0.3730677795410156, 0.37321624755859373, 0.37303192138671876, 0.3729459228515625, 0.37312921142578126, 0.37362380981445314, 0.3736708984375, 0.3731230773925781, 0.3730134887695313, 0.37328897094726565, 0.3731169128417969, 0.3731312561035156, 0.3731087341308594, 0.37309234619140624, 0.373032958984375, 0.3732090759277344, 0.37307699584960935, 0.37311898803710936, 0.37305856323242187, 0.37314151000976564, 0.37311386108398437, 0.37317938232421877, 0.3735777282714844, 0.373285888671875, 0.3730687866210938, 0.37312716674804686, 0.37326028442382814, 0.37316915893554686, 0.37328179931640626, 0.37309439086914065, 0.37309747314453123, 0.3731374206542969, 0.37349169921875, 0.37324798583984375, 0.37313330078125, 0.37298379516601565, 0.3733196716308594, 0.3731261291503906, 0.37324697875976565, 0.3732305908203125, 0.3732357177734375, 0.8020367431640625, 0.3730831298828125, 0.37295513916015627, 0.37298074340820314, 0.37294488525390623, 0.37317529296875, 0.3729490051269531, 0.373243896484375, 0.37304934692382813, 0.3731097717285156, 0.3729971313476563, 0.37312716674804686, 0.37314663696289063, 0.3731455993652344, 0.37314968872070314, 0.37305035400390624, 0.37300634765625, 0.37323162841796875, 0.37325006103515623, 0.3732787170410156, 0.37298379516601565, 0.3730206604003906, 0.3730124816894531, 0.3732725830078125, 0.3729930114746094, 0.37306369018554686, 0.37321829223632813, 0.373106689453125, 0.37307699584960935, 0.37313534545898436, 0.3731169128417969, 0.3728875427246094, 0.3730513916015625, 0.37305856323242187, 0.3730903015136719, 0.37302886962890625, 0.37321829223632813, 0.37333709716796876, 0.37323876953125, 0.3732428894042969, 0.3732152404785156, 0.37314663696289063, 0.3731722106933594, 0.37318246459960935, 0.37324594116210935, 0.3732203369140625, 0.3731445617675781, 0.3730882568359375, 0.3731036071777344, 0.37331045532226564, 0.3732971496582031, 0.37316915893554686, 0.3732838439941406, 0.373170166015625, 0.3732203369140625, 0.37314764404296874, 0.3732561950683594, 0.3732285461425781, 0.37336474609375, 0.3734241333007812, 0.3732285461425781, 0.37331558227539063, 0.37341082763671873, 0.8008192138671875, 0.3731537780761719, 0.3729490051269531, 0.3731722106933594, 0.3728486328125, 0.37320498657226564, 0.3729776611328125, 0.37345382690429685, 0.37316299438476563, 0.37307699584960935, 0.3730483093261719, 0.37314968872070314, 0.37296435546875, 0.37352346801757813, 0.3729213562011719, 0.3730483093261719, 0.37297357177734375, 0.3730882568359375, 0.37315277099609373, 0.3732305908203125, 0.37476864624023437, 0.3735941162109375, 0.3732490234375, 0.37320294189453124, 0.37308416748046874, 0.37313433837890625, 0.37309234619140624, 0.3731558532714844, 0.37313433837890625, 0.3730145263671875, 0.373001220703125, 0.37291622924804685, 0.3731251220703125, 0.3731097717285156, 0.3731036071777344, 0.37313330078125, 0.37301657104492186, 0.37316915893554686, 0.37326028442382814, 0.3732715454101562, 0.37318246459960935, 0.3732090759277344, 0.3732561950683594, 0.37324594116210935, 0.3733391418457031, 0.37328485107421877, 0.3731107788085937, 0.37339852905273435, 0.37339544677734376, 0.37350503540039065, 0.37321829223632813, 0.37319781494140625, 0.3731671142578125, 0.3730677795410156, 0.3731875915527344, 0.3731445617675781, 0.3732213745117188, 0.37312408447265627, 0.3731660766601563, 0.37319476318359374, 0.3737794494628906, 0.37323263549804686, 0.373138427734375, 0.8008683471679687, 0.3731507263183594, 0.3730595703125, 0.37318350219726565, 0.372990966796875, 0.3731230773925781, 0.37294488525390623, 0.3730247802734375, 0.3730350036621094, 0.37301556396484375, 0.37292645263671875, 0.37320703125, 0.3731435546875, 0.3733483581542969, 0.37306369018554686, 0.37319064331054685, 0.37317733764648436, 0.37296743774414065, 0.3731036071777344, 0.3730093994140625, 0.37298587036132813, 0.3731640319824219, 0.37479833984375, 0.37319064331054685, 0.3733125, 0.3731199951171875, 0.3731148681640625, 0.373222412109375, 0.37292645263671875, 0.37314047241210935, 0.3729407958984375, 0.3731517333984375, 0.3729623107910156, 0.3731589050292969, 0.3731578979492188, 0.373185546875, 0.373222412109375, 0.3731374206542969, 0.3735080871582031, 0.37351629638671874, 0.37323876953125, 0.37323468017578126, 0.3731015625, 0.3731660766601563, 0.3731865539550781, 0.373096435546875, 0.3731885986328125, 0.3730247802734375, 0.3732367248535156, 0.37322750854492187, 0.373796875, 0.3730247802734375, 0.37314663696289063, 0.3732643737792969, 0.37317428588867185, 0.37375588989257813, 0.37332992553710936, 0.3731722106933594, 0.3731589050292969, 0.37317938232421877, 0.373212158203125, 0.37326849365234377, 0.37335244750976565, 0.80138134765625, 0.37311590576171877, 0.3728486328125, 0.3731302490234375, 0.3730831298828125, 0.3731087341308594, 0.3730544738769531, 0.3732490234375, 0.37309439086914065, 0.3732695007324219, 0.3730985107421875, 0.37333505249023435, 0.37291622924804685, 0.37313638305664065, 0.3729623107910156, 0.372990966796875, 0.37310772705078127, 0.37303604125976564, 0.37300634765625, 0.3731773681640625, 0.37315582275390624, 0.37308721923828125, 0.3732561950683594, 0.373043212890625, 0.37327667236328127, 0.3731517333984375, 0.373106689453125, 0.373212158203125, 0.3733309326171875, 0.3735859069824219, 0.373148681640625, 0.37302169799804685, 0.37313638305664065, 0.37306674194335937, 0.3730452575683594, 0.3731445617675781, 0.3729623107910156, 0.37312820434570315, 0.3735152587890625, 0.3732142028808594, 0.37327462768554687, 0.3731619873046875, 0.37321115112304687, 0.37318450927734376, 0.3730882568359375, 0.3731927185058594, 0.3731322937011719, 0.3731619873046875, 0.37333197021484377, 0.37327462768554687, 0.37304934692382813, 0.37325515747070315, 0.37335244750976565, 0.3730247802734375, 0.3731005554199219, 0.37331353759765623, 0.37335140991210936, 0.37335757446289064, 0.3732643737792969, 0.37328692626953125, 0.37321011352539063, 0.37358694458007813, 0.3731374206542969, 0.800611328125, 0.3731507263183594, 0.3735572509765625, 0.373285888671875, 0.37302374267578126, 0.37313534545898436, 0.3731374206542969, 0.37296331787109377, 0.37312716674804686, 0.37298074340820314, 0.3729541015625, 0.37317938232421877, 0.3731005554199219, 0.37316094970703123, 0.37308621215820315, 0.37306573486328126, 0.37290292358398436, 0.373423095703125, 0.37339340209960936, 0.3730483093261719, 0.37310772705078127, 0.37330227661132814, 0.37319476318359374, 0.3731537780761719, 0.3731230773925781, 0.37313946533203124, 0.3732295532226563, 0.37302783203125, 0.3731537780761719, 0.3735255126953125, 0.373106689453125, 0.37300634765625, 0.373074951171875, 0.3731148681640625, 0.37321319580078127, 0.37321728515625, 0.3734394836425781, 0.3732971496582031, 0.3732418518066406, 0.3731732482910156, 0.37319168090820315, 0.37337088012695313, 0.3732561950683594, 0.37321319580078127, 0.3733145751953125, 0.3732695007324219, 0.3732213745117188, 0.3731312561035156, 0.37326028442382814, 0.3731537780761719, 0.3732295532226563, 0.3731005554199219, 0.37324594116210935, 0.37309951782226564, 0.3731435546875, 0.37312408447265627, 0.37325006103515623, 0.3732408447265625, 0.37334014892578127, 0.373291015625, 0.37328897094726565, 0.373180419921875, 0.373243896484375, 0.8022302856445312, 0.3732367248535156, 0.3730473022460937, 0.37313638305664065, 0.3729766540527344, 0.373064697265625, 0.37281484985351565, 0.3731517333984375, 0.3731517333984375, 0.37317120361328127, 0.3729776611328125, 0.3731445617675781, 0.37313330078125, 0.37301556396484375, 0.3731087341308594, 0.37299609375, 0.3729981384277344, 0.3730073547363281, 0.3732428894042969, 0.3732244873046875, 0.3732152099609375, 0.3731445617675781, 0.37313330078125, 0.3734200439453125, 0.3730196533203125, 0.37302169799804685, 0.37305856323242187, 0.3733114929199219, 0.37323468017578126, 0.373148681640625, 0.3729070129394531, 0.3729428405761719, 0.37319476318359374, 0.3731937255859375, 0.3731722106933594, 0.37321115112304687, 0.37312203979492187, 0.37307699584960935, 0.37318450927734376, 0.373254150390625, 0.37303909301757815, 0.37332171630859373, 0.37332992553710936, 0.3732735900878906, 0.37309234619140624, 0.373327880859375, 0.37318246459960935, 0.3731937255859375, 0.373138427734375, 0.3731312561035156, 0.3730073547363281, 0.3731507263183594, 0.3735469970703125, 0.3731732482910156, 0.37316915893554686, 0.3732090759277344, 0.3731619873046875, 0.3731087341308594, 0.3732336730957031, 0.37324594116210935, 0.3736606750488281, 0.3738367919921875, 0.37335244750976565]",tokens/s,2.6364884031443228,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -9208,7 +9948,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c10a8-2191c7c010eac6cd5292d4ec;a70deaef-cc89-4768-9686-49e4c5fbc8f0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d42f7-77f4d5820389e974744f4d4b;6890e4ad-93d7-457e-a3a8-aa4bf2be4b95) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9239,6 +9979,7 @@ OSError: s is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2280.251392,3257.401344,0.0,2610.95424,2376.633344,s,10,1.3862317810058593,0.13862317810058594,0.00141741642608869,0.13817852783203124,0.13898810577392579,0.14091418075561524,0.1424550407409668,"[0.1428402557373047, 0.1380400390625, 0.13832354736328126, 0.13791488647460937, 0.13817552185058593, 0.13818153381347656, 0.1379407958984375, 0.13822525024414062, 0.1380298614501953, 0.13856008911132814]",tokens/s,1846.733017578378,kWh,1.6310952320309356e-06,8.937665289211085e-07,7.707140868904123e-06,1.0232002629856168e-05,tokens/kWh,25019540.090129808,MB,2280.251392,3257.401344,0.0,2610.95424,2525.751808,s,10,74.85740771484375,7.485740771484375,0.0031337016640348436,7.485281982421874,7.487884423828125,7.490807446289063,7.493145864257812,"[7.49373046875, 7.48317431640625, 7.48723486328125, 7.482859375, 7.48208740234375, 7.486955078125, 7.48550634765625, 7.4850576171875, 7.4844873046875, 7.48631494140625]",tokens/s,8.416000757064355,kWh,8.838547692882235e-05,4.8441651965326174e-05,0.00040339438245649535,0.0005402215113506439,tokens/kWh,116618.82889944439,,s,629,75.97926604461664,0.12079374569891366,0.016430053385745564,0.11878092956542968,0.11900436248779296,0.11913031616210937,0.25699758911132814,"[0.12148735809326172, 0.12057292938232422, 0.11934105682373047, 0.11900825500488281, 0.11877887725830077, 0.11872563171386719, 0.11900313568115234, 0.11889254760742188, 0.11883110046386719, 0.11867955017089844, 0.11865599822998046, 0.11875020599365234, 0.11877785491943359, 0.11879730987548828, 0.11893145751953126, 0.11859251403808593, 0.11864575958251954, 0.11864371490478516, 0.11889971160888672, 0.11905126190185547, 0.11882701110839844, 0.11864780426025391, 0.11892530822753906, 0.11885158538818359, 0.11881676483154296, 0.1188331527709961, 0.11881779479980469, 0.11872358703613281, 0.11903692626953125, 0.11903794860839843, 0.11879424285888672, 0.11873689270019532, 0.11872767639160156, 0.11870310211181641, 0.11868364715576171, 0.11869798278808594, 0.11872051239013671, 0.11880140686035157, 0.11876863861083985, 0.11880038452148438, 0.11885363006591797, 0.11878604888916015, 0.1187747802734375, 0.11874508666992188, 0.11871743774414062, 0.11891814422607422, 0.11883827209472657, 0.11879014587402344, 0.11884544372558593, 0.11880038452148438, 0.119119873046875, 0.11894579315185547, 0.11882701110839844, 0.11877581024169923, 0.11902361297607422, 0.1188853759765625, 0.11901337432861328, 0.11886386871337891, 0.11883213043212891, 0.11893965148925781, 0.11898060607910156, 0.11884544372558593, 0.25700762939453126, 0.1187215347290039, 0.11884953308105468, 0.118793212890625, 0.11871743774414062, 0.11865190124511718, 0.11872665405273437, 0.11887923431396484, 0.11878707122802734, 0.11871743774414062, 0.11865599822998046, 0.11864883422851563, 0.11860889434814453, 0.11867852783203126, 0.11873587036132813, 0.11875328063964843, 0.11871027374267579, 0.11867648315429688, 0.11879936218261719, 0.11867340850830078, 0.1187440643310547, 0.11865805053710937, 0.1188331527709961, 0.11903385925292968, 0.11867750549316407, 0.11875328063964843, 0.118687744140625, 0.11879833221435547, 0.11875122833251953, 0.11872051239013671, 0.11871539306640624, 0.11883213043212891, 0.11874713897705078, 0.11873792266845704, 0.11877785491943359, 0.11873484802246094, 0.1188136978149414, 0.11876454162597656, 0.11867955017089844, 0.11872972869873047, 0.11878502655029297, 0.11877785491943359, 0.11875635528564453, 0.11875737762451172, 0.11869593811035156, 0.11877273559570313, 0.11876761627197266, 0.11885670471191406, 0.11873689270019532, 0.11874816131591796, 0.119299072265625, 0.11888333129882812, 0.11880754852294922, 0.11899187469482422, 0.11877069091796875, 0.11894783782958984, 0.11879116821289062, 0.11896729278564454, 0.1188136978149414, 0.11876249694824219, 0.11882905578613281, 0.11883519744873047, 0.11889356994628907, 0.2573066101074219, 0.11885158538818359, 0.11878092956542968, 0.11885363006591797, 0.11884748840332031, 0.11876249694824219, 0.11912908935546875, 0.11910655975341797, 0.11878809356689453, 0.11891200256347656, 0.11891097259521484, 0.11889663696289063, 0.11876557159423828, 0.11887820434570312, 0.11878912353515625, 0.11880550384521485, 0.11888742065429687, 0.11874201965332032, 0.11874610900878907, 0.11900621032714843, 0.11888435363769531, 0.11894989013671875, 0.11894477081298828, 0.11900415802001953, 0.11870310211181641, 0.11884236907958984, 0.11869184112548828, 0.11884339141845703, 0.118793212890625, 0.11888435363769531, 0.11882495880126953, 0.11875532531738281, 0.11875020599365234, 0.11875328063964843, 0.11879116821289062, 0.11881779479980469, 0.11889561462402344, 0.11876351928710938, 0.11884441375732421, 0.11896729278564454, 0.11886080169677735, 0.1189191665649414, 0.11933900451660157, 0.1189017562866211, 0.11875328063964843, 0.1187061767578125, 0.11873587036132813, 0.1187747802734375, 0.11875635528564453, 0.11884646606445312, 0.11881676483154296, 0.11880754852294922, 0.11878502655029297, 0.11884646606445312, 0.11881881713867187, 0.11887411499023437, 0.11868978881835937, 0.11886284637451172, 0.11876659393310547, 0.11876659393310547, 0.11881983947753906, 0.11880038452148438, 0.11876249694824219, 0.2568867797851562, 0.11862322998046874, 0.11860582733154297, 0.1186846694946289, 0.1186662368774414, 0.1187583999633789, 0.11907891082763672, 0.11880038452148438, 0.11864883422851563, 0.11862732696533203, 0.11873177337646484, 0.11859865570068359, 0.1187041244506836, 0.11873382568359375, 0.11864883422851563, 0.11872972869873047, 0.11863654327392578, 0.11863756561279297, 0.11869696044921875, 0.11869184112548828, 0.11872051239013671, 0.11913113403320312, 0.118866943359375, 0.11936153411865234, 0.11871334075927735, 0.11872767639160156, 0.11900006103515624, 0.1192806396484375, 0.11870105743408203, 0.1187041244506836, 0.11878809356689453, 0.11877069091796875, 0.11866214752197266, 0.11875942230224609, 0.118761474609375, 0.11870207977294922, 0.11872255706787109, 0.1187092514038086, 0.11868057250976563, 0.11887513732910156, 0.11874201965332032, 0.11875122833251953, 0.11880140686035157, 0.11873280334472656, 0.11872563171386719, 0.11872563171386719, 0.11866726684570313, 0.11876966094970703, 0.118761474609375, 0.11882086181640625, 0.118830078125, 0.11870515441894532, 0.11879730987548828, 0.11883929443359376, 0.11884236907958984, 0.1188823013305664, 0.11876044464111328, 0.11883417510986329, 0.11877887725830077, 0.11874816131591796, 0.11887513732910156, 0.11880857849121093, 0.11886182403564453, 0.25713458251953125, 0.11862322998046874, 0.11855974578857421, 0.11867750549316407, 0.1186723861694336, 0.11865907287597656, 0.11861811065673829, 0.11866111755371093, 0.11866111755371093, 0.1186355209350586, 0.11870207977294922, 0.11861606597900391, 0.11865907287597656, 0.1186355209350586, 0.11862220764160156, 0.11868364715576171, 0.11856690979003906, 0.11864780426025391, 0.11866111755371093, 0.11873689270019532, 0.11864473724365235, 0.1186170883178711, 0.1186170883178711, 0.11881983947753906, 0.11864883422851563, 0.11870310211181641, 0.11898675537109375, 0.1188505630493164, 0.11869388580322265, 0.11911577606201172, 0.1186170883178711, 0.118761474609375, 0.11871949005126953, 0.1187409896850586, 0.11874713897705078, 0.11871231842041016, 0.1186846694946289, 0.11867340850830078, 0.11867443084716797, 0.11871437072753906, 0.11872767639160156, 0.11932672119140625, 0.11876249694824219, 0.11871743774414062, 0.11873177337646484, 0.11873075103759766, 0.11874508666992188, 0.11923046112060547, 0.11879730987548828, 0.11878502655029297, 0.11888435363769531, 0.11883827209472657, 0.11881779479980469, 0.11898368072509766, 0.11900927734375, 0.1190666275024414, 0.11874918365478515, 0.11893247985839844, 0.11886489868164063, 0.11890995025634765, 0.1188136978149414, 0.118761474609375, 0.11898265838623047, 0.25697177124023435, 0.11863756561279297, 0.11881574249267578, 0.1188485107421875, 0.1187430419921875, 0.1186529312133789, 0.11871437072753906, 0.11902464294433594, 0.11879014587402344, 0.1185802230834961, 0.11868569946289062, 0.11872870635986328, 0.11875430297851562, 0.11874918365478515, 0.11903180694580077, 0.11878809356689453, 0.1187215347290039, 0.11869388580322265, 0.1190645751953125, 0.11893145751953126, 0.11888025665283203, 0.11871846771240234, 0.11874201965332032, 0.11916390228271484, 0.11881267547607421, 0.11875942230224609, 0.11871027374267579, 0.11877069091796875, 0.11881983947753906, 0.11884134674072265, 0.11873689270019532, 0.11876659393310547, 0.11889766693115235, 0.11874508666992188, 0.118761474609375, 0.11871129608154297, 0.11911373138427735, 0.11876044464111328, 0.11881881713867187, 0.11877375793457032, 0.1187768325805664, 0.11875430297851562, 0.11875225830078125, 0.11871641540527343, 0.11877375793457032, 0.11888025665283203, 0.11917721557617188, 0.11915161895751954, 0.11898162841796875, 0.11890688323974609, 0.11887615966796874, 0.11886080169677735, 0.11896832275390624, 0.11905023956298828, 0.11883929443359376, 0.11891506958007812, 0.11887104034423829, 0.11888025665283203, 0.11872767639160156, 0.11872255706787109, 0.11872767639160156, 0.1191178207397461, 0.11907071685791015, 0.2572810363769531, 0.11864371490478516, 0.11879936218261719, 0.11882701110839844, 0.11868876647949218, 0.11884953308105468, 0.11872051239013671, 0.11873996734619141, 0.11865190124511718, 0.11866214752197266, 0.118830078125, 0.11855872344970703, 0.11869286346435547, 0.11869081878662109, 0.11881267547607421, 0.11879424285888672, 0.11861913299560548, 0.11868569946289062, 0.1187215347290039, 0.1187799072265625, 0.11875225830078125, 0.11871129608154297, 0.11886796569824219, 0.11885568237304688, 0.11875737762451172, 0.11871129608154297, 0.11866726684570313, 0.11888435363769531, 0.11878809356689453, 0.11875328063964843, 0.11872665405273437, 0.118940673828125, 0.1188853759765625, 0.11914444732666016, 0.11903078460693359, 0.11889766693115235, 0.11882905578613281, 0.1193175048828125, 0.1194629135131836, 0.11916390228271484, 0.11871437072753906, 0.1188362274169922, 0.11889868927001954, 0.11877171325683594, 0.11894783782958984, 0.11882598114013672, 0.11881983947753906, 0.11883110046386719, 0.11869593811035156, 0.11869286346435547, 0.11880448150634766, 0.11876761627197266, 0.11887820434570312, 0.11883827209472657, 0.11880754852294922, 0.11889356994628907, 0.1188884506225586, 0.11900518035888671, 0.11882393646240234, 0.11872563171386719, 0.1188485107421875, 0.11869696044921875, 0.11872051239013671, 0.25712435913085935, 0.11903794860839843, 0.11959500885009766, 0.11911679840087891, 0.11868876647949218, 0.1186355209350586, 0.11873382568359375, 0.11865907287597656, 0.11861913299560548, 0.11864575958251954, 0.11867955017089844, 0.11856896209716797, 0.11867340850830078, 0.11858841705322265, 0.11869286346435547, 0.11870105743408203, 0.11860173034667969, 0.11890483093261718, 0.11876351928710938, 0.11874918365478515, 0.11870105743408203, 0.11868364715576171, 0.11892838287353516, 0.11915878295898437, 0.11867750549316407, 0.11876966094970703, 0.11884544372558593, 0.11876249694824219, 0.11886489868164063, 0.11876454162597656, 0.11879014587402344, 0.11874508666992188, 0.1188116455078125, 0.11873382568359375, 0.11884339141845703, 0.11875328063964843, 0.118793212890625, 0.11881267547607421, 0.11897344207763672, 0.11890380859375, 0.11877171325683594, 0.11874201965332032, 0.11877273559570313, 0.1187430419921875, 0.11879424285888672, 0.11870105743408203, 0.11880345916748047, 0.11893043518066407, 0.11869900512695312, 0.1187215347290039, 0.11876761627197266, 0.11889459228515625, 0.1188362274169922, 0.11887820434570312, 0.11884441375732421, 0.11891609954833984, 0.11884748840332031, 0.11893657684326171, 0.11885465240478515, 0.11897548675537109, 0.11905126190185547, 0.11878195190429687, 0.11880652618408204, 0.25719192504882815, 0.11870310211181641, 0.11854438018798828, 0.11866419219970703, 0.11863346862792969, 0.1186723861694336, 0.11862937927246094, 0.11873587036132813, 0.11876454162597656, 0.1187440643310547, 0.11862322998046874, 0.11850956726074219, 0.11866419219970703, 0.11865395355224609, 0.11865395355224609, 0.11879424285888672, 0.11863142395019531, 0.11870003509521485, 0.11862220764160156, 0.11868876647949218, 0.11871437072753906, 0.1188505630493164, 0.11876351928710938, 0.1194629135131836, 0.11897138977050781, 0.11900518035888671, 0.11874508666992188, 0.11873075103759766, 0.11873792266845704, 0.11888025665283203, 0.11870822143554688, 0.11880960083007812, 0.11874201965332032, 0.11898470306396484, 0.11879424285888672, 0.11886386871337891, 0.11871231842041016, 0.11913625335693359, 0.11880038452148438, 0.11878399658203125, 0.11880754852294922, 0.11878707122802734, 0.11881983947753906, 0.11889561462402344, 0.11877069091796875, 0.11868876647949218, 0.11882803344726563, 0.11891609954833984, 0.11885260772705078, 0.11894374084472656, 0.1189375991821289, 0.11888435363769531, 0.11889766693115235, 0.11880242919921875, 0.11900109100341796, 0.11900825500488281, 0.11882393646240234, 0.11897548675537109, 0.11886489868164063, 0.11870822143554688, 0.11882803344726563, 0.11881574249267578, 0.1188106231689453, 0.257470458984375, 0.11861504364013672, 0.11862220764160156, 0.11870105743408203, 0.11876454162597656, 0.11863756561279297, 0.11868978881835937, 0.11875942230224609, 0.11868364715576171, 0.11868876647949218, 0.1187092514038086, 0.11873177337646484, 0.1186344985961914, 0.11864064025878907, 0.11893350219726563, 0.11897344207763672, 0.11875020599365234, 0.11871846771240234, 0.11872972869873047, 0.11889459228515625, 0.11897344207763672, 0.11886489868164063, 0.11876557159423828, 0.11885568237304688, 0.11884031677246094, 0.11881574249267578, 0.11884441375732421, 0.11887104034423829, 0.11884134674072265, 0.1187747802734375, 0.11868569946289062, 0.11880140686035157, 0.11876863861083985, 0.11873382568359375, 0.11895193481445313, 0.11897241973876953, 0.1189191665649414, 0.1187799072265625, 0.118761474609375, 0.11878912353515625, 0.11878195190429687, 0.11875020599365234, 0.11874918365478515, 0.11920793914794922, 0.11907071685791015, 0.11895398712158203, 0.11883929443359376, 0.11883929443359376, 0.11869593811035156, 0.11901747131347656, 0.11879936218261719, 0.11887411499023437, 0.11884646606445312, 0.11897856140136719, 0.11885670471191406, 0.11900415802001953, 0.11876557159423828, 0.11950182342529297, 0.11893555450439453, 0.118761474609375, 0.11876966094970703, 0.11880242919921875, 0.11889459228515625]",tokens/s,8.278574310399858,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1506.42688,1579.679744,0.0,933.23264,856.69632,s,10,0.43564921951293945,0.043564921951293946,0.0031757046407324095,0.04331248092651367,0.04437063980102539,0.0485523281097412,0.051897678756713875,"[0.052734016418457035, 0.04333430480957031, 0.04164771270751953, 0.043380638122558594, 0.0414692497253418, 0.04140153503417969, 0.04329619216918945, 0.04344137573242188, 0.04332876968383789, 0.04161542510986328]",tokens/s,5876.287355368403,kWh,4.969496102560134e-07,2.72305043992154e-07,1.379101243335609e-06,2.1483558975837764e-06,tokens/kWh,119160889.63095888,MB,1506.42688,1579.679744,0.0,933.23264,899.62496,s,10,24.820336181640624,2.4820336181640625,0.03505314468873128,2.4815734863281254,2.524392333984375,2.52884619140625,2.53240927734375,"[2.533300048828125, 2.512392822265625, 2.510242431640625, 2.523402587890625, 2.427984130859375, 2.441885986328125, 2.461017822265625, 2.4780390625, 2.48510791015625, 2.44696337890625]",tokens/s,25.382412042670286,kWh,2.9558689719627776e-05,1.6199240900285592e-05,7.306973609346903e-05,0.00011882766671338238,tokens/kWh,530179.5595461687,,s,629,25.16415895462036,0.04000661201052522,0.005123259677494279,0.03999129486083984,0.040338432312011716,0.04051681213378906,0.07998427032470704,"[0.04113715362548828, 0.041027584075927735, 0.041578495025634765, 0.04084121704101563, 0.04151398468017578, 0.041063423156738284, 0.04110131072998047, 0.0394700813293457, 0.039803905487060545, 0.03899084854125977, 0.03885260772705078, 0.04075110244750976, 0.04097536087036133, 0.04030361557006836, 0.04015513610839844, 0.04008652877807617, 0.0400261116027832, 0.04022272109985352, 0.04007628631591797, 0.040078399658203125, 0.04018272018432617, 0.040068096160888675, 0.040422401428222655, 0.04001484680175781, 0.04015411376953125, 0.04017356872558594, 0.04025958251953125, 0.04034048080444336, 0.04012134552001953, 0.04011212921142578, 0.04016537475585937, 0.040008705139160154, 0.040130561828613284, 0.04007731246948242, 0.04001484680175781, 0.04027699279785156, 0.03999129486083984, 0.04007731246948242, 0.040084480285644535, 0.0402503662109375, 0.04009983825683594, 0.04006399917602539, 0.04001279830932617, 0.040207359313964845, 0.04010092926025391, 0.04016838455200195, 0.04013158416748047, 0.040172542572021484, 0.040498241424560544, 0.04003936004638672, 0.040325119018554685, 0.04010291290283203, 0.040225791931152347, 0.04005376052856445, 0.04014796829223633, 0.039678016662597654, 0.04023392105102539, 0.04024422454833984, 0.03986227035522461, 0.04005376052856445, 0.040235008239746094, 0.04025446319580078, 0.08356147003173828, 0.04013158416748047, 0.040151039123535154, 0.04020537567138672, 0.04016428756713867, 0.03999334335327148, 0.040237056732177735, 0.0403394546508789, 0.04013260650634766, 0.04007219314575195, 0.04012236785888672, 0.04005785751342773, 0.03998822402954102, 0.040136703491210936, 0.040052734375, 0.040161312103271486, 0.03975164794921875, 0.040303680419921876, 0.04009260940551758, 0.04006707382202149, 0.04007628631591797, 0.04026265716552734, 0.04016128158569336, 0.040051712036132815, 0.040156158447265625, 0.04031795120239258, 0.04012748718261719, 0.04016230392456055, 0.03998617553710938, 0.04028518295288086, 0.04020633697509766, 0.04005068969726563, 0.0399288330078125, 0.040256511688232424, 0.039979007720947264, 0.04008550262451172, 0.04009164810180664, 0.04038860702514648, 0.040223743438720705, 0.04016844940185547, 0.04011724853515625, 0.04027699279785156, 0.04012441635131836, 0.04024729537963867, 0.040101886749267575, 0.04006399917602539, 0.04013158416748047, 0.04007526397705078, 0.039964672088623046, 0.04025958251953125, 0.04006707382202149, 0.040025089263916014, 0.04005068969726563, 0.03987968063354492, 0.03824639892578125, 0.038312961578369144, 0.03836006546020508, 0.038322208404541015, 0.03826992034912109, 0.03886899185180664, 0.03874611282348633, 0.0382751350402832, 0.038287296295166015, 0.08017305755615234, 0.03826483154296875, 0.038435840606689455, 0.03826073455810547, 0.03853107070922852, 0.04071014404296875, 0.03881062316894531, 0.03902771377563476, 0.040190975189208986, 0.04033331298828125, 0.04007526397705078, 0.04009471893310547, 0.0400711669921875, 0.04029747009277344, 0.03996364974975586, 0.040139774322509765, 0.04016128158569336, 0.04015206527709961, 0.04011315155029297, 0.040158206939697266, 0.039951358795166016, 0.04007526397705078, 0.04027699279785156, 0.03998720169067383, 0.04024729537963867, 0.03996672058105469, 0.0404398078918457, 0.040035327911376956, 0.040134654998779294, 0.04028108978271484, 0.03998822402954102, 0.039959552764892575, 0.040130561828613284, 0.04015411376953125, 0.04040499114990234, 0.04033433532714844, 0.04011929702758789, 0.040313854217529296, 0.04006614303588867, 0.039992225646972655, 0.04012543869018555, 0.04003942489624023, 0.04003635025024414, 0.04030976104736328, 0.04017561721801758, 0.03996364974975586, 0.04004048156738281, 0.04014076614379883, 0.040002559661865236, 0.04034969711303711, 0.04014284896850586, 0.04008345413208008, 0.04016844940185547, 0.03999948883056641, 0.039951358795166016, 0.03837644958496094, 0.03825766372680664, 0.03821158218383789, 0.038324222564697266, 0.039932926177978514, 0.040046592712402344, 0.04010291290283203, 0.04031283187866211, 0.08356454467773437, 0.040120319366455076, 0.040197120666503904, 0.04001484680175781, 0.04011520004272461, 0.0399738883972168, 0.04001587295532227, 0.040022014617919925, 0.040030208587646485, 0.03988684844970703, 0.040529918670654294, 0.04023910522460938, 0.039923713684082034, 0.0401797103881836, 0.04036505508422852, 0.040136703491210936, 0.039977985382080077, 0.04037222290039062, 0.040443904876708986, 0.04031999969482422, 0.04006092834472656, 0.04047359848022461, 0.040342529296875, 0.04001279830932617, 0.040022014617919925, 0.040134654998779294, 0.04007219314575195, 0.04011212921142578, 0.040000511169433595, 0.04015209579467773, 0.04024316787719726, 0.0400711669921875, 0.04026265716552734, 0.04001792144775391, 0.04007526397705078, 0.0404398078918457, 0.04011724853515625, 0.04022784042358398, 0.040081409454345705, 0.04112998580932617, 0.04065484619140625, 0.04024524688720703, 0.04050841522216797, 0.04051763153076172, 0.04026163101196289, 0.040182785034179686, 0.04026572799682617, 0.04014080047607422, 0.040338432312011716, 0.04013363265991211, 0.04012236785888672, 0.039981056213378906, 0.04031488037109375, 0.04002304077148437, 0.04011520004272461, 0.040033279418945314, 0.04029849624633789, 0.040041473388671874, 0.03832729721069336, 0.038545406341552735, 0.03859763336181641, 0.03841331100463867, 0.03827199935913086, 0.07982694244384765, 0.03820544052124023, 0.0380579833984375, 0.03835903930664063, 0.038199295043945314, 0.03845939254760742, 0.04019404983520508, 0.04046745681762695, 0.03990630340576172, 0.04013363265991211, 0.03996364974975586, 0.03836006546020508, 0.03817062377929688, 0.038171646118164065, 0.0381921272277832, 0.03843379211425781, 0.03821670532226563, 0.03819007873535156, 0.03816755294799805, 0.038141952514648435, 0.038299648284912106, 0.038242305755615234, 0.038491134643554685, 0.038255615234375, 0.03834982299804687, 0.038201343536376955, 0.03825151824951172, 0.038152191162109376, 0.03830886459350586, 0.03816755294799805, 0.038171646118164065, 0.03821263885498047, 0.03831600189208984, 0.03889561462402344, 0.03994112014770508, 0.04019302368164063, 0.03997183990478516, 0.04004761505126953, 0.03841843032836914, 0.03847679901123047, 0.038371326446533204, 0.038234111785888675, 0.03837849426269531, 0.038258689880371094, 0.03819007873535156, 0.03830886459350586, 0.03834982299804687, 0.03825254440307617, 0.03839590454101562, 0.038346752166748044, 0.03817779159545898, 0.03827199935913086, 0.03827609634399414, 0.03823311996459961, 0.03817987060546875, 0.03817059326171875, 0.0383221435546875, 0.038166526794433595, 0.03820646286010742, 0.03827609634399414, 0.03827507019042969, 0.038204414367675785, 0.03846553421020508, 0.07984639739990235, 0.03822284698486328, 0.03814604949951172, 0.038381568908691405, 0.038329345703125, 0.03812659072875976, 0.0381030387878418, 0.0381921272277832, 0.03803033447265625, 0.03826483154296875, 0.03832831954956055, 0.038163455963134765, 0.038128639221191404, 0.03808256149291992, 0.03830476760864258, 0.038152191162109376, 0.03814912033081055, 0.03819724655151367, 0.03816150283813476, 0.03834255981445313, 0.038214656829833986, 0.0388587532043457, 0.03990937423706055, 0.03986841583251953, 0.04010291290283203, 0.040019966125488284, 0.03821670532226563, 0.038198272705078126, 0.03840409469604492, 0.038588417053222655, 0.03825664138793945, 0.038371326446533204, 0.03809075164794922, 0.03829862213134766, 0.03820236968994141, 0.038201343536376955, 0.038217727661132815, 0.038234111785888675, 0.038209537506103515, 0.03817574310302734, 0.038247425079345705, 0.0382371826171875, 0.03834470367431641, 0.03826892852783203, 0.03811840057373047, 0.03825459289550781, 0.03821363067626953, 0.039229438781738284, 0.04001279830932617, 0.040089599609375, 0.0400076789855957, 0.040134654998779294, 0.040187904357910156, 0.040088607788085935, 0.039887840270996094, 0.04031283187866211, 0.04015411376953125, 0.04027084732055664, 0.04039273452758789, 0.03996566390991211, 0.03829145431518555, 0.03822079849243164, 0.04009471893310547, 0.08352470397949219, 0.0401376953125, 0.040349632263183596, 0.04077977752685547, 0.04024524688720703, 0.04009676742553711, 0.040515583038330076, 0.040248321533203124, 0.04014796829223633, 0.04009574508666992, 0.04011929702758789, 0.040210430145263674, 0.040223743438720705, 0.0399738883972168, 0.04003839874267578, 0.0382371826171875, 0.038629375457763675, 0.038491134643554685, 0.03825766372680664, 0.03826483154296875, 0.03846656036376953, 0.038354942321777344, 0.038373374938964845, 0.03835289764404297, 0.03834265518188477, 0.03841535949707031, 0.03825664138793945, 0.038163455963134765, 0.038193153381347655, 0.038416385650634766, 0.03832217788696289, 0.03832320022583008, 0.0391464958190918, 0.040323070526123043, 0.04012851333618164, 0.04029849624633789, 0.04007526397705078, 0.040306751251220706, 0.03839788818359375, 0.0381399040222168, 0.038373374938964845, 0.03852492904663086, 0.038386688232421876, 0.038383617401123046, 0.03821670532226563, 0.03838054275512695, 0.03832012939453125, 0.038214656829833986, 0.03820032119750977, 0.038507518768310545, 0.03825664138793945, 0.03829145431518555, 0.038245376586914064, 0.03831193542480469, 0.038317054748535154, 0.0382740478515625, 0.038214656829833986, 0.038345729827880856, 0.03892124938964844, 0.03984790420532226, 0.04002918243408203, 0.040172542572021484, 0.04021247863769531, 0.08194969940185547, 0.03841331100463867, 0.038258689880371094, 0.038373374938964845, 0.038317054748535154, 0.038228992462158204, 0.03824127960205078, 0.038424575805664066, 0.03839487838745117, 0.038250495910644534, 0.038561790466308594, 0.03838873672485352, 0.04093439865112305, 0.040123390197753905, 0.04016742324829101, 0.040235008239746094, 0.04011212921142578, 0.04035276794433594, 0.04093132781982422, 0.03868569564819336, 0.038564895629882814, 0.0386354866027832, 0.03827199935913086, 0.038193153381347655, 0.038217727661132815, 0.038247425079345705, 0.03840518569946289, 0.038314945220947264, 0.03830988693237305, 0.03851468658447266, 0.03837952041625976, 0.03823308944702149, 0.03839184188842774, 0.0384859848022461, 0.0382105598449707, 0.0382740478515625, 0.03916595077514649, 0.04033740615844727, 0.038247425079345705, 0.03826176071166992, 0.039144447326660156, 0.041452545166015625, 0.040622081756591794, 0.04063641738891602, 0.040891391754150394, 0.04014591979980469, 0.040700927734375, 0.04041318511962891, 0.040308734893798825, 0.04019404983520508, 0.03999129486083984, 0.04006707382202149, 0.040828929901123044, 0.04010905456542969, 0.04021145629882812, 0.04004249572753906, 0.04000460815429688, 0.04015718460083008, 0.04007014465332031, 0.04011315155029297, 0.03992473602294922, 0.040338432312011716, 0.04014899063110351, 0.08364339447021485, 0.03997491073608399, 0.04009267044067383, 0.03998207855224609, 0.040376319885253906, 0.04017766571044922, 0.04027494430541992, 0.04038451385498047, 0.0402503662109375, 0.04045414352416992, 0.04019814300537109, 0.0397916145324707, 0.03818700790405274, 0.0382105598449707, 0.03836108779907227, 0.03888435363769531, 0.04011929702758789, 0.040153087615966795, 0.039880702972412106, 0.03999846267700195, 0.04047872161865235, 0.040325119018554685, 0.040395774841308595, 0.040190975189208986, 0.040041473388671874, 0.040049663543701174, 0.03856281661987305, 0.0383375358581543, 0.038228992462158204, 0.038348800659179685, 0.03831398391723633, 0.03893657684326172, 0.040150016784667966, 0.04014796829223633, 0.03996876907348633, 0.040226814270019534, 0.03987865447998047, 0.04001792144775391, 0.040079360961914064, 0.04005376052856445, 0.040172542572021484, 0.04024422454833984, 0.04010291290283203, 0.04010905456542969, 0.0401162223815918, 0.040218624114990234, 0.03831808090209961, 0.03854848098754883, 0.038324222564697266, 0.038430721282958984, 0.038348800659179685, 0.03827302551269531, 0.03822796630859375, 0.039934974670410156, 0.04017766571044922, 0.038391807556152346, 0.038354942321777344, 0.03833446502685547, 0.03852799987792969, 0.03838771057128906, 0.038245376586914064, 0.03821875381469726, 0.03834163284301758, 0.08003788757324219, 0.0383631362915039, 0.03848601531982422, 0.03860172653198242, 0.038424575805664066, 0.03824127960205078, 0.03828224182128906, 0.03827814483642578, 0.03832217788696289, 0.038422527313232424, 0.03840921783447265, 0.0383559684753418, 0.03847782516479492, 0.038231040954589846, 0.03817375946044922, 0.038246337890625, 0.03848294448852539, 0.038414337158203124, 0.038351871490478515, 0.03842764663696289, 0.038779937744140625, 0.0384532470703125, 0.038394847869873044, 0.038340606689453126, 0.03857612609863281, 0.0383375358581543, 0.03840716934204102, 0.03837855911254883, 0.03835078430175781, 0.038474750518798825, 0.03840921783447265, 0.038712318420410154, 0.0397844467163086, 0.04009881591796875, 0.04036812973022461, 0.0402606086730957, 0.04012441635131836, 0.04048793411254883, 0.04021350479125976, 0.04022886276245117, 0.04016025543212891, 0.03828326416015625, 0.03829350280761719, 0.038520832061767575, 0.03837747192382813, 0.038302719116210936, 0.03847884750366211, 0.03830476760864258, 0.038267902374267575, 0.038665214538574216, 0.0383744010925293, 0.038373374938964845, 0.038384639739990234, 0.038225921630859375, 0.03869081497192383, 0.038553600311279294, 0.03964313507080078, 0.04035276794433594, 0.03975884628295898, 0.04010700988769531, 0.040389633178710936, 0.03976704025268555, 0.04012748718261719]",tokens/s,24.995868176413264,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1281.970176,1017.643008,0.0,371.195904,277.410816,s,10,0.40465283584594725,0.04046528358459473,0.0012749464270020827,0.040339471817016595,0.041213354873657224,0.04256080570220947,0.04363876636505127,"[0.04390825653076172, 0.039191009521484375, 0.04020032119750976, 0.04062038421630859, 0.04001875305175781, 0.04091392135620117, 0.04051433563232422, 0.03929123306274414, 0.03951599884033203, 0.040478622436523434]",tokens/s,6326.4106246733445,kWh,4.743588816529138e-07,2.5992338141804797e-07,1.2597421947630295e-06,1.994024457833991e-06,tokens/kWh,128383580.7500977,MB,1281.970176,1017.643008,0.0,371.195904,323.048448,s,10,24.280882568359374,2.428088256835937,0.01672270399851807,2.436138427734375,2.4378572509765624,2.4385178588867187,2.4390463452148436,"[2.426797119140625, 2.4161357421875, 2.43644775390625, 2.436491455078125, 2.4358291015625, 2.439178466796875, 2.381922119140625, 2.4365849609375, 2.433785400390625, 2.43771044921875]",tokens/s,25.94633857423941,kWh,2.789970010339496e-05,1.5289993593269173e-05,6.758836843923807e-05,0.0001107780621359022,tokens/kWh,568704.65853349,,s,629,24.602083320617677,0.03911301004867675,0.004804808080024372,0.03870412826538086,0.0389910530090332,0.03930890274047851,0.07752167083740236,"[0.039656448364257815, 0.038916095733642575, 0.03911679840087891, 0.038441982269287106, 0.03871846389770508, 0.03863347244262695, 0.03846553421020508, 0.038747135162353515, 0.038983680725097655, 0.038621185302734375, 0.039468032836914066, 0.03973324966430664, 0.040008705139160154, 0.03956224060058594, 0.038986751556396484, 0.03861708831787109, 0.03892736053466797, 0.038964225769042966, 0.03865497589111328, 0.03835289764404297, 0.03786240005493164, 0.03894784164428711, 0.03873689651489258, 0.038623233795166016, 0.03889459228515625, 0.03893350219726562, 0.038599681854248044, 0.038431808471679686, 0.03866719818115234, 0.03852288055419922, 0.03881881713867188, 0.037525505065917966, 0.03853823852539062, 0.03861196899414063, 0.03871027374267578, 0.0388218879699707, 0.03872870254516601, 0.038642688751220705, 0.03880550384521484, 0.0386693115234375, 0.037972991943359374, 0.03834265518188477, 0.038626304626464845, 0.03872358322143555, 0.03853311920166016, 0.038572032928466796, 0.03861913681030273, 0.03886796951293945, 0.039011329650878904, 0.038809600830078124, 0.03837952041625976, 0.037607425689697264, 0.03740979385375977, 0.03724185562133789, 0.03723164749145508, 0.03744559860229492, 0.037335041046142575, 0.037253120422363284, 0.03730022430419922, 0.037302272796630856, 0.03739648056030274, 0.037308414459228514, 0.0768358383178711, 0.03727462387084961, 0.03721932983398438, 0.03744255828857422, 0.03722444915771484, 0.03723263931274414, 0.03723468780517578, 0.037351425170898435, 0.03726028823852539, 0.037222400665283206, 0.03719782257080078, 0.037326847076416016, 0.03750092697143555, 0.038975486755371096, 0.03851468658447266, 0.038324222564697266, 0.03836415863037109, 0.03865702438354492, 0.03859251022338867, 0.03796889495849609, 0.038454273223876956, 0.03858124923706055, 0.03827609634399414, 0.039785472869873044, 0.03924787139892578, 0.03859251022338867, 0.03877068710327149, 0.03864780807495117, 0.038196224212646485, 0.03844095993041992, 0.03861913681030273, 0.038670337677001954, 0.038452224731445314, 0.037917697906494144, 0.03871744155883789, 0.038529022216796875, 0.038520832061767575, 0.038778881072998046, 0.03852799987792969, 0.038389759063720705, 0.03772422409057617, 0.03858425521850586, 0.03871641540527344, 0.03861094284057617, 0.03887411117553711, 0.03873996734619141, 0.03890790557861328, 0.03827711868286133, 0.038902782440185545, 0.03904307174682617, 0.0387922248840332, 0.038702049255371095, 0.03786137771606445, 0.038809600830078124, 0.03896934509277344, 0.03885977554321289, 0.038787071228027346, 0.038757377624511716, 0.03866009521484375, 0.038470657348632815, 0.038882305145263675, 0.03882393646240234, 0.038712318420410154, 0.07939891052246094, 0.03869286346435547, 0.038795265197753906, 0.03887513732910156, 0.03879423904418945, 0.039034881591796876, 0.03891302490234375, 0.038836223602294925, 0.03861708831787109, 0.03800678253173828, 0.03891097640991211, 0.03804876708984375, 0.038882305145263675, 0.038168575286865236, 0.037981182098388674, 0.03827097702026367, 0.03884646224975586, 0.038865921020507815, 0.0377077751159668, 0.03886284637451172, 0.03846460723876953, 0.0388955192565918, 0.0387583999633789, 0.038612991333007815, 0.038899711608886715, 0.038785022735595705, 0.03849932861328125, 0.037820415496826174, 0.038653953552246094, 0.03862220764160156, 0.038809600830078124, 0.03883724975585937, 0.0387512321472168, 0.03888844680786133, 0.03887411117553711, 0.038833152770996096, 0.03888947296142578, 0.03813888168334961, 0.03830380630493164, 0.038789054870605466, 0.038876190185546874, 0.038718433380126954, 0.03883520126342774, 0.03773132705688476, 0.03853110504150391, 0.03896828842163086, 0.03829248046875, 0.037789695739746096, 0.038752254486083985, 0.03887206268310547, 0.03885158538818359, 0.038711296081542966, 0.038781951904296875, 0.03914342498779297, 0.03878400039672852, 0.038970367431640625, 0.038795265197753906, 0.03885772705078125, 0.03893145751953125, 0.03871027374267578, 0.038882305145263675, 0.03935232162475586, 0.03893657684326172, 0.0801976318359375, 0.040052734375, 0.039428096771240234, 0.038870014190673825, 0.038866943359375, 0.03884646224975586, 0.03879219055175781, 0.038886398315429685, 0.038757377624511716, 0.04078496170043945, 0.039140289306640624, 0.03867238235473633, 0.03890687942504883, 0.038779903411865234, 0.03772723388671875, 0.03863865661621094, 0.038687679290771486, 0.03894681549072266, 0.03893964767456055, 0.03887411117553711, 0.03868371200561523, 0.03804358291625977, 0.03752959823608398, 0.03767091369628906, 0.03757567977905273, 0.03858227157592774, 0.038712318420410154, 0.038975486755371096, 0.03836723327636719, 0.03893862533569336, 0.038735870361328126, 0.038797374725341796, 0.038236095428466794, 0.03860889434814453, 0.03884236907958984, 0.03882495880126953, 0.038738975524902346, 0.03886486434936524, 0.03886489486694336, 0.038742015838623044, 0.03873894500732422, 0.03862015914916992, 0.038100990295410156, 0.03866624069213867, 0.03854131317138672, 0.038796287536621094, 0.0379607048034668, 0.03861708831787109, 0.038790145874023435, 0.03870515060424805, 0.03848908615112305, 0.03781836700439453, 0.0387665901184082, 0.038441982269287106, 0.037978111267089845, 0.038556671142578124, 0.03888127899169922, 0.03868569564819336, 0.03868876647949219, 0.03892428970336914, 0.037594112396240234, 0.038866943359375, 0.038779903411865234, 0.07831346893310547, 0.03814604949951172, 0.03882700729370117, 0.038877185821533204, 0.039695358276367186, 0.03878604888916016, 0.0389826545715332, 0.038836223602294925, 0.038972415924072266, 0.03871744155883789, 0.03882393646240234, 0.03850649642944336, 0.03882495880126953, 0.0381921272277832, 0.03886796951293945, 0.038828033447265625, 0.039008255004882815, 0.0389222412109375, 0.038781951904296875, 0.03867750549316406, 0.038834175109863284, 0.03868876647949219, 0.03847577667236328, 0.03872358322143555, 0.03866726303100586, 0.03881062316894531, 0.03909939193725586, 0.038662143707275394, 0.038950912475585936, 0.038830078125, 0.038580257415771486, 0.038977504730224606, 0.03872870254516601, 0.038736961364746095, 0.03875628662109375, 0.038795265197753906, 0.038596607208251955, 0.03853414535522461, 0.03887206268310547, 0.03856076812744141, 0.03790950393676758, 0.038529022216796875, 0.03870412826538086, 0.03858432006835937, 0.037351425170898435, 0.03753881454467774, 0.038594558715820314, 0.038742015838623044, 0.038214656829833986, 0.038109184265136715, 0.03885363388061523, 0.03869696044921875, 0.038675457000732424, 0.038441982269287106, 0.03867750549316406, 0.03830476760864258, 0.0388587532043457, 0.038814720153808595, 0.038749183654785156, 0.038161407470703124, 0.03876147079467773, 0.03878297424316406, 0.03930521774291992, 0.07981977844238282, 0.03822796630859375, 0.03854131317138672, 0.03871027374267578, 0.03872153472900391, 0.03880550384521484, 0.038801406860351564, 0.0385873908996582, 0.03875635147094727, 0.03867852783203125, 0.03861606216430664, 0.038575103759765625, 0.03855052947998047, 0.03796489715576172, 0.03806198501586914, 0.03942092895507812, 0.04028108978271484, 0.038964225769042966, 0.03882393646240234, 0.03866419219970703, 0.03899084854125977, 0.03872051239013672, 0.037768192291259765, 0.03894169616699219, 0.0388935661315918, 0.03912704086303711, 0.03989913558959961, 0.038953983306884765, 0.03872972869873047, 0.03805184173583984, 0.037454849243164064, 0.03857100677490234, 0.038653953552246094, 0.03874508666992187, 0.03873689651489258, 0.038778881072998046, 0.03867340850830078, 0.03877580642700195, 0.03796480178833008, 0.03889152145385742, 0.0387665901184082, 0.03850137710571289, 0.03892633438110352, 0.03906969451904297, 0.0388403205871582, 0.03877171325683594, 0.03902873611450195, 0.03871744155883789, 0.03862015914916992, 0.03813683319091797, 0.038904830932617186, 0.03887308883666992, 0.038991870880126955, 0.03880243301391602, 0.03870207977294922, 0.03907174301147461, 0.03927449417114258, 0.038986751556396484, 0.0388218879699707, 0.03859763336181641, 0.03830579376220703, 0.03789209747314453, 0.03869081497192383, 0.07953612518310547, 0.03946086502075195, 0.03886796951293945, 0.03824127960205078, 0.037698593139648434, 0.037787647247314454, 0.03757564926147461, 0.03742924880981445, 0.037425151824951174, 0.03743027114868164, 0.037174270629882815, 0.037351425170898435, 0.0374200325012207, 0.037800960540771485, 0.037493759155273435, 0.037550079345703126, 0.03741593551635742, 0.037563392639160156, 0.037478462219238284, 0.037314495086669924, 0.037454849243164064, 0.037424129486083986, 0.03731763076782227, 0.03743129730224609, 0.03770265579223633, 0.03746201705932617, 0.03763411331176758, 0.03753977584838867, 0.037443584442138675, 0.03744255828857422, 0.0373196792602539, 0.03894784164428711, 0.03799244689941406, 0.03865497589111328, 0.03867750549316406, 0.038719486236572266, 0.03785830307006836, 0.038100990295410156, 0.03767705535888672, 0.03843379211425781, 0.038870014190673825, 0.03885055923461914, 0.0385054702758789, 0.03881881713867188, 0.03787366485595703, 0.03811635208129883, 0.03767193603515625, 0.03761459350585938, 0.03747123336791992, 0.037459968566894535, 0.03746105575561524, 0.037547969818115236, 0.037591041564941405, 0.03749990463256836, 0.03776716613769531, 0.037759998321533206, 0.03740467071533203, 0.037466110229492186, 0.03745587158203125, 0.03762176132202148, 0.03740467071533203, 0.03756851196289063, 0.037424129486083986, 0.07716556549072266, 0.03832524871826172, 0.03913113784790039, 0.03907583999633789, 0.03930316925048828, 0.04006604766845703, 0.03922431945800781, 0.03863961410522461, 0.03853311920166016, 0.03850035095214844, 0.038386688232421876, 0.03758899307250976, 0.0377077751159668, 0.03966054534912109, 0.03961654281616211, 0.039029727935791014, 0.038882305145263675, 0.03892940902709961, 0.038949951171875, 0.038799297332763674, 0.03883827209472656, 0.03873177719116211, 0.038470657348632815, 0.03772825622558594, 0.03750096130371094, 0.03739542388916016, 0.03849523162841797, 0.038932479858398435, 0.03892736053466797, 0.038724609375, 0.03894784164428711, 0.03870003128051758, 0.03879731369018555, 0.03885772705078125, 0.039008255004882815, 0.03877580642700195, 0.03863449478149414, 0.03875020980834961, 0.038730751037597655, 0.038609920501708986, 0.03889152145385742, 0.03879935836791992, 0.03799859237670898, 0.03885055923461914, 0.03883827209472656, 0.03868364715576172, 0.0378419189453125, 0.03827097702026367, 0.038882366180419924, 0.03889350509643555, 0.03880550384521484, 0.03875328063964844, 0.0384983024597168, 0.03806719970703125, 0.03889766311645508, 0.038747135162353515, 0.03880652618408203, 0.03879731369018555, 0.038317054748535154, 0.03869696044921875, 0.038836223602294925, 0.038951934814453124, 0.03869081497192383, 0.07766015625, 0.03868057632446289, 0.03878092956542969, 0.03884236907958984, 0.03863961410522461, 0.03917420959472656, 0.03880441665649414, 0.03875942230224609, 0.03831193542480469, 0.03883724975585937, 0.03871334457397461, 0.038863872528076174, 0.03870003128051758, 0.0387061767578125, 0.03883929443359375, 0.03847679901123047, 0.0389939193725586, 0.039126014709472655, 0.03876249694824219, 0.03862835311889649, 0.038225921630859375, 0.037928958892822266, 0.038201343536376955, 0.038752254486083985, 0.03788288116455078, 0.038316032409667966, 0.03892838287353516, 0.03883827209472656, 0.038575103759765625, 0.03866624069213867, 0.038742015838623044, 0.03868569564819336, 0.038523902893066404, 0.03863142395019531, 0.03812351989746094, 0.03848396682739258, 0.0384266242980957, 0.0380579833984375, 0.03916595077514649, 0.03902873611450195, 0.03880038452148438, 0.03798527908325195, 0.03868979263305664, 0.03873484802246094, 0.03874303817749023, 0.03897139358520508, 0.03858534240722656, 0.03883724975585937, 0.03884134292602539, 0.03871027374267578, 0.038441982269287106, 0.03866726303100586, 0.038435840606689455, 0.037664768218994144, 0.03777433776855469, 0.03874611282348633, 0.03870412826538086, 0.039122943878173826, 0.03954585647583008, 0.0387061767578125, 0.03868569564819336, 0.038558719635009765, 0.03892326354980469, 0.0795125732421875, 0.03869388961791992, 0.038509567260742186, 0.03903078460693359, 0.03869900894165039, 0.03888844680786133, 0.03870924758911133, 0.03882495880126953, 0.03866828918457031, 0.03871334457397461, 0.03749068832397461, 0.0382371826171875, 0.038708225250244144, 0.03870412826538086, 0.038577152252197267, 0.039311359405517575, 0.03893862533569336, 0.03851878356933594, 0.03875328063964844, 0.03947212982177734, 0.03978035354614258, 0.038626304626464845, 0.03877171325683594, 0.038569984436035154, 0.0387512321472168, 0.038697982788085936, 0.03871027374267578, 0.038152191162109376, 0.03879423904418945, 0.0386324462890625, 0.03985612869262695, 0.038866943359375, 0.038002784729003904, 0.03856582260131836, 0.03872560119628906, 0.038675457000732424, 0.038768638610839845, 0.03867750549316406, 0.03897651290893555, 0.03865599822998047, 0.03887104034423828, 0.03846963119506836, 0.038752254486083985, 0.0383109130859375, 0.039003135681152344, 0.03877273559570313, 0.03880755233764648, 0.038665214538574216, 0.0388853759765625, 0.038697982788085936, 0.03744563293457031, 0.038452224731445314, 0.03876761627197266, 0.038814720153808595, 0.03867750549316406, 0.03863961410522461, 0.03885977554321289, 0.03878604888916016, 0.03861920166015625, 0.039180225372314456, 0.038986751556396484, 0.03813785552978516, 0.037392383575439454]",tokens/s,25.56694048234806,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2029.686784,5274.861568,0.0,4628.414464,4279.64672,s,10,3.174773986816406,0.3174773986816406,0.0009221487872443522,0.3173043212890625,0.31849222106933595,0.31901463470458985,0.319432565612793,"[0.31953704833984375, 0.31614645385742185, 0.31721881103515626, 0.3171442565917969, 0.31738983154296874, 0.3167364807128906, 0.3175615539550781, 0.31837612915039065, 0.31800189208984375, 0.3166615295410156]",tokens/s,806.356613299302,kWh,3.736858639038272e-06,2.0476319108638563e-06,1.8145144724437956e-05,2.392963527434008e-05,tokens/kWh,10698031.836469762,MB,2029.686784,5274.861568,0.0,4628.414464,4463.185408,s,10,176.95450000000002,17.69545,0.0031517961497735305,17.6950830078125,17.699231054687502,17.70004228515625,17.700691269531248,"[17.69897265625, 17.69905078125, 17.700853515625, 17.69500390625, 17.695162109375, 17.69576171875, 17.693421875, 17.69222265625, 17.693814453125, 17.690236328125]",tokens/s,3.5602372361256704,kWh,0.00020892207810655237,0.00011450671390637579,0.0009965308336823604,0.0013199596256952884,tokens/kWh,47728.732586661325,,s,629,179.5270961608886,0.2854166870602364,0.03770346789831642,0.28084429931640625,0.2811535339355469,0.2813114379882813,0.5979956518554688,"[0.2810900573730469, 0.28075418090820314, 0.2808637390136719, 0.28086273193359373, 0.2813634643554688, 0.2807715759277344, 0.2808360900878906, 0.28084429931640625, 0.28089447021484376, 0.28069683837890624, 0.2808883056640625, 0.28103372192382814, 0.28076544189453123, 0.28069683837890624, 0.2809046936035156, 0.28098458862304687, 0.2806855773925781, 0.28079000854492187, 0.2813388671875, 0.2808258666992188, 0.2808616943359375, 0.28091802978515623, 0.2808821716308594, 0.28074496459960935, 0.28068453979492186, 0.28067019653320313, 0.28085964965820315, 0.28079205322265627, 0.2806651000976563, 0.28080743408203124, 0.2807859191894531, 0.2809405517578125, 0.28102041625976565, 0.28240179443359376, 0.28073675537109377, 0.2808053894042969, 0.28098355102539063, 0.28075518798828125, 0.2811678771972656, 0.2814289855957031, 0.28092312622070315, 0.2808555603027344, 0.28077566528320314, 0.2810408935546875, 0.28091595458984375, 0.2808340454101563, 0.28091189575195313, 0.2809783935546875, 0.28075726318359373, 0.28088116455078127, 0.2808842163085937, 0.28079718017578126, 0.2806927490234375, 0.28096408081054686, 0.28156112670898437, 0.28095278930664064, 0.28083203125, 0.281001953125, 0.28085964965820315, 0.28099172973632813, 0.28073165893554686, 0.28094259643554687, 0.59812353515625, 0.2807715759277344, 0.28079205322265627, 0.2807306213378906, 0.28069989013671875, 0.28109515380859373, 0.280732666015625, 0.2807408752441406, 0.28080743408203124, 0.28074188232421876, 0.2810009460449219, 0.2807080993652344, 0.28133273315429685, 0.28084326171875, 0.28075518798828125, 0.28082382202148437, 0.28070196533203123, 0.2806036376953125, 0.28080230712890625, 0.28076544189453123, 0.28069989013671875, 0.28076031494140624, 0.28054937744140623, 0.28094772338867186, 0.2809251708984375, 0.28090264892578126, 0.2812037048339844, 0.2809169921875, 0.2808821716308594, 0.28092312622070315, 0.2808821716308594, 0.2807029724121094, 0.28090673828125, 0.2810491027832031, 0.28080126953125, 0.2807142333984375, 0.28088626098632813, 0.2810163269042969, 0.28109619140625, 0.2811351013183594, 0.28115045166015623, 0.2812467346191406, 0.2808637390136719, 0.2808412170410156, 0.2813102111816406, 0.28072039794921877, 0.28143719482421875, 0.2813716430664063, 0.2811709289550781, 0.28095489501953125, 0.280875, 0.2809036865234375, 0.28120269775390627, 0.2809466857910156, 0.2809661560058594, 0.28132965087890627, 0.28103067016601563, 0.28087808227539063, 0.28130508422851563, 0.2808392028808594, 0.28099169921875, 0.28102554321289064, 0.2814218139648437, 0.5983283081054688, 0.280806396484375, 0.28079000854492187, 0.28083917236328126, 0.28077362060546873, 0.28097125244140625, 0.2809374694824219, 0.28083096313476563, 0.2809302978515625, 0.2808637390136719, 0.28075418090820314, 0.28077362060546873, 0.28114227294921873, 0.2807705688476563, 0.28075314331054685, 0.28077362060546873, 0.28073675537109377, 0.280890380859375, 0.28105523681640626, 0.28109722900390627, 0.28098458862304687, 0.28080126953125, 0.28088626098632813, 0.2811689453125, 0.2808769836425781, 0.2808493957519531, 0.28083917236328126, 0.28073370361328126, 0.280880126953125, 0.2808493957519531, 0.2811535339355469, 0.28096511840820315, 0.28092108154296874, 0.2810409240722656, 0.281107421875, 0.28079000854492187, 0.2810408935546875, 0.2809927673339844, 0.2809743347167969, 0.28115866088867186, 0.28111154174804687, 0.2810992736816406, 0.28116583251953126, 0.28090573120117185, 0.2810828857421875, 0.2810828857421875, 0.2809743347167969, 0.2810828857421875, 0.281702392578125, 0.28085964965820315, 0.2809661560058594, 0.2809671630859375, 0.2809518127441406, 0.2811023254394531, 0.281312255859375, 0.281206787109375, 0.2811064453125, 0.2809405517578125, 0.28098355102539063, 0.2809098205566406, 0.28094259643554687, 0.2807930908203125, 0.2810654602050781, 0.5982269287109375, 0.28070199584960936, 0.2806517333984375, 0.2808186950683594, 0.2806579284667969, 0.2811094970703125, 0.2807715759277344, 0.2807982177734375, 0.28059237670898435, 0.2807500915527344, 0.28073675537109377, 0.2805841979980469, 0.2806558837890625, 0.2806814575195313, 0.2806732788085938, 0.2806876220703125, 0.28078387451171877, 0.28072659301757813, 0.2806312255859375, 0.28082278442382813, 0.28084326171875, 0.28062823486328126, 0.28076544189453123, 0.280748046875, 0.28071014404296873, 0.2810214538574219, 0.2813992919921875, 0.28084530639648436, 0.28090264892578126, 0.2808401794433594, 0.2808299560546875, 0.2809927673339844, 0.28081253051757815, 0.28067636108398436, 0.2807152709960937, 0.2807992248535156, 0.2808186950683594, 0.2809610290527344, 0.2807982177734375, 0.280875, 0.280995849609375, 0.2812569580078125, 0.2808719482421875, 0.2807510986328125, 0.2810491027832031, 0.28073165893554686, 0.280985595703125, 0.28093548583984373, 0.2809241027832031, 0.280806396484375, 0.28093646240234377, 0.2808463439941406, 0.2809794616699219, 0.28087808227539063, 0.2808883056640625, 0.2814617614746094, 0.28086782836914065, 0.28073165893554686, 0.2811760559082031, 0.2809938049316406, 0.28129177856445314, 0.28095187377929687, 0.2817914123535156, 0.5984921875, 0.28099789428710936, 0.28095077514648437, 0.281122802734375, 0.2807726135253906, 0.2808883056640625, 0.28075418090820314, 0.28093646240234377, 0.28081253051757815, 0.28075213623046874, 0.2807726135253906, 0.28081253051757815, 0.28140032958984373, 0.28087808227539063, 0.28151806640625, 0.2811494445800781, 0.2811781005859375, 0.2809241638183594, 0.2809599914550781, 0.28065176391601565, 0.2806476745605469, 0.28059033203125, 0.28050125122070313, 0.2807224426269531, 0.28059039306640626, 0.2806568298339844, 0.2806446228027344, 0.2807173156738281, 0.280627197265625, 0.28063540649414065, 0.2808340454101563, 0.2807080993652344, 0.2810368041992187, 0.2810357666015625, 0.28083712768554686, 0.2806732788085938, 0.2807879638671875, 0.2809333801269531, 0.2808832092285156, 0.280869873046875, 0.2805882873535156, 0.28103067016601563, 0.2808279113769531, 0.2807982177734375, 0.2809129638671875, 0.28075820922851563, 0.28076953125, 0.2810152893066406, 0.28108184814453124, 0.2806855773925781, 0.2812528686523437, 0.2812149658203125, 0.28091494750976564, 0.28111154174804687, 0.28087908935546874, 0.2808330383300781, 0.2809251708984375, 0.28099172973632813, 0.28070401000976564, 0.2810091552734375, 0.2808401794433594, 0.2807623901367188, 0.2810029907226563, 0.5976668090820313, 0.28066201782226563, 0.280774658203125, 0.28086578369140625, 0.28072857666015627, 0.2810163879394531, 0.28058001708984376, 0.2806947937011719, 0.28056781005859377, 0.2807726135253906, 0.28080230712890625, 0.280553466796875, 0.28076544189453123, 0.280690673828125, 0.2808268737792969, 0.280875, 0.28081048583984375, 0.28086477661132814, 0.28081048583984375, 0.2810245056152344, 0.2807193603515625, 0.2807080993652344, 0.2806855773925781, 0.2810071105957031, 0.2810624084472656, 0.28088525390625, 0.28093646240234377, 0.2807060546875, 0.28088116455078127, 0.2806405029296875, 0.28068658447265626, 0.2808965148925781, 0.28075930786132813, 0.2807490539550781, 0.28122213745117186, 0.280764404296875, 0.28102655029296875, 0.28092825317382814, 0.28138290405273436, 0.28117196655273435, 0.28144537353515625, 0.2809876403808594, 0.281027587890625, 0.28075424194335935, 0.2808401184082031, 0.28082278442382813, 0.28102655029296875, 0.28107366943359374, 0.28075314331054685, 0.28084735107421877, 0.28093438720703123, 0.28076544189453123, 0.28106341552734376, 0.28107879638671873, 0.28101837158203125, 0.2814535827636719, 0.28100811767578127, 0.28086477661132814, 0.2809016418457031, 0.28101119995117185, 0.2808832092285156, 0.28088528442382815, 0.28095895385742187, 0.598540283203125, 0.2806599731445312, 0.2808401794433594, 0.2807408752441406, 0.28097842407226564, 0.2807705688476563, 0.28113101196289064, 0.28089547729492187, 0.28100607299804686, 0.28068453979492186, 0.2810439758300781, 0.28067019653320313, 0.28120269775390627, 0.2807859191894531, 0.280658935546875, 0.2809333801269531, 0.28086578369140625, 0.2808279113769531, 0.28084530639648436, 0.2808913879394531, 0.28075314331054685, 0.28059237670898435, 0.28081561279296874, 0.28066815185546873, 0.28071832275390624, 0.2806855773925781, 0.28078182983398436, 0.2810224914550781, 0.2808012390136719, 0.28073370361328126, 0.28072857666015627, 0.2806651000976563, 0.2808084411621094, 0.280742919921875, 0.2807459716796875, 0.2807408752441406, 0.28083096313476563, 0.280958984375, 0.2808985595703125, 0.2808555603027344, 0.2810992736816406, 0.2811351013183594, 0.2806947937011719, 0.28073370361328126, 0.2807859191894531, 0.2808637390136719, 0.2812098693847656, 0.2812231750488281, 0.28102041625976565, 0.2809169921875, 0.2809938049316406, 0.28092007446289063, 0.2809825134277344, 0.2808616943359375, 0.28068453979492186, 0.28086782836914065, 0.2807234497070312, 0.28062619018554685, 0.2807685241699219, 0.28067123413085937, 0.2807296142578125, 0.2808002624511719, 0.28117300415039065, 0.5992601318359375, 0.2805698547363281, 0.2806220703125, 0.28086578369140625, 0.2806927490234375, 0.28068658447265626, 0.28074188232421876, 0.28050225830078124, 0.28068658447265626, 0.2806220703125, 0.28081460571289063, 0.2807357482910156, 0.28070706176757815, 0.28100311279296875, 0.28070492553710935, 0.2808493957519531, 0.2806118469238281, 0.2809036865234375, 0.2807828369140625, 0.2809046936035156, 0.2806640625, 0.28067636108398436, 0.28069888305664065, 0.28089959716796875, 0.2806876220703125, 0.28073165893554686, 0.28076544189453123, 0.2807173156738281, 0.2807705688476563, 0.2804951171875, 0.2809938049316406, 0.2808279113769531, 0.28095693969726565, 0.28122726440429685, 0.2810163269042969, 0.28069989013671875, 0.2808842163085937, 0.28082278442382813, 0.28080126953125, 0.28084429931640625, 0.28069580078125, 0.2806753234863281, 0.28080230712890625, 0.2806804504394531, 0.2807798461914062, 0.2811206970214844, 0.28091802978515623, 0.28079718017578126, 0.28101937866210935, 0.2810828857421875, 0.2809466857910156, 0.2809190368652344, 0.2809415588378906, 0.28108389282226565, 0.28117913818359375, 0.28109515380859373, 0.2810357666015625, 0.2808545227050781, 0.280890380859375, 0.2808606872558594, 0.28100811767578127, 0.28085760498046874, 0.28095489501953125, 0.5989181518554687, 0.28075518798828125, 0.28063436889648435, 0.2807193603515625, 0.2805350341796875, 0.2809671630859375, 0.2804613037109375, 0.28068453979492186, 0.28063641357421876, 0.2805770263671875, 0.2807633972167969, 0.28062924194335936, 0.2807562255859375, 0.2807296142578125, 0.2805565490722656, 0.2806087646484375, 0.2806476745605469, 0.2805893249511719, 0.28073983764648436, 0.28069171142578125, 0.2810378112792969, 0.28060671997070313, 0.28059442138671875, 0.28090673828125, 0.28125799560546877, 0.28063436889648435, 0.28085861206054685, 0.280511474609375, 0.28059442138671875, 0.2806200256347656, 0.2808002624511719, 0.2806661071777344, 0.280890380859375, 0.28054629516601565, 0.2807828369140625, 0.28066302490234374, 0.280732666015625, 0.2807224426269531, 0.28073165893554686, 0.2807142333984375, 0.28113101196289064, 0.280958984375, 0.28107366943359374, 0.2807439270019531, 0.2809518127441406, 0.28086578369140625, 0.28080435180664065, 0.2811535339355469, 0.2808248291015625, 0.28094064331054686, 0.2808831176757812, 0.2809313354492188, 0.2811975708007812, 0.281122802734375, 0.2807705688476563, 0.2809292907714844, 0.281385986328125, 0.2810009460449219, 0.2827796630859375, 0.28104193115234377, 0.2813173828125, 0.2812630920410156, 0.2812651672363281, 0.597570556640625, 0.2809108581542969, 0.28094259643554687, 0.28096307373046875, 0.280585205078125, 0.2806241149902344, 0.28071218872070314, 0.28067840576171876, 0.2805749816894531, 0.2805186767578125, 0.28062612915039065, 0.280669189453125, 0.2806609802246094, 0.2807214050292969, 0.28059954833984374, 0.28069888305664065, 0.28063540649414065, 0.28063128662109377, 0.2805401611328125, 0.28073983764648436, 0.28051763916015626, 0.28082278442382813, 0.2806200256347656, 0.2806640625, 0.28071218872070314, 0.2807879638671875, 0.28067123413085937, 0.2807459716796875, 0.2807715759277344, 0.2807060546875, 0.28080743408203124, 0.28084326171875, 0.28085247802734375, 0.2809169921875, 0.2808770446777344, 0.28057907104492186, 0.28081048583984375, 0.28083712768554686, 0.28081765747070314, 0.28080743408203124, 0.28084326171875, 0.2808401794433594, 0.28075314331054685, 0.2806026306152344, 0.28084640502929686, 0.28075103759765624, 0.2807500915527344, 0.2811566162109375, 0.2810798034667969, 0.28084429931640625, 0.2807982177734375, 0.28074188232421876, 0.28069989013671875, 0.28078695678710935, 0.280806396484375, 0.2809016418457031, 0.28089447021484376, 0.2807982177734375, 0.28089547729492187, 0.28093438720703123, 0.2809241638183594, 0.2807132568359375, 0.28090365600585937]",tokens/s,3.50364938469401,, @@ -9401,7 +10142,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c105b-53b68b62257666bc6308ea5d;bae415a9-ea0c-4ad7-9372-1ad80f39e9fb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d429d-255edcd86e59f77963184a55;9be29526-012a-4860-961a-aeb5f950e207) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9471,7 +10212,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c14c6-1d5c50cf73709c275b3ac7e0;f4c4bf31-4a77-473e-ac1c-74606b121575) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46dc-0574d69847aedc013c7d67fe;101b824b-5fe9-4216-abb3-73aa3b5c249f) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9541,7 +10282,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1478-53a62f2c0641af8b7d580ced;0a69a37e-e4ba-4318-a3f5-f950758ffac1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d468e-589ee7e43b85fad33f4db78a;bfc33719-09f6-410a-84e4-0d2b2eec5306) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9641,7 +10382,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1369-1ce85d5b672b8c1a5a27256f;b693e070-1717-41b4-add2-eb7e03da73aa) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4584-2862f7927f20c13e0137ed65;67d7afca-9ed3-4949-aac5-cc81ec6d349e) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9712,7 +10453,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c13df-05af99fa275b23386e8a3d7c;9037856d-7216-4f89-ad52-9df2600d8119) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45f8-3e1a1de308fcccf938ca13cf;28e4513f-4022-4e17-86e2-4d863f7d8ab5) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9866,6 +10607,7 @@ ChildProcessError: Traceback (most recent call last): TypeError: DeciCoderAttention.forward() got an unexpected keyword argument 'cache_position' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2108.653568,5971.116032,0.0,5324.668928,5080.876032,s,10,3.254842498779297,0.3254842498779297,0.001114065237433989,0.3250286560058594,0.3257514434814453,0.3272673233032226,0.3284800271606445,"[0.328783203125, 0.32502313232421876, 0.3253096313476562, 0.3250040283203125, 0.32493109130859377, 0.3254145812988281, 0.3249727783203125, 0.32496087646484373, 0.3250341796875, 0.32540899658203126]",tokens/s,786.5203926027474,kWh,3.84116784859729e-06,2.1047766923490397e-06,1.920148489522528e-05,2.514742943617161e-05,tokens/kWh,10179966.928618725,MB,2108.653568,5971.116032,0.0,5324.668928,5290.118656,s,10,177.13133984375,17.713133984375002,0.001836879786445902,17.712749023437503,17.7155212890625,17.71599501953125,17.71637400390625,"[17.7094375, 17.71198828125, 17.71232421875, 17.712689453125, 17.712490234375, 17.713939453125, 17.71377734375, 17.71280859375, 17.71646875, 17.715416015625]",tokens/s,3.5566828577920293,kWh,0.00020915095838113933,0.00011463212397823582,0.0010249259615171765,0.0013487090438765517,tokens/kWh,46711.335025174216,,s,629,179.77634902954114,0.2858129555318617,0.038628015635411614,0.2811412353515625,0.2814228515625,0.28152381591796877,0.6061751391601563,"[0.2810777587890625, 0.28092620849609373, 0.28088525390625, 0.280838134765625, 0.2809333801269531, 0.28067941284179687, 0.28087091064453124, 0.2809241638183594, 0.2810931091308594, 0.28108184814453124, 0.28116070556640627, 0.28089447021484376, 0.2808821716308594, 0.28095077514648437, 0.2809036865234375, 0.280943603515625, 0.2808340454101563, 0.2809927673339844, 0.28108184814453124, 0.281091064453125, 0.28095794677734376, 0.2809938049316406, 0.2811002807617187, 0.28091595458984375, 0.28110848999023436, 0.2811279296875, 0.2810521545410156, 0.2811770935058594, 0.28101223754882815, 0.281059326171875, 0.2810245056152344, 0.2809323425292969, 0.28119961547851563, 0.2811125793457031, 0.28143206787109376, 0.281101318359375, 0.28115045166015623, 0.28138803100585935, 0.2812610473632812, 0.28120166015625, 0.28122726440429685, 0.28131329345703127, 0.28121701049804687, 0.28117913818359375, 0.2809989013671875, 0.2813675537109375, 0.28137677001953126, 0.28117913818359375, 0.2810777587890625, 0.28124774169921873, 0.281164794921875, 0.281196533203125, 0.28134707641601564, 0.2811832275390625, 0.2812538757324219, 0.28116888427734377, 0.28106854248046875, 0.28122726440429685, 0.281122802734375, 0.28126617431640627, 0.28131329345703127, 0.28124774169921873, 0.6063380737304688, 0.28098458862304687, 0.2808115234375, 0.2807510986328125, 0.28091802978515623, 0.28093646240234377, 0.28100302124023435, 0.28108084106445314, 0.2809866333007813, 0.2808965148925781, 0.2809385070800781, 0.2809046936035156, 0.2808934326171875, 0.28086273193359373, 0.28120269775390627, 0.2810859375, 0.28100811767578127, 0.28084326171875, 0.2810449829101562, 0.2810654602050781, 0.28101937866210935, 0.28106341552734376, 0.28127743530273436, 0.28122113037109375, 0.2810931091308594, 0.28118118286132815, 0.28112384033203125, 0.28111358642578127, 0.28134194946289065, 0.28118118286132815, 0.2811453552246094, 0.2811934814453125, 0.2813255615234375, 0.2810931091308594, 0.2812149658203125, 0.28111154174804687, 0.2811781005859375, 0.28105624389648437, 0.2810224609375, 0.28145050048828124, 0.28102859497070315, 0.2810921020507812, 0.2813788146972656, 0.28130712890625, 0.2813788146972656, 0.2810214538574219, 0.2815672302246094, 0.28127435302734377, 0.2810726318359375, 0.28103067016601563, 0.28111154174804687, 0.28136038208007813, 0.28119143676757813, 0.28112896728515624, 0.28117913818359375, 0.28128256225585935, 0.2813081665039063, 0.28129995727539064, 0.28135321044921874, 0.28134707641601564, 0.28152935791015626, 0.2813255615234375, 0.28147198486328123, 0.6066227416992187, 0.2809098205566406, 0.2808719482421875, 0.28102554321289064, 0.28104702758789063, 0.2808401794433594, 0.28098968505859373, 0.2811484069824219, 0.28116378784179685, 0.28091802978515623, 0.28097842407226564, 0.2810368041992187, 0.28097125244140625, 0.2808842163085937, 0.28110336303710937, 0.28096511840820315, 0.2808688659667969, 0.281027587890625, 0.2810245056152344, 0.28101119995117185, 0.2810521545410156, 0.2808965148925781, 0.281069580078125, 0.2811453552246094, 0.28131329345703127, 0.28105624389648437, 0.2812549133300781, 0.28084225463867185, 0.28120065307617187, 0.28095281982421877, 0.2810378112792969, 0.28143515014648435, 0.2811975708007812, 0.28118936157226565, 0.2811351013183594, 0.2813869934082031, 0.28113714599609374, 0.28141876220703127, 0.28113201904296875, 0.2815682678222656, 0.2815467529296875, 0.28107672119140625, 0.281275390625, 0.2812549133300781, 0.2810992736816406, 0.28117196655273435, 0.28111154174804687, 0.28140032958984373, 0.2811883544921875, 0.28148223876953127, 0.281238525390625, 0.2814341125488281, 0.28122830200195315, 0.28124465942382815, 0.2812200927734375, 0.2811576232910156, 0.2811269226074219, 0.2812252197265625, 0.2812303466796875, 0.28119961547851563, 0.28121908569335935, 0.28127743530273436, 0.2814617614746094, 0.6061670532226563, 0.2809108581542969, 0.28080230712890625, 0.2814484558105469, 0.28111358642578127, 0.28132147216796877, 0.28121701049804687, 0.2810378112792969, 0.2809323425292969, 0.2810368041992187, 0.2813255615234375, 0.28106341552734376, 0.28104702758789063, 0.28093438720703123, 0.28106036376953125, 0.281196533203125, 0.28113919067382814, 0.2808115234375, 0.2810900573730469, 0.2811146240234375, 0.2812098693847656, 0.281169921875, 0.2809169921875, 0.28105010986328127, 0.28096511840820315, 0.2809661560058594, 0.28101119995117185, 0.2809743347167969, 0.2809354248046875, 0.28098150634765623, 0.28091290283203124, 0.28116888427734377, 0.28102655029296875, 0.280943603515625, 0.281064453125, 0.2811545715332031, 0.2809794616699219, 0.2811832275390625, 0.28138189697265625, 0.2813788146972656, 0.281302001953125, 0.2813061218261719, 0.2811617431640625, 0.28119143676757813, 0.281169921875, 0.28117913818359375, 0.2815242309570313, 0.2812149658203125, 0.28117913818359375, 0.28131942749023436, 0.2813040771484375, 0.2812733459472656, 0.281275390625, 0.28110336303710937, 0.2812569580078125, 0.2811883544921875, 0.28134808349609375, 0.2814064636230469, 0.2813460388183594, 0.2813624267578125, 0.28149041748046877, 0.28145458984375, 0.28124365234375, 0.6059878540039062, 0.2807142333984375, 0.28078695678710935, 0.2806753234863281, 0.2809016418457031, 0.28094772338867186, 0.280922119140625, 0.2809876403808594, 0.28105523681640626, 0.2810040283203125, 0.2814208068847656, 0.28126617431640627, 0.2815948791503906, 0.281312255859375, 0.2813163452148438, 0.2814361572265625, 0.2813061218261719, 0.28089959716796875, 0.2814156799316406, 0.2811781005859375, 0.2810296325683594, 0.28103167724609374, 0.28088934326171877, 0.28106649780273435, 0.28101937866210935, 0.2810941467285156, 0.2810859375, 0.2814464111328125, 0.2812600402832031, 0.2811463623046875, 0.28102041625976565, 0.2810408935546875, 0.28117300415039065, 0.2810572814941406, 0.280922119140625, 0.28111871337890626, 0.28110336303710937, 0.28109619140625, 0.2811473999023438, 0.28106036376953125, 0.28131121826171873, 0.2811351013183594, 0.281248779296875, 0.28103067016601563, 0.28100506591796875, 0.2814095458984375, 0.2813429870605469, 0.2812538757324219, 0.2815825805664062, 0.2811955261230469, 0.28126412963867187, 0.28140237426757814, 0.2811934814453125, 0.28137777709960937, 0.28109619140625, 0.28129278564453125, 0.2812129211425781, 0.28120065307617187, 0.2812610473632812, 0.2811412353515625, 0.2812590026855469, 0.2810992736816406, 0.28120065307617187, 0.6068009033203124, 0.2809374694824219, 0.281122802734375, 0.2810439758300781, 0.28092825317382814, 0.2809692077636719, 0.280853515625, 0.281069580078125, 0.2810142822265625, 0.2809354248046875, 0.28104702758789063, 0.2809108581542969, 0.28106854248046875, 0.28096307373046875, 0.28088626098632813, 0.2809968566894531, 0.28083712768554686, 0.281101318359375, 0.2811217956542969, 0.28138290405273436, 0.281101318359375, 0.281059326171875, 0.2809938049316406, 0.2811545715332031, 0.28104193115234377, 0.2813081665039063, 0.28128460693359375, 0.28130508422851563, 0.2814761047363281, 0.28104193115234377, 0.2812200927734375, 0.28116888427734377, 0.2809876403808594, 0.2812538757324219, 0.28137985229492185, 0.28124774169921873, 0.2811678771972656, 0.28140032958984373, 0.2810624084472656, 0.28133273315429685, 0.2811002807617187, 0.28105831909179685, 0.2811576232910156, 0.2810726318359375, 0.28113919067382814, 0.2811269226074219, 0.28103372192382814, 0.28134808349609375, 0.28107672119140625, 0.2812047424316406, 0.2810511474609375, 0.2811412353515625, 0.2813204345703125, 0.2812037048339844, 0.2814218139648437, 0.2817884216308594, 0.28124978637695314, 0.281523193359375, 0.2813183898925781, 0.28128460693359375, 0.2815395812988281, 0.28148736572265626, 0.28139520263671874, 0.6065490112304688, 0.2811002807617187, 0.28092620849609373, 0.28095489501953125, 0.2808719482421875, 0.28084326171875, 0.2811023254394531, 0.281069580078125, 0.2810491027832031, 0.2809333801269531, 0.280995849609375, 0.2810798034667969, 0.2810992736816406, 0.2809169921875, 0.28097329711914065, 0.28101837158203125, 0.28094259643554687, 0.2809354248046875, 0.28100607299804686, 0.2810429382324219, 0.28095794677734376, 0.280880126953125, 0.2807767028808594, 0.28106649780273435, 0.2810859375, 0.28120574951171873, 0.2813941650390625, 0.28095281982421877, 0.2810931091308594, 0.2811975708007812, 0.28105831909179685, 0.28106341552734376, 0.2813562927246094, 0.28138803100585935, 0.2810921020507812, 0.28107879638671873, 0.2809671630859375, 0.28129791259765624, 0.28090573120117185, 0.28116070556640627, 0.2810798034667969, 0.28148837280273437, 0.2814392395019531, 0.2813061218261719, 0.2811269226074219, 0.2812876892089844, 0.28159078979492186, 0.2817607727050781, 0.2814699401855469, 0.2814566345214844, 0.2813061218261719, 0.2814146423339844, 0.28142489624023437, 0.28127947998046876, 0.28125595092773437, 0.2811934814453125, 0.28141055297851564, 0.2812682189941406, 0.2812600402832031, 0.2815703125, 0.2814689331054688, 0.28122113037109375, 0.28161843872070313, 0.6061782836914062, 0.2809692077636719, 0.28095693969726565, 0.2809661560058594, 0.28118014526367185, 0.2808770446777344, 0.2809989013671875, 0.28118118286132815, 0.28119961547851563, 0.281017333984375, 0.28093646240234377, 0.28108901977539064, 0.2810705871582031, 0.28113409423828123, 0.28105831909179685, 0.28111358642578127, 0.28118426513671874, 0.2812620849609375, 0.2812037048339844, 0.28109619140625, 0.28100607299804686, 0.281069580078125, 0.281059326171875, 0.28090573120117185, 0.2810521545410156, 0.28109515380859373, 0.28112997436523435, 0.28092108154296874, 0.2811074523925781, 0.2809927673339844, 0.28092724609375, 0.281491455078125, 0.28131942749023436, 0.281628662109375, 0.2811770935058594, 0.2811514892578125, 0.2810828857421875, 0.28110540771484377, 0.2809518127441406, 0.28126617431640627, 0.28103884887695313, 0.28114227294921873, 0.2810705871582031, 0.28101324462890626, 0.281322509765625, 0.28095489501953125, 0.2812886962890625, 0.2812037048339844, 0.2814197692871094, 0.2811985778808594, 0.28113714599609374, 0.2812467346191406, 0.28106854248046875, 0.2812129211425781, 0.28124978637695314, 0.2813388671875, 0.2813460388183594, 0.2814832763671875, 0.28139620971679685, 0.2815467529296875, 0.28133477783203126, 0.28151397705078124, 0.28122418212890626, 0.6063093872070312, 0.2810449829101562, 0.2814197692871094, 0.28110848999023436, 0.2810439758300781, 0.28139007568359375, 0.2809241638183594, 0.2809989013671875, 0.28083096313476563, 0.28112384033203125, 0.2810460205078125, 0.281302001953125, 0.28109515380859373, 0.2808883056640625, 0.28115966796875, 0.2810040283203125, 0.28084225463867185, 0.2809395141601562, 0.28089447021484376, 0.2809190368652344, 0.281059326171875, 0.28120779418945313, 0.2809241638183594, 0.28102655029296875, 0.28089752197265627, 0.2810101623535156, 0.28102859497070315, 0.28097537231445313, 0.28098355102539063, 0.2810992736816406, 0.2810654602050781, 0.28129995727539064, 0.2813839416503906, 0.2814535827636719, 0.2811074523925781, 0.28135833740234373, 0.28134707641601564, 0.2813183898925781, 0.28143515014648435, 0.2816296997070312, 0.28170855712890625, 0.2813061218261719, 0.2814228515625, 0.28137472534179686, 0.28131942749023436, 0.28127743530273436, 0.2817423400878906, 0.281481201171875, 0.2812590026855469, 0.2812630920410156, 0.2812231750488281, 0.2812467346191406, 0.28122113037109375, 0.2812037048339844, 0.2813644714355469, 0.281638916015625, 0.2814474182128906, 0.28127435302734377, 0.28134194946289065, 0.2812262268066406, 0.28141876220703127, 0.28152935791015626, 0.28133477783203126, 0.6068182983398438, 0.2809968566894531, 0.2809323425292969, 0.2809938049316406, 0.281027587890625, 0.28121701049804687, 0.28108389282226565, 0.2811351013183594, 0.28117401123046876, 0.28138494873046876, 0.28109722900390627, 0.2813368225097656, 0.28140850830078123, 0.28154779052734374, 0.28113409423828123, 0.2812129211425781, 0.28104806518554687, 0.2809405517578125, 0.2812672119140625, 0.281064453125, 0.28112588500976565, 0.28105523681640626, 0.28098150634765623, 0.2810439758300781, 0.28095590209960936, 0.28117300415039065, 0.28090777587890625, 0.2813061218261719, 0.2813839416503906, 0.28095077514648437, 0.2811576232910156, 0.2810654602050781, 0.2809599914550781, 0.28118118286132815, 0.2810726318359375, 0.28104193115234377, 0.28117300415039065, 0.28109515380859373, 0.281133056640625, 0.2813736877441406, 0.28131329345703127, 0.28105831909179685, 0.28130712890625, 0.2811975708007812, 0.28105624389648437, 0.2814986267089844, 0.281385986328125, 0.2814115905761719, 0.2812958679199219, 0.2812630920410156, 0.2812508239746094, 0.2813562927246094, 0.28127847290039065, 0.28153753662109376, 0.2815887451171875, 0.28126925659179686, 0.28113714599609374, 0.2811975708007812, 0.2814228515625, 0.28128564453125, 0.2811975708007812, 0.2811351013183594, 0.28143820190429686]",tokens/s,3.4987917120101386,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -9895,7 +10637,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664beea7-5eaa742e1b6dcac94a92386b;6942daa8-5021-4206-a63b-8492f0ca65ee) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3fb5-04fab0b34cf6f330727aecd1;4b070a13-ad61-4910-85df-5954d792a68e) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -9984,7 +10726,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1391-70146ae254f38334517043dc;a26d6935-75ae-420a-a26d-a0dc50e3b5e6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45aa-31d92a125d1e0e83161f6d1c;a798d76f-f7e9-4f21-b6e5-694f5bd2a9e7) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10044,7 +10786,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee7f-331221e56686f36417efc1e3;ad6c48ee-bf11-4cf7-b682-a4b331d24c30) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3f8f-57ee9248082d2b64319e1ff8;24e000b0-7b60-4b81-812d-5661667fc00f) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -10345,7 +11087,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664becf6-23243aa829ccfade7f157971;c73d19c1-8c3b-476d-863b-288462fea5f9) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3e0f-0faa211b7fb12e5864674569;a9f123fd-37ac-4db5-8f63-7d8d21405076) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -10430,7 +11172,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c131a-777f17940e72ca1f7ee6866f;a91ae6fe-3d63-41ba-adc7-89a000ba8263) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4537-4637550a21fa6ca1754a2eec;1745e3c5-03ec-484e-9d60-366ee85b1de2) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10546,7 +11288,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1213-67b578c05bc9c4142c300f43;76e36579-4cd7-42e7-b5db-d7c407be527f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4471-5243bad23663200c400e14f1;4d105ca7-327e-4c35-9135-1ee18878c7aa) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10620,6 +11362,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1852.063744,3368.5504,0.0,2722.103296,2585.357312,s,10,0.7361118698120117,0.07361118698120116,0.0003642046363895853,0.07361937713623047,0.07414464721679688,0.07420198898315429,0.07424786239624023,"[0.07425933074951171, 0.07323929595947265, 0.07386649322509765, 0.07366835021972656, 0.07318646240234375, 0.07325574493408203, 0.07413190460205078, 0.07326553344726562, 0.07365193939208985, 0.0735868148803711]",tokens/s,3477.732264599637,kWh,8.660967405115021e-07,4.745775110883439e-07,4.11042299636542e-06,5.451097247965265e-06,tokens/kWh,46963022.00360804,MB,1852.452864,3368.5504,0.0,2722.103296,2628.769792,s,10,38.912985107421875,3.8912985107421876,0.002125353517438014,3.89062548828125,3.8938447753906247,3.895329248046875,3.896516826171875,"[3.896813720703125, 3.89010009765625, 3.890788330078125, 3.890443603515625, 3.88959716796875, 3.890462646484375, 3.890949951171875, 3.890897216796875, 3.893514892578125, 3.88941748046875]",tokens/s,16.189968419560802,kWh,4.5877188203519475e-05,2.5143288266442475e-05,0.00021536239147063786,0.00028638286794059977,tokens/kWh,219985.22625685547,,s,629,39.51227389144898,0.06281760555079327,0.008801190571361054,0.061709312438964846,0.06187274169921875,0.06235340805053711,0.13552583007812502,"[0.06250291061401367, 0.0624824333190918, 0.06248860931396484, 0.06241788864135742, 0.06238412857055664, 0.062355457305908205, 0.06268928146362304, 0.06286848068237305, 0.061916160583496097, 0.061690879821777345, 0.06164377593994141, 0.06167552185058594, 0.06169702529907226, 0.061725696563720706, 0.06180147171020508, 0.06190796661376953, 0.06164275360107422, 0.06171340942382812, 0.06235033416748047, 0.06176665496826172, 0.06160793685913086, 0.06171750259399414, 0.061628414154052735, 0.06169497680664063, 0.06164275360107422, 0.061661182403564455, 0.06167244720458984, 0.0616806411743164, 0.061723648071289064, 0.061810817718505856, 0.06176448059082031, 0.06172876739501953, 0.06162124633789062, 0.06174105453491211, 0.06160793685913086, 0.06174924850463867, 0.06167859268188477, 0.0617625617980957, 0.06171340942382812, 0.062069759368896485, 0.06172172927856445, 0.06168972778320313, 0.06166220855712891, 0.06177689743041992, 0.06163558578491211, 0.06175948715209961, 0.061769729614257814, 0.06181683349609375, 0.06176665496826172, 0.061758464813232425, 0.06166425704956055, 0.061740032196044924, 0.061720577239990235, 0.06168371200561523, 0.06170640182495117, 0.06162825775146484, 0.061742080688476565, 0.06170009613037109, 0.06172774505615235, 0.061704193115234375, 0.061769729614257814, 0.06176870346069336, 0.13554074096679689, 0.06162432098388672, 0.06162432098388672, 0.06162134552001953, 0.06155356979370117, 0.06167654418945313, 0.06153932952880859, 0.06170521545410156, 0.06169804763793945, 0.06160486221313476, 0.06173798370361328, 0.061633537292480466, 0.06171964645385742, 0.06159964752197265, 0.061695999145507815, 0.061778942108154294, 0.06163148880004883, 0.061780990600585936, 0.061868030548095705, 0.06245068740844727, 0.06177689743041992, 0.06273433685302734, 0.06171750259399414, 0.061677566528320314, 0.06168678283691406, 0.061638656616210936, 0.061682689666748045, 0.06182297515869141, 0.061917182922363284, 0.06169702529907226, 0.06170624160766602, 0.06173286437988281, 0.061692928314208986, 0.061625343322753906, 0.06172774505615235, 0.06163455963134765, 0.061687808990478515, 0.061682815551757815, 0.06168051147460937, 0.061712383270263675, 0.06174822235107422, 0.061663230895996096, 0.06169001770019531, 0.061602657318115234, 0.06268313598632813, 0.06171955108642578, 0.061693984985351565, 0.061719520568847656, 0.06171955108642578, 0.061679615020751956, 0.06182809448242187, 0.06170521545410156, 0.061693950653076174, 0.06171340942382812, 0.06160179138183594, 0.06178508758544922, 0.06174822235107422, 0.061707393646240234, 0.061671295166015626, 0.06169721603393555, 0.06164358520507812, 0.06166527938842774, 0.061709312438964846, 0.13617971801757814, 0.06163046264648438, 0.06165401458740234, 0.06161407852172852, 0.06157516860961914, 0.061656063079833984, 0.06166016006469727, 0.061707263946533204, 0.061707263946533204, 0.06162944030761719, 0.06173388671875, 0.06161932754516602, 0.06166515350341797, 0.061652992248535154, 0.06172979354858398, 0.06173798370361328, 0.06173183822631836, 0.06171340942382812, 0.06171750259399414, 0.061639678955078124, 0.06162432098388672, 0.06173183822631836, 0.06170828628540039, 0.06276300811767578, 0.061687808990478515, 0.061636608123779295, 0.06173183822631836, 0.06161612701416016, 0.0616806411743164, 0.06161203384399414, 0.06170009613037109, 0.06170316696166992, 0.06180966567993164, 0.061895679473876954, 0.06174105453491211, 0.061710334777832034, 0.06173081588745117, 0.06171660614013672, 0.061677440643310544, 0.061674495697021485, 0.0617441291809082, 0.061692928314208986, 0.0617625617980957, 0.06161612701416016, 0.06178611373901367, 0.06169926452636719, 0.061711166381835936, 0.0628408317565918, 0.06211075210571289, 0.062286815643310546, 0.06192435073852539, 0.061740032196044924, 0.0617534065246582, 0.06165497589111328, 0.061690879821777345, 0.06172671890258789, 0.06164889526367188, 0.06174310302734375, 0.06171340942382812, 0.061661182403564455, 0.061671424865722656, 0.061740032196044924, 0.06165708923339844, 0.1359667205810547, 0.06163763046264648, 0.06169497680664063, 0.061628414154052735, 0.061633537292480466, 0.061764606475830076, 0.06156185531616211, 0.061704193115234375, 0.0617891845703125, 0.06193459320068359, 0.06180044937133789, 0.061663230895996096, 0.0622295036315918, 0.06165913772583008, 0.06168166351318359, 0.061690879821777345, 0.06174924850463867, 0.06168371200561523, 0.061770751953125, 0.0616519660949707, 0.06170111846923828, 0.06165913772583008, 0.061707263946533204, 0.0617441291809082, 0.06170009613037109, 0.06163148880004883, 0.06163558578491211, 0.061644798278808595, 0.06166425704956055, 0.06169190216064453, 0.061710334777832034, 0.06169497680664063, 0.061644798278808595, 0.06169497680664063, 0.061671520233154295, 0.06164572906494141, 0.06277836990356446, 0.06170009613037109, 0.06170009613037109, 0.06167859268188477, 0.06167660903930664, 0.06171231842041015, 0.06174924850463867, 0.06165094375610351, 0.06181478500366211, 0.061690879821777345, 0.06172671890258789, 0.06169190216064453, 0.06179020690917969, 0.06173388671875, 0.06178508758544922, 0.061685760498046874, 0.061876224517822265, 0.062241790771484375, 0.06170828628540039, 0.06173081588745117, 0.06157823944091797, 0.061774848937988285, 0.06169497680664063, 0.06177382278442383, 0.06168678283691406, 0.06180556869506836, 0.06197145462036133, 0.13539430236816405, 0.06167552185058594, 0.061671424865722656, 0.06164070510864258, 0.061591552734375, 0.06165401458740234, 0.06159667205810547, 0.06172876739501953, 0.06168678283691406, 0.06160281753540039, 0.061707263946533204, 0.06159667205810547, 0.0617154541015625, 0.06167244720458984, 0.061693950653076174, 0.0616519660949707, 0.06165094375610351, 0.06250291061401367, 0.06174720001220703, 0.06172159957885742, 0.06170009613037109, 0.061649921417236325, 0.061625343322753906, 0.06166732788085937, 0.06170111846923828, 0.061608959197998046, 0.06172774505615235, 0.06161305618286133, 0.06164889526367188, 0.061679615020751956, 0.06164070510864258, 0.06172265625, 0.061671390533447265, 0.061677566528320314, 0.061658111572265625, 0.061710334777832034, 0.06173183822631836, 0.061690879821777345, 0.061764606475830076, 0.06166835021972656, 0.06195308685302734, 0.06167443084716797, 0.06182809448242187, 0.06171244812011719, 0.061844417572021484, 0.06171750259399414, 0.06164582443237305, 0.06171340942382812, 0.06173593521118164, 0.06290636825561523, 0.06180249786376953, 0.06168678283691406, 0.061709312438964846, 0.06173183822631836, 0.061699073791503904, 0.061742080688476565, 0.061656063079833984, 0.061811710357666014, 0.061736961364746094, 0.061709312438964846, 0.061710334777832034, 0.061693950653076174, 0.06191308975219727, 0.13563392639160157, 0.0618526725769043, 0.061897727966308595, 0.06170316696166992, 0.06156803131103516, 0.061701087951660155, 0.06166128158569336, 0.06172867202758789, 0.0617154541015625, 0.061690879821777345, 0.06177996826171875, 0.06166527938842774, 0.06173183822631836, 0.061636608123779295, 0.0617523193359375, 0.06169702529907226, 0.0616673583984375, 0.061728736877441404, 0.061797374725341796, 0.06189875030517578, 0.0617441291809082, 0.06164275360107422, 0.061674495697021485, 0.06159360122680664, 0.06164796829223633, 0.06164368057250977, 0.061792255401611325, 0.06162739181518555, 0.061645023345947264, 0.06166812896728516, 0.06170115280151367, 0.06169699096679687, 0.061623294830322264, 0.06330368041992188, 0.06182092666625977, 0.06176051330566406, 0.06183628845214844, 0.06172774505615235, 0.06193971252441406, 0.061797374725341796, 0.0616888313293457, 0.06170329666137695, 0.06172966384887695, 0.06168166351318359, 0.06176371383666992, 0.061695873260498045, 0.061770751953125, 0.061723648071289064, 0.06177689743041992, 0.06173491287231445, 0.06175641632080078, 0.06165708923339844, 0.06165401458740234, 0.06177382278442383, 0.06164684677124024, 0.06175641632080078, 0.06164377593994141, 0.06168166351318359, 0.06175027084350586, 0.061699073791503904, 0.06170009613037109, 0.061710334777832034, 0.06166527938842774, 0.13643980407714842, 0.061716545104980466, 0.061607872009277344, 0.06166732788085937, 0.0616099853515625, 0.06168371200561523, 0.06158540725708008, 0.06174105453491211, 0.061677566528320314, 0.061644798278808595, 0.061795326232910154, 0.06170624160766602, 0.06171648025512695, 0.06165919876098633, 0.06173894500732422, 0.061732929229736326, 0.06168876647949219, 0.06168377685546875, 0.0616693115234375, 0.06164275360107422, 0.06164377593994141, 0.06178201675415039, 0.06199398422241211, 0.06298828887939453, 0.06168985748291016, 0.06167871856689453, 0.061652862548828125, 0.06170316696166992, 0.061723648071289064, 0.06167244720458984, 0.06170934295654297, 0.061823966979980466, 0.06195199966430664, 0.061687808990478515, 0.06180352020263672, 0.061764606475830076, 0.061685760498046874, 0.061674495697021485, 0.061693950653076174, 0.06170624160766602, 0.06166835021972656, 0.06171443176269531, 0.06177382278442383, 0.06163148880004883, 0.061774848937988285, 0.0616734733581543, 0.06170624160766602, 0.061707263946533204, 0.06180044937133789, 0.06171340942382812, 0.06172671890258789, 0.061739009857177736, 0.06166220855712891, 0.06182505416870117, 0.06167343902587891, 0.061732929229736326, 0.06168364715576172, 0.06287052917480469, 0.06172166442871094, 0.06172563171386719, 0.06172159957885742, 0.0617441291809082, 0.061709312438964846, 0.13548748779296876, 0.061685760498046874, 0.06184550476074219, 0.06171340942382812, 0.061636608123779295, 0.06169295883178711, 0.06171337509155273, 0.06171340942382812, 0.06171443176269531, 0.06175743865966797, 0.06177996826171875, 0.06164684677124024, 0.0617410888671875, 0.06167958450317383, 0.06168371200561523, 0.06167859268188477, 0.06231347274780273, 0.061669376373291014, 0.06177791976928711, 0.06170111846923828, 0.06169190216064453, 0.06170214462280273, 0.06165401458740234, 0.06170009613037109, 0.06169190216064453, 0.06161612701416016, 0.061682689666748045, 0.06170521545410156, 0.06170316696166992, 0.06166220855712891, 0.06175436782836914, 0.0616888313293457, 0.06170624160766602, 0.0616734733581543, 0.061725696563720706, 0.061830142974853515, 0.06189158248901367, 0.06173798370361328, 0.06172876739501953, 0.06174310302734375, 0.061740032196044924, 0.06168371200561523, 0.06173183822631836, 0.06170316696166992, 0.06178201675415039, 0.061722686767578125, 0.06168569564819336, 0.06171852874755859, 0.061712383270263675, 0.06175955200195313, 0.061817790985107424, 0.062246910095214845, 0.06236262512207031, 0.06189673614501953, 0.06172975921630859, 0.061764606475830076, 0.06163455963134765, 0.06180681610107422, 0.06167631912231445, 0.06176870346069336, 0.06178611373901367, 0.06176563262939453, 0.0617891845703125, 0.13594931030273438, 0.06170828628540039, 0.061732894897460935, 0.061699039459228514, 0.061656063079833984, 0.0616376953125, 0.06158432006835937, 0.06172467041015625, 0.06168473434448242, 0.06169497680664063, 0.06175641632080078, 0.06178620910644531, 0.061687713623046876, 0.0616806411743164, 0.06187007904052735, 0.06170214462280273, 0.06175641632080078, 0.06246201705932617, 0.06178400039672852, 0.061663230895996096, 0.06180352020263672, 0.06171750259399414, 0.06183731079101563, 0.061625343322753906, 0.061712383270263675, 0.061710334777832034, 0.06167244720458984, 0.06171443176269531, 0.06174723052978515, 0.06172361755371094, 0.061871135711669925, 0.061722591400146486, 0.06174617767333984, 0.061699073791503904, 0.061830142974853515, 0.06177996826171875, 0.061927425384521485, 0.06182195281982422, 0.061843456268310545, 0.061658206939697265, 0.06172662353515625, 0.06186086273193359, 0.06174310302734375, 0.06176153564453125, 0.06186905670166016, 0.06178815841674805, 0.06171340942382812, 0.06176563262939453, 0.061828289031982425, 0.0616824951171875, 0.061835262298583986, 0.0634439697265625, 0.06188544082641602, 0.06187519836425781, 0.06173081588745117, 0.06184243011474609, 0.06188544082641602, 0.06187212753295898, 0.06179020690917969, 0.06177587127685547, 0.0616888313293457, 0.06170521545410156, 0.0617441291809082, 0.13600973510742187, 0.062129150390625, 0.06186905670166016, 0.0616888313293457, 0.06164889526367188, 0.061658111572265625, 0.06158335876464844, 0.06171852874755859, 0.06171340942382812, 0.06164684677124024, 0.061730880737304684, 0.06165702438354492, 0.06167552185058594, 0.061641727447509766, 0.061739009857177736, 0.06163455963134765, 0.06174617767333984, 0.061677566528320314, 0.06170828628540039, 0.061725696563720706, 0.061671424865722656, 0.06169702529907226, 0.06174515151977539, 0.061661182403564455, 0.06163046264648438, 0.06156185531616211, 0.06298624038696289, 0.06177280044555664, 0.06197350311279297, 0.061710334777832034, 0.06174105453491211, 0.061628414154052735, 0.06164787292480469, 0.0615997428894043, 0.061770751953125, 0.06168678283691406, 0.06174515151977539, 0.06170111846923828, 0.06171955108642578, 0.061655040740966796, 0.06164377593994141, 0.06171852874755859, 0.061695999145507815, 0.061661182403564455, 0.06176153564453125, 0.06170316696166992, 0.061671424865722656, 0.061693023681640625, 0.061720481872558595, 0.06173593521118164, 0.06173081588745117, 0.061682689666748045, 0.061692928314208986, 0.06175539016723633, 0.061644798278808595, 0.06165708923339844, 0.06165404891967773, 0.06168368148803711, 0.06167552185058594, 0.061703231811523436, 0.061744064331054685, 0.06178303909301758, 0.06170316696166992]",tokens/s,15.919104066954871,, +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1490.198528,1841.823744,0.0,1195.37664,1022.155264,s,10,0.6871841278076172,0.06871841278076171,0.00033427755700674906,0.06862833786010741,0.06882823104858399,0.06926301155090332,0.06961083595275878,"[0.06969779205322266, 0.06856297302246094, 0.06858707427978515, 0.06850640106201172, 0.0686982421875, 0.06862300872802735, 0.0686336669921875, 0.06864895629882813, 0.0687316131591797, 0.06849440002441406]",tokens/s,3725.347976483958,kWh,8.097485796501648e-07,4.4370366405853753e-07,3.761924622931496e-06,5.0153768666401985e-06,tokens/kWh,51043023.64649507,MB,1490.198528,1841.823744,0.0,1195.37664,1091.597312,s,10,35.169312255859374,3.5169312255859375,0.00128683836079331,3.516599853515625,3.5173736328125,3.5190165283203125,3.5203308447265624,"[3.520659423828125, 3.51630859375, 3.515842041015625, 3.516979248046875, 3.5166650390625, 3.51653466796875, 3.516714599609375, 3.516245361328125, 3.517008544921875, 3.516354736328125]",tokens/s,17.91334432179688,kWh,4.1521653622797086e-05,2.2756081263433605e-05,0.0001967916524864682,0.0002610693873726989,tokens/kWh,241315.15622726805,,s,629,35.7307012748718,0.056805566414740584,0.008168059089268582,0.05580799865722656,0.055887052917480466,0.055967333984375,0.12452544342041016,"[0.05626367950439453, 0.05630361557006836, 0.05626367950439453, 0.056338432312011716, 0.05623910522460938, 0.05629337692260742, 0.05628313446044922, 0.056346622467041016, 0.05584896087646484, 0.05588991928100586, 0.05574758529663086, 0.055757823944091796, 0.05575884628295898, 0.055839744567871094, 0.05575680160522461, 0.055742462158203124, 0.0557209587097168, 0.05613260650634765, 0.05585203170776367, 0.0557844467163086, 0.055812095642089846, 0.05583769607543945, 0.05573529434204102, 0.05577523040771484, 0.05584588623046875, 0.05580595016479492, 0.0557946891784668, 0.055825408935546876, 0.05578137588500977, 0.05588684844970703, 0.055812095642089846, 0.055757823944091796, 0.05581414413452149, 0.05580595016479492, 0.05578137588500977, 0.05587353515625, 0.05582745742797852, 0.055779327392578126, 0.05571379089355469, 0.05586943817138672, 0.055806976318359375, 0.05578137588500977, 0.05583871841430664, 0.055760894775390625, 0.05582233428955078, 0.05581721496582031, 0.05575372695922851, 0.05586022567749024, 0.055790592193603515, 0.0557946891784668, 0.05580492782592773, 0.055793663024902344, 0.055809024810791016, 0.055818241119384764, 0.05578342437744141, 0.05583462524414062, 0.0558653450012207, 0.055798782348632815, 0.05585715103149414, 0.05593804931640625, 0.055790592193603515, 0.055820289611816405, 0.12452658843994141, 0.055755775451660154, 0.055787521362304686, 0.05579776000976563, 0.05577830505371094, 0.05574758529663086, 0.05581414413452149, 0.055787521362304686, 0.05573017501831055, 0.05575680160522461, 0.05580492782592773, 0.05572915267944336, 0.05578649520874023, 0.05576806259155274, 0.0557946891784668, 0.05575167846679688, 0.05578956985473633, 0.05579673767089844, 0.05581926345825195, 0.05581721496582031, 0.05590323257446289, 0.05579776000976563, 0.05584588623046875, 0.055739391326904295, 0.055790592193603515, 0.05578137588500977, 0.055836673736572265, 0.05580083084106445, 0.055825408935546876, 0.05577011108398437, 0.055809024810791016, 0.05585715103149414, 0.055793663024902344, 0.05574758529663086, 0.055787521362304686, 0.05576499176025391, 0.055809024810791016, 0.055766014099121096, 0.05583462524414062, 0.05594828796386719, 0.05584896087646484, 0.05576294326782227, 0.05584281539916992, 0.05577830505371094, 0.055858177185058595, 0.0560076789855957, 0.05585408020019531, 0.05578342437744141, 0.05575065612792969, 0.055859233856201174, 0.05579670333862305, 0.055782398223876956, 0.05586739349365234, 0.05582950210571289, 0.055809024810791016, 0.05578342437744141, 0.0558551025390625, 0.0561715202331543, 0.055932926177978515, 0.05579776000976563, 0.05574553680419922, 0.0558131217956543, 0.05583052825927735, 0.12455526733398438, 0.0557916145324707, 0.05582950210571289, 0.05575065612792969, 0.05580799865722656, 0.05575987243652344, 0.0558131217956543, 0.055723007202148435, 0.05578137588500977, 0.055793663024902344, 0.05577523040771484, 0.05576806259155274, 0.055795711517333986, 0.05574655914306641, 0.05582745742797852, 0.05574860763549805, 0.05578137588500977, 0.05572608184814453, 0.05576806259155274, 0.055809024810791016, 0.0557916145324707, 0.05582438278198242, 0.05580083084106445, 0.05575475311279297, 0.05576192092895508, 0.05580083084106445, 0.055844863891601565, 0.05580492782592773, 0.055823360443115234, 0.05579673767089844, 0.05576294326782227, 0.055785472869873044, 0.05581619262695312, 0.05576806259155274, 0.05587251281738281, 0.05579776000976563, 0.0557844467163086, 0.05576294326782227, 0.055839744567871094, 0.05573017501831055, 0.055815166473388675, 0.0558766098022461, 0.05590220642089844, 0.05579776000976563, 0.055818241119384764, 0.055841793060302736, 0.05584588623046875, 0.05583564758300781, 0.05581619262695312, 0.05581619262695312, 0.05587148666381836, 0.05586636734008789, 0.055820289611816405, 0.05580595016479492, 0.05584998321533203, 0.055793663024902344, 0.055785472869873044, 0.05583052825927735, 0.055785472869873044, 0.0557916145324707, 0.05583564758300781, 0.05590323257446289, 0.05586636734008789, 0.12449382019042969, 0.05575372695922851, 0.05576499176025391, 0.05578649520874023, 0.05577318572998047, 0.05608038330078125, 0.0558766098022461, 0.055744510650634765, 0.055757823944091796, 0.05576704025268555, 0.05586739349365234, 0.0557762565612793, 0.05626163101196289, 0.05574860763549805, 0.05581107330322266, 0.05573836898803711, 0.0557916145324707, 0.05576396942138672, 0.05574348831176758, 0.05576499176025391, 0.0557844467163086, 0.05575475311279297, 0.05585203170776367, 0.05574041748046875, 0.05575884628295898, 0.05578342437744141, 0.05584998321533203, 0.0558131217956543, 0.05589913558959961, 0.05588684844970703, 0.05580492782592773, 0.0557762565612793, 0.05579776000976563, 0.05576806259155274, 0.05586431884765625, 0.0557916145324707, 0.05583052825927735, 0.055790592193603515, 0.05582438278198242, 0.05580799865722656, 0.055820289611816405, 0.05578137588500977, 0.055798782348632815, 0.0558551025390625, 0.055766014099121096, 0.055825408935546876, 0.055982078552246094, 0.055766014099121096, 0.055793663024902344, 0.05594214248657227, 0.05587251281738281, 0.055806976318359375, 0.05583257675170898, 0.055965694427490234, 0.05590630340576172, 0.0557946891784668, 0.05584588623046875, 0.05582438278198242, 0.05581926345825195, 0.05581414413452149, 0.055809024810791016, 0.055823360443115234, 0.05583564758300781, 0.12459212493896485, 0.05578956985473633, 0.05584588623046875, 0.055777278900146485, 0.055769088745117185, 0.05573734283447265, 0.055818241119384764, 0.055785472869873044, 0.05583052825927735, 0.0557844467163086, 0.05582438278198242, 0.055782398223876956, 0.05585100936889648, 0.05584281539916992, 0.05582131195068359, 0.05576499176025391, 0.0557844467163086, 0.055757823944091796, 0.05580287933349609, 0.055757823944091796, 0.055841793060302736, 0.05579673767089844, 0.055818241119384764, 0.055736320495605465, 0.055801856994628904, 0.05581721496582031, 0.05588582229614258, 0.05578649520874023, 0.05585919952392578, 0.055790592193603515, 0.055836673736572265, 0.05586329650878906, 0.055874561309814455, 0.0558653450012207, 0.055793663024902344, 0.05587046432495117, 0.05577523040771484, 0.05574860763549805, 0.055839744567871094, 0.05576396942138672, 0.05587558364868164, 0.05575884628295898, 0.055861248016357425, 0.055810047149658204, 0.056005630493164066, 0.055798782348632815, 0.055818241119384764, 0.055913471221923826, 0.05584691238403321, 0.05581721496582031, 0.05577830505371094, 0.05587865447998047, 0.05581414413452149, 0.05581414413452149, 0.0557916145324707, 0.05582339096069336, 0.05581923294067383, 0.05581107330322266, 0.055856128692626954, 0.05586636734008789, 0.05587148666381836, 0.055801856994628904, 0.05587558364868164, 0.12469657897949218, 0.055790592193603515, 0.055731201171875, 0.05574860763549805, 0.055801856994628904, 0.055787521362304686, 0.0557844467163086, 0.0557496337890625, 0.055774208068847655, 0.0557844467163086, 0.05589913558959961, 0.055757823944091796, 0.055774208068847655, 0.05576806259155274, 0.05582745742797852, 0.05572710418701172, 0.055798782348632815, 0.055744510650634765, 0.055760894775390625, 0.05584281539916992, 0.05578035354614258, 0.05584384155273438, 0.05585919952392578, 0.05575680160522461, 0.055766014099121096, 0.05579673767089844, 0.055812095642089846, 0.0557946891784668, 0.05587353515625, 0.055825408935546876, 0.055828479766845705, 0.055782398223876956, 0.055831550598144535, 0.055787521362304686, 0.055877632141113284, 0.055788543701171874, 0.05577318572998047, 0.0557916145324707, 0.055841793060302736, 0.055771137237548826, 0.055823360443115234, 0.05577318572998047, 0.05582131195068359, 0.055801856994628904, 0.05582950210571289, 0.055844863891601565, 0.05596672058105469, 0.055787521362304686, 0.05583052825927735, 0.055897087097167966, 0.055853057861328125, 0.05585100936889648, 0.05584896087646484, 0.05590220642089844, 0.055806976318359375, 0.056011775970458984, 0.05590118408203125, 0.05583564758300781, 0.055809024810791016, 0.055826431274414064, 0.055774208068847655, 0.055810047149658204, 0.055856128692626954, 0.12463922882080078, 0.05575680160522461, 0.055785472869873044, 0.05579673767089844, 0.05581107330322266, 0.05583257675170898, 0.055858177185058595, 0.055795711517333986, 0.055752704620361325, 0.055806976318359375, 0.055839744567871094, 0.05576294326782227, 0.05579980850219726, 0.05575065612792969, 0.055801856994628904, 0.05574860763549805, 0.05582438278198242, 0.055736320495605465, 0.05581107330322266, 0.05579673767089844, 0.055809024810791016, 0.055801856994628904, 0.05583769607543945, 0.05578956985473633, 0.05581619262695312, 0.05584998321533203, 0.05587046432495117, 0.05583359909057617, 0.055967742919921876, 0.055815166473388675, 0.05589503860473633, 0.055825408935546876, 0.05581619262695312, 0.05591551971435547, 0.055801856994628904, 0.05583769607543945, 0.05586329650878906, 0.05574860763549805, 0.05582131195068359, 0.05574041748046875, 0.05583769607543945, 0.055839744567871094, 0.05586841583251953, 0.05581721496582031, 0.05584588623046875, 0.05578137588500977, 0.055820289611816405, 0.055841793060302736, 0.05583462524414062, 0.0557916145324707, 0.055788543701171874, 0.05590118408203125, 0.055795711517333986, 0.05579673767089844, 0.0557844467163086, 0.05580492782592773, 0.055831550598144535, 0.05579264068603516, 0.0558551025390625, 0.05587251281738281, 0.055856128692626954, 0.05582032012939453, 0.05584585571289063, 0.12463410949707031, 0.05575065612792969, 0.05573734283447265, 0.05575475311279297, 0.05578342437744141, 0.055742462158203124, 0.055790592193603515, 0.05573734283447265, 0.05578649520874023, 0.05577830505371094, 0.0558551025390625, 0.055788543701171874, 0.0557946891784668, 0.055787521362304686, 0.055820289611816405, 0.055739391326904295, 0.055820289611816405, 0.05573427200317383, 0.05605068969726563, 0.055803905487060546, 0.05579776000976563, 0.055766014099121096, 0.055853057861328125, 0.055741439819335936, 0.055752704620361325, 0.05580083084106445, 0.05584076690673828, 0.0558131217956543, 0.05589606475830078, 0.05578137588500977, 0.05597081756591797, 0.055836673736572265, 0.05578956985473633, 0.05579980850219726, 0.05587353515625, 0.05579673767089844, 0.055774208068847655, 0.05579980850219726, 0.055815166473388675, 0.05575372695922851, 0.05580492782592773, 0.05586022567749024, 0.055844863891601565, 0.055806976318359375, 0.055806976318359375, 0.055812095642089846, 0.05589606475830078, 0.055772159576416014, 0.05577318572998047, 0.05588684844970703, 0.05584793472290039, 0.05581721496582031, 0.055810047149658204, 0.05586329650878906, 0.05579673767089844, 0.055772159576416014, 0.05586227035522461, 0.05584281539916992, 0.055785472869873044, 0.05581619262695312, 0.0557946891784668, 0.05583257675170898, 0.05588787078857422, 0.1247088623046875, 0.05576806259155274, 0.05583462524414062, 0.055766014099121096, 0.055772159576416014, 0.05578956985473633, 0.05581414413452149, 0.05579673767089844, 0.055742462158203124, 0.0557762565612793, 0.05581619262695312, 0.05573529434204102, 0.05583564758300781, 0.05576704025268555, 0.055815166473388675, 0.05571583938598633, 0.05580595016479492, 0.05576704025268555, 0.055790592193603515, 0.055782398223876956, 0.05589811325073242, 0.0557946891784668, 0.055809024810791016, 0.05574758529663086, 0.055798782348632815, 0.05577830505371094, 0.05583052825927735, 0.05578137588500977, 0.05585408020019531, 0.055790592193603515, 0.05580799865722656, 0.05587251281738281, 0.0557844467163086, 0.055795711517333986, 0.05579673767089844, 0.05584588623046875, 0.055782398223876956, 0.055806976318359375, 0.05589606475830078, 0.055777278900146485, 0.055844863891601565, 0.0557762565612793, 0.056202239990234375, 0.05586431884765625, 0.05582438278198242, 0.05581414413452149, 0.055820289611816405, 0.05581926345825195, 0.055769088745117185, 0.05581619262695312, 0.05580595016479492, 0.05586943817138672, 0.05609164810180664, 0.055853057861328125, 0.0558551025390625, 0.055856128692626954, 0.056139774322509765, 0.05579980850219726, 0.05579264068603516, 0.05577318572998047, 0.05584384155273438, 0.055788543701171874, 0.05585408020019531, 0.12452249908447266, 0.055742462158203124, 0.05573734283447265, 0.055742462158203124, 0.055787521362304686, 0.05576806259155274, 0.0557946891784668, 0.05572608184814453, 0.055795711517333986, 0.055779327392578126, 0.055890945434570315, 0.055760894775390625, 0.05575475311279297, 0.05573222351074219, 0.055815166473388675, 0.055723007202148435, 0.055858177185058595, 0.05573529434204102, 0.055725055694580077, 0.055790592193603515, 0.05591961669921875, 0.05581619262695312, 0.05587865447998047, 0.05575884628295898, 0.05578956985473633, 0.05578956985473633, 0.05581414413452149, 0.055790592193603515, 0.05590937423706055, 0.05582438278198242, 0.055806976318359375, 0.05578342437744141, 0.055907329559326174, 0.055831550598144535, 0.05587558364868164, 0.055785472869873044, 0.05579264068603516, 0.055812095642089846, 0.05582950210571289, 0.05575372695922851, 0.05583871841430664, 0.0558551025390625, 0.055823360443115234, 0.05580287933349609, 0.0557946891784668, 0.05583359909057617, 0.05591756820678711, 0.055779327392578126, 0.055826431274414064, 0.05596160125732422, 0.055982078552246094, 0.055801856994628904, 0.05582233428955078, 0.05587251281738281, 0.05585100936889648, 0.0557762565612793, 0.055826431274414064, 0.055815166473388675, 0.05581926345825195, 0.05585919952392578, 0.05578137588500977, 0.05581619262695312, 0.05582950210571289]",tokens/s,17.60390861520409,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -10689,7 +11432,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1456-197e78d2202a675007ebb633;f4a217f6-29b5-4354-8092-ab4df8a61acb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d466c-7ae192436aed31d34a93895d;ae5f62be-f893-4d4e-891f-3b644203b165) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10760,7 +11503,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c11eb-0f21c6f43e3eda67701cf190;41ce62fd-1eb2-4dc1-95b5-dcdbc0e24b4f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d444a-491f67ea57b8edd56842aa9d;ac6b51aa-eadf-4baf-b50e-37a2db71530e) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10831,7 +11574,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12b6-736d7698680864a45e7fc9ff;bda1052f-34a0-49df-9365-c050772748e5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44e8-4559000a77f06e566e316fc3;6ad9bda6-325a-428a-ab51-8942434f889e) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10862,8 +11605,140 @@ OSError: r is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5008.822272,8173.125632,0.0,7526.678528,6833.744896,s,10,3.2637095642089844,0.3263709564208984,0.0007368089127270137,0.3261606750488281,0.32729655456542966,0.3276723556518555,0.3279729965209961,"[0.32596209716796876, 0.32586428833007813, 0.3263592529296875, 0.32655685424804687, 0.326583251953125, 0.32592489624023435, 0.32721304321289063, 0.32552175903320313, 0.32567596435546875, 0.32804815673828125]",tokens/s,784.3835211545423,kWh,3.849155682816727e-06,2.109159659243408e-06,2.032348400070999e-05,2.6281799342770124e-05,tokens/kWh,9740581.177917834,MB,5008.822272,8173.125632,0.0,7526.678528,7110.584832,s,10,177.045357421875,17.7045357421875,0.0012845826660970417,17.70458984375,17.706225781249998,17.70631015625,17.70637765625,"[17.703931640625, 17.701494140625, 17.704443359375, 17.70496875, 17.70412890625, 17.70639453125, 17.70620703125, 17.704736328125, 17.70418359375, 17.704869140625]",tokens/s,3.558410167733434,kWh,0.00020893763350596564,0.00011451522852683938,0.0010782542004162915,0.0014017070624490963,tokens/kWh,44945.196958574845,,s,629,179.70182278442405,0.285694471835332,0.038836008660610016,0.2809876403808594,0.28119777221679687,0.2813642761230469,0.6074044165039063,"[0.2808002624511719, 0.2810859375, 0.2808913879394531, 0.280943603515625, 0.2809088134765625, 0.2810040283203125, 0.28078695678710935, 0.28094772338867186, 0.2809241638183594, 0.2809722900390625, 0.2809620361328125, 0.2810992736816406, 0.28092007446289063, 0.28101119995117185, 0.2808463439941406, 0.281027587890625, 0.2808821716308594, 0.2812037048339844, 0.2810091552734375, 0.2810931091308594, 0.28097125244140625, 0.2809722900390625, 0.2810142822265625, 0.280943603515625, 0.28114544677734377, 0.28108175659179685, 0.28101937866210935, 0.28090777587890625, 0.2809098205566406, 0.2811453552246094, 0.28107467651367185, 0.281064453125, 0.2811750793457031, 0.28106442260742187, 0.2810439758300781, 0.28124465942382815, 0.2813736877441406, 0.28123956298828123, 0.28094259643554687, 0.28106451416015626, 0.2809650573730469, 0.2809836120605469, 0.2809210205078125, 0.28089547729492187, 0.2809016418457031, 0.280958984375, 0.28140850830078123, 0.28111666870117186, 0.2809743347167969, 0.2808883056640625, 0.28079623413085936, 0.28105413818359376, 0.2809292907714844, 0.2810009460449219, 0.28097329711914065, 0.28109619140625, 0.2809098205566406, 0.28071221923828127, 0.280933349609375, 0.28117401123046876, 0.28113409423828123, 0.28092620849609373, 0.60747265625, 0.28070806884765626, 0.2811351013183594, 0.28077362060546873, 0.28089959716796875, 0.2807675476074219, 0.2810275268554687, 0.28092312622070315, 0.2809046936035156, 0.2808084411621094, 0.2810439758300781, 0.2808105163574219, 0.28097430419921876, 0.2808217468261719, 0.28103271484375, 0.28088729858398437, 0.2810675048828125, 0.28095693969726565, 0.28103167724609374, 0.28106036376953125, 0.28104193115234377, 0.2808299560546875, 0.2809108581542969, 0.28080743408203124, 0.28094873046875, 0.280958984375, 0.2808985595703125, 0.2808197021484375, 0.2809333801269531, 0.28095489501953125, 0.2808002624511719, 0.28090264892578126, 0.2809466857910156, 0.28094158935546876, 0.2808708801269531, 0.28107467651367185, 0.2808638000488281, 0.28104391479492186, 0.2809292907714844, 0.2809661560058594, 0.28094772338867186, 0.28103271484375, 0.28111053466796876, 0.2809088134765625, 0.28127651977539064, 0.2812466125488281, 0.2810214538574219, 0.2809333801269531, 0.2810347595214844, 0.28096408081054686, 0.28108184814453124, 0.2808186950683594, 0.2811372985839844, 0.28084515380859376, 0.28100827026367187, 0.28106539916992185, 0.2810152282714844, 0.2809989013671875, 0.2808842163085937, 0.2811985778808594, 0.2813501586914062, 0.2810224609375, 0.28091802978515623, 0.6079191284179688, 0.2806446228027344, 0.28111154174804687, 0.28073370361328126, 0.2810019836425781, 0.28069903564453125, 0.281046875, 0.2809036865234375, 0.28084225463867185, 0.2807582702636719, 0.28101119995117185, 0.28089959716796875, 0.28108697509765623, 0.28107672119140625, 0.280932373046875, 0.28086679077148435, 0.2810224609375, 0.28107571411132815, 0.2811197509765625, 0.28087091064453124, 0.2810859375, 0.2809395141601562, 0.2808412170410156, 0.28205978393554687, 0.2808739929199219, 0.28123648071289065, 0.28117913818359375, 0.28086273193359373, 0.28080435180664065, 0.2808360900878906, 0.28089242553710936, 0.28096820068359374, 0.2810460205078125, 0.28101544189453126, 0.28080010986328124, 0.2809333801269531, 0.280880126953125, 0.28094464111328127, 0.28085147094726565, 0.2810408935546875, 0.28090060424804686, 0.28079513549804686, 0.28095794677734376, 0.28093438720703123, 0.28107467651367185, 0.28103067016601563, 0.28095693969726565, 0.28259121704101564, 0.2810726318359375, 0.28107366943359374, 0.2812630920410156, 0.2811709289550781, 0.2813378601074219, 0.2809466857910156, 0.28098046875, 0.28088116455078127, 0.2810010681152344, 0.2809774169921875, 0.28094351196289064, 0.28106137084960936, 0.28105523681640626, 0.28106649780273435, 0.28105010986328127, 0.6080983276367188, 0.2805442810058594, 0.2809896545410156, 0.2807879638671875, 0.2809374694824219, 0.28062310791015627, 0.28097537231445313, 0.28071218872070314, 0.280995849609375, 0.2809415588378906, 0.2810777587890625, 0.2808299560546875, 0.28200244140625, 0.28103067016601563, 0.28094259643554687, 0.28076031494140624, 0.28103372192382814, 0.28095693969726565, 0.28099789428710936, 0.28095590209960936, 0.28118731689453125, 0.28101324462890626, 0.2808268737792969, 0.28090060424804686, 0.2810726318359375, 0.28087408447265627, 0.2808984680175781, 0.2808555603027344, 0.2809251708984375, 0.28087100219726563, 0.28081143188476565, 0.28094772338867186, 0.281238525390625, 0.28138494873046876, 0.2810378112792969, 0.28088116455078127, 0.2808719482421875, 0.28103988647460937, 0.28095794677734376, 0.28113714599609374, 0.28095693969726565, 0.2808350830078125, 0.2807224426269531, 0.2808401794433594, 0.283072509765625, 0.2809395141601562, 0.28103167724609374, 0.2809651489257812, 0.2811422424316406, 0.2808913879394531, 0.280922119140625, 0.2810408935546875, 0.28119448852539064, 0.2809671630859375, 0.28113409423828123, 0.28099072265625, 0.28125595092773437, 0.280995849609375, 0.2812262268066406, 0.2814044189453125, 0.28109722900390627, 0.28100811767578127, 0.2810019836425781, 0.6083676147460938, 0.281069580078125, 0.2809661865234375, 0.2808504028320313, 0.2810521545410156, 0.28073675537109377, 0.28086477661132814, 0.2808186950683594, 0.2809518127441406, 0.2808637390136719, 0.28104193115234377, 0.28080743408203124, 0.28095489501953125, 0.2810368041992187, 0.28096408081054686, 0.28084844970703127, 0.2810756530761719, 0.28226763916015624, 0.2809692077636719, 0.28088626098632813, 0.28098355102539063, 0.2808197021484375, 0.2807982177734375, 0.28103277587890624, 0.28112786865234374, 0.280961181640625, 0.2809997863769531, 0.28112399291992185, 0.2811482543945312, 0.2809599914550781, 0.28090060424804686, 0.2811207580566406, 0.28099172973632813, 0.2809938049316406, 0.2809968566894531, 0.28118731689453125, 0.280995849609375, 0.2809426574707031, 0.28101626586914064, 0.2809979553222656, 0.28103265380859377, 0.280943603515625, 0.28089447021484376, 0.28097329711914065, 0.2810726318359375, 0.28112384033203125, 0.28100811767578127, 0.2809580078125, 0.28122515869140624, 0.28084429931640625, 0.2810142822265625, 0.2808360900878906, 0.2810214538574219, 0.2808760375976562, 0.2810133666992187, 0.2810059509277344, 0.28116583251953126, 0.28140237426757814, 0.2807633972167969, 0.28105523681640626, 0.28104702758789063, 0.28084735107421877, 0.2810245056152344, 0.6076272583007812, 0.28068353271484375, 0.28097637939453124, 0.2806927490234375, 0.28078695678710935, 0.28059033203125, 0.28082278442382813, 0.2807767028808594, 0.2808760375976562, 0.28086578369140625, 0.2808350830078125, 0.28088626098632813, 0.2813931579589844, 0.2808126525878906, 0.28106640625, 0.28102655029296875, 0.2811617431640625, 0.28116583251953126, 0.2809968566894531, 0.28094259643554687, 0.28096307373046875, 0.28095489501953125, 0.28097637939453124, 0.28106137084960936, 0.2811781005859375, 0.28100607299804686, 0.28082177734375, 0.28104290771484375, 0.28098150634765623, 0.281091064453125, 0.2810624084472656, 0.28119244384765624, 0.28088729858398437, 0.2811064453125, 0.2811351013183594, 0.2812886962890625, 0.28093438720703123, 0.28102554321289064, 0.28370022583007815, 0.2810654602050781, 0.2810654602050781, 0.280869873046875, 0.2810214538574219, 0.28114227294921873, 0.28109515380859373, 0.28090060424804686, 0.28101324462890626, 0.2810245056152344, 0.28122726440429685, 0.2809302978515625, 0.28107366943359374, 0.28088934326171877, 0.28120883178710937, 0.28091390991210935, 0.2811975708007812, 0.2811269226074219, 0.2810368041992187, 0.2808280639648438, 0.2809659729003906, 0.2812037048339844, 0.28127130126953126, 0.28098968505859373, 0.28096511840820315, 0.60883251953125, 0.28060467529296873, 0.2809794616699219, 0.28092312622070315, 0.2813982849121094, 0.2813009948730469, 0.2809886779785156, 0.2808883056640625, 0.28107366943359374, 0.280922119140625, 0.2809456787109375, 0.2808084411621094, 0.2810163269042969, 0.28097537231445313, 0.2810071105957031, 0.2808258666992188, 0.2810491027832031, 0.28098968505859373, 0.2814863891601562, 0.2812784118652344, 0.281064453125, 0.28083511352539064, 0.28078179931640623, 0.28249298095703124, 0.2810910034179688, 0.2808832092285156, 0.28091494750976564, 0.2809292907714844, 0.28089141845703125, 0.28101425170898436, 0.2809743347167969, 0.2809722900390625, 0.2810408935546875, 0.28118426513671874, 0.2812303466796875, 0.2811094970703125, 0.28078387451171877, 0.28088626098632813, 0.28119451904296877, 0.28095895385742187, 0.28112588500976565, 0.28113006591796874, 0.2810581970214844, 0.280953857421875, 0.28095077514648437, 0.28101223754882815, 0.281027587890625, 0.28109515380859373, 0.2811064453125, 0.28092825317382814, 0.28113818359375, 0.28082891845703123, 0.28119143676757813, 0.28105426025390623, 0.2811329956054687, 0.28106854248046875, 0.2810572814941406, 0.28092724609375, 0.2808821716308594, 0.2811207580566406, 0.2811412353515625, 0.2810296325683594, 0.28099172973632813, 0.60706103515625, 0.2807059936523438, 0.2810726318359375, 0.2809098205566406, 0.28105624389648437, 0.2806220703125, 0.2809917907714844, 0.28088113403320314, 0.28103372192382814, 0.28095281982421877, 0.28107879638671873, 0.2810849304199219, 0.2812733459472656, 0.2809743347167969, 0.2812303466796875, 0.28095281982421877, 0.2824908752441406, 0.2809866333007813, 0.2809456787109375, 0.280880126953125, 0.2809169921875, 0.2809302978515625, 0.2807879638671875, 0.2809661560058594, 0.2810798034667969, 0.28095489501953125, 0.28097537231445313, 0.2809948120117188, 0.2809794616699219, 0.28092724609375, 0.28095794677734376, 0.2811883544921875, 0.28089959716796875, 0.2809692077636719, 0.28097552490234373, 0.2809906005859375, 0.28083914184570313, 0.2809456787109375, 0.2810152893066406, 0.28097125244140625, 0.2810501708984375, 0.2807622985839844, 0.280890380859375, 0.28096511840820315, 0.2809323425292969, 0.28090264892578126, 0.28089752197265627, 0.28092108154296874, 0.2811678771972656, 0.2809251708984375, 0.2810941467285156, 0.2808832092285156, 0.2812651672363281, 0.2808053894042969, 0.281017333984375, 0.28113409423828123, 0.28139724731445315, 0.28102041625976565, 0.2808330383300781, 0.2810378112792969, 0.2812538757324219, 0.28141259765625, 0.2812620849609375, 0.6072289428710937, 0.28075930786132813, 0.28142489624023437, 0.28085861206054685, 0.280919189453125, 0.2806700439453125, 0.28098770141601564, 0.2808155517578125, 0.2809251708984375, 0.28085861206054685, 0.2810214538574219, 0.28077163696289065, 0.280939453125, 0.2808053894042969, 0.2809876403808594, 0.2808883056640625, 0.28102349853515624, 0.2819471435546875, 0.28139111328125, 0.2810654602050781, 0.28137985229492185, 0.28105010986328127, 0.28085861206054685, 0.2810859375, 0.28121600341796876, 0.2809887390136719, 0.28093536376953127, 0.280958984375, 0.28101223754882815, 0.2809333801269531, 0.2809938049316406, 0.28112384033203125, 0.28098458862304687, 0.281059326171875, 0.2811770935058594, 0.28115045166015623, 0.28095794677734376, 0.2810523071289063, 0.28106121826171876, 0.2809232177734375, 0.28085751342773435, 0.28083096313476563, 0.2809395141601562, 0.2809108581542969, 0.2809722900390625, 0.2809190673828125, 0.28105520629882813, 0.28091390991210935, 0.2810091552734375, 0.2810040283203125, 0.28116790771484373, 0.28094573974609377, 0.2811063232421875, 0.280995849609375, 0.2812528686523437, 0.2810511474609375, 0.2810040283203125, 0.2809139709472656, 0.2808759765625, 0.2811085205078125, 0.2810316467285156, 0.28102349853515624, 0.2808760375976562, 0.6096005249023437, 0.28075726318359373, 0.2809169921875, 0.28073370361328126, 0.2812037048339844, 0.2810357666015625, 0.2810040283203125, 0.2807848815917969, 0.281017333984375, 0.2811248779296875, 0.28101223754882815, 0.2807848815917969, 0.28090673828125, 0.28089547729492187, 0.2810491027832031, 0.28088729858398437, 0.2809088134765625, 0.28079617309570315, 0.28105523681640626, 0.28090060424804686, 0.28096820068359374, 0.28085861206054685, 0.28080230712890625, 0.280880126953125, 0.28098355102539063, 0.2810572814941406, 0.2831247253417969, 0.281059326171875, 0.2810521545410156, 0.28110848999023436, 0.2811576232910156, 0.2811760559082031, 0.2811463623046875, 0.280911865234375, 0.2809190368652344, 0.2809876403808594, 0.28097024536132814, 0.2809886779785156, 0.2810624084472656, 0.28105523681640626, 0.28116378784179685, 0.28104193115234377, 0.28093646240234377, 0.2808760375976562, 0.28102859497070315, 0.28101837158203125, 0.2808401794433594, 0.28090573120117185, 0.28106854248046875, 0.28096820068359374, 0.2809661560058594, 0.2807940979003906, 0.28126412963867187, 0.28086273193359373, 0.28110540771484377, 0.2811617431640625, 0.281143310546875, 0.28087295532226564, 0.28085043334960935, 0.2810019836425781, 0.2811822204589844, 0.28111871337890626, 0.28107366943359374]",tokens/s,3.5002427368506384,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4708.687872,21498.42944,0.0,20851.982336,20236.731392,s,10,15.782414306640625,1.5782414306640624,0.0005368111432327376,1.5781434326171875,1.5784585083007812,1.5790858825683594,1.579587781982422,"[1.578103759765625, 1.5775478515625, 1.5778953857421876, 1.5797132568359376, 1.57818310546875, 1.5782744140625, 1.57828564453125, 1.578103759765625, 1.578319091796875, 1.577988037109375]",tokens/s,162.20585458353173,kWh,1.8640190545055602e-05,1.0214868406037568e-05,9.259768518919964e-05,0.0001214527441402928,tokens/kWh,2107815.6925321394,MB,4708.687872,21498.42944,0.0,20851.982336,20339.707392,s,10,881.7457578125,88.17457578125,0.002115503395001207,88.17539453124999,88.17674609375,88.17697460937501,88.17715742187501,"[88.17125, 88.17175, 88.1749609375, 88.175828125, 88.177203125, 88.175890625, 88.1766953125, 88.173328125, 88.176578125, 88.1722734375]",tokens/s,0.714491671117251,kWh,0.001040854390843047,0.0005704774625729078,0.005178987532075602,0.006790319385491556,tokens/kWh,9277.914104395162,,s,629,894.5504398193359,1.4221787596491826,0.18751446646952302,1.3995601806640625,1.4000109375,1.4001393798828126,2.978033017578125,"[1.399194580078125, 1.399298095703125, 1.3991290283203126, 1.3992857666015626, 1.3990655517578126, 1.399204833984375, 1.3991751708984375, 1.3994710693359376, 1.3990738525390625, 1.3990277099609374, 1.39915771484375, 1.399203857421875, 1.3990020751953125, 1.39932470703125, 1.39911376953125, 1.3995806884765625, 1.399319580078125, 1.3996319580078125, 1.399352294921875, 1.3991802978515624, 1.399151611328125, 1.39943017578125, 1.3990972900390626, 1.3995540771484376, 1.3996278076171875, 1.399510009765625, 1.39930322265625, 1.3995018310546874, 1.399568359375, 1.39934619140625, 1.399161865234375, 1.3999493408203125, 1.3994700927734376, 1.399572509765625, 1.39960107421875, 1.399762939453125, 1.3996871337890624, 1.399688232421875, 1.3997169189453125, 1.4000067138671874, 1.399546875, 1.399795654296875, 1.399826416015625, 1.3999605712890626, 1.3998192138671874, 1.399867431640625, 1.399826416015625, 1.399509033203125, 1.399413818359375, 1.3999349365234375, 1.399510986328125, 1.3996646728515625, 1.399531494140625, 1.4000762939453124, 1.3997393798828126, 1.40002099609375, 1.39964208984375, 1.39989306640625, 1.399784423828125, 1.4000015869140625, 1.4001776123046874, 1.4001162109375, 2.977680419921875, 1.3987635498046875, 1.3990216064453125, 1.398865966796875, 1.399183349609375, 1.3990440673828124, 1.399251953125, 1.399235595703125, 1.399419921875, 1.3989498291015625, 1.3994588623046875, 1.3995325927734374, 1.39915576171875, 1.399416748046875, 1.3992857666015626, 1.399309326171875, 1.3996400146484376, 1.39901025390625, 1.3994168701171874, 1.39928369140625, 1.3992008056640626, 1.399120849609375, 1.39950390625, 1.3994332275390624, 1.3995540771484376, 1.399593017578125, 1.3996851806640624, 1.3992059326171875, 1.3998223876953124, 1.3995396728515626, 1.399499755859375, 1.3992294921875, 1.399562255859375, 1.399331787109375, 1.3999698486328125, 1.39949365234375, 1.3995172119140624, 1.3994915771484375, 1.399731201171875, 1.3996185302734374, 1.399804931640625, 1.3992376708984375, 1.3995550537109376, 1.399384033203125, 1.39986328125, 1.3994813232421874, 1.401822265625, 1.399848876953125, 1.400058837890625, 1.399477294921875, 1.399910400390625, 1.3995653076171874, 1.399931884765625, 1.3996185302734374, 1.399846923828125, 1.399761962890625, 1.3999759521484374, 1.3998602294921876, 1.3998929443359376, 1.39964208984375, 1.4000107421875, 1.399794677734375, 1.4000025634765625, 2.978083740234375, 1.398897705078125, 1.3991322021484376, 1.3995980224609375, 1.3994290771484375, 1.399194580078125, 1.399624755859375, 1.399552001953125, 1.3993800048828124, 1.3990799560546876, 1.399405517578125, 1.399300048828125, 1.3992899169921875, 1.3992335205078126, 1.39951513671875, 1.3993482666015624, 1.399700439453125, 1.399189453125, 1.3992960205078124, 1.3992008056640626, 1.3995550537109376, 1.399152587890625, 1.3994906005859375, 1.3997373046875, 1.399525390625, 1.3993953857421875, 1.3996871337890624, 1.3994229736328125, 1.39953662109375, 1.39951513671875, 1.3995233154296876, 1.3995028076171876, 1.399846923828125, 1.399299072265625, 1.3997659912109375, 1.3994556884765625, 1.3997015380859374, 1.39947412109375, 1.3996082763671875, 1.39980078125, 1.3997916259765626, 1.3995919189453125, 1.39965234375, 1.3994854736328124, 1.39974755859375, 1.400369140625, 1.3997813720703125, 1.3998275146484376, 1.3997486572265625, 1.3997486572265625, 1.39992578125, 1.3999349365234375, 1.3997034912109374, 1.3998878173828124, 1.39997900390625, 1.400004638671875, 1.4001182861328125, 1.399736328125, 1.400037353515625, 1.3998233642578124, 1.399794677734375, 1.399687255859375, 1.400122314453125, 2.979734619140625, 1.3991536865234375, 1.399083984375, 1.3987276611328125, 1.39987255859375, 1.3991597900390624, 1.3992386474609375, 1.399719970703125, 1.3995755615234375, 1.3991290283203126, 1.3991075439453124, 1.39930419921875, 1.3992325439453126, 1.3994864501953126, 1.39919775390625, 1.39900830078125, 1.399330810546875, 1.3992427978515625, 1.39966357421875, 1.39919873046875, 1.39934619140625, 1.39936767578125, 1.3994219970703126, 1.399363525390625, 1.3997291259765625, 1.3998028564453124, 1.399531494140625, 1.3992847900390626, 1.399729248046875, 1.3994423828125, 1.399742431640625, 1.3993021240234376, 1.39958984375, 1.3994332275390624, 1.399636962890625, 1.3993062744140625, 1.399445556640625, 1.3994761962890625, 1.399825439453125, 1.3996768798828125, 1.399867431640625, 1.3995806884765625, 1.3999698486328125, 1.3994281005859375, 1.4019144287109375, 1.39976806640625, 1.400162353515625, 1.3998765869140626, 1.3997568359375, 1.3996256103515625, 1.399816162109375, 1.3998919677734376, 1.399845947265625, 1.3995601806640625, 1.3998387451171874, 1.4001397705078125, 1.400004638671875, 1.4000650634765626, 1.40010693359375, 1.39981103515625, 1.3998050537109374, 1.3996778564453125, 1.4001243896484374, 2.97879443359375, 1.399288818359375, 1.3990625, 1.3989530029296875, 1.3993800048828124, 1.3989437255859376, 1.399468017578125, 1.39931640625, 1.3996451416015625, 1.3991168212890626, 1.39941064453125, 1.399319580078125, 1.39913525390625, 1.3994024658203126, 1.399150634765625, 1.3994117431640625, 1.399594970703125, 1.3994075927734375, 1.399793701171875, 1.399268310546875, 1.3994659423828124, 1.3996287841796875, 1.3994556884765625, 1.3993697509765626, 1.3996676025390624, 1.399447509765625, 1.3997608642578125, 1.3995079345703125, 1.399666748046875, 1.3995478515625, 1.3996953125, 1.399482421875, 1.3994608154296875, 1.3996329345703125, 1.3997855224609375, 1.39945166015625, 1.3995386962890626, 1.39945166015625, 1.3997578125, 1.3999656982421875, 1.4002125244140624, 1.3995765380859375, 1.4000814208984376, 1.3997659912109375, 1.39985302734375, 1.399521240234375, 1.399901123046875, 1.39972705078125, 1.400004638671875, 1.399415771484375, 1.3997598876953126, 1.399835693359375, 1.3999913330078124, 1.399677001953125, 1.4000362548828125, 1.4001612548828124, 1.4003128662109374, 1.39993603515625, 1.400158203125, 1.4000721435546875, 1.4003721923828125, 1.399899169921875, 1.3997158203125, 2.978334716796875, 1.39909326171875, 1.3990604248046874, 1.39932568359375, 1.399351318359375, 1.39915771484375, 1.3992540283203125, 1.3992642822265624, 1.399413818359375, 1.3991741943359375, 1.3993359375, 1.399224365234375, 1.3994229736328125, 1.39940966796875, 1.399456787109375, 1.39919873046875, 1.3996123046875, 1.3993431396484375, 1.3997291259765625, 1.3994117431640625, 1.3995386962890626, 1.3994404296875, 1.3996083984375, 1.399300048828125, 1.3998018798828125, 1.4001162109375, 1.399901123046875, 1.3998970947265625, 1.399531494140625, 1.399182373046875, 1.3994619140625, 1.399109619140625, 1.3996697998046874, 1.3993175048828126, 1.399574462890625, 1.3992847900390626, 1.39966162109375, 1.399520263671875, 1.401324462890625, 1.399489501953125, 1.399690185546875, 1.3993994140625, 1.399900146484375, 1.39951611328125, 1.399921630859375, 1.399678955078125, 1.4001673583984375, 1.39990625, 1.3998018798828125, 1.3995079345703125, 1.400015869140625, 1.3994803466796875, 1.399951416015625, 1.3997076416015626, 1.40027294921875, 1.3999246826171876, 1.3999246826171876, 1.399732177734375, 1.3998970947265625, 1.399783447265625, 1.4001387939453125, 1.3993502197265626, 1.4002431640625, 2.9787392578125, 1.398930419921875, 1.3988720703125, 1.3989642333984376, 1.3989478759765626, 1.3991597900390624, 1.3990850830078125, 1.400115234375, 1.3992796630859374, 1.3990533447265625, 1.39926220703125, 1.3995345458984374, 1.39920703125, 1.3993245849609375, 1.399505859375, 1.39932666015625, 1.3994383544921876, 1.399141357421875, 1.3998828125, 1.39945166015625, 1.3994423828125, 1.39908203125, 1.399351318359375, 1.401248779296875, 1.3996380615234374, 1.399615478515625, 1.3997967529296875, 1.3995499267578124, 1.399447509765625, 1.3995377197265626, 1.39965234375, 1.3995089111328125, 1.3995694580078124, 1.39930419921875, 1.3994619140625, 1.399435302734375, 1.399462890625, 1.3994710693359376, 1.39966259765625, 1.3994649658203124, 1.3999195556640625, 1.399552001953125, 1.399803955078125, 1.3996103515625, 1.399816162109375, 1.39976708984375, 1.400056884765625, 1.399909423828125, 1.39981103515625, 1.3995601806640625, 1.3999759521484374, 1.399825439453125, 1.4004111328125, 1.399698486328125, 1.3998018798828125, 1.4003282470703124, 1.4000977783203126, 1.4000067138671874, 1.399920654296875, 1.3997823486328125, 1.400100830078125, 1.3995079345703125, 1.4000343017578125, 2.979209228515625, 1.3990235595703124, 1.398892578125, 1.3989365234375, 1.39919970703125, 1.3989918212890624, 1.3991229248046875, 1.3992652587890626, 1.3993604736328125, 1.3993175048828126, 1.3992017822265626, 1.39951416015625, 1.3994075927734375, 1.399320556640625, 1.39930419921875, 1.39944140625, 1.4017669677734375, 1.3992376708984375, 1.399351318359375, 1.3993123779296874, 1.3993021240234376, 1.39919873046875, 1.399609375, 1.399462890625, 1.399446533203125, 1.3993922119140625, 1.3996553955078126, 1.3996207275390624, 1.399635986328125, 1.399510986328125, 1.3993389892578125, 1.3993994140625, 1.39949462890625, 1.3993338623046876, 1.399525390625, 1.39961962890625, 1.3995970458984375, 1.3995335693359374, 1.39974658203125, 1.3998438720703126, 1.4000506591796875, 1.3993870849609376, 1.40001171875, 1.3994864501953126, 1.3999493408203125, 1.39968408203125, 1.39966162109375, 1.399667724609375, 1.3997752685546876, 1.3997281494140625, 1.3997158203125, 1.399551025390625, 1.39983154296875, 1.39955810546875, 1.4000404052734374, 1.39997998046875, 1.3996800537109375, 1.3997813720703125, 1.3998623046875, 1.3999267578125, 1.3998714599609374, 1.399816162109375, 1.3999677734375, 2.9783203125, 1.3990379638671875, 1.398981689453125, 1.3989110107421876, 1.3995018310546874, 1.399352294921875, 1.39950390625, 1.399098388671875, 1.3995130615234375, 1.399405517578125, 1.3993851318359376, 1.3991854248046875, 1.39917724609375, 1.399246826171875, 1.3993748779296875, 1.3995396728515626, 1.399609375, 1.401608154296875, 1.39957763671875, 1.399351318359375, 1.3994659423828124, 1.3992908935546875, 1.39940966796875, 1.399413818359375, 1.3994700927734376, 1.3996348876953124, 1.3998448486328126, 1.3994188232421876, 1.3997701416015624, 1.3996114501953125, 1.3994290771484375, 1.39945166015625, 1.3996492919921875, 1.39970458984375, 1.39961865234375, 1.3996348876953124, 1.39974755859375, 1.399419921875, 1.3996390380859376, 1.399568359375, 1.40006396484375, 1.3995345458984374, 1.399699462890625, 1.39949462890625, 1.3997711181640624, 1.39957763671875, 1.400057861328125, 1.399718994140625, 1.3997783203125, 1.3994967041015625, 1.3998909912109374, 1.399602294921875, 1.3997977294921875, 1.3996103515625, 1.4001929931640624, 1.400183837890625, 1.4000916748046874, 1.3998070068359374, 1.400088623046875, 1.3998140869140625, 1.399877685546875, 1.3998448486328126, 1.4000291748046876, 2.977902587890625, 1.3989754638671874, 1.3993359375, 1.399014404296875, 1.39926123046875, 1.399205810546875, 1.3993922119140625, 1.39930322265625, 1.39965234375, 1.3993604736328125, 1.3991485595703126, 1.3991854248046875, 1.3994864501953126, 1.399546875, 1.399447509765625, 1.3990963134765626, 1.3993790283203125, 1.399372802734375, 1.399635986328125, 1.3994639892578125, 1.399376953125, 1.398993896484375, 1.399372802734375, 1.3994066162109375, 1.3993277587890625, 1.3993963623046874, 1.3998489990234375, 1.3994404296875, 1.3996298828125, 1.3993543701171876, 1.3996400146484376, 1.3993819580078124, 1.399593994140625, 1.39938818359375, 1.399609375, 1.3995673828125, 1.399614501953125, 1.399294921875, 1.39947412109375, 1.399741455078125, 1.399953369140625, 1.3997691650390625, 1.3998253173828126, 1.399690185546875, 1.399609375, 1.3995694580078124, 1.3996502685546874, 1.3995181884765624, 1.3998612060546876, 1.3997015380859374, 1.399868408203125, 1.399719970703125, 1.3997659912109375, 1.399816162109375, 1.400052734375, 1.3999656982421875, 1.399974853515625, 1.39974755859375, 1.4001602783203124, 1.4000721435546875, 1.3998294677734375, 1.399690185546875, 1.3998980712890625]",tokens/s,0.7031464878906475,, +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5008.822272,8173.125632,0.0,7526.678528,6833.744896,s,10,3.2637095642089844,0.3263709564208984,0.0007368089127270137,0.3261606750488281,0.32729655456542966,0.3276723556518555,0.3279729965209961,"[0.32596209716796876, 0.32586428833007813, 0.3263592529296875, 0.32655685424804687, 0.326583251953125, 0.32592489624023435, 0.32721304321289063, 0.32552175903320313, 0.32567596435546875, 0.32804815673828125]",tokens/s,784.3835211545423,kWh,3.849155682816727e-06,2.109159659243408e-06,2.032348400070999e-05,2.6281799342770124e-05,tokens/kWh,9740581.177917834,MB,5008.822272,8173.125632,0.0,7526.678528,7110.584832,s,10,177.045357421875,17.7045357421875,0.0012845826660970417,17.70458984375,17.706225781249998,17.70631015625,17.70637765625,"[17.703931640625, 17.701494140625, 17.704443359375, 17.70496875, 17.70412890625, 17.70639453125, 17.70620703125, 17.704736328125, 17.70418359375, 17.704869140625]",tokens/s,3.558410167733434,kWh,0.00020893763350596564,0.00011451522852683938,0.0010782542004162915,0.0014017070624490963,tokens/kWh,44945.196958574845,,s,629,179.70182278442405,0.285694471835332,0.038836008660610016,0.2809876403808594,0.28119777221679687,0.2813642761230469,0.6074044165039063,"[0.2808002624511719, 0.2810859375, 0.2808913879394531, 0.280943603515625, 0.2809088134765625, 0.2810040283203125, 0.28078695678710935, 0.28094772338867186, 0.2809241638183594, 0.2809722900390625, 0.2809620361328125, 0.2810992736816406, 0.28092007446289063, 0.28101119995117185, 0.2808463439941406, 0.281027587890625, 0.2808821716308594, 0.2812037048339844, 0.2810091552734375, 0.2810931091308594, 0.28097125244140625, 0.2809722900390625, 0.2810142822265625, 0.280943603515625, 0.28114544677734377, 0.28108175659179685, 0.28101937866210935, 0.28090777587890625, 0.2809098205566406, 0.2811453552246094, 0.28107467651367185, 0.281064453125, 0.2811750793457031, 0.28106442260742187, 0.2810439758300781, 0.28124465942382815, 0.2813736877441406, 0.28123956298828123, 0.28094259643554687, 0.28106451416015626, 0.2809650573730469, 0.2809836120605469, 0.2809210205078125, 0.28089547729492187, 0.2809016418457031, 0.280958984375, 0.28140850830078123, 0.28111666870117186, 0.2809743347167969, 0.2808883056640625, 0.28079623413085936, 0.28105413818359376, 0.2809292907714844, 0.2810009460449219, 0.28097329711914065, 0.28109619140625, 0.2809098205566406, 0.28071221923828127, 0.280933349609375, 0.28117401123046876, 0.28113409423828123, 0.28092620849609373, 0.60747265625, 0.28070806884765626, 0.2811351013183594, 0.28077362060546873, 0.28089959716796875, 0.2807675476074219, 0.2810275268554687, 0.28092312622070315, 0.2809046936035156, 0.2808084411621094, 0.2810439758300781, 0.2808105163574219, 0.28097430419921876, 0.2808217468261719, 0.28103271484375, 0.28088729858398437, 0.2810675048828125, 0.28095693969726565, 0.28103167724609374, 0.28106036376953125, 0.28104193115234377, 0.2808299560546875, 0.2809108581542969, 0.28080743408203124, 0.28094873046875, 0.280958984375, 0.2808985595703125, 0.2808197021484375, 0.2809333801269531, 0.28095489501953125, 0.2808002624511719, 0.28090264892578126, 0.2809466857910156, 0.28094158935546876, 0.2808708801269531, 0.28107467651367185, 0.2808638000488281, 0.28104391479492186, 0.2809292907714844, 0.2809661560058594, 0.28094772338867186, 0.28103271484375, 0.28111053466796876, 0.2809088134765625, 0.28127651977539064, 0.2812466125488281, 0.2810214538574219, 0.2809333801269531, 0.2810347595214844, 0.28096408081054686, 0.28108184814453124, 0.2808186950683594, 0.2811372985839844, 0.28084515380859376, 0.28100827026367187, 0.28106539916992185, 0.2810152282714844, 0.2809989013671875, 0.2808842163085937, 0.2811985778808594, 0.2813501586914062, 0.2810224609375, 0.28091802978515623, 0.6079191284179688, 0.2806446228027344, 0.28111154174804687, 0.28073370361328126, 0.2810019836425781, 0.28069903564453125, 0.281046875, 0.2809036865234375, 0.28084225463867185, 0.2807582702636719, 0.28101119995117185, 0.28089959716796875, 0.28108697509765623, 0.28107672119140625, 0.280932373046875, 0.28086679077148435, 0.2810224609375, 0.28107571411132815, 0.2811197509765625, 0.28087091064453124, 0.2810859375, 0.2809395141601562, 0.2808412170410156, 0.28205978393554687, 0.2808739929199219, 0.28123648071289065, 0.28117913818359375, 0.28086273193359373, 0.28080435180664065, 0.2808360900878906, 0.28089242553710936, 0.28096820068359374, 0.2810460205078125, 0.28101544189453126, 0.28080010986328124, 0.2809333801269531, 0.280880126953125, 0.28094464111328127, 0.28085147094726565, 0.2810408935546875, 0.28090060424804686, 0.28079513549804686, 0.28095794677734376, 0.28093438720703123, 0.28107467651367185, 0.28103067016601563, 0.28095693969726565, 0.28259121704101564, 0.2810726318359375, 0.28107366943359374, 0.2812630920410156, 0.2811709289550781, 0.2813378601074219, 0.2809466857910156, 0.28098046875, 0.28088116455078127, 0.2810010681152344, 0.2809774169921875, 0.28094351196289064, 0.28106137084960936, 0.28105523681640626, 0.28106649780273435, 0.28105010986328127, 0.6080983276367188, 0.2805442810058594, 0.2809896545410156, 0.2807879638671875, 0.2809374694824219, 0.28062310791015627, 0.28097537231445313, 0.28071218872070314, 0.280995849609375, 0.2809415588378906, 0.2810777587890625, 0.2808299560546875, 0.28200244140625, 0.28103067016601563, 0.28094259643554687, 0.28076031494140624, 0.28103372192382814, 0.28095693969726565, 0.28099789428710936, 0.28095590209960936, 0.28118731689453125, 0.28101324462890626, 0.2808268737792969, 0.28090060424804686, 0.2810726318359375, 0.28087408447265627, 0.2808984680175781, 0.2808555603027344, 0.2809251708984375, 0.28087100219726563, 0.28081143188476565, 0.28094772338867186, 0.281238525390625, 0.28138494873046876, 0.2810378112792969, 0.28088116455078127, 0.2808719482421875, 0.28103988647460937, 0.28095794677734376, 0.28113714599609374, 0.28095693969726565, 0.2808350830078125, 0.2807224426269531, 0.2808401794433594, 0.283072509765625, 0.2809395141601562, 0.28103167724609374, 0.2809651489257812, 0.2811422424316406, 0.2808913879394531, 0.280922119140625, 0.2810408935546875, 0.28119448852539064, 0.2809671630859375, 0.28113409423828123, 0.28099072265625, 0.28125595092773437, 0.280995849609375, 0.2812262268066406, 0.2814044189453125, 0.28109722900390627, 0.28100811767578127, 0.2810019836425781, 0.6083676147460938, 0.281069580078125, 0.2809661865234375, 0.2808504028320313, 0.2810521545410156, 0.28073675537109377, 0.28086477661132814, 0.2808186950683594, 0.2809518127441406, 0.2808637390136719, 0.28104193115234377, 0.28080743408203124, 0.28095489501953125, 0.2810368041992187, 0.28096408081054686, 0.28084844970703127, 0.2810756530761719, 0.28226763916015624, 0.2809692077636719, 0.28088626098632813, 0.28098355102539063, 0.2808197021484375, 0.2807982177734375, 0.28103277587890624, 0.28112786865234374, 0.280961181640625, 0.2809997863769531, 0.28112399291992185, 0.2811482543945312, 0.2809599914550781, 0.28090060424804686, 0.2811207580566406, 0.28099172973632813, 0.2809938049316406, 0.2809968566894531, 0.28118731689453125, 0.280995849609375, 0.2809426574707031, 0.28101626586914064, 0.2809979553222656, 0.28103265380859377, 0.280943603515625, 0.28089447021484376, 0.28097329711914065, 0.2810726318359375, 0.28112384033203125, 0.28100811767578127, 0.2809580078125, 0.28122515869140624, 0.28084429931640625, 0.2810142822265625, 0.2808360900878906, 0.2810214538574219, 0.2808760375976562, 0.2810133666992187, 0.2810059509277344, 0.28116583251953126, 0.28140237426757814, 0.2807633972167969, 0.28105523681640626, 0.28104702758789063, 0.28084735107421877, 0.2810245056152344, 0.6076272583007812, 0.28068353271484375, 0.28097637939453124, 0.2806927490234375, 0.28078695678710935, 0.28059033203125, 0.28082278442382813, 0.2807767028808594, 0.2808760375976562, 0.28086578369140625, 0.2808350830078125, 0.28088626098632813, 0.2813931579589844, 0.2808126525878906, 0.28106640625, 0.28102655029296875, 0.2811617431640625, 0.28116583251953126, 0.2809968566894531, 0.28094259643554687, 0.28096307373046875, 0.28095489501953125, 0.28097637939453124, 0.28106137084960936, 0.2811781005859375, 0.28100607299804686, 0.28082177734375, 0.28104290771484375, 0.28098150634765623, 0.281091064453125, 0.2810624084472656, 0.28119244384765624, 0.28088729858398437, 0.2811064453125, 0.2811351013183594, 0.2812886962890625, 0.28093438720703123, 0.28102554321289064, 0.28370022583007815, 0.2810654602050781, 0.2810654602050781, 0.280869873046875, 0.2810214538574219, 0.28114227294921873, 0.28109515380859373, 0.28090060424804686, 0.28101324462890626, 0.2810245056152344, 0.28122726440429685, 0.2809302978515625, 0.28107366943359374, 0.28088934326171877, 0.28120883178710937, 0.28091390991210935, 0.2811975708007812, 0.2811269226074219, 0.2810368041992187, 0.2808280639648438, 0.2809659729003906, 0.2812037048339844, 0.28127130126953126, 0.28098968505859373, 0.28096511840820315, 0.60883251953125, 0.28060467529296873, 0.2809794616699219, 0.28092312622070315, 0.2813982849121094, 0.2813009948730469, 0.2809886779785156, 0.2808883056640625, 0.28107366943359374, 0.280922119140625, 0.2809456787109375, 0.2808084411621094, 0.2810163269042969, 0.28097537231445313, 0.2810071105957031, 0.2808258666992188, 0.2810491027832031, 0.28098968505859373, 0.2814863891601562, 0.2812784118652344, 0.281064453125, 0.28083511352539064, 0.28078179931640623, 0.28249298095703124, 0.2810910034179688, 0.2808832092285156, 0.28091494750976564, 0.2809292907714844, 0.28089141845703125, 0.28101425170898436, 0.2809743347167969, 0.2809722900390625, 0.2810408935546875, 0.28118426513671874, 0.2812303466796875, 0.2811094970703125, 0.28078387451171877, 0.28088626098632813, 0.28119451904296877, 0.28095895385742187, 0.28112588500976565, 0.28113006591796874, 0.2810581970214844, 0.280953857421875, 0.28095077514648437, 0.28101223754882815, 0.281027587890625, 0.28109515380859373, 0.2811064453125, 0.28092825317382814, 0.28113818359375, 0.28082891845703123, 0.28119143676757813, 0.28105426025390623, 0.2811329956054687, 0.28106854248046875, 0.2810572814941406, 0.28092724609375, 0.2808821716308594, 0.2811207580566406, 0.2811412353515625, 0.2810296325683594, 0.28099172973632813, 0.60706103515625, 0.2807059936523438, 0.2810726318359375, 0.2809098205566406, 0.28105624389648437, 0.2806220703125, 0.2809917907714844, 0.28088113403320314, 0.28103372192382814, 0.28095281982421877, 0.28107879638671873, 0.2810849304199219, 0.2812733459472656, 0.2809743347167969, 0.2812303466796875, 0.28095281982421877, 0.2824908752441406, 0.2809866333007813, 0.2809456787109375, 0.280880126953125, 0.2809169921875, 0.2809302978515625, 0.2807879638671875, 0.2809661560058594, 0.2810798034667969, 0.28095489501953125, 0.28097537231445313, 0.2809948120117188, 0.2809794616699219, 0.28092724609375, 0.28095794677734376, 0.2811883544921875, 0.28089959716796875, 0.2809692077636719, 0.28097552490234373, 0.2809906005859375, 0.28083914184570313, 0.2809456787109375, 0.2810152893066406, 0.28097125244140625, 0.2810501708984375, 0.2807622985839844, 0.280890380859375, 0.28096511840820315, 0.2809323425292969, 0.28090264892578126, 0.28089752197265627, 0.28092108154296874, 0.2811678771972656, 0.2809251708984375, 0.2810941467285156, 0.2808832092285156, 0.2812651672363281, 0.2808053894042969, 0.281017333984375, 0.28113409423828123, 0.28139724731445315, 0.28102041625976565, 0.2808330383300781, 0.2810378112792969, 0.2812538757324219, 0.28141259765625, 0.2812620849609375, 0.6072289428710937, 0.28075930786132813, 0.28142489624023437, 0.28085861206054685, 0.280919189453125, 0.2806700439453125, 0.28098770141601564, 0.2808155517578125, 0.2809251708984375, 0.28085861206054685, 0.2810214538574219, 0.28077163696289065, 0.280939453125, 0.2808053894042969, 0.2809876403808594, 0.2808883056640625, 0.28102349853515624, 0.2819471435546875, 0.28139111328125, 0.2810654602050781, 0.28137985229492185, 0.28105010986328127, 0.28085861206054685, 0.2810859375, 0.28121600341796876, 0.2809887390136719, 0.28093536376953127, 0.280958984375, 0.28101223754882815, 0.2809333801269531, 0.2809938049316406, 0.28112384033203125, 0.28098458862304687, 0.281059326171875, 0.2811770935058594, 0.28115045166015623, 0.28095794677734376, 0.2810523071289063, 0.28106121826171876, 0.2809232177734375, 0.28085751342773435, 0.28083096313476563, 0.2809395141601562, 0.2809108581542969, 0.2809722900390625, 0.2809190673828125, 0.28105520629882813, 0.28091390991210935, 0.2810091552734375, 0.2810040283203125, 0.28116790771484373, 0.28094573974609377, 0.2811063232421875, 0.280995849609375, 0.2812528686523437, 0.2810511474609375, 0.2810040283203125, 0.2809139709472656, 0.2808759765625, 0.2811085205078125, 0.2810316467285156, 0.28102349853515624, 0.2808760375976562, 0.6096005249023437, 0.28075726318359373, 0.2809169921875, 0.28073370361328126, 0.2812037048339844, 0.2810357666015625, 0.2810040283203125, 0.2807848815917969, 0.281017333984375, 0.2811248779296875, 0.28101223754882815, 0.2807848815917969, 0.28090673828125, 0.28089547729492187, 0.2810491027832031, 0.28088729858398437, 0.2809088134765625, 0.28079617309570315, 0.28105523681640626, 0.28090060424804686, 0.28096820068359374, 0.28085861206054685, 0.28080230712890625, 0.280880126953125, 0.28098355102539063, 0.2810572814941406, 0.2831247253417969, 0.281059326171875, 0.2810521545410156, 0.28110848999023436, 0.2811576232910156, 0.2811760559082031, 0.2811463623046875, 0.280911865234375, 0.2809190368652344, 0.2809876403808594, 0.28097024536132814, 0.2809886779785156, 0.2810624084472656, 0.28105523681640626, 0.28116378784179685, 0.28104193115234377, 0.28093646240234377, 0.2808760375976562, 0.28102859497070315, 0.28101837158203125, 0.2808401794433594, 0.28090573120117185, 0.28106854248046875, 0.28096820068359374, 0.2809661560058594, 0.2807940979003906, 0.28126412963867187, 0.28086273193359373, 0.28110540771484377, 0.2811617431640625, 0.281143310546875, 0.28087295532226564, 0.28085043334960935, 0.2810019836425781, 0.2811822204589844, 0.28111871337890626, 0.28107366943359374]",tokens/s,3.5002427368506384,, +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 117, in run + _ = backend.generate(self.inputs, self.config.generate_kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 369, in generate + return self.pretrained_model.generate(**inputs, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1736, in generate + result = self._sample( + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2375, in _sample + outputs = self( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 900, in forward + transformer_outputs = self.transformer( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 797, in forward + outputs = block( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 453, in forward + attn_outputs = self.self_attention( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-7b/898df1396f35e447d5fe44e0a3ccaaaa69f30d36/modeling_falcon.py"", line 291, in forward + fused_qkv = self.query_key_value(hidden_states) # [batch_size, seq_length, 3 x hidden_size] + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/auto_gptq/nn_modules/qlinear/qlinear_cuda_old.py"", line 316, in forward + weight = weight.reshape(-1, self.group_size, weight.shape[2]) +RuntimeError: shape '[-1, 128, 4672]' is invalid for input of size 21229568 + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4708.687872,21498.42944,0.0,20851.982336,20236.731392,s,10,15.782414306640625,1.5782414306640624,0.0005368111432327376,1.5781434326171875,1.5784585083007812,1.5790858825683594,1.579587781982422,"[1.578103759765625, 1.5775478515625, 1.5778953857421876, 1.5797132568359376, 1.57818310546875, 1.5782744140625, 1.57828564453125, 1.578103759765625, 1.578319091796875, 1.577988037109375]",tokens/s,162.20585458353173,kWh,1.8640190545055602e-05,1.0214868406037568e-05,9.259768518919964e-05,0.0001214527441402928,tokens/kWh,2107815.6925321394,MB,4708.687872,21498.42944,0.0,20851.982336,20339.707392,s,10,881.7457578125,88.17457578125,0.002115503395001207,88.17539453124999,88.17674609375,88.17697460937501,88.17715742187501,"[88.17125, 88.17175, 88.1749609375, 88.175828125, 88.177203125, 88.175890625, 88.1766953125, 88.173328125, 88.176578125, 88.1722734375]",tokens/s,0.714491671117251,kWh,0.001040854390843047,0.0005704774625729078,0.005178987532075602,0.006790319385491556,tokens/kWh,9277.914104395162,,s,629,894.5504398193359,1.4221787596491826,0.18751446646952302,1.3995601806640625,1.4000109375,1.4001393798828126,2.978033017578125,"[1.399194580078125, 1.399298095703125, 1.3991290283203126, 1.3992857666015626, 1.3990655517578126, 1.399204833984375, 1.3991751708984375, 1.3994710693359376, 1.3990738525390625, 1.3990277099609374, 1.39915771484375, 1.399203857421875, 1.3990020751953125, 1.39932470703125, 1.39911376953125, 1.3995806884765625, 1.399319580078125, 1.3996319580078125, 1.399352294921875, 1.3991802978515624, 1.399151611328125, 1.39943017578125, 1.3990972900390626, 1.3995540771484376, 1.3996278076171875, 1.399510009765625, 1.39930322265625, 1.3995018310546874, 1.399568359375, 1.39934619140625, 1.399161865234375, 1.3999493408203125, 1.3994700927734376, 1.399572509765625, 1.39960107421875, 1.399762939453125, 1.3996871337890624, 1.399688232421875, 1.3997169189453125, 1.4000067138671874, 1.399546875, 1.399795654296875, 1.399826416015625, 1.3999605712890626, 1.3998192138671874, 1.399867431640625, 1.399826416015625, 1.399509033203125, 1.399413818359375, 1.3999349365234375, 1.399510986328125, 1.3996646728515625, 1.399531494140625, 1.4000762939453124, 1.3997393798828126, 1.40002099609375, 1.39964208984375, 1.39989306640625, 1.399784423828125, 1.4000015869140625, 1.4001776123046874, 1.4001162109375, 2.977680419921875, 1.3987635498046875, 1.3990216064453125, 1.398865966796875, 1.399183349609375, 1.3990440673828124, 1.399251953125, 1.399235595703125, 1.399419921875, 1.3989498291015625, 1.3994588623046875, 1.3995325927734374, 1.39915576171875, 1.399416748046875, 1.3992857666015626, 1.399309326171875, 1.3996400146484376, 1.39901025390625, 1.3994168701171874, 1.39928369140625, 1.3992008056640626, 1.399120849609375, 1.39950390625, 1.3994332275390624, 1.3995540771484376, 1.399593017578125, 1.3996851806640624, 1.3992059326171875, 1.3998223876953124, 1.3995396728515626, 1.399499755859375, 1.3992294921875, 1.399562255859375, 1.399331787109375, 1.3999698486328125, 1.39949365234375, 1.3995172119140624, 1.3994915771484375, 1.399731201171875, 1.3996185302734374, 1.399804931640625, 1.3992376708984375, 1.3995550537109376, 1.399384033203125, 1.39986328125, 1.3994813232421874, 1.401822265625, 1.399848876953125, 1.400058837890625, 1.399477294921875, 1.399910400390625, 1.3995653076171874, 1.399931884765625, 1.3996185302734374, 1.399846923828125, 1.399761962890625, 1.3999759521484374, 1.3998602294921876, 1.3998929443359376, 1.39964208984375, 1.4000107421875, 1.399794677734375, 1.4000025634765625, 2.978083740234375, 1.398897705078125, 1.3991322021484376, 1.3995980224609375, 1.3994290771484375, 1.399194580078125, 1.399624755859375, 1.399552001953125, 1.3993800048828124, 1.3990799560546876, 1.399405517578125, 1.399300048828125, 1.3992899169921875, 1.3992335205078126, 1.39951513671875, 1.3993482666015624, 1.399700439453125, 1.399189453125, 1.3992960205078124, 1.3992008056640626, 1.3995550537109376, 1.399152587890625, 1.3994906005859375, 1.3997373046875, 1.399525390625, 1.3993953857421875, 1.3996871337890624, 1.3994229736328125, 1.39953662109375, 1.39951513671875, 1.3995233154296876, 1.3995028076171876, 1.399846923828125, 1.399299072265625, 1.3997659912109375, 1.3994556884765625, 1.3997015380859374, 1.39947412109375, 1.3996082763671875, 1.39980078125, 1.3997916259765626, 1.3995919189453125, 1.39965234375, 1.3994854736328124, 1.39974755859375, 1.400369140625, 1.3997813720703125, 1.3998275146484376, 1.3997486572265625, 1.3997486572265625, 1.39992578125, 1.3999349365234375, 1.3997034912109374, 1.3998878173828124, 1.39997900390625, 1.400004638671875, 1.4001182861328125, 1.399736328125, 1.400037353515625, 1.3998233642578124, 1.399794677734375, 1.399687255859375, 1.400122314453125, 2.979734619140625, 1.3991536865234375, 1.399083984375, 1.3987276611328125, 1.39987255859375, 1.3991597900390624, 1.3992386474609375, 1.399719970703125, 1.3995755615234375, 1.3991290283203126, 1.3991075439453124, 1.39930419921875, 1.3992325439453126, 1.3994864501953126, 1.39919775390625, 1.39900830078125, 1.399330810546875, 1.3992427978515625, 1.39966357421875, 1.39919873046875, 1.39934619140625, 1.39936767578125, 1.3994219970703126, 1.399363525390625, 1.3997291259765625, 1.3998028564453124, 1.399531494140625, 1.3992847900390626, 1.399729248046875, 1.3994423828125, 1.399742431640625, 1.3993021240234376, 1.39958984375, 1.3994332275390624, 1.399636962890625, 1.3993062744140625, 1.399445556640625, 1.3994761962890625, 1.399825439453125, 1.3996768798828125, 1.399867431640625, 1.3995806884765625, 1.3999698486328125, 1.3994281005859375, 1.4019144287109375, 1.39976806640625, 1.400162353515625, 1.3998765869140626, 1.3997568359375, 1.3996256103515625, 1.399816162109375, 1.3998919677734376, 1.399845947265625, 1.3995601806640625, 1.3998387451171874, 1.4001397705078125, 1.400004638671875, 1.4000650634765626, 1.40010693359375, 1.39981103515625, 1.3998050537109374, 1.3996778564453125, 1.4001243896484374, 2.97879443359375, 1.399288818359375, 1.3990625, 1.3989530029296875, 1.3993800048828124, 1.3989437255859376, 1.399468017578125, 1.39931640625, 1.3996451416015625, 1.3991168212890626, 1.39941064453125, 1.399319580078125, 1.39913525390625, 1.3994024658203126, 1.399150634765625, 1.3994117431640625, 1.399594970703125, 1.3994075927734375, 1.399793701171875, 1.399268310546875, 1.3994659423828124, 1.3996287841796875, 1.3994556884765625, 1.3993697509765626, 1.3996676025390624, 1.399447509765625, 1.3997608642578125, 1.3995079345703125, 1.399666748046875, 1.3995478515625, 1.3996953125, 1.399482421875, 1.3994608154296875, 1.3996329345703125, 1.3997855224609375, 1.39945166015625, 1.3995386962890626, 1.39945166015625, 1.3997578125, 1.3999656982421875, 1.4002125244140624, 1.3995765380859375, 1.4000814208984376, 1.3997659912109375, 1.39985302734375, 1.399521240234375, 1.399901123046875, 1.39972705078125, 1.400004638671875, 1.399415771484375, 1.3997598876953126, 1.399835693359375, 1.3999913330078124, 1.399677001953125, 1.4000362548828125, 1.4001612548828124, 1.4003128662109374, 1.39993603515625, 1.400158203125, 1.4000721435546875, 1.4003721923828125, 1.399899169921875, 1.3997158203125, 2.978334716796875, 1.39909326171875, 1.3990604248046874, 1.39932568359375, 1.399351318359375, 1.39915771484375, 1.3992540283203125, 1.3992642822265624, 1.399413818359375, 1.3991741943359375, 1.3993359375, 1.399224365234375, 1.3994229736328125, 1.39940966796875, 1.399456787109375, 1.39919873046875, 1.3996123046875, 1.3993431396484375, 1.3997291259765625, 1.3994117431640625, 1.3995386962890626, 1.3994404296875, 1.3996083984375, 1.399300048828125, 1.3998018798828125, 1.4001162109375, 1.399901123046875, 1.3998970947265625, 1.399531494140625, 1.399182373046875, 1.3994619140625, 1.399109619140625, 1.3996697998046874, 1.3993175048828126, 1.399574462890625, 1.3992847900390626, 1.39966162109375, 1.399520263671875, 1.401324462890625, 1.399489501953125, 1.399690185546875, 1.3993994140625, 1.399900146484375, 1.39951611328125, 1.399921630859375, 1.399678955078125, 1.4001673583984375, 1.39990625, 1.3998018798828125, 1.3995079345703125, 1.400015869140625, 1.3994803466796875, 1.399951416015625, 1.3997076416015626, 1.40027294921875, 1.3999246826171876, 1.3999246826171876, 1.399732177734375, 1.3998970947265625, 1.399783447265625, 1.4001387939453125, 1.3993502197265626, 1.4002431640625, 2.9787392578125, 1.398930419921875, 1.3988720703125, 1.3989642333984376, 1.3989478759765626, 1.3991597900390624, 1.3990850830078125, 1.400115234375, 1.3992796630859374, 1.3990533447265625, 1.39926220703125, 1.3995345458984374, 1.39920703125, 1.3993245849609375, 1.399505859375, 1.39932666015625, 1.3994383544921876, 1.399141357421875, 1.3998828125, 1.39945166015625, 1.3994423828125, 1.39908203125, 1.399351318359375, 1.401248779296875, 1.3996380615234374, 1.399615478515625, 1.3997967529296875, 1.3995499267578124, 1.399447509765625, 1.3995377197265626, 1.39965234375, 1.3995089111328125, 1.3995694580078124, 1.39930419921875, 1.3994619140625, 1.399435302734375, 1.399462890625, 1.3994710693359376, 1.39966259765625, 1.3994649658203124, 1.3999195556640625, 1.399552001953125, 1.399803955078125, 1.3996103515625, 1.399816162109375, 1.39976708984375, 1.400056884765625, 1.399909423828125, 1.39981103515625, 1.3995601806640625, 1.3999759521484374, 1.399825439453125, 1.4004111328125, 1.399698486328125, 1.3998018798828125, 1.4003282470703124, 1.4000977783203126, 1.4000067138671874, 1.399920654296875, 1.3997823486328125, 1.400100830078125, 1.3995079345703125, 1.4000343017578125, 2.979209228515625, 1.3990235595703124, 1.398892578125, 1.3989365234375, 1.39919970703125, 1.3989918212890624, 1.3991229248046875, 1.3992652587890626, 1.3993604736328125, 1.3993175048828126, 1.3992017822265626, 1.39951416015625, 1.3994075927734375, 1.399320556640625, 1.39930419921875, 1.39944140625, 1.4017669677734375, 1.3992376708984375, 1.399351318359375, 1.3993123779296874, 1.3993021240234376, 1.39919873046875, 1.399609375, 1.399462890625, 1.399446533203125, 1.3993922119140625, 1.3996553955078126, 1.3996207275390624, 1.399635986328125, 1.399510986328125, 1.3993389892578125, 1.3993994140625, 1.39949462890625, 1.3993338623046876, 1.399525390625, 1.39961962890625, 1.3995970458984375, 1.3995335693359374, 1.39974658203125, 1.3998438720703126, 1.4000506591796875, 1.3993870849609376, 1.40001171875, 1.3994864501953126, 1.3999493408203125, 1.39968408203125, 1.39966162109375, 1.399667724609375, 1.3997752685546876, 1.3997281494140625, 1.3997158203125, 1.399551025390625, 1.39983154296875, 1.39955810546875, 1.4000404052734374, 1.39997998046875, 1.3996800537109375, 1.3997813720703125, 1.3998623046875, 1.3999267578125, 1.3998714599609374, 1.399816162109375, 1.3999677734375, 2.9783203125, 1.3990379638671875, 1.398981689453125, 1.3989110107421876, 1.3995018310546874, 1.399352294921875, 1.39950390625, 1.399098388671875, 1.3995130615234375, 1.399405517578125, 1.3993851318359376, 1.3991854248046875, 1.39917724609375, 1.399246826171875, 1.3993748779296875, 1.3995396728515626, 1.399609375, 1.401608154296875, 1.39957763671875, 1.399351318359375, 1.3994659423828124, 1.3992908935546875, 1.39940966796875, 1.399413818359375, 1.3994700927734376, 1.3996348876953124, 1.3998448486328126, 1.3994188232421876, 1.3997701416015624, 1.3996114501953125, 1.3994290771484375, 1.39945166015625, 1.3996492919921875, 1.39970458984375, 1.39961865234375, 1.3996348876953124, 1.39974755859375, 1.399419921875, 1.3996390380859376, 1.399568359375, 1.40006396484375, 1.3995345458984374, 1.399699462890625, 1.39949462890625, 1.3997711181640624, 1.39957763671875, 1.400057861328125, 1.399718994140625, 1.3997783203125, 1.3994967041015625, 1.3998909912109374, 1.399602294921875, 1.3997977294921875, 1.3996103515625, 1.4001929931640624, 1.400183837890625, 1.4000916748046874, 1.3998070068359374, 1.400088623046875, 1.3998140869140625, 1.399877685546875, 1.3998448486328126, 1.4000291748046876, 2.977902587890625, 1.3989754638671874, 1.3993359375, 1.399014404296875, 1.39926123046875, 1.399205810546875, 1.3993922119140625, 1.39930322265625, 1.39965234375, 1.3993604736328125, 1.3991485595703126, 1.3991854248046875, 1.3994864501953126, 1.399546875, 1.399447509765625, 1.3990963134765626, 1.3993790283203125, 1.399372802734375, 1.399635986328125, 1.3994639892578125, 1.399376953125, 1.398993896484375, 1.399372802734375, 1.3994066162109375, 1.3993277587890625, 1.3993963623046874, 1.3998489990234375, 1.3994404296875, 1.3996298828125, 1.3993543701171876, 1.3996400146484376, 1.3993819580078124, 1.399593994140625, 1.39938818359375, 1.399609375, 1.3995673828125, 1.399614501953125, 1.399294921875, 1.39947412109375, 1.399741455078125, 1.399953369140625, 1.3997691650390625, 1.3998253173828126, 1.399690185546875, 1.399609375, 1.3995694580078124, 1.3996502685546874, 1.3995181884765624, 1.3998612060546876, 1.3997015380859374, 1.399868408203125, 1.399719970703125, 1.3997659912109375, 1.399816162109375, 1.400052734375, 1.3999656982421875, 1.399974853515625, 1.39974755859375, 1.4001602783203124, 1.4000721435546875, 1.3998294677734375, 1.399690185546875, 1.3998980712890625]",tokens/s,0.7031464878906475,, +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status + response.raise_for_status() + File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status + raise HTTPError(http_error_msg, response=self) +requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error + metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata + r = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper + response = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper + hf_raise_for_status(response) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status + raise HfHubHTTPError(message, response=response) from e +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4aa7-34acad0a1b0bd2f218242c91;bcc119f6-4e32-4f94-9c96-76fa966e660c) + +403 Forbidden: Authorization error.. +Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. +If you are trying to create or update content,make sure you have a token with the `write` role. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 399, in cached_file + resolved_file = hf_hub_download( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download + return _hf_hub_download_to_cache_dir( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir + _raise_on_head_call_error(head_call_error, force_download, local_files_only) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error + raise LocalEntryNotFoundError( +huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ + super().__init__(config) + File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ + self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) + File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config + return AutoConfig.from_pretrained(model, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict + resolved_config_file = cached_file( + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 442, in cached_file + raise EnvironmentError( +OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. +Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -10903,7 +11778,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12e9-1a3bc2b25c55bfc440769276;c06a8fb9-0e25-44db-b7ca-f716af2cb42b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4510-3f75da72213387d27a6bff1b;83b18d0a-9bc4-4107-8113-a39628cf26cd) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10965,7 +11840,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee57-6ad03dca41083e801160bcdb;c4f597f6-bc61-40a4-8c04-05a5f8a41e06) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3f67-7e645e407752c7161697deff;55187485-9dbf-4887-b6ba-20146ec99c48) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -11092,7 +11967,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1277-56a968f57540f7521394be49;92977195-894a-4384-9237-54e18dca45ca) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44c0-1da2803d3d41108f528a0594;adc3a332-7fc5-4ca9-9250-45b0fff23b42) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11165,7 +12040,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664beedd-32866ea645249b4543ce3790;f12613af-55d7-4e28-892c-795fbf9ec3ea) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d3fdc-3f3e01be778eb000189e20ee;666de4b5-3116-4734-8e33-553ddcad9c39) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11238,7 +12113,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1408-57ea1ad376423646399391c1;9435d010-633e-419c-bf5e-48ba92fb297d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d461f-7df667e649c655c6333785ca;5092a282-e965-48f1-baeb-f2863e5da726) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11268,6 +12143,63 @@ Traceback (most recent call last): OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 117, in run + _ = backend.generate(self.inputs, self.config.generate_kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 369, in generate + return self.pretrained_model.generate(**inputs, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1736, in generate + result = self._sample( + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2375, in _sample + outputs = self( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 900, in forward + transformer_outputs = self.transformer( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 797, in forward + outputs = block( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 477, in forward + mlp_output = self.mlp(mlp_layernorm_out) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/tiiuae/falcon-40b/4a70170c215b36a3cce4b4253f6d0612bb7d4146/modeling_falcon.py"", line 409, in forward + x = self.act(self.dense_h_to_4h(x)) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/auto_gptq/nn_modules/qlinear/qlinear_cuda_old.py"", line 314, in forward + ).to(torch.int16 if self.bits == 8 else torch.int8) +torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 256.00 MiB. GPU + ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,6168.621056,7055.343616,0.0,6408.896512,6098.739712,s,10,4.280724426269532,0.4280724426269532,0.0004871990956745782,0.42789128112792973,0.4287641021728516,0.428794465637207,0.4288187564086914,"[0.4278382568359375, 0.4275146789550781, 0.4288248291015625, 0.4283955078125, 0.42875735473632814, 0.4279443054199219, 0.42762127685546875, 0.4285697326660156, 0.42756597900390625, 0.4276925048828125]",tokens/s,598.0296195405716,kWh,5.0498818771706694e-06,2.7671110361779933e-06,2.4970529235667182e-05,3.278752214901584e-05,tokens/kWh,7807848.328291077,MB,6168.621056,7055.343616,0.0,6408.896512,6157.953024,s,10,235.18207421875002,23.518207421875,0.0011249789087652983,23.517884765625,23.5197904296875,23.519849316406248,23.519896425781248,"[23.51977734375, 23.517072265625, 23.517470703125, 23.51779296875, 23.5169453125, 23.5179765625, 23.51966015625, 23.51699609375, 23.519908203125, 23.518474609375]",tokens/s,2.6787755916040514,kWh,0.00027765461982124387,0.00015217842519419265,0.0013375406718835342,0.0017673737168989706,tokens/kWh,35646.11117479988,,s,629,238.66193222045916,0.37943073484969636,0.05086073881524738,0.373291015625,0.3734941833496094,0.3736278991699219,0.8010455517578124,"[0.373222412109375, 0.37314764404296874, 0.37317633056640626, 0.373254150390625, 0.3736207275390625, 0.3733463134765625, 0.3731026000976562, 0.37307186889648436, 0.3732715454101562, 0.37323162841796875, 0.3731568603515625, 0.3733329772949219, 0.373170166015625, 0.3731435546875, 0.37321932983398437, 0.3731650695800781, 0.37336575317382814, 0.37309951782226564, 0.37331045532226564, 0.3736504211425781, 0.3732428894042969, 0.37334527587890626, 0.373185546875, 0.3730626525878906, 0.37320294189453124, 0.3733186645507813, 0.3732336730957031, 0.3731455993652344, 0.37332171630859373, 0.373148681640625, 0.3734599609375, 0.3735726013183594, 0.37363507080078123, 0.37325515747070315, 0.3733125, 0.37346612548828123, 0.3734200439453125, 0.37349169921875, 0.37328692626953125, 0.3732725830078125, 0.3733391418457031, 0.3733258361816406, 0.3733166198730469, 0.37393716430664065, 0.3732735900878906, 0.3732674560546875, 0.3736258544921875, 0.37329510498046875, 0.3732920227050781, 0.3734814758300781, 0.373496826171875, 0.3734415283203125, 0.37342105102539064, 0.3733739624023438, 0.3731875915527344, 0.3734466552734375, 0.3733831787109375, 0.37326641845703123, 0.37344561767578127, 0.3734722595214844, 0.3734394836425781, 0.3734302673339844, 0.801101806640625, 0.37349581909179685, 0.3731026000976562, 0.373170166015625, 0.3731128234863281, 0.3731015625, 0.3729213562011719, 0.3732490234375, 0.373106689453125, 0.37314047241210935, 0.3731589050292969, 0.3733483581542969, 0.3731568603515625, 0.37321829223632813, 0.3730565185546875, 0.37321829223632813, 0.37324798583984375, 0.3733483581542969, 0.3731199951171875, 0.37336474609375, 0.3733145751953125, 0.3732490234375, 0.3734937744140625, 0.3732203369140625, 0.3731046447753906, 0.37321624755859373, 0.37317428588867185, 0.3732991943359375, 0.37330841064453124, 0.37321624755859373, 0.3730206604003906, 0.3731875915527344, 0.373423095703125, 0.3732705383300781, 0.3733360595703125, 0.3732991943359375, 0.37313433837890625, 0.37337701416015623, 0.37330535888671873, 0.3732777099609375, 0.3732367248535156, 0.37339956665039065, 0.3733729248046875, 0.3734087829589844, 0.3733186645507813, 0.373243896484375, 0.37347634887695313, 0.37349169921875, 0.37362380981445314, 0.37352139282226565, 0.3734313049316406, 0.3732735900878906, 0.3733166198730469, 0.3734732666015625, 0.3732695007324219, 0.3733340148925781, 0.3733780517578125, 0.37346917724609374, 0.3733145751953125, 0.37323876953125, 0.37345074462890626, 0.37327975463867186, 0.37330535888671873, 0.8023336791992187, 0.37309133911132814, 0.3731537780761719, 0.3732090759277344, 0.373212158203125, 0.3730411376953125, 0.372990966796875, 0.3732254638671875, 0.3729438781738281, 0.3730821228027344, 0.3731005554199219, 0.37333197021484377, 0.3732285461425781, 0.37328076171875, 0.37311386108398437, 0.37319064331054685, 0.37324594116210935, 0.37323876953125, 0.3731261291503906, 0.37317938232421877, 0.37323980712890625, 0.3731374206542969, 0.3732008972167969, 0.3735142517089844, 0.373243896484375, 0.37305035400390624, 0.37411737060546874, 0.37328076171875, 0.3731927185058594, 0.37317120361328127, 0.373254150390625, 0.37319985961914065, 0.3736309814453125, 0.3731927185058594, 0.3732060241699219, 0.3733749694824219, 0.3733780517578125, 0.37336474609375, 0.373327880859375, 0.37334323120117185, 0.37330740356445313, 0.37332275390625, 0.37334426879882815, 0.373401611328125, 0.37332376098632813, 0.3732561950683594, 0.3734097900390625, 0.37332376098632813, 0.3733309326171875, 0.3733739624023438, 0.37343539428710937, 0.3732613220214844, 0.37328997802734376, 0.37364324951171873, 0.3733729248046875, 0.37339340209960936, 0.37357568359375, 0.3732725830078125, 0.37343435668945313, 0.37338009643554687, 0.3734364013671875, 0.37323980712890625, 0.37351220703125, 0.802039794921875, 0.3732705383300781, 0.37316299438476563, 0.3731927185058594, 0.3730821228027344, 0.37314764404296874, 0.3729827880859375, 0.37317938232421877, 0.3731251220703125, 0.37312820434570315, 0.37312921142578126, 0.3732674560546875, 0.37320703125, 0.3732695007324219, 0.3732234191894531, 0.3731322937011719, 0.37319781494140625, 0.3730882568359375, 0.37317529296875, 0.3731640319824219, 0.373212158203125, 0.3732060241699219, 0.37308621215820315, 0.37323468017578126, 0.373222412109375, 0.37314047241210935, 0.37319168090820315, 0.37375692749023437, 0.37372314453125, 0.3731814270019531, 0.37317633056640626, 0.37309234619140624, 0.3733421936035156, 0.37338623046875, 0.3732152404785156, 0.3733944396972656, 0.37317633056640626, 0.3732981872558594, 0.3733196716308594, 0.3736330261230469, 0.3733114929199219, 0.3732418518066406, 0.373475341796875, 0.3732254638671875, 0.37330841064453124, 0.37334323120117185, 0.3735429077148438, 0.37338522338867186, 0.3733411865234375, 0.3737456665039062, 0.3734886474609375, 0.3733125, 0.3734097900390625, 0.3733340148925781, 0.37344869995117186, 0.373423095703125, 0.37322650146484376, 0.3733544921875, 0.37341592407226565, 0.37342105102539064, 0.3734323120117187, 0.37352139282226565, 0.37334426879882815, 0.802165771484375, 0.37316094970703123, 0.37318450927734376, 0.3732357177734375, 0.37309133911132814, 0.373180419921875, 0.37320193481445313, 0.37312820434570315, 0.37309747314453123, 0.3732991943359375, 0.37303192138671876, 0.37303402709960937, 0.37311279296875, 0.3731374206542969, 0.37302987670898435, 0.373185546875, 0.3731568603515625, 0.373138427734375, 0.37330841064453124, 0.37307904052734375, 0.37332275390625, 0.37319168090820315, 0.3731671142578125, 0.37323162841796875, 0.373291015625, 0.3731671142578125, 0.3733247985839844, 0.37319680786132814, 0.3731988525390625, 0.3732408447265625, 0.3732971496582031, 0.3731169128417969, 0.37325210571289064, 0.37314663696289063, 0.37325515747070315, 0.37322445678710936, 0.3734722595214844, 0.37343743896484377, 0.37350299072265625, 0.3734527893066406, 0.37328280639648437, 0.3733247985839844, 0.3734149169921875, 0.3734937744140625, 0.3734497375488281, 0.37342718505859374, 0.3736278991699219, 0.37341592407226565, 0.3733821411132813, 0.3734886474609375, 0.37328280639648437, 0.37341287231445314, 0.3734722595214844, 0.3733463134765625, 0.37333709716796876, 0.37325006103515623, 0.3733166198730469, 0.3733974914550781, 0.37348455810546877, 0.37356646728515625, 0.3734149169921875, 0.3733473205566406, 0.3733616638183594, 0.8014735107421875, 0.3731681213378906, 0.373327880859375, 0.3731148681640625, 0.3732234191894531, 0.37305752563476563, 0.3733063659667969, 0.3736422424316406, 0.37333810424804686, 0.3731097717285156, 0.3731251220703125, 0.3732336730957031, 0.37313534545898436, 0.3732090759277344, 0.3730411376953125, 0.37321624755859373, 0.3731558532714844, 0.37325515747070315, 0.373180419921875, 0.3731578979492188, 0.3731957702636719, 0.3736954956054688, 0.373222412109375, 0.37328692626953125, 0.37320294189453124, 0.37312921142578126, 0.37324697875976565, 0.3734835205078125, 0.37332275390625, 0.373254150390625, 0.37303399658203124, 0.37323980712890625, 0.3733760070800781, 0.3732357177734375, 0.37328897094726565, 0.373507080078125, 0.3731005554199219, 0.37347122192382814, 0.3732090759277344, 0.3733739624023438, 0.3733360595703125, 0.37336575317382814, 0.3731988525390625, 0.3733831787109375, 0.373317626953125, 0.3737794494628906, 0.3734589538574219, 0.37355316162109375, 0.37336270141601563, 0.3734169616699219, 0.37338827514648437, 0.3731497192382813, 0.37337493896484375, 0.373359619140625, 0.3732234191894531, 0.3732510986328125, 0.373265380859375, 0.373391357421875, 0.37323263549804686, 0.3733186645507813, 0.37346099853515624, 0.3735173034667969, 0.37349581909179685, 0.8010035400390625, 0.37311386108398437, 0.37337905883789063, 0.37324594116210935, 0.37309234619140624, 0.3731230773925781, 0.3730473022460937, 0.3733667907714844, 0.373212158203125, 0.3731978759765625, 0.3731783142089844, 0.3731865539550781, 0.37313433837890625, 0.37311697387695314, 0.3730810546875, 0.37312203979492187, 0.37326849365234377, 0.37339340209960936, 0.37326849365234377, 0.3732285461425781, 0.37314047241210935, 0.3731937255859375, 0.375251953125, 0.37332376098632813, 0.3734794311523438, 0.3732060241699219, 0.37325927734375, 0.37302783203125, 0.3732367248535156, 0.3730954284667969, 0.3731783752441406, 0.37331353759765623, 0.37304421997070314, 0.3731589050292969, 0.37371697998046877, 0.3734466552734375, 0.37335552978515624, 0.37338623046875, 0.37326028442382814, 0.3732725830078125, 0.37330841064453124, 0.3732674560546875, 0.3734364013671875, 0.373465087890625, 0.3733944396972656, 0.3733421936035156, 0.3734200439453125, 0.3733207092285156, 0.3734446105957031, 0.3736135559082031, 0.37353164672851563, 0.373454833984375, 0.3734169616699219, 0.3735173034667969, 0.37333197021484377, 0.37332171630859373, 0.3734517822265625, 0.3732705383300781, 0.3733811340332031, 0.373170166015625, 0.37344561767578127, 0.3734814758300781, 0.37345382690429685, 0.8020367431640625, 0.3732572021484375, 0.373148681640625, 0.3730821228027344, 0.37318350219726565, 0.3732367248535156, 0.3730616455078125, 0.3733196716308594, 0.37334527587890626, 0.3730780029296875, 0.3730626525878906, 0.3732080688476562, 0.3732008972167969, 0.37330023193359374, 0.3731322937011719, 0.373185546875, 0.37332791137695315, 0.3730851440429688, 0.3731671142578125, 0.37337905883789063, 0.37325311279296874, 0.3733207092285156, 0.3732787170410156, 0.3732213745117188, 0.373180419921875, 0.37319781494140625, 0.373291015625, 0.37346917724609374, 0.3733094482421875, 0.3732991943359375, 0.37332275390625, 0.37309747314453123, 0.373291015625, 0.3733145751953125, 0.37317120361328127, 0.37335552978515624, 0.3732654113769531, 0.37330227661132814, 0.37337191772460937, 0.37339852905273435, 0.373349365234375, 0.3732940673828125, 0.37326641845703123, 0.37327462768554687, 0.3731599426269531, 0.3735521240234375, 0.37336575317382814, 0.3732777099609375, 0.3734036560058594, 0.373222412109375, 0.373423095703125, 0.37342105102539064, 0.373496826171875, 0.37328793334960936, 0.373391357421875, 0.3733166198730469, 0.37332171630859373, 0.37341287231445314, 0.3732561950683594, 0.37350299072265625, 0.3734026184082031, 0.37336575317382814, 0.37333197021484377, 0.8009021606445312, 0.3735941162109375, 0.3736401977539062, 0.37326641845703123, 0.3731814270019531, 0.3732643737792969, 0.37304934692382813, 0.3731558532714844, 0.37303909301757815, 0.3731322937011719, 0.37307086181640625, 0.37325823974609373, 0.37325006103515623, 0.37372210693359376, 0.3732510681152344, 0.3730616455078125, 0.3732971496582031, 0.3732336730957031, 0.3733114929199219, 0.37319781494140625, 0.3732080688476562, 0.3732654113769531, 0.3731937255859375, 0.3732643737792969, 0.3733186645507813, 0.3733309326171875, 0.37337191772460937, 0.3734323120117187, 0.37324493408203124, 0.37316094970703123, 0.3732295532226563, 0.3733974914550781, 0.3733391418457031, 0.3731015625, 0.3731507263183594, 0.373291015625, 0.37332275390625, 0.37336370849609374, 0.37342926025390627, 0.3732940673828125, 0.37340570068359374, 0.37344052124023436, 0.3733309326171875, 0.3735900573730469, 0.3732919921875, 0.37349169921875, 0.3736330261230469, 0.3733186645507813, 0.37347430419921873, 0.37389004516601565, 0.37367706298828124, 0.37339340209960936, 0.37343435668945313, 0.3733974914550781, 0.3732715454101562, 0.37323468017578126, 0.3734026184082031, 0.3732787170410156, 0.37342926025390627, 0.37337905883789063, 0.3733831787109375, 0.3733616638183594, 0.37340057373046875, 0.8010618896484375, 0.3733186645507813, 0.3731128234863281, 0.3733063659667969, 0.37329510498046875, 0.3731568603515625, 0.37312103271484376, 0.3736278991699219, 0.37314663696289063, 0.37328997802734376, 0.37308416748046874, 0.3732357177734375, 0.37325210571289064, 0.3732336730957031, 0.3731988525390625, 0.37321728515625, 0.37312103271484376, 0.3731599426269531, 0.37319064331054685, 0.3733544921875, 0.37318246459960935, 0.3732090759277344, 0.3734046630859375, 0.37324697875976565, 0.3732572021484375, 0.3732213745117188, 0.373170166015625, 0.3734589538574219, 0.37343743896484377, 0.37326028442382814, 0.37303604125976564, 0.3733974914550781, 0.373327880859375, 0.37322750854492187, 0.37323776245117185, 0.37330535888671873, 0.3732715454101562, 0.3732613220214844, 0.37321728515625, 0.37325927734375, 0.3732367248535156, 0.3734118347167969, 0.3732510681152344, 0.373296142578125, 0.3736750183105469, 0.373580810546875, 0.3734517822265625, 0.3732705383300781, 0.37349786376953126, 0.3734149169921875, 0.3733677978515625, 0.37328897094726565, 0.3733463134765625, 0.3733125, 0.3732991943359375, 0.373401611328125, 0.373749755859375, 0.37331353759765623, 0.37323162841796875, 0.37353268432617187, 0.3736278991699219, 0.37345382690429685, 0.37330023193359374]",tokens/s,2.635527141458715,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -11309,7 +12241,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c10af-4a05847074bb074c2b52da3d;9f5e34c4-21d5-473a-8937-4e7ad96d7bd7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d42fe-488290042f0ae7c90c15a1f7;a78db277-9161-432a-b522-59fa52114b47) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11340,6 +12272,7 @@ OSError: s is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2263.904256,3257.401344,0.0,2610.95424,2376.633344,s,10,1.3856711425781252,0.1385671142578125,0.0011278118977040663,0.13823102569580079,0.13877978973388672,0.1403439888000488,0.1415953480529785,"[0.14190818786621093, 0.13788294982910157, 0.13812364196777344, 0.13839004516601563, 0.1379633026123047, 0.13809609985351562, 0.1382872009277344, 0.13817485046386718, 0.13841267395019533, 0.13843218994140624]",tokens/s,1847.4802002710144,kWh,1.6318407709195736e-06,8.94172149389916e-07,7.834545841452268e-06,1.036055876176176e-05,tokens/kWh,24709092.037085123,MB,2263.904256,3257.401344,0.0,2610.95424,2525.751808,s,10,74.90619775390626,7.490619775390625,0.002892086826923777,7.489885986328124,7.49303408203125,7.495257275390625,7.497035830078125,"[7.49748046875, 7.489609375, 7.4925400390625, 7.49201220703125, 7.4921337890625, 7.49016259765625, 7.48744384765625, 7.4889111328125, 7.4876474609375, 7.4882568359375]",tokens/s,8.410519007649755,kWh,8.842178812740963e-05,4.8461578623475653e-05,0.000419604657220747,0.0005564880239716324,tokens/kWh,113209.98347883853,,s,629,76.03107839202893,0.12087611827031607,0.016429380185563336,0.11885874938964844,0.11916492614746094,0.11930234832763673,0.2569760400390625,"[0.11889971160888672, 0.11878399658203125, 0.11925606536865234, 0.11902464294433594, 0.1188485107421875, 0.1191731185913086, 0.11938406372070312, 0.11878912353515625, 0.1189191665649414, 0.11888333129882812, 0.11871743774414062, 0.11923353576660156, 0.11888639831542969, 0.11883213043212891, 0.11902259063720703, 0.11896422576904297, 0.11902054595947266, 0.11877887725830077, 0.11896115112304688, 0.11885772705078125, 0.11894477081298828, 0.11872358703613281, 0.11921715545654298, 0.11882086181640625, 0.11885568237304688, 0.11888639831542969, 0.11897856140136719, 0.11893145751953126, 0.11888127899169922, 0.11925401306152343, 0.11921920013427735, 0.11904819488525391, 0.11908710479736329, 0.11901644897460938, 0.11889971160888672, 0.1191229476928711, 0.11946189117431641, 0.11878502655029297, 0.1189375991821289, 0.11886386871337891, 0.11896422576904297, 0.11879014587402344, 0.11913318634033203, 0.11885874938964844, 0.11890380859375, 0.11873587036132813, 0.11920076751708984, 0.11944755554199218, 0.1193543701171875, 0.11916902160644531, 0.11922637176513672, 0.11924479675292969, 0.11955506896972656, 0.11894886779785156, 0.11902259063720703, 0.11879833221435547, 0.1190645751953125, 0.11874201965332032, 0.11884134674072265, 0.11911577606201172, 0.11912703704833984, 0.11885363006591797, 0.2569912414550781, 0.1188823013305664, 0.11865805053710937, 0.11874201965332032, 0.11862834930419922, 0.11875737762451172, 0.11892633819580078, 0.11885772705078125, 0.11878092956542968, 0.1186662368774414, 0.11865190124511718, 0.11869184112548828, 0.11876557159423828, 0.11876044464111328, 0.1187962875366211, 0.118830078125, 0.11929804992675781, 0.11897138977050781, 0.11878297424316406, 0.11876044464111328, 0.11873689270019532, 0.11866726684570313, 0.11890892791748046, 0.11899903869628906, 0.11870310211181641, 0.11877069091796875, 0.11901747131347656, 0.11906764984130859, 0.11878399658203125, 0.11872563171386719, 0.11886796569824219, 0.11909939575195312, 0.11939737701416016, 0.11898880004882813, 0.11885363006591797, 0.11896832275390624, 0.11878195190429687, 0.11885977935791016, 0.11884236907958984, 0.1190635528564453, 0.11891712188720703, 0.11886386871337891, 0.11883519744873047, 0.11875635528564453, 0.11882291412353516, 0.11877273559570313, 0.118761474609375, 0.11913728332519531, 0.1192591323852539, 0.11893965148925781, 0.11885465240478515, 0.11907891082763672, 0.11896217346191407, 0.11902566528320313, 0.11883519744873047, 0.11901337432861328, 0.11892633819580078, 0.11896012878417969, 0.11881676483154296, 0.11883417510986329, 0.1189959716796875, 0.11886899566650391, 0.1189027862548828, 0.25737112426757813, 0.11880550384521485, 0.11871641540527343, 0.11895295715332031, 0.11919257354736328, 0.11897241973876953, 0.1187440643310547, 0.11885670471191406, 0.11871437072753906, 0.1191383056640625, 0.1187440643310547, 0.11878399658203125, 0.11873792266845704, 0.11884134674072265, 0.11891609954833984, 0.11900211334228515, 0.11896217346191407, 0.1190799331665039, 0.11884953308105468, 0.11889561462402344, 0.11884646606445312, 0.11894374084472656, 0.1189222412109375, 0.11903590393066406, 0.1187747802734375, 0.11887718200683593, 0.11879936218261719, 0.11874610900878907, 0.11870207977294922, 0.11871949005126953, 0.11879219055175781, 0.11888435363769531, 0.11882086181640625, 0.11873587036132813, 0.11910963439941406, 0.11885158538818359, 0.11942809295654297, 0.11903180694580077, 0.1191884765625, 0.11949874877929688, 0.11893452453613282, 0.11903385925292968, 0.11915264129638672, 0.1187962875366211, 0.11881881713867187, 0.11878502655029297, 0.11915058898925782, 0.11906047821044922, 0.11891712188720703, 0.11893657684326171, 0.11876966094970703, 0.11901747131347656, 0.11956940460205077, 0.11894169616699218, 0.11885363006591797, 0.11905228424072266, 0.11891814422607422, 0.11893043518066407, 0.11877887725830077, 0.11883929443359376, 0.11885977935791016, 0.11886284637451172, 0.11882803344726563, 0.25746636962890623, 0.11874201965332032, 0.11870207977294922, 0.1187747802734375, 0.1188362274169922, 0.11872358703613281, 0.11868160247802734, 0.11879014587402344, 0.11885260772705078, 0.11863756561279297, 0.11864985656738282, 0.11870207977294922, 0.11877375793457032, 0.1189744644165039, 0.11873382568359375, 0.11887206268310548, 0.11878092956542968, 0.11883519744873047, 0.1186693115234375, 0.11877887725830077, 0.11887513732910156, 0.1188362274169922, 0.11870310211181641, 0.11920178985595703, 0.11923865509033203, 0.1191884765625, 0.11896320343017579, 0.1189017562866211, 0.11871641540527343, 0.11927449798583985, 0.1189908447265625, 0.1191720962524414, 0.11884544372558593, 0.1189201889038086, 0.11882803344726563, 0.11889356994628907, 0.11882291412353516, 0.11890688323974609, 0.11884748840332031, 0.11882598114013672, 0.11903385925292968, 0.11936255645751953, 0.11917823791503906, 0.11911373138427735, 0.11909222412109376, 0.11881676483154296, 0.11902156829833985, 0.11887411499023437, 0.11882189178466797, 0.11884646606445312, 0.11913215637207031, 0.11890688323974609, 0.11885670471191406, 0.11890073394775391, 0.11893350219726563, 0.11897138977050781, 0.11930111694335938, 0.11934003448486329, 0.11905228424072266, 0.11880960083007812, 0.11903282928466796, 0.11903180694580077, 0.11902156829833985, 0.25693695068359373, 0.11868978881835937, 0.11935539245605468, 0.11892121887207031, 0.1187799072265625, 0.1186662368774414, 0.11860377502441406, 0.11915878295898437, 0.11897344207763672, 0.11868672180175781, 0.11873484802246094, 0.11875737762451172, 0.1185955810546875, 0.11869900512695312, 0.11919872283935547, 0.11896320343017579, 0.11864268493652344, 0.118940673828125, 0.11869081878662109, 0.11894681549072265, 0.11879116821289062, 0.11888127899169922, 0.11871949005126953, 0.11904307556152344, 0.11921715545654298, 0.11943424224853516, 0.11901030731201172, 0.1189969940185547, 0.11901439666748047, 0.11949056243896485, 0.11904102325439453, 0.1190830078125, 0.11907481384277344, 0.11894681549072265, 0.11885260772705078, 0.1188700180053711, 0.11913932800292969, 0.11917619323730469, 0.11892326354980469, 0.1189017562866211, 0.11899494171142579, 0.11891506958007812, 0.11873280334472656, 0.11906969451904297, 0.11905945587158204, 0.1188116455078125, 0.11880345916748047, 0.1189201889038086, 0.1188106231689453, 0.11904819488525391, 0.11894783782958984, 0.11883417510986329, 0.11885465240478515, 0.11891609954833984, 0.11878912353515625, 0.118940673828125, 0.11884236907958984, 0.11901849365234375, 0.11890483093261718, 0.11881472015380859, 0.11884544372558593, 0.11896832275390624, 0.11879424285888672, 0.256927734375, 0.11882086181640625, 0.11866316986083984, 0.11866521453857422, 0.11861094665527344, 0.1186344985961914, 0.11877887725830077, 0.11860684967041016, 0.11878195190429687, 0.11867955017089844, 0.11874508666992188, 0.11869798278808594, 0.11891506958007812, 0.11881267547607421, 0.11930726623535157, 0.11886080169677735, 0.1187409896850586, 0.1189222412109375, 0.11888947296142578, 0.1187430419921875, 0.11915264129638672, 0.11876863861083985, 0.11934207916259766, 0.11926118469238281, 0.11912089538574219, 0.11884339141845703, 0.11878399658203125, 0.11898470306396484, 0.11916287994384765, 0.11876454162597656, 0.11894989013671875, 0.11937177276611328, 0.11890585327148437, 0.11885158538818359, 0.11880550384521485, 0.11898572540283203, 0.11883519744873047, 0.11871641540527343, 0.11882086181640625, 0.11889356994628907, 0.11893965148925781, 0.11879424285888672, 0.1188362274169922, 0.11884441375732421, 0.11884544372558593, 0.11874508666992188, 0.11869798278808594, 0.11907174682617187, 0.11909734344482421, 0.11896115112304688, 0.11882803344726563, 0.11883827209472657, 0.11891506958007812, 0.11888639831542969, 0.1188505630493164, 0.11917005157470703, 0.1190830078125, 0.11898982238769532, 0.11903078460693359, 0.11894886779785156, 0.11903590393066406, 0.11890995025634765, 0.11884441375732421, 0.257412109375, 0.11882803344726563, 0.11874610900878907, 0.11872972869873047, 0.11867648315429688, 0.11894169616699218, 0.11878604888916015, 0.11885670471191406, 0.11871539306640624, 0.11885874938964844, 0.11888639831542969, 0.11874918365478515, 0.1187041244506836, 0.11886182403564453, 0.11865907287597656, 0.11878092956542968, 0.11879424285888672, 0.11882701110839844, 0.1187409896850586, 0.11877375793457032, 0.1187215347290039, 0.11869081878662109, 0.11869081878662109, 0.11883724975585938, 0.11875122833251953, 0.11888947296142578, 0.11891097259521484, 0.1188106231689453, 0.11873587036132813, 0.11875430297851562, 0.11876351928710938, 0.1187768325805664, 0.1188823013305664, 0.11892530822753906, 0.11900927734375, 0.11892530822753906, 0.11881574249267578, 0.11896524810791016, 0.11876863861083985, 0.11886592102050782, 0.11877581024169923, 0.1187440643310547, 0.11917926025390625, 0.11894271850585937, 0.11872767639160156, 0.11896524810791016, 0.11880857849121093, 0.11917721557617188, 0.11886386871337891, 0.11879424285888672, 0.1189191665649414, 0.11890892791748046, 0.11880754852294922, 0.11891404724121094, 0.11881472015380859, 0.11905433654785157, 0.11876863861083985, 0.11896832275390624, 0.1190666275024414, 0.11895091247558594, 0.11884031677246094, 0.11912601470947265, 0.1188362274169922, 0.25716326904296877, 0.11867955017089844, 0.11894681549072265, 0.11877069091796875, 0.11869593811035156, 0.11866111755371093, 0.11864166259765625, 0.11885158538818359, 0.11880960083007812, 0.11866316986083984, 0.1186723861694336, 0.11862220764160156, 0.11868978881835937, 0.11866521453857422, 0.11872972869873047, 0.11913932800292969, 0.11882701110839844, 0.1188362274169922, 0.11876044464111328, 0.11930316925048828, 0.11890585327148437, 0.11863040161132812, 0.11860173034667969, 0.11883519744873047, 0.11885158538818359, 0.11885363006591797, 0.11886899566650391, 0.11885158538818359, 0.1187061767578125, 0.11873382568359375, 0.11869491577148437, 0.11889561462402344, 0.11880242919921875, 0.11904102325439453, 0.11893350219726563, 0.11880242919921875, 0.11878502655029297, 0.11876249694824219, 0.11886080169677735, 0.118830078125, 0.11884031677246094, 0.11890483093261718, 0.11884134674072265, 0.11888333129882812, 0.11888025665283203, 0.1187430419921875, 0.11880038452148438, 0.11896115112304688, 0.11887308502197266, 0.11889459228515625, 0.11896627044677735, 0.11931238555908204, 0.1190113296508789, 0.11975577545166016, 0.11912601470947265, 0.11896217346191407, 0.11889663696289063, 0.1190492172241211, 0.11892121887207031, 0.11889663696289063, 0.11886284637451172, 0.11887206268310548, 0.11946495819091797, 0.25736398315429687, 0.11885260772705078, 0.11883827209472657, 0.11901952362060547, 0.11892838287353516, 0.11871027374267579, 0.1186693115234375, 0.1189969940185547, 0.11868364715576171, 0.1187041244506836, 0.11863040161132812, 0.11893657684326171, 0.11866419219970703, 0.11878399658203125, 0.11916390228271484, 0.1189201889038086, 0.11865702056884765, 0.11877581024169923, 0.11886592102050782, 0.11892121887207031, 0.11877887725830077, 0.11868262481689452, 0.118724609375, 0.11902361297607422, 0.11877069091796875, 0.11880960083007812, 0.11872563171386719, 0.11882803344726563, 0.11896012878417969, 0.11869798278808594, 0.11872665405273437, 0.11884441375732421, 0.11890892791748046, 0.11885158538818359, 0.11878604888916015, 0.11873689270019532, 0.11888742065429687, 0.11892736053466797, 0.1187215347290039, 0.11918335723876954, 0.11902464294433594, 0.11885465240478515, 0.11872051239013671, 0.11881472015380859, 0.11879936218261719, 0.11916083526611328, 0.11879833221435547, 0.11898777770996094, 0.1188106231689453, 0.11898265838623047, 0.11880140686035157, 0.11889254760742188, 0.11883929443359376, 0.11892633819580078, 0.11894886779785156, 0.11896217346191407, 0.11891302490234375, 0.11896012878417969, 0.11883827209472657, 0.11883724975585938, 0.11882086181640625, 0.11895091247558594, 0.11880242919921875, 0.2574468994140625, 0.11882086181640625, 0.11873792266845704, 0.11878297424316406, 0.11871949005126953, 0.11923865509033203, 0.11882291412353516, 0.11870310211181641, 0.11867750549316407, 0.11930521392822266, 0.11877785491943359, 0.11883110046386719, 0.1186693115234375, 0.1187041244506836, 0.11871846771240234, 0.11876454162597656, 0.11869388580322265, 0.11886592102050782, 0.11914342498779297, 0.11890483093261718, 0.11874201965332032, 0.11870310211181641, 0.11866521453857422, 0.11937382507324219, 0.11894271850585937, 0.11889049530029297, 0.11881574249267578, 0.11886592102050782, 0.11874816131591796, 0.11883519744873047, 0.11876044464111328, 0.11896217346191407, 0.11878092956542968, 0.11879833221435547, 0.11886592102050782, 0.11879116821289062, 0.1187799072265625, 0.11876863861083985, 0.11878604888916015, 0.11882495880126953, 0.11895398712158203, 0.11887718200683593, 0.11878297424316406, 0.11880857849121093, 0.11879116821289062, 0.11880448150634766, 0.1188362274169922, 0.1188362274169922, 0.11876249694824219, 0.11883417510986329, 0.11885977935791016, 0.11889663696289063, 0.11884134674072265, 0.11900723266601562, 0.11903590393066406, 0.11901337432861328, 0.11885670471191406, 0.11921305847167969, 0.11894783782958984, 0.11908096313476563, 0.11896729278564454, 0.11892326354980469, 0.11877069091796875]",tokens/s,8.2729327704228,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1488.359424,1579.679744,0.0,933.23264,856.69632,s,10,0.4187196159362792,0.041871961593627924,0.0014808966097482894,0.04126091194152832,0.043131636810302736,0.0445920425415039,0.04576036712646484,"[0.04605244827270508, 0.04131564712524414, 0.0411776008605957, 0.04146476745605469, 0.04118272018432617, 0.04280710220336914, 0.0412061767578125, 0.0415596809387207, 0.04101833724975586, 0.04093513488769531]",tokens/s,6113.876452326468,kWh,4.934550676908758e-07,2.7027020769764927e-07,1.3917580578494868e-06,2.155483333238012e-06,tokens/kWh,118766865.90539834,MB,1488.359424,1579.679744,0.0,933.23264,899.62496,s,10,24.019910400390625,2.4019910400390625,0.019429569356012314,2.394274169921875,2.414185791015625,2.436057373046875,2.453554638671875,"[2.457928955078125, 2.409325439453125, 2.399566650390625, 2.391402587890625, 2.39863720703125, 2.391660400390625, 2.391578369140625, 2.391282958984375, 2.391639892578125, 2.396887939453125]",tokens/s,26.228241050797365,kWh,2.945148775767949e-05,1.6140594842595416e-05,7.348580184415957e-05,0.00011907788444443448,tokens/kWh,529065.4960316985,,s,629,24.35256624984743,0.03871632154188778,0.004963175047223502,0.037943294525146484,0.03889213485717773,0.03936890869140625,0.079117353515625,"[0.040226814270019534, 0.04027084732055664, 0.040318977355957034, 0.03974041748046875, 0.040825855255126955, 0.04090367889404297, 0.03897753524780274, 0.039828479766845705, 0.04030361557006836, 0.03983871841430664, 0.04004761505126953, 0.04024524688720703, 0.03980287933349609, 0.039346176147460936, 0.038795265197753906, 0.039277568817138675, 0.03892326354980469, 0.03963289642333984, 0.039182334899902346, 0.038902782440185545, 0.039244800567626956, 0.03907788848876953, 0.0392806396484375, 0.03914035034179687, 0.03907788848876953, 0.03934003067016602, 0.03828940963745117, 0.038212608337402344, 0.03859251022338867, 0.038125568389892575, 0.03932876968383789, 0.03894169616699219, 0.03829043197631836, 0.03913216018676758, 0.03880038452148438, 0.038793216705322264, 0.039011329650878904, 0.039003135681152344, 0.039147518157958985, 0.038547454833984376, 0.03806617736816406, 0.03796889495849609, 0.03785420989990235, 0.03794432067871094, 0.03905023956298828, 0.03965235137939453, 0.03999129486083984, 0.039384063720703126, 0.039397377014160156, 0.0378419189453125, 0.03808563232421875, 0.037797889709472655, 0.037797889709472655, 0.03783987045288086, 0.037891071319580076, 0.03788083267211914, 0.03912499237060547, 0.038760448455810545, 0.03786342239379883, 0.038950912475585936, 0.0378419189453125, 0.03783270263671875, 0.0792985610961914, 0.03790643310546875, 0.03781017684936523, 0.038370304107666016, 0.03831193542480469, 0.03775078582763672, 0.03785420989990235, 0.038435840606689455, 0.038365184783935545, 0.03781017684936523, 0.037905406951904294, 0.03861094284057617, 0.03815731048583984, 0.037917697906494144, 0.037776382446289065, 0.03875532913208008, 0.03827711868286133, 0.03894784164428711, 0.0380313606262207, 0.038877185821533204, 0.037923839569091795, 0.03808870315551758, 0.03812351989746094, 0.03917004776000976, 0.0390645751953125, 0.03829862213134766, 0.03843379211425781, 0.038542335510253906, 0.03800166320800781, 0.0378869743347168, 0.038740993499755856, 0.03850239944458008, 0.03803955078125, 0.038438911437988284, 0.03892838287353516, 0.03817267227172851, 0.038012928009033206, 0.03801190567016602, 0.039065601348876954, 0.03803033447265625, 0.03791462326049805, 0.03789619064331055, 0.03887923049926758, 0.038371326446533204, 0.03790848159790039, 0.037904384613037106, 0.03786956787109375, 0.037988353729248046, 0.037943294525146484, 0.03816755294799805, 0.03861503982543945, 0.03803340911865234, 0.037966846466064456, 0.03835084915161133, 0.03848396682739258, 0.03804159927368164, 0.03789926528930664, 0.038539264678955076, 0.038340606689453126, 0.03802521514892578, 0.03793203353881836, 0.038679550170898434, 0.03826892852783203, 0.07922994995117187, 0.03890790557861328, 0.03783679962158203, 0.03792281723022461, 0.03792588806152344, 0.03827507019042969, 0.03808153533935547, 0.03808768081665039, 0.03786137771606445, 0.03799552154541016, 0.0380313606262207, 0.03805388641357422, 0.03804569625854492, 0.03840409469604492, 0.03806412887573242, 0.037905406951904294, 0.03795455932617187, 0.03794636917114258, 0.03813478469848633, 0.03800985717773438, 0.03793203353881836, 0.03811123275756836, 0.038125568389892575, 0.03792793655395508, 0.03794739151000977, 0.03824435043334961, 0.0380847053527832, 0.0379257926940918, 0.03787161636352539, 0.037819393157958986, 0.03803647994995117, 0.037956607818603515, 0.039212032318115236, 0.03805184173583984, 0.0379504623413086, 0.0379156494140625, 0.04073574447631836, 0.038391807556152346, 0.03786649703979492, 0.03888947296142578, 0.0400711669921875, 0.03781222534179687, 0.03783679962158203, 0.037981182098388674, 0.03775078582763672, 0.03779891204833984, 0.03775590515136719, 0.0378419189453125, 0.03782144165039063, 0.037781505584716796, 0.03789311981201172, 0.03784396743774414, 0.03780710220336914, 0.038095870971679685, 0.037797889709472655, 0.03782451248168945, 0.03784601593017578, 0.03794124984741211, 0.037806079864501956, 0.037781505584716796, 0.037884929656982425, 0.03780710220336914, 0.03832627105712891, 0.07955046081542969, 0.03795251083374023, 0.03794947052001953, 0.03773539352416992, 0.03782860946655273, 0.03780710220336914, 0.037874687194824216, 0.037789695739746096, 0.037867584228515626, 0.03773331069946289, 0.038191104888916014, 0.03783679962158203, 0.037872638702392575, 0.03788185501098633, 0.037850112915039064, 0.0378869743347168, 0.03800371170043945, 0.037870655059814455, 0.038172607421875, 0.037921791076660154, 0.03788390350341797, 0.03808256149291992, 0.03796377563476563, 0.037819393157958986, 0.03782553482055664, 0.03784806442260742, 0.038128639221191404, 0.038128639221191404, 0.03777228927612305, 0.03794944000244141, 0.03797196960449219, 0.03817062377929688, 0.03783168029785156, 0.037855232238769534, 0.037928958892822266, 0.03805286407470703, 0.0382371826171875, 0.03802828979492188, 0.037988353729248046, 0.037961727142333986, 0.038179840087890625, 0.03786240005493164, 0.037978111267089845, 0.0379156494140625, 0.038059009552001956, 0.038112255096435545, 0.037910526275634765, 0.03792588806152344, 0.03811942291259766, 0.03815935897827148, 0.037978111267089845, 0.037874687194824216, 0.03809996795654297, 0.03806924819946289, 0.03789311981201172, 0.037894176483154296, 0.037800926208496094, 0.038100990295410156, 0.037956607818603515, 0.037953536987304685, 0.037889022827148434, 0.03792899322509766, 0.03823203277587891, 0.07917670440673828, 0.03791462326049805, 0.03787366485595703, 0.03793305587768555, 0.03789926528930664, 0.03786240005493164, 0.038117374420166016, 0.03861913681030273, 0.03791974258422852, 0.038007808685302735, 0.03791257476806641, 0.03793817520141601, 0.03802521514892578, 0.0379422721862793, 0.03785830307006836, 0.038046718597412106, 0.037959678649902344, 0.037958656311035156, 0.03790028762817383, 0.037817344665527344, 0.03794124984741211, 0.03805184173583984, 0.03784294509887695, 0.03805388641357422, 0.03809894561767578, 0.03805286407470703, 0.03784499359130859, 0.03791974258422852, 0.03803955078125, 0.03802828979492188, 0.038161407470703124, 0.03802214431762695, 0.037894142150878905, 0.03818598556518555, 0.037917697906494144, 0.03795455932617187, 0.037975040435791016, 0.03828224182128906, 0.038002689361572264, 0.037910526275634765, 0.03793100738525391, 0.03803852844238281, 0.03797708892822266, 0.03793612670898437, 0.0378869743347168, 0.0380313606262207, 0.03825971221923828, 0.037953536987304685, 0.03790643310546875, 0.038029312133789066, 0.03808153533935547, 0.03808870315551758, 0.0379791374206543, 0.03979673767089844, 0.03898777770996094, 0.03879116821289062, 0.037969921112060545, 0.03789926528930664, 0.03830579376220703, 0.038193153381347655, 0.038182910919189454, 0.03849523162841797, 0.03826278305053711, 0.08257740783691406, 0.03822489547729492, 0.03793920135498047, 0.03778559875488281, 0.03784499359130859, 0.037872638702392575, 0.038141952514648435, 0.037921791076660154, 0.03786342239379883, 0.03808256149291992, 0.03808051300048828, 0.0378419189453125, 0.03813273620605469, 0.038711296081542966, 0.038114303588867186, 0.03786240005493164, 0.03781119918823242, 0.03850137710571289, 0.03799347305297852, 0.0378152961730957, 0.0378869743347168, 0.037989376068115234, 0.03785728073120117, 0.037850112915039064, 0.0379607048034668, 0.03787782287597656, 0.03834259033203125, 0.03791462326049805, 0.03792588806152344, 0.03778047943115234, 0.038484992980957033, 0.03798015975952149, 0.03783782577514649, 0.03778559875488281, 0.03781017684936523, 0.0379791374206543, 0.03785625457763672, 0.03782553482055664, 0.0378603515625, 0.037923839569091795, 0.0378152961730957, 0.03780505752563477, 0.03792998504638672, 0.03794124984741211, 0.03778355026245117, 0.03781119918823242, 0.03775385665893555, 0.03779891204833984, 0.03803238296508789, 0.03786956787109375, 0.03816960144042969, 0.037814273834228515, 0.03785625457763672, 0.038043647766113284, 0.03777849578857422, 0.037824447631835935, 0.037713920593261716, 0.03772108840942383, 0.03790233612060547, 0.0377784309387207, 0.037776382446289065, 0.0378152961730957, 0.037921791076660154, 0.07919411468505859, 0.03782758331298828, 0.03773747253417969, 0.03775385665893555, 0.03801804733276367, 0.037800960540771485, 0.03785728073120117, 0.03783679962158203, 0.037719039916992186, 0.038024192810058595, 0.03775590515136719, 0.037768192291259765, 0.037823486328125, 0.037910526275634765, 0.03777433776855469, 0.03780505752563477, 0.03785113525390625, 0.0377968635559082, 0.037874687194824216, 0.03781222534179687, 0.03783475112915039, 0.03781119918823242, 0.037754878997802735, 0.03797094345092773, 0.03784499359130859, 0.03780710220336914, 0.03785728073120117, 0.03789209747314453, 0.0379791374206543, 0.03779891204833984, 0.03774771118164062, 0.03775078582763672, 0.03789926528930664, 0.03783987045288086, 0.037738529205322266, 0.037819358825683595, 0.03775795364379883, 0.037889022827148434, 0.037773311614990236, 0.03777740859985351, 0.03778662490844727, 0.03790950393676758, 0.03822079849243164, 0.04011008071899414, 0.03838054275512695, 0.0379576301574707, 0.03801599884033203, 0.03980595016479492, 0.03857920074462891, 0.03786956787109375, 0.03796384048461914, 0.03790534210205078, 0.03787059020996094, 0.037891071319580076, 0.037838848114013675, 0.03804774475097656, 0.03783270263671875, 0.03785113525390625, 0.038819839477539066, 0.038046718597412106, 0.03789823913574219, 0.03790233612060547, 0.03783679962158203, 0.08027750396728515, 0.038125568389892575, 0.03789619064331055, 0.03792588806152344, 0.03793100738525391, 0.03866726303100586, 0.03827916717529297, 0.03803443145751953, 0.03789619064331055, 0.03781222534179687, 0.03788185501098633, 0.03791257476806641, 0.03783065414428711, 0.03784505462646484, 0.03773846435546875, 0.03788899230957031, 0.03809280014038086, 0.03783475112915039, 0.03798323059082031, 0.03775385665893555, 0.037966846466064456, 0.038214656829833986, 0.03794124984741211, 0.03781119918823242, 0.037868545532226565, 0.03785113525390625, 0.03793100738525391, 0.0380211181640625, 0.03793817520141601, 0.037855232238769534, 0.03802828979492188, 0.0378081283569336, 0.03788185501098633, 0.03851264190673828, 0.03827097702026367, 0.03790643310546875, 0.03789619064331055, 0.03783065414428711, 0.037907455444335936, 0.03800985717773438, 0.03793920135498047, 0.03778867340087891, 0.03791872024536133, 0.037806079864501956, 0.03793920135498047, 0.03792793655395508, 0.037855232238769534, 0.03791974258422852, 0.03787161636352539, 0.03796582412719727, 0.038024192810058595, 0.03782860946655273, 0.03795251083374023, 0.03783782577514649, 0.037956607818603515, 0.03788083267211914, 0.03803964614868164, 0.03777833557128906, 0.037766143798828124, 0.03796275329589844, 0.03784703826904297, 0.03799859237670898, 0.03801804733276367, 0.07892582702636719, 0.03798527908325195, 0.037719039916992186, 0.03776921463012695, 0.03804467010498047, 0.037907455444335936, 0.03788288116455078, 0.03789516830444336, 0.03778867340087891, 0.037966846466064456, 0.03824844741821289, 0.03791155242919922, 0.03777228927612305, 0.03783065414428711, 0.03782758331298828, 0.03794739151000977, 0.03860070419311523, 0.0380948486328125, 0.037868545532226565, 0.03780505752563477, 0.0378152961730957, 0.037822463989257815, 0.03793407821655274, 0.03778771209716797, 0.03782342529296875, 0.037855232238769534, 0.03776921463012695, 0.038046718597412106, 0.037803009033203126, 0.03780198287963867, 0.037838848114013675, 0.03789516830444336, 0.03798015975952149, 0.0378603515625, 0.03887923049926758, 0.03785625457763672, 0.03788390350341797, 0.03847884750366211, 0.03800064086914062, 0.03786342239379883, 0.03780710220336914, 0.037826591491699216, 0.03794736099243164, 0.0378787841796875, 0.03783065414428711, 0.03801702499389648, 0.03784601593017578, 0.03804569625854492, 0.03787571334838867, 0.03788288116455078, 0.03784806442260742, 0.03797094345092773, 0.038019073486328124, 0.03792281723022461, 0.037956607818603515, 0.03784601593017578, 0.03790848159790039, 0.03802316665649414, 0.037868545532226565, 0.037868545532226565, 0.03931238555908203, 0.03824435043334961, 0.038171646118164065, 0.07896473693847657, 0.03796480178833008, 0.038031391143798825, 0.038429664611816405, 0.03807231903076172, 0.03814297485351562, 0.04003839874267578, 0.038354942321777344, 0.038171646118164065, 0.03779481506347656, 0.03799347305297852, 0.03781631851196289, 0.037822463989257815, 0.03803852844238281, 0.037795841217041014, 0.03802009582519531, 0.03805184173583984, 0.037812255859375, 0.038042591094970706, 0.0378787841796875, 0.03813785552978516, 0.03794636917114258, 0.038004737854003906, 0.03808153533935547, 0.03786240005493164, 0.03785318374633789, 0.03793817520141601, 0.0379156494140625, 0.03786137771606445, 0.03794124984741211, 0.037972991943359374, 0.03801599884033203, 0.03807743835449219, 0.038757377624511716, 0.038007808685302735, 0.03791155242919922, 0.03793407821655274, 0.03837849426269531, 0.038100990295410156, 0.037975040435791016, 0.037891071319580076, 0.03786751937866211, 0.03800678253173828, 0.03780505752563477, 0.03791360092163086, 0.037907455444335936, 0.03789209747314453, 0.03788288116455078, 0.037894142150878905, 0.03790335845947266, 0.0380497932434082, 0.0378787841796875, 0.0379504623413086, 0.03782963180541992, 0.037920768737792966, 0.03815731048583984, 0.03894476699829102, 0.03841228866577148, 0.0381102066040039, 0.03795455932617187, 0.03790848159790039, 0.03784396743774414, 0.03802521514892578]",tokens/s,25.828900065262776,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1279.995904,1017.643008,0.0,371.195904,277.410816,s,10,0.4175716514587402,0.04175716514587402,0.001329291485763711,0.04133342361450196,0.042088745498657223,0.04389397258758545,0.04533815425872803,"[0.045699199676513674, 0.04128799819946289, 0.04128377532958984, 0.04097267150878906, 0.04157187271118164, 0.04107036972045899, 0.04121916961669922, 0.04137884902954102, 0.04168758392333984, 0.04140016174316406]",tokens/s,6130.684377296505,kWh,4.643787057311447e-07,2.5445859858069414e-07,1.253729761153002e-06,1.972567065464841e-06,tokens/kWh,129780124.83427167,MB,1279.995904,1017.643008,0.0,371.195904,323.048448,s,10,24.960871826171868,2.496087182617187,0.007120166670465486,2.4940638427734374,2.4991390625000003,2.5077969482421874,2.5147232568359374,"[2.516454833984375, 2.49401806640625, 2.495659423828125, 2.49662109375, 2.49185595703125, 2.493477294921875, 2.4905234375, 2.49093701171875, 2.494109619140625, 2.497215087890625]",tokens/s,25.239503026470214,kWh,2.763858284111377e-05,1.5146819142039876e-05,6.740021405084962e-05,0.00011018561603400326,tokens/kWh,571762.4701626954,,s,629,25.29028705596926,0.04020713363429131,0.0049559405951164134,0.03952540969848633,0.039936614990234376,0.040598938751220705,0.08109604888916017,"[0.04100912094116211, 0.04127743911743164, 0.04177305603027344, 0.04085964965820312, 0.041427967071533206, 0.04092620849609375, 0.04130918502807617, 0.040700927734375, 0.042106880187988284, 0.04071731185913086, 0.040460289001464846, 0.04082995223999023, 0.03992473602294922, 0.039828479766845705, 0.03968511962890625, 0.039523326873779296, 0.03983257675170898, 0.039553024291992187, 0.039913471221923826, 0.039607295989990236, 0.03934515380859375, 0.03961139297485351, 0.03963187026977539, 0.039706623077392575, 0.03935334396362305, 0.03937484741210937, 0.03951923370361328, 0.039556095123291016, 0.03966873550415039, 0.03936972808837891, 0.03945574569702148, 0.039840831756591796, 0.0411473274230957, 0.039992321014404295, 0.03950080108642578, 0.0393994255065918, 0.03961958312988281, 0.039554046630859374, 0.03956838226318359, 0.039792640686035156, 0.03939740753173828, 0.03990217590332031, 0.03950284957885742, 0.03952537536621094, 0.039384063720703126, 0.039395328521728515, 0.0393809928894043, 0.03959091186523438, 0.03945676803588867, 0.039395328521728515, 0.03952947235107422, 0.03961139297485351, 0.03948441696166992, 0.0394967041015625, 0.03940147018432617, 0.039337982177734376, 0.04001484680175781, 0.03948953628540039, 0.03933695983886719, 0.039452671051025394, 0.03969023895263672, 0.03961958312988281, 0.08116633605957031, 0.03934003067016602, 0.03947724914550781, 0.03954687881469727, 0.03953664016723633, 0.039618560791015625, 0.039354366302490236, 0.039488510131835936, 0.039809024810791016, 0.03934310531616211, 0.039357440948486325, 0.03944243240356445, 0.03957657623291016, 0.03940761566162109, 0.039419902801513675, 0.03942399978637695, 0.0395489273071289, 0.039482368469238284, 0.03993907165527344, 0.03930419158935547, 0.0394967041015625, 0.03933491134643555, 0.039373825073242184, 0.03941785430908203, 0.039444480895996094, 0.039467041015625, 0.039427040100097656, 0.03946086502075195, 0.039478271484375, 0.039577598571777346, 0.03945471954345703, 0.039298046112060545, 0.039298046112060545, 0.03948339080810547, 0.039370750427246096, 0.03942502212524414, 0.03943833541870117, 0.039602176666259765, 0.03958988952636719, 0.03945881652832031, 0.039532543182373044, 0.04032716751098633, 0.04127542495727539, 0.040283103942871094, 0.03961139297485351, 0.03957657623291016, 0.03948953628540039, 0.03951615905761719, 0.039367679595947266, 0.03977833557128906, 0.039686111450195315, 0.04003123092651367, 0.04087910461425781, 0.040207359313964845, 0.03972403335571289, 0.03946393585205078, 0.039387134552001955, 0.039554046630859374, 0.03936972808837891, 0.03947724914550781, 0.0394005126953125, 0.03951507186889648, 0.03940662384033203, 0.08121955108642578, 0.039504894256591795, 0.039790592193603515, 0.03975270462036133, 0.03982131195068359, 0.039556129455566406, 0.039650272369384766, 0.03966259384155273, 0.03948339080810547, 0.03957145690917969, 0.03962265777587891, 0.039591934204101564, 0.03957452774047852, 0.03968716812133789, 0.03951103973388672, 0.039772159576416014, 0.03950592041015625, 0.039828479766845705, 0.03968819046020508, 0.039564289093017575, 0.039532543182373044, 0.03951615905761719, 0.03973324966430664, 0.03973222351074219, 0.03958476638793945, 0.03959807968139648, 0.03997183990478516, 0.03969740676879883, 0.03954380798339844, 0.0401080322265625, 0.03958784103393555, 0.03946188735961914, 0.03951718521118164, 0.03965542221069336, 0.03964416122436523, 0.039585792541503906, 0.039599105834960936, 0.039634944915771485, 0.0393891830444336, 0.03955712127685547, 0.03965542221069336, 0.03983161544799805, 0.03946899032592773, 0.039577598571777346, 0.039572479248046875, 0.03957555389404297, 0.039698463439941406, 0.03968406295776367, 0.039599105834960936, 0.03945779037475586, 0.03973529434204102, 0.038989822387695314, 0.03940147018432617, 0.03963904190063477, 0.03952435302734375, 0.03980287933349609, 0.039634944915771485, 0.03947724914550781, 0.03947417449951172, 0.0395786247253418, 0.039591934204101564, 0.03949465560913086, 0.03950899124145508, 0.08075161743164062, 0.03947520065307617, 0.03970457458496094, 0.03956329727172852, 0.03949052810668945, 0.03964108657836914, 0.03945369720458984, 0.03959091186523438, 0.03958476638793945, 0.03949465560913086, 0.03940147018432617, 0.03942502212524414, 0.039354366302490236, 0.039539710998535156, 0.03950387191772461, 0.03942297744750976, 0.039923713684082034, 0.03959603118896484, 0.03958169555664062, 0.03936870574951172, 0.039532543182373044, 0.039547904968261716, 0.039572479248046875, 0.03954687881469727, 0.03948748779296875, 0.03956531143188476, 0.04043366241455078, 0.040976383209228515, 0.039962623596191404, 0.039689216613769535, 0.039373825073242184, 0.03946188735961914, 0.03932364654541016, 0.03958476638793945, 0.039539710998535156, 0.03945471954345703, 0.039446529388427735, 0.039629825592041014, 0.039375873565673826, 0.03932467269897461, 0.039597057342529295, 0.040233985900878906, 0.04049612808227539, 0.03984896087646484, 0.039357440948486325, 0.03942195129394531, 0.03954483032226563, 0.039456832885742185, 0.03950073623657226, 0.03953049468994141, 0.03962060928344727, 0.039444480895996094, 0.039754753112792966, 0.03976806259155274, 0.04006911849975586, 0.039629825592041014, 0.039419902801513675, 0.03955199813842773, 0.03977830505371094, 0.039602176666259765, 0.040025089263916014, 0.03968819046020508, 0.03958169555664062, 0.08158719635009766, 0.03952844619750977, 0.03982643127441406, 0.03968000030517578, 0.03958476638793945, 0.03935334396362305, 0.039572479248046875, 0.040210430145263674, 0.03953561782836914, 0.039654399871826174, 0.03932057571411133, 0.03947417449951172, 0.03926323318481445, 0.039636993408203126, 0.039689216613769535, 0.03954278564453125, 0.039504894256591795, 0.0395601921081543, 0.03949055862426758, 0.0393963508605957, 0.03964518356323242, 0.039634944915771485, 0.039430145263671876, 0.03942297744750976, 0.03927347183227539, 0.039599105834960936, 0.039367679595947266, 0.039485439300537106, 0.039806976318359374, 0.040528896331787106, 0.039532543182373044, 0.039362560272216796, 0.03942502212524414, 0.039651329040527344, 0.03954483032226563, 0.03994521713256836, 0.040018943786621096, 0.03948748779296875, 0.039367679595947266, 0.039653377532958986, 0.03947110366821289, 0.03956633758544922, 0.03923660659790039, 0.039523326873779296, 0.03932876968383789, 0.039406593322753904, 0.039462913513183595, 0.03944140625, 0.03933184051513672, 0.039863296508789066, 0.03953664016723633, 0.039504894256591795, 0.039191551208496093, 0.040030208587646485, 0.03945881652832031, 0.03943324661254883, 0.03956835174560547, 0.0393891830444336, 0.03934822463989258, 0.039390209197998044, 0.03933292770385742, 0.03960108947753906, 0.03945881652832031, 0.08104959869384766, 0.03950592041015625, 0.03941891098022461, 0.039413726806640625, 0.03958988952636719, 0.03938102340698242, 0.03919769668579102, 0.03935843276977539, 0.03926323318481445, 0.03933184051513672, 0.03955507278442383, 0.039395328521728515, 0.0393963508605957, 0.039449600219726565, 0.03967795181274414, 0.039828479766845705, 0.039708671569824217, 0.0393994255065918, 0.03942195129394531, 0.03932160186767578, 0.039772159576416014, 0.0416184310913086, 0.03949363327026367, 0.03941273498535156, 0.0393114242553711, 0.039466945648193356, 0.03957555389404297, 0.039329792022705076, 0.039397377014160156, 0.03927859115600586, 0.039390239715576175, 0.039481311798095706, 0.03924889755249023, 0.039397377014160156, 0.03929600143432617, 0.039414783477783204, 0.039723007202148435, 0.039465984344482424, 0.04114329528808594, 0.04011625671386719, 0.039448543548583986, 0.03947315216064453, 0.03952537536621094, 0.039370750427246096, 0.03939123153686523, 0.03942604827880859, 0.039684097290039064, 0.03952742385864258, 0.040645633697509766, 0.039919647216796875, 0.03952329635620117, 0.03967078399658203, 0.03952540969848633, 0.03946083068847656, 0.039446529388427735, 0.03928268814086914, 0.03962879943847656, 0.039470142364501956, 0.03950073623657226, 0.03953868865966797, 0.03960422515869141, 0.03969945526123047, 0.03978854370117187, 0.08111411285400391, 0.040289279937744144, 0.0393072624206543, 0.03955199813842773, 0.039501823425292966, 0.0393891830444336, 0.0393359375, 0.03960934448242188, 0.03942604827880859, 0.039554046630859374, 0.03937279891967774, 0.03945369720458984, 0.03937177658081055, 0.039611423492431644, 0.03978031921386719, 0.0395335693359375, 0.039449600219726565, 0.03945574569702148, 0.03949158477783203, 0.039651329040527344, 0.039408641815185545, 0.039446529388427735, 0.039370750427246096, 0.039618560791015625, 0.03969126510620117, 0.03963391876220703, 0.039793663024902344, 0.03998822402954102, 0.03951513671875, 0.03942502212524414, 0.03936460876464844, 0.03952844619750977, 0.03944550323486328, 0.039468032836914066, 0.03973836898803711, 0.039488510131835936, 0.039398399353027344, 0.03951103973388672, 0.039433216094970705, 0.03953049468994141, 0.039384063720703126, 0.039610366821289066, 0.03942195129394531, 0.03937484741210937, 0.039444480895996094, 0.039605247497558595, 0.0394700813293457, 0.0394598388671875, 0.03948134231567383, 0.03951103973388672, 0.039702529907226565, 0.03990528106689453, 0.03952844619750977, 0.03959807968139648, 0.039534591674804685, 0.03949772644042969, 0.03940966415405273, 0.039390209197998044, 0.03936153411865234, 0.03949260711669922, 0.03947724914550781, 0.039478271484375, 0.03961964797973633, 0.08132498931884766, 0.039577598571777346, 0.039392257690429686, 0.039580673217773435, 0.039521278381347655, 0.03939328002929687, 0.039190528869628906, 0.03920588684082031, 0.039411712646484375, 0.0393891830444336, 0.03947110366821289, 0.03920588684082031, 0.03942911911010742, 0.03978137588500977, 0.03956531143188476, 0.03957452774047852, 0.039624702453613284, 0.03947110366821289, 0.03937996673583984, 0.03949676895141602, 0.03960723114013672, 0.03957350540161133, 0.03959299087524414, 0.04049200057983399, 0.040079360961914064, 0.03965235137939453, 0.03947315216064453, 0.039523326873779296, 0.03956633758544922, 0.03941785430908203, 0.03943219375610352, 0.03935539245605469, 0.03939328002929687, 0.03963596725463867, 0.03949772644042969, 0.039539710998535156, 0.03936460876464844, 0.039382015228271484, 0.03985203170776367, 0.04112588882446289, 0.03990835189819336, 0.039303199768066406, 0.03935740661621094, 0.03930828857421875, 0.039629825592041014, 0.039370750427246096, 0.03955507278442383, 0.039414783477783204, 0.039392257690429686, 0.03963391876220703, 0.039408641815185545, 0.03947727966308594, 0.03938198471069336, 0.03943116760253906, 0.039287841796875, 0.03957244873046875, 0.03941686248779297, 0.03938608169555664, 0.03933695983886719, 0.03938816070556641, 0.03966566467285156, 0.03951001739501953, 0.03967795181274414, 0.0816015396118164, 0.03950899124145508, 0.03951308822631836, 0.039567359924316405, 0.03965951919555664, 0.03955820846557617, 0.03948230361938477, 0.03946086502075195, 0.03948339080810547, 0.03967385482788086, 0.039572479248046875, 0.03956326293945313, 0.03950387191772461, 0.03951103973388672, 0.03965446472167969, 0.03958572769165039, 0.03975372695922851, 0.039676959991455076, 0.03963081741333008, 0.03968307113647461, 0.039874561309814455, 0.03963596725463867, 0.039359489440917966, 0.03943936157226562, 0.03984998321533203, 0.03992166519165039, 0.039825408935546876, 0.03975987243652344, 0.03933695983886719, 0.03944755172729492, 0.03970355224609375, 0.03947520065307617, 0.03924070358276367, 0.039408641815185545, 0.03939328002929687, 0.0395786247253418, 0.039427135467529295, 0.039470016479492186, 0.03935641479492188, 0.039585792541503906, 0.03967488098144531, 0.039550975799560545, 0.039375873565673826, 0.039498752593994144, 0.03950387191772461, 0.03968307113647461, 0.0395601921081543, 0.03951513671875, 0.039567359924316405, 0.03954278564453125, 0.040025089263916014, 0.03950694274902344, 0.03945369720458984, 0.03981414413452149, 0.03958169555664062, 0.03971583938598633, 0.040190975189208986, 0.039646209716796874, 0.03933184051513672, 0.03954380798339844, 0.03963289642333984, 0.03962777709960937, 0.039547904968261716, 0.081259521484375, 0.03940454483032227, 0.03937996673583984, 0.03957452774047852, 0.039414783477783204, 0.039419902801513675, 0.03935644912719727, 0.03931337738037109, 0.03946700668334961, 0.0394117431640625, 0.03942396926879883, 0.039762943267822266, 0.039406593322753904, 0.039550975799560545, 0.039481407165527345, 0.03942700958251953, 0.0395489273071289, 0.03993600082397461, 0.04035379028320313, 0.04098361587524414, 0.03966969680786133, 0.03950080108642578, 0.039616512298583983, 0.03977830505371094, 0.039462913513183595, 0.03949260711669922, 0.039669761657714846, 0.03949977493286133, 0.03933388900756836, 0.039332862854003905, 0.039398399353027344, 0.03946700668334961, 0.039672832489013675, 0.039769088745117184, 0.040182785034179686, 0.04107059097290039, 0.04016742324829101, 0.03948441696166992, 0.03967078399658203, 0.03942604827880859, 0.03934207916259766, 0.03936460876464844, 0.03937279891967774, 0.03931545639038086, 0.04087807846069336, 0.04114636611938476, 0.039907329559326174, 0.03947110366821289, 0.03962879943847656, 0.03942707061767578, 0.03959609603881836, 0.03944236755371094, 0.039609375, 0.03951305770874024, 0.03944857788085938, 0.03932160186767578, 0.03956531143188476, 0.03942399978637695, 0.03932160186767578, 0.03932057571411133, 0.039365631103515625, 0.03976499176025391, 0.03956326293945313]",tokens/s,24.871208405344607,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2019.69664,5274.861568,0.0,4628.414464,4279.64672,s,10,3.1774854736328124,0.31774854736328123,0.0007817722146641278,0.3174602966308594,0.3187982666015625,0.3189695983886719,0.3191066638183594,"[0.31848922729492185, 0.3174812927246094, 0.31692730712890627, 0.31720437622070313, 0.3174393005371094, 0.31661932373046875, 0.3174072570800781, 0.31876019287109375, 0.31914093017578127, 0.3180162658691406]",tokens/s,805.6685140634671,kWh,3.737864001757569e-06,2.0481879523572387e-06,1.80164467048112e-05,2.3802498658926007e-05,tokens/kWh,10755173.38193397,MB,2019.69664,5274.861568,0.0,4628.414464,4463.185408,s,10,176.9617734375,17.69617734375,0.0029728792469339506,17.696136718749997,17.6995603515625,17.699568261718753,17.699574589843753,"[17.69631640625, 17.695939453125, 17.698544921875, 17.693357421875, 17.690775390625, 17.69253125, 17.699216796875, 17.69955859375, 17.699576171875, 17.69595703125]",tokens/s,3.5600909041664055,kWh,0.00020889662316689894,0.00011449271197744792,0.0009934635569229903,0.0013168528920673372,tokens/kWh,47841.33473033258,,s,629,179.53831649780273,0.2854345254337086,0.0377383923728263,0.28085964965820315,0.28117113647460934,0.2812770324707031,0.5980694702148438,"[0.2809251708984375, 0.280669189453125, 0.2807091064453125, 0.2809108581542969, 0.28111666870117186, 0.28064971923828125, 0.2810654602050781, 0.28062106323242186, 0.2806558837890625, 0.2807910461425781, 0.28068453979492186, 0.2808770446777344, 0.2806640625, 0.2807613525390625, 0.28071832275390624, 0.28078387451171877, 0.2807152709960937, 0.28067431640625, 0.28070501708984374, 0.2807490539550781, 0.28121600341796876, 0.2806056823730469, 0.28095077514648437, 0.28073471069335937, 0.28082278442382813, 0.2807633972167969, 0.2808299560546875, 0.2809405517578125, 0.28083712768554686, 0.28074700927734375, 0.28070401000976564, 0.28072857666015627, 0.2808084411621094, 0.2823638916015625, 0.2806558837890625, 0.2807715759277344, 0.2807767028808594, 0.2810624084472656, 0.28082278442382813, 0.2810224609375, 0.28100811767578127, 0.2810163269042969, 0.2808545227050781, 0.28095077514648437, 0.2809190368652344, 0.28101324462890626, 0.2807828369140625, 0.2812323913574219, 0.2809671630859375, 0.28106137084960936, 0.2808330383300781, 0.2810439758300781, 0.28064971923828125, 0.28111770629882815, 0.2812231750488281, 0.2810408935546875, 0.28118118286132815, 0.28099072265625, 0.2808832092285156, 0.2812590026855469, 0.2809385070800781, 0.2808832092285156, 0.5983795166015625, 0.28064871215820314, 0.2806937561035156, 0.2806876220703125, 0.2805964660644531, 0.28077362060546873, 0.28084735107421877, 0.2808135681152344, 0.28077978515625, 0.2806302795410156, 0.2807633972167969, 0.2808637390136719, 0.281501708984375, 0.2808135681152344, 0.28127743530273436, 0.2808412170410156, 0.2809886779785156, 0.2807439270019531, 0.2807940979003906, 0.28082382202148437, 0.28080126953125, 0.28092620849609373, 0.2807767028808594, 0.28085964965820315, 0.2809610290527344, 0.2808668212890625, 0.28077978515625, 0.2808545227050781, 0.2811832275390625, 0.28089959716796875, 0.28109515380859373, 0.28077566528320314, 0.2809292907714844, 0.28073983764648436, 0.28083712768554686, 0.2807214050292969, 0.28081561279296874, 0.28083096313476563, 0.28084735107421877, 0.2808258666992188, 0.28108901977539064, 0.28077566528320314, 0.28080230712890625, 0.2808279113769531, 0.28091494750976564, 0.280700927734375, 0.28091494750976564, 0.28119143676757813, 0.28127642822265625, 0.2812569580078125, 0.2810091552734375, 0.2808279113769531, 0.28108697509765623, 0.2808934326171875, 0.28086273193359373, 0.28114431762695313, 0.28093438720703123, 0.28095489501953125, 0.2810040283203125, 0.2808350830078125, 0.28096307373046875, 0.2807848815917969, 0.2808197021484375, 0.5980743408203125, 0.28070401000976564, 0.2806609802246094, 0.28080743408203124, 0.280690673828125, 0.2807357482910156, 0.28078900146484376, 0.28069888305664065, 0.28073883056640625, 0.28077566528320314, 0.280637451171875, 0.2809395141601562, 0.28084225463867185, 0.280669189453125, 0.28070196533203123, 0.28089242553710936, 0.2807613525390625, 0.28102554321289064, 0.280958984375, 0.28090777587890625, 0.28084530639648436, 0.2811146240234375, 0.2807060546875, 0.2808821716308594, 0.28075314331054685, 0.2806824951171875, 0.2807562255859375, 0.2807132263183594, 0.2805975036621094, 0.280627197265625, 0.2807439270019531, 0.280848388671875, 0.2807879638671875, 0.2806599731445312, 0.2806077575683594, 0.28076031494140624, 0.28118426513671874, 0.28088525390625, 0.2811043701171875, 0.281143310546875, 0.2812733459472656, 0.2809036865234375, 0.28103372192382814, 0.28068453979492186, 0.28123956298828123, 0.28111566162109375, 0.28099072265625, 0.28118118286132815, 0.2812323913574219, 0.28117300415039065, 0.2811197509765625, 0.2810224609375, 0.28101937866210935, 0.28089447021484376, 0.28143719482421875, 0.2813931579589844, 0.28145767211914063, 0.28113409423828123, 0.2810245056152344, 0.2813102111816406, 0.2813634643554688, 0.2808913879394531, 0.2811832275390625, 0.5980569458007813, 0.280806396484375, 0.28072549438476563, 0.28071218872070314, 0.2806876220703125, 0.2807807922363281, 0.28076235961914064, 0.2810705871582031, 0.2806241149902344, 0.2806855773925781, 0.28086578369140625, 0.2805698547363281, 0.2807582702636719, 0.28055859375, 0.28054937744140623, 0.28078387451171877, 0.2807562255859375, 0.2807173156738281, 0.28059033203125, 0.2807162780761719, 0.2807500915527344, 0.28070196533203123, 0.2806824951171875, 0.2808340454101563, 0.28079718017578126, 0.28069171142578125, 0.2811125793457031, 0.2808883056640625, 0.28106341552734376, 0.28073471069335937, 0.28064254760742186, 0.2807705688476563, 0.2809989013671875, 0.28072857666015627, 0.28084225463867185, 0.28058624267578125, 0.28073370361328126, 0.28079513549804686, 0.28098458862304687, 0.28078900146484376, 0.2809722900390625, 0.2808965148925781, 0.2811125793457031, 0.28095693969726565, 0.281196533203125, 0.28067636108398436, 0.28092312622070315, 0.2809385070800781, 0.28081561279296874, 0.28086477661132814, 0.2808913879394531, 0.28067431640625, 0.28132455444335935, 0.280922119140625, 0.28108697509765623, 0.2814136352539062, 0.2810091552734375, 0.28069989013671875, 0.280838134765625, 0.28130712890625, 0.2808555603027344, 0.2809016418457031, 0.2813931579589844, 0.5983590698242187, 0.2807306213378906, 0.28087808227539063, 0.2805196838378906, 0.2805125122070313, 0.2804551696777344, 0.2805780334472656, 0.2806200256347656, 0.2805002136230469, 0.28068453979492186, 0.28058010864257815, 0.2806302795410156, 0.2806098022460938, 0.2805196838378906, 0.2809016418457031, 0.28069989013671875, 0.28064971923828125, 0.28055755615234373, 0.28064666748046874, 0.28067019653320313, 0.2806671447753906, 0.28063641357421876, 0.28048590087890624, 0.2806886291503906, 0.2806722412109375, 0.2805166015625, 0.2806087646484375, 0.2805964660644531, 0.28062106323242186, 0.2810009460449219, 0.2807500915527344, 0.2806753234863281, 0.28076235961914064, 0.28126412963867187, 0.28134194946289065, 0.281164794921875, 0.28111666870117186, 0.2810347595214844, 0.28092620849609373, 0.2809241638183594, 0.28083096313476563, 0.28105523681640626, 0.2809354248046875, 0.2808258666992188, 0.281017333984375, 0.2810296325683594, 0.2807910461425781, 0.2810521545410156, 0.28102859497070315, 0.2807940979003906, 0.2811197509765625, 0.28095590209960936, 0.2808842163085937, 0.2810460205078125, 0.2808913879394531, 0.2809292907714844, 0.2807777404785156, 0.28074188232421876, 0.28088525390625, 0.280995849609375, 0.28099172973632813, 0.28075213623046874, 0.28117196655273435, 0.5974200439453125, 0.28051559448242186, 0.28070501708984374, 0.28070501708984374, 0.280669189453125, 0.28124365234375, 0.2808965148925781, 0.2807500915527344, 0.2808545227050781, 0.2804951171875, 0.280748046875, 0.28090264892578126, 0.28080435180664065, 0.2806118469238281, 0.28069580078125, 0.2808616943359375, 0.28078387451171877, 0.2806773681640625, 0.28072039794921877, 0.2808463439941406, 0.28071014404296873, 0.2810408935546875, 0.28070501708984374, 0.28084530639648436, 0.2805821533203125, 0.2807767028808594, 0.2807459716796875, 0.2806599731445312, 0.28074188232421876, 0.28069580078125, 0.2806732788085938, 0.2806507568359375, 0.2808084411621094, 0.2809466857910156, 0.2810572814941406, 0.2805780334472656, 0.28073883056640625, 0.28082891845703123, 0.2810071105957031, 0.28080331420898436, 0.28094976806640626, 0.28130303955078123, 0.28097842407226564, 0.2809968566894531, 0.2808493957519531, 0.2807193603515625, 0.2809456787109375, 0.2811535339355469, 0.2808330383300781, 0.28075418090820314, 0.280853515625, 0.2809169921875, 0.28076953125, 0.280890380859375, 0.2809374694824219, 0.281280517578125, 0.28133172607421875, 0.28101324462890626, 0.28093438720703123, 0.28075314331054685, 0.2808053894042969, 0.28079205322265627, 0.2809333801269531, 0.5982566528320312, 0.2806384582519531, 0.28069989013671875, 0.28062823486328126, 0.2805821533203125, 0.28079205322265627, 0.2808053894042969, 0.28074496459960935, 0.28080743408203124, 0.2807490539550781, 0.2807940979003906, 0.2807173156738281, 0.28145458984375, 0.281248779296875, 0.28092825317382814, 0.2808186950683594, 0.280995849609375, 0.2806609802246094, 0.2807244873046875, 0.2807807922363281, 0.28073779296875, 0.28084326171875, 0.2809169921875, 0.2808883056640625, 0.2809671630859375, 0.2806528015136719, 0.28088525390625, 0.2807490539550781, 0.28113714599609374, 0.2809948120117188, 0.28080331420898436, 0.28099993896484377, 0.28106137084960936, 0.2807500915527344, 0.28100302124023435, 0.280774658203125, 0.28079718017578126, 0.28105319213867186, 0.2808555603027344, 0.2810224609375, 0.2812200927734375, 0.28123544311523435, 0.2811412353515625, 0.28102554321289064, 0.28113201904296875, 0.2809661560058594, 0.280995849609375, 0.2812538757324219, 0.2814095458984375, 0.2813102111816406, 0.28100506591796875, 0.28089447021484376, 0.28107672119140625, 0.28094464111328127, 0.28103884887695313, 0.28113613891601563, 0.2810777587890625, 0.28092312622070315, 0.2810982360839844, 0.280875, 0.2811453552246094, 0.28104193115234377, 0.28097842407226564, 0.6000148315429688, 0.28073370361328126, 0.2808883056640625, 0.28085964965820315, 0.2808688659667969, 0.28098355102539063, 0.280806396484375, 0.28103988647460937, 0.2807173156738281, 0.2808053894042969, 0.28107879638671873, 0.2807080993652344, 0.2809938049316406, 0.28080947875976564, 0.28094259643554687, 0.28093646240234377, 0.2807214050292969, 0.2808412170410156, 0.28081048583984375, 0.2815068054199219, 0.28095590209960936, 0.28094976806640626, 0.28077978515625, 0.2809466857910156, 0.2811985778808594, 0.2808299560546875, 0.28084530639648436, 0.28066201782226563, 0.2807275390625, 0.280890380859375, 0.28104702758789063, 0.28076748657226563, 0.2809395141601562, 0.2808934326171875, 0.2809876403808594, 0.28116683959960936, 0.28092007446289063, 0.2807060546875, 0.2810091552734375, 0.28097024536132814, 0.280890380859375, 0.28086782836914065, 0.28100811767578127, 0.28064358520507815, 0.2809088134765625, 0.2808115234375, 0.2807173156738281, 0.28098150634765623, 0.28097537231445313, 0.2808688659667969, 0.2809374694824219, 0.28094976806640626, 0.2810654602050781, 0.28078387451171877, 0.2811709289550781, 0.2815887451171875, 0.2812333984375, 0.28107879638671873, 0.2810439758300781, 0.281069580078125, 0.28113818359375, 0.28109515380859373, 0.2812569580078125, 0.6002237548828125, 0.280869873046875, 0.2811043701171875, 0.28084225463867185, 0.2808985595703125, 0.28092007446289063, 0.28068658447265626, 0.28079513549804686, 0.28076953125, 0.28095281982421877, 0.2808002624511719, 0.28057089233398436, 0.28077978515625, 0.28081460571289063, 0.2807982177734375, 0.2810357666015625, 0.28085760498046874, 0.28081561279296874, 0.28080947875976564, 0.2812375183105469, 0.28106341552734376, 0.2807726135253906, 0.2809108581542969, 0.2809692077636719, 0.2808115234375, 0.28083712768554686, 0.2811781005859375, 0.28099172973632813, 0.2809948120117188, 0.2808340454101563, 0.28090264892578126, 0.28101324462890626, 0.2809876403808594, 0.2817751159667969, 0.2810347595214844, 0.2808299560546875, 0.281091064453125, 0.28080435180664065, 0.28114227294921873, 0.2809620361328125, 0.2809395141601562, 0.28100811767578127, 0.28102041625976565, 0.2808565673828125, 0.28097024536132814, 0.2810491027832031, 0.281064453125, 0.281091064453125, 0.2807859191894531, 0.2808688659667969, 0.28093646240234377, 0.28071832275390624, 0.2809395141601562, 0.2809599914550781, 0.2808821716308594, 0.28085247802734375, 0.28103372192382814, 0.2807777404785156, 0.2810357666015625, 0.28092108154296874, 0.28099789428710936, 0.28096820068359374, 0.28156927490234374, 0.5991085815429688, 0.280742919921875, 0.2808258666992188, 0.2806661071777344, 0.28075418090820314, 0.2806241149902344, 0.2807306213378906, 0.2808197021484375, 0.2806886291503906, 0.2807244873046875, 0.28067019653320313, 0.2805350341796875, 0.2807224426269531, 0.280869873046875, 0.28091390991210935, 0.2807828369140625, 0.2808197021484375, 0.28051046752929687, 0.2806558837890625, 0.280774658203125, 0.28073165893554686, 0.2806384582519531, 0.280711181640625, 0.28101223754882815, 0.28087091064453124, 0.2806732788085938, 0.28076748657226563, 0.28072653198242187, 0.28075418090820314, 0.28080230712890625, 0.2811269226074219, 0.28098046875, 0.28079513549804686, 0.281059326171875, 0.2811566162109375, 0.2811617431640625, 0.2810163269042969, 0.28065689086914064, 0.2807767028808594, 0.2808002624511719, 0.28075726318359373, 0.28086782836914065, 0.28077566528320314, 0.2806456298828125, 0.28085964965820315, 0.28072039794921877, 0.2807767028808594, 0.2809374694824219, 0.28083096313476563, 0.2811514892578125, 0.2812262268066406, 0.2808821716308594, 0.2809948120117188, 0.2811064453125, 0.28104193115234377, 0.2811760559082031, 0.28112997436523435, 0.28082891845703123, 0.2809241638183594, 0.28104806518554687, 0.2811535339355469, 0.2811269226074219, 0.28101223754882815]",tokens/s,3.5034304223728086,, @@ -11502,7 +12435,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1061-44f2a7c5331d0271539651cf;7431494a-b54e-45d1-89e6-c42d3eda5562) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d42a7-201bd12f6eb7a75a7ab60766;e38774ef-a8bb-41b6-b3a9-219c38692455) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11572,7 +12505,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c14cd-32dc1fea097b1f0e0305f9b3;d0d652f6-d20a-41db-b778-be1421044fa1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46e3-601eb589507c38ac1b38a847;2df42ab0-a98b-4b4a-a9d0-300e698205f6) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11642,7 +12575,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c147f-491adf87747530ac038e71b1;09f9dce5-3983-4a27-9e82-b413e454af13) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4695-5bcd078e11234c5378b66406;f2a9b307-b66f-4271-a200-9db9a181acdf) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11742,7 +12675,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1376-137ce063793d32e53aa6454b;9db7907e-075a-46b5-96f7-23591cd1c21c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4590-33bc598177b3f8c53b396a2b;5bb057c2-8ee2-48b1-bc0c-88a3215cc01f) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11841,7 +12774,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c13ed-69932902638a072633dd7da7;80841a25-bc28-4c5b-af92-a90ea550865b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4605-3020e19f1154ec3a1188dfd0;e91ec65d-9f8d-41b5-be82-221446c928f3) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12050,6 +12983,35 @@ ChildProcessError: Traceback (most recent call last): return forward_call(*args, **kwargs) TypeError: DeciCoderAttention.forward() got an unexpected keyword argument 'cache_position' +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 563, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1478, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1644, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,google/recurrentgemma-2b,google/recurrentgemma-2b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch @@ -12080,7 +13042,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664beeb5-696661ac49b44382768d604c;3d2e96aa-f913-4bb2-9d0a-567d8d53f607) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3fc2-5d332f63391b159b04c9ed40;00176382-abc1-47a6-8811-cf0f709a7bea) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -12285,7 +13247,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c139e-7cb93f8437ef9c297574cc5f;ca4275bb-d00b-4f45-b21d-4559ab5598fd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45b7-3a4e09315ecb0ff14557cfa3;8557c8c6-2efd-436e-a910-1c04101076c5) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12345,7 +13307,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee8d-7a1224ae48847829423ac303;d443d06d-d91d-44d7-be18-a1df2390e5da) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3f9c-35cafe5d1f5d953649f7fc07;2a14995e-0c2e-4b42-8162-3e229f720b20) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -12870,7 +13832,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bed04-0d08b07e715916062b4c30c3;d38cb924-3207-44d5-8ac1-7922d476b270) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3e1b-1b35ab3f1532214d1f39ad7b;2d8c8f28-0894-4f4f-8082-cc668ad1a36d) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -12955,7 +13917,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1327-68ee233600e856ce4025aa6b;46ad2473-dfcb-4dd7-861f-377a339f543e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4544-3605361d062088d158e609a5;7f682eaf-35b5-41ed-8cd7-15e02c140239) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13099,7 +14061,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1224-372afea44c37fd7822bef93f;86036a40-991d-410b-ad99-007ede9dd690) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d447e-07a415bb447be1bf1cad07a1;1e97f6a5-b71b-4e9a-ba2a-763f0cc3a612) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13173,6 +14135,35 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1857.773568,3395.813376,0.0,2749.366272,2585.357312,s,10,0.7326278381347655,0.07326278381347656,0.0010202652967553705,0.07289332580566407,0.0744323989868164,0.07474466476440429,0.0749944773864746,"[0.07307369232177735, 0.07505693054199218, 0.07257308959960937, 0.07189148712158203, 0.07271295928955078, 0.07210118103027344, 0.072621826171875, 0.07393405151367187, 0.07429961395263672, 0.07436300659179687]",tokens/s,3494.2707152892713,kWh,8.526288467250187e-07,4.671987187216898e-07,4.050754639481646e-06,5.370582204928354e-06,tokens/kWh,47667085.28641079,MB,1857.773568,3395.813376,0.0,2749.366272,2628.768768,s,10,38.62795434570313,3.8627954345703124,0.0013787928567573532,3.86274462890625,3.8645461914062498,3.86462001953125,3.86467908203125,"[3.862239990234375, 3.863882568359375, 3.86469384765625, 3.86452978515625, 3.863611328125, 3.862596923828125, 3.862892333984375, 3.8602333984375, 3.86094921875, 3.862324951171875]",tokens/s,16.309432137197284,kWh,4.555938902852252e-05,2.4969072835359344e-05,0.00021447355906873103,0.00028500202093261287,tokens/kWh,221051.06410770328,,s,629,39.2255651473999,0.062361788787599216,0.008780807090293767,0.06126895904541016,0.061400267791748044,0.06155386810302734,0.1344632177734375,"[0.06139596939086914, 0.06134783935546875, 0.06130278396606445, 0.06122700881958008, 0.06143606567382812, 0.061158241271972655, 0.061230079650878906, 0.06156902313232422, 0.06164582443237305, 0.06149529647827148, 0.06133760070800781, 0.06120243072509766, 0.06127718353271484, 0.0615997428894043, 0.0614205436706543, 0.06143385696411133, 0.06148198318481445, 0.061162494659423826, 0.061267967224121096, 0.06151065444946289, 0.06127414321899414, 0.061287391662597654, 0.06125568008422851, 0.06121779251098633, 0.06121574401855469, 0.061246463775634766, 0.06121270370483398, 0.061201377868652346, 0.06132428741455078, 0.061178878784179686, 0.06119833755493164, 0.06125260925292969, 0.06123519897460938, 0.06116454315185547, 0.06141747283935547, 0.06122089767456055, 0.06124540710449219, 0.06128332901000977, 0.061276161193847656, 0.0612044792175293, 0.0612935676574707, 0.06123110580444336, 0.061246463775634766, 0.06124249649047851, 0.06127705764770508, 0.06118608093261719, 0.06124233627319336, 0.06122598266601562, 0.061317119598388675, 0.0612044792175293, 0.0613201904296875, 0.06122393417358398, 0.06131302261352539, 0.06132735824584961, 0.061297664642333986, 0.061230079650878906, 0.061290496826171874, 0.06120345687866211, 0.06123519897460938, 0.06125568008422851, 0.061211647033691405, 0.06121062469482422, 0.13685043334960936, 0.061385726928710936, 0.06136422348022461, 0.061391902923583985, 0.0613969612121582, 0.06137651062011719, 0.06124031829833984, 0.061290496826171874, 0.06117171096801758, 0.06124748611450195, 0.06125056076049805, 0.061244415283203124, 0.06116352081298828, 0.06120243072509766, 0.061238273620605466, 0.061238273620605466, 0.061163585662841795, 0.06130579376220703, 0.061200382232666016, 0.06135603332519531, 0.06155263900756836, 0.061240478515625, 0.061153247833251954, 0.061241214752197265, 0.06119116973876953, 0.06119014358520508, 0.06122102355957031, 0.06126063919067383, 0.06113894271850586, 0.06125363159179688, 0.061257728576660155, 0.06135398483276367, 0.06126182556152344, 0.06128844833374023, 0.06118092727661133, 0.06130790328979492, 0.061292545318603515, 0.06130585479736328, 0.06130585479736328, 0.06157823944091797, 0.0615546875, 0.06145843124389649, 0.061260799407958984, 0.06128742218017578, 0.061306880950927733, 0.06136832046508789, 0.06350233459472657, 0.061344768524169924, 0.061143039703369144, 0.061300830841064455, 0.061209503173828124, 0.06128025436401367, 0.061257728576660155, 0.06132035064697266, 0.06122991943359375, 0.06137548828125, 0.061472766876220705, 0.061297664642333986, 0.06125673675537109, 0.06129660797119141, 0.0612935676574707, 0.06125158309936524, 0.06124448013305664, 0.134383544921875, 0.06131916809082031, 0.0611409912109375, 0.0613488655090332, 0.06120044708251953, 0.06135084915161133, 0.06120755386352539, 0.06130380630493164, 0.0612751350402832, 0.06122905731201172, 0.06122700881958008, 0.06135295867919922, 0.061241344451904295, 0.061300735473632816, 0.06121577453613281, 0.061286369323730466, 0.0612751350402832, 0.06132940673828125, 0.06118809509277344, 0.06131814575195312, 0.06130790328979492, 0.0613488655090332, 0.061249534606933595, 0.06230227279663086, 0.06123923110961914, 0.06123929595947265, 0.06117478561401367, 0.061221889495849606, 0.06119833755493164, 0.061257728576660155, 0.061176830291748044, 0.06125158309936524, 0.061233150482177735, 0.06127718353271484, 0.061344768524169924, 0.06139494323730469, 0.06125574493408203, 0.061348800659179685, 0.06131097412109375, 0.061314048767089846, 0.061292545318603515, 0.06137855911254883, 0.06121267318725586, 0.06131609725952149, 0.0612935676574707, 0.0612977294921875, 0.06115423965454102, 0.062395423889160154, 0.06123737716674805, 0.06142038345336914, 0.06166016006469727, 0.06140723037719727, 0.061308929443359375, 0.06141439819335937, 0.061271041870117185, 0.061306880950927733, 0.06125568008422851, 0.061437950134277344, 0.061267967224121096, 0.06134067153930664, 0.06126396942138672, 0.061334430694580076, 0.06136627197265625, 0.1344952697753906, 0.061371360778808594, 0.061200382232666016, 0.06137036895751953, 0.06131916809082031, 0.061306880950927733, 0.06132940673828125, 0.06137651062011719, 0.06126182556152344, 0.06136217498779297, 0.06131097412109375, 0.06130995178222656, 0.06125056076049805, 0.061431808471679686, 0.06132223892211914, 0.06132940673828125, 0.061211647033691405, 0.0613570556640625, 0.06121984100341797, 0.061379585266113285, 0.061233150482177735, 0.06123724746704102, 0.061211647033691405, 0.061295616149902345, 0.06129151916503906, 0.06138675308227539, 0.06125260925292969, 0.06131302261352539, 0.061254657745361325, 0.06133452987670898, 0.06119116973876953, 0.061399040222167967, 0.06132940673828125, 0.06128025436401367, 0.06132223892211914, 0.06143590545654297, 0.06123519897460938, 0.06140108871459961, 0.061257728576660155, 0.06137753677368164, 0.061502464294433595, 0.06135910415649414, 0.061244415283203124, 0.06137753677368164, 0.061966335296630856, 0.06137651062011719, 0.06124748611450195, 0.06135910415649414, 0.06123724746704102, 0.061400062561035154, 0.06116761779785156, 0.06129983901977539, 0.06128524780273437, 0.061574142456054685, 0.06160192108154297, 0.06146547317504883, 0.061254657745361325, 0.0613939208984375, 0.06138777542114258, 0.061290496826171874, 0.06130380630493164, 0.061260799407958984, 0.0612782096862793, 0.13449420166015624, 0.0613201904296875, 0.06128128051757813, 0.06128128051757813, 0.06118608093261719, 0.06126895904541016, 0.061241344451904295, 0.061216766357421876, 0.06117171096801758, 0.061315071105957034, 0.06134272003173828, 0.06174822235107422, 0.06120140838623047, 0.061178878784179686, 0.061195297241210936, 0.0614153938293457, 0.06117068862915039, 0.0621393928527832, 0.06123110580444336, 0.06123519897460938, 0.0611860466003418, 0.061274112701416014, 0.06114713668823242, 0.06122716903686523, 0.061267807006835935, 0.061276161193847656, 0.06118297576904297, 0.06133555221557617, 0.06133452987670898, 0.061262847900390625, 0.061224960327148435, 0.06130278396606445, 0.06118092727661133, 0.06130694580078125, 0.06124435043334961, 0.06132940673828125, 0.061295616149902345, 0.06131513595581055, 0.06114809417724609, 0.061308929443359375, 0.06128947067260742, 0.06133350372314453, 0.06126387023925781, 0.06140825653076172, 0.061385726928710936, 0.06133248138427734, 0.06132121658325195, 0.061254657745361325, 0.06115532684326172, 0.06133964920043945, 0.061189121246337894, 0.061254657745361325, 0.061254657745361325, 0.06132940673828125, 0.061244415283203124, 0.06130278396606445, 0.06126387023925781, 0.06329753494262695, 0.06128639984130859, 0.061312000274658204, 0.06132735824584961, 0.06121062469482422, 0.061227104187011716, 0.13391145324707032, 0.06119424057006836, 0.061162494659423826, 0.06187417602539062, 0.06159462356567383, 0.061346817016601565, 0.061265918731689455, 0.061241344451904295, 0.061262847900390625, 0.06127718353271484, 0.06129663848876953, 0.06137446212768555, 0.061271041870117185, 0.06124236679077148, 0.061238273620605466, 0.061267967224121096, 0.06116556930541992, 0.061265918731689455, 0.06117478561401367, 0.06121779251098633, 0.06124544143676758, 0.06132735824584961, 0.061238273620605466, 0.061195392608642575, 0.06117772674560547, 0.06121984100341797, 0.06117478561401367, 0.061249534606933595, 0.06135398483276367, 0.06128025436401367, 0.061143039703369144, 0.061276161193847656, 0.061193214416503904, 0.06121984100341797, 0.061216766357421876, 0.06132428741455078, 0.06124544143676758, 0.061282302856445314, 0.06337228775024414, 0.06132940673828125, 0.061230079650878906, 0.06132326507568359, 0.06124249649047851, 0.06130777740478516, 0.06127836990356445, 0.061315937042236326, 0.06120755386352539, 0.061228031158447264, 0.061224960327148435, 0.061297664642333986, 0.061158401489257816, 0.06132428741455078, 0.061193214416503904, 0.06124560165405273, 0.06118076705932617, 0.06130176162719726, 0.06120550537109375, 0.06126694488525391, 0.06125363159179688, 0.061369342803955076, 0.061314048767089846, 0.06129971313476563, 0.06124236679077148, 0.13450445556640625, 0.061317119598388675, 0.061284351348876956, 0.06129971313476563, 0.06124860763549805, 0.06139587020874023, 0.06129663848876953, 0.06127628707885742, 0.061279102325439455, 0.06135603332519531, 0.0612147216796875, 0.061383678436279294, 0.06137241744995117, 0.06130483245849609, 0.061230079650878906, 0.06132326507568359, 0.061260799407958984, 0.061249534606933595, 0.061295616149902345, 0.06130278396606445, 0.06117171096801758, 0.061284351348876956, 0.061284351348876956, 0.0623636474609375, 0.061454334259033204, 0.061413375854492185, 0.061369342803955076, 0.06145024108886719, 0.061213695526123046, 0.061308929443359375, 0.06139801788330078, 0.06125568008422851, 0.06120243072509766, 0.06139596939086914, 0.06120652770996094, 0.061292545318603515, 0.061486080169677736, 0.061244415283203124, 0.06117478561401367, 0.061338623046875, 0.061300735473632816, 0.06128025436401367, 0.061260799407958984, 0.061298686981201174, 0.06127308654785156, 0.06131302261352539, 0.061295616149902345, 0.061325313568115235, 0.061228031158447264, 0.06130380630493164, 0.06117174530029297, 0.06120867156982422, 0.06118182373046875, 0.061200382232666016, 0.06121267318725586, 0.061249534606933595, 0.06119731140136719, 0.061259777069091796, 0.06121779251098633, 0.0612853775024414, 0.06123519897460938, 0.061363201141357425, 0.06137651062011719, 0.13581520080566406, 0.06123721694946289, 0.061230079650878906, 0.06120243072509766, 0.06118809509277344, 0.06125888061523437, 0.06128627014160156, 0.061183998107910156, 0.06116966247558594, 0.061292545318603515, 0.06119116973876953, 0.06121267318725586, 0.06119424057006836, 0.0612567024230957, 0.06120345687866211, 0.0611778564453125, 0.062005279541015625, 0.061244384765625, 0.06121267318725586, 0.06127001571655273, 0.061156352996826174, 0.06120140838623047, 0.06130585479736328, 0.06121984100341797, 0.06111231994628906, 0.0612567024230957, 0.06117068862915039, 0.061274112701416014, 0.06124031829833984, 0.0612567024230957, 0.06120959854125976, 0.06128332901000977, 0.06125568008422851, 0.06122393417358398, 0.061233150482177735, 0.0613089599609375, 0.06115020751953125, 0.06122288131713867, 0.06142156982421875, 0.06144102478027344, 0.061252639770507815, 0.06141129684448242, 0.061379585266113285, 0.061317119598388675, 0.06144524765014649, 0.06125657653808594, 0.061233150482177735, 0.061290496826171874, 0.0611778564453125, 0.06124851226806641, 0.06120755386352539, 0.061276161193847656, 0.06118809509277344, 0.06129151916503906, 0.06121267318725586, 0.06132326507568359, 0.061453311920166016, 0.06124748611450195, 0.06111743927001953, 0.06126387023925781, 0.06130790328979492, 0.061262847900390625, 0.06124748611450195, 0.13622784423828124, 0.06128128051757813, 0.06116864013671875, 0.06118195343017578, 0.06121881484985352, 0.06123212814331055, 0.061211742401123044, 0.061231006622314454, 0.06120550537109375, 0.061178878784179686, 0.061222911834716794, 0.061193214416503904, 0.06135910415649414, 0.061623294830322264, 0.061295616149902345, 0.06120652770996094, 0.06120761489868164, 0.06214854431152344, 0.061200382232666016, 0.06128128051757813, 0.06116147232055664, 0.06121062469482422, 0.06120755386352539, 0.06122393417358398, 0.06119424057006836, 0.06125475311279297, 0.061219745635986325, 0.06122598266601562, 0.061143039703369144, 0.06130585479736328, 0.06125363159179688, 0.06124748611450195, 0.06132735824584961, 0.06121881484985352, 0.06116966247558594, 0.061338623046875, 0.0612567024230957, 0.061250625610351564, 0.061302753448486326, 0.061341663360595704, 0.061221889495849606, 0.06131820678710938, 0.061247425079345705, 0.061279232025146485, 0.061300735473632816, 0.06132735824584961, 0.0612935676574707, 0.061274112701416014, 0.06122393417358398, 0.06123622512817383, 0.061162494659423826, 0.06129663848876953, 0.06119424057006836, 0.061271041870117185, 0.061315071105957034, 0.06127001571655273, 0.061238273620605466, 0.06128128051757813, 0.06128844833374023, 0.061314048767089846, 0.06154751968383789, 0.06133248138427734, 0.061330432891845706, 0.13624729919433592, 0.061328384399414064, 0.061246463775634766, 0.06124031829833984, 0.061189121246337894, 0.061243392944335937, 0.061193214416503904, 0.06124851226806641, 0.061246463775634766, 0.06123219299316406, 0.061181888580322266, 0.06125158309936524, 0.06116761779785156, 0.06128742218017578, 0.061243392944335937, 0.06128639984130859, 0.06118406295776367, 0.06136620712280273, 0.06136115264892578, 0.06126396942138672, 0.061195167541503906, 0.0612751350402832, 0.06120243072509766, 0.06128332901000977, 0.06121779251098633, 0.061230079650878906, 0.0632248306274414, 0.06133350372314453, 0.06123212814331055, 0.06128639984130859, 0.061257728576660155, 0.06139596939086914, 0.061276161193847656, 0.06122700881958008, 0.06123519897460938, 0.06130380630493164, 0.06121881484985352, 0.06122905731201172, 0.06126387023925781, 0.06125568008422851, 0.061244415283203124, 0.06136627197265625, 0.06116147232055664, 0.061262847900390625, 0.06133145523071289, 0.06130294418334961, 0.0612022705078125, 0.06134783935546875, 0.06124236679077148, 0.06139801788330078, 0.06120969772338867, 0.061249439239501956, 0.06118204879760742, 0.06134364700317383, 0.06126489639282227, 0.06138374328613281, 0.06129779052734375, 0.06130259323120117, 0.06122086334228516, 0.06133555221557617, 0.061274112701416014, 0.061267967224121096, 0.061358081817626954]",tokens/s,16.035460486964933,, +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1478, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1644, in _check_and_enable_sdpa + raise ValueError( +ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -13242,7 +14233,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1465-7307e46056aa8f0657c7813c;3a765831-0842-411c-8cee-d1d8a4fe5ff5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d467a-1eb50b0f339c76d03ab9bb90;72474fca-b4bf-4d03-8c00-24fcbc912629) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13341,7 +14332,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c11f8-385beaa0288014512ff1d19e;22b287b0-69de-4be7-9ef0-790400e39dcd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4457-2215852a2fafb3fb1972dfdc;c7aef44c-732d-4b7c-8366-a90a129357f4) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13440,7 +14431,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12c8-19b17d1229fe45a52c46d30d;a5f2b29c-da6c-436c-bf83-70efffb9be19) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44f5-109b6a1f6754d83d2f0a0cc7;c714a862-630f-484e-a24d-fcc240ba898d) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13472,7 +14463,111 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5005.856768,8173.125632,0.0,7526.678528,6833.613824,s,10,3.237135192871093,0.32371351928710934,0.0011265830391930785,0.32340528869628904,0.3252778289794922,0.32533260955810545,0.3253764340209961,"[0.3226540832519531, 0.32538739013671875, 0.3225751647949219, 0.32281320190429685, 0.323228515625, 0.3222604675292969, 0.32358206176757814, 0.32464373779296873, 0.32526565551757813, 0.32472491455078123]",tokens/s,790.8227020106239,kWh,3.805233683023188e-06,2.085111826499997e-06,1.9868748533876518e-05,2.57590940433997e-05,tokens/kWh,9938237.717859311,MB,5005.856768,8173.125632,0.0,7526.678528,7110.583808,s,10,176.72496484375003,17.672496484375003,0.0017080130721937633,17.6719814453125,17.674641601562502,17.67482177734375,17.674965917968752,"[17.67114453125, 17.670958984375, 17.673966796875, 17.671052734375, 17.6746015625, 17.675001953125, 17.67048828125, 17.67075, 17.672818359375, 17.674181640625]",tokens/s,3.5648613683811425,kWh,0.00020860287739584842,0.00011433028744598232,0.00106939106731773,0.0013923242321595608,tokens/kWh,45248.081262138214,,s,629,179.35808807373044,0.2851479937579181,0.038540759752562596,0.2804643859863281,0.2806615966796875,0.280817041015625,0.6038926318359376,"[0.2805963134765625, 0.28048486328125, 0.28035687255859376, 0.28048077392578125, 0.2802708435058594, 0.28063641357421876, 0.280880126953125, 0.2805626831054687, 0.28056576538085937, 0.28044699096679687, 0.28062106323242186, 0.2806651000976563, 0.2804029541015625, 0.28052276611328125, 0.28037017822265625, 0.28041830444335936, 0.28037835693359375, 0.28043264770507814, 0.2805278625488281, 0.28084326171875, 0.280748046875, 0.28032305908203126, 0.2804572143554688, 0.2806138916015625, 0.2806814575195313, 0.2805565490722656, 0.28059237670898435, 0.28052276611328125, 0.2805063781738281, 0.2803814697265625, 0.28054525756835935, 0.2805565490722656, 0.28050546264648435, 0.280545166015625, 0.2803732604980469, 0.2802689514160156, 0.2804836730957031, 0.28037631225585935, 0.28042648315429686, 0.2804510803222656, 0.28022171020507813, 0.28037530517578124, 0.2806087646484375, 0.2803619689941406, 0.2803292236328125, 0.28030975341796877, 0.28040704345703127, 0.2805073852539062, 0.2803589172363281, 0.2804869079589844, 0.28044699096679687, 0.2804418640136719, 0.28030670166015625, 0.28054425048828124, 0.2804132080078125, 0.2802564697265625, 0.28043878173828124, 0.28012237548828123, 0.2805473327636719, 0.28069381713867186, 0.28056365966796876, 0.28044390869140623, 0.606454833984375, 0.28040594482421877, 0.28052581787109376, 0.2802616271972656, 0.28048382568359376, 0.2802401428222656, 0.28023602294921873, 0.28026571655273436, 0.2802964477539063, 0.28036309814453125, 0.28032196044921875, 0.28032000732421875, 0.28035482788085936, 0.28039883422851564, 0.2804223937988281, 0.2805524597167969, 0.28039678955078123, 0.280553466796875, 0.28060159301757814, 0.28041012573242186, 0.2802832336425781, 0.2802912292480469, 0.2802391052246094, 0.2806220703125, 0.2804019165039062, 0.28050433349609377, 0.28030157470703126, 0.2803660888671875, 0.2802759704589844, 0.28037631225585935, 0.28027188110351564, 0.2805688171386719, 0.2806599731445312, 0.280518798828125, 0.28085235595703123, 0.2807992248535156, 0.2804234313964844, 0.28058111572265626, 0.2803384399414062, 0.2803363952636719, 0.2804286193847656, 0.28044281005859373, 0.2804090881347656, 0.2804664306640625, 0.2803998718261719, 0.2803660888671875, 0.2832220153808594, 0.2804951171875, 0.28065899658203125, 0.28072747802734377, 0.2807408752441406, 0.280563720703125, 0.2804142150878906, 0.28033535766601564, 0.28055755615234373, 0.280489990234375, 0.2803722229003906, 0.28031402587890625, 0.28017034912109373, 0.2805073852539062, 0.28048895263671875, 0.2806773681640625, 0.2802821044921875, 0.6037001953125, 0.280263671875, 0.28055755615234373, 0.2806200256347656, 0.28048486328125, 0.28029132080078123, 0.2803927001953125, 0.28048077392578125, 0.28041522216796877, 0.2804715576171875, 0.28054425048828124, 0.2804725646972656, 0.28047360229492185, 0.2803722229003906, 0.280458251953125, 0.2804316101074219, 0.28048895263671875, 0.2804490356445313, 0.28044699096679687, 0.28046234130859377, 0.28052377319335936, 0.280511474609375, 0.28041830444335936, 0.28209765625, 0.2803117980957031, 0.2805626831054687, 0.2804756774902344, 0.28051861572265624, 0.28057907104492186, 0.2806241149902344, 0.2804572143554688, 0.28048281860351565, 0.2805565490722656, 0.28042245483398437, 0.28039572143554686, 0.28050433349609377, 0.2804633483886719, 0.2805053405761719, 0.28054629516601565, 0.2807029724121094, 0.2806098022460938, 0.2805504150390625, 0.2805821533203125, 0.28057089233398436, 0.28033444213867187, 0.28032501220703127, 0.28029541015625, 0.2819450988769531, 0.2804808349609375, 0.28046636962890625, 0.28050225830078124, 0.2803804016113281, 0.2805422058105469, 0.28032211303710936, 0.28040594482421877, 0.28035174560546877, 0.28062924194335936, 0.28064666748046874, 0.28025140380859376, 0.2804951171875, 0.2808965148925781, 0.2807439270019531, 0.2803025817871094, 0.6039674682617188, 0.2802984924316406, 0.2803036193847656, 0.2803732604980469, 0.2804992065429687, 0.2802586059570312, 0.28033941650390626, 0.2803660888671875, 0.28037530517578124, 0.28076544189453123, 0.2807685241699219, 0.2807357482910156, 0.2816512145996094, 0.2805882873535156, 0.28059341430664064, 0.28039678955078123, 0.2803189697265625, 0.280300537109375, 0.2803343505859375, 0.28028619384765624, 0.28039883422851564, 0.2803978271484375, 0.28032101440429685, 0.2804572143554688, 0.2802882690429688, 0.280342529296875, 0.2803343505859375, 0.28048794555664064, 0.2802626647949219, 0.28064364624023436, 0.2803271179199219, 0.2806200256347656, 0.28050433349609377, 0.280521728515625, 0.2802626647949219, 0.2803947448730469, 0.28026776123046876, 0.2804962463378906, 0.28034445190429685, 0.2803445739746094, 0.2804142150878906, 0.2802759704589844, 0.28033740234375, 0.28036300659179686, 0.2823341979980469, 0.2804653930664063, 0.28042547607421875, 0.2806005859375, 0.28057089233398436, 0.280553466796875, 0.28051046752929687, 0.28042752075195315, 0.28049099731445315, 0.28036505126953126, 0.2804695129394531, 0.28058724975585936, 0.28055960083007814, 0.28043777465820313, 0.2803456115722656, 0.28051864624023437, 0.28044082641601564, 0.2805299072265625, 0.28034048461914063, 0.6044927978515625, 0.28044390869140623, 0.28047564697265626, 0.2803916931152344, 0.2804715576171875, 0.2804223937988281, 0.28053607177734374, 0.2804244384765625, 0.28046029663085936, 0.2805473327636719, 0.28041830444335936, 0.2804009094238281, 0.28044287109375, 0.28053707885742185, 0.2805196838378906, 0.28043060302734374, 0.2806947937011719, 0.28244479370117187, 0.2806507568359375, 0.2805596923828125, 0.2804827575683594, 0.2806190185546875, 0.2803179626464844, 0.280352783203125, 0.2803558349609375, 0.28041830444335936, 0.28029541015625, 0.2804592590332031, 0.2803466186523437, 0.2804244384765625, 0.2802903137207031, 0.28051763916015626, 0.28042752075195315, 0.28032205200195315, 0.28036300659179686, 0.2805196838378906, 0.28048178100585935, 0.28048178100585935, 0.2802964477539063, 0.28041317749023437, 0.2806640625, 0.2805135498046875, 0.28036300659179686, 0.28051776123046873, 0.2803956604003906, 0.28043777465820313, 0.2803589172363281, 0.28044082641601564, 0.28042547607421875, 0.28058111572265626, 0.2805688171386719, 0.2803609619140625, 0.2805853576660156, 0.28058303833007814, 0.28053094482421875, 0.2804971618652344, 0.28056781005859377, 0.28342782592773436, 0.2802606201171875, 0.280511474609375, 0.28044699096679687, 0.28046234130859377, 0.28032101440429685, 0.6034566040039062, 0.2803240051269531, 0.28044287109375, 0.28026470947265625, 0.28046646118164065, 0.2802431640625, 0.2804776916503906, 0.28053912353515625, 0.280553466796875, 0.2805084228515625, 0.28041317749023437, 0.280406005859375, 0.28044699096679687, 0.2805145568847656, 0.28044000244140627, 0.2804590759277344, 0.2803117980957031, 0.28043060302734374, 0.2803947448730469, 0.28041317749023437, 0.2804346923828125, 0.28048281860351565, 0.2803292236328125, 0.28043060302734374, 0.2803916931152344, 0.28090573120117185, 0.281034912109375, 0.2807693786621094, 0.2806179809570313, 0.2805002136230469, 0.2804552001953125, 0.28051556396484373, 0.2805329895019531, 0.28052890014648435, 0.2803896179199219, 0.2805248107910156, 0.28049612426757814, 0.28070196533203123, 0.2834462585449219, 0.28048492431640626, 0.28036190795898436, 0.28037118530273436, 0.2805299072265625, 0.28084225463867185, 0.28050128173828126, 0.2805862121582031, 0.2804244384765625, 0.2804561767578125, 0.2806599731445312, 0.28053607177734374, 0.28061489868164063, 0.28050335693359374, 0.28050936889648437, 0.2803722229003906, 0.2803978271484375, 0.28043777465820313, 0.2805206909179688, 0.2805084228515625, 0.2803292236328125, 0.28083203125, 0.28075106811523437, 0.2806098022460938, 0.2802431945800781, 0.6047293701171875, 0.2803046264648438, 0.28038552856445315, 0.2803200378417969, 0.28035171508789064, 0.280205322265625, 0.28043878173828124, 0.280405029296875, 0.2803721923828125, 0.2802923583984375, 0.2804090881347656, 0.2803650817871094, 0.2803547973632812, 0.2803897705078125, 0.28044378662109376, 0.28044082641601564, 0.28047360229492185, 0.28044699096679687, 0.28070501708984374, 0.2806456298828125, 0.280531982421875, 0.2805053405761719, 0.2803169250488281, 0.2821160888671875, 0.2803343505859375, 0.2804653930664063, 0.2802472839355469, 0.28062310791015627, 0.2803384399414062, 0.28051559448242186, 0.28032513427734373, 0.2805145568847656, 0.28086477661132814, 0.2805626831054687, 0.28067840576171876, 0.28092825317382814, 0.28034866333007813, 0.280458251953125, 0.28032205200195315, 0.28049615478515627, 0.28045614624023435, 0.2804019165039062, 0.28041317749023437, 0.2805063781738281, 0.28031900024414064, 0.2803783264160156, 0.2802769775390625, 0.2803916931152344, 0.2805688171386719, 0.2805893249511719, 0.2807664489746094, 0.28051864624023437, 0.280453125, 0.28029440307617187, 0.28048794555664064, 0.2804981689453125, 0.28035687255859376, 0.28037530517578124, 0.2801438598632813, 0.280474609375, 0.2804223937988281, 0.28052276611328125, 0.2804583129882813, 0.605824951171875, 0.28028619384765624, 0.2806384582519531, 0.2806241149902344, 0.2806609802246094, 0.280263671875, 0.2804490356445313, 0.2804356994628906, 0.2804572143554688, 0.2802821044921875, 0.28043264770507814, 0.2805053405761719, 0.28048794555664064, 0.28051864624023437, 0.280521728515625, 0.28047052001953127, 0.28164608764648436, 0.2804326782226563, 0.2803670654296875, 0.2803937377929687, 0.2804633483886719, 0.2804572143554688, 0.2804356994628906, 0.2803651123046875, 0.2801980895996094, 0.2803189697265625, 0.28040606689453124, 0.2805523986816406, 0.28038861083984373, 0.2804223937988281, 0.28030670166015625, 0.28060159301757814, 0.28061489868164063, 0.28049099731445315, 0.2803456115722656, 0.2804869079589844, 0.2803640441894531, 0.28059854125976563, 0.280479736328125, 0.28065484619140624, 0.2805299072265625, 0.2803129577636719, 0.28043557739257813, 0.2806098022460938, 0.28044595336914063, 0.28049612426757814, 0.28042752075195315, 0.28031488037109376, 0.2805555114746094, 0.28026571655273436, 0.28036813354492185, 0.28044287109375, 0.280352783203125, 0.28028005981445314, 0.28044287109375, 0.28049612426757814, 0.2805616760253906, 0.28052276611328125, 0.280416259765625, 0.28058624267578125, 0.28044082641601564, 0.28082891845703123, 0.28066815185546873, 0.6064496459960937, 0.28037631225585935, 0.28064358520507815, 0.2803916931152344, 0.28039883422851564, 0.28029336547851563, 0.2804223937988281, 0.2804695129394531, 0.28044287109375, 0.2806394958496094, 0.28072970581054685, 0.28055029296875, 0.28057598876953127, 0.2806005859375, 0.28058624267578125, 0.280553466796875, 0.2804592590332031, 0.2818191223144531, 0.2805145568847656, 0.28053912353515625, 0.28071633911132815, 0.2805329895019531, 0.28040093994140625, 0.2804130859375, 0.28032000732421875, 0.2806302795410156, 0.28032205200195315, 0.2805473327636719, 0.28040499877929687, 0.2805749816894531, 0.2803865661621094, 0.2806026306152344, 0.28049099731445315, 0.2805350341796875, 0.2804643859863281, 0.2805053405761719, 0.28049099731445315, 0.2805206909179688, 0.28050125122070313, 0.28054937744140623, 0.2804787292480469, 0.28038861083984373, 0.28054425048828124, 0.2805473327636719, 0.280352783203125, 0.28046847534179686, 0.2804111328125, 0.28038348388671874, 0.2803814392089844, 0.2805504150390625, 0.28066815185546873, 0.2804643859863281, 0.28052581787109376, 0.28044595336914063, 0.2804920349121094, 0.2805177001953125, 0.2804632873535156, 0.280416259765625, 0.28042034912109376, 0.2804696044921875, 0.2805472412109375, 0.2804971618652344, 0.28035687255859376, 0.6061209716796875, 0.28034356689453127, 0.28048382568359376, 0.2802329711914063, 0.28037835693359375, 0.28030975341796877, 0.2804234313964844, 0.2804244384765625, 0.2806826477050781, 0.2807540283203125, 0.28043264770507814, 0.280531982421875, 0.2804561767578125, 0.28045208740234373, 0.2804736633300781, 0.2804694519042969, 0.28052276611328125, 0.2806302795410156, 0.2805452880859375, 0.2804142150878906, 0.28073370361328126, 0.2807664489746094, 0.28032818603515625, 0.280453125, 0.2803537902832031, 0.28041934204101565, 0.28231680297851564, 0.2805032958984375, 0.2803619689941406, 0.280521728515625, 0.2803240966796875, 0.28040704345703127, 0.2804725646972656, 0.2804316101074219, 0.2803804016113281, 0.2805718994140625, 0.2806169738769531, 0.28048281860351565, 0.2804500427246094, 0.28053912353515625, 0.28075314331054685, 0.2806179809570313, 0.2803937377929687, 0.28062310791015627, 0.2805350341796875, 0.2806814575195313, 0.2804490356445313, 0.2805248107910156, 0.2805606384277344, 0.2803978271484375, 0.28048281860351565, 0.28046234130859377, 0.2805616760253906, 0.28059033203125, 0.28056781005859377, 0.28047564697265626, 0.28059033203125, 0.28059442138671875, 0.28028619384765624, 0.28051763916015626, 0.28055960083007814, 0.28064358520507815, 0.28044390869140623]",tokens/s,3.5069508532084197,, +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1478, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1644, in _check_and_enable_sdpa + raise ValueError( +ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4660.158464,21498.42944,0.0,20851.982336,20236.60032,s,10,15.67305651855469,1.5673056518554689,0.002935213598468352,1.5664246826171875,1.5689432861328125,1.5722443481445312,1.5748851977539062,"[1.5660777587890624, 1.5652294921875, 1.5653187255859375, 1.56764404296875, 1.5667716064453125, 1.565794677734375, 1.568209716796875, 1.5673927001953125, 1.57554541015625, 1.5650723876953125]",tokens/s,163.3376359594774,kWh,1.8488901803890867e-05,1.0131977631626796e-05,9.191171241820228e-05,0.00012053259185371994,tokens/kWh,2123906.8708543596,MB,4660.158464,21498.42944,0.0,20851.982336,20339.706368,s,10,881.3746874999999,88.13746875000001,0.00269603704029487,88.13719531250001,88.13986015625,88.141726953125,88.14322039062499,"[88.1357109375, 88.1359765625, 88.13715625, 88.137234375, 88.1394453125, 88.1378203125, 88.1392890625, 88.1333984375, 88.14359375, 88.1350625]",tokens/s,0.7147924814893211,kWh,0.001040502367599143,0.0005702873234177241,0.005180215005279792,0.006791004696296659,tokens/kWh,9276.97782838463,,s,629,894.0823742675777,1.4214346172775487,0.18625247039114828,1.3989376220703125,1.3992808349609374,1.3994753662109376,2.9653164257812503,"[1.39905224609375, 1.3987666015625, 1.3988515625, 1.3986263427734376, 1.398855712890625, 1.3987164306640625, 1.3989693603515625, 1.39875537109375, 1.3986160888671875, 1.3986785888671875, 1.3988690185546875, 1.39879833984375, 1.398836181640625, 1.3988291015625, 1.398834228515625, 1.39947412109375, 1.3986580810546876, 1.398688720703125, 1.398931396484375, 1.3984696044921876, 1.398876220703125, 1.398803466796875, 1.399109619140625, 1.3990972900390626, 1.3988321533203125, 1.3986181640625, 1.398677490234375, 1.3991065673828125, 1.3987225341796874, 1.39873486328125, 1.398603759765625, 1.3990164794921875, 1.398830078125, 1.3991229248046875, 1.3992918701171875, 1.4000015869140625, 1.3992899169921875, 1.3990687255859375, 1.3990062255859375, 1.399046142578125, 1.3990379638671875, 1.3989140625, 1.3989693603515625, 1.398824951171875, 1.3988116455078126, 1.3989478759765626, 1.3991649169921876, 1.3989530029296875, 1.399182373046875, 1.3990277099609374, 1.399468017578125, 1.399045166015625, 1.3990164794921875, 1.3987860107421874, 1.399066650390625, 1.3993809814453124, 1.3991322021484376, 1.399140380859375, 1.3991536865234375, 1.3990225830078125, 1.3993870849609376, 1.3994906005859375, 2.964931640625, 1.3986201171875, 1.3987666015625, 1.39865087890625, 1.3985853271484374, 1.398708251953125, 1.39879833984375, 1.398951904296875, 1.398729736328125, 1.399015380859375, 1.3991639404296874, 1.39902978515625, 1.3989161376953125, 1.3987215576171874, 1.3987850341796875, 1.3988013916015625, 1.3988004150390625, 1.3990830078125, 1.3986212158203124, 1.3988321533203125, 1.398856689453125, 1.3988372802734375, 1.398662109375, 1.3987337646484375, 1.398878173828125, 1.39877783203125, 1.398846435546875, 1.398877197265625, 1.4004705810546876, 1.3986488037109375, 1.399041015625, 1.398792236328125, 1.3986795654296875, 1.398814697265625, 1.3986632080078125, 1.398877197265625, 1.399088134765625, 1.398846435546875, 1.398697998046875, 1.398998046875, 1.398794189453125, 1.3989376220703125, 1.3990277099609374, 1.398767578125, 1.398866943359375, 1.398981689453125, 1.4009364013671874, 1.39877587890625, 1.399131103515625, 1.3991177978515625, 1.3994281005859375, 1.399003173828125, 1.398744140625, 1.3990062255859375, 1.398962158203125, 1.3992396240234375, 1.3992283935546874, 1.3991383056640625, 1.3989990234375, 1.39947314453125, 1.399541748046875, 1.3994761962890625, 1.39911474609375, 2.96489990234375, 1.3986221923828126, 1.3986396484375, 1.3987462158203126, 1.398728759765625, 1.3984818115234374, 1.3987225341796874, 1.3987666015625, 1.3989918212890624, 1.3991875, 1.398984619140625, 1.399055419921875, 1.39894482421875, 1.3992437744140624, 1.3988853759765625, 1.3988741455078124, 1.3987921142578126, 1.3989744873046874, 1.4005791015625, 1.3989959716796876, 1.3986417236328126, 1.39873583984375, 1.3984593505859375, 1.3988751220703124, 1.398887451171875, 1.3985894775390626, 1.3989530029296875, 1.3985628662109375, 1.398551513671875, 1.398708251953125, 1.3987327880859375, 1.3988291015625, 1.3988648681640625, 1.3988126220703125, 1.3988331298828125, 1.3991240234375, 1.3989898681640625, 1.3990860595703125, 1.3987808837890625, 1.3990594482421874, 1.3988106689453126, 1.398973388671875, 1.401247802734375, 1.398792236328125, 1.3992130126953124, 1.399266357421875, 1.3992427978515625, 1.399146484375, 1.398971435546875, 1.398993896484375, 1.3989744873046874, 1.3990277099609374, 1.3990123291015626, 1.3992764892578125, 1.39936767578125, 1.39896826171875, 1.399108642578125, 1.3990491943359376, 1.399204833984375, 1.3993646240234374, 1.399277587890625, 1.399076904296875, 1.39888330078125, 2.967499755859375, 1.3984256591796875, 1.3987880859375, 1.3985330810546874, 1.39865087890625, 1.3985106201171875, 1.3986744384765626, 1.3990687255859375, 1.398646728515625, 1.3988494873046875, 1.3989488525390625, 1.3988157958984375, 1.4003507080078126, 1.3992181396484376, 1.3986282958984375, 1.3989796142578126, 1.3991424560546875, 1.3989110107421876, 1.39889453125, 1.398781982421875, 1.398676513671875, 1.39867236328125, 1.3985843505859374, 1.3988526611328125, 1.398802490234375, 1.3987880859375, 1.3986365966796874, 1.398794189453125, 1.3989488525390625, 1.3988382568359374, 1.398718505859375, 1.3987451171875, 1.3987327880859375, 1.398834228515625, 1.3988956298828126, 1.3989713134765625, 1.3988956298828126, 1.3992642822265624, 1.398856689453125, 1.399077880859375, 1.39888232421875, 1.39892529296875, 1.3987440185546876, 1.3989969482421876, 1.4024007568359376, 1.399172119140625, 1.3992745361328125, 1.3992008056640626, 1.3990318603515626, 1.398909912109375, 1.3992008056640626, 1.3993646240234374, 1.39901953125, 1.399173095703125, 1.3989775390625, 1.399046142578125, 1.399246826171875, 1.3991126708984376, 1.399083984375, 1.399047119140625, 1.3991341552734375, 1.399214111328125, 1.3988863525390625, 2.96650537109375, 1.3986048583984374, 1.3984266357421875, 1.39875537109375, 1.3986641845703125, 1.398703125, 1.3987225341796874, 1.39884130859375, 1.3988424072265624, 1.398752197265625, 1.398993896484375, 1.399047119140625, 1.3987225341796874, 1.39919775390625, 1.3986856689453124, 1.399119873046875, 1.3985965576171875, 1.400858642578125, 1.398835205078125, 1.3988424072265624, 1.3985096435546875, 1.3990072021484374, 1.3985545654296876, 1.3986754150390626, 1.3990328369140625, 1.3986468505859375, 1.3989324951171875, 1.3988331298828125, 1.3989283447265626, 1.398687744140625, 1.39865087890625, 1.3986221923828126, 1.3990799560546876, 1.3989017333984375, 1.39886083984375, 1.39890283203125, 1.39919873046875, 1.3989652099609375, 1.39868359375, 1.398814697265625, 1.3989857177734375, 1.3991802978515624, 1.3991793212890624, 1.3992017822265626, 1.398865966796875, 1.39908203125, 1.39892529296875, 1.399057373046875, 1.3991065673828125, 1.39901953125, 1.4028165283203125, 1.399258056640625, 1.39905029296875, 1.399616455078125, 1.3996502685546874, 1.399204833984375, 1.399150634765625, 1.399098388671875, 1.3990655517578126, 1.3990645751953126, 1.3991474609375, 1.3992633056640624, 1.39909423828125, 2.965466064453125, 1.398435791015625, 1.3989212646484375, 1.3987410888671874, 1.398793212890625, 1.398793212890625, 1.3986417236328126, 1.398877197265625, 1.398681640625, 1.398823974609375, 1.398592529296875, 1.398836181640625, 1.3988443603515626, 1.398813720703125, 1.398644775390625, 1.398635498046875, 1.3989150390625, 1.3987430419921876, 1.399067626953125, 1.399125, 1.398625244140625, 1.3987430419921876, 1.3987952880859376, 1.3990604248046874, 1.398804443359375, 1.3987532958984374, 1.39879736328125, 1.3987962646484375, 1.398583251953125, 1.398846435546875, 1.398750244140625, 1.3987994384765625, 1.3991485595703126, 1.3989007568359375, 1.3991126708984376, 1.3992816162109376, 1.3991004638671876, 1.3990328369140625, 1.4033602294921874, 1.398982666015625, 1.3990400390625, 1.3989600830078126, 1.3988065185546874, 1.3990594482421874, 1.398823974609375, 1.3988914794921874, 1.398825927734375, 1.3992816162109376, 1.3990482177734376, 1.3992335205078126, 1.3989591064453124, 1.399271484375, 1.3991065673828125, 1.399330810546875, 1.398982666015625, 1.3989150390625, 1.39902978515625, 1.3993175048828126, 1.399251953125, 1.3993543701171876, 1.399087158203125, 1.399087158203125, 1.39928271484375, 2.967877685546875, 1.399, 1.3988577880859374, 1.398720458984375, 1.3989058837890624, 1.3989591064453124, 1.398603759765625, 1.398509521484375, 1.3988433837890626, 1.398835205078125, 1.398892578125, 1.3990687255859375, 1.3989376220703125, 1.3988812255859375, 1.3986109619140625, 1.398836181640625, 1.3987666015625, 1.3988116455078126, 1.398899658203125, 1.398931396484375, 1.3985228271484376, 1.398972412109375, 1.3989959716796876, 1.3989703369140625, 1.3989365234375, 1.3988116455078126, 1.398856689453125, 1.39875537109375, 1.3987542724609374, 1.398794189453125, 1.3987860107421874, 1.3986048583984374, 1.3989365234375, 1.398892578125, 1.3988424072265624, 1.398866943359375, 1.4031943359375, 1.399446533203125, 1.3989171142578125, 1.3989058837890624, 1.39900927734375, 1.3987225341796874, 1.399125, 1.39943115234375, 1.3992969970703124, 1.398981689453125, 1.399066650390625, 1.39924169921875, 1.39911474609375, 1.3990235595703124, 1.3991434326171874, 1.3992274169921874, 1.398771728515625, 1.3991884765625, 1.3990921630859374, 1.3990921630859374, 1.3991424560546875, 1.399141357421875, 1.399182373046875, 1.399150634765625, 1.399173095703125, 1.3995130615234375, 1.3992161865234376, 2.966830078125, 1.39867236328125, 1.3989918212890624, 1.3987451171875, 1.399013427734375, 1.3989908447265624, 1.3986611328125, 1.3989530029296875, 1.3987686767578125, 1.3985955810546875, 1.3984910888671875, 1.398908935546875, 1.3986058349609376, 1.3991700439453125, 1.3991044921875, 1.3990963134765626, 1.398783935546875, 1.3988414306640624, 1.39865087890625, 1.398973388671875, 1.398529052734375, 1.398624267578125, 1.3986068115234376, 1.39877783203125, 1.3988494873046875, 1.398686767578125, 1.3989161376953125, 1.3988074951171876, 1.398720458984375, 1.3985545654296876, 1.398549560546875, 1.398677490234375, 1.399013427734375, 1.3988321533203125, 1.3990225830078125, 1.3991875, 1.399309326171875, 1.3990697021484375, 1.3987850341796875, 1.3989110107421876, 1.3988515625, 1.39879833984375, 1.3990164794921875, 1.3991669921875, 1.3989970703125, 1.398814697265625, 1.3990440673828124, 1.3993902587890625, 1.39905029296875, 1.3986529541015624, 1.399108642578125, 1.3995069580078126, 1.3990020751953125, 1.3991290283203126, 1.399109619140625, 1.3991793212890624, 1.3990748291015624, 1.3990533447265625, 1.399342041015625, 1.3992509765625, 1.3991875, 1.3994813232421874, 1.3992847900390626, 2.975549560546875, 1.3990369873046875, 1.398686767578125, 1.3988699951171875, 1.3988116455078126, 1.398876220703125, 1.398803466796875, 1.398488037109375, 1.398929443359375, 1.398972412109375, 1.398601806640625, 1.3989036865234374, 1.3985341796875, 1.398978515625, 1.398749267578125, 1.398830078125, 1.3989969482421876, 1.398941650390625, 1.3987880859375, 1.398920166015625, 1.3986058349609376, 1.3989754638671874, 1.3987789306640626, 1.3987738037109374, 1.398740966796875, 1.39875537109375, 1.398782958984375, 1.398865966796875, 1.3988935546875, 1.3988218994140624, 1.39879931640625, 1.398435791015625, 1.39890283203125, 1.39907275390625, 1.3990400390625, 1.39919970703125, 1.3989857177734375, 1.39919873046875, 1.399015380859375, 1.3991219482421875, 1.3990953369140624, 1.3992161865234376, 1.39879736328125, 1.399045166015625, 1.3990635986328126, 1.3989110107421876, 1.3989007568359375, 1.3991126708984376, 1.3991710205078125, 1.3991290283203126, 1.3995233154296876, 1.399393310546875, 1.4048031005859376, 1.3995601806640625, 1.39926318359375, 1.399203857421875, 1.399208984375, 1.3992806396484374, 1.3992008056640626, 1.3996329345703125, 1.399182373046875, 1.3990205078125, 1.399141357421875, 2.966223876953125, 1.398709228515625, 1.399329833984375, 1.3987266845703126, 1.398455322265625, 1.398873046875, 1.39873583984375, 1.39871533203125, 1.39875537109375, 1.3985904541015626, 1.398762451171875, 1.398834228515625, 1.3986702880859374, 1.3991434326171874, 1.3988095703125, 1.3991116943359374, 1.398962158203125, 1.3986590576171876, 1.398698974609375, 1.398835205078125, 1.39871435546875, 1.3988116455078126, 1.398834228515625, 1.398676513671875, 1.3988095703125, 1.398972412109375, 1.399234619140625, 1.3989007568359375, 1.3986314697265625, 1.3986949462890625, 1.398951904296875, 1.3991597900390624, 1.3991392822265625, 1.3987696533203124, 1.3990225830078125, 1.398940673828125, 1.3991015625, 1.3990347900390625, 1.39875634765625, 1.3991424560546875, 1.3989007568359375, 1.3987020263671874, 1.3989140625, 1.399394287109375, 1.3992960205078124, 1.3988515625, 1.3991485595703126, 1.399234619140625, 1.398983642578125, 1.3988310546875, 1.3990318603515626, 1.3990963134765626, 1.3992110595703124, 1.399140380859375, 1.3990972900390626, 1.39926123046875, 1.3992274169921874, 1.3991915283203125, 1.3993380126953125, 1.3993133544921874, 1.3990963134765626, 1.399372802734375, 1.3989632568359376]",tokens/s,0.7035145956380915,, +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 304, in hf_raise_for_status + response.raise_for_status() + File ""/usr/local/lib/python3.10/dist-packages/requests/models.py"", line 1024, in raise_for_status + raise HTTPError(http_error_msg, response=self) +requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1722, in _get_metadata_or_catch_error + metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1645, in get_hf_file_metadata + r = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 372, in _request_wrapper + response = _request_wrapper( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 396, in _request_wrapper + hf_raise_for_status(response) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status + raise HfHubHTTPError(message, response=response) from e +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4ab5-171b328a778f712479b07012;1bdc15d3-d38d-4d92-8329-1bff2dac1900) + +403 Forbidden: Authorization error.. +Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. +If you are trying to create or update content,make sure you have a token with the `write` role. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 399, in cached_file + resolved_file = hf_hub_download( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py"", line 114, in _inner_fn + return fn(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1221, in hf_hub_download + return _hf_hub_download_to_cache_dir( + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1325, in _hf_hub_download_to_cache_dir + _raise_on_head_call_error(head_call_error, force_download, local_files_only) + File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py"", line 1826, in _raise_on_head_call_error + raise LocalEntryNotFoundError( +huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. + +The above exception was the direct cause of the following exception: + +Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 37, in __init__ + super().__init__(config) + File ""/workspace/optimum_benchmark/backends/base.py"", line 62, in __init__ + self.pretrained_config = get_transformers_pretrained_config(self.config.model, **self.config.model_kwargs) + File ""/workspace/optimum_benchmark/backends/transformers_utils.py"", line 22, in get_transformers_pretrained_config + return AutoConfig.from_pretrained(model, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py"", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py"", line 689, in _get_config_dict + resolved_config_file = cached_file( + File ""/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py"", line 442, in cached_file + raise EnvironmentError( +OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like tiiuae/falcon-180B is not the path to a directory containing a file named config.json. +Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'. + +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,a,a,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -13512,7 +14607,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c12f8-354c821c2d64d2df5ea093a6;9c92f26d-23ac-48c3-be3a-a7377c86059a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d451d-66a1cb99491a095836922aa2;8a607815-958e-4c4f-bcf9-eea0d39a11a3) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13602,7 +14697,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664bee64-7596919177ffab9402952018;c9d9369a-1f10-4d32-a79a-e321c51a6e7a) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d3f74-3ae7b77e193f422736b4cee0;4c5f342a-47d1-4889-a8bd-128116016a03) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -13785,7 +14880,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c128e-15e8c5c4284bc5a20f7f897b;fc17249e-48c2-42df-a661-7034b664a4ae) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d44cd-4f1b75687f647e6b09e31f79;03a49401-4633-44ab-8d68-83d7252f6915) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13914,7 +15009,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664beef0-7311dd056fcf5f11075807c2;ca4d9bdb-e5ac-438c-b20f-8056ee940f1f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d3fe9-4d65a84a6ebb1969338f8738;38485712-97de-4698-904b-4ecd2dac5e77) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14043,7 +15138,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c1415-39f8e93320808a697549544d;958c20d7-9e60-4c05-a70b-83fbcd2663a3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d462c-5d8d0f2434749b5f32d831fa;38d66fc7-9b20-46a9-9a70-c9cb9e6ff37c) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14073,6 +15168,35 @@ Traceback (most recent call last): OSError: B is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1478, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1644, in _check_and_enable_sdpa + raise ValueError( +ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5928.198144,7053.246464,0.0,6406.79936,6098.60864,s,10,4.240164489746094,0.42401644897460944,0.0007862412146943955,0.4238206481933594,0.4252529205322265,0.42526194305419923,0.42526916107177737,"[0.4231962585449219, 0.42358792114257815, 0.42403680419921874, 0.42408154296875, 0.4247743225097656, 0.4236044921875, 0.4229784240722656, 0.4252709655761719, 0.42525091552734373, 0.4233828430175781]",tokens/s,603.7501625681733,kWh,4.996894775993295e-06,2.738078027505253e-06,2.4908191222833276e-05,3.264316402633182e-05,tokens/kWh,7842377.037761901,MB,5928.198144,7055.343616,0.0,6408.896512,6157.952,s,10,234.891478515625,23.4891478515625,0.0008632158708629218,23.4891904296875,23.490145703125002,23.4903619140625,23.4905348828125,"[23.490578125, 23.48784375, 23.4893828125, 23.49009765625, 23.4881875, 23.49003515625, 23.489291015625, 23.48875, 23.48822265625, 23.48908984375]",tokens/s,2.6820896355254216,kWh,0.0002773190428689122,0.00015199443763651288,0.0013502960478285586,0.0017796095283339835,tokens/kWh,35401.02421174306,,s,629,238.3356229553223,0.3789119601833422,0.050402910658677205,0.3728250732421875,0.3730047119140625,0.3731396545410156,0.7967181420898437,"[0.3726909484863281, 0.3727615966796875, 0.37275033569335936, 0.37272882080078124, 0.372853759765625, 0.3727329406738281, 0.3727196044921875, 0.37270938110351565, 0.37320193481445313, 0.37283941650390623, 0.3727615966796875, 0.37314047241210935, 0.37314047241210935, 0.3727513732910156, 0.3729571838378906, 0.3727513732910156, 0.37278411865234373, 0.37285888671875, 0.3729561462402344, 0.37282611083984374, 0.37271142578125, 0.3728527221679688, 0.3729326171875, 0.3729141845703125, 0.37295822143554686, 0.372853759765625, 0.37274215698242186, 0.3729397888183594, 0.37267864990234373, 0.37283428955078124, 0.3728598937988281, 0.372780029296875, 0.37270834350585935, 0.3727237243652344, 0.37280563354492186, 0.373138427734375, 0.372895751953125, 0.37279437255859377, 0.37294692993164064, 0.3730411376953125, 0.3728271484375, 0.37274520874023437, 0.37282302856445315, 0.37274215698242186, 0.37277490234375, 0.3727185974121094, 0.3728598937988281, 0.3727492980957031, 0.37287014770507815, 0.3729827880859375, 0.3730401306152344, 0.3728189392089844, 0.3728404541015625, 0.3727615966796875, 0.37286297607421875, 0.3739125671386719, 0.3729490051269531, 0.37281280517578125, 0.37288858032226563, 0.37284146118164063, 0.3727615966796875, 0.3727492980957031, 0.796822509765625, 0.37277081298828124, 0.3728988037109375, 0.37286297607421875, 0.37280767822265626, 0.37276980590820313, 0.37276568603515625, 0.372664306640625, 0.37277798461914063, 0.3728087158203125, 0.37276364135742185, 0.37271755981445315, 0.3729407958984375, 0.3727431640625, 0.3728045959472656, 0.37276364135742185, 0.37279232788085936, 0.3727605895996094, 0.3727933349609375, 0.37284146118164063, 0.3728793640136719, 0.3727667236328125, 0.3727462463378906, 0.3726837768554688, 0.37286810302734374, 0.3728844909667969, 0.37282815551757814, 0.3727462463378906, 0.3726909484863281, 0.37285479736328125, 0.3729295349121094, 0.372664306640625, 0.3727810668945313, 0.37266738891601564, 0.3728240661621094, 0.37283428955078124, 0.37280667114257815, 0.37290496826171876, 0.3727431640625, 0.37279437255859377, 0.3728773193359375, 0.3727790222167969, 0.37273907470703127, 0.3729797058105469, 0.3727872009277344, 0.3729725341796875, 0.37286297607421875, 0.3727892456054688, 0.3729284973144531, 0.37281997680664064, 0.37299508666992187, 0.3726796875, 0.37273907470703127, 0.37301861572265627, 0.3727605895996094, 0.3729971313476563, 0.3728527221679688, 0.3729121398925781, 0.37293875122070314, 0.3727738952636719, 0.3729407958984375, 0.37281689453125, 0.37280667114257815, 0.7972147216796875, 0.37270834350585935, 0.37275851440429686, 0.37316915893554686, 0.3728302001953125, 0.37271142578125, 0.37290292358398436, 0.3729121398925781, 0.3727564697265625, 0.3728250732421875, 0.37277490234375, 0.3727329406738281, 0.37278515625, 0.372853759765625, 0.37280667114257815, 0.372890625, 0.3730401306152344, 0.37271755981445315, 0.37296640014648436, 0.37314151000976564, 0.37271450805664064, 0.37271551513671874, 0.3727831115722656, 0.37286297607421875, 0.37275341796875, 0.3726489562988281, 0.3726499938964844, 0.37285580444335936, 0.37283428955078124, 0.372684814453125, 0.37318450927734376, 0.3728486328125, 0.3727667236328125, 0.3728619384765625, 0.37320703125, 0.37286297607421875, 0.3729203186035156, 0.372780029296875, 0.37272064208984373, 0.372780029296875, 0.37288858032226563, 0.3727615966796875, 0.37287832641601565, 0.37283328247070313, 0.37265716552734374, 0.3728660583496094, 0.3727339477539062, 0.3727984619140625, 0.37280255126953127, 0.37279437255859377, 0.3728988037109375, 0.372843505859375, 0.3729090576171875, 0.372811767578125, 0.3728015441894531, 0.3728568420410156, 0.3730687866210938, 0.37280767822265626, 0.3727984619140625, 0.3731435546875, 0.3727820739746094, 0.3728988037109375, 0.37287014770507815, 0.79723828125, 0.372748291015625, 0.3729121398925781, 0.372780029296875, 0.37285784912109377, 0.37321624755859373, 0.37275851440429686, 0.37264080810546873, 0.3728076477050781, 0.37328793334960936, 0.3728517150878906, 0.37312921142578126, 0.3726581726074219, 0.3728322448730469, 0.3728465881347656, 0.3728465881347656, 0.37285479736328125, 0.3727339477539062, 0.37266943359375, 0.37281484985351565, 0.372738037109375, 0.37278616333007814, 0.3727052917480469, 0.3727339477539062, 0.37282815551757814, 0.3726755981445313, 0.37273599243164063, 0.3728240661621094, 0.37270834350585935, 0.37299200439453123, 0.37282611083984374, 0.3730687866210938, 0.37305035400390624, 0.3729971313476563, 0.37284762573242186, 0.3728947143554687, 0.37294692993164064, 0.373064697265625, 0.37282302856445315, 0.37282815551757814, 0.37269708251953126, 0.3728568420410156, 0.3730114440917969, 0.37281484985351565, 0.372706298828125, 0.37283328247070313, 0.3727677307128906, 0.37275033569335936, 0.37276364135742185, 0.37286810302734374, 0.37285580444335936, 0.3730206604003906, 0.3728721923828125, 0.37273703002929687, 0.37278411865234373, 0.3727810668945313, 0.37282815551757814, 0.37268991088867187, 0.372843505859375, 0.3728404541015625, 0.37287014770507815, 0.37319781494140625, 0.3732561950683594, 0.7978936157226563, 0.37297357177734375, 0.3727278137207031, 0.3729295349121094, 0.37272268676757814, 0.3728271484375, 0.3726704711914062, 0.3726305236816406, 0.3727124328613281, 0.373001220703125, 0.3726376953125, 0.3726407775878906, 0.3728486328125, 0.372822021484375, 0.37276980590820313, 0.37266226196289065, 0.3728844909667969, 0.37276980590820313, 0.37276980590820313, 0.3727831115722656, 0.37270733642578124, 0.3727339477539062, 0.37273703002929687, 0.372748291015625, 0.372895751953125, 0.3728742370605469, 0.3728404541015625, 0.3729459228515625, 0.3727339477539062, 0.37284249877929687, 0.37284146118164063, 0.3728271484375, 0.3727185974121094, 0.37319064331054685, 0.37279232788085936, 0.3728660583496094, 0.3728517150878906, 0.37299508666992187, 0.3727933349609375, 0.37292645263671875, 0.3728650207519531, 0.37288037109375, 0.3727933349609375, 0.3727595520019531, 0.372701171875, 0.3728240661621094, 0.37269503784179686, 0.37287115478515626, 0.37288858032226563, 0.37272882080078124, 0.3728097229003906, 0.37293157958984374, 0.3728937072753906, 0.37281689453125, 0.37296844482421876, 0.3728373718261719, 0.37314151000976564, 0.37274420166015626, 0.37282302856445315, 0.3729407958984375, 0.37278515625, 0.3729490051269531, 0.3727882385253906, 0.7968818969726562, 0.37289166259765627, 0.3727564697265625, 0.37314047241210935, 0.37275851440429686, 0.37275341796875, 0.3727810668945313, 0.37276467895507814, 0.3728670654296875, 0.37284146118164063, 0.3728240661621094, 0.37266738891601564, 0.37273703002929687, 0.3727615966796875, 0.3727687683105469, 0.3727124328613281, 0.37284249877929687, 0.37286297607421875, 0.3728650207519531, 0.3729090576171875, 0.372642822265625, 0.372811767578125, 0.3728302001953125, 0.37274520874023437, 0.3727267761230469, 0.3728302001953125, 0.3726663818359375, 0.37289675903320313, 0.3727339477539062, 0.3727339477539062, 0.37296127319335937, 0.37266021728515625, 0.37277593994140623, 0.3729090576171875, 0.3729879150390625, 0.37280667114257815, 0.37285580444335936, 0.3727400817871094, 0.373080078125, 0.3728445434570313, 0.37341592407226565, 0.37341900634765623, 0.3727411193847656, 0.37296844482421876, 0.3728793640136719, 0.3728035888671875, 0.37276364135742185, 0.3726755981445313, 0.3727882385253906, 0.37276467895507814, 0.3730616455078125, 0.3729203186035156, 0.37335140991210936, 0.37285479736328125, 0.3728292236328125, 0.3729039001464844, 0.3727032470703125, 0.3728773193359375, 0.3730513916015625, 0.372701171875, 0.3728517150878906, 0.3728353271484375, 0.3729141845703125, 0.796095458984375, 0.37281381225585936, 0.3726960754394531, 0.37284249877929687, 0.37278616333007814, 0.3727964172363281, 0.3726796875, 0.37272268676757814, 0.3727769470214844, 0.3726714782714844, 0.37295822143554686, 0.37281689453125, 0.3726376953125, 0.37286810302734374, 0.37288037109375, 0.37296640014648436, 0.3727790222167969, 0.37272882080078124, 0.37284146118164063, 0.37298175048828125, 0.37272882080078124, 0.37291110229492186, 0.37273599243164063, 0.3727267761230469, 0.37271038818359375, 0.372853759765625, 0.37274725341796877, 0.3727984619140625, 0.372885498046875, 0.37299508666992187, 0.37281997680664064, 0.3729213562011719, 0.37272882080078124, 0.373032958984375, 0.37307391357421876, 0.3728363647460938, 0.37282815551757814, 0.3728087158203125, 0.3728824462890625, 0.37300634765625, 0.3728322448730469, 0.3730145263671875, 0.37303704833984375, 0.37271755981445315, 0.3727237243652344, 0.37280563354492186, 0.37287014770507815, 0.3728302001953125, 0.3729704895019531, 0.3728486328125, 0.3729725341796875, 0.372890625, 0.37290292358398436, 0.37265921020507814, 0.37278515625, 0.37288140869140624, 0.37280050659179687, 0.372917236328125, 0.37298074340820314, 0.37283328247070313, 0.37277798461914063, 0.3730595703125, 0.37283123779296873, 0.7982418212890625, 0.3727196044921875, 0.3727431640625, 0.37278411865234373, 0.3728455810546875, 0.3728035888671875, 0.3726776428222656, 0.3727595520019531, 0.3727492980957031, 0.3727349853515625, 0.3727912902832031, 0.37254962158203125, 0.3728793640136719, 0.3726776428222656, 0.3727882385253906, 0.37269503784179686, 0.37284762573242186, 0.3728517150878906, 0.37276263427734374, 0.37303091430664065, 0.3731087341308594, 0.37288958740234374, 0.37265408325195315, 0.372632568359375, 0.37281585693359376, 0.37261721801757813, 0.3727052917480469, 0.3728619384765625, 0.37265921020507814, 0.37284249877929687, 0.37281689453125, 0.3733667907714844, 0.3728865356445312, 0.3727513732910156, 0.37287115478515626, 0.3729981384277344, 0.37291928100585936, 0.3728742370605469, 0.372780029296875, 0.372959228515625, 0.3728271484375, 0.3727431640625, 0.372748291015625, 0.3727544250488281, 0.3728250732421875, 0.3729203186035156, 0.3729070129394531, 0.37317428588867185, 0.3730114440917969, 0.37277285766601564, 0.37303704833984375, 0.3728404541015625, 0.37297357177734375, 0.3727964172363281, 0.37282614135742187, 0.3728813781738281, 0.37282302856445315, 0.3727564697265625, 0.3728721923828125, 0.3728721923828125, 0.37279232788085936, 0.373080078125, 0.37286810302734374, 0.7984179077148438, 0.37283328247070313, 0.372748291015625, 0.37297357177734375, 0.37281381225585936, 0.3728302001953125, 0.37263360595703127, 0.37311386108398437, 0.3727513732910156, 0.3728035888671875, 0.37269403076171875, 0.37263156127929686, 0.372674560546875, 0.3728650207519531, 0.37274420166015626, 0.37281280517578125, 0.37275546264648435, 0.3726295166015625, 0.3728558349609375, 0.37276873779296876, 0.3728271484375, 0.37277490234375, 0.37270834350585935, 0.3729121398925781, 0.3726858215332031, 0.37277490234375, 0.372706298828125, 0.37280667114257815, 0.37266534423828124, 0.3727595520019531, 0.3728875427246094, 0.37261822509765624, 0.3728506774902344, 0.3729121398925781, 0.3732254638671875, 0.3727953796386719, 0.37293057250976563, 0.37280767822265626, 0.3729571838378906, 0.3729776611328125, 0.37303399658203124, 0.37287832641601565, 0.37280667114257815, 0.3728670654296875, 0.372790283203125, 0.37291622924804685, 0.37279232788085936, 0.3727677307128906, 0.3727677307128906, 0.37282098388671875, 0.37280255126953127, 0.37281585693359376, 0.3730176086425781, 0.3728824462890625, 0.3729018859863281, 0.37278616333007814, 0.3728271484375, 0.3729571838378906, 0.37275851440429686, 0.37271038818359375, 0.37286911010742185, 0.3729203186035156, 0.37288140869140624, 0.7964497680664062, 0.3727339477539062, 0.37287115478515626, 0.3727738952636719, 0.37274725341796877, 0.37273907470703127, 0.37268685913085936, 0.3725946960449219, 0.37271450805664064, 0.37275341796875, 0.3733821411132813, 0.3728527221679688, 0.3727769470214844, 0.3728189392089844, 0.372885498046875, 0.37269403076171875, 0.3727492980957031, 0.37275033569335936, 0.3726684265136719, 0.3727964172363281, 0.37269503784179686, 0.37292340087890624, 0.3727769470214844, 0.37295309448242187, 0.3729715270996094, 0.3729725341796875, 0.3730043029785156, 0.3729541015625, 0.3728486328125, 0.372990966796875, 0.3728445434570313, 0.37290802001953127, 0.3728527221679688, 0.3727892456054688, 0.3727872009277344, 0.3730176086425781, 0.3727687683105469, 0.3729141845703125, 0.3727267761230469, 0.3728240661621094, 0.3727892456054688, 0.37288345336914064, 0.3727738952636719, 0.37278515625, 0.3728455810546875, 0.3726356506347656, 0.3727677307128906, 0.3729776611328125, 0.37292340087890624, 0.37301043701171877, 0.373001220703125, 0.3728322448730469, 0.37296438598632814, 0.3727707824707031, 0.37279437255859377, 0.3728455810546875, 0.3728271484375, 0.3728291931152344, 0.3728353271484375, 0.37280050659179687, 0.37286911010742185, 0.3729459228515625, 0.37285888671875]",tokens/s,2.6391354854994153,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,s,s,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -14114,7 +15238,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c10bc-712a361b095f879d03598236;498eff1a-ef89-4c19-80ad-4e85e8b41f66) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d430c-1948004e647d2d9e2d917af3;83a7837c-a298-4320-96a4-27bbc234f70c) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14144,6 +15268,35 @@ Traceback (most recent call last): OSError: s is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` +",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): + File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch + benchmark_report = Benchmark.launch(benchmark_config) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch + report = launcher.launch(worker=cls.run, worker_args=[config]) + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch + raise ChildProcessError(response[""traceback""]) +ChildProcessError: Traceback (most recent call last): + File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target + report = worker(*worker_args) + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run + backend: Backend = backend_factory(backend_config) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ + self.load_model_with_no_weights() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights + self.load_model_from_pretrained() + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained + self.pretrained_model = self.automodel_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 563, in from_pretrained + return model_class.from_pretrained( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3620, in from_pretrained + config = cls._autoset_attn_implementation( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1478, in _autoset_attn_implementation + config = cls._check_and_enable_sdpa( + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1644, in _check_and_enable_sdpa + raise ValueError( +ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` + ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch @@ -14466,7 +15619,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c106e-3a49ec5668aefd1d6c092352;889de78a-027a-4504-8a3c-2cf0df247062) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d42b4-7c15790b11e1693d70b580db;955d30e0-3551-45d4-a50a-64d07f8f9c50) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14536,7 +15689,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c14da-6964852b3178f7ae3347f0ea;cb98d63b-5c13-461d-a943-ac451e88f45c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46f0-4f381d0120c10a524a28701f;51a09c2b-cf60-49a5-b893-d6c0be84a1ab) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14606,7 +15759,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664c148c-0d9a3c2a37baf7694dce8ea7;0053dc3a-47d4-4692-b57a-5fe702e699a8) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46a2-0e1566c026efc0785a00e8ae;526ed273-5f97-4bb0-a5d9-e29103d1197a) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`.