diff --git "a/perf-df-bnb-1xA10.csv" "b/perf-df-bnb-1xA10.csv" --- "a/perf-df-bnb-1xA10.csv" +++ "b/perf-df-bnb-1xA10.csv" @@ -242,7 +242,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 192.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -525,7 +525,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1319.202816,9043.443712,0.0,8396.996608,8008.01792,s,10,5.896726806640625,0.5896726806640624,0.00026212653204851826,0.5896717224121093,0.5898943115234374,0.5900787170410156,0.5902262414550781,"[0.5902631225585937, 0.5893323364257812, 0.5893238525390625, 0.5895547485351562, 0.5897279663085937, 0.5898533325195312, 0.5898179321289062, 0.5895100708007812, 0.5897154541015625, 0.5896279907226563]",tokens/s,434.1391561700035,kWh,6.965565515889062e-06,3.816821717876283e-06,4.512784655976448e-05,5.591023379352982e-05,tokens/kWh,4578768.190191783,MB,1319.202816,9043.443712,0.0,8396.996608,8008.02048,s,10,51.720919921874994,5.172091992187501,0.03199561273793778,5.16110205078125,5.200138427734375,5.230424438476563,5.254653247070312,"[5.26071044921875, 5.1531064453125, 5.16850439453125, 5.193408203125, 5.15474853515625, 5.16189208984375, 5.1687548828125, 5.15084326171875, 5.1486396484375, 5.16031201171875]",tokens/s,12.18075782394477,kWh,6.24684377345774e-05,3.423757327319167e-05,0.00017972266665383635,0.0002764286776616054,tokens/kWh,227906.88915830385,,s,629,56.94001155090334,0.09052466065326442,0.07012736713909555,0.08166912078857422,0.08367902984619141,0.08570038452148437,0.6721920239257813,"[0.08598121643066406, 0.0854783706665039, 0.08584909057617188, 0.08569856262207032, 0.08566067504882813, 0.08607027435302735, 0.08577433776855468, 0.08586547088623046, 0.08603545379638672, 0.08604774475097657, 0.08603033447265625, 0.08688127899169922, 0.08623923492431641, 0.08137932586669921, 0.0814039077758789, 0.08132403564453125, 0.08151859283447266, 0.08160665893554687, 0.08168447875976563, 0.0814172134399414, 0.08171417236328125, 0.08234598541259766, 0.08227839660644531, 0.08186367797851563, 0.0812564468383789, 0.0812943344116211, 0.08116223907470703, 0.08145919799804688, 0.08139369964599609, 0.08133628845214844, 0.08130867004394532, 0.08153395080566406, 0.08149094390869141, 0.08115923309326172, 0.08133010864257813, 0.08142355346679687, 0.0837619857788086, 0.08571107482910156, 0.08674486541748047, 0.08162509155273437, 0.08396800231933593, 0.08661507415771484, 0.08570159912109375, 0.08567910766601562, 0.08584806060791016, 0.08146739196777343, 0.08154316711425781, 0.08158719635009766, 0.08162918090820312, 0.08164044952392578, 0.0811151351928711, 0.08137318420410156, 0.08232243347167968, 0.08588604736328125, 0.08569436645507812, 0.08584416198730468, 0.08571695709228516, 0.08523868560791016, 0.08626579284667969, 0.08208895874023438, 0.08409407806396485, 0.08153791809082031, 0.6719406127929688, 0.08185753631591797, 0.08154009246826172, 0.08195686340332031, 0.08159232330322265, 0.08155967712402344, 0.08148467254638672, 0.08141619110107422, 0.0815146255493164, 0.08282201385498048, 0.08156569671630859, 0.08121651458740234, 0.08132300567626953, 0.08270130920410156, 0.0816332778930664, 0.08157081604003906, 0.08167219543457031, 0.08201119995117187, 0.08188921356201172, 0.08139263916015625, 0.08153190612792968, 0.08142642974853516, 0.08134143829345702, 0.08124313354492188, 0.0814755859375, 0.08129740905761719, 0.08164249420166016, 0.08143360137939454, 0.08155033874511719, 0.08404582214355469, 0.0816742706298828, 0.08162505340576172, 0.08150527954101562, 0.08175513458251953, 0.08145203399658203, 0.0815964126586914, 0.08145619201660156, 0.0814048614501953, 0.0815810546875, 0.0817419204711914, 0.08134550476074219, 0.0814970245361328, 0.08176537322998047, 0.0815626220703125, 0.08154930877685547, 0.08182374572753906, 0.08233164978027344, 0.08217929840087891, 0.0831455078125, 0.08158822631835938, 0.081797119140625, 0.08196915435791016, 0.08143360137939454, 0.081691650390625, 0.08263270568847657, 0.0817786865234375, 0.08154323577880859, 0.08337709045410156, 0.08180121612548828, 0.08236851501464844, 0.08175730895996093, 0.08172940826416016, 0.08236953735351563, 0.672289794921875, 0.08184941101074218, 0.08308217620849609, 0.08367616271972657, 0.08152371215820313, 0.08200102233886719, 0.08170995330810547, 0.08176844787597656, 0.08257433319091798, 0.08175615692138671, 0.08157593536376953, 0.08134963226318359, 0.08312422180175781, 0.08147666931152343, 0.08125536346435547, 0.08136294555664063, 0.08140083312988282, 0.08147865295410156, 0.08164051055908203, 0.08156665802001953, 0.08184524536132813, 0.08193331146240235, 0.08301773071289062, 0.08472370910644532, 0.08209308624267578, 0.08115196990966797, 0.0827688980102539, 0.08197017669677735, 0.08144486236572265, 0.0812410888671875, 0.08203775787353515, 0.08156671905517578, 0.08194969940185547, 0.0813311996459961, 0.08171622467041016, 0.08153497314453124, 0.08227123260498047, 0.08154326629638672, 0.08151952362060547, 0.08161280059814453, 0.08145827484130859, 0.08139561462402344, 0.08492339324951172, 0.0834703369140625, 0.08132608032226563, 0.08184320068359376, 0.08130457305908204, 0.08137318420410156, 0.0819261474609375, 0.08215961456298829, 0.08194969940185547, 0.0821770248413086, 0.08171212768554688, 0.08147763061523437, 0.08170393371582031, 0.0817080307006836, 0.08141824340820313, 0.08207564544677734, 0.08254668426513671, 0.08159539031982421, 0.0829276123046875, 0.08584601593017578, 0.08180941009521485, 0.6725437622070313, 0.0816701431274414, 0.08299417877197265, 0.0818976287841797, 0.08269808197021485, 0.08372633361816406, 0.08145817565917969, 0.0814326400756836, 0.08147142028808593, 0.08147148895263671, 0.08324813079833984, 0.08439603424072266, 0.08220569610595703, 0.08173567962646484, 0.08186163330078125, 0.08427827453613282, 0.08266764831542969, 0.08259059143066406, 0.08248422241210937, 0.08257862091064454, 0.08179078674316406, 0.08185654449462891, 0.08350921630859375, 0.08204185485839843, 0.08292658996582031, 0.08164659118652344, 0.08164864349365235, 0.08335257720947266, 0.08254873657226562, 0.0838123550415039, 0.08271974182128906, 0.08367513275146485, 0.08213811492919922, 0.08257023620605469, 0.08335871887207032, 0.08214527893066406, 0.08226934051513672, 0.08150729370117188, 0.08136998748779296, 0.0828927993774414, 0.08226406097412109, 0.08139059448242188, 0.08316416168212891, 0.08163123321533203, 0.08356147003173828, 0.08335052490234375, 0.08262348937988281, 0.08280166625976562, 0.08164556884765625, 0.08133837127685548, 0.08139059448242188, 0.08324095916748046, 0.0815810546875, 0.08414022064208984, 0.08362595367431641, 0.08281378936767578, 0.08182169342041015, 0.08148070526123047, 0.08191999816894531, 0.08212070465087891, 0.08173567962646484, 0.08164249420166016, 0.08254166412353516, 0.672660400390625, 0.08245452880859375, 0.0815288314819336, 0.08148377227783203, 0.08647894287109376, 0.08233155059814454, 0.08190271759033203, 0.08214412689208984, 0.08166194915771484, 0.08165171051025391, 0.08193344116210938, 0.0821164779663086, 0.08232345581054687, 0.0821229476928711, 0.08210208129882812, 0.08151859283447266, 0.08135782623291016, 0.08166912078857422, 0.08189132690429687, 0.08124825286865234, 0.08137728118896484, 0.08065740966796875, 0.08134355163574218, 0.08176121520996094, 0.08308019256591796, 0.08202252960205078, 0.0816833953857422, 0.08122777557373047, 0.08128921508789062, 0.08144070434570312, 0.0815841293334961, 0.0813803482055664, 0.08121036529541016, 0.081006591796875, 0.081342529296875, 0.08246675109863282, 0.081112060546875, 0.08053043365478516, 0.08086630249023438, 0.08098099517822266, 0.08133324432373047, 0.08138751983642578, 0.08494284820556641, 0.0818524169921875, 0.08139775848388672, 0.08141423797607422, 0.08229161834716797, 0.08346316528320312, 0.08133529663085938, 0.08110489654541016, 0.08143158721923828, 0.08158306884765625, 0.08111615753173829, 0.08149606323242188, 0.08271052551269531, 0.08220262145996093, 0.08205824279785157, 0.08132915496826172, 0.08152371215820313, 0.0815637435913086, 0.0813855972290039, 0.0814549789428711, 0.08308521270751953, 0.6730864868164063, 0.0817100830078125, 0.08181043243408204, 0.08168755340576171, 0.08148480224609375, 0.08149606323242188, 0.08168038177490235, 0.08150630187988281, 0.08187904357910156, 0.08181356811523438, 0.08159430694580078, 0.08318873596191406, 0.08122982025146484, 0.08196828460693359, 0.08185533142089843, 0.08158515167236328, 0.0813117446899414, 0.08160768127441406, 0.08167731475830078, 0.08148582458496094, 0.08156380462646484, 0.08165462493896485, 0.08164044952392578, 0.08193433380126953, 0.08137113952636718, 0.0821443862915039, 0.08154303741455078, 0.0830044174194336, 0.08234188842773438, 0.08136192321777344, 0.08123801422119141, 0.08144179534912109, 0.08117964935302735, 0.08163737487792969, 0.08142150115966797, 0.08256800079345702, 0.08272793579101563, 0.08183602905273438, 0.08191603088378906, 0.08245855712890625, 0.08282310485839844, 0.0830904312133789, 0.08287026977539062, 0.08189746856689453, 0.08201728057861328, 0.08198973083496094, 0.08177654266357422, 0.08173670196533203, 0.08145919799804688, 0.08115711975097656, 0.0815288314819336, 0.0815934066772461, 0.08163219451904297, 0.08135785675048827, 0.0816885757446289, 0.08129737854003906, 0.08578253173828125, 0.08414617919921875, 0.08186675262451172, 0.0821801300048828, 0.08173257446289063, 0.08175103759765626, 0.08183193969726563, 0.67270654296875, 0.08198348999023437, 0.08186675262451172, 0.08158310699462891, 0.08135065460205078, 0.08432742309570312, 0.08253446197509766, 0.08141919708251953, 0.08127999877929687, 0.08125440216064453, 0.08135372924804687, 0.08108748626708985, 0.08207974243164062, 0.08278733062744141, 0.08315296173095703, 0.08211756896972656, 0.08262860870361328, 0.08196812438964844, 0.08262879943847656, 0.08169964599609375, 0.08189337921142578, 0.08180223846435547, 0.08169676971435547, 0.08429977416992188, 0.08342527770996094, 0.08201728057861328, 0.08161491394042969, 0.0815687026977539, 0.08260198211669922, 0.08168061065673828, 0.08142211151123047, 0.08114482879638672, 0.08310169219970703, 0.0812410888671875, 0.0819783706665039, 0.08209510040283204, 0.08369049835205078, 0.08197232055664062, 0.08267356872558594, 0.08201216125488281, 0.08308748626708984, 0.08237664031982422, 0.0816219482421875, 0.08173056030273437, 0.08605184173583984, 0.08263475036621094, 0.08151961517333985, 0.08128819274902344, 0.08154930877685547, 0.0816209945678711, 0.08137420654296874, 0.0812390365600586, 0.08094310760498047, 0.08128511810302734, 0.08175513458251953, 0.08166316986083984, 0.08166585540771484, 0.08154637145996094, 0.08137305450439453, 0.08089513397216797, 0.08154716491699218, 0.08147859191894531, 0.08262060546875, 0.671777587890625, 0.08148684692382813, 0.08177049255371094, 0.08174079895019531, 0.08170496368408203, 0.08164761352539063, 0.08150732421875, 0.08097792053222656, 0.08322156524658203, 0.08251590728759765, 0.08136192321777344, 0.0814233627319336, 0.08132915496826172, 0.08111001586914063, 0.08148889923095703, 0.08284979248046875, 0.08191078186035156, 0.08224870300292969, 0.08213507080078125, 0.08157488250732423, 0.08164761352539063, 0.08164454650878906, 0.08128717041015625, 0.08139170837402344, 0.08169052886962891, 0.08132198333740234, 0.08176640319824219, 0.08152473449707032, 0.08308019256591796, 0.08210431671142578, 0.08133529663085938, 0.0813803482055664, 0.08153292846679687, 0.08147148895263671, 0.08141433715820312, 0.08146617889404296, 0.08148992156982422, 0.08351232147216797, 0.08391474914550781, 0.08213811492919922, 0.08200192260742188, 0.08190573120117188, 0.08143353271484376, 0.08160460662841797, 0.0814172134399414, 0.08180223846435547, 0.08131590270996093, 0.0816434555053711, 0.08135372924804687, 0.0815810546875, 0.08159846496582031, 0.08168653106689452, 0.08158310699462891, 0.08139263916015625, 0.08135167694091797, 0.08126976013183594, 0.08123187255859375, 0.083272705078125, 0.08229686737060547, 0.08144175720214844, 0.08157389068603516, 0.0814755859375, 0.08128205108642578, 0.6725498657226563, 0.08201216125488281, 0.08160665893554687, 0.08182681274414062, 0.08139366149902344, 0.08449961853027343, 0.08289469146728516, 0.08166912078857422, 0.08166400146484375, 0.08163126373291016, 0.08308940887451172, 0.081882080078125, 0.08147456359863281, 0.0811509780883789, 0.08164556884765625, 0.0813117446899414, 0.08126054382324219, 0.08160358428955078, 0.08188009643554688, 0.08243401336669921, 0.08137932586669921, 0.08149298858642579, 0.08213401794433593, 0.08148377227783203, 0.0813005142211914, 0.08151651000976562, 0.0826050567626953, 0.08456192016601563, 0.08146534729003906, 0.08109670257568359, 0.08133334350585937, 0.08120207977294921, 0.08153600311279297, 0.08106086730957031, 0.08127999877929687, 0.08195993804931641, 0.0819240951538086, 0.0815462417602539, 0.08128921508789062, 0.08280166625976562, 0.08131686401367187, 0.08106086730957031, 0.08126771545410157, 0.08141516876220703, 0.0818370590209961, 0.0813117446899414, 0.08142438507080078, 0.08106291198730468, 0.0813311996459961, 0.08158924865722657, 0.08149913787841796, 0.08117555236816407, 0.08124518585205077, 0.08163737487792969, 0.0808304672241211, 0.08384931182861328, 0.08154822540283203, 0.08121449279785156, 0.08134547424316406, 0.0814919662475586, 0.08126361846923828, 0.0817786865234375, 0.08144179534912109, 0.6729789428710937, 0.08176332855224609, 0.08148172760009766, 0.08134451293945312, 0.0810466537475586, 0.08107711791992188, 0.08121139526367188, 0.08119206237792968, 0.08125222778320312, 0.08112947082519531, 0.08264908599853515, 0.08139161682128906, 0.08172560119628906, 0.08166896057128906, 0.08164864349365235, 0.08168755340576171, 0.08153804779052734, 0.08195686340332031, 0.08161689758300782, 0.08461824035644532, 0.08384614562988281, 0.08173465728759766, 0.08185139465332031, 0.08169369506835937, 0.08177561950683594, 0.08152780914306641, 0.08148582458496094, 0.08169983673095703, 0.08150425720214843, 0.0814755859375, 0.08133222198486328, 0.081544189453125, 0.08147148895263671, 0.08166706848144531, 0.0822590103149414, 0.08144697570800781, 0.08130547332763671, 0.08122982025146484, 0.08153600311279297, 0.08408985900878906, 0.08272998046875, 0.0818687973022461, 0.08162918090820312, 0.08164351654052734, 0.08125151824951173, 0.08148870086669922, 0.0817479705810547, 0.08159436798095702, 0.08132403564453125, 0.08240332794189453, 0.08200908660888671, 0.08223846435546875, 0.08379698944091797, 0.08200816345214844, 0.08158403015136718, 0.08173977661132813, 0.0823634262084961, 0.08420246124267577, 0.0824985580444336, 0.0819210205078125, 0.08133222198486328, 0.08211558532714844, 0.08303308868408203]",tokens/s,11.046713600289413,,, -4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -554,7 +554,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -992,7 +992,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1844.416512,3485.990912,0.0,2839.543808,2585.78432,s,10,0.7208391494750976,0.07208391494750976,0.0010593011326744488,0.07179596710205079,0.07217364807128906,0.07371277770996093,0.07494408142089844,"[0.07525190734863281, 0.07183161926269531, 0.07159577941894531, 0.07166025543212891, 0.07179535675048829, 0.07181980895996094, 0.07180057525634766, 0.07164816284179687, 0.07163910675048828, 0.07179657745361329]",tokens/s,3551.416431618825,kWh,8.471489603084233e-07,4.6419818784540567e-07,5.0629107963571074e-06,6.374257944510935e-06,tokens/kWh,40161537.58265294,MB,1844.416512,3485.990912,0.0,2839.543808,2585.78688,s,10,32.42097436523437,3.242097436523437,0.05245766106895201,3.241113037109375,3.316388037109375,3.3204720947265627,3.3237393408203126,"[3.32455615234375, 3.250596923828125, 3.186911376953125, 3.31548046875, 3.21188330078125, 3.263601806640625, 3.143630615234375, 3.2239912109375, 3.268693359375, 3.231629150390625]",tokens/s,19.431865091493396,kWh,3.705561255533543e-05,2.030822340021716e-05,6.564659021564359e-05,0.00012301042617119618,tokens/kWh,512151.7090943298,,s,629,33.011150894165056,0.05248195690646271,0.008535835568901745,0.05236326217651367,0.05292093353271485,0.053374565887451174,0.12176461975097656,"[0.05524684906005859, 0.05259366226196289, 0.05517107009887695, 0.054330368041992184, 0.054488063812255856, 0.05532364654541016, 0.053797889709472656, 0.0525404167175293, 0.054201343536376956, 0.05319680023193359, 0.05262233734130859, 0.052686847686767575, 0.05267763137817383, 0.05260595321655273, 0.05269913482666016, 0.05261209487915039, 0.05247385787963867, 0.05271449661254883, 0.052654079437255856, 0.052798465728759764, 0.05261004638671875, 0.05266636657714844, 0.05250259017944336, 0.05289056015014648, 0.05259571075439453, 0.05266022491455078, 0.052446208953857425, 0.05264384078979492, 0.05271449661254883, 0.052613121032714844, 0.052569087982177735, 0.052416511535644535, 0.05210009765625, 0.05234688186645508, 0.05247385787963867, 0.05256192016601562, 0.05225779342651367, 0.05242985534667969, 0.05269807815551758, 0.05286195373535156, 0.05210009765625, 0.05267763137817383, 0.053465087890625, 0.05284662246704101, 0.05278307342529297, 0.052572158813476565, 0.05175296020507812, 0.05228441619873047, 0.052104190826416014, 0.05242675018310547, 0.0519659538269043, 0.05216153717041016, 0.05163417434692383, 0.05204070281982422, 0.05218406295776367, 0.05212160110473633, 0.05211347198486328, 0.052210624694824216, 0.052157440185546876, 0.052152320861816405, 0.0522608642578125, 0.052300830841064454, 0.12214678192138671, 0.049549312591552735, 0.04994662475585938, 0.049710079193115236, 0.04982681655883789, 0.04973567962646484, 0.049775615692138675, 0.04985343933105469, 0.05024051284790039, 0.04974489593505859, 0.049740798950195314, 0.049701889038085936, 0.04986777496337891, 0.04990467071533203, 0.049756126403808595, 0.049870849609375, 0.052242431640625, 0.053852161407470706, 0.05294182586669922, 0.05282099151611328, 0.05307187271118164, 0.05291110229492187, 0.049944576263427735, 0.049893375396728515, 0.049926143646240234, 0.04974796676635742, 0.05142835235595703, 0.052598785400390625, 0.052765697479248044, 0.05271039962768555, 0.0525404167175293, 0.052673534393310545, 0.05262540817260742, 0.05266124725341797, 0.052675582885742187, 0.05256294250488281, 0.052939777374267576, 0.05256198501586914, 0.0526181755065918, 0.05269708633422852, 0.05283638381958008, 0.05257622528076172, 0.05260902404785156, 0.05266841506958008, 0.052634624481201174, 0.05249433517456055, 0.04990259170532227, 0.04996198272705078, 0.04988927841186523, 0.04979916763305664, 0.050508800506591796, 0.053085182189941404, 0.05285478210449219, 0.052841472625732425, 0.05267251205444336, 0.052744190216064454, 0.05290086364746094, 0.052782081604003904, 0.05286604690551758, 0.05284044647216797, 0.0529172477722168, 0.05260800170898437, 0.05270220947265625, 0.12197682952880859, 0.04973567962646484, 0.04974387359619141, 0.04969369506835938, 0.049683456420898435, 0.04958617782592773, 0.04965683364868164, 0.04984320068359375, 0.04964863967895508, 0.04955033493041992, 0.049616897583007816, 0.04984012985229492, 0.0496517105102539, 0.04964147186279297, 0.04971724700927734, 0.0496629753112793, 0.04968550491333008, 0.04956159973144531, 0.04961999893188476, 0.049701854705810546, 0.049699840545654295, 0.0496629753112793, 0.049719295501708984, 0.04972032165527344, 0.049721343994140625, 0.051194881439208986, 0.05389926528930664, 0.052751361846923826, 0.0529172477722168, 0.05268172836303711, 0.05184000015258789, 0.049926143646240234, 0.04982067108154297, 0.049748062133789066, 0.0497099838256836, 0.049676319122314454, 0.04992816162109375, 0.04969372940063477, 0.049805278778076174, 0.04970086288452148, 0.049691646575927735, 0.04959641647338867, 0.04969267272949219, 0.04974796676635742, 0.0496629753112793, 0.05259468841552734, 0.05344563293457031, 0.05293772888183594, 0.052873279571533205, 0.052592575073242186, 0.05235200119018555, 0.05231513595581055, 0.052529151916503904, 0.05229875183105469, 0.05240627288818359, 0.052478977203369144, 0.052511745452880856, 0.05270732879638672, 0.049732608795166014, 0.049593345642089844, 0.0497151985168457, 0.04960563278198242, 0.049827838897705076, 0.1239019546508789, 0.053005313873291014, 0.053164031982421874, 0.05292031860351563, 0.049751041412353515, 0.05084262466430664, 0.053678081512451174, 0.05344972610473633, 0.052792320251464846, 0.05290086364746094, 0.053103614807128906, 0.052923393249511716, 0.05286195373535156, 0.05250969696044922, 0.052618240356445314, 0.05253222274780273, 0.05254143905639649, 0.052617279052734375, 0.052555713653564456, 0.05255372619628906, 0.052552703857421876, 0.05273702239990234, 0.052739070892333983, 0.05299507141113281, 0.053655582427978514, 0.05288035202026367, 0.052343807220458984, 0.05209088134765625, 0.052653057098388675, 0.05268377685546875, 0.05274009704589844, 0.05270528030395508, 0.052744190216064454, 0.05284864044189453, 0.05259980773925781, 0.05269708633422852, 0.05268889617919922, 0.05264384078979492, 0.05269708633422852, 0.052822017669677736, 0.052724735260009765, 0.05260800170898437, 0.05270016098022461, 0.05285171127319336, 0.05296537780761719, 0.0526929931640625, 0.052155391693115234, 0.05222707366943359, 0.05282815933227539, 0.05264691162109375, 0.052838401794433595, 0.05254143905639649, 0.05269094467163086, 0.05239091110229492, 0.0523612174987793, 0.05253324890136719, 0.05239910507202149, 0.052367359161376956, 0.05241753768920898, 0.05258956909179688, 0.05262438583374023, 0.052673534393310545, 0.05260595321655273, 0.12182425689697265, 0.04960563278198242, 0.04967116928100586, 0.049585151672363284, 0.04964147186279297, 0.049509376525878904, 0.04957081604003906, 0.04926259231567383, 0.04953804779052735, 0.04956467056274414, 0.04944998550415039, 0.04948992156982422, 0.049530879974365234, 0.05156351852416992, 0.05317324829101563, 0.052806655883789064, 0.05310464096069336, 0.054540287017822264, 0.053236736297607425, 0.05276979064941406, 0.049713153839111325, 0.0491069450378418, 0.0495810546875, 0.04995481491088867, 0.04960671997070312, 0.049584129333496096, 0.0494376335144043, 0.04950732803344727, 0.04937420654296875, 0.049734657287597656, 0.04969267272949219, 0.04956777572631836, 0.049570785522460935, 0.04955136108398438, 0.04973567962646484, 0.049539070129394534, 0.04963328170776367, 0.04967833709716797, 0.049729534149169925, 0.049688575744628906, 0.04953395080566406, 0.052332542419433595, 0.05268377685546875, 0.052664321899414064, 0.052762622833251956, 0.05271859359741211, 0.04965785598754883, 0.04962303924560547, 0.05341286468505859, 0.05290703964233399, 0.05259465789794922, 0.05336166381835938, 0.05402009582519531, 0.05293056106567383, 0.052969470977783206, 0.052913150787353515, 0.052775936126708986, 0.052340736389160154, 0.052391937255859375, 0.052313087463378906, 0.0523581428527832, 0.05241244888305664, 0.05260796737670898, 0.12193177795410157, 0.049742847442626956, 0.049692737579345704, 0.04952364730834961, 0.049509376525878904, 0.04954111862182617, 0.049393665313720705, 0.04952371215820312, 0.049584129333496096, 0.049600513458251956, 0.049582080841064455, 0.04947148895263672, 0.04946432113647461, 0.05095935821533203, 0.053016574859619144, 0.05274726486206055, 0.052798465728759764, 0.05246876907348633, 0.049756126403808595, 0.04968038558959961, 0.052094974517822266, 0.05160857772827149, 0.05253529739379883, 0.05257932662963867, 0.052519935607910156, 0.05246368026733399, 0.052520896911621096, 0.05939814376831055, 0.05282918548583984, 0.05238784027099609, 0.05234688186645508, 0.05241548919677735, 0.05263872146606445, 0.05251891326904297, 0.05241753768920898, 0.0524400634765625, 0.052915199279785156, 0.05268787384033203, 0.052514816284179686, 0.052580352783203124, 0.05246054458618164, 0.05246361541748047, 0.05234790420532227, 0.05239295959472656, 0.04925747299194336, 0.05096345520019531, 0.052532257080078124, 0.05258134460449219, 0.05242367935180664, 0.05237145614624023, 0.05232332611083984, 0.052380672454833986, 0.052536319732666016, 0.05247590255737305, 0.05253734588623047, 0.05246361541748047, 0.05244825744628906, 0.052357120513916014, 0.05246976089477539, 0.05230185699462891, 0.05226697540283203, 0.052350975036621096, 0.052367359161376956, 0.12207411193847656, 0.04981657409667969, 0.049740798950195314, 0.04961996841430664, 0.04959027099609375, 0.04964352035522461, 0.04969881439208984, 0.04981353759765625, 0.049662944793701175, 0.05032857513427735, 0.04981248092651367, 0.052457473754882813, 0.04957593536376953, 0.04948787307739258, 0.04952371215820312, 0.04986675262451172, 0.04953190231323242, 0.04944998550415039, 0.049438720703125, 0.04950835037231445, 0.04942233657836914, 0.04948070526123047, 0.049600513458251956, 0.04957798385620117, 0.049462272644042966, 0.049495040893554686, 0.04945100784301758, 0.04960358428955078, 0.049414142608642575, 0.049549312591552735, 0.049497150421142576, 0.049874881744384765, 0.04965273666381836, 0.04952371215820312, 0.049486846923828126, 0.04955340957641602, 0.04945305633544922, 0.049547264099121094, 0.04956159973144531, 0.04962614440917969, 0.04954825592041016, 0.04950527954101563, 0.049511425018310545, 0.049503231048583986, 0.04956979370117188, 0.049498111724853515, 0.04944793701171875, 0.04991385650634766, 0.04964761734008789, 0.04951551818847656, 0.04956671905517578, 0.0529090576171875, 0.05213183975219727, 0.05245542526245117, 0.05224038314819336, 0.052590591430664066, 0.049604606628417966, 0.049576961517333984, 0.04991999816894531, 0.049791999816894535, 0.04999782562255859, 0.04991692733764649, 0.049631233215332034, 0.12161126708984375, 0.049600513458251956, 0.0498073616027832, 0.04977766418457031, 0.049582080841064455, 0.04939059066772461, 0.04973056030273437, 0.049719295501708984, 0.049713153839111325, 0.0496363525390625, 0.04959743881225586, 0.04951347351074219, 0.0494919662475586, 0.04945100784301758, 0.04947763061523437, 0.049650688171386716, 0.04972959899902344, 0.04990867233276367, 0.049585151672363284, 0.04953497695922852, 0.05236633682250977, 0.05469491195678711, 0.0527718391418457, 0.052706302642822264, 0.05229161453247071, 0.04960559844970703, 0.051649566650390624, 0.052454368591308594, 0.05246566390991211, 0.05216255950927735, 0.04950732803344727, 0.04968243026733398, 0.04954521560668945, 0.04941107177734375, 0.04964966583251953, 0.049742847442626956, 0.04985651016235351, 0.04976537704467773, 0.04974796676635742, 0.049964031219482424, 0.05108736038208008, 0.05258137512207031, 0.0529356803894043, 0.052944896697998046, 0.052534271240234375, 0.05246464157104492, 0.05250764846801758, 0.052208641052246096, 0.052506622314453126, 0.052547584533691405, 0.05199359893798828, 0.052032512664794923, 0.05243494415283203, 0.05261721420288086, 0.05302169418334961, 0.05401702499389648, 0.05312204742431641, 0.052947967529296876, 0.052926464080810545, 0.05290393447875977, 0.052894718170166014, 0.05305449676513672, 0.05279945755004883, 0.12194303894042968, 0.04976947021484375, 0.04957593536376953, 0.049620990753173826, 0.04975516891479492, 0.04991279983520508, 0.05176422500610352, 0.05237247848510742, 0.05247488021850586, 0.05231001663208008, 0.05236326217651367, 0.05241241455078125, 0.052550655364990234, 0.05254348754882812, 0.05261209487915039, 0.05019340896606445, 0.052787200927734375, 0.052762622833251956, 0.04973158264160156, 0.04930252838134765, 0.04962713623046875, 0.04965478515625, 0.049477695465087894, 0.049627071380615236, 0.04969472122192383, 0.049614849090576174, 0.049552383422851565, 0.04970393753051758, 0.049519615173339845, 0.04952883148193359, 0.049667072296142575, 0.05924966430664062, 0.05293363189697266, 0.0527011833190918, 0.05258444976806641, 0.05290598297119141, 0.05312204742431641, 0.05292031860351563, 0.052722686767578124, 0.05310464096069336, 0.052620288848876956, 0.052582401275634766, 0.05259571075439453, 0.052730911254882815, 0.05249737548828125, 0.05262745666503906, 0.05271756744384765, 0.05287628936767578, 0.05449523162841797, 0.0533831672668457, 0.05295308685302735, 0.05283430480957031, 0.052754432678222656, 0.052877311706542966, 0.05288652801513672, 0.05319270324707031, 0.05282304000854492, 0.05281484985351562, 0.05264281463623047, 0.05270016098022461, 0.052803585052490234, 0.05272883224487305, 0.052566017150878906, 0.12157030487060547, 0.04963225555419922, 0.052506622314453126, 0.052890625, 0.05269606399536133, 0.050800640106201174, 0.05253017425537109, 0.05267865753173828, 0.05251891326904297, 0.05282918548583984, 0.05267148971557617, 0.05241753768920898, 0.052410369873046876, 0.05290502548217774, 0.049600448608398434, 0.05183488082885742, 0.05243801498413086, 0.05250559997558594, 0.052340736389160154, 0.05268076705932617, 0.05265299224853515, 0.04959436798095703, 0.049479679107666014, 0.0495810546875, 0.04951859283447266, 0.05051391983032227, 0.04963840103149414, 0.05029785537719727, 0.05257523345947265, 0.05287324905395508, 0.05248099136352539, 0.0525219841003418, 0.05217279815673828, 0.05221376037597656, 0.05236019134521484, 0.04950732803344727, 0.050272254943847655, 0.052590591430664066, 0.04943564987182617, 0.04949401473999023, 0.04942540740966797, 0.049535999298095705, 0.049511425018310545, 0.04950531387329102, 0.049619937896728514, 0.04952576065063476, 0.049393665313720705, 0.04959539031982422, 0.04957081604003906, 0.04990156936645508, 0.04955852890014648, 0.0494284782409668, 0.05068492889404297, 0.05280460739135742, 0.052574207305908206, 0.052819007873535155, 0.05239603042602539, 0.05237241744995117, 0.05239603042602539, 0.05248921585083008, 0.05264896011352539, 0.052732959747314456, 0.052745185852050784]",tokens/s,19.054167545281807,,, 4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1256.689664,1745.354752,0.0,1098.907648,1041.442816,s,10,0.6699706268310546,0.06699706268310546,0.0003317573314529018,0.0668515510559082,0.06714975357055664,0.06755207633972168,0.06787393455505371,"[0.06795439910888672, 0.06681501007080078, 0.06701734161376953, 0.06706034851074219, 0.06685266876220704, 0.06683049774169922, 0.06698662567138672, 0.06680048370361329, 0.06685043334960937, 0.06680281829833984]",tokens/s,3821.0630398958538,kWh,7.923062565150085e-07,4.33971366208425e-07,4.665207435866996e-06,5.891485058590429e-06,tokens/kWh,43452541.66888262,MB,1256.689664,1745.354752,0.0,1098.907648,1041.445376,s,10,21.635458007812503,2.1635458007812503,0.028737143827813845,2.157843994140625,2.1950017333984375,2.2086248901367185,2.219523415527344,"[2.160812744140625, 2.13723681640625, 2.191974365234375, 2.146611328125, 2.154875244140625, 2.144602294921875, 2.222248046875, 2.179533447265625, 2.179539794921875, 2.11802392578125]",tokens/s,29.118865880838243,kWh,2.5123077208245243e-05,1.3768126637089156e-05,4.6496583493526774e-05,8.53877873388612e-05,tokens/kWh,737810.428908114,,s,629,22.202548202514663,0.035298168843425515,0.007982089678382267,0.03353190231323242,0.035698072052001956,0.03589630813598633,0.10071298095703125,"[0.03492761611938477, 0.034754558563232424, 0.034855934143066404, 0.03446169662475586, 0.03524300765991211, 0.03501465606689453, 0.03721830368041992, 0.03706265640258789, 0.03686604690551758, 0.03769139099121094, 0.03671142578125, 0.03425075149536133, 0.04055039978027344, 0.03762278366088867, 0.03601203155517578, 0.03361894226074219, 0.03357900619506836, 0.03350425720214844, 0.033498111724853515, 0.03351551818847656, 0.033448959350585936, 0.033454078674316406, 0.03351859283447266, 0.033516544342041016, 0.033309696197509765, 0.03351347351074219, 0.03370393753051758, 0.03333324813842774, 0.03341721725463867, 0.03346944046020508, 0.033538047790527346, 0.033455169677734375, 0.03348787307739258, 0.033520576477050784, 0.033452030181884765, 0.03405619049072266, 0.03373158264160156, 0.0334694709777832, 0.033554401397705075, 0.033432575225830076, 0.03331379318237305, 0.0334284782409668, 0.03338547134399414, 0.03341721725463867, 0.033939456939697264, 0.035695648193359374, 0.035856353759765626, 0.03574169540405273, 0.03339263916015625, 0.03349094390869141, 0.03345100784301758, 0.033686527252197264, 0.033807361602783206, 0.03326566314697266, 0.03361280059814453, 0.03389235305786133, 0.033514495849609374, 0.03356982421875, 0.03349910354614258, 0.03346944046020508, 0.033391616821289063, 0.03340185546875, 0.10070118713378906, 0.03346022415161133, 0.03314790344238281, 0.0332677116394043, 0.033290241241455076, 0.03330867385864258, 0.033317886352539065, 0.03304963302612305, 0.033358814239501954, 0.03337318420410156, 0.03349401473999023, 0.033360897064208986, 0.03347148895263672, 0.03338751983642578, 0.03339059066772461, 0.033188865661621096, 0.03307120132446289, 0.033419166564941406, 0.03340697479248047, 0.033442817687988284, 0.03346636962890625, 0.03339059066772461, 0.03340902328491211, 0.03562700653076172, 0.034587646484375, 0.035639297485351565, 0.035525630950927735, 0.035592254638671876, 0.03535763168334961, 0.036942848205566405, 0.03385036849975586, 0.033426433563232424, 0.034748416900634765, 0.03342233657836914, 0.03336294555664063, 0.033328128814697267, 0.03341516876220703, 0.03343155288696289, 0.03405619049072266, 0.03563212966918945, 0.035638431549072265, 0.0333105583190918, 0.0333496322631836, 0.03343155288696289, 0.033445888519287106, 0.03342745590209961, 0.03362713623046875, 0.03345510482788086, 0.033462272644042966, 0.033448959350585936, 0.033527809143066405, 0.03358924865722656, 0.033501182556152344, 0.033380352020263675, 0.033255424499511715, 0.033465343475341795, 0.03337932968139649, 0.03318783950805664, 0.03514265441894531, 0.035705856323242184, 0.035743743896484374, 0.0355860481262207, 0.035833854675292966, 0.10111186981201171, 0.03332601547241211, 0.033320960998535154, 0.03322777557373047, 0.033259521484375, 0.03307724761962891, 0.03330355072021484, 0.034609153747558595, 0.03668787384033203, 0.03571814346313477, 0.035552257537841796, 0.0356126708984375, 0.035383296966552735, 0.03533926391601563, 0.033326080322265625, 0.03450777435302734, 0.035560447692871096, 0.03528192138671875, 0.035383296966552735, 0.035814399719238284, 0.03556358337402344, 0.03547948837280274, 0.0355153923034668, 0.03562188720703125, 0.03550207901000976, 0.03548876953125, 0.035535873413085936, 0.035460094451904296, 0.03546623992919922, 0.03548876953125, 0.0353331184387207, 0.03349606323242187, 0.033326080322265625, 0.03334041595458984, 0.033337345123291014, 0.034541568756103515, 0.03563212966918945, 0.03340697479248047, 0.0334284782409668, 0.03342745590209961, 0.033501182556152344, 0.03340390396118164, 0.03343667221069336, 0.03313971328735352, 0.03392208099365234, 0.03590345764160156, 0.03547340774536133, 0.03579084777832031, 0.035550209045410154, 0.03549184036254883, 0.03551747131347656, 0.03547545623779297, 0.03581744003295898, 0.03562700653076172, 0.03568230438232422, 0.033230846405029296, 0.03584000015258789, 0.03482726287841797, 0.035563552856445316, 0.03546006393432617, 0.03545702362060547, 0.03558195114135742, 0.035585025787353515, 0.10148659515380859, 0.03544575881958008, 0.03547955322265625, 0.03527782440185547, 0.03540582275390625, 0.035662879943847654, 0.03550204849243164, 0.03341823959350586, 0.03511603164672852, 0.033393726348876954, 0.03335673522949219, 0.0332677116394043, 0.03330867385864258, 0.03324316787719726, 0.033262561798095704, 0.033320960998535154, 0.033083393096923826, 0.03333529663085937, 0.033358848571777344, 0.033053695678710936, 0.03323801422119141, 0.03288780975341797, 0.033293312072753906, 0.03338444900512695, 0.033309696197509765, 0.03334348678588867, 0.03339980697631836, 0.03331071853637695, 0.033277950286865234, 0.033306625366210936, 0.03324620819091797, 0.03335987091064453, 0.03321036911010742, 0.033363006591796876, 0.03338134384155273, 0.03361481475830078, 0.033448959350585936, 0.035111934661865234, 0.03602534484863281, 0.03507712173461914, 0.03546112060546875, 0.035659774780273434, 0.03555737686157227, 0.03556147384643555, 0.035602558135986326, 0.03579379272460938, 0.035806209564208984, 0.03552460861206055, 0.03342950439453125, 0.03347251129150391, 0.033377281188964845, 0.03337420654296875, 0.03343564987182617, 0.03334348678588867, 0.03343667221069336, 0.03541401672363281, 0.03561983871459961, 0.03536896133422852, 0.033463294982910154, 0.033339393615722655, 0.03337932968139649, 0.03352268981933594, 0.03342438507080078, 0.10061824035644532, 0.03328409576416016, 0.03476172637939453, 0.03331379318237305, 0.03328105545043945, 0.033218528747558596, 0.03319705581665039, 0.03334348678588867, 0.03341516876220703, 0.0332492790222168, 0.033301502227783206, 0.03333631896972656, 0.03333222579956055, 0.033271808624267575, 0.03564748764038086, 0.03541404724121094, 0.035486686706542966, 0.03534643173217773, 0.03322060775756836, 0.03331584167480469, 0.033279998779296875, 0.03327078247070313, 0.034531326293945314, 0.03341107177734375, 0.03331692886352539, 0.03331679916381836, 0.03329945755004883, 0.033317886352539065, 0.03335270309448242, 0.03326163101196289, 0.03329324722290039, 0.033247230529785156, 0.035422206878662106, 0.03546112060546875, 0.03546623992919922, 0.035743743896484374, 0.03546316909790039, 0.03332198333740234, 0.03489081573486328, 0.03601606369018555, 0.03565260696411133, 0.03543654251098633, 0.035536895751953124, 0.03549593734741211, 0.03448012924194336, 0.03536793518066406, 0.03488051223754883, 0.03418521499633789, 0.03545395278930664, 0.0354856948852539, 0.03554304122924805, 0.03561471939086914, 0.035544063568115236, 0.033339393615722655, 0.03335475158691406, 0.03340083312988281, 0.033307647705078124, 0.033274879455566404, 0.03328102493286133, 0.033326080322265625, 0.03470438385009766, 0.035740673065185545, 0.03331891250610351, 0.10077088165283203, 0.033294273376464845, 0.03333529663085937, 0.03332403182983398, 0.03329228973388672, 0.03333020782470703, 0.03333833694458008, 0.033442817687988284, 0.03325132751464844, 0.033452030181884765, 0.03339468765258789, 0.033364990234375, 0.03344486236572266, 0.03451289749145508, 0.035590145111083986, 0.03514572906494141, 0.0356864013671875, 0.035552257537841796, 0.03555737686157227, 0.03539865493774414, 0.03655680084228516, 0.036364288330078126, 0.03383705520629883, 0.03348582458496094, 0.033430526733398434, 0.03344486236572266, 0.033452030181884765, 0.03341926574707031, 0.03346432113647461, 0.03520716857910156, 0.03345612716674805, 0.033491008758544924, 0.033418209075927734, 0.03345507049560547, 0.033354816436767576, 0.033352638244628904, 0.033432575225830076, 0.033582080841064454, 0.03342438507080078, 0.03354009628295898, 0.033358848571777344, 0.03348275375366211, 0.0333496322631836, 0.03343974304199219, 0.03336601638793945, 0.03338956832885742, 0.033247230529785156, 0.03333529663085937, 0.03338854217529297, 0.03337830352783203, 0.03342233657836914, 0.0333568000793457, 0.03348787307739258, 0.03333631896972656, 0.03353190231323242, 0.035056640625, 0.03565363311767578, 0.035547134399414065, 0.0355788803100586, 0.035334144592285156, 0.03560140609741211, 0.035896320343017575, 0.03614617538452149, 0.10100940704345703, 0.033432575225830076, 0.03553177642822265, 0.03560345458984375, 0.03562393569946289, 0.035776512145996094, 0.03584204864501953, 0.03568435287475586, 0.035767295837402346, 0.03547443389892578, 0.035492862701416016, 0.03551129531860352, 0.03566592025756836, 0.036144126892089845, 0.035563518524169925, 0.03561676788330078, 0.03564134216308594, 0.035533824920654294, 0.035760128021240234, 0.03461734390258789, 0.035827713012695314, 0.03544575881958008, 0.03530547332763672, 0.035487743377685545, 0.03562192153930664, 0.03589628982543945, 0.03566387176513672, 0.03556147384643555, 0.035699710845947266, 0.03625983810424805, 0.03580518341064453, 0.03566387176513672, 0.03353395080566406, 0.03339980697631836, 0.03344998550415039, 0.03319705581665039, 0.033481727600097655, 0.03339263916015625, 0.03489689636230469, 0.03641753768920898, 0.035811328887939455, 0.03562496185302735, 0.03563417434692383, 0.03554816055297851, 0.03559427261352539, 0.035501022338867184, 0.03568742370605469, 0.035573760986328126, 0.03565875244140625, 0.035642368316650394, 0.03502899169921875, 0.03556969451904297, 0.035501022338867184, 0.03567718505859375, 0.03562700653076172, 0.035620864868164064, 0.035567680358886716, 0.035562431335449216, 0.035639297485351565, 0.03352576065063476, 0.03551232147216797, 0.035522560119628906, 0.03346432113647461, 0.10072576141357421, 0.0354969596862793, 0.03568025588989258, 0.03563417434692383, 0.035574783325195314, 0.03567718505859375, 0.035579902648925785, 0.03338444900512695, 0.033463294982910154, 0.03346124649047852, 0.034549758911132815, 0.03573555374145508, 0.035655681610107424, 0.03563827133178711, 0.035506175994873046, 0.03558297729492187, 0.035588096618652344, 0.036219905853271485, 0.03565465545654297, 0.03577241516113281, 0.03572326278686523, 0.03560345458984375, 0.035697662353515625, 0.03567411041259766, 0.035655681610107424, 0.03553484725952148, 0.03570687866210937, 0.03564031982421875, 0.035571712493896485, 0.03345305633544922, 0.03345100784301758, 0.03351039886474609, 0.0357283821105957, 0.033416191101074216, 0.03324313735961914, 0.03377356719970703, 0.03292979049682617, 0.033290241241455076, 0.03339468765258789, 0.03355136108398438, 0.033511425018310545, 0.03337932968139649, 0.033481727600097655, 0.03344998550415039, 0.03338547134399414, 0.03342540740966797, 0.03344793701171875, 0.03346022415161133, 0.03337420654296875, 0.03323289489746094, 0.0334284782409668, 0.0343900146484375, 0.0356577262878418, 0.035566593170166014, 0.0356495361328125, 0.03353190231323242, 0.03350630569458008, 0.03343360137939453, 0.033565696716308595, 0.035705856323242184, 0.0356126708984375, 0.035416065216064455, 0.0356126708984375, 0.10088652801513671, 0.0334202880859375, 0.03332198333740234, 0.03423027038574219, 0.03553996658325195, 0.03549798583984375, 0.03558399963378906, 0.035748863220214845, 0.03550822448730469, 0.03550822448730469, 0.03619123077392578, 0.035929088592529294, 0.03545600128173828, 0.03551027297973633, 0.03555635070800781, 0.03566592025756836, 0.03518873596191406, 0.035503166198730465, 0.033024959564208985, 0.03332505416870117, 0.03338137435913086, 0.03325558471679688, 0.033774433135986326, 0.03563212966918945, 0.03558399963378906, 0.0354785270690918, 0.03546419143676758, 0.03547340774536133, 0.03564134216308594, 0.03551334381103516, 0.033350654602050785, 0.033446910858154294, 0.03525222396850586, 0.035642368316650394, 0.03326976013183594, 0.03338240051269531, 0.03337318420410156, 0.03533110427856445, 0.03550921630859375, 0.035643390655517575, 0.03338649749755859, 0.034890750885009765, 0.035507198333740234, 0.033347583770751955, 0.03336908721923828, 0.0342476806640625, 0.035676158905029294, 0.03531161499023437, 0.033377281188964845, 0.03369779205322266, 0.03554816055297851, 0.035817470550537106, 0.035487743377685545, 0.03556249618530274, 0.03480780792236328, 0.033347583770751955, 0.03327897644042969, 0.03324620819091797, 0.0333383674621582, 0.033331199645996096, 0.033127422332763674, 0.03335782241821289, 0.033355777740478515, 0.10071756744384766, 0.03344793701171875, 0.03344076919555664, 0.03351551818847656, 0.03352883148193359, 0.03343155288696289, 0.033328128814697267, 0.033508350372314456, 0.03346124649047852, 0.03334143829345703, 0.033410049438476565, 0.03324006271362305, 0.03339263916015625, 0.033462272644042966, 0.033363967895507815, 0.033445888519287106, 0.03335168075561523, 0.033486846923828126, 0.03344384002685547, 0.03340800094604492, 0.0334202880859375, 0.03332710266113281, 0.03337420654296875, 0.033452030181884765, 0.03323392105102539, 0.03331686401367188, 0.03351347351074219, 0.03338137435913086, 0.03340800094604492, 0.033391616821289063, 0.03339263916015625, 0.0334571533203125, 0.03340800094604492, 0.033363967895507815, 0.03336908721923828, 0.03337318420410156, 0.03339878463745117, 0.033276927947998046, 0.03330054473876953, 0.03349702453613281, 0.03333631896972656, 0.03341516876220703, 0.033298431396484376, 0.033377281188964845, 0.033274879455566404, 0.03415859222412109, 0.03479449462890625, 0.03467673492431641, 0.03343667221069336, 0.03342950439453125, 0.03362508773803711, 0.03506790542602539, 0.03521023941040039, 0.03476070404052734, 0.034840576171875, 0.034895870208740236, 0.03480166244506836, 0.03470131301879883, 0.03331174468994141, 0.03301375961303711, 0.033307647705078124, 0.0333834228515625, 0.0333834228515625]",tokens/s,28.330081496175282,,, -4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1865,7 +1865,7 @@ If this is a private repository, make sure to pass a token having permission to 4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1429.356544,1598.554112,0.0,952.107008,854.79936,s,10,0.53831734085083,0.053831734085083,0.002254951571002322,0.05407975959777832,0.05562794380187988,0.05734848346710204,0.05872491519927979,"[0.05906902313232422, 0.05372809600830078, 0.05172111892700195, 0.05172230529785156, 0.05472201538085938, 0.05464009475708008, 0.05152070236206055, 0.051516960144042966, 0.055245601654052734, 0.054431423187255856]",tokens/s,4755.559231946397,kWh,6.15175628034329e-07,3.3708533112530853e-07,1.9185938381145913e-06,2.870854797274229e-06,tokens/kWh,89172047.37873283,MB,1429.356544,1598.554112,0.0,952.107008,877.2224,s,10,25.505871826171873,2.5505871826171878,0.028116069009882463,2.5422601318359375,2.5838009033203124,2.6005984497070314,2.6140364868164063,"[2.5606396484375, 2.536609130859375, 2.52938232421875, 2.580068115234375, 2.61739599609375, 2.5553193359375, 2.525668701171875, 2.528643798828125, 2.5479111328125, 2.524233642578125]",tokens/s,24.70019469609149,kWh,3.0264432317646285e-05,1.6586069742802462e-05,4.851686317688493e-05,9.536736523733369e-05,tokens/kWh,660603.339970823,,s,629,25.943555118560788,0.04124571560979458,0.0064890861182490805,0.03989503860473633,0.042374362945556646,0.04255313949584961,0.0920839422607422,"[0.04125900650024414, 0.04211507034301758, 0.04153343963623047, 0.04111974334716797, 0.042240001678466796, 0.041599998474121096, 0.041731071472167966, 0.0414730224609375, 0.042033153533935545, 0.041373695373535156, 0.04116787338256836, 0.04285747146606445, 0.04190924835205078, 0.041414657592773435, 0.041425918579101564, 0.041384960174560545, 0.040025089263916014, 0.03996057510375976, 0.0403056640625, 0.04034867095947266, 0.043480064392089846, 0.04030156707763672, 0.04022169494628906, 0.03991558456420898, 0.03989190292358399, 0.0398837776184082, 0.03992268753051758, 0.03991244888305664, 0.039918590545654296, 0.03982745742797852, 0.03989811325073242, 0.039962623596191404, 0.0396759033203125, 0.04005484771728516, 0.03993183898925781, 0.03983052825927735, 0.039932926177978514, 0.039876609802246096, 0.03996470260620117, 0.03998204803466797, 0.039943168640136716, 0.039880702972412106, 0.03998207855224609, 0.04002406311035156, 0.03998617553710938, 0.03986739349365234, 0.04007526397705078, 0.039863296508789066, 0.042264575958251956, 0.041785343170166016, 0.04133990478515625, 0.039929855346679685, 0.039959552764892575, 0.039787521362304686, 0.04157132720947266, 0.039997440338134765, 0.04130099105834961, 0.04157340621948242, 0.03992982482910156, 0.03995238494873047, 0.040032257080078126, 0.04006707382202149, 0.0940052490234375, 0.03971788787841797, 0.041586688995361325, 0.04148428726196289, 0.04171571350097656, 0.041335807800292966, 0.04132147216796875, 0.040986625671386716, 0.039777278900146484, 0.03990835189819336, 0.03989503860473633, 0.039890945434570314, 0.03964108657836914, 0.0398653450012207, 0.040114273071289064, 0.04143196868896484, 0.041494560241699216, 0.041375713348388674, 0.040371200561523435, 0.042234878540039066, 0.04178739166259766, 0.040079360961914064, 0.040406017303466796, 0.0400076789855957, 0.039907329559326174, 0.039684097290039064, 0.03981414413452149, 0.04132352066040039, 0.04186316680908203, 0.040338432312011716, 0.03991244888305664, 0.03987558364868164, 0.03989811325073242, 0.03994521713256836, 0.039918590545654296, 0.040019966125488284, 0.039636993408203126, 0.03990528106689453, 0.03983462524414062, 0.039926784515380856, 0.03989503860473633, 0.039932926177978514, 0.03989503860473633, 0.039910400390625, 0.03990425491333008, 0.039923713684082034, 0.04003635025024414, 0.03975584030151367, 0.03987756729125977, 0.03985203170776367, 0.03985612869262695, 0.039926784515380856, 0.03992272186279297, 0.03998204803466797, 0.04028006362915039, 0.03990739059448242, 0.039924671173095706, 0.03992268753051758, 0.039755775451660154, 0.03993907165527344, 0.03988787078857422, 0.03993907165527344, 0.039877632141113284, 0.09206988525390625, 0.040153217315673825, 0.04049497604370117, 0.041761791229248044, 0.04114227294921875, 0.039846912384033206, 0.03978956985473633, 0.03987046432495117, 0.03981619262695312, 0.03978854370117187, 0.039820289611816405, 0.03975167846679688, 0.04019200134277344, 0.03993907165527344, 0.03971481704711914, 0.04026367950439453, 0.042229759216308595, 0.04194406509399414, 0.04149964904785156, 0.0399554557800293, 0.040046592712402344, 0.04090265655517578, 0.041491455078125, 0.04133785629272461, 0.041050144195556644, 0.04154262542724609, 0.041368576049804685, 0.04000972747802734, 0.03989724731445313, 0.03975465774536133, 0.03983359909057617, 0.03983865737915039, 0.03989100646972656, 0.03992671966552734, 0.03991654586791992, 0.03990528106689453, 0.03977523040771484, 0.03993600082397461, 0.039847934722900394, 0.03991142272949219, 0.03967385482788086, 0.039777278900146484, 0.03971583938598633, 0.03983871841430664, 0.03970355224609375, 0.039809024810791016, 0.03973017501831055, 0.03974041748046875, 0.03970969772338867, 0.039790592193603515, 0.03971788787841797, 0.039846912384033206, 0.03988787078857422, 0.03996876907348633, 0.039847934722900394, 0.039828479766845705, 0.039779327392578126, 0.03986435317993164, 0.039704544067382816, 0.039760894775390625, 0.039640064239501956, 0.03980915069580078, 0.03974643325805664, 0.09208940887451172, 0.03994822311401367, 0.04165324783325195, 0.04278681564331055, 0.039894016265869144, 0.03986943817138672, 0.03982438278198242, 0.03988172912597656, 0.0396308479309082, 0.03978649520874023, 0.039809024810791016, 0.039798782348632815, 0.039853057861328124, 0.03984384155273438, 0.039975936889648435, 0.03991244888305664, 0.03982745742797852, 0.03984998321533203, 0.03988479995727539, 0.03964416122436523, 0.03977328109741211, 0.03988163375854492, 0.03982131195068359, 0.04015718460083008, 0.039932926177978514, 0.041090049743652345, 0.04196150588989258, 0.042425312042236325, 0.04245094299316406, 0.0423741455078125, 0.04208127975463867, 0.04234035110473633, 0.042188800811767575, 0.03983462524414062, 0.03983052825927735, 0.03994112014770508, 0.04080640029907227, 0.04031078338623047, 0.04196249771118164, 0.042589183807373046, 0.04249702453613281, 0.04226662445068359, 0.042243072509765625, 0.04251443099975586, 0.04236185455322266, 0.042375232696533205, 0.042337215423583985, 0.04292403030395508, 0.04236083221435547, 0.04246323013305664, 0.04218368148803711, 0.04006707382202149, 0.03986022567749024, 0.03984076690673828, 0.03964211273193359, 0.03988787078857422, 0.03982745742797852, 0.040190975189208986, 0.040458240509033204, 0.04232806396484375, 0.04258816146850586, 0.04239974212646484, 0.04267724609375, 0.09750527954101562, 0.042894336700439455, 0.04238336181640625, 0.04238848114013672, 0.04226662445068359, 0.04254207992553711, 0.042294273376464846, 0.042759166717529294, 0.04245708847045898, 0.042436607360839845, 0.04295577621459961, 0.04297625732421875, 0.04256051254272461, 0.042420223236083986, 0.042310657501220705, 0.0397916145324707, 0.03972403335571289, 0.03984281539916992, 0.03969843292236328, 0.04023091125488281, 0.04167270278930664, 0.04262400054931641, 0.042434558868408204, 0.042517505645751956, 0.04235161590576172, 0.04244275283813476, 0.042343425750732425, 0.04261580657958984, 0.04241305541992187, 0.04211609649658203, 0.042446849822998046, 0.042313728332519535, 0.04233932876586914, 0.042452991485595705, 0.04236492919921875, 0.04247859191894531, 0.04234035110473633, 0.04226969528198242, 0.04277350234985351, 0.04266495895385742, 0.04234854507446289, 0.042193920135498046, 0.039798782348632815, 0.03971891021728516, 0.03970355224609375, 0.039793663024902344, 0.03987558364868164, 0.03981619262695312, 0.03975884628295898, 0.03974348831176758, 0.039711742401123046, 0.039725055694580076, 0.03969638442993164, 0.03971788787841797, 0.039800830841064457, 0.03953049468994141, 0.03981926345825195, 0.039977985382080077, 0.04165119934082031, 0.04205055999755859, 0.04246732711791992, 0.04246015930175781, 0.04238848114013672, 0.09493913269042968, 0.03971993637084961, 0.03997491073608399, 0.040051712036132815, 0.03979776000976563, 0.03985612869262695, 0.03971686553955078, 0.04187955093383789, 0.04288409423828125, 0.04252774429321289, 0.04225331115722656, 0.04229836654663086, 0.042076160430908206, 0.04232806396484375, 0.039744510650634765, 0.04253696060180664, 0.042927104949951174, 0.04258099365234375, 0.04219289779663086, 0.04234239959716797, 0.03973843383789062, 0.039862209320068356, 0.03969126510620117, 0.03969740676879883, 0.03973126220703125, 0.03983763122558594, 0.0397496337890625, 0.03971891021728516, 0.03971891021728516, 0.03983257675170898, 0.039779327392578126, 0.03996876907348633, 0.03971891021728516, 0.03965849685668945, 0.039624702453613284, 0.039798782348632815, 0.0397209587097168, 0.03976192092895508, 0.03970150375366211, 0.039754753112792966, 0.04007731246948242, 0.03989606475830078, 0.04212940979003906, 0.042352638244628905, 0.039686145782470705, 0.03972198486328125, 0.03988889694213867, 0.039654399871826174, 0.03977318572998047, 0.039744510650634765, 0.03979673767089844, 0.03982131195068359, 0.03975785446166992, 0.03991446304321289, 0.0399257926940918, 0.03965948867797851, 0.039836673736572265, 0.04205670547485352, 0.042310657501220705, 0.04226047897338867, 0.04223078536987305, 0.042143775939941404, 0.0396973762512207, 0.09212825775146484, 0.03988787078857422, 0.039657505035400394, 0.04330089569091797, 0.04303251266479492, 0.04241100692749023, 0.03984281539916992, 0.03975270462036133, 0.039744510650634765, 0.03966054534912109, 0.039790592193603515, 0.0397946891784668, 0.039806976318359374, 0.039706623077392575, 0.03975987243652344, 0.03953158569335938, 0.039758785247802735, 0.03986431884765625, 0.03992473602294922, 0.03975065612792969, 0.039686145782470705, 0.0397844467163086, 0.03975065612792969, 0.03969638442993164, 0.03972608184814453, 0.03986739349365234, 0.039669761657714846, 0.039946239471435545, 0.041613311767578126, 0.03981721496582031, 0.0398623046875, 0.03979062271118164, 0.03978131103515625, 0.039790592193603515, 0.03965542221069336, 0.03979980850219727, 0.039798782348632815, 0.039572479248046875, 0.03976192092895508, 0.039853057861328124, 0.039815166473388675, 0.039861248016357424, 0.039762943267822266, 0.039684097290039064, 0.039750686645507814, 0.03978543853759765, 0.03968204879760742, 0.03984694290161133, 0.03951203155517578, 0.03973222351074219, 0.03970048141479492, 0.039913471221923826, 0.039918590545654296, 0.03985408020019531, 0.040139774322509765, 0.042816543579101564, 0.0423771858215332, 0.042226688385009765, 0.03980799865722656, 0.039513118743896486, 0.03971990585327148, 0.03978649520874023, 0.03967180633544922, 0.09183232116699219, 0.03976806259155274, 0.03975680160522461, 0.03984384155273438, 0.039918590545654296, 0.039708671569824217, 0.03956121444702149, 0.03977318572998047, 0.03974860763549805, 0.03966259384155273, 0.03981926345825195, 0.04078694534301758, 0.04238848114013672, 0.04241616058349609, 0.042312671661376956, 0.039948287963867186, 0.03978342437744141, 0.039534591674804685, 0.03973222351074219, 0.039656448364257815, 0.039689216613769535, 0.03975987243652344, 0.03971788787841797, 0.03981619262695312, 0.04029030227661133, 0.04252979278564453, 0.039861248016357424, 0.03984384155273438, 0.040013824462890625, 0.039897087097167966, 0.03946700668334961, 0.03974348831176758, 0.03977523040771484, 0.03972710418701172, 0.03974041748046875, 0.039672832489013675, 0.04113612747192383, 0.04235673522949219, 0.042409984588623044, 0.04086476898193359, 0.04203519821166992, 0.03952640151977539, 0.03970764923095703, 0.03969843292236328, 0.039728126525878905, 0.039772159576416014, 0.03976192092895508, 0.039672832489013675, 0.03979267120361328, 0.03962467193603516, 0.03969331359863281, 0.03983257675170898, 0.03966668701171875, 0.03975683212280273, 0.03977724838256836, 0.03966572952270508, 0.03979872131347656, 0.03973427200317383, 0.03973023986816406, 0.03980588912963867, 0.039771137237548826, 0.0407347183227539, 0.04059852981567383, 0.0980316162109375, 0.04196659088134766, 0.04169728088378906, 0.04141366577148437, 0.039808990478515625, 0.04006195068359375, 0.04150067138671875, 0.0413306884765625, 0.03990835189819336, 0.040576000213623044, 0.04159795379638672, 0.04132969665527344, 0.04052886581420898, 0.041611263275146484, 0.04017663955688477, 0.041403392791748046, 0.04139519882202149, 0.04142489624023438, 0.04151193618774414, 0.03978342437744141, 0.041427967071533206, 0.0413573112487793, 0.041283584594726565, 0.04180070495605469, 0.039946239471435545, 0.03980192184448242, 0.03995027160644531, 0.0397844467163086, 0.0406558723449707, 0.04172185516357422, 0.04208025741577148, 0.03975680160522461, 0.039841793060302735, 0.03966566467285156, 0.039823360443115234, 0.039897087097167966, 0.03978956985473633, 0.03986739349365234, 0.039900192260742186, 0.039884769439697265, 0.03975270462036133, 0.03993395233154297, 0.03980799865722656, 0.039863296508789066, 0.03965951919555664, 0.0397946891784668, 0.03980595016479492, 0.04016025543212891, 0.039910400390625, 0.039815166473388675, 0.03983871841430664, 0.039820289611816405, 0.039964672088623046, 0.0397946891784668, 0.0397916145324707, 0.03956531143188476, 0.03986022567749024, 0.039801856994628904, 0.03984896087646484, 0.03979776000976563, 0.039927806854248044, 0.039853057861328124, 0.041455615997314454, 0.09666464233398438, 0.041374656677246095, 0.04113510513305664, 0.04130099105834961, 0.03983257675170898, 0.03989606475830078, 0.04361318588256836, 0.04183865737915039, 0.04197267150878906, 0.03978956985473633, 0.039806976318359374, 0.039841793060302735, 0.039618560791015625, 0.04114022445678711, 0.03982950210571289, 0.03974348831176758, 0.03987865447998047, 0.039702590942382814, 0.03982636642456055, 0.03976396942138672, 0.039782398223876955, 0.039755809783935545, 0.03979977416992188, 0.039698463439941406, 0.039704544067382816, 0.03986636734008789, 0.039772159576416014, 0.039795711517333986, 0.039801856994628904, 0.03985715103149414, 0.039757823944091795, 0.03981824111938476, 0.03993190383911133, 0.04024115371704102, 0.039689216613769535, 0.039779327392578126, 0.0398837776184082, 0.040033279418945314, 0.03981824111938476, 0.03988479995727539, 0.03985919952392578, 0.03990118408203125, 0.039820289611816405, 0.039787521362304686, 0.0398837776184082, 0.039684097290039064, 0.03966873550415039, 0.03974553680419922, 0.03967078399658203, 0.0397496337890625, 0.03966873550415039, 0.039910400390625, 0.03974655914306641, 0.03984384155273438, 0.039815166473388675, 0.03982233428955078, 0.039782398223876955, 0.03995443344116211, 0.040190975189208986, 0.03966054534912109, 0.039777278900146484, 0.039815166473388675, 0.03982131195068359]",tokens/s,24.244942419244413,,, 4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1202.454528,1013.448704,0.0,367.0016,281.15456,s,10,0.5170420455932617,0.05170420455932617,0.0019927500328859337,0.05184947204589844,0.052662210845947265,0.054818768692016594,0.05654401496887207,"[0.05697532653808594, 0.04963488006591797, 0.04988550567626953, 0.04992681503295898, 0.051834590911865236, 0.05218297576904297, 0.052009918212890624, 0.051902305603027346, 0.05082537460327148, 0.05186435317993164]",tokens/s,4951.2414354283665,kWh,5.898392195527629e-07,3.2320095488374595e-07,1.816576909313418e-06,2.7296170837499265e-06,tokens/kWh,93786048.4256309,MB,1202.454528,1013.448704,0.0,367.0016,298.884608,s,10,25.31195751953125,2.531195751953125,0.027080945151457755,2.535255737304688,2.5684645263671873,2.569234338378906,2.569850187988281,"[2.5165126953125, 2.488560546875, 2.500708740234375, 2.545905029296875, 2.56829345703125, 2.570004150390625, 2.529230224609375, 2.5492568359375, 2.54128125, 2.50220458984375]",tokens/s,24.889422302241083,kWh,2.952885702301811e-05,1.618289650339376e-05,4.531898857688704e-05,9.103074210329892e-05,tokens/kWh,692073.8922298306,,s,629,25.731172348022454,0.04090806414629962,0.006223221476213263,0.040169471740722655,0.04110868377685547,0.04131000213623047,0.08985792602539062,"[0.04097443389892578, 0.04092099380493164, 0.04067532730102539, 0.04037836837768555, 0.04136038589477539, 0.04064767837524414, 0.04053504180908203, 0.04054937744140625, 0.041608192443847655, 0.04034969711303711, 0.040497150421142575, 0.04194406509399414, 0.04154777526855469, 0.04120371246337891, 0.04062515258789062, 0.04007321548461914, 0.038499393463134766, 0.039467967987060544, 0.040567806243896484, 0.03928166580200195, 0.03928780746459961, 0.039397377014160156, 0.039370750427246096, 0.03927552032470703, 0.039346176147460936, 0.039283710479736327, 0.040357887268066404, 0.039943168640136716, 0.038932479858398435, 0.03952947235107422, 0.0421847038269043, 0.03991654586791992, 0.03948441696166992, 0.03948646545410156, 0.03939430236816406, 0.03951615905761719, 0.039406593322753904, 0.03911475372314453, 0.03890073776245117, 0.03867647933959961, 0.03882495880126953, 0.03925503921508789, 0.03964211273193359, 0.03979167938232422, 0.03971065521240234, 0.03978956985473633, 0.03966259384155273, 0.03968307113647461, 0.039370750427246096, 0.0390748176574707, 0.03890995025634766, 0.03899596786499023, 0.039617599487304686, 0.03964614486694336, 0.03966054534912109, 0.04105215835571289, 0.04097228622436523, 0.04091904067993164, 0.04106752014160156, 0.03951308822631836, 0.03893350219726562, 0.03877273559570313, 0.0895436782836914, 0.03944550323486328, 0.03952435302734375, 0.03971481704711914, 0.04039168167114258, 0.040839168548583986, 0.040755199432373046, 0.040645633697509766, 0.039316478729248046, 0.038793216705322264, 0.038863872528076174, 0.03873996734619141, 0.0388485107421875, 0.0388587532043457, 0.03888332748413086, 0.039468032836914066, 0.03953664016723633, 0.039406593322753904, 0.03945369720458984, 0.039414783477783204, 0.03939430236816406, 0.039126014709472655, 0.03879635238647461, 0.038766529083251955, 0.03878604888916016, 0.038790145874023435, 0.03884339141845703, 0.039504894256591795, 0.03941068649291992, 0.039462913513183595, 0.03945881652832031, 0.03952025604248047, 0.039277568817138675, 0.03884543991088867, 0.03862732696533203, 0.038811649322509766, 0.03890790557861328, 0.039311359405517575, 0.04003635025024414, 0.04153241729736328, 0.04081151962280274, 0.040978431701660156, 0.03948441696166992, 0.0393256950378418, 0.038949951171875, 0.03863238525390625, 0.038796287536621094, 0.03911171340942383, 0.039620574951171876, 0.039446529388427735, 0.039332862854003905, 0.039408641815185545, 0.04063846588134765, 0.040807422637939454, 0.040583168029785156, 0.04003123092651367, 0.03964723205566406, 0.03955507278442383, 0.03943731307983398, 0.039556095123291016, 0.03941785430908203, 0.03952947235107422, 0.03945676803588867, 0.0897628173828125, 0.039223297119140625, 0.039145503997802734, 0.03871535873413086, 0.03946393585205078, 0.0395335693359375, 0.03964211273193359, 0.03944857788085938, 0.03943526458740235, 0.03952025604248047, 0.03956121444702149, 0.039523326873779296, 0.03931545639038086, 0.038937694549560545, 0.038661022186279294, 0.03890995025634766, 0.03911065673828125, 0.03942297744750976, 0.03948339080810547, 0.039539710998535156, 0.039518207550048826, 0.03949055862426758, 0.04011008071899414, 0.03949158477783203, 0.039163902282714845, 0.038899711608886715, 0.038814720153808595, 0.03906662368774414, 0.039019519805908204, 0.040226814270019534, 0.04003635025024414, 0.03969945526123047, 0.039488510131835936, 0.040392704010009765, 0.04088524627685547, 0.0409620475769043, 0.039723007202148435, 0.03948441696166992, 0.03924889755249023, 0.0401448974609375, 0.0398551025390625, 0.040338432312011716, 0.04078079986572265, 0.0395417594909668, 0.04012748718261719, 0.04076240158081055, 0.04079305648803711, 0.040782848358154294, 0.040681472778320314, 0.03963596725463867, 0.03890892791748047, 0.03962575912475586, 0.03942396926879883, 0.03960115051269531, 0.03949055862426758, 0.03968307113647461, 0.03962777709960937, 0.03951103973388672, 0.04046131134033203, 0.04066304016113281, 0.040199169158935545, 0.03986841583251953, 0.040046592712402344, 0.08989491271972656, 0.03958681488037109, 0.041444351196289066, 0.041183231353759765, 0.0408166389465332, 0.04088115310668945, 0.040354816436767575, 0.040033279418945314, 0.04027494430541992, 0.04094668960571289, 0.042428417205810545, 0.04171980667114258, 0.041090049743652345, 0.04101324844360352, 0.04104294586181641, 0.04104499053955078, 0.04093132781982422, 0.0402872314453125, 0.03892940902709961, 0.03899596786499023, 0.0390041618347168, 0.039002174377441405, 0.03939424133300781, 0.03970560073852539, 0.03962060928344727, 0.03969740676879883, 0.039678974151611326, 0.04041932678222656, 0.04183244705200195, 0.04082694244384766, 0.04070291137695312, 0.039798782348632815, 0.03979673767089844, 0.03960319900512695, 0.039634944915771485, 0.039605247497558595, 0.04269772720336914, 0.04129894256591797, 0.041181182861328124, 0.040427616119384766, 0.03999430465698242, 0.04013872146606445, 0.04005478286743164, 0.04005478286743164, 0.04014694213867188, 0.040185855865478515, 0.04058828735351563, 0.04087091064453125, 0.04093132781982422, 0.040970241546630856, 0.04096614456176758, 0.04090675354003906, 0.04063033676147461, 0.040285118103027345, 0.0388587532043457, 0.03944755172729492, 0.03974863815307617, 0.03973936080932617, 0.03994214248657227, 0.04105420684814453, 0.0409876480102539, 0.041366527557373044, 0.04110131072998047, 0.09274060821533203, 0.04002816009521484, 0.040736766815185545, 0.04066099166870117, 0.041047039031982424, 0.040956928253173826, 0.04108390426635742, 0.04081356811523437, 0.04103577423095703, 0.04096307373046875, 0.040412158966064454, 0.039868446350097654, 0.040812511444091794, 0.041460735321044925, 0.04113100814819336, 0.04103372955322265, 0.04107161712646484, 0.04098867034912109, 0.04110540771484375, 0.04104499053955078, 0.04071219253540039, 0.04029337692260742, 0.04007833480834961, 0.041078784942626956, 0.04112384033203125, 0.04108083343505859, 0.04111872100830078, 0.04101939010620117, 0.04102963256835938, 0.04109209442138672, 0.041009151458740234, 0.04047769546508789, 0.04021145629882812, 0.040136703491210936, 0.04022476959228516, 0.040621055603027346, 0.04110847854614258, 0.041109504699707033, 0.0411596794128418, 0.04108492660522461, 0.041040897369384766, 0.041111553192138675, 0.04038041687011719, 0.04013772964477539, 0.040253440856933595, 0.040190975189208986, 0.04028211212158203, 0.04105625534057617, 0.041111553192138675, 0.04111052703857422, 0.04107571029663086, 0.04114022445678711, 0.040927230834960936, 0.04041318511962891, 0.040068096160888675, 0.0402790412902832, 0.04030054473876953, 0.040215553283691405, 0.040248321533203124, 0.04087705612182617, 0.041204734802246096, 0.041024574279785155, 0.0410049934387207, 0.09393049621582031, 0.04108390426635742, 0.04032819366455078, 0.04015923309326172, 0.04032716751098633, 0.04052380752563477, 0.04114326477050781, 0.04104499053955078, 0.041352191925048826, 0.04147609710693359, 0.041095169067382815, 0.04119756698608398, 0.040812545776367185, 0.04028108978271484, 0.04007219314575195, 0.04013772964477539, 0.040275966644287106, 0.04044902420043945, 0.040932350158691407, 0.04099379348754883, 0.04102044677734375, 0.0410101432800293, 0.04106752014160156, 0.04068044662475586, 0.04120576095581055, 0.040134654998779294, 0.04042956924438477, 0.04101222229003906, 0.04095078277587891, 0.04131737518310547, 0.04108390426635742, 0.041032703399658206, 0.04146585464477539, 0.041171966552734376, 0.040551422119140625, 0.03997183990478516, 0.04030361557006836, 0.040123390197753905, 0.04072447967529297, 0.041040897369384766, 0.0409989128112793, 0.0408985595703125, 0.041009151458740234, 0.04114739227294922, 0.04116889572143555, 0.04042444610595703, 0.03994214248657227, 0.040169471740722655, 0.040242176055908206, 0.04086067199707031, 0.040981502532958985, 0.040892414093017575, 0.04101222229003906, 0.041253887176513675, 0.040871936798095705, 0.04069580841064453, 0.0400948486328125, 0.0407694091796875, 0.041055233001708984, 0.0409989128112793, 0.040853504180908204, 0.04093952178955078, 0.04099071884155273, 0.0934830093383789, 0.04100403213500976, 0.040561664581298826, 0.04012144088745117, 0.04020630264282227, 0.04009983825683594, 0.0400516471862793, 0.040065025329589846, 0.040522815704345704, 0.04116166305541992, 0.0411146240234375, 0.04106649780273437, 0.041188350677490236, 0.04127641677856445, 0.041219070434570314, 0.040973312377929685, 0.04041318511962891, 0.040637439727783206, 0.041692161560058595, 0.042656768798828126, 0.041234432220458986, 0.039656448364257815, 0.04013772964477539, 0.04099071884155273, 0.04094259262084961, 0.04091494369506836, 0.0404326400756836, 0.040102977752685544, 0.040726463317871095, 0.03986841583251953, 0.0396124153137207, 0.0396308479309082, 0.039593982696533206, 0.03948748779296875, 0.03951103973388672, 0.03963391876220703, 0.03963596725463867, 0.039256065368652344, 0.03893964767456055, 0.03884339141845703, 0.03895603179931641, 0.038972415924072266, 0.04047052764892578, 0.04105215835571289, 0.041186302185058594, 0.04051251220703125, 0.03959296035766602, 0.039823360443115234, 0.039346176147460936, 0.0390010871887207, 0.0387327995300293, 0.03908403015136719, 0.03934822463989258, 0.03971891021728516, 0.039580673217773435, 0.039728126525878905, 0.03959807968139648, 0.039755775451660154, 0.03956531143188476, 0.03930931091308594, 0.038863872528076174, 0.039654399871826174, 0.04077056121826172, 0.095098876953125, 0.04260454559326172, 0.04101222229003906, 0.040910846710205076, 0.04077977752685547, 0.040801280975341796, 0.039686145782470705, 0.03862527847290039, 0.03878092956542969, 0.03888025665283203, 0.03885670471191406, 0.03940556716918946, 0.039605247497558595, 0.039547904968261716, 0.03953670501708984, 0.039568321228027344, 0.03990323257446289, 0.041447425842285154, 0.041134078979492186, 0.04087398529052735, 0.040215553283691405, 0.03986841583251953, 0.0400200309753418, 0.03888636779785156, 0.03921404647827149, 0.04169113540649414, 0.04326604843139648, 0.04101836776733398, 0.04120576095581055, 0.04089753723144531, 0.040591358184814456, 0.03983257675170898, 0.03971993637084961, 0.039975936889648435, 0.04011724853515625, 0.03925299072265625, 0.04067942428588867, 0.04077772903442383, 0.04084633636474609, 0.04097536087036133, 0.04073062515258789, 0.040637439727783206, 0.04014591979980469, 0.04007731246948242, 0.040866817474365234, 0.04088627243041992, 0.04087910461425781, 0.04081868743896484, 0.04071731185913086, 0.040869888305664064, 0.04072857666015625, 0.04081356811523437, 0.04069068908691406, 0.03971891021728516, 0.039984127044677735, 0.04077260971069336, 0.04091392135620117, 0.04083200073242187, 0.040768512725830076, 0.04081151962280274, 0.04079513549804688, 0.04077363204956055, 0.040855552673339846, 0.09118720245361328, 0.040318977355957034, 0.04097945785522461, 0.040774654388427735, 0.04084940719604492, 0.040687614440917966, 0.04078387069702148, 0.04069171142578125, 0.040683521270751956, 0.04055862426757813, 0.04007011032104492, 0.039847969055175785, 0.039939041137695315, 0.040395774841308595, 0.04084326553344726, 0.04079411315917969, 0.04084428787231445, 0.04060160064697266, 0.040733695983886715, 0.0408279037475586, 0.04062412643432617, 0.040000511169433595, 0.03898470306396484, 0.0387696647644043, 0.03987558364868164, 0.040120319366455076, 0.04076236724853516, 0.040834049224853515, 0.04075212860107422, 0.04081049728393555, 0.03948441696166992, 0.039741439819335936, 0.04008550262451172, 0.03974348831176758, 0.03990323257446289, 0.03990937423706055, 0.039929855346679685, 0.039910400390625, 0.04002918243408203, 0.0401162223815918, 0.040574977874755856, 0.04083919906616211, 0.04075516891479492, 0.040768512725830076, 0.040854591369628906, 0.04043667221069336, 0.03999334335327148, 0.039962623596191404, 0.03974655914306641, 0.040008705139160154, 0.04051251220703125, 0.040921089172363284, 0.04077772903442383, 0.040791038513183595, 0.040733695983886715, 0.040834049224853515, 0.040253440856933595, 0.03975884628295898, 0.039863296508789066, 0.0399738883972168, 0.0399554557800293, 0.04032921600341797, 0.04071116638183594, 0.09313996887207031, 0.040855552673339846, 0.040766464233398435, 0.04003430557250977, 0.039390209197998044, 0.040068096160888675, 0.040008705139160154, 0.04105830383300781, 0.040812545776367185, 0.040871936798095705, 0.040853504180908204, 0.04092620849609375, 0.041021503448486325, 0.039309249877929685, 0.03959091186523438, 0.04001177597045898, 0.04004761505126953, 0.04047052764892578, 0.04080844879150391, 0.040880126953125, 0.04120883178710937, 0.040888320922851565, 0.040866817474365234, 0.040981502532958985, 0.03926630401611328, 0.03882393646240234, 0.03863868713378906, 0.03880028915405274, 0.03913420867919922, 0.03946905517578125, 0.039465984344482424, 0.03941068649291992, 0.039468032836914066, 0.03947724914550781, 0.03954483032226563, 0.039373825073242184, 0.038899711608886715, 0.03860172653198242, 0.03873081588745117, 0.03911468887329102, 0.03947417449951172, 0.03952947235107422, 0.039387134552001955, 0.039521278381347655, 0.03944755172729492, 0.03945369720458984, 0.03941888046264649, 0.03891814422607422, 0.038569984436035154, 0.038811649322509766, 0.03884134292602539, 0.03885670471191406, 0.039501823425292966, 0.03961139297485351, 0.039465984344482424, 0.039521278381347655, 0.03947212982177734, 0.03948134231567383, 0.039430206298828124, 0.03946387100219727, 0.03934105682373047, 0.03885772705078125, 0.038623233795166016]",tokens/s,24.44505798230142,,, 4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1214.881792,5247.598592,0.0,4601.151488,4334.80704,s,10,3.267789489746094,0.3267789489746094,0.0035483168358907903,0.32566836547851563,0.32736670532226564,0.3323732025146484,0.3363784002685547,"[0.33737969970703124, 0.3258516845703125, 0.32539251708984374, 0.3251649169921875, 0.32531024169921874, 0.3255181884765625, 0.3258467102050781, 0.32525283813476563, 0.326254150390625, 0.32581854248046876]",tokens/s,783.4041966390286,kWh,3.8510935334321854e-06,2.110222071508272e-06,2.452015402470975e-05,3.048146962965021e-05,tokens/kWh,8398545.18533389,MB,1214.881792,5247.598592,0.0,4601.151488,4334.8096,s,10,33.965556640625,3.3965556640625003,0.04749440117816425,3.391852783203125,3.4535903076171874,3.4631746704101563,3.470842160644531,"[3.45146044921875, 3.472759033203125, 3.420389404296875, 3.389158935546875, 3.394546630859375, 3.442658447265625, 3.352416259765625, 3.31369580078125, 3.370570068359375, 3.357901611328125]",tokens/s,18.5482018347516,kWh,3.9719256113506986e-05,2.176858996747532e-05,0.00010147150232409086,0.0001629593484050732,tokens/kWh,386599.48396086437,,s,629,36.84095593643186,0.0585706771644386,0.03867820960308495,0.053100543975830077,0.05589114761352539,0.05641809997558594,0.37857341430664065,"[0.055300094604492187, 0.054234111785888675, 0.054659072875976565, 0.05691392135620117, 0.05697951889038086, 0.055758785247802735, 0.05543526458740235, 0.05571379089355469, 0.05566156768798828, 0.05568716812133789, 0.05569638442993164, 0.05563187026977539, 0.055498752593994144, 0.05557452774047852, 0.052566017150878906, 0.05276671981811523, 0.05560319900512695, 0.055946239471435545, 0.05564521789550781, 0.05556528091430664, 0.05611212921142578, 0.055940097808837894, 0.05656063842773437, 0.055861248016357425, 0.05574553680419922, 0.05571891021728516, 0.05281689453125, 0.05274214553833008, 0.05278822326660156, 0.05266022491455078, 0.05244927978515625, 0.05258137512207031, 0.05256192016601562, 0.052706302642822264, 0.05278822326660156, 0.05296537780761719, 0.05343027114868164, 0.052806655883789064, 0.05204291152954101, 0.05247472000122071, 0.052706432342529294, 0.05603417587280273, 0.05768806457519531, 0.05599641418457031, 0.055690238952636716, 0.052836353302001954, 0.0529172477722168, 0.0528455696105957, 0.05245849609375, 0.05604556655883789, 0.05572608184814453, 0.055769088745117185, 0.055757823944091796, 0.055586814880371094, 0.0563507194519043, 0.05612134552001953, 0.05545574569702148, 0.055684097290039064, 0.055695358276367186, 0.05663641738891602, 0.05572915267944336, 0.055777278900146485, 0.3792537536621094, 0.05259366226196289, 0.05505228805541992, 0.052840511322021486, 0.05606496047973633, 0.055856128692626954, 0.05580492782592773, 0.05567180633544922, 0.05566668701171875, 0.05737472152709961, 0.057040897369384766, 0.055769088745117185, 0.05585408020019531, 0.05595852661132812, 0.055624702453613284, 0.05595852661132812, 0.05557555389404297, 0.05566873550415039, 0.0557209587097168, 0.05554483032226563, 0.05552640151977539, 0.055534591674804686, 0.055787616729736325, 0.05575465774536133, 0.05560019302368164, 0.05554886245727539, 0.05486284637451172, 0.0564316177368164, 0.055600128173828124, 0.055550975799560545, 0.055419902801513675, 0.052803585052490234, 0.052529151916503904, 0.053945343017578126, 0.05564211273193359, 0.05560422515869141, 0.05565030288696289, 0.05549772644042969, 0.05546803283691406, 0.055613441467285155, 0.05564313507080078, 0.05566668701171875, 0.05537279891967774, 0.05552537536621094, 0.05552742385864258, 0.05635891342163086, 0.05552947235107422, 0.055654399871826174, 0.05580083084106445, 0.05566054534912109, 0.05566566467285156, 0.05559807968139648, 0.055669761657714846, 0.0555417594909668, 0.052442142486572266, 0.05273392105102539, 0.052523006439208986, 0.05242265701293945, 0.052572158813476565, 0.052547584533691405, 0.05389929580688477, 0.05557347106933594, 0.055446529388427736, 0.3786106872558594, 0.052354049682617185, 0.05242879867553711, 0.05246464157104492, 0.05244927978515625, 0.052880382537841795, 0.05246774291992187, 0.052547550201416014, 0.05239807891845703, 0.0522762222290039, 0.05253836822509766, 0.052391937255859375, 0.0523612174987793, 0.05238681411743164, 0.05596160125732422, 0.05570969772338867, 0.052544513702392576, 0.05265919876098633, 0.05549260711669922, 0.05607219314575195, 0.055527454376220704, 0.05556732940673828, 0.05329612731933594, 0.055632896423339843, 0.05562572860717773, 0.05559296035766602, 0.05552435302734375, 0.05253734588623047, 0.054591487884521485, 0.05555507278442383, 0.05553670501708984, 0.05556320190429687, 0.05669375991821289, 0.05573017501831055, 0.05563596725463867, 0.05701324844360352, 0.05600665664672851, 0.05550592041015625, 0.05547520065307617, 0.055640064239501956, 0.055462913513183595, 0.05542707061767578, 0.05405081558227539, 0.05551411056518555, 0.055460865020751954, 0.05562060928344727, 0.05550694274902344, 0.055608318328857424, 0.052501502990722655, 0.05248921585083008, 0.05472051239013672, 0.0524400634765625, 0.05249126434326172, 0.05247180938720703, 0.05535232162475586, 0.055641086578369144, 0.053147647857666014, 0.05546803283691406, 0.0556308479309082, 0.05564211273193359, 0.05267865753173828, 0.052592704772949216, 0.05268576049804687, 0.38248037719726563, 0.05272780990600586, 0.052528129577636716, 0.052512767791748044, 0.05253324890136719, 0.052552703857421876, 0.05257113647460938, 0.052393985748291017, 0.05253734588623047, 0.05237247848510742, 0.05260697555541992, 0.05272678375244141, 0.052547584533691405, 0.05331148910522461, 0.052941886901855466, 0.055717823028564456, 0.05567897415161133, 0.05532160186767578, 0.05460483169555664, 0.05629238510131836, 0.05336159896850586, 0.055539710998535156, 0.05257113647460938, 0.05264179229736328, 0.0526110725402832, 0.052462593078613284, 0.05254553604125976, 0.052534271240234375, 0.05255680084228516, 0.052416511535644535, 0.052528129577636716, 0.0526192626953125, 0.05325209426879883, 0.05353267288208008, 0.05246566390991211, 0.05348969650268555, 0.05567382431030273, 0.056443904876708986, 0.05561548614501953, 0.055465984344482425, 0.055787521362304686, 0.05550796890258789, 0.055341056823730465, 0.052529151916503904, 0.052601856231689455, 0.05267865753173828, 0.05259366226196289, 0.052653057098388675, 0.05270732879638672, 0.052523006439208986, 0.053897216796875, 0.055790592193603515, 0.05558272171020508, 0.05574041748046875, 0.05576192092895508, 0.05576297760009766, 0.06015593719482422, 0.0565093765258789, 0.05280153656005859, 0.05269401550292969, 0.05258137512207031, 0.05243596649169922, 0.05244518280029297, 0.3784775695800781, 0.05257318496704102, 0.05247590255737305, 0.052485118865966796, 0.05244211196899414, 0.05507788848876953, 0.05566361618041992, 0.05563715362548828, 0.056770401000976564, 0.05591244888305664, 0.053940223693847655, 0.05559910583496094, 0.05649817657470703, 0.05583052825927735, 0.05552640151977539, 0.05569228744506836, 0.05266739273071289, 0.05254860687255859, 0.0526110725402832, 0.052542465209960934, 0.0525404167175293, 0.05242265701293945, 0.05233868789672851, 0.05255174255371094, 0.05561747360229492, 0.055572479248046876, 0.05551103973388672, 0.05257625579833984, 0.05258342361450195, 0.052523006439208986, 0.052519935607910156, 0.05259980773925781, 0.05219635009765625, 0.052329471588134766, 0.05232537460327148, 0.052519935607910156, 0.05242777633666992, 0.05276671981811523, 0.05409996795654297, 0.05563596725463867, 0.055362560272216796, 0.05542911911010742, 0.056629249572753906, 0.05582438278198242, 0.05552435302734375, 0.05269606399536133, 0.052523006439208986, 0.05264281463623047, 0.05238886260986328, 0.052456478118896484, 0.05256803131103516, 0.05245951843261719, 0.05255680084228516, 0.05269401550292969, 0.05262847900390625, 0.054525951385498046, 0.05570560073852539, 0.05543526458740235, 0.055539710998535156, 0.055638015747070314, 0.05277286529541016, 0.05540454483032226, 0.055803905487060546, 0.37906234741210937, 0.052725696563720705, 0.052670463562011716, 0.05258342361450195, 0.052582401275634766, 0.053100543975830077, 0.05568921661376953, 0.05588582229614258, 0.055774208068847655, 0.05563391876220703, 0.057197566986083984, 0.05605580902099609, 0.05584384155273438, 0.05751295852661133, 0.05628313446044922, 0.055981056213378906, 0.05581619262695312, 0.055752704620361325, 0.056048641204833986, 0.05587353515625, 0.05610700988769531, 0.05582950210571289, 0.055782398223876956, 0.056150016784667967, 0.05602304077148437, 0.055929855346679686, 0.05595238494873047, 0.055877632141113284, 0.05668659210205078, 0.055482368469238284, 0.0526295051574707, 0.05260287857055664, 0.052509727478027346, 0.05247382354736328, 0.0526192626953125, 0.05264691162109375, 0.052499454498291014, 0.0524769287109375, 0.05240422439575195, 0.052640766143798826, 0.052653057098388675, 0.052808704376220705, 0.0526827507019043, 0.052636672973632816, 0.052395008087158204, 0.05274521636962891, 0.05230182266235352, 0.05548953628540039, 0.05547724914550781, 0.055766014099121096, 0.05570764923095703, 0.052613121032714844, 0.052499454498291014, 0.05474611282348633, 0.05524991989135742, 0.055769088745117185, 0.0554700813293457, 0.05639782333374024, 0.05604048156738281, 0.056163295745849606, 0.057729022979736325, 0.05588275146484375, 0.055585792541503906, 0.37909912109375, 0.05600255966186524, 0.05552640151977539, 0.05533695983886719, 0.05566054534912109, 0.052752384185791014, 0.05250559997558594, 0.05276364898681641, 0.05251081466674805, 0.05256489562988281, 0.05272678375244141, 0.05279436874389649, 0.05549363327026367, 0.05556838226318359, 0.055585792541503906, 0.05662822341918945, 0.055826431274414064, 0.055498752593994144, 0.05552742385864258, 0.05543833541870117, 0.052506622314453126, 0.05261209487915039, 0.05257932662963867, 0.05262438583374023, 0.05245849609375, 0.05258342361450195, 0.05251071929931641, 0.05243392181396484, 0.05264691162109375, 0.05260902404785156, 0.052636672973632816, 0.0527011833190918, 0.052450302124023435, 0.05264486312866211, 0.05638655853271484, 0.054452224731445314, 0.05222092819213867, 0.05244313430786133, 0.052193279266357424, 0.05262643051147461, 0.05246566390991211, 0.05245132827758789, 0.05237247848510742, 0.052490238189697266, 0.05231001663208008, 0.05241753768920898, 0.052101119995117184, 0.053324832916259765, 0.05256700897216797, 0.0523612174987793, 0.05247283172607422, 0.052364288330078126, 0.05231513595581055, 0.05243910217285156, 0.0520805778503418, 0.05227315139770508, 0.052499454498291014, 0.05250764846801758, 0.052375553131103515, 0.052511745452880856, 0.052496383666992184, 0.05258956909179688, 0.05243904113769531, 0.37831475830078126, 0.05246054458618164, 0.052278270721435545, 0.05240323257446289, 0.05254140853881836, 0.05254655838012695, 0.05252608108520508, 0.05252608108520508, 0.05246771240234375, 0.05256806564331055, 0.0523765754699707, 0.05254143905639649, 0.05266944122314453, 0.05241241455078125, 0.0524318733215332, 0.052729854583740236, 0.052567039489746094, 0.05253734588623047, 0.0525404167175293, 0.05203046417236328, 0.05258649444580078, 0.053454849243164064, 0.05414297485351562, 0.05286297607421875, 0.05267763137817383, 0.05264281463623047, 0.05255987167358398, 0.052367359161376956, 0.05261721420288086, 0.05236838531494141, 0.05251583862304687, 0.05251686477661133, 0.052528129577636716, 0.05250457763671875, 0.05259571075439453, 0.05269708633422852, 0.052506622314453126, 0.05234995269775391, 0.05248819351196289, 0.052651008605957034, 0.052511745452880856, 0.05263359832763672, 0.052514816284179686, 0.052511745452880856, 0.052498432159423826, 0.05253324890136719, 0.05247385787963867, 0.052775936126708986, 0.05272063827514648, 0.05266841506958008, 0.052542465209960934, 0.05250764846801758, 0.052567039489746094, 0.05261209487915039, 0.052441089630126954, 0.052857856750488284, 0.05248921585083008, 0.05248102569580078, 0.05257523345947265, 0.05267763137817383, 0.05261721420288086, 0.05252608108520508, 0.053008384704589843, 0.37954354858398437, 0.05399552154541016, 0.05383987045288086, 0.054414337158203124, 0.054488063812255856, 0.05287936019897461, 0.0525219841003418, 0.05268582534790039, 0.053661697387695315, 0.054024192810058595, 0.054217727661132815, 0.052620288848876956, 0.053408767700195314, 0.0541214714050293, 0.05240627288818359, 0.052975616455078124, 0.05419417572021484, 0.05247180938720703, 0.05241551971435547, 0.05302780914306641, 0.05266534423828125, 0.05351833724975586, 0.05331763076782227, 0.05436723327636719, 0.053918720245361325, 0.05417267227172851, 0.052523006439208986, 0.05361971282958984, 0.05290188980102539, 0.05314355087280274, 0.05268070220947266, 0.05405491256713867, 0.05295820617675781, 0.0532490234375, 0.05294899368286133, 0.05421670532226563, 0.052340736389160154, 0.052831230163574217, 0.05444716644287109, 0.05442348861694336, 0.05497958374023437, 0.05670809555053711, 0.053951488494873044, 0.0524769287109375, 0.05194649505615234, 0.052590591430664066, 0.052765697479248044, 0.055825408935546876, 0.05435084915161133, 0.05287628936767578, 0.05422182464599609, 0.05264281463623047, 0.05254860687255859, 0.05398015975952149, 0.05281075286865235, 0.0534917106628418, 0.052514816284179686, 0.05335244750976562, 0.0540682258605957, 0.0541921272277832, 0.05562787246704102, 0.05419203186035156, 0.05260902404785156, 0.379261962890625, 0.05339136123657227, 0.05404467010498047, 0.05422284698486328, 0.0540682258605957, 0.054168575286865236, 0.055019519805908204, 0.054232063293457033, 0.05310771179199219, 0.053136383056640625, 0.054139904022216793, 0.05375897598266602, 0.052615169525146485, 0.05253324890136719, 0.05255475234985352, 0.05357670211791992, 0.05721702575683594, 0.05342105484008789, 0.05472256088256836, 0.052725791931152344, 0.055411678314208984, 0.052741119384765625, 0.053250049591064455, 0.05307494354248047, 0.05418598556518555, 0.052776958465576174, 0.052789249420166016, 0.053776382446289066, 0.05424947357177735, 0.053101566314697264, 0.05259980773925781, 0.05290188980102539, 0.052573249816894534, 0.05260591888427734, 0.0524881591796875, 0.053116928100585936, 0.05360128021240235, 0.05386444854736328, 0.0526346549987793, 0.05257212829589844, 0.05351628875732422, 0.05268582534790039, 0.05249753570556641, 0.052568958282470706, 0.05263980865478515, 0.05284038543701172, 0.05410713577270508, 0.05288857650756836, 0.05304422378540039, 0.05255168151855469, 0.05258956909179688, 0.05232230377197265, 0.052719615936279295, 0.05258752059936524, 0.052634624481201174, 0.05341491317749023, 0.054287391662597655, 0.055008289337158206, 0.052591552734375, 0.05232537460327148, 0.05260697555541992, 0.052517887115478515, 0.052618240356445314]",tokens/s,17.073389764514342,,, -4bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2431,7 +2431,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") RuntimeError: Isolated process exited with non-zero code -9 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2714,7 +2714,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1303.953408,14865.137664,0.0,14210.301952,13895.148544,s,10,1.4190867767333983,0.14190867767333984,0.004303490677899202,0.14121475219726565,0.1443467758178711,0.14890522994995117,0.15255199325561525,"[0.15346368408203126, 0.14007859802246095, 0.13893321228027344, 0.1388474884033203, 0.13808953857421874, 0.14333378601074218, 0.14273135375976562, 0.1427906951904297, 0.14235090637207032, 0.13846751403808594]",tokens/s,1803.977066076871,kWh,1.6544308329437987e-06,9.062882622705121e-07,5.129062745222148e-06,7.68978184043646e-06,tokens/kWh,33290931.434989817,MB,1303.953408,14865.137664,0.0,14210.301952,14156.562432,s,10,87.40294921874998,8.740294921874998,0.04671273288724903,8.740237792968749,8.794755859375,8.80359765625,8.81067109375,"[8.792791015625, 8.7309091796875, 8.7116328125, 8.74956640625, 8.7035205078125, 8.812439453125, 8.71580859375, 8.787935546875, 8.7504609375, 8.647884765625]",tokens/s,7.207994760259763,kWh,0.00010446664855895955,5.72529480816926e-05,0.00021741389306677742,0.00037913348970742953,tokens/kWh,166168.38583322187,,s,629,88.517223388672,0.14072690522841316,0.016961712681533493,0.13787440490722655,0.14250925903320313,0.14326845092773438,0.277751923828125,"[0.14591990661621093, 0.1507440643310547, 0.1459220428466797, 0.14518170166015626, 0.14182911682128907, 0.1432995910644531, 0.14274867248535156, 0.1425059814453125, 0.1432299499511719, 0.14488780212402344, 0.14327090454101563, 0.14485920715332032, 0.1422550354003906, 0.1388062744140625, 0.13970329284667968, 0.14023167419433594, 0.1380966339111328, 0.13666610717773436, 0.1386024932861328, 0.13734501647949218, 0.13717811584472656, 0.13633331298828125, 0.1366763458251953, 0.1370408935546875, 0.1370091552734375, 0.13818879699707032, 0.13815811157226562, 0.139334716796875, 0.1387109375, 0.13674803161621094, 0.13736959838867188, 0.13699789428710937, 0.138250244140625, 0.13854106140136718, 0.13770751953125, 0.13822157287597656, 0.13775257873535157, 0.13871002197265625, 0.13841935729980467, 0.14331068420410156, 0.1434245147705078, 0.13857586669921876, 0.13701734924316405, 0.1361623077392578, 0.1361561584472656, 0.13606707763671874, 0.1359605712890625, 0.13637327575683594, 0.1362288360595703, 0.14020608520507813, 0.1376880645751953, 0.13642547607421876, 0.1396459503173828, 0.1367060546875, 0.1379983367919922, 0.13877250671386718, 0.13760610961914063, 0.13782528686523438, 0.13681356811523437, 0.13767578125, 0.1381201934814453, 0.14281829833984375, 0.2773677978515625, 0.1363732452392578, 0.1368290557861328, 0.13752102661132812, 0.13585816955566407, 0.13624114990234376, 0.13731634521484376, 0.13802496337890624, 0.13710540771484375, 0.13816831970214843, 0.1396142120361328, 0.13733786010742188, 0.13729696655273438, 0.13768287658691405, 0.13835877990722656, 0.13711683654785156, 0.13759779357910157, 0.1399500732421875, 0.1368248291015625, 0.136310791015625, 0.13644390869140624, 0.1398271942138672, 0.13964697265625, 0.14173695373535156, 0.13692723083496094, 0.13697535705566405, 0.13739520263671876, 0.13816217041015624, 0.14288589477539063, 0.1395978240966797, 0.1420113983154297, 0.14300775146484376, 0.13986508178710938, 0.14189276123046876, 0.1417009582519531, 0.14293504333496093, 0.1434818572998047, 0.13927935791015625, 0.13991424560546875, 0.1361674499511719, 0.13704803466796875, 0.1370603485107422, 0.13745050048828125, 0.13662208557128908, 0.13761331176757813, 0.13770137023925783, 0.13660569763183594, 0.13610188293457032, 0.1414645690917969, 0.14177900695800782, 0.13941548156738282, 0.13859738159179688, 0.1369108428955078, 0.13653094482421874, 0.13665184020996093, 0.13755897521972657, 0.13800968933105467, 0.13786512756347657, 0.13773823547363281, 0.14345216369628908, 0.13765017700195312, 0.14006271362304687, 0.14240255737304688, 0.27790130615234376, 0.13772288513183595, 0.13748941040039062, 0.13696205139160156, 0.13761843872070312, 0.13724876403808595, 0.13747305297851561, 0.1380771484375, 0.13865267944335938, 0.13724159240722655, 0.1408511962890625, 0.13779661560058593, 0.13777407836914063, 0.13772390747070312, 0.13757235717773436, 0.13803213500976563, 0.13767376708984375, 0.13639369201660156, 0.13750886535644533, 0.13770870971679688, 0.13556105041503907, 0.13758674621582032, 0.1391185302734375, 0.13787750244140626, 0.14007705688476563, 0.13999717712402343, 0.13615827941894532, 0.13773715209960938, 0.13817753601074217, 0.13825843811035157, 0.14264524841308593, 0.1383362579345703, 0.13762457275390624, 0.1374115905761719, 0.13887692260742188, 0.13895884704589845, 0.13658522033691406, 0.13696818542480468, 0.13601075744628907, 0.13641836547851563, 0.1361357116699219, 0.13722000122070313, 0.1370531768798828, 0.13794406127929687, 0.13904598999023438, 0.14032681274414063, 0.1375283203125, 0.14054092407226562, 0.1357854766845703, 0.1356769256591797, 0.13854515075683593, 0.1373922882080078, 0.13656764221191406, 0.13741055297851562, 0.13703167724609375, 0.14017330932617186, 0.14381465148925782, 0.14204415893554687, 0.14266586303710938, 0.14118284606933593, 0.1430312957763672, 0.14095155334472656, 0.13676864624023438, 0.2794064636230469, 0.13988761901855468, 0.1387071075439453, 0.1426082305908203, 0.13757029724121095, 0.13681050109863283, 0.13792973327636718, 0.13646336364746095, 0.13666304016113281, 0.13718016052246093, 0.13650944519042968, 0.137091064453125, 0.13762765502929689, 0.14315110778808593, 0.1377996826171875, 0.13688729858398438, 0.1368606719970703, 0.13788262939453125, 0.1360332794189453, 0.13717401123046874, 0.13610394287109376, 0.13564825439453124, 0.13609983825683594, 0.1376573486328125, 0.13771273803710937, 0.13781494140625, 0.13664358520507813, 0.1365278778076172, 0.13747506713867189, 0.13764002990722657, 0.1380484161376953, 0.1377628173828125, 0.1355919952392578, 0.1373738250732422, 0.13951980590820312, 0.13880422973632814, 0.13897319030761718, 0.14277529907226563, 0.14031564331054688, 0.14326374816894533, 0.1412464599609375, 0.14298316955566406, 0.1393756103515625, 0.1412689971923828, 0.14172773742675782, 0.1408389129638672, 0.13946981811523437, 0.13907865905761718, 0.13965426635742187, 0.1423338165283203, 0.1418270721435547, 0.1440563201904297, 0.14029209899902345, 0.1428326416015625, 0.1414072265625, 0.13760716247558594, 0.13732147216796875, 0.13762969970703126, 0.1378488311767578, 0.13775564575195312, 0.14051327514648437, 0.13807513427734375, 0.1412884521484375, 0.2787891845703125, 0.14149728393554686, 0.13724159240722655, 0.13787033081054687, 0.13697637939453125, 0.1375979461669922, 0.13696940612792968, 0.1378404541015625, 0.13731634521484376, 0.13822976684570312, 0.13733786010742188, 0.144648193359375, 0.14218853759765626, 0.14068428039550782, 0.13770361328125, 0.13728749084472655, 0.14057586669921876, 0.13756710815429687, 0.13622271728515625, 0.14043443298339844, 0.14218035888671876, 0.14099559020996094, 0.13730816650390626, 0.13795123291015626, 0.13842022705078125, 0.13932135009765625, 0.13601382446289062, 0.137533447265625, 0.136998046875, 0.13655638122558594, 0.13713818359375, 0.13742898559570313, 0.1375283203125, 0.13586959838867188, 0.13665673828125, 0.13771160888671874, 0.13751295471191408, 0.1377505340576172, 0.13680863952636718, 0.1375631103515625, 0.13572285461425782, 0.13541696166992187, 0.13530723571777345, 0.13541162109375, 0.13543936157226563, 0.13632920837402343, 0.13749459838867187, 0.14063302612304687, 0.13696922302246095, 0.13742898559570313, 0.1371576385498047, 0.13693644714355468, 0.13892710876464845, 0.1394186248779297, 0.13753753662109375, 0.13826661682128907, 0.13721498107910157, 0.13759706115722656, 0.1395004119873047, 0.1389311981201172, 0.13763481140136719, 0.1439621124267578, 0.14216703796386718, 0.2822441101074219, 0.13755497741699219, 0.13755606079101562, 0.13892092895507813, 0.14105194091796874, 0.13857574462890626, 0.1363681335449219, 0.137027587890625, 0.13898342895507812, 0.14138983154296875, 0.1370061798095703, 0.13853993225097655, 0.13715557861328126, 0.13582237243652343, 0.13538914489746093, 0.13571994018554687, 0.13568205261230468, 0.13672550964355468, 0.13748736572265624, 0.14119833374023438, 0.14181683349609375, 0.14316236877441407, 0.14196432495117187, 0.1428121337890625, 0.14399795532226561, 0.13968077087402345, 0.1395722198486328, 0.13850009155273438, 0.14041191101074219, 0.1407078399658203, 0.14267289733886718, 0.14208921813964845, 0.1393428497314453, 0.14171443176269533, 0.14254386901855468, 0.13946981811523437, 0.13768191528320312, 0.13940736389160155, 0.14149530029296875, 0.14094032287597658, 0.1399480895996094, 0.13665997314453124, 0.14062991333007813, 0.1400126647949219, 0.13895155334472656, 0.14079692077636718, 0.13884931945800782, 0.14146556091308593, 0.14195404052734376, 0.14105087280273437, 0.14262886047363282, 0.14131517028808593, 0.14327389526367187, 0.1404067840576172, 0.14420582580566407, 0.1419141082763672, 0.14139801025390625, 0.14084300231933594, 0.13938175964355468, 0.14063002014160156, 0.14176768493652345, 0.13854421997070313, 0.13876316833496094, 0.2822369384765625, 0.1377587127685547, 0.13734713745117189, 0.13792352294921875, 0.14130892944335938, 0.14072627258300782, 0.13784780883789063, 0.1379471435546875, 0.1371514892578125, 0.13558889770507812, 0.13786006164550782, 0.1424486389160156, 0.1412464599609375, 0.13649305725097657, 0.13695079040527344, 0.1371893768310547, 0.1442027587890625, 0.13810896301269532, 0.1370930938720703, 0.13750271606445313, 0.13759078979492187, 0.13750067138671876, 0.13704605102539064, 0.13695074462890625, 0.1373184051513672, 0.13759385681152345, 0.13765426635742187, 0.14313880920410157, 0.1411215362548828, 0.13895578002929687, 0.13746791076660156, 0.13600665283203126, 0.13847459411621094, 0.13707887268066407, 0.13622256469726562, 0.1363353271484375, 0.13635072326660155, 0.1374167022705078, 0.13548246765136718, 0.13568623352050782, 0.13712979125976563, 0.13921804809570312, 0.14218739318847656, 0.13581004333496094, 0.13725286865234376, 0.1372559356689453, 0.1372446746826172, 0.13790924072265626, 0.13759286499023438, 0.1378795166015625, 0.1394370574951172, 0.14192947387695312, 0.13752217102050782, 0.138176513671875, 0.13972479248046876, 0.13746278381347657, 0.13853497314453125, 0.13926188659667968, 0.13917286682128907, 0.13750579833984375, 0.14155577087402343, 0.14105491638183593, 0.1414635467529297, 0.28619674682617186, 0.13978111267089843, 0.13974322509765624, 0.13723545837402343, 0.13919027709960938, 0.1400647735595703, 0.13714227294921874, 0.13971250915527345, 0.13940019226074218, 0.13821951293945312, 0.13812940979003907, 0.13958348083496094, 0.1415905303955078, 0.14374607849121093, 0.14276502990722656, 0.14258995056152343, 0.14252236938476562, 0.1413201904296875, 0.1425254364013672, 0.1425059814453125, 0.1422009582519531, 0.1406165771484375, 0.14199296569824219, 0.1392926788330078, 0.14494105529785156, 0.143214599609375, 0.14069145202636718, 0.14048153686523437, 0.13947494506835936, 0.13968896484375, 0.13696630859375, 0.1370141143798828, 0.13718630981445312, 0.13730816650390626, 0.1432647705078125, 0.13996237182617188, 0.13929676818847656, 0.13996954345703125, 0.1362872314453125, 0.1367562255859375, 0.13746687316894532, 0.13686579895019532, 0.13806694030761718, 0.13770445251464844, 0.13717607116699218, 0.13785087585449218, 0.13803109741210937, 0.13723033142089844, 0.139399169921875, 0.14039141845703124, 0.1374371795654297, 0.13712588500976564, 0.1375283203125, 0.13702450561523438, 0.13633331298828125, 0.1370777587890625, 0.13929779052734376, 0.1379358673095703, 0.13786335754394533, 0.13783226013183594, 0.1412515869140625, 0.1430804443359375, 0.14021017456054688, 0.28120883178710937, 0.14227250671386718, 0.1414993896484375, 0.13919129943847655, 0.1388062744140625, 0.13942373657226562, 0.13952922058105469, 0.14154342651367188, 0.13901210021972657, 0.1374187469482422, 0.1379246063232422, 0.13622784423828124, 0.13784268188476562, 0.13748736572265624, 0.1374371795654297, 0.13920172119140625, 0.13772172546386718, 0.137376708984375, 0.13687110900878907, 0.13813433837890626, 0.1421281280517578, 0.1435156555175781, 0.13793894958496095, 0.14134169006347655, 0.13749452209472657, 0.13775071716308593, 0.13787440490722655, 0.13807395935058595, 0.13885542297363282, 0.13752114868164061, 0.13638041687011718, 0.13760934448242187, 0.14170303344726562, 0.13675212097167969, 0.13925479125976561, 0.13805874633789061, 0.13569331359863282, 0.13589312744140625, 0.1371155242919922, 0.137806884765625, 0.13889532470703125, 0.1421271057128906, 0.14063104248046876, 0.14199090576171874, 0.13774029541015625, 0.1400647735595703, 0.14305792236328124, 0.13771058654785157, 0.13801266479492189, 0.13784371948242188, 0.1377638397216797, 0.14315827941894532, 0.14059417724609374, 0.1373666229248047, 0.13784771728515624, 0.137762939453125, 0.13954342651367188, 0.13911244201660156, 0.14111538696289064, 0.13837619018554687, 0.14165093994140626, 0.13799320983886718, 0.13760205078125, 0.27648001098632813, 0.13766143798828126, 0.13727845764160157, 0.13720063781738281, 0.13741055297851562, 0.1386395263671875, 0.13755087280273437, 0.13770323181152344, 0.13796354675292968, 0.13789811706542968, 0.1375281524658203, 0.13773516845703124, 0.1377638397216797, 0.1368995819091797, 0.13734194946289063, 0.14268722534179687, 0.13812428283691405, 0.13760205078125, 0.13783961486816407, 0.13758770751953125, 0.13757542419433594, 0.13781607055664064, 0.13793382263183593, 0.14104473876953125, 0.1422274627685547, 0.14168576049804688, 0.13832205200195313, 0.1364048614501953, 0.1395589141845703, 0.14106214904785155, 0.13713714599609375, 0.13594940185546875, 0.13518634033203125, 0.13545677185058594, 0.13569331359863282, 0.135625732421875, 0.1352899475097656, 0.1355293731689453, 0.13509324645996093, 0.13559193420410157, 0.13658111572265624, 0.13711465454101562, 0.13629437255859375, 0.1363017578125, 0.1368041534423828, 0.1368392639160156, 0.13643766784667968, 0.13604454040527345, 0.1365944366455078, 0.13677976989746093, 0.13877452087402345, 0.13843353271484374, 0.13798208618164062, 0.13785586547851564, 0.13603021240234375, 0.13525196838378906, 0.1352478790283203, 0.13637017822265626, 0.1360098876953125, 0.13554159545898437, 0.13530419921875, 0.1354281005859375, 0.1353492431640625]",tokens/s,7.105961709147976,,,True -8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3119,7 +3119,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp5_kqe5mb/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3419,7 +3419,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp_vjhh9wo/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3844,7 +3844,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1883.611136,1453.850624,0.0,799.014912,732.325888,s,10,0.5456827468872071,0.0545682746887207,0.002603940000442501,0.05431547164916992,0.05621721267700195,0.058992221832275385,0.06121222915649414,"[0.06176723098754883, 0.054209888458251955, 0.052554111480712894, 0.05268876647949219, 0.05442105484008789, 0.05264287948608398, 0.05448508834838867, 0.05270828628540039, 0.05560054397583008, 0.054604896545410155]",tokens/s,4691.370607927895,kWh,6.298350386022112e-07,3.451214845542981e-07,1.0450495949041792e-06,2.0200061180606883e-06,tokens/kWh,126732289.42780302,MB,1883.611136,1453.850624,0.0,799.014912,775.103488,s,10,34.065859375,3.4065859375,0.021326349723627,3.4142813720703122,3.4274317138671875,3.430554357910156,3.433052473144531,"[3.41606298828125, 3.433677001953125, 3.40022412109375, 3.412499755859375, 3.390482421875, 3.42195458984375, 3.421415283203125, 3.376487060546875, 3.366318359375, 3.42673779296875]",tokens/s,18.49358893503622,kWh,3.9972726362394104e-05,2.1907063218565998e-05,6.090855552349357e-05,0.00012278834510445367,tokens/kWh,513078.01197586884,,s,629,34.49621915054321,0.05484295572423404,0.006478630529547525,0.05452288055419922,0.05484912643432617,0.05521797027587891,0.10582065307617189,"[0.05363916778564453, 0.053016574859619144, 0.052956161499023435, 0.055152641296386716, 0.05471641540527344, 0.05458227157592774, 0.05455462265014648, 0.05454848098754883, 0.054860801696777345, 0.055795711517333986, 0.05509222412109375, 0.05499903869628906, 0.05428736114501953, 0.054765567779541016, 0.054609920501708986, 0.05461196899414063, 0.0552355842590332, 0.05407231903076172, 0.05327667236328125, 0.0531517448425293, 0.05347942352294922, 0.05429452896118164, 0.05466214370727539, 0.05535232162475586, 0.054803455352783206, 0.054621185302734375, 0.054558719635009766, 0.0544716796875, 0.053789695739746096, 0.054550529479980465, 0.054627361297607424, 0.05475635147094727, 0.052840415954589846, 0.0527749137878418, 0.052660255432128905, 0.0526520004272461, 0.05285171127319336, 0.052721664428710936, 0.05276671981811523, 0.05278822326660156, 0.052604927062988284, 0.05453926467895508, 0.054641727447509766, 0.05463961410522461, 0.05458015823364258, 0.054776832580566405, 0.05392998504638672, 0.054010879516601565, 0.0533309440612793, 0.05421158218383789, 0.054847488403320314, 0.054719486236572266, 0.05502671813964844, 0.05542294311523437, 0.053856254577636715, 0.05466419219970703, 0.05462220764160156, 0.05497753524780274, 0.05477180862426758, 0.05473168182373047, 0.05473177719116211, 0.05462015914916992, 0.10931302642822266, 0.0547061767578125, 0.05429452896118164, 0.05360537719726562, 0.054728702545166014, 0.054599681854248044, 0.05445939254760742, 0.05439078521728516, 0.054768638610839845, 0.054578174591064454, 0.0545269775390625, 0.05453107070922852, 0.054712318420410154, 0.05462015914916992, 0.05453004837036133, 0.05438566589355469, 0.054578174591064454, 0.05449932861328125, 0.054642688751220705, 0.05465190505981445, 0.05455974578857422, 0.055772159576416014, 0.05582233428955078, 0.054847488403320314, 0.05430374526977539, 0.05270732879638672, 0.05264588928222656, 0.05461913681030273, 0.054558719635009766, 0.05481983947753906, 0.05461606216430664, 0.05487820816040039, 0.05448704147338867, 0.05461401748657226, 0.05456076812744141, 0.05447782516479492, 0.054507518768310545, 0.054550529479980465, 0.05465190505981445, 0.05470310211181641, 0.0544532470703125, 0.05410201644897461, 0.0543201904296875, 0.05441632080078125, 0.054566913604736325, 0.05446553421020508, 0.05402316665649414, 0.053789695739746096, 0.0545167350769043, 0.054681598663330076, 0.05453823852539062, 0.054669376373291015, 0.054470592498779294, 0.054024192810058595, 0.054583297729492185, 0.05456281661987305, 0.054555648803710936, 0.05448191833496094, 0.05464780807495117, 0.05456076812744141, 0.054558719635009766, 0.054578174591064454, 0.05452288055419922, 0.10562969970703125, 0.052664321899414064, 0.05281587219238281, 0.0526561279296875, 0.053303295135498044, 0.05283020782470703, 0.052657150268554685, 0.052686847686767575, 0.052279296875, 0.05262745666503906, 0.052689918518066405, 0.052588638305664064, 0.05264374542236328, 0.05430374526977539, 0.056197120666503904, 0.05495808029174805, 0.054712318420410154, 0.05448089599609375, 0.054572032928466796, 0.05453619384765625, 0.054542335510253906, 0.05445939254760742, 0.0548935661315918, 0.054795265197753906, 0.05466729736328125, 0.05454438400268555, 0.05459657669067383, 0.05447577667236328, 0.05458124923706055, 0.05460684967041016, 0.054793216705322265, 0.05449523162841797, 0.05466316986083984, 0.054725631713867184, 0.05448601531982422, 0.05454848098754883, 0.05425664138793945, 0.053513214111328124, 0.05256294250488281, 0.052324352264404295, 0.05257113647460938, 0.0543375358581543, 0.05464371109008789, 0.05450239944458008, 0.05449420928955078, 0.05391974258422851, 0.054452224731445314, 0.05398220825195312, 0.05449523162841797, 0.054525951385498046, 0.05372927856445313, 0.052706302642822264, 0.052787200927734375, 0.052326400756835936, 0.052744190216064454, 0.055120895385742184, 0.05465599822998047, 0.05456281661987305, 0.054781951904296876, 0.05473177719116211, 0.05520076751708984, 0.054695934295654294, 0.05449318313598633, 0.10589491271972656, 0.05317331314086914, 0.05264992141723633, 0.052618240356445314, 0.05274828720092774, 0.05276467132568359, 0.05274009704589844, 0.052789249420166016, 0.05268889617919922, 0.05274214553833008, 0.05258137512207031, 0.05262847900390625, 0.052539390563964845, 0.05255782318115235, 0.05269504165649414, 0.05284044647216797, 0.05259775924682617, 0.05255680084228516, 0.05604556655883789, 0.0548853759765625, 0.054719486236572266, 0.0547327995300293, 0.054593536376953126, 0.054523902893066405, 0.05471641540527344, 0.05479219055175781, 0.05467136001586914, 0.05465599822998047, 0.054543392181396484, 0.05474812698364258, 0.05463046264648438, 0.054561729431152346, 0.05457721710205078, 0.05503071975708008, 0.0551649284362793, 0.054709247589111325, 0.05475430297851563, 0.05448191833496094, 0.054645759582519535, 0.054659072875976565, 0.05459865570068359, 0.05468262481689453, 0.0547696647644043, 0.05469696044921875, 0.05404876708984375, 0.053961727142333986, 0.056630271911621094, 0.05424844741821289, 0.05411840057373047, 0.05474508666992187, 0.05444812774658203, 0.05566463851928711, 0.055229438781738284, 0.054724609375, 0.0546058235168457, 0.05464371109008789, 0.05485567855834961, 0.054588417053222656, 0.054540287017822264, 0.054573089599609374, 0.054555614471435546, 0.054763519287109375, 0.054834175109863284, 0.10957209777832032, 0.05471744155883789, 0.05447372817993164, 0.05456486511230469, 0.054687744140625, 0.054679550170898435, 0.054727680206298826, 0.05451264190673828, 0.054724609375, 0.054740993499755856, 0.05483212661743164, 0.0544901123046875, 0.0545167350769043, 0.054488063812255856, 0.055298046112060545, 0.05463961410522461, 0.05509529495239258, 0.054555648803710936, 0.054621185302734375, 0.05448191833496094, 0.05446451187133789, 0.054586368560791014, 0.05460684967041016, 0.054686721801757814, 0.05526220703125, 0.05471641540527344, 0.05453823852539062, 0.05445939254760742, 0.05451776123046875, 0.054586368560791014, 0.05462220764160156, 0.05464271926879883, 0.0544931526184082, 0.05366579055786133, 0.0544983024597168, 0.05413683319091797, 0.05384806442260742, 0.054709247589111325, 0.05406105422973633, 0.05336576080322265, 0.051888126373291016, 0.05226803207397461, 0.05283327865600586, 0.05266329574584961, 0.05267865753173828, 0.0526192626953125, 0.05234688186645508, 0.052416511535644535, 0.05257625579833984, 0.052526111602783206, 0.05265507125854492, 0.05269913482666016, 0.05253529739379883, 0.052770816802978515, 0.052722686767578124, 0.052132865905761716, 0.05251583862304687, 0.052201473236083984, 0.05236326217651367, 0.052544513702392576, 0.05243392181396484, 0.05288345718383789, 0.052836353302001954, 0.10562764739990234, 0.05262540817260742, 0.05267148971557617, 0.05258752059936524, 0.05264179229736328, 0.05259468841552734, 0.05368832015991211, 0.05526630401611328, 0.05540966415405273, 0.05446451187133789, 0.05446758270263672, 0.054346752166748044, 0.05452288055419922, 0.0538152961730957, 0.054296577453613284, 0.05328998565673828, 0.054438911437988284, 0.05464678573608398, 0.05448089599609375, 0.05440512084960938, 0.05469081497192383, 0.05517311859130859, 0.055155712127685545, 0.054659072875976565, 0.0545976333618164, 0.05470515060424805, 0.05473177719116211, 0.054555648803710936, 0.054766590118408204, 0.05462732696533203, 0.054809600830078124, 0.054563838958740236, 0.054602752685546874, 0.05480243301391602, 0.055354366302490236, 0.05497651290893555, 0.054779903411865234, 0.05453004837036133, 0.05459865570068359, 0.05442355346679688, 0.054456382751464846, 0.05465798568725586, 0.054752254486083986, 0.05455769729614258, 0.05359513473510742, 0.054742015838623044, 0.05471641540527344, 0.054542335510253906, 0.05464371109008789, 0.05457715225219727, 0.05480243301391602, 0.05503180694580078, 0.054333438873291014, 0.05533695983886719, 0.05476761627197266, 0.05264284896850586, 0.052663326263427734, 0.053578689575195314, 0.05449420928955078, 0.05401702499389648, 0.05333913421630859, 0.05427609634399414, 0.05470207977294922, 0.10963148498535157, 0.054697982788085936, 0.0545976333618164, 0.05464780807495117, 0.0547061767578125, 0.05476147079467773, 0.05467647933959961, 0.054981632232666014, 0.054563838958740236, 0.05468364715576172, 0.05445734405517578, 0.05613881683349609, 0.054710208892822264, 0.05471750259399414, 0.05452076721191406, 0.05465497589111328, 0.05489561462402344, 0.05467647933959961, 0.054553600311279295, 0.05483520126342773, 0.05466009521484375, 0.05483321762084961, 0.054682559967041015, 0.05454131317138672, 0.05398220825195312, 0.054320159912109374, 0.054513633728027346, 0.05487923049926758, 0.05454336166381836, 0.054742015838623044, 0.054637569427490235, 0.05451980972290039, 0.05478096008300781, 0.05448905563354492, 0.054594558715820314, 0.05526732635498047, 0.054487102508544924, 0.05454739379882813, 0.05599846267700195, 0.055067649841308595, 0.054542335510253906, 0.054542335510253906, 0.05461606216430664, 0.054834175109863284, 0.054657024383544923, 0.05465292739868164, 0.05458432006835937, 0.05451264190673828, 0.054414337158203124, 0.05461708831787109, 0.05266329574584961, 0.05345792007446289, 0.053146625518798826, 0.052724735260009765, 0.05257318496704102, 0.05267660903930664, 0.052770816802978515, 0.05267763137817383, 0.05255168151855469, 0.052462593078613284, 0.052517887115478515, 0.05269606399536133, 0.05282304000854492, 0.10590822601318359, 0.052776958465576174, 0.052585472106933595, 0.052580352783203124, 0.05257523345947265, 0.05279641723632812, 0.05279436874389649, 0.054599681854248044, 0.055035903930664064, 0.054563838958740236, 0.052798465728759764, 0.05253734588623047, 0.052621311187744144, 0.05249228668212891, 0.05257113647460938, 0.05256806564331055, 0.0526295051574707, 0.05347225570678711, 0.05465190505981445, 0.05444607925415039, 0.05454848098754883, 0.054212608337402345, 0.05226291275024414, 0.05259980773925781, 0.052654079437255856, 0.0526295051574707, 0.052590591430664066, 0.05264998245239258, 0.05257625579833984, 0.05257011032104492, 0.05264998245239258, 0.05269606399536133, 0.05249740982055664, 0.05260083389282227, 0.052703231811523435, 0.05255168151855469, 0.05347635269165039, 0.0544532470703125, 0.05450239944458008, 0.054575103759765625, 0.054392833709716794, 0.05458432006835937, 0.0545689582824707, 0.05442457580566406, 0.054553600311279295, 0.054447105407714844, 0.054527008056640625, 0.053981151580810544, 0.054537216186523435, 0.05445119857788086, 0.0545269775390625, 0.05447987365722656, 0.05444300842285156, 0.0544716796875, 0.05497753524780274, 0.05481676864624024, 0.05473382568359375, 0.05262335968017578, 0.052773887634277344, 0.05430579376220703, 0.05451366424560547, 0.05462835311889649, 0.05449420928955078, 0.11093913269042968, 0.05476454544067383, 0.05466828918457031, 0.05442457580566406, 0.05462732696533203, 0.05452288055419922, 0.05449728012084961, 0.05445632171630859, 0.05507891082763672, 0.05463552093505859, 0.054455295562744144, 0.05248409652709961, 0.052716545104980465, 0.052603904724121096, 0.05268070220947266, 0.05262847900390625, 0.05270425415039062, 0.05265510559082031, 0.05267251205444336, 0.05268787384033203, 0.05273702239990234, 0.051912704467773435, 0.05276467132568359, 0.052670463562011716, 0.052736000061035154, 0.052636672973632816, 0.05271039962768555, 0.05264281463623047, 0.052915199279785156, 0.05271551895141602, 0.05350092697143555, 0.052977664947509766, 0.05269401550292969, 0.052819969177246094, 0.05262438583374023, 0.05271347045898438, 0.052618240356445314, 0.05267967987060547, 0.05212876892089844, 0.05285171127319336, 0.05233356857299805, 0.05228953552246094, 0.052031551361083984, 0.052687808990478514, 0.052689918518066405, 0.05286809539794922, 0.05250867080688477, 0.05555916976928711, 0.05485977554321289, 0.05484134292602539, 0.05461612701416016, 0.054203327178955076, 0.0533309440612793, 0.05378867340087891, 0.053591041564941405, 0.053400577545166014, 0.053800960540771485, 0.054642688751220705, 0.05466214370727539, 0.0545689582824707, 0.054580223083496096, 0.054525951385498046, 0.05461196899414063, 0.10980966186523437, 0.05515980911254883, 0.053141502380371096, 0.05294598388671875, 0.05289875030517578, 0.05460684967041016, 0.05526630401611328, 0.05472256088256836, 0.05456076812744141, 0.052939777374267576, 0.05215027236938476, 0.052673534393310545, 0.054345729827880856, 0.05453414535522461, 0.0545167350769043, 0.05453004837036133, 0.05454131317138672, 0.05462015914916992, 0.0546058235168457, 0.05460076904296875, 0.05474399948120117, 0.054970367431640625, 0.054604801177978515, 0.05420646286010742, 0.054110240936279294, 0.05450953674316406, 0.054545406341552735, 0.054158336639404295, 0.053992446899414064, 0.05471334457397461, 0.05452185440063476, 0.05469491195678711, 0.05486185455322266, 0.055264225006103514, 0.05460070419311523, 0.054537216186523435, 0.05451366424560547, 0.054594558715820314, 0.054332416534423826, 0.05461913681030273, 0.05456793594360351, 0.05461503982543945, 0.054681598663330076, 0.05484134292602539, 0.05452089691162109, 0.054454208374023434, 0.053981182098388675, 0.05386652755737305, 0.05448905563354492, 0.05477888107299805, 0.05444300842285156, 0.05450649642944336, 0.05453619384765625, 0.054542335510253906, 0.054626304626464846, 0.05460172653198242, 0.05398015975952149, 0.05419007873535156, 0.05467446517944336, 0.054416351318359375, 0.05445017623901367, 0.05455462265014648, 0.054768638610839845]",tokens/s,18.233882306203263,,,True -8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4084,7 +4084,7 @@ ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please req ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1808.71168,1156.05504,0.0,501.219328,421.38624,s,10,0.6557844429016113,0.06557844429016113,0.0016121189522829652,0.06561726379394531,0.06635521469116211,0.0679624698638916,0.0692482740020752,"[0.0695697250366211, 0.06572547149658203, 0.06325740814208984, 0.0656346893310547, 0.06348710250854492, 0.065998046875, 0.06559983825683594, 0.06528044891357422, 0.06556521606445312, 0.06566649627685547]",tokens/s,3903.7217605726005,kWh,7.511196445815172e-07,4.1158045541803256e-07,1.186247432822891e-06,2.348947532822441e-06,tokens/kWh,108984980.04865877,MB,1808.71168,1156.05504,0.0,501.219328,446.103552,s,10,40.87055737304687,4.087055737304687,0.03552877711462539,4.07553369140625,4.138192431640625,4.150490747070313,4.160329399414063,"[4.088535400390625, 4.05095263671875, 4.05692822265625, 4.060334228515625, 4.056707275390625, 4.1627890625, 4.0762265625, 4.0748408203125, 4.13545947265625, 4.10778369140625]",tokens/s,15.414519411850977,kWh,4.73587211557833e-05,2.5955100413705194e-05,7.1150199325175e-05,0.00014446402089466349,tokens/kWh,436094.74255141144,,s,629,41.38968462753295,0.06580236029814461,0.007976604406274616,0.06478950500488281,0.0663504867553711,0.0667641860961914,0.13004037841796884,"[0.06647500610351563, 0.06594048309326171, 0.06670642852783203, 0.0657940444946289, 0.06640544128417969, 0.06511097717285157, 0.06726963043212891, 0.06630912017822266, 0.06468812561035156, 0.06544281768798828, 0.0649881591796875, 0.0648642578125, 0.06588006591796874, 0.06432665252685547, 0.06496256256103515, 0.06455398559570312, 0.06496460723876953, 0.06480076599121094, 0.06372966384887695, 0.06439628601074218, 0.06593536376953125, 0.06751744079589844, 0.06517247772216797, 0.0638658561706543, 0.06520028686523438, 0.06555324554443359, 0.0660367660522461, 0.06627529907226562, 0.06627225494384766, 0.06573875427246094, 0.06626918029785156, 0.06558310699462891, 0.06619545745849609, 0.06600601959228515, 0.06492876434326172, 0.06601113891601562, 0.06298419189453125, 0.06373683166503906, 0.06383308792114258, 0.0637132797241211, 0.06276300811767578, 0.06293196868896485, 0.0629309425354004, 0.06337945556640626, 0.06385667037963867, 0.06395183944702149, 0.06389452743530273, 0.06307635116577148, 0.06384230422973633, 0.06393548965454102, 0.06405427551269531, 0.06402355194091797, 0.06359654235839844, 0.0634439697265625, 0.06391807937622071, 0.06398259353637695, 0.0638914566040039, 0.06402764892578125, 0.06408191680908203, 0.06410137939453125, 0.06633164978027344, 0.06668287658691406, 0.13226495361328125, 0.06864383697509766, 0.0682239990234375, 0.06662348937988281, 0.06603981018066406, 0.06596607971191407, 0.06607667541503906, 0.06637158203125, 0.06506803131103515, 0.06397446441650391, 0.0638719367980957, 0.0637470703125, 0.06387302398681641, 0.06536908721923829, 0.06394879913330079, 0.06390579223632813, 0.0639109115600586, 0.06369382476806641, 0.06373273468017578, 0.06322796630859374, 0.06329030227661132, 0.06360883331298828, 0.06372147369384766, 0.06385971069335937, 0.06380854415893554, 0.06394057464599609, 0.06319206237792968, 0.0640184326171875, 0.06390169525146484, 0.06392934417724609, 0.06338252639770507, 0.06395084762573242, 0.06565376281738282, 0.06521241760253907, 0.06612480163574219, 0.06378905487060547, 0.06396928024291992, 0.06390681457519531, 0.06374297714233398, 0.06336000061035156, 0.0635863037109375, 0.06385868835449218, 0.06493901062011719, 0.06605209350585937, 0.06598860931396484, 0.0660664291381836, 0.06334361648559571, 0.06390579223632813, 0.06382592010498046, 0.06373273468017578, 0.0638392333984375, 0.06390681457519531, 0.06295654296875, 0.06260531234741211, 0.06271590423583985, 0.06283366394042969, 0.06332012939453124, 0.06383302307128906, 0.06400819396972657, 0.06387712097167969, 0.06275686264038086, 0.06304358291625976, 0.06282649612426758, 0.12765491485595704, 0.06377164840698242, 0.06407376098632812, 0.06329443359375, 0.06278041458129882, 0.06285209655761718, 0.06543158721923828, 0.06593225860595703, 0.06463488006591797, 0.0638361587524414, 0.06394572830200196, 0.06346444702148438, 0.06323199844360351, 0.06396416091918945, 0.0637594223022461, 0.06376646423339843, 0.06392115020751953, 0.06400614166259766, 0.06361702346801758, 0.06380748748779297, 0.06612480163574219, 0.06605516815185547, 0.06586573028564453, 0.06602444458007813, 0.06591180419921874, 0.06580735778808594, 0.06597017669677735, 0.06595276641845703, 0.06422630310058594, 0.06387923049926758, 0.06385862350463867, 0.06366412734985352, 0.06368153762817383, 0.06372147369384766, 0.06370918273925781, 0.06375423812866211, 0.06393036651611328, 0.06383001708984375, 0.06357913589477539, 0.06380339050292969, 0.06384230422973633, 0.0637655029296875, 0.06393344116210938, 0.06383103942871093, 0.06371839904785156, 0.06355148696899414, 0.06387916946411133, 0.06375116729736328, 0.06379417419433593, 0.06519398498535156, 0.06758604431152344, 0.0640716781616211, 0.06368153762817383, 0.06352691268920899, 0.06363443374633788, 0.06386278533935547, 0.06388121414184571, 0.06588722991943359, 0.06606950378417968, 0.06595276641845703, 0.06572032165527343, 0.06603571319580079, 0.06594252777099609, 0.1321994171142578, 0.06549298858642578, 0.06610636901855468, 0.06590668487548829, 0.06593126678466797, 0.06627430725097656, 0.06602547454833985, 0.06605516815185547, 0.0662456283569336, 0.065627197265625, 0.0636568946838379, 0.0638474235534668, 0.06618418884277344, 0.06595279693603516, 0.06288687896728516, 0.06363750457763671, 0.06335692977905273, 0.06328627014160157, 0.06587391662597657, 0.06594252777099609, 0.06367129516601562, 0.06365695953369141, 0.0636753921508789, 0.0635750389099121, 0.0645027847290039, 0.0667883529663086, 0.0647557144165039, 0.06379929733276367, 0.06379315185546874, 0.06860697937011719, 0.06873395538330078, 0.0663214111328125, 0.06610329437255859, 0.06382185745239258, 0.06380233764648438, 0.06537010955810547, 0.06403699493408203, 0.06367219161987304, 0.06359347152709961, 0.06378291320800782, 0.06386383819580078, 0.06376956939697266, 0.06383718490600586, 0.0637583351135254, 0.06374911880493164, 0.06351257705688476, 0.06287360000610352, 0.0637501449584961, 0.06380134582519531, 0.06379622268676757, 0.06363443374633788, 0.06351667022705078, 0.06266576004028321, 0.06279779052734374, 0.06356787109375, 0.06376755142211914, 0.06393548965454102, 0.06386790466308594, 0.06378598403930665, 0.06286643218994141, 0.0628039665222168, 0.06282444763183594, 0.06272204971313476, 0.12783001708984376, 0.06387420654296876, 0.06265740966796875, 0.0627465934753418, 0.06269030380249023, 0.06373068618774413, 0.06374092864990234, 0.06368767929077149, 0.06360371017456054, 0.06272716903686523, 0.06413107299804688, 0.06379827117919921, 0.06572134399414062, 0.06603263854980469, 0.06595378875732422, 0.06459391784667969, 0.06392012786865234, 0.0637122573852539, 0.06591590118408203, 0.0652759017944336, 0.06600498962402344, 0.06611865234375, 0.06459801483154297, 0.06371942520141602, 0.0627691535949707, 0.06261967849731445, 0.0634552001953125, 0.06431231689453125, 0.06457651519775391, 0.06386790466308594, 0.06384128189086914, 0.06380748748779297, 0.06375116729736328, 0.06383824157714844, 0.0642815704345703, 0.06434815979003906, 0.06391603088378907, 0.06373478317260742, 0.06390169525146484, 0.06378086471557617, 0.06377881622314453, 0.06387200164794922, 0.0636497917175293, 0.06413311767578125, 0.06321152114868164, 0.06381465530395508, 0.06373279953002929, 0.06379308700561523, 0.065723388671875, 0.06371532821655274, 0.06381468963623046, 0.06257251358032226, 0.06630400085449219, 0.06505369567871094, 0.06634291076660156, 0.06622617340087891, 0.06616783905029297, 0.06545712280273437, 0.06599577331542969, 0.06638905334472656, 0.06629984283447266, 0.06604083251953125, 0.06650367736816407, 0.131819580078125, 0.06617081451416015, 0.06610739135742187, 0.06624870300292969, 0.06627021026611328, 0.06614323425292969, 0.06607154846191406, 0.06444236755371094, 0.06461337280273438, 0.0649175033569336, 0.06518374633789062, 0.06532511901855469, 0.06700230407714844, 0.06687232208251953, 0.0659261474609375, 0.06596812438964844, 0.06615961456298829, 0.06648831939697265, 0.06616371154785156, 0.06616169738769531, 0.06632752227783204, 0.06636953735351563, 0.06549298858642578, 0.06629888153076172, 0.06623334503173828, 0.06634188842773438, 0.06623846435546875, 0.06625689697265626, 0.06459391784667969, 0.06456012725830078, 0.06583602905273438, 0.06627737426757813, 0.06631423950195313, 0.06607052612304687, 0.06511923217773438, 0.06618726348876953, 0.06640025329589844, 0.06632755279541015, 0.06647808074951173, 0.06629682922363281, 0.06607161712646484, 0.06473926544189453, 0.06478028869628906, 0.06522470092773437, 0.0663900146484375, 0.06617292785644531, 0.06606950378417968, 0.06661631774902343, 0.06606233978271485, 0.06657843017578124, 0.06643814086914063, 0.06793727874755859, 0.06630092620849609, 0.06613094329833985, 0.06616678619384765, 0.06719078063964844, 0.067525634765625, 0.06655693054199219, 0.06621798706054688, 0.06641158294677735, 0.06627833557128907, 0.06627021026611328, 0.06662553405761719, 0.13253225708007813, 0.0662415008544922, 0.06643302154541016, 0.06621900939941407, 0.0659261474609375, 0.0661585922241211, 0.06614118194580078, 0.06634905242919922, 0.0661944351196289, 0.06482841491699219, 0.06490316772460937, 0.0647383041381836, 0.06463488006591797, 0.06460415649414063, 0.06459391784667969, 0.06473420715332032, 0.06504755401611328, 0.06633984375, 0.06620569610595703, 0.06458470153808593, 0.0644485092163086, 0.06268415832519532, 0.06272512054443359, 0.06281830215454101, 0.06280294418334961, 0.06306918334960937, 0.06307737731933594, 0.06412287902832031, 0.06373683166503906, 0.06379315185546874, 0.06385766220092773, 0.06381875228881836, 0.06361497497558594, 0.06274764633178712, 0.06318694305419922, 0.06378496170043946, 0.06465638732910156, 0.065939453125, 0.06598963165283203, 0.06543257904052735, 0.06275481414794921, 0.06258687973022461, 0.0627589111328125, 0.06263705444335937, 0.06261964797973633, 0.06484480285644531, 0.06616786956787109, 0.06611859130859375, 0.06577254486083985, 0.0659609603881836, 0.06596304321289062, 0.06632035064697266, 0.06618009948730469, 0.06374195098876953, 0.06303948974609375, 0.06268928146362304, 0.06360371017456054, 0.06478950500488281, 0.06601113891601562, 0.0660869140625, 0.06603878021240234, 0.06551654052734375, 0.06597734069824218, 0.13089996337890625, 0.06601625823974609, 0.06581964874267578, 0.064289794921875, 0.06424678039550781, 0.06588006591796874, 0.0659261474609375, 0.06546739196777343, 0.0636119041442871, 0.06598348999023437, 0.06655590057373047, 0.06307430267333984, 0.06406758117675782, 0.06614425659179687, 0.06658150482177734, 0.06394982528686523, 0.06379622268676757, 0.06276710510253906, 0.06260531234741211, 0.06311936187744141, 0.06290329742431641, 0.06282547378540039, 0.06291459274291993, 0.06354224014282227, 0.06307430267333984, 0.06413414764404297, 0.06305484771728516, 0.06265036773681641, 0.06321152114868164, 0.06376140975952148, 0.06377676773071289, 0.06370099258422851, 0.06383820724487305, 0.0635750389099121, 0.06380031967163086, 0.06381158447265625, 0.06360377502441407, 0.06520313262939453, 0.06782361602783203, 0.06647808074951173, 0.06569574737548828, 0.06589542388916016, 0.06591078186035157, 0.06612691497802735, 0.06618617248535157, 0.06605926513671875, 0.06602342224121094, 0.06262169647216796, 0.06363647842407226, 0.06358732986450195, 0.06500556945800781, 0.06587494659423829, 0.06599680328369141, 0.06581247711181641, 0.06388735961914062, 0.06294937515258789, 0.06566092681884765, 0.06612684631347657, 0.06594560241699218, 0.06567219543457031, 0.064574462890625, 0.066334716796875, 0.06607462310791015, 0.1322229766845703, 0.0659559326171875, 0.06698998260498047, 0.06614630126953125, 0.06614015960693359, 0.0662138900756836, 0.06633369445800781, 0.06382694244384765, 0.06436966705322265, 0.06616985321044921, 0.0662343978881836, 0.06618621063232422, 0.06712422180175781, 0.06681804656982422, 0.06616166687011718, 0.0637020149230957, 0.06377471923828125, 0.0642551040649414, 0.06880652618408203, 0.06609305572509766, 0.0659947509765625, 0.06427852630615234, 0.06317158508300781, 0.06366822433471679, 0.06341017532348633, 0.06269132614135742, 0.06339388656616211, 0.06360976028442383, 0.06566297912597656, 0.0662999038696289, 0.06626509094238281, 0.06594457244873046, 0.06650064086914062, 0.06512230682373046, 0.0653567657470703, 0.06575513458251953, 0.06604185485839843, 0.06467603302001954, 0.06296153640747071, 0.06383817672729492, 0.06521036529541016, 0.06594761657714844, 0.0658687973022461, 0.06615756988525391, 0.06584627532958984, 0.065833984375, 0.06607360076904296, 0.06596915435791016, 0.06672486114501953, 0.06636953735351563, 0.06747135925292969, 0.06687129974365234, 0.06696243286132812, 0.06617190551757812, 0.06605721282958985, 0.06605619049072266, 0.0662476806640625, 0.06578380584716798, 0.06672793579101563, 0.06578892517089843, 0.06591283416748046, 0.06683859252929687, 0.06599059295654297, 0.1323714599609375, 0.06605017852783203, 0.06388825607299804, 0.06411468505859375, 0.06296063995361328, 0.06544998168945312, 0.0643399658203125, 0.06294022369384765, 0.06413510131835938, 0.06377881622314453, 0.06354534530639648, 0.06437888336181641, 0.06439836883544922, 0.063868896484375, 0.0638105583190918, 0.06322687911987304, 0.06573567962646484, 0.06596710205078125, 0.06547148895263671, 0.06523699188232422, 0.06448127746582032, 0.06565785980224609, 0.06558003234863281, 0.06586163330078125, 0.06360268783569335, 0.06614835357666016, 0.06699827575683594, 0.06628150177001953, 0.06599164581298828, 0.06567731475830078, 0.06608076477050781, 0.0660674591064453, 0.06603775787353515, 0.06575718688964843, 0.06550630187988281, 0.06302617645263672, 0.06520524597167969, 0.06578482818603516, 0.06434508514404297, 0.06618726348876953, 0.06635622406005859, 0.06434918212890625, 0.06579923248291016, 0.06643910217285157, 0.0655964126586914, 0.06613094329833985, 0.0647014389038086, 0.06580633544921875, 0.06578688049316406, 0.06450176239013672, 0.06430617523193359, 0.06544895935058594, 0.06582886505126953, 0.06639103698730468, 0.06591693115234375, 0.06648831939697265, 0.0654755859375, 0.06568447875976563, 0.06577254486083985, 0.06593023681640625, 0.0659947509765625, 0.06377779388427735, 0.0650588150024414]",tokens/s,15.197023259790218,,,True 8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5080.854528,7927.758848,0.0,7272.923136,7060.84352,s,10,0.8602653045654296,0.08602653045654297,0.0025361049561959228,0.08472780609130859,0.08768762664794921,0.09032965316772461,0.09244327438354492,"[0.0929716796875, 0.08426844787597657, 0.08474726104736328, 0.08670726776123047, 0.08441555023193359, 0.08710050964355469, 0.08439910125732422, 0.08443769836425781, 0.0847083511352539, 0.08650943756103516]",tokens/s,2975.8261624804286,kWh,1.0147501883456963e-06,5.560239239154422e-07,2.7211750164445202e-06,4.291949128705659e-06,tokens/kWh,59646559.71521335,MB,5080.854528,7927.758848,0.0,7272.923136,7224.15616,s,10,53.64037744140626,5.364037744140626,0.02511353395653809,5.358508544921875,5.382963232421875,5.403990405273437,5.420812143554687,"[5.37023388671875, 5.34612939453125, 5.33742626953125, 5.346783203125, 5.3775791015625, 5.343830078125, 5.37829052734375, 5.34442578125, 5.37066162109375, 5.425017578125]",tokens/s,11.744883799301686,kWh,6.422513438355097e-05,3.519960208537062e-05,0.00013451865699755494,0.00023394339346647652,tokens/kWh,269295.91413756995,,s,629,54.31733248901367,0.08635505960097563,0.01029378417205034,0.08457215881347656,0.08698565216064452,0.08747437896728516,0.1694535711669922,"[0.0887910385131836, 0.0874967041015625, 0.08779673767089843, 0.08604876708984376, 0.08760012817382813, 0.08511385345458984, 0.08487935638427735, 0.08468685150146485, 0.08430387115478516, 0.084552734375, 0.08408470153808593, 0.08337920379638672, 0.08380416107177735, 0.08345394897460938, 0.0840447998046875, 0.08320822143554688, 0.08370684814453125, 0.08456396484375, 0.08479641723632812, 0.08475238037109376, 0.0848056640625, 0.0846673583984375, 0.08696627044677735, 0.08729702758789062, 0.08472370910644532, 0.08434175872802735, 0.08479644775390625, 0.08467453002929687, 0.08494284820556641, 0.08478208160400391, 0.08446361541748047, 0.0852520980834961, 0.08500019073486328, 0.08454348754882812, 0.08480461120605469, 0.08428953552246093, 0.08452512359619141, 0.08459161376953125, 0.08577529907226562, 0.08476057434082031, 0.08581427001953125, 0.08782438659667968, 0.08731340789794922, 0.08747519683837891, 0.08732672119140625, 0.08773222351074218, 0.08470323181152344, 0.08627507019042968, 0.08368844604492187, 0.08532275390625, 0.08466432189941406, 0.08470732879638672, 0.08457215881347656, 0.08468380737304687, 0.08470012664794922, 0.08483328247070313, 0.08458342742919922, 0.08427417755126954, 0.08411756896972657, 0.0846990737915039, 0.0845322265625, 0.0847093734741211, 0.1693429718017578, 0.0846919708251953, 0.08443186950683594, 0.08481075286865235, 0.08499302673339844, 0.0843325424194336, 0.08423628997802735, 0.08731549072265625, 0.08740656280517578, 0.08654643249511719, 0.08414822387695313, 0.08826982116699218, 0.08635699462890625, 0.08446873474121094, 0.08433049774169922, 0.08421580505371094, 0.08452607727050782, 0.08448716735839844, 0.08463155364990234, 0.08643385314941407, 0.08746284484863281, 0.08428543853759765, 0.08424143981933593, 0.08640713500976563, 0.08449638366699219, 0.0844134750366211, 0.08404579162597656, 0.08447487640380859, 0.08468991851806641, 0.084316162109375, 0.08420352172851563, 0.08447078704833984, 0.08352460479736327, 0.08428953552246093, 0.0842455062866211, 0.08463155364990234, 0.08463155364990234, 0.08431001281738282, 0.08437149047851562, 0.08444105529785156, 0.08414002990722656, 0.08449024200439453, 0.08370687866210938, 0.08459059143066407, 0.08445132446289062, 0.08438988494873047, 0.08446463775634766, 0.08600780487060547, 0.08529203033447266, 0.08448409271240234, 0.087910400390625, 0.08655564880371094, 0.08440422058105469, 0.08447180938720703, 0.08421068572998047, 0.08463155364990234, 0.08440525054931641, 0.08426092529296875, 0.08422291564941406, 0.08452403259277344, 0.08405709075927735, 0.08436640167236328, 0.08445331573486328, 0.16951298522949218, 0.08455677032470703, 0.08439705657958985, 0.08411443328857422, 0.08448307037353515, 0.0843130874633789, 0.08614604949951171, 0.08464383697509766, 0.08430592346191407, 0.08448512268066406, 0.08423014068603515, 0.08450559997558593, 0.08418508911132813, 0.08677996826171876, 0.08428844451904297, 0.08391168212890625, 0.0839014434814453, 0.08454144287109375, 0.08425472259521484, 0.08450355529785156, 0.08425676727294922, 0.0843468780517578, 0.08680038452148438, 0.0868505630493164, 0.08427008056640625, 0.08447289276123048, 0.08418911743164062, 0.08447590637207031, 0.08380723571777343, 0.08619519805908203, 0.08443084716796875, 0.08457421112060547, 0.08475955200195312, 0.0844390411376953, 0.0843468780517578, 0.08468685150146485, 0.08441958618164062, 0.08440013122558594, 0.08482819366455079, 0.08458338928222656, 0.08425472259521484, 0.08437554931640626, 0.08550093078613281, 0.08701747131347656, 0.08452607727050782, 0.08368742370605468, 0.08446669006347657, 0.08440525054931641, 0.0839925765991211, 0.08394547271728516, 0.08449440002441407, 0.08449836730957032, 0.08441241455078125, 0.0850094757080078, 0.0864234848022461, 0.08449433898925782, 0.08454348754882812, 0.08461312103271484, 0.08445337677001953, 0.08442880249023438, 0.08428953552246093, 0.08747315216064454, 0.08568934631347656, 0.17373799133300782, 0.08697650909423828, 0.08698060607910156, 0.08464691162109375, 0.08418406677246094, 0.0842086410522461, 0.08439910125732422, 0.08439807891845703, 0.08427212524414063, 0.0862033920288086, 0.08444723510742187, 0.08414720153808594, 0.0845660171508789, 0.0853391342163086, 0.08661094665527344, 0.08463257598876953, 0.08424755096435547, 0.08450662231445312, 0.08503091430664063, 0.08479334259033203, 0.0844400634765625, 0.08608972930908203, 0.08449433898925782, 0.08428749084472656, 0.08555827331542969, 0.08610201263427734, 0.08555315399169922, 0.08432640075683594, 0.08432537841796875, 0.08466534423828125, 0.08454144287109375, 0.08437554931640626, 0.08434381103515624, 0.08430694580078125, 0.0842086410522461, 0.08424960327148437, 0.08497869110107421, 0.08424755096435547, 0.08461516571044922, 0.08451277160644531, 0.08753971099853515, 0.08475750732421874, 0.08603648376464844, 0.0842608642578125, 0.0843499526977539, 0.08426802825927734, 0.08419123077392578, 0.08419430541992187, 0.0843345947265625, 0.08449740600585938, 0.08442675018310547, 0.08428134155273438, 0.08445849609375, 0.0856463394165039, 0.08487628936767579, 0.08445951843261719, 0.08416665649414062, 0.08425472259521484, 0.08414217376708985, 0.08416860961914062, 0.08704000091552734, 0.08639692687988282, 0.08425267028808593, 0.16917709350585938, 0.08591769409179688, 0.08596479797363281, 0.08453427124023437, 0.08435814666748047, 0.08520806121826172, 0.08494284820556641, 0.08447487640380859, 0.08437350463867188, 0.08441036987304687, 0.08436640167236328, 0.08445331573486328, 0.08462131500244141, 0.08456703948974609, 0.08505241394042969, 0.0865771484375, 0.08424140930175782, 0.08388505554199219, 0.08556134033203125, 0.08651570892333985, 0.08540876770019531, 0.08605398559570313, 0.08698563385009765, 0.08720076751708984, 0.08475238037109376, 0.08444931030273438, 0.08445436859130859, 0.08442572784423828, 0.084495361328125, 0.08439705657958985, 0.08418816375732421, 0.08419020843505859, 0.08398745727539063, 0.08673894500732422, 0.08669900512695312, 0.08741484832763671, 0.08733894348144532, 0.0863846435546875, 0.08430284881591797, 0.08443289947509766, 0.08521932983398438, 0.08450150299072266, 0.08376428985595703, 0.08443488311767579, 0.08565555572509766, 0.08704204559326172, 0.08431513977050781, 0.08441241455078125, 0.08677481842041015, 0.08713008117675781, 0.0844615707397461, 0.08833535766601562, 0.08720588684082031, 0.08686489868164063, 0.08670105743408203, 0.08830156707763671, 0.08466329956054687, 0.08437760162353515, 0.08672767639160156, 0.0860231704711914, 0.0844615707397461, 0.08428134155273438, 0.08479948425292969, 0.17416806030273438, 0.0869713897705078, 0.08536780548095703, 0.08551219177246094, 0.084569091796875, 0.08421376037597657, 0.08439603424072266, 0.08691609954833984, 0.08447487640380859, 0.08457017517089843, 0.08480352020263672, 0.0845107192993164, 0.08435097503662109, 0.08695193481445312, 0.08678604888916015, 0.08448102569580078, 0.0856657943725586, 0.08437657928466796, 0.08425984191894531, 0.08475341033935548, 0.0838707504272461, 0.08407548522949218, 0.08431513977050781, 0.08435097503662109, 0.08438169860839843, 0.08448102569580078, 0.0853934097290039, 0.08368230438232421, 0.0842987823486328, 0.08434172821044922, 0.08750796508789062, 0.08506777954101563, 0.08410316467285156, 0.08448204803466797, 0.08371507263183593, 0.08448102569580078, 0.08432434844970703, 0.0840110092163086, 0.08406425476074218, 0.08588288116455078, 0.08428646087646484, 0.08457421112060547, 0.08426191711425782, 0.08681673431396485, 0.08428134155273438, 0.08440217590332032, 0.08434278106689454, 0.08619026947021484, 0.08341587066650391, 0.08537702178955078, 0.08439705657958985, 0.08442675018310547, 0.08432434844970703, 0.08420146942138672, 0.08438374328613281, 0.08493260955810547, 0.087119873046875, 0.08462847900390626, 0.08418303680419922, 0.08432537841796875, 0.08432844543457031, 0.0843694076538086, 0.08450867462158203, 0.16949658203125, 0.08441139221191406, 0.08437350463867188, 0.0844185562133789, 0.08415948486328124, 0.08431922912597656, 0.08414412689208985, 0.08444115447998046, 0.08700819396972656, 0.08555830383300782, 0.08441545867919922, 0.08663654327392578, 0.08463565063476562, 0.08444620513916015, 0.08434073638916016, 0.08452301025390625, 0.08437964630126953, 0.08504934692382812, 0.08464486694335938, 0.0854466552734375, 0.08517938995361328, 0.08565042877197265, 0.08716287994384765, 0.0859310073852539, 0.08557670593261718, 0.08746189117431641, 0.08470118713378906, 0.08449638366699219, 0.08458137512207031, 0.08655059051513672, 0.0852581787109375, 0.0862033920288086, 0.08450969696044922, 0.08508108520507812, 0.08606617736816406, 0.08696217346191407, 0.08549581146240234, 0.08659967803955078, 0.0895129623413086, 0.08702259063720703, 0.08899788665771484, 0.0860579833984375, 0.08444313812255859, 0.08444927978515625, 0.08687513732910156, 0.08636109161376954, 0.08433561706542969, 0.0842946548461914, 0.08449228668212891, 0.08514969635009766, 0.08465305328369141, 0.0846192626953125, 0.08565042877197265, 0.08535346984863282, 0.08459059143066407, 0.0844933090209961, 0.0854824981689453, 0.08594432067871094, 0.08414617919921875, 0.08424447631835938, 0.08569344329833985, 0.08707276916503906, 0.08450457763671874, 0.1707509765625, 0.08539852905273437, 0.08444416046142578, 0.08491212463378907, 0.08615731048583984, 0.08502374267578125, 0.08677581024169922, 0.08456294250488282, 0.08636518096923829, 0.08442982482910157, 0.0842260513305664, 0.08445951843261719, 0.08464998626708985, 0.08599449920654297, 0.08399052429199219, 0.08418102264404297, 0.08421372985839844, 0.08464179229736328, 0.0846714859008789, 0.08444927978515625, 0.08919347381591797, 0.08483123016357422, 0.08437042999267579, 0.08449740600585938, 0.08637439727783203, 0.08696422576904297, 0.08431718444824218, 0.08420556640625, 0.08429261016845703, 0.08435609436035156, 0.08457628631591797, 0.08444822692871094, 0.08485683441162109, 0.084279296875, 0.08430284881591797, 0.08450559997558593, 0.084347900390625, 0.08384819030761718, 0.0839731216430664, 0.08436121368408203, 0.0843499526977539, 0.08408576202392579, 0.08471756744384766, 0.08442572784423828, 0.08523980712890625, 0.08464691162109375, 0.08448921966552735, 0.08452198028564453, 0.08434585571289062, 0.08386252593994141, 0.08443289947509766, 0.08413798522949219, 0.08454144287109375, 0.08449945831298829, 0.08412159729003907, 0.08694374084472656, 0.0868076171875, 0.0843939208984375, 0.08431721496582031, 0.08441136169433594, 0.08444313812255859, 0.08431410980224609, 0.08559308624267578, 0.17149337768554687, 0.08675122833251953, 0.08436531066894531, 0.0842239990234375, 0.08473190307617187, 0.08543231964111328, 0.08451481628417969, 0.08446873474121094, 0.08421376037597657, 0.08467967987060547, 0.08418303680419922, 0.08445235443115234, 0.08440115356445313, 0.0844554214477539, 0.08520191955566406, 0.08485887908935547, 0.08418819427490234, 0.08437551879882813, 0.08422707366943359, 0.08439199829101562, 0.08411129760742188, 0.0836157455444336, 0.08414412689208985, 0.08456192016601563, 0.08438784027099609, 0.08462438201904297, 0.08500838470458984, 0.08472882843017578, 0.08454656219482422, 0.08468070220947266, 0.08762572479248047, 0.08477798461914063, 0.08412979125976562, 0.0843663330078125, 0.08642765045166016, 0.08730009460449219, 0.08429977416992188, 0.08445030212402344, 0.08415545654296876, 0.08668972778320312, 0.08602931213378906, 0.08559616088867188, 0.085274658203125, 0.0845250244140625, 0.08619417572021484, 0.08754176330566406, 0.08730521392822266, 0.0871014404296875, 0.08410214233398437, 0.08429875183105469, 0.08392396545410157, 0.08654950714111329, 0.08695807647705078, 0.08679116821289062, 0.08776089477539062, 0.08659865570068359, 0.0849459228515625, 0.08680038452148438, 0.08509030151367188, 0.08599756622314453, 0.08622489929199219, 0.08581222534179687, 0.08574156951904296, 0.1734819793701172, 0.08658329772949219, 0.08708403015136719, 0.08722124481201173, 0.08616242980957031, 0.08642662048339844, 0.08513228607177735, 0.08639488220214844, 0.08455168151855469, 0.08803839874267579, 0.0892221450805664, 0.08678092956542968, 0.0852655029296875, 0.0853380126953125, 0.08673894500732422, 0.08706150054931641, 0.08644915008544922, 0.08673689270019531, 0.08636627197265626, 0.08598316955566407, 0.0853780517578125, 0.08668262481689454, 0.08578662109375, 0.08685465240478515, 0.08570880126953125, 0.08561357116699218, 0.0835041275024414, 0.08297379302978515, 0.08333712005615235, 0.0866355209350586, 0.08714342498779297, 0.08611225891113282, 0.08581324768066406, 0.08675020599365234, 0.08588082885742188, 0.0868823013305664, 0.08579379272460938, 0.08552345275878906, 0.0838318099975586, 0.08332492828369141, 0.08361881256103515, 0.08655878448486329, 0.08584799957275391, 0.08591155242919922, 0.08685465240478515, 0.08651673889160157, 0.08710758209228515, 0.08700723266601562, 0.08584806060791016, 0.08690380859375, 0.08668672180175781, 0.08698572540283203, 0.08597401428222656, 0.08868659210205078, 0.08886374664306641, 0.08608358764648437, 0.08676044464111328, 0.08451789093017578, 0.08377241516113282, 0.08544255828857422, 0.0859259490966797, 0.08663033294677734, 0.08651776123046875]",tokens/s,11.580097386542736,,,True -8bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4448,7 +4448,7 @@ OSError: 8 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4576,7 +4576,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4644,7 +4644,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") RuntimeError: Isolated process exited with non-zero code -9 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5071,7 +5071,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1278.656512,14865.137664,0.0,14210.301952,13895.148544,s,10,1.3928687133789064,0.13928687133789064,0.0048535291321430245,0.13745452880859374,0.142009619140625,0.1476316253662109,0.1521292303466797,"[0.15325363159179686, 0.13677174377441406, 0.13689894104003905, 0.13689718627929687, 0.13801011657714843, 0.14076028442382812, 0.13649462890625, 0.13609942626953125, 0.13939091491699218, 0.13829183959960936]",tokens/s,1837.9334501597032,kWh,1.6569359618940472e-06,9.075599060148913e-07,5.1513852939442396e-06,7.715881161853178e-06,tokens/kWh,33178323.334688928,MB,1278.656512,14865.137664,0.0,14210.301952,14156.562432,s,10,86.32789160156251,8.632789160156252,0.07122559729372871,8.61228759765625,8.687139453124999,8.759019921875,8.816524296875,"[8.830900390625, 8.5951943359375, 8.6222919921875, 8.602283203125, 8.6282822265625, 8.5773349609375, 8.576921875, 8.596681640625, 8.671166015625, 8.6268349609375]",tokens/s,7.29775728692298,kWh,0.0001028236737828932,5.635546536206068e-05,0.00021441406813605697,0.0003735932072810109,tokens/kWh,168632.6163650304,,s,629,87.41687489318848,0.13897754355037914,0.016552903379217204,0.1363056640625,0.13998633117675782,0.1419325439453125,0.273337587890625,"[0.14988787841796875, 0.14766496276855468, 0.14396205139160156, 0.14188954162597656, 0.14043545532226562, 0.1401712646484375, 0.14181068420410156, 0.1413376007080078, 0.14149530029296875, 0.1412210235595703, 0.1421392364501953, 0.14165196228027344, 0.14433689880371095, 0.14238310241699217, 0.13620428466796874, 0.13577626037597657, 0.14025625610351564, 0.1366138916015625, 0.13676441955566407, 0.13585203552246095, 0.13596263122558594, 0.13545266723632812, 0.1352120361328125, 0.13816432189941405, 0.14305885314941405, 0.13969509887695314, 0.13598226928710938, 0.13899655151367188, 0.142308349609375, 0.14245703125, 0.13762232971191407, 0.13979359436035158, 0.13923513793945314, 0.13666610717773436, 0.13895680236816407, 0.14275584411621095, 0.13469696044921875, 0.1425807342529297, 0.14284902954101564, 0.1402941436767578, 0.14227264404296874, 0.13976051330566405, 0.1420779571533203, 0.1423953857421875, 0.13646336364746095, 0.14172979736328126, 0.14220716857910157, 0.1360013427734375, 0.136015869140625, 0.1365985870361328, 0.1367643585205078, 0.14221824645996095, 0.1405338592529297, 0.14239222717285155, 0.14214656066894532, 0.14200933837890625, 0.14196121215820312, 0.1406965789794922, 0.13656370544433594, 0.13660159301757813, 0.14127615356445314, 0.1366620178222656, 0.27431228637695315, 0.13553263854980468, 0.13589280700683593, 0.1360834503173828, 0.13625958251953124, 0.13628329467773437, 0.1362071990966797, 0.13589503479003906, 0.13628005981445312, 0.13602517700195313, 0.1362277374267578, 0.13591552734375, 0.13624114990234376, 0.1354598388671875, 0.136193115234375, 0.13627894592285156, 0.13619815063476562, 0.136053955078125, 0.13597779846191407, 0.13675929260253905, 0.13648281860351563, 0.13886770629882814, 0.13945549011230468, 0.13645106506347657, 0.13632716369628906, 0.13611212158203126, 0.13627597045898437, 0.136268798828125, 0.1364071044921875, 0.13589190673828125, 0.137417724609375, 0.135984130859375, 0.13632716369628906, 0.13707469177246093, 0.13632722473144532, 0.13605574035644533, 0.13619815063476562, 0.1356431427001953, 0.13591346740722657, 0.13595033264160156, 0.1353861083984375, 0.13606195068359375, 0.1363937225341797, 0.13646438598632812, 0.14046003723144532, 0.13906944274902344, 0.13594009399414062, 0.1369733123779297, 0.1363651885986328, 0.13627891540527343, 0.13615718078613281, 0.13612646484375, 0.13638864135742187, 0.13624726867675782, 0.13630873107910157, 0.1359779815673828, 0.13711360168457032, 0.13616435241699218, 0.13599232482910156, 0.1360650177001953, 0.1363619842529297, 0.13574861145019532, 0.13643878173828125, 0.27481396484375, 0.13824409484863281, 0.13750886535644533, 0.13692005920410155, 0.13676850891113282, 0.13617561340332032, 0.1396961212158203, 0.13664057922363282, 0.1363086700439453, 0.13667942810058595, 0.1359667205810547, 0.1362677764892578, 0.13631794738769532, 0.13649737548828125, 0.1363586883544922, 0.13851852416992189, 0.13667225646972656, 0.13651455688476563, 0.13778329467773437, 0.13668453979492187, 0.1362821502685547, 0.1366814422607422, 0.13598822021484375, 0.13631907653808595, 0.1363618927001953, 0.136521728515625, 0.13604672241210938, 0.13624525451660155, 0.1365902099609375, 0.1361561584472656, 0.13815411376953124, 0.13858087158203125, 0.136380615234375, 0.13662086486816405, 0.1362155456542969, 0.13575987243652343, 0.13573222351074218, 0.13595237731933593, 0.13494886779785156, 0.13592166137695313, 0.1359431610107422, 0.13755699157714843, 0.1393931884765625, 0.13692726135253908, 0.13622149658203125, 0.13768293762207032, 0.13615309143066406, 0.13818572998046874, 0.1368289337158203, 0.13640605163574218, 0.13609776306152344, 0.1363609619140625, 0.13732147216796875, 0.13992672729492187, 0.13795103454589844, 0.13635789489746095, 0.13770034790039062, 0.1362677764892578, 0.13658316040039062, 0.13668045043945312, 0.13752012634277344, 0.13808026123046874, 0.1362872314453125, 0.2736363525390625, 0.13601075744628907, 0.13585714721679687, 0.13596978759765624, 0.13635789489746095, 0.1356431427001953, 0.13778242492675782, 0.13615394592285157, 0.13602610778808594, 0.13648690795898438, 0.13629440307617188, 0.1367093048095703, 0.13611398315429687, 0.13660467529296874, 0.13592781066894533, 0.13755084228515624, 0.1397760009765625, 0.1375979461669922, 0.13798399353027344, 0.13640089416503906, 0.13628620910644532, 0.13761740112304688, 0.1365278778076172, 0.1362073669433594, 0.1361274871826172, 0.13660671997070312, 0.1363752899169922, 0.13646540832519533, 0.13638552856445313, 0.13623507690429687, 0.13634042358398438, 0.1364285430908203, 0.1362135009765625, 0.136089599609375, 0.13598104858398438, 0.13554083251953125, 0.1363681335449219, 0.13629849243164063, 0.13584375, 0.1363609619140625, 0.13906533813476563, 0.13712588500976564, 0.13691410827636719, 0.1379653778076172, 0.13544873046875, 0.13545660400390624, 0.1361029052734375, 0.13567385864257814, 0.1364490203857422, 0.1360579833984375, 0.13600038146972657, 0.1376040954589844, 0.13709721374511719, 0.13590745544433594, 0.13611097717285156, 0.13692416381835937, 0.136553466796875, 0.13648281860351563, 0.13595750427246095, 0.13591757202148438, 0.13638656616210937, 0.13590631103515624, 0.13884825134277343, 0.2743162841796875, 0.13775360107421875, 0.13637222290039064, 0.1361634521484375, 0.13614682006835938, 0.135910400390625, 0.13654541015625, 0.1392670440673828, 0.13641615295410156, 0.13630464172363282, 0.13682278442382811, 0.13541273498535156, 0.13584384155273438, 0.1365780487060547, 0.13630157470703125, 0.13613363647460938, 0.13642857360839844, 0.1363916473388672, 0.13606297302246093, 0.136163330078125, 0.13590631103515624, 0.1364644470214844, 0.13839045715332032, 0.13620326232910157, 0.13613465881347656, 0.14064230346679688, 0.13726719665527343, 0.1364521026611328, 0.1363548126220703, 0.13647769165039061, 0.1363056640625, 0.13631494140625, 0.13608953857421874, 0.1365893096923828, 0.1368269500732422, 0.13752825927734375, 0.13940121459960939, 0.13748445129394532, 0.13712777709960938, 0.13680128479003906, 0.13577113342285158, 0.13611007690429688, 0.13852467346191405, 0.1398814697265625, 0.13714533996582032, 0.13800344848632812, 0.14050918579101562, 0.13650636291503906, 0.13727845764160157, 0.13809365844726562, 0.1357464599609375, 0.13630873107910157, 0.13618893432617188, 0.13631488037109374, 0.13585203552246095, 0.13553050231933594, 0.13572505187988282, 0.13615411376953124, 0.13697654724121094, 0.13603619384765625, 0.13699891662597657, 0.14267289733886718, 0.13998182678222656, 0.28226150512695314, 0.13753138732910156, 0.1362617950439453, 0.13632511901855468, 0.1359727020263672, 0.13584690856933593, 0.1362001953125, 0.13828608703613282, 0.13946981811523437, 0.1373982696533203, 0.13763481140136719, 0.13620428466796874, 0.13881344604492188, 0.1380301513671875, 0.1374699249267578, 0.13884104919433593, 0.13694464111328125, 0.1385912322998047, 0.13573324584960939, 0.13662106323242187, 0.13518130493164063, 0.13542092895507812, 0.13529498291015624, 0.13547622680664062, 0.13579161071777343, 0.13545062255859375, 0.13542707824707031, 0.13543731689453126, 0.13538304138183593, 0.13520384216308592, 0.134384765625, 0.13530201721191407, 0.13512617492675782, 0.13532861328125, 0.13919949340820312, 0.13607635498046874, 0.13525088500976562, 0.13523660278320312, 0.1347870788574219, 0.13481062316894532, 0.13569638061523437, 0.13493043518066405, 0.13509939575195312, 0.135077880859375, 0.13534927368164062, 0.1358294677734375, 0.13546290588378906, 0.1353318328857422, 0.13538201904296876, 0.13558578491210938, 0.13471334838867188, 0.13525401306152343, 0.1353564147949219, 0.13520486450195313, 0.13578445434570313, 0.13556121826171874, 0.1356021728515625, 0.1355284423828125, 0.14007296752929688, 0.1354905548095703, 0.13551513671875, 0.1351884765625, 0.13512806701660157, 0.2725693359375, 0.13558700561523437, 0.13558560180664062, 0.1356390380859375, 0.13567385864257814, 0.13566157531738282, 0.14110003662109374, 0.13586636352539064, 0.13542195129394533, 0.13578346252441406, 0.1359073028564453, 0.13568722534179686, 0.13587776184082032, 0.13565625, 0.13624732971191406, 0.13585302734375, 0.13591961669921876, 0.1406863708496094, 0.13640188598632813, 0.13642343139648438, 0.1364295654296875, 0.13577626037597657, 0.13762661743164062, 0.13568614196777343, 0.13732659912109374, 0.13610086059570312, 0.13570559692382814, 0.13572198486328124, 0.13613775634765626, 0.13586941528320312, 0.13610394287109376, 0.1359349822998047, 0.1360168914794922, 0.13558169555664062, 0.1359811553955078, 0.13574748229980468, 0.13579673767089845, 0.13681356811523437, 0.1367060546875, 0.13589817810058594, 0.13676025390625, 0.13588796997070313, 0.135553955078125, 0.1356155548095703, 0.13906019592285157, 0.13608444213867188, 0.13550694274902345, 0.1356104278564453, 0.13594412231445313, 0.1358673858642578, 0.1358019256591797, 0.13565242004394532, 0.13578842163085938, 0.1360834503173828, 0.13585203552246095, 0.1357117462158203, 0.13608447265625, 0.13566053771972655, 0.13594522094726563, 0.13564842224121093, 0.13598104858398438, 0.1348739471435547, 0.13597286987304688, 0.2720789794921875, 0.13575975036621094, 0.13560421752929688, 0.13812428283691405, 0.1402603454589844, 0.1372620849609375, 0.13582234191894532, 0.13575885009765626, 0.13576499938964845, 0.13611827087402345, 0.1351577606201172, 0.13579571533203125, 0.13611827087402345, 0.13599449157714844, 0.1354371795654297, 0.13947097778320314, 0.13811801147460936, 0.1360199737548828, 0.13594522094726563, 0.13585408020019532, 0.13577113342285158, 0.13573837280273438, 0.13565235900878905, 0.1361878967285156, 0.136195068359375, 0.13603634643554688, 0.13957017517089843, 0.1381591033935547, 0.13590733337402344, 0.1359226837158203, 0.13579878234863282, 0.138756103515625, 0.13628314208984374, 0.13658828735351564, 0.13879295349121093, 0.13949542236328125, 0.13971559143066406, 0.1381898193359375, 0.13689651489257812, 0.13621043395996094, 0.1349785614013672, 0.13623808288574218, 0.13557875061035157, 0.13460263061523436, 0.134898681640625, 0.13534002685546875, 0.135510009765625, 0.13546188354492186, 0.1353533477783203, 0.1354475555419922, 0.13550186157226562, 0.13543215942382814, 0.13552639770507813, 0.13712998962402342, 0.13973504638671874, 0.13576396179199218, 0.13535948181152344, 0.13556736755371093, 0.13554080200195312, 0.13528790283203124, 0.13553750610351561, 0.13526527404785157, 0.13937152099609376, 0.27988070678710936, 0.13893222045898437, 0.13684121704101562, 0.14009138488769532, 0.13784268188476562, 0.13598617553710937, 0.13791334533691407, 0.13850828552246094, 0.13901119995117187, 0.13893417358398438, 0.14032687377929687, 0.13661593627929688, 0.1395476531982422, 0.13814373779296876, 0.13520281982421875, 0.13813865661621094, 0.13614997863769532, 0.13586329650878906, 0.13634457397460936, 0.13633433532714845, 0.13620428466796874, 0.13833010864257814, 0.1364920654296875, 0.13657801818847656, 0.14219993591308594, 0.13754969787597657, 0.13599334716796874, 0.13613568115234376, 0.13823692321777345, 0.13782937622070313, 0.13634970092773438, 0.1360343017578125, 0.13588595581054688, 0.136532958984375, 0.13688412475585937, 0.13863731384277345, 0.13929574584960938, 0.13791436767578125, 0.1364838409423828, 0.13624832153320313, 0.13730918884277343, 0.13695283508300782, 0.13647769165039061, 0.13553056335449218, 0.13752825927734375, 0.137776123046875, 0.13736857604980468, 0.13668658447265625, 0.13617971801757814, 0.13834751892089844, 0.13937767028808593, 0.1379921875, 0.1359974365234375, 0.1367183380126953, 0.13781114196777344, 0.1385582733154297, 0.13805874633789061, 0.13638041687011718, 0.14069453430175782, 0.1405491180419922, 0.13930598449707032, 0.13724278259277345, 0.1391768035888672, 0.27513037109375, 0.13619203186035156, 0.1360589141845703, 0.13624415588378908, 0.1358133087158203, 0.1359870147705078, 0.13598104858398438, 0.13620838928222656, 0.135657470703125, 0.1387130889892578, 0.13655859375, 0.13592678833007812, 0.13561447143554686, 0.13624319458007814, 0.13576806640625, 0.13599334716796874, 0.13577113342285158, 0.13600563049316405, 0.13578034973144532, 0.13589299011230468, 0.13587455749511718, 0.13686476135253905, 0.136732666015625, 0.13627903747558595, 0.13618380737304686, 0.1364285430908203, 0.1362442169189453, 0.1364715576171875, 0.13552024841308594, 0.13628314208984374, 0.13929676818847656, 0.13876121520996093, 0.13615309143066406, 0.14000434875488282, 0.13895173645019532, 0.13639468383789063, 0.13584793090820313, 0.1353123779296875, 0.13633638000488282, 0.13655859375, 0.13620127868652343, 0.13616429138183594, 0.13556636047363282, 0.13449110412597656, 0.13782528686523438, 0.1363998718261719, 0.13610195922851562, 0.1353399658203125, 0.13577113342285158, 0.13993881225585938, 0.14040473937988282, 0.1381273651123047, 0.13764515686035156, 0.1361714172363281, 0.1377198028564453, 0.13922918701171874, 0.14060954284667967, 0.13710643005371093, 0.13558784484863282, 0.14004019165039064, 0.1410232391357422, 0.14098739624023438, 0.13884722900390625]",tokens/s,7.195407074075256,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5100,7 +5100,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5128,7 +5128,7 @@ RuntimeError: Isolated process exited with non-zero code -9 raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") RuntimeError: Isolated process exited with non-zero code -9 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5187,7 +5187,7 @@ ValueError: CodeGenForCausalLM does not support an attention implementation thro ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,53592.666112,16670.785536,0.0,16015.949824,15921.204224,s,10,7.127675476074219,0.7127675476074219,0.01149264894687282,0.7140398864746094,0.7276809753417969,0.7280174591064453,0.7282866461181641,"[0.71203759765625, 0.7283539428710938, 0.6990225830078125, 0.7211824951171875, 0.7001098022460938, 0.7160421752929688, 0.719827880859375, 0.727606201171875, 0.6935702514648437, 0.7099225463867187]",tokens/s,359.1633778211795,kWh,8.344990798720607e-06,4.572616197328898e-06,1.2488269249866234e-05,2.5405876245915736e-05,tokens/kWh,10076408.997747311,MB,53592.666112,16670.785536,0.0,16015.949824,15921.206784,s,10,445.65144140625,44.565144140625,0.404306927303512,44.448330078125,45.166128125,45.304851171875,45.415829609375,"[45.44357421875, 44.64917578125, 44.11805859375, 44.2393203125, 44.24665234375, 44.30507421875, 44.33498046875, 45.13530078125, 44.617625, 44.5616796875]",tokens/s,1.4136608601826561,kWh,0.0005204199726658839,0.000285235793628708,0.0007449961145149397,0.0015506518808095315,tokens/kWh,40628.07441158895,,s,629,451.34379284667943,0.7175576992793,0.08493602610246886,0.7082782592773438,0.72226123046875,0.7242967041015624,1.3919694287109377,"[0.7255736083984375, 0.7212359619140625, 0.7245496215820313, 0.7233116455078125, 0.7230648193359375, 0.7220479736328125, 0.7137260131835937, 0.7248711547851563, 0.724379638671875, 0.7255408935546875, 0.7253944091796874, 0.725496826171875, 0.7250953979492187, 0.7236444091796875, 0.7275110473632812, 0.7256217651367187, 0.72698779296875, 0.7253258056640625, 0.7233290405273437, 0.7241144409179687, 0.7251845092773438, 0.725128173828125, 0.714851318359375, 0.7212083129882813, 0.7143311157226563, 0.7219824829101562, 0.7221483764648438, 0.7216537475585938, 0.7222804565429688, 0.7123056640625, 0.7169105834960937, 0.7230873413085938, 0.7169515380859375, 0.7236065063476562, 0.7231190795898438, 0.7204198608398438, 0.7217203369140625, 0.7216404418945312, 0.72226611328125, 0.7051611938476563, 0.7228477172851563, 0.7264594116210937, 0.7243192138671875, 0.7231190795898438, 0.7252920532226562, 0.723989501953125, 0.724262939453125, 0.695920654296875, 0.7241328735351562, 0.7280496826171875, 0.706951171875, 0.7245916137695313, 0.7246786499023438, 0.7198187255859375, 0.7036958618164062, 0.7181096801757813, 0.7239239501953125, 0.7240703735351562, 0.7235389404296875, 0.7227525024414062, 0.7108710327148438, 0.7217899780273438, 1.4524375, 0.7067391967773438, 0.7191275634765625, 0.7226368408203125, 0.7199580078125, 0.72167529296875, 0.7217029418945312, 0.7170509033203125, 0.719899658203125, 0.7225620727539063, 0.7176365966796875, 0.7201474609375, 0.7185131225585938, 0.7155138549804687, 0.7081370239257813, 0.7127091064453125, 0.72131884765625, 0.72052734375, 0.7083407592773437, 0.7227289428710938, 0.7237437744140625, 0.72176025390625, 0.7206737670898438, 0.6994114379882812, 0.6926550903320312, 0.7082721557617188, 0.6912849731445313, 0.7200347900390625, 0.7174215698242188, 0.6940436401367187, 0.69676953125, 0.6918922119140625, 0.6978641967773438, 0.694908935546875, 0.7011716918945312, 0.69071875, 0.6904483642578125, 0.7062036743164063, 0.7022745361328125, 0.6918450927734375, 0.699926513671875, 0.6986731567382812, 0.7079505615234375, 0.7032371215820312, 0.6960722045898438, 0.7045447387695313, 0.7077488403320312, 0.7051837158203125, 0.7070075073242188, 0.710645751953125, 0.7024578857421875, 0.706423828125, 0.7130634155273438, 0.70959716796875, 0.70384228515625, 0.7136839599609375, 0.7106539306640625, 0.7176406860351563, 0.7080068969726563, 0.713153564453125, 0.6919813232421875, 0.6948587646484375, 0.7084544067382812, 1.3894317626953125, 0.6915389404296876, 0.7015577392578125, 0.6911651611328125, 0.690165771484375, 0.6914816284179688, 0.6974034423828125, 0.7106436767578125, 0.6942320556640625, 0.6952366333007812, 0.6996592407226563, 0.6954711303710938, 0.6958806762695312, 0.7094773559570312, 0.6908098754882812, 0.6911528930664063, 0.6902302856445313, 0.6937047119140625, 0.6901381225585937, 0.6935819091796875, 0.703395751953125, 0.7118079833984375, 0.7109713745117188, 0.7025889282226563, 0.7020175170898437, 0.7004733276367188, 0.6917447509765625, 0.7035228271484375, 0.6960486450195312, 0.70552880859375, 0.7067381591796875, 0.7056066284179687, 0.710044677734375, 0.7103150024414062, 0.711426025390625, 0.7129098510742188, 0.6989434814453125, 0.7000647583007813, 0.7007918090820312, 0.7004098510742187, 0.697217041015625, 0.7007958984375, 0.6997033081054688, 0.6927605590820313, 0.6928076782226562, 0.6937569580078125, 0.6907647705078125, 0.6925404663085938, 0.6920078735351562, 0.703889404296875, 0.7010293579101563, 0.7013058471679687, 0.6958612670898437, 0.6920745849609375, 0.6905845336914063, 0.70081640625, 0.7138611450195312, 0.711014404296875, 0.71436083984375, 0.713466796875, 0.716062744140625, 0.7142328491210937, 0.70786767578125, 1.431025634765625, 0.7076649169921875, 0.704828369140625, 0.708400146484375, 0.6967101440429687, 0.6960855102539063, 0.7068538818359374, 0.7067555541992188, 0.7016141967773437, 0.7125431518554688, 0.700564453125, 0.7082403564453125, 0.6999664916992188, 0.7036201782226562, 0.710879150390625, 0.7103314208984375, 0.7003729858398438, 0.6915963134765625, 0.6904381713867187, 0.6911273193359375, 0.6961541137695313, 0.6923499755859375, 0.69075146484375, 0.7082782592773438, 0.69370166015625, 0.7015720825195313, 0.6899036865234375, 0.6902424926757813, 0.6996387939453125, 0.6934896850585938, 0.7028172607421875, 0.7096422119140625, 0.7123814697265625, 0.7098613891601563, 0.7124951171875, 0.7158292236328125, 0.7061626586914063, 0.7083079833984375, 0.70917529296875, 0.7128709106445312, 0.7124019165039063, 0.6923427734375, 0.7018711547851563, 0.698140625, 0.715873291015625, 0.703626220703125, 0.70523291015625, 0.71003857421875, 0.7071589965820313, 0.7117864990234375, 0.6946099243164062, 0.70748779296875, 0.6912439575195313, 0.6920478515625, 0.6906951904296875, 0.6920386352539063, 0.6971259155273437, 0.6944010009765625, 0.701539306640625, 0.6965729370117187, 0.7084338989257812, 0.7089285278320312, 0.6917120361328125, 1.392956298828125, 0.6897039794921875, 0.7090360107421875, 0.6931896362304687, 0.70025830078125, 0.7102003173828125, 0.7080089721679688, 0.701665283203125, 0.7045130004882812, 0.7013294067382813, 0.7061616821289063, 0.705196044921875, 0.6906838989257813, 0.6974269409179688, 0.7108444213867188, 0.7117875366210937, 0.7111321411132813, 0.7121766357421875, 0.7053352661132812, 0.7101563110351562, 0.6991390991210937, 0.7133009643554687, 0.7077877807617188, 0.7040645141601563, 0.69035107421875, 0.6899118041992187, 0.689249267578125, 0.6928046264648438, 0.71171484375, 0.7106478271484375, 0.7127449340820312, 0.7135928344726562, 0.7123671264648438, 0.7051868286132813, 0.7122144775390625, 0.7146076049804687, 0.709739501953125, 0.6987212524414063, 0.7089705200195312, 0.7040215454101563, 0.6997094116210938, 0.699683837890625, 0.69285888671875, 0.6886297607421875, 0.6914150390625, 0.6925332641601563, 0.6909890747070313, 0.6912481079101562, 0.6907320556640625, 0.6927216796875, 0.6915532836914062, 0.701533203125, 0.69283740234375, 0.692980712890625, 0.7023831176757812, 0.6914293823242188, 0.700474365234375, 0.6968370971679687, 0.71233740234375, 0.71347509765625, 0.7166781616210938, 0.7148134155273438, 0.7160411987304688, 1.4289151611328126, 0.6955325317382812, 0.6953052368164062, 0.692052978515625, 0.7058472900390625, 0.6911641845703125, 0.70571826171875, 0.6948782348632813, 0.691051513671875, 0.6908047485351563, 0.7034111938476563, 0.7099883422851563, 0.697037841796875, 0.6940374755859375, 0.7124285278320313, 0.7082250366210937, 0.7124572143554687, 0.7143598022460937, 0.7101328125, 0.6945299682617188, 0.7015025024414062, 0.704259033203125, 0.6985942993164063, 0.7057039184570313, 0.7063224487304688, 0.704321533203125, 0.69368115234375, 0.7010426635742187, 0.7125791015625, 0.7105331420898438, 0.7042662353515625, 0.7042785034179687, 0.7002726440429687, 0.7055953369140625, 0.704611328125, 0.7138621215820312, 0.71011328125, 0.7018137817382812, 0.710160400390625, 0.7036774291992187, 0.705523681640625, 0.7112182006835938, 0.7116452026367187, 0.70494921875, 0.700062744140625, 0.7130439453125, 0.7110287475585938, 0.7117793579101562, 0.6937733154296875, 0.696748046875, 0.6906859741210938, 0.6916853637695313, 0.69648486328125, 0.6916188354492188, 0.6906173706054688, 0.6900479736328125, 0.6963834838867188, 0.7100416259765625, 0.713185302734375, 0.7102945556640625, 0.7123804321289062, 0.7075655517578125, 0.7153110961914062, 1.41534716796875, 0.6952191772460937, 0.7029258422851562, 0.6983966674804688, 0.7045549926757813, 0.7002542724609375, 0.699189208984375, 0.71530908203125, 0.6991575317382812, 0.6983895263671875, 0.7023472900390625, 0.7032371215820312, 0.70215576171875, 0.6975580444335937, 0.6901176147460938, 0.6900111083984375, 0.68868505859375, 0.6900009155273438, 0.6894612426757812, 0.6902435913085937, 0.6902640380859375, 0.7043768310546875, 0.7031654663085938, 0.7112161254882813, 0.7124100952148438, 0.7125964965820313, 0.7104470825195313, 0.7130715942382813, 0.708515869140625, 0.6899261474609375, 0.68954931640625, 0.6919782104492187, 0.694086669921875, 0.6907914428710937, 0.69389208984375, 0.7029248046875, 0.7086295166015625, 0.70033203125, 0.6978324584960938, 0.6945126342773438, 0.69344970703125, 0.6994759521484375, 0.706914306640625, 0.6980198364257812, 0.708443115234375, 0.7137771606445312, 0.6892994384765625, 0.6904463500976562, 0.689138671875, 0.6954066162109375, 0.7219671020507813, 0.722260009765625, 0.7195719604492188, 0.7222763671875, 0.7202744140625, 0.7216005249023437, 0.722387939453125, 0.7215103759765625, 0.720226318359375, 0.7207874755859375, 0.7213209838867187, 0.7214581909179687, 0.7217469482421875, 1.42337841796875, 0.7240621948242187, 0.7218779907226562, 0.7150274658203125, 0.7181455078125, 0.6992332763671875, 0.717083740234375, 0.7217899169921875, 0.7200143432617188, 0.7218565063476563, 0.7160719604492187, 0.7213772583007813, 0.7195402221679688, 0.7249326171875, 0.690555908203125, 0.7115612182617187, 0.7213373413085937, 0.7208335571289063, 0.7211345825195312, 0.7220838623046875, 0.7225200805664063, 0.7187169189453125, 0.7207577514648438, 0.721586181640625, 0.7188264770507813, 0.71381298828125, 0.7213588256835938, 0.7263856811523437, 0.7214356079101563, 0.7246417846679688, 0.7238932495117187, 0.718761962890625, 0.7213424682617188, 0.7176724243164062, 0.721596435546875, 0.7207587890625, 0.7093667602539062, 0.7189494018554687, 0.6898718872070313, 0.69104443359375, 0.6902578735351562, 0.7035914306640625, 0.720300048828125, 0.7240693969726563, 0.7204300537109375, 0.7221739501953125, 0.719899658203125, 0.7207147216796875, 0.7212882080078125, 0.7198812255859375, 0.7206256713867187, 0.7191695556640625, 0.715915283203125, 0.7201822509765625, 0.7202273559570312, 0.7202211303710937, 0.6893793334960937, 0.7027394409179688, 0.7177861328125, 0.7208673095703125, 0.7189872436523438, 0.7194183959960937, 0.7195914306640625, 1.3814937744140625, 0.7049287719726562, 0.7155353393554688, 0.7175731201171875, 0.68709375, 0.7163494262695312, 0.7188336791992187, 0.7206522827148437, 0.7190947875976562, 0.718688232421875, 0.7189657592773437, 0.7205867309570313, 0.7144601440429688, 0.6903009033203125, 0.6952529907226562, 0.698081298828125, 0.6930339965820312, 0.6908262329101562, 0.6966180419921875, 0.6964141235351563, 0.7180994262695313, 0.7190281982421876, 0.7095634155273437, 0.6912828979492187, 0.690714599609375, 0.694160400390625, 0.7181824340820312, 0.7190885620117188, 0.7019171752929687, 0.6902866821289062, 0.6949180908203125, 0.6963466186523437, 0.6958377075195312, 0.7111454467773437, 0.7185950927734375, 0.7016458129882812, 0.714861572265625, 0.7039610595703125, 0.6974904174804688, 0.710371337890625, 0.7163740234375, 0.7177113647460938, 0.7179694213867187, 0.717907958984375, 0.7192576293945312, 0.7063377685546876, 0.7086356201171875, 0.7209031982421875, 0.7211950073242187, 0.7132743530273438, 0.7218779907226562, 0.70689794921875, 0.7037071533203125, 0.6996551513671875, 0.7013928833007812, 0.7203563232421875, 0.7183206176757813, 0.7195709228515625, 0.7184322509765625, 0.6994892578125, 0.7177011108398438, 0.7173232421875, 0.6947051391601563, 1.4297733154296874, 0.71096728515625, 0.7161384887695312, 0.7194634399414063, 0.6989844360351563, 0.7201853637695312, 0.719120361328125, 0.6948587646484375, 0.6928762817382812, 0.69334130859375, 0.6918604125976563, 0.6926243896484375, 0.7062702026367188, 0.6956707763671875, 0.7146905517578125, 0.718424072265625, 0.7171133422851562, 0.6943836059570313, 0.6950523071289062, 0.72003173828125, 0.71356005859375, 0.7142328491210937, 0.7044229125976562, 0.6922567749023437, 0.693190673828125, 0.6924758911132812, 0.7026697998046875, 0.6950604858398437, 0.7184701538085938, 0.7030877075195312, 0.6978170166015625, 0.7199293212890625, 0.7171512451171875, 0.7184619750976563, 0.7007611083984375, 0.7018014526367188, 0.7160801391601562, 0.7034695434570313, 0.6984356079101562, 0.7046102905273437, 0.7172556762695312, 0.6988626098632813, 0.70894287109375, 0.7222907104492188, 0.7052902221679688, 0.71887158203125, 0.71828173828125, 0.7132344360351562, 0.7128657836914063, 0.71899853515625, 0.7183267822265625, 0.7189278564453125, 0.693190673828125, 0.693011474609375, 0.7165470581054687, 0.6971566162109375, 0.7110062255859375, 0.7240038452148437, 0.7132200927734375, 0.6925977783203126, 0.6917017822265625, 0.711172119140625, 0.7061053466796875]",tokens/s,1.3936161524075055,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5216,7 +5216,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5424,7 +5424,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1257.271296,8173.125632,0.0,7518.28992,7334.49728,s,10,1.1121479263305662,0.11121479263305664,0.0023814823720823245,0.11015043258666993,0.11502404174804688,0.11586635665893555,0.11654020858764648,"[0.11670867156982422, 0.10984870147705078, 0.1088564453125, 0.11011993408203125, 0.11122988891601562, 0.11483686065673829, 0.10990399932861328, 0.11040982055664063, 0.1101809310913086, 0.11005267333984375]",tokens/s,2301.852064272145,kWh,1.3091683569699708e-06,7.173647082339273e-07,3.425182837824217e-06,5.451715903028115e-06,tokens/kWh,46957692.6886096,MB,1257.271296,8173.125632,0.0,7518.28992,7485.806592,s,10,69.45216015625,6.945216015625,0.03570031437080937,6.941093505859374,6.994362890625,6.9972219726562495,6.99950923828125,"[6.895080078125, 6.94246826171875, 6.9937275390625, 6.93971875, 7.0000810546875, 6.98611962890625, 6.9457822265625, 6.914953125, 6.93530224609375, 6.89892724609375]",tokens/s,9.070992156077759,kWh,8.340076308947637e-05,4.570957993626566e-05,0.00016399845110097504,0.000293108794126717,tokens/kWh,214937.25627612454,,s,629,70.33750126647953,0.11182432633780526,0.013263304791932896,0.10958748626708985,0.11358883819580079,0.11438264007568359,0.21947199951171875,"[0.10853785705566406, 0.10911743927001953, 0.10968985748291016, 0.10979958343505859, 0.11122672271728516, 0.10912358093261719, 0.10954854583740234, 0.10924134063720703, 0.1092220458984375, 0.10932003021240234, 0.10912073516845704, 0.10907020568847656, 0.1088941421508789, 0.10874262237548828, 0.10958866882324218, 0.11017504119873046, 0.10943590545654297, 0.10930601501464844, 0.10942038726806641, 0.10924134063720703, 0.10951475524902343, 0.10944825744628907, 0.10908774566650391, 0.10924230194091797, 0.10933248138427734, 0.10894643402099609, 0.10940723419189453, 0.10915142059326172, 0.10929644775390625, 0.10974208068847656, 0.10932530975341796, 0.10925977325439454, 0.1092116470336914, 0.10959769439697266, 0.10926182556152343, 0.10874982452392579, 0.10950553894042969, 0.109233154296875, 0.10920448303222656, 0.1093438720703125, 0.10996006774902344, 0.10966233825683594, 0.10952998352050781, 0.10909593963623047, 0.1090723876953125, 0.10988543701171875, 0.11014246368408204, 0.10933964538574219, 0.1089280014038086, 0.1093570556640625, 0.10944409942626954, 0.10953830718994141, 0.10932121276855469, 0.1089269790649414, 0.10965100860595703, 0.11289798736572265, 0.10945433807373046, 0.10934899139404297, 0.10942041778564453, 0.10965196990966797, 0.10889727783203125, 0.10944306945800782, 0.21951181030273437, 0.10946867370605469, 0.10920365142822265, 0.10915206146240235, 0.10887168121337891, 0.11283763122558593, 0.11002674865722656, 0.10913382720947265, 0.10944512176513672, 0.1084958724975586, 0.10887987518310546, 0.10966937255859376, 0.10958541107177734, 0.10915532684326172, 0.10813148498535156, 0.10894115447998047, 0.10918502044677734, 0.10935091400146485, 0.11128438568115234, 0.11376214599609374, 0.10910521697998046, 0.11015878295898437, 0.11337728118896484, 0.11337830352783203, 0.11346329498291016, 0.1136824951171875, 0.11284575653076172, 0.11328409576416015, 0.11337120056152344, 0.11316432189941407, 0.11394960021972657, 0.113544189453125, 0.11197030639648438, 0.10917494201660156, 0.10904764556884766, 0.10931199645996094, 0.10897017669677735, 0.10901599884033203, 0.1089227523803711, 0.10942156982421875, 0.10901209259033204, 0.10914189147949219, 0.10888601684570312, 0.10905907440185547, 0.10892594909667969, 0.10893004608154297, 0.10896192169189453, 0.1092442855834961, 0.10953113555908203, 0.11112857818603515, 0.11319910430908203, 0.1128632354736328, 0.10916671752929688, 0.10917874908447266, 0.10883379364013672, 0.10915225219726563, 0.10922598266601563, 0.10958131408691406, 0.10922598266601563, 0.10881433868408204, 0.1090355224609375, 0.10894131469726563, 0.10907750701904297, 0.2183372802734375, 0.10915737915039063, 0.11211468505859375, 0.11267993927001953, 0.11483545684814453, 0.1097441635131836, 0.10974921417236329, 0.10912358093261719, 0.1094655990600586, 0.109412353515625, 0.10974720001220703, 0.1098270721435547, 0.1097000961303711, 0.11293593597412109, 0.1132759017944336, 0.11354828643798828, 0.11353292846679687, 0.11330457305908204, 0.1147484130859375, 0.1138892822265625, 0.10949632263183594, 0.10929254150390624, 0.10903142547607422, 0.10922700500488282, 0.10923011016845703, 0.10980553436279297, 0.10820301055908203, 0.11203193664550781, 0.11292550659179687, 0.1146961898803711, 0.11323596954345704, 0.1134940185546875, 0.11805696105957031, 0.11398451232910156, 0.11375206756591796, 0.11362815856933593, 0.11326464080810547, 0.1133680648803711, 0.11396198272705078, 0.11297382354736328, 0.1134950408935547, 0.1109749755859375, 0.10946867370605469, 0.10942156982421875, 0.10922393798828126, 0.10915430450439453, 0.10926092529296876, 0.10945011138916015, 0.10929151916503907, 0.10918399810791016, 0.10907647705078125, 0.10938265228271485, 0.1093355484008789, 0.1087979507446289, 0.11060326385498047, 0.1089269790649414, 0.10936729431152344, 0.10921990203857422, 0.10919929504394531, 0.10910105895996093, 0.10929663848876953, 0.10946969604492188, 0.1091031036376953, 0.2195650634765625, 0.1092863998413086, 0.10983628845214843, 0.10940723419189453, 0.10983526611328125, 0.10937657928466797, 0.11131590270996093, 0.1100021743774414, 0.11015679931640625, 0.11035852813720703, 0.11043225860595703, 0.10973388671875, 0.10992230224609376, 0.10938470458984376, 0.11009331512451172, 0.11044454193115234, 0.11013529968261719, 0.10993049621582031, 0.11011174774169921, 0.10978406524658203, 0.10939289855957031, 0.1098792953491211, 0.11011481475830077, 0.11115315246582032, 0.10974617767333984, 0.10993663787841797, 0.1094277114868164, 0.1096079330444336, 0.10927206420898437, 0.10971238708496094, 0.11074662780761718, 0.10997657775878907, 0.11023052978515625, 0.11088588714599609, 0.10970214080810547, 0.10850201416015624, 0.10940415954589844, 0.10951475524902343, 0.10909184265136719, 0.10937139129638672, 0.10916556549072266, 0.1098087387084961, 0.11038198089599609, 0.10981068420410156, 0.10933248138427734, 0.115378173828125, 0.11064422607421875, 0.11104358673095703, 0.11056947326660156, 0.11062477111816406, 0.112, 0.11066470336914062, 0.10972467041015625, 0.10975027465820313, 0.109264892578125, 0.10971340942382812, 0.11165081787109375, 0.10943488311767578, 0.1120153579711914, 0.11232771301269531, 0.10975740814208984, 0.11008204650878907, 0.11190067291259766, 0.22177690124511718, 0.10909490966796875, 0.10954854583740234, 0.10981887817382813, 0.10983225250244141, 0.10969696044921876, 0.10940415954589844, 0.10918707275390625, 0.10898636627197265, 0.10910822296142578, 0.10902732849121094, 0.10907341003417968, 0.10883379364013672, 0.10934783935546875, 0.1092116470336914, 0.10958553314208984, 0.10909273529052735, 0.10941337585449219, 0.11207475280761718, 0.11443523406982421, 0.11385327911376954, 0.11339981079101563, 0.10900991821289062, 0.11181568145751954, 0.11428659057617188, 0.1143521270751953, 0.11432550048828125, 0.11373772430419922, 0.11423641967773437, 0.11468800354003907, 0.11422003173828126, 0.11440128326416016, 0.11581855773925781, 0.114831298828125, 0.11461631774902344, 0.10952703857421875, 0.10920448303222656, 0.10923725128173828, 0.1092147216796875, 0.10930483245849609, 0.10826956939697266, 0.10965110778808594, 0.10926882934570313, 0.10897100830078126, 0.1092669448852539, 0.10960076904296875, 0.10913702392578126, 0.1092842254638672, 0.10949427032470703, 0.10912461090087891, 0.11292467498779298, 0.11453145599365235, 0.1093180160522461, 0.10825215911865234, 0.10807705688476563, 0.10928230285644532, 0.1099777603149414, 0.11510460662841797, 0.11429273223876953, 0.11425587463378906, 0.11452416229248047, 0.11407257843017578, 0.11408809661865234, 0.22967485046386718, 0.11443814086914063, 0.11416780853271484, 0.10911129760742187, 0.10907852935791015, 0.10933760070800781, 0.1091409912109375, 0.11380326080322266, 0.11439615631103515, 0.11469004821777344, 0.11373875427246094, 0.11358003234863281, 0.11431423950195313, 0.11489801788330078, 0.11466127777099609, 0.11431935882568359, 0.11436236572265625, 0.113728515625, 0.11315507507324218, 0.11451699066162109, 0.1169100799560547, 0.11009945678710938, 0.11031961822509766, 0.11431436920166016, 0.114347900390625, 0.11405824279785157, 0.10894028472900391, 0.10882064056396484, 0.10885104370117188, 0.10847743988037109, 0.11119615936279297, 0.11461222076416015, 0.11459276580810547, 0.10855628967285157, 0.1089617919921875, 0.10799423980712891, 0.10873535919189453, 0.10893619537353516, 0.10926898956298828, 0.1090355224609375, 0.10907548522949219, 0.10887062072753906, 0.10884403228759766, 0.10876640319824218, 0.10888275146484375, 0.10893824005126954, 0.10893106842041016, 0.10873856353759766, 0.1093570556640625, 0.10934989166259766, 0.10941852569580078, 0.10904367828369141, 0.10888703918457031, 0.10799520111083985, 0.10922294616699219, 0.10902006530761718, 0.10892813110351562, 0.10951052856445312, 0.10831974029541015, 0.10967346954345703, 0.10863104248046875, 0.10841600036621094, 0.10805248260498047, 0.21936962890625, 0.10905072021484374, 0.10877951812744141, 0.10939612579345703, 0.11030210876464844, 0.10973494720458984, 0.10921564483642578, 0.10962655639648437, 0.10945414733886719, 0.10931507110595703, 0.11006777954101563, 0.1095608673095703, 0.1089606704711914, 0.10962022399902344, 0.10924134063720703, 0.11083878326416016, 0.11015270233154296, 0.11127808380126954, 0.10987007904052734, 0.11152178955078125, 0.11219558715820313, 0.11285298919677735, 0.11157094573974609, 0.11294412994384766, 0.11082857513427734, 0.11008934020996093, 0.11014230346679688, 0.10940415954589844, 0.10990489959716797, 0.11050291442871094, 0.111246337890625, 0.10958748626708985, 0.10894332885742188, 0.11398758697509766, 0.11205939483642578, 0.11079373168945313, 0.10986102294921875, 0.11015663909912109, 0.10997350311279297, 0.10968370819091797, 0.10954150390625, 0.10995289611816406, 0.10949136352539063, 0.11010543823242187, 0.1095198745727539, 0.10970317077636718, 0.11119411468505859, 0.11122278594970703, 0.11173580932617187, 0.11004108428955078, 0.11049983978271484, 0.10967346954345703, 0.1093355484008789, 0.10924031829833984, 0.10934067535400391, 0.10920345306396484, 0.11221417236328125, 0.10958022308349609, 0.11011062622070313, 0.1108153305053711, 0.1098628158569336, 0.11008112335205078, 0.11116124725341797, 0.2201324462890625, 0.10866678619384766, 0.10969395446777344, 0.11017830657958984, 0.10882355499267578, 0.10950348663330078, 0.10896588897705078, 0.10886354827880859, 0.1088941421508789, 0.11057782745361328, 0.10997449493408203, 0.11107315063476562, 0.11106610870361328, 0.11104768371582031, 0.11013734436035157, 0.11028173065185547, 0.11172659301757812, 0.11009126281738281, 0.10883993530273438, 0.10915634918212891, 0.10995609283447266, 0.11161497497558594, 0.11087564849853515, 0.11196006774902344, 0.11103135681152344, 0.10995196533203125, 0.10915734100341797, 0.10895769500732422, 0.10926387023925781, 0.1091962890625, 0.11077855682373047, 0.11132498931884766, 0.1108255386352539, 0.11048544311523438, 0.10963455963134766, 0.10920755004882812, 0.10976870727539062, 0.10876739501953125, 0.10921660614013672, 0.10960076904296875, 0.11107123565673828, 0.10986614227294922, 0.10970915222167969, 0.10937139129638672, 0.10984754943847656, 0.10955570983886719, 0.10974720001220703, 0.10898534393310547, 0.10956390380859375, 0.10907033538818359, 0.10920252990722656, 0.10917673492431641, 0.10951475524902343, 0.10841190338134765, 0.10893619537353516, 0.10921369934082031, 0.10864230346679687, 0.10815283203125, 0.10967961883544922, 0.10897929382324219, 0.10957405090332031, 0.1100882568359375, 0.10977069091796875, 0.22062387084960938, 0.10923622131347656, 0.10986700439453125, 0.10914508819580078, 0.10911129760742187, 0.10907443237304687, 0.10923827362060547, 0.10929878234863281, 0.10911427307128906, 0.10916556549072266, 0.1092300796508789, 0.11362406158447266, 0.10924966430664063, 0.10959356689453124, 0.10878352355957031, 0.10925055694580078, 0.10984857940673828, 0.10921183776855468, 0.10818335723876953, 0.1083904037475586, 0.1098076171875, 0.11136307525634766, 0.10997760009765625, 0.11015372467041015, 0.1097359390258789, 0.11153202819824219, 0.11074150085449219, 0.1112125473022461, 0.11117977905273438, 0.11013529968261719, 0.1104168930053711, 0.10958643341064453, 0.11040358734130859, 0.1110640640258789, 0.11034735870361329, 0.11026428985595703, 0.11137529754638673, 0.11061555480957032, 0.11044457244873047, 0.11145929718017578, 0.11038822174072266, 0.11043840026855468, 0.11042098999023438, 0.11041494750976563, 0.11013417816162109, 0.11002982330322265, 0.1102572479248047, 0.11092982482910156, 0.11161087799072265, 0.11030937957763672, 0.10974002838134765, 0.11052134704589844, 0.10913484954833984, 0.11095859527587891, 0.10994073486328125, 0.11006156921386719, 0.10990819549560547, 0.11116726684570312, 0.11080499267578126, 0.10931199645996094, 0.10945331573486328, 0.10948095703125, 0.10905497741699219, 0.21976371765136718, 0.10950656127929688, 0.10864640045166016, 0.10917488098144532, 0.10903849792480469, 0.10977279663085937, 0.10949222564697265, 0.10977177429199218, 0.11040255737304687, 0.11123609924316406, 0.10997254180908203, 0.11027552032470703, 0.109949951171875, 0.1099335708618164, 0.11008512115478515, 0.10886246490478516, 0.1095198745727539, 0.10939801788330078, 0.10926387023925781, 0.1093355484008789, 0.10925772857666016, 0.10940108489990234, 0.10928742218017579, 0.11002162933349609, 0.11044863891601563, 0.10954956817626953, 0.10945024108886718, 0.10851737976074219, 0.10840585327148437, 0.10940815734863281, 0.10896998596191407, 0.10927616119384766, 0.10900070190429688, 0.10885222625732421, 0.10935100555419922, 0.10901801300048829, 0.1083135986328125, 0.11027769470214843, 0.11111929321289063, 0.11061248016357422, 0.10994483184814453, 0.10949855804443359, 0.10993952178955078, 0.10911949157714844, 0.10915235137939452, 0.10907433319091797, 0.10904064178466796, 0.10966937255859376, 0.10942259216308593, 0.11000422668457031, 0.10967961883544922, 0.10946355438232422, 0.10929663848876953, 0.10973900604248046, 0.10924031829833984, 0.10853580474853515, 0.10929766082763671, 0.10929561614990234, 0.10963471984863281, 0.10994057464599609, 0.10967756652832031, 0.10927718353271484, 0.10911334228515625]",tokens/s,8.942598026292986,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5573,7 +5573,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5772,7 +5772,7 @@ OSError: M is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5901,7 +5901,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6200,7 +6200,7 @@ Traceback (most recent call last): OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6337,7 +6337,7 @@ If this is a private repository, make sure to pass a token having permission to raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") RuntimeError: Isolated process exited with non-zero code -9 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6465,7 +6465,7 @@ OSError: google/recurrentgemma-7b is not a local folder and is not a valid model If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6494,7 +6494,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6818,7 +6818,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6888,7 +6888,7 @@ ValueError: GPT2LMHeadModel does not support an attention implementation through raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") RuntimeError: Isolated process exited with non-zero code -9 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7190,7 +7190,7 @@ OSError: 0 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7490,7 +7490,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 192.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7773,7 +7773,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1278.62784,9043.443712,0.0,8396.996608,8007.886848,s,10,5.840388916015625,0.5840388916015625,0.00031264013819856207,0.58402783203125,0.5843611145019532,0.5845285186767578,0.5846624420166016,"[0.5846959228515625, 0.5836339111328125, 0.583645263671875, 0.58408447265625, 0.583726318359375, 0.5843239135742188, 0.5839649047851563, 0.5841549682617188, 0.5841880493164062, 0.58397119140625]",tokens/s,438.32697390749433,kWh,6.9056205911400885e-06,3.784001771903907e-06,4.5127597830444617e-05,5.581722019348861e-05,tokens/kWh,4586398.231810616,MB,1278.939136,9043.443712,0.0,8396.996608,8007.889408,s,10,50.830269531249996,5.083026953125,0.03837433694757469,5.0823950195312495,5.143792578125,5.144825732421875,5.145652255859375,"[5.0555966796875, 5.036369140625, 5.14585888671875, 5.090703125, 5.06720361328125, 5.09485107421875, 5.14356298828125, 5.02436962890625, 5.09766748046875, 5.0740869140625]",tokens/s,12.394189639554865,kWh,6.210882268386123e-05,3.4040414997440776e-05,0.00018252019539875467,0.00027866943308005673,tokens/kWh,226074.3107117214,,s,629,56.00362285614008,0.08903596638496047,0.0694670532025038,0.08005836486816406,0.08282109374999999,0.0833206298828125,0.6644542456054687,"[0.08291020965576172, 0.08056524658203125, 0.08002457427978515, 0.07959040069580078, 0.0796693115234375, 0.07963231658935546, 0.08058265686035156, 0.07995398712158203, 0.07994054412841797, 0.07959449768066407, 0.07967436981201172, 0.08162611389160156, 0.08048947143554687, 0.08033484649658203, 0.08009318542480469, 0.07970816040039062, 0.07974400329589844, 0.08013619232177735, 0.07984230041503906, 0.079494140625, 0.07952486419677735, 0.07967747497558594, 0.07991907501220703, 0.08194457244873046, 0.07954841613769531, 0.08305152130126953, 0.0806789093017578, 0.08011980438232422, 0.07993138885498047, 0.0792647705078125, 0.07913075256347656, 0.07912230682373046, 0.07862374114990234, 0.08174591827392579, 0.08013619232177735, 0.07978803253173829, 0.08090726470947265, 0.0791021728515625, 0.07962089538574219, 0.08049459075927734, 0.080321533203125, 0.07973273468017578, 0.07954246520996094, 0.07965580749511719, 0.07988320159912109, 0.07952384185791016, 0.07985356903076171, 0.07994163513183594, 0.08069222259521484, 0.08007782745361328, 0.07992934417724609, 0.0801269760131836, 0.08028057861328125, 0.07971532440185547, 0.07988326263427735, 0.08096870422363281, 0.07999078369140625, 0.07964672088623047, 0.08389119720458985, 0.08078643035888672, 0.07970304107666015, 0.0828436508178711, 0.6643097534179687, 0.07870771026611328, 0.07853465270996093, 0.07933235168457031, 0.08077823638916015, 0.07940096282958985, 0.07982899475097656, 0.07922688293457031, 0.07879167938232422, 0.08144486236572265, 0.07970304107666015, 0.07991193389892579, 0.07885107421875, 0.079685791015625, 0.07888390350341797, 0.07853340911865235, 0.07944601440429687, 0.07975017547607421, 0.07975939178466797, 0.0790025634765625, 0.07950745391845702, 0.07938457489013671, 0.07933849334716797, 0.07951974487304687, 0.07970918273925781, 0.07965081787109375, 0.07987814331054688, 0.0804290542602539, 0.08215756988525391, 0.08098918151855469, 0.07969996643066406, 0.07955865478515625, 0.08080998229980468, 0.07991705322265626, 0.07958220672607422, 0.07961615753173829, 0.07944892883300782, 0.079678466796875, 0.07969280242919922, 0.07978905487060547, 0.07977267456054687, 0.07979110717773437, 0.08038297271728516, 0.0799836196899414, 0.0799447021484375, 0.08022630310058594, 0.08001126098632813, 0.07973580932617187, 0.079963134765625, 0.0801976318359375, 0.07974002838134765, 0.08232947540283203, 0.08232556915283203, 0.08240326690673828, 0.07957708740234375, 0.079889404296875, 0.07975424194335938, 0.07997132873535157, 0.08189849853515625, 0.0800083236694336, 0.07979814147949219, 0.08032780456542969, 0.07981247711181641, 0.6645104370117187, 0.07945116424560547, 0.07980745697021484, 0.08049372863769531, 0.08225571441650391, 0.08387789154052734, 0.08159334564208984, 0.08176742553710938, 0.08168358612060547, 0.08321932983398438, 0.08420556640625, 0.0833095703125, 0.08297471618652344, 0.08393727874755859, 0.08226719665527343, 0.08292755126953125, 0.08297062683105469, 0.08239718627929687, 0.08227225494384766, 0.08287744140625, 0.08160562896728515, 0.08292262268066407, 0.08278412628173829, 0.08336895751953124, 0.083093505859375, 0.08215961456298829, 0.08168141174316407, 0.08241356658935547, 0.08153190612792968, 0.08139263916015625, 0.08263270568847657, 0.0827146224975586, 0.08284671783447266, 0.08262963104248047, 0.08158739471435547, 0.08068998718261719, 0.08006147003173827, 0.07953199768066406, 0.08057344055175782, 0.08217497253417969, 0.08129334259033204, 0.07968150329589843, 0.08222822570800781, 0.08382361602783203, 0.08077005004882812, 0.08278937530517579, 0.08146636962890624, 0.08007884979248046, 0.07962009429931641, 0.0794972152709961, 0.07905075073242188, 0.08247808074951171, 0.08189542388916016, 0.07984246063232422, 0.08182665252685548, 0.07992626953125, 0.0792063980102539, 0.07983718109130859, 0.07929138946533203, 0.08002355194091797, 0.0825907211303711, 0.08137318420410156, 0.0836648941040039, 0.6666311645507812, 0.08276377868652343, 0.0831283187866211, 0.08124518585205077, 0.08178079986572266, 0.08299005126953125, 0.08289686584472657, 0.08132406616210938, 0.08445436859130859, 0.0831272964477539, 0.08397516632080078, 0.08128819274902344, 0.07980032348632812, 0.07996109008789062, 0.08095142364501953, 0.08318252563476562, 0.07973983764648437, 0.07981568145751954, 0.079857666015625, 0.07976038360595702, 0.07997132873535157, 0.07947468566894532, 0.08273101043701171, 0.07991500854492188, 0.08001945495605468, 0.08004096221923829, 0.08232653045654297, 0.08080793762207031, 0.0833280029296875, 0.0800030746459961, 0.08251801300048828, 0.08140198516845704, 0.0787720947265625, 0.078906494140625, 0.07899021148681641, 0.07861350250244141, 0.078814208984375, 0.07860838317871094, 0.07851929473876954, 0.0791377944946289, 0.07885107421875, 0.07858073425292969, 0.07909894561767578, 0.07961798095703125, 0.07994367980957032, 0.07981568145751954, 0.07975846099853516, 0.08274419403076172, 0.08241458892822266, 0.08255181121826172, 0.08268902587890625, 0.08263782501220703, 0.08236339569091797, 0.08257331085205079, 0.07978905487060547, 0.08213094329833984, 0.07936409759521484, 0.08117453002929688, 0.07902207946777344, 0.07913075256347656, 0.07941107177734374, 0.08015574645996094, 0.07954013061523438, 0.664648681640625, 0.08135475158691406, 0.08279551696777344, 0.08241458892822266, 0.08013833618164062, 0.08009513854980468, 0.08086835479736328, 0.07994060516357422, 0.07903855895996094, 0.07974800109863281, 0.07907635498046875, 0.07896883392333984, 0.07869036865234375, 0.07892070770263672, 0.07922271728515624, 0.07875686645507812, 0.07878457641601562, 0.07889810943603516, 0.07901081848144531, 0.07868927764892578, 0.07878041839599609, 0.07896166229248047, 0.07916441345214843, 0.08148070526123047, 0.07964374542236329, 0.08013609313964844, 0.08106393432617187, 0.08005635070800782, 0.07976445007324219, 0.08264601898193359, 0.0807045135498047, 0.0799477767944336, 0.07977267456054687, 0.07972566223144531, 0.079891357421875, 0.08099430084228515, 0.08259481811523438, 0.08105590057373047, 0.07976131439208985, 0.08161068725585938, 0.08006451416015625, 0.082585693359375, 0.0798749771118164, 0.08114892578125, 0.08266652679443359, 0.08251593780517578, 0.08365977478027344, 0.0827320327758789, 0.07970201873779297, 0.07986790466308594, 0.07999897766113281, 0.0803594207763672, 0.07997952270507812, 0.07975647735595703, 0.07944898986816407, 0.0803439712524414, 0.08300236511230469, 0.08279551696777344, 0.08125234985351562, 0.08137625885009765, 0.07991705322265626, 0.08000511932373047, 0.08003276824951172, 0.6658119506835938, 0.07989965057373047, 0.08387174224853515, 0.08333312225341796, 0.08225791931152343, 0.08073420715332032, 0.07924018859863281, 0.07894322967529296, 0.08025599670410156, 0.08036659240722656, 0.07964466857910156, 0.08242278289794921, 0.08162611389160156, 0.07858380889892579, 0.08090521240234375, 0.07936409759521484, 0.08161804962158203, 0.07985858917236328, 0.07985456085205078, 0.0797306900024414, 0.07977584075927735, 0.08092867279052735, 0.0805580825805664, 0.08249241638183594, 0.08108134460449219, 0.07992217254638671, 0.07977788543701173, 0.08294697570800781, 0.0823900146484375, 0.08240025329589844, 0.08117772674560547, 0.07950323486328124, 0.08071782684326172, 0.0806666259765625, 0.0800716781616211, 0.07977273559570312, 0.0814273910522461, 0.08270028686523437, 0.08002662658691406, 0.07989759826660156, 0.08005836486816406, 0.08213196563720702, 0.08035020446777344, 0.07973273468017578, 0.07981568145751954, 0.08250163269042969, 0.07970201873779297, 0.08253132629394531, 0.08337407684326172, 0.08212889862060548, 0.07963136291503906, 0.08239411163330078, 0.08214937591552735, 0.07937433624267579, 0.0797870101928711, 0.08129228973388672, 0.0810855712890625, 0.0787608642578125, 0.08034198760986327, 0.08283443450927734, 0.08003276824951172, 0.08074240112304687, 0.08194457244873046, 0.6642647094726563, 0.0794993896484375, 0.0797940444946289, 0.08199167633056641, 0.08071167755126953, 0.08071174621582031, 0.08195065307617187, 0.08254873657226562, 0.08254054260253907, 0.08443405151367188, 0.08217689514160156, 0.08004720306396484, 0.08074435424804688, 0.08279039764404297, 0.082482177734375, 0.08118275451660156, 0.08172029113769531, 0.0808663330078125, 0.08280879974365235, 0.08311090850830079, 0.08071577453613281, 0.08034099578857422, 0.08230809783935547, 0.08026624298095703, 0.08235929870605468, 0.08276377868652343, 0.08247193908691407, 0.0801269760131836, 0.08438886260986328, 0.08388607788085937, 0.08226509094238281, 0.0826961898803711, 0.08268697357177734, 0.08248320007324218, 0.08162406158447266, 0.08251100921630859, 0.0821912612915039, 0.08259782409667969, 0.08264002990722656, 0.08281801605224609, 0.08292953491210937, 0.08166400146484375, 0.0824453125, 0.08283647918701172, 0.08246476745605469, 0.08106495666503906, 0.08014540863037109, 0.0838656005859375, 0.08050688171386719, 0.07997337341308594, 0.08045465850830077, 0.08002047729492187, 0.08176947021484375, 0.07960575866699218, 0.07915641784667969, 0.08000083160400391, 0.08100454711914062, 0.08247193908691407, 0.08261222076416015, 0.08169471740722656, 0.08025804901123047, 0.08014848327636719, 0.07995807647705078, 0.6648954467773438, 0.08304946899414062, 0.08025702667236329, 0.07931305694580078, 0.07981142425537109, 0.07934464263916016, 0.07946556854248046, 0.07945206451416016, 0.07950233459472657, 0.07922898864746093, 0.07977772521972656, 0.07961190032958984, 0.07945830535888672, 0.07940300750732422, 0.07993446350097656, 0.07959347534179688, 0.08008822631835938, 0.08004489898681641, 0.0799477767944336, 0.07949017333984375, 0.07983500671386719, 0.07953817749023437, 0.07943475341796875, 0.07895756530761719, 0.07932723236083984, 0.0797675552368164, 0.07883468627929688, 0.07859814453125, 0.07889715576171875, 0.07878860473632812, 0.07981977844238282, 0.0805898208618164, 0.07976959991455078, 0.07911833953857422, 0.079087646484375, 0.07890735626220703, 0.07930572509765625, 0.07932723236083984, 0.0800030746459961, 0.08006758117675782, 0.07967948913574219, 0.07968256378173828, 0.07989043426513671, 0.07972557067871094, 0.07980748748779297, 0.07971430206298828, 0.07981986999511718, 0.07977565002441406, 0.07982387542724609, 0.07960275268554687, 0.0798094711303711, 0.07985254669189454, 0.07983411407470703, 0.0797635498046875, 0.07966812896728516, 0.08000316619873046, 0.07971625518798828, 0.08070963287353515, 0.07998770904541015, 0.07977779388427735, 0.08144793701171875, 0.08023757171630859, 0.08048230743408204, 0.6673449096679688, 0.08292153930664062, 0.08312620544433594, 0.08225791931152343, 0.08246991729736328, 0.07978902435302734, 0.08033484649658203, 0.0824668197631836, 0.08165888214111328, 0.0805775375366211, 0.08234086608886719, 0.08295014190673829, 0.08257331085205079, 0.08338845062255859, 0.08021807861328124, 0.07925247955322266, 0.07979417419433593, 0.08004812622070312, 0.07905587005615235, 0.07926783752441406, 0.07897293090820312, 0.07880397033691407, 0.078993408203125, 0.07901798248291016, 0.0785090560913086, 0.0800184326171875, 0.07927398681640625, 0.07920435333251953, 0.0795125732421875, 0.08040447998046875, 0.07980032348632812, 0.08294400024414063, 0.08279961395263671, 0.07977062225341797, 0.07976448059082031, 0.07961804962158203, 0.08125030517578125, 0.08011587524414063, 0.08056829071044921, 0.08093682861328125, 0.08159334564208984, 0.07977471923828125, 0.08167529296875, 0.08220054626464844, 0.08044544219970703, 0.07977881622314453, 0.08283340454101562, 0.08029593658447266, 0.08064205169677735, 0.08343244934082031, 0.07977062225341797, 0.08119808197021484, 0.08234700775146485, 0.08169267272949218, 0.07956684875488282, 0.08198963165283203, 0.0824791030883789, 0.08178278350830079, 0.07999282836914062, 0.08173567962646484, 0.07975628662109376, 0.08161500549316406, 0.08327356719970704, 0.6670172119140625, 0.08281702423095703, 0.08259891510009766, 0.08108860778808594, 0.080001953125, 0.08203059387207032, 0.08035123443603516, 0.08262963104248047, 0.07965491485595703, 0.07912652587890626, 0.07961395263671875, 0.07980032348632812, 0.08106905364990234, 0.08265113830566406, 0.08179507446289062, 0.08528998565673829, 0.08327986907958984, 0.08008719635009766, 0.07972438049316406, 0.08000921630859376, 0.07971430206298828, 0.07974400329589844, 0.08374066925048829, 0.0815472640991211, 0.0802498550415039, 0.08239933013916016, 0.08024361419677735, 0.07971635437011719, 0.08020275115966796, 0.07980850982666016, 0.07986495971679687, 0.07986675262451172, 0.07992425537109375, 0.07984226989746093, 0.07981056213378906, 0.07977267456054687, 0.07981362915039063, 0.07961293029785156, 0.07974604797363281, 0.0794788818359375, 0.08003065490722656, 0.08018531036376954, 0.08001638031005859, 0.07974297332763672, 0.0799836196899414, 0.07964774322509766, 0.0835389404296875, 0.08009625244140625, 0.07970816040039062, 0.08071987152099609, 0.08007286071777343, 0.07983087921142579, 0.07964672088623047, 0.0823377914428711, 0.07989055633544923, 0.07954220581054687, 0.07942649841308594, 0.07968681335449218, 0.07980323028564452, 0.07957810974121093, 0.07954227447509765, 0.07969075012207032, 0.07958834838867188]",tokens/s,11.231416253476137,,, -4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7802,7 +7802,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8296,7 +8296,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpjy906s3s/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9237,7 +9237,7 @@ ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please req ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1163.595776,1013.448704,0.0,367.0016,281.023488,s,10,0.4671973457336425,0.046719734573364265,0.0011570082478510138,0.04616308784484863,0.04822501564025879,0.048944972038269044,0.04952093715667724,"[0.049664928436279294, 0.04617475128173828, 0.048065025329589846, 0.04615142440795898, 0.04686195373535156, 0.04602249526977539, 0.046215328216552734, 0.04605571365356445, 0.04610313415527344, 0.04588259124755859]",tokens/s,5479.483184948359,kWh,5.603492016727012e-07,3.0704557188358005e-07,1.7518036657914398e-06,2.6191984393477212e-06,tokens/kWh,97739826.10639979,MB,1163.595776,1013.448704,0.0,367.0016,298.883584,s,10,23.53951806640625,2.3539518066406253,0.02066314872443343,2.346360595703125,2.384883740234375,2.386660009765625,2.388081025390625,"[2.37219873046875, 2.384489013671875, 2.388436279296875, 2.340317626953125, 2.3628896484375, 2.337561279296875, 2.341433349609375, 2.3329326171875, 2.3279716796875, 2.351287841796875]",tokens/s,26.76350459778896,kWh,2.7878720072726313e-05,1.5278461452286933e-05,4.256164845160901e-05,8.571882997662227e-05,tokens/kWh,734961.035016247,,s,629,23.918093322753908,0.038025585568766146,0.005597171597052502,0.03703705596923828,0.038441985321044925,0.03860725631713867,0.08344375122070312,"[0.03803852844238281, 0.0382740478515625, 0.03829350280761719, 0.03800985717773438, 0.03834777450561523, 0.03843993759155274, 0.03794124984741211, 0.038307838439941407, 0.038882366180419924, 0.04038547134399414, 0.0395335693359375, 0.03956838226318359, 0.04239769744873047, 0.04064460754394531, 0.03885158538818359, 0.0383631362915039, 0.0383559684753418, 0.036988929748535154, 0.03714355087280274, 0.03709747314453125, 0.03706982421875, 0.037000190734863284, 0.036923393249511716, 0.03702783966064453, 0.03697151947021484, 0.036959232330322264, 0.03696230316162109, 0.036934654235839845, 0.037370880126953124, 0.03696332931518555, 0.037574657440185545, 0.03700940704345703, 0.03705548858642578, 0.036994049072265625, 0.03704115295410156, 0.03687116622924805, 0.03696025466918945, 0.03684454345703125, 0.03705036926269531, 0.0370247688293457, 0.036929534912109374, 0.03713433456420898, 0.037039134979248045, 0.03718755340576172, 0.03721011352539062, 0.03724800109863281, 0.037184513092041016, 0.0372305908203125, 0.03714048004150391, 0.037340160369873046, 0.03721932983398438, 0.03755929565429687, 0.03708313751220703, 0.03727052688598633, 0.037136383056640625, 0.03713740921020508, 0.03725823974609375, 0.03719987106323242, 0.037177345275878904, 0.03729919815063477, 0.037303295135498044, 0.03718348693847656, 0.08368537902832031, 0.0370882568359375, 0.03700428771972656, 0.03704217529296875, 0.03706572723388672, 0.03709439849853516, 0.036999168395996096, 0.03710771179199219, 0.03708108901977539, 0.0371517448425293, 0.03712921524047851, 0.0370964469909668, 0.037136383056640625, 0.03711078262329102, 0.037198848724365234, 0.0369172477722168, 0.03712614440917969, 0.0370780143737793, 0.037136383056640625, 0.03702783966064453, 0.03716304016113281, 0.03708924865722656, 0.036947967529296875, 0.03715071868896484, 0.03827097702026367, 0.03841740798950195, 0.03755110549926758, 0.037015552520751956, 0.03714355087280274, 0.03705241775512695, 0.03820646286010742, 0.03845529556274414, 0.038426654815673825, 0.038419422149658204, 0.03842764663696289, 0.03850239944458008, 0.03845939254760742, 0.03841641616821289, 0.038408161163330075, 0.038432769775390625, 0.03845017623901367, 0.03847884750366211, 0.03835801696777344, 0.03837235260009766, 0.038424575805664066, 0.038474750518798825, 0.03832524871826172, 0.038412353515625, 0.038470592498779294, 0.03872358322143555, 0.038209537506103515, 0.03858432006835937, 0.03842559814453125, 0.03840921783447265, 0.038403072357177735, 0.03843993759155274, 0.03840512084960938, 0.038558719635009765, 0.03856076812744141, 0.038487041473388675, 0.03853619384765625, 0.03842051315307617, 0.03853718566894531, 0.08692018890380859, 0.038454273223876956, 0.03861094284057617, 0.03851878356933594, 0.0384634895324707, 0.038525951385498046, 0.0385269775390625, 0.03856076812744141, 0.03888435363769531, 0.038889537811279296, 0.03855251312255859, 0.03807436752319336, 0.038642688751220705, 0.038752254486083985, 0.038556671142578124, 0.038563838958740236, 0.03845017623901367, 0.03842355346679688, 0.03845119857788086, 0.038529022216796875, 0.03855155181884766, 0.03860172653198242, 0.03840921783447265, 0.038422527313232424, 0.03867238235473633, 0.038569984436035154, 0.0395335693359375, 0.03974758529663086, 0.03854950332641602, 0.03836928176879883, 0.038182910919189454, 0.038389759063720705, 0.037166080474853515, 0.03708313751220703, 0.036751361846923826, 0.03750400161743164, 0.03920896148681641, 0.03729817581176758, 0.03704422378540039, 0.03711795043945312, 0.03697459030151367, 0.036983806610107424, 0.036915199279785156, 0.037182464599609374, 0.03705241775512695, 0.036975616455078124, 0.036835391998291014, 0.03715372848510742, 0.036956161499023435, 0.0370145263671875, 0.036915199279785156, 0.036985855102539066, 0.036967422485351564, 0.036992000579833983, 0.03690086364746094, 0.03713846588134766, 0.03875017547607422, 0.037133312225341795, 0.03691622543334961, 0.03701145553588867, 0.038316032409667966, 0.03706163024902344, 0.036893695831298826, 0.08358297729492188, 0.03701964950561523, 0.03696332931518555, 0.036910079956054685, 0.03696230316162109, 0.036822017669677735, 0.036944896697998046, 0.03696537780761719, 0.03691526412963867, 0.03704313659667969, 0.0373309440612793, 0.03702374267578125, 0.037005313873291014, 0.03789619064331055, 0.03707699203491211, 0.03742105484008789, 0.036997119903564454, 0.036913150787353514, 0.037082111358642575, 0.03712921524047851, 0.03707699203491211, 0.03715686416625977, 0.036983806610107424, 0.036928512573242187, 0.037238784790039066, 0.036972545623779295, 0.037013504028320314, 0.03707596969604492, 0.03713228988647461, 0.03693670272827149, 0.03709337615966797, 0.03693894577026367, 0.036943679809570314, 0.037005313873291014, 0.037048320770263675, 0.036988929748535154, 0.03708927917480469, 0.03729100799560547, 0.03712614440917969, 0.03694387054443359, 0.037043201446533204, 0.03701964950561523, 0.037028865814208986, 0.03678412628173828, 0.03706163024902344, 0.036999168395996096, 0.03909222412109375, 0.0383375358581543, 0.038348800659179685, 0.03747840118408203, 0.03827916717529297, 0.03701862335205078, 0.03709439849853516, 0.03715999984741211, 0.037053375244140624, 0.036759552001953126, 0.03706060791015625, 0.037103614807128905, 0.03692031860351563, 0.036931583404541016, 0.03708620834350586, 0.036959232330322264, 0.03689267349243164, 0.08558284759521484, 0.03703500747680664, 0.03761971282958984, 0.03717331314086914, 0.03696428680419922, 0.03712716674804688, 0.037000190734863284, 0.03735039901733399, 0.03775590515136719, 0.03703910446166992, 0.036994049072265625, 0.037101566314697264, 0.03697971343994141, 0.037577728271484374, 0.03808256149291992, 0.03833446502685547, 0.03837542343139649, 0.038368255615234374, 0.03834368133544922, 0.03841331100463867, 0.03831193542480469, 0.03829660797119141, 0.03825353622436523, 0.038214656829833986, 0.037924896240234374, 0.03696547317504883, 0.0373072624206543, 0.03739136123657227, 0.03703910446166992, 0.03698483276367188, 0.03702579116821289, 0.037005313873291014, 0.03697971343994141, 0.036994049072265625, 0.03692544174194336, 0.036983806610107424, 0.037032958984375, 0.03678310394287109, 0.03704012680053711, 0.0369879035949707, 0.037144577026367184, 0.03695929718017578, 0.03798009490966797, 0.03834470367431641, 0.038267902374267575, 0.0382371826171875, 0.03833760070800781, 0.03831289672851562, 0.038329345703125, 0.038100990295410156, 0.03827609634399414, 0.037138431549072266, 0.03688550567626953, 0.037359615325927735, 0.037032958984375, 0.03762790298461914, 0.03760332870483398, 0.03706880187988281, 0.037000190734863284, 0.03702272033691406, 0.03696844863891602, 0.036953086853027346, 0.037152801513671875, 0.0834436798095703, 0.03685171127319336, 0.036947967529296875, 0.03693363189697266, 0.03692236709594727, 0.0370145263671875, 0.036999168395996096, 0.03806412887573242, 0.03839078521728516, 0.03811328125, 0.03750707244873047, 0.036795391082763675, 0.03697459030151367, 0.036928512573242187, 0.036959232330322264, 0.037256191253662106, 0.03716198348999023, 0.0369879035949707, 0.03698175811767578, 0.036964351654052735, 0.036969470977783206, 0.03703807830810547, 0.03692134475708008, 0.03673088073730469, 0.036985855102539066, 0.036967422485351564, 0.03695718383789062, 0.036985855102539066, 0.03688652801513672, 0.037160961151123044, 0.03701964950561523, 0.036877311706542966, 0.03700326538085937, 0.037179393768310545, 0.037615615844726565, 0.03835801696777344, 0.036985855102539066, 0.03703500747680664, 0.03696230316162109, 0.03696844863891602, 0.03697151947021484, 0.0369951057434082, 0.036846561431884764, 0.03836723327636719, 0.0373125114440918, 0.037157886505126955, 0.03683737564086914, 0.036945919036865234, 0.036952129364013674, 0.03695302581787109, 0.036934654235839845, 0.03704118347167969, 0.036977630615234375, 0.03694079971313476, 0.03695001602172852, 0.03692748641967773, 0.036947967529296875, 0.03690291213989258, 0.03684249496459961, 0.03698483276367188, 0.03706675338745117, 0.03695001602172852, 0.0370186882019043, 0.08373343658447266, 0.03699609756469727, 0.036988929748535154, 0.03689267349243164, 0.03697663879394531, 0.0370964469909668, 0.036863998413085936, 0.036975616455078124, 0.0369541130065918, 0.03684864044189453, 0.037108734130859376, 0.03701862335205078, 0.03702169418334961, 0.03710464096069336, 0.03711590576171875, 0.03706982421875, 0.03699609756469727, 0.036983806610107424, 0.037285888671875, 0.038042625427246096, 0.03699302291870117, 0.03704217529296875, 0.037171199798583986, 0.037028865814208986, 0.03702272033691406, 0.03697459030151367, 0.03704217529296875, 0.03697868728637695, 0.03697459030151367, 0.037045310974121094, 0.03697964859008789, 0.03701145553588867, 0.03688652801513672, 0.037010433197021485, 0.03699097442626953, 0.036953086853027346, 0.037054462432861326, 0.03701657485961914, 0.03993395233154297, 0.038919166564941404, 0.03830579376220703, 0.038316032409667966, 0.03747020721435547, 0.03695532989501953, 0.036810558319091795, 0.03705241775512695, 0.037217281341552735, 0.03750092697143555, 0.037200897216796876, 0.03711897659301758, 0.03704217529296875, 0.03701760101318359, 0.03696332931518555, 0.03698080062866211, 0.037045215606689455, 0.03683734512329102, 0.03688243103027344, 0.03695820617675781, 0.0369172477722168, 0.03704729461669922, 0.03702374267578125, 0.036967422485351564, 0.03693875122070313, 0.08346521759033203, 0.03698483276367188, 0.0369244155883789, 0.0367534065246582, 0.036908031463623044, 0.036956161499023435, 0.0369356803894043, 0.0368455696105957, 0.036999168395996096, 0.03702272033691406, 0.036980735778808595, 0.03696332931518555, 0.03690598297119141, 0.03693056106567383, 0.03687936019897461, 0.036772865295410156, 0.036972545623779295, 0.03696537780761719, 0.03704729461669922, 0.036945919036865234, 0.036972545623779295, 0.036951038360595705, 0.037005313873291014, 0.036931583404541016, 0.03696230316162109, 0.036994049072265625, 0.0369541130065918, 0.036790271759033204, 0.03719475173950195, 0.03707289505004883, 0.03682099151611328, 0.036913150787353514, 0.03749683380126953, 0.03820544052124023, 0.03815628814697265, 0.0382105598449707, 0.03713740921020508, 0.036939777374267575, 0.036915199279785156, 0.0370145263671875, 0.03706777572631836, 0.036746238708496096, 0.0369244155883789, 0.03686809539794922, 0.037048320770263675, 0.036931583404541016, 0.03693670272827149, 0.03696537780761719, 0.03699302291870117, 0.03704627227783203, 0.03698175811767578, 0.03697459030151367, 0.03689471817016601, 0.03681792068481445, 0.036951038360595705, 0.0370145263671875, 0.03688550567626953, 0.03695513534545898, 0.03731052780151367, 0.03704620742797852, 0.03695929718017578, 0.03697248077392578, 0.03693875122070313, 0.08344377899169922, 0.03677177429199219, 0.03709542465209961, 0.03705753707885742, 0.03696230316162109, 0.03700735855102539, 0.03692748641967773, 0.03697868728637695, 0.03692748641967773, 0.03690291213989258, 0.03694079971313476, 0.036863998413085936, 0.0370382080078125, 0.03672563171386719, 0.036961280822753906, 0.03697151947021484, 0.03692236709594727, 0.036929534912109374, 0.03697151947021484, 0.03686502456665039, 0.036860927581787106, 0.03703705596923828, 0.036835391998291014, 0.03689875030517578, 0.03679846572875976, 0.03678412628173828, 0.036929534912109374, 0.03693670272827149, 0.0368455696105957, 0.036983806610107424, 0.036967422485351564, 0.03689779281616211, 0.036926464080810545, 0.03690086364746094, 0.03829759979248047, 0.03696230316162109, 0.03669504165649414, 0.03691622543334961, 0.03679743957519531, 0.03689267349243164, 0.03695001602172852, 0.03690598297119141, 0.03686502456665039, 0.03692236709594727, 0.03696844863891602, 0.036910079956054685, 0.03692236709594727, 0.03697459030151367, 0.036842655181884766, 0.03685257720947266, 0.03689984130859375, 0.03681587219238281, 0.03703705596923828, 0.0370247688293457, 0.03685683059692383, 0.03693363189697266, 0.036908031463623044, 0.03697049713134765, 0.03744255828857422, 0.03701964950561523, 0.03685990524291992, 0.03689574432373047, 0.036898815155029296, 0.08322457885742188, 0.03692544174194336, 0.037465087890625, 0.03834268951416016, 0.03717014312744141, 0.03688652801513672, 0.03682918548583984, 0.03694079971313476, 0.03679641723632812, 0.03685273742675781, 0.036499614715576174, 0.03629961776733399, 0.03678617477416992, 0.03683327865600586, 0.03686297607421875, 0.03689984130859375, 0.03688652801513672, 0.03695206451416016, 0.03687628936767578, 0.03704934310913086, 0.03689267349243164, 0.03689478302001953, 0.0367891845703125, 0.03672063827514648, 0.03685580825805664, 0.03686809539794922, 0.036819969177246094, 0.03729305648803711, 0.03830169677734375, 0.03824025726318359, 0.038112255096435545, 0.03822489547729492, 0.03721830368041992, 0.03701862335205078, 0.036857856750488284, 0.03697663879394531, 0.03696640014648438, 0.03701657485961914, 0.03687526321411133, 0.03727360153198242, 0.03696230316162109, 0.03703705596923828, 0.036929534912109374, 0.03711078262329102, 0.036923393249511716, 0.037015552520751956, 0.03719270324707031, 0.039554046630859374, 0.03838873672485352, 0.03845939254760742, 0.038168575286865236, 0.03887104034423828, 0.03853414535522461, 0.037528575897216795, 0.0369879035949707, 0.03750912094116211, 0.038250495910644534, 0.03698995208740234, 0.03824025726318359, 0.03704329681396484, 0.037685150146484374, 0.03813785552978516, 0.03806412887573242]",tokens/s,26.2980828577007,,, 4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1176.236032,5247.598592,0.0,4601.151488,4334.675968,s,10,3.1631250305175787,0.31631250305175784,0.00022168596924691957,0.31634973144531253,0.3165389404296875,0.3165441589355469,0.3165483337402344,"[0.31653778076171873, 0.3162330932617187, 0.31654937744140627, 0.31643914794921874, 0.31627874755859375, 0.3157511291503906, 0.3162454833984375, 0.31642071533203125, 0.31621484375, 0.3164547119140625]",tokens/s,809.326212306287,kWh,3.738922491255734e-06,2.048716852237931e-06,2.411586130656243e-05,2.9903500650056095e-05,tokens/kWh,8560870.614976637,MB,1176.236032,5247.598592,0.0,4601.151488,4334.678528,s,10,31.734150146484378,3.173415014648438,0.04143043606077721,3.1842030029296877,3.2168704833984374,3.2169435424804687,3.2170019897460937,"[3.1371982421875, 3.08391748046875, 3.132481689453125, 3.216854248046875, 3.208049560546875, 3.2170166015625, 3.168064697265625, 3.177404541015625, 3.19100146484375, 3.20216162109375]",tokens/s,19.85243017670015,kWh,3.66868678915004e-05,2.0104653791335068e-05,9.635306840184085e-05,0.0001531445900846763,tokens/kWh,411375.941945884,,s,629,34.52952166748047,0.0548959009021947,0.0375264663982122,0.050331649780273435,0.0519317512512207,0.05230387268066406,0.3654816918945313,"[0.049326080322265625, 0.049463294982910154, 0.04948992156982422, 0.04940604782104492, 0.05011654281616211, 0.052609054565429685, 0.052092864990234376, 0.0489697265625, 0.04905574417114258, 0.04862464141845703, 0.04899123382568359, 0.04877721786499024, 0.04927897644042969, 0.0490557746887207, 0.04926665496826172, 0.049329151153564454, 0.049007614135742186, 0.04905267333984375, 0.049138687133789063, 0.04894003295898437, 0.04890726470947265, 0.048919551849365236, 0.049170433044433595, 0.04901580810546875, 0.04929024124145508, 0.05164851379394531, 0.052337665557861325, 0.05171712112426758, 0.05176115036010742, 0.051501056671142575, 0.051571712493896485, 0.04910489654541016, 0.04950425720214844, 0.04897587203979492, 0.04907014465332031, 0.04910483169555664, 0.0500398063659668, 0.04906502532958985, 0.04831532669067383, 0.048712703704833986, 0.04919398498535156, 0.04902297592163086, 0.04914585494995117, 0.04898406219482422, 0.048928768157958984, 0.048729087829589846, 0.04889395141601562, 0.048173057556152345, 0.05331763076782227, 0.05189734268188476, 0.04906086349487305, 0.048936958312988284, 0.052135936737060545, 0.05109145736694336, 0.051789825439453124, 0.0488540153503418, 0.051504127502441405, 0.05199052810668945, 0.04906291198730469, 0.04908544158935547, 0.04912639999389649, 0.049018878936767575, 0.3654072265625, 0.04859699249267578, 0.04905267333984375, 0.049023998260498046, 0.049462272644042966, 0.04878643035888672, 0.049097728729248044, 0.04836556625366211, 0.04893491363525391, 0.04986675262451172, 0.049426433563232425, 0.04912844848632812, 0.04902809524536133, 0.04909568023681641, 0.04930867385864258, 0.04921139144897461, 0.04876800155639648, 0.0484290885925293, 0.04903318405151367, 0.04900556945800781, 0.048546817779541014, 0.048996353149414064, 0.04876697540283203, 0.04880691146850586, 0.04930355072021484, 0.04892160034179688, 0.04907827377319336, 0.04904755020141602, 0.04905779266357422, 0.04895129776000977, 0.04845260620117187, 0.048395263671875, 0.04894828796386719, 0.04880070495605469, 0.049081344604492184, 0.04924518585205078, 0.04914688110351562, 0.0489881591796875, 0.049014785766601565, 0.04882534408569336, 0.049097728729248044, 0.049002494812011715, 0.04878438568115234, 0.04833280181884766, 0.04919500732421875, 0.048863231658935545, 0.049070079803466796, 0.048998401641845706, 0.04883967971801758, 0.04880179214477539, 0.04921548843383789, 0.04843212890625, 0.04830822372436523, 0.0484136962890625, 0.048514049530029295, 0.0491776008605957, 0.04942233657836914, 0.049111038208007815, 0.04896051025390625, 0.04899430465698242, 0.049127422332763675, 0.04865331268310547, 0.04943155288696289, 0.3668848571777344, 0.04947251129150391, 0.04883763122558594, 0.048979969024658204, 0.049119232177734375, 0.04912639999389649, 0.04909465789794922, 0.0491069450378418, 0.04897894287109375, 0.04893491363525391, 0.049023998260498046, 0.04887142562866211, 0.04900556945800781, 0.04895948791503906, 0.04910182571411133, 0.049173503875732424, 0.049393665313720705, 0.04918272018432617, 0.04909363174438477, 0.04904550552368164, 0.04900044631958008, 0.04912025451660156, 0.04897587203979492, 0.04906598281860351, 0.04904447937011719, 0.04894828796386719, 0.04892870330810547, 0.04903833770751953, 0.04894515228271484, 0.04881817626953125, 0.049154048919677736, 0.04902406311035156, 0.049087425231933594, 0.04906905746459961, 0.04899020767211914, 0.04910182571411133, 0.049522720336914065, 0.04881404876708984, 0.0488458251953125, 0.049081344604492184, 0.04900454330444336, 0.04894617462158203, 0.04903219223022461, 0.04901785659790039, 0.04900044631958008, 0.049084415435791014, 0.04905984115600586, 0.04926464080810547, 0.05324390411376953, 0.05209190368652344, 0.05203046417236328, 0.05140480041503906, 0.04917657470703125, 0.051130367279052735, 0.052101119995117184, 0.0525219841003418, 0.052291584014892575, 0.051416065216064455, 0.051860481262207034, 0.05178777694702148, 0.05181644821166992, 0.05204991912841797, 0.05178879928588867, 0.36726272583007813, 0.04961177444458008, 0.0512911376953125, 0.05200588989257812, 0.05116723251342774, 0.05160243225097656, 0.05197619247436523, 0.05205196762084961, 0.051705856323242184, 0.05210009765625, 0.05150310516357422, 0.05182463836669922, 0.05140377426147461, 0.05038284683227539, 0.051227649688720706, 0.05105561447143555, 0.05234175872802734, 0.05228339385986328, 0.05160038375854492, 0.05044736099243164, 0.049081344604492184, 0.05074431991577148, 0.04886220932006836, 0.050547710418701174, 0.05145702362060547, 0.051708927154541014, 0.05186969757080078, 0.05135769653320312, 0.051590145111083986, 0.05146623992919922, 0.04899327850341797, 0.05083443069458008, 0.04898611068725586, 0.05069926452636719, 0.04894412612915039, 0.050508800506591796, 0.05116928100585937, 0.05155123138427734, 0.049484798431396484, 0.05170380783081055, 0.05133414459228516, 0.05164543914794922, 0.05431500625610351, 0.051931137084960936, 0.051637248992919924, 0.05175910568237305, 0.04948582458496094, 0.05162905502319336, 0.0513259506225586, 0.04898406219482422, 0.052127742767333986, 0.051863552093505856, 0.05156351852416992, 0.05145702362060547, 0.049495040893554686, 0.05168742370605469, 0.05192192077636719, 0.0504002571105957, 0.049549312591552735, 0.05090508651733398, 0.04910899353027344, 0.051763198852539063, 0.05104127883911133, 0.3655106506347656, 0.050769920349121096, 0.049014785766601565, 0.05167001724243164, 0.04905062484741211, 0.0504535026550293, 0.05113139343261719, 0.050029567718505856, 0.051471359252929685, 0.05019443130493164, 0.0514600944519043, 0.05162905502319336, 0.0515195198059082, 0.05131465530395508, 0.04901683044433594, 0.05071155166625976, 0.05128908920288086, 0.05137408065795898, 0.04911308670043945, 0.05129830551147461, 0.048328704833984375, 0.050282497406005856, 0.05157273483276367, 0.05154304122924805, 0.04890521621704102, 0.0509224967956543, 0.05154816055297851, 0.05140377426147461, 0.05144371032714844, 0.051530750274658206, 0.051509246826171876, 0.05153696060180664, 0.051349441528320314, 0.051381248474121094, 0.049081344604492184, 0.050216960906982425, 0.051901439666748046, 0.05149798583984375, 0.05147443389892578, 0.051469345092773434, 0.051388385772705075, 0.051388416290283206, 0.051533824920654295, 0.051402751922607424, 0.05139558410644531, 0.048992286682128905, 0.04943049621582031, 0.05150207901000976, 0.05152665710449219, 0.051515392303466793, 0.051332096099853515, 0.051550209045410154, 0.05013708877563477, 0.049271808624267575, 0.04944384002685547, 0.05204070281982422, 0.05294694519042969, 0.05240627288818359, 0.051908672332763674, 0.05154707336425781, 0.051676158905029294, 0.05153177642822265, 0.051560447692871096, 0.36518310546875, 0.049238910675048826, 0.049148929595947265, 0.04931071853637695, 0.04899532699584961, 0.04968550491333008, 0.04921139144897461, 0.05128908920288086, 0.05162905502319336, 0.05139353561401367, 0.05159833526611328, 0.051814399719238284, 0.05161062240600586, 0.05162393569946289, 0.051522560119628906, 0.051560447692871096, 0.050179073333740234, 0.05207244873046875, 0.0513259506225586, 0.05152460861206055, 0.05170687866210937, 0.051765247344970705, 0.04937113571166992, 0.05161574554443359, 0.051661823272705076, 0.05231206512451172, 0.05114265441894531, 0.051864574432373044, 0.051351551055908204, 0.05171507263183594, 0.05182361602783203, 0.051716094970703126, 0.04993843078613281, 0.05160243225097656, 0.05125632095336914, 0.04917657470703125, 0.050958335876464846, 0.04966092681884766, 0.04936703872680664, 0.05073100662231445, 0.052122623443603515, 0.05184102249145508, 0.05166080093383789, 0.05268384170532227, 0.052374462127685546, 0.05087846374511719, 0.051806209564208984, 0.051821632385253905, 0.05177542495727539, 0.051627006530761715, 0.05180313491821289, 0.05178675079345703, 0.04925439834594727, 0.0492410888671875, 0.04880998229980469, 0.04935782241821289, 0.05232025527954102, 0.05278515243530273, 0.05175398254394531, 0.05167513656616211, 0.051743743896484375, 0.051507198333740234, 0.05155430221557617, 0.3660758972167969, 0.050388992309570314, 0.05113446426391602, 0.05162598419189453, 0.04928921508789062, 0.049170433044433595, 0.04955955123901367, 0.049334270477294925, 0.05116416168212891, 0.049304576873779295, 0.04921855926513672, 0.04851814270019531, 0.04916121673583984, 0.049263614654541016, 0.04922880172729492, 0.0526376953125, 0.05345382308959961, 0.049721343994140625, 0.04922784042358398, 0.051360702514648436, 0.04924825668334961, 0.05102796936035156, 0.049255424499511716, 0.04929536056518555, 0.049279998779296875, 0.04924415969848633, 0.04936601638793945, 0.04922060775756836, 0.050405376434326174, 0.051566593170166014, 0.05176422500610352, 0.05237760162353516, 0.05340364837646484, 0.0518922233581543, 0.05222809600830078, 0.04931379318237305, 0.05008076858520508, 0.05142121505737305, 0.05104841613769531, 0.04860825729370117, 0.04925132751464844, 0.05118566513061523, 0.051689472198486325, 0.051813377380371096, 0.05157785415649414, 0.04933631896972656, 0.049339393615722656, 0.04929843139648438, 0.050331649780273435, 0.04918783950805664, 0.049259521484375, 0.04927897644042969, 0.04922060775756836, 0.04920729446411133, 0.049258495330810545, 0.04967731094360352, 0.050028545379638675, 0.051507198333740234, 0.05155433654785156, 0.05167305755615234, 0.050710529327392576, 0.05175296020507812, 0.049306625366210936, 0.36612506103515624, 0.04942745590209961, 0.04921241760253906, 0.049140735626220705, 0.049081344604492184, 0.05012377548217774, 0.051736640930175784, 0.05177644729614258, 0.05170278549194336, 0.05188710403442383, 0.04920832061767578, 0.04938547134399414, 0.05068598556518555, 0.05159731292724609, 0.05157270431518555, 0.051724288940429686, 0.049642494201660156, 0.05167622375488281, 0.05174777603149414, 0.05130444717407227, 0.05182361602783203, 0.049274879455566405, 0.048985088348388675, 0.049170433044433595, 0.04934860610961914, 0.05040332794189453, 0.05165363311767578, 0.04930252838134765, 0.04917145538330078, 0.049906688690185545, 0.04945612716674805, 0.04867583847045898, 0.04913772964477539, 0.04887545776367187, 0.04923187255859375, 0.04917452621459961, 0.04922265625, 0.0492410888671875, 0.05051702499389649, 0.049186782836914064, 0.049306625366210936, 0.04935270309448242, 0.050108417510986325, 0.05163417434692383, 0.05176422500610352, 0.05215334320068359, 0.05348044967651367, 0.052375553131103515, 0.05211852645874023, 0.05194956970214844, 0.05207244873046875, 0.0519731216430664, 0.049240062713623044, 0.0500398063659668, 0.05119385528564453, 0.05156147384643555, 0.05197516632080078, 0.05172019195556641, 0.05165055847167969, 0.048927745819091796, 0.049185791015625, 0.049107967376708986, 0.05043711853027344, 0.3661455383300781, 0.049414142608642575, 0.051817470550537106, 0.05160140609741211, 0.051934207916259766, 0.05161062240600586, 0.04934860610961914, 0.05106687927246094, 0.051676158905029294, 0.0514252815246582, 0.05184511947631836, 0.049075199127197267, 0.048979969024658204, 0.05180313491821289, 0.05154304122924805, 0.0510382080078125, 0.05098188781738281, 0.051829761505126956, 0.04919398498535156, 0.05170073699951172, 0.05008076858520508, 0.051738624572753904, 0.05166899108886719, 0.05161779022216797, 0.05109145736694336, 0.0507770881652832, 0.05037875366210937, 0.05235302352905274, 0.05175807952880859, 0.05164851379394531, 0.05192396926879883, 0.05156966400146484, 0.05164646530151367, 0.05006438446044922, 0.05252505493164063, 0.05164646530151367, 0.05206835174560547, 0.049672191619873046, 0.04910899353027344, 0.049084415435791014, 0.04921241760253906, 0.05140377426147461, 0.04913459014892578, 0.04897280120849609, 0.04897382354736328, 0.04911718368530273, 0.050857982635498046, 0.05160140609741211, 0.05158195114135742, 0.051367935180664064, 0.05120102310180664, 0.04905984115600586, 0.048729087829589846, 0.0487720947265625, 0.05086105728149414, 0.05155737686157227, 0.051576831817626956, 0.04906291198730469, 0.04913459014892578, 0.049097728729248044, 0.049337345123291014, 0.05103308868408203, 0.04916326522827148, 0.3670015869140625, 0.05172326278686523, 0.04929536056518555, 0.04998246383666992, 0.051542015075683595, 0.052160511016845705, 0.050126911163330075, 0.0516453742980957, 0.05007360076904297, 0.048998401641845706, 0.05123379135131836, 0.051863552093505856, 0.04964352035522461, 0.05156966400146484, 0.04935065460205078, 0.049293312072753906, 0.04991385650634766, 0.05186764907836914, 0.04906086349487305, 0.04916326522827148, 0.050078720092773435, 0.05166592025756836, 0.05129830551147461, 0.05249740982055664, 0.05102182388305664, 0.052063297271728516, 0.0517825927734375, 0.05185740661621094, 0.049293312072753906, 0.049219585418701174, 0.05063884735107422, 0.04929228973388672, 0.049307647705078124, 0.04968038558959961, 0.051896320343017575, 0.05139353561401367, 0.051958782196044925, 0.049037311553955076, 0.0512624626159668, 0.049519615173339845, 0.051481601715087894, 0.051697696685791016, 0.051786720275878904, 0.04920729446411133, 0.05130035018920898, 0.05168844985961914, 0.05171916961669922, 0.05164748764038086, 0.049347583770751956, 0.04920012664794922, 0.05144063949584961, 0.05171916961669922, 0.05189836883544922, 0.05204582214355469, 0.051716094970703126, 0.051694591522216796, 0.052135936737060545, 0.05168332672119141, 0.051806209564208984, 0.051714046478271485, 0.05174169540405273, 0.04931891250610351, 0.05037363052368164]",tokens/s,18.216296363942554,,, -4bit-bnb-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9663,7 +9663,7 @@ OSError: 8 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9791,7 +9791,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9891,7 +9891,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 192.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10318,7 +10318,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1275.252736,9043.443712,0.0,8396.996608,8007.886848,s,10,5.839258605957031,0.5839258605957032,0.0002902213580805191,0.5840317687988281,0.5842310180664062,0.5842669982910156,0.5842957824707031,"[0.5842230224609375, 0.583387939453125, 0.5835930786132812, 0.5836356201171875, 0.5837993774414062, 0.584034912109375, 0.5840615844726562, 0.584302978515625, 0.5841914672851563, 0.5840286254882813]",tokens/s,438.411821217914,kWh,6.90310146705604e-06,3.7826373528029457e-06,4.4987813767999934e-05,5.5673552587858914e-05,tokens/kWh,4598233.597469897,MB,1275.252736,9043.443712,0.0,8396.996608,8007.889408,s,10,50.588525878906246,5.058852587890625,0.08395309153330552,5.082033935546875,5.14226640625,5.142390771484375,5.142490263671875,"[5.09013134765625, 5.0336044921875, 5.0739365234375, 4.90679248046875, 4.9339384765625, 4.99420849609375, 5.14251513671875, 5.14223876953125, 5.13380517578125, 5.13735498046875]",tokens/s,12.453416838198269,kWh,5.875015757518049e-05,3.219780382033908e-05,0.0001767818636476009,0.0002677298250431205,tokens/kWh,235311.84838989543,,s,629,55.761538002014184,0.0886510938028842,0.06938948591757807,0.08127062225341797,0.08205045776367187,0.08251166534423827,0.6628051928710939,"[0.08312934112548828, 0.08194662475585937, 0.08161094665527344, 0.08106793975830077, 0.07799388885498047, 0.07790182495117187, 0.07877426910400391, 0.08109056091308593, 0.08273715209960937, 0.08016281890869141, 0.08193740844726563, 0.08150323486328125, 0.08199372863769532, 0.08215654754638672, 0.08155033874511719, 0.08119091033935547, 0.08179199981689453, 0.082050048828125, 0.08167219543457031, 0.08016793823242188, 0.08142848205566407, 0.08155049896240234, 0.08174269104003906, 0.08181247711181641, 0.08187088012695312, 0.08173872375488281, 0.08148172760009766, 0.0813589096069336, 0.08177247619628907, 0.08165068817138672, 0.08111923217773437, 0.08182272338867187, 0.08138854217529297, 0.08169779205322265, 0.0813803482055664, 0.07805133056640624, 0.0778260498046875, 0.07720857238769531, 0.07690342712402344, 0.07678157043457032, 0.07862886047363281, 0.07793670654296875, 0.0814273910522461, 0.08033689880371093, 0.08093901062011719, 0.08202957153320313, 0.08053657531738281, 0.0800348129272461, 0.08112844848632812, 0.08078438568115234, 0.08056320190429687, 0.08157695770263672, 0.0806645736694336, 0.08098508453369141, 0.08129843139648438, 0.08099737548828125, 0.08169779205322265, 0.08143769836425781, 0.07781171417236328, 0.08082841491699219, 0.081438720703125, 0.0816732177734375, 0.6624255981445313, 0.0779161605834961, 0.07806361389160156, 0.07813433837890625, 0.07830214691162109, 0.0823377914428711, 0.081797119140625, 0.08149298858642579, 0.08155651092529297, 0.08154927825927734, 0.08165785980224609, 0.08036557006835937, 0.08090419006347656, 0.07792230224609376, 0.07803699493408203, 0.07802674865722656, 0.0782429428100586, 0.07808396911621093, 0.07814553833007812, 0.07801856231689454, 0.08006758117675782, 0.08151551818847656, 0.08212806701660157, 0.08106681823730469, 0.08180429077148438, 0.07889520263671874, 0.0796589126586914, 0.08029491424560548, 0.07896268463134766, 0.08166092681884765, 0.08164761352539063, 0.08132198333740234, 0.07873126220703125, 0.07970713806152344, 0.08085606384277344, 0.08102194976806641, 0.08094617462158203, 0.08055091094970702, 0.07810355377197266, 0.07798681640625, 0.07718297576904297, 0.07773900604248046, 0.07992934417724609, 0.08185139465332031, 0.08153292846679687, 0.0834672622680664, 0.08181657409667968, 0.08205209350585937, 0.08122675323486328, 0.08143974304199218, 0.08141516876220703, 0.08156979370117187, 0.08153292846679687, 0.08144588470458984, 0.07833497619628907, 0.07812608337402344, 0.07809228515625, 0.0776039047241211, 0.07791200256347657, 0.07793670654296875, 0.07808812713623046, 0.0779940185546875, 0.0787721939086914, 0.6652927856445312, 0.081797119140625, 0.0824463348388672, 0.08166400146484375, 0.0812943344116211, 0.08171417236328125, 0.081797119140625, 0.0810403823852539, 0.08148684692382813, 0.08060108947753906, 0.08116223907470703, 0.08115200042724609, 0.0798914566040039, 0.0795688934326172, 0.08124313354492188, 0.08029286193847657, 0.0800747528076172, 0.08036761474609375, 0.08075775909423828, 0.0814571533203125, 0.08046080017089843, 0.07997849273681641, 0.08161705780029296, 0.08034492492675781, 0.08042819213867188, 0.08051900482177735, 0.08177254486083985, 0.08252928161621094, 0.08148172760009766, 0.08164659118652344, 0.08396185302734376, 0.0818309097290039, 0.08160460662841797, 0.08195891571044922, 0.08126873779296875, 0.08167024230957032, 0.08137206268310547, 0.08147968292236328, 0.08164351654052734, 0.08213196563720702, 0.08143666839599609, 0.08147865295410156, 0.08106508636474609, 0.08147750091552734, 0.08155238342285157, 0.08134758758544922, 0.08133837127685548, 0.08147161865234374, 0.08138739013671875, 0.08059391784667969, 0.0777871322631836, 0.07810969543457032, 0.07796633911132812, 0.07770623779296874, 0.07752191925048828, 0.0778045425415039, 0.0777237777709961, 0.07775443267822266, 0.07790265655517578, 0.07787725067138672, 0.07770111846923829, 0.07686246490478515, 0.0769587173461914, 0.6624778442382813, 0.07796444702148438, 0.07821807861328126, 0.07775949096679688, 0.07773388671875, 0.07772476959228515, 0.07772560119628906, 0.07763455963134766, 0.0778260498046875, 0.07856435394287109, 0.07809945678710938, 0.07831346893310547, 0.07798992156982422, 0.07810249328613281, 0.07798579406738282, 0.07784243011474609, 0.07784243011474609, 0.07813426971435547, 0.07801344299316407, 0.07783628845214843, 0.07761116790771484, 0.07766512298583984, 0.0778792953491211, 0.07784754943847656, 0.07777996826171875, 0.07789875030517578, 0.07791513824462891, 0.07812525177001953, 0.07811872100830078, 0.07833497619628907, 0.0784682846069336, 0.07776441955566406, 0.07800729370117188, 0.0782008285522461, 0.07801548767089844, 0.07788761901855469, 0.07785254669189454, 0.07780147552490234, 0.07757933044433593, 0.07708358764648438, 0.07733452606201172, 0.07772467041015625, 0.0777175064086914, 0.07655321502685547, 0.07728444671630859, 0.0779682846069336, 0.07754649353027344, 0.07680921936035157, 0.07681641387939453, 0.08104351806640625, 0.07825296020507813, 0.07756390380859375, 0.07738982391357421, 0.07754137420654297, 0.07784960174560547, 0.07721778869628906, 0.07747071838378906, 0.07846300506591797, 0.07957193756103516, 0.0775393295288086, 0.07772569274902344, 0.07767654418945312, 0.07778816223144532, 0.6633749389648438, 0.07927085113525391, 0.07809945678710938, 0.07801958465576173, 0.07804313659667969, 0.07777484893798828, 0.07775949096679688, 0.0777349090576172, 0.07792332458496094, 0.07786495971679687, 0.07802572631835937, 0.07779955291748047, 0.07796211242675781, 0.07779328155517579, 0.0778967056274414, 0.0776632308959961, 0.07759974670410157, 0.07993762969970702, 0.07793859100341796, 0.07785062408447266, 0.07779634857177735, 0.07753727722167969, 0.07777689361572265, 0.0774307861328125, 0.07764803314208985, 0.07781667327880859, 0.0776263656616211, 0.07756697845458985, 0.07763251495361329, 0.07775641632080078, 0.07764275360107421, 0.07765007781982422, 0.0776343994140625, 0.07750566101074219, 0.07771532440185547, 0.07787007904052734, 0.07760588836669922, 0.07744818878173829, 0.07757212829589843, 0.07950128173828125, 0.0798361587524414, 0.08110797119140625, 0.08121753692626953, 0.0811171875, 0.08120524597167969, 0.08123801422119141, 0.07770022583007813, 0.07786994934082031, 0.07789977264404296, 0.07779020690917969, 0.07773811340332032, 0.07815257263183593, 0.07773190307617188, 0.07772665405273438, 0.07764173126220703, 0.07759053039550781, 0.0777871322631836, 0.0775546875, 0.0775516128540039, 0.07761116790771484, 0.07986672210693359, 0.08118374633789062, 0.0815606689453125, 0.6631760864257813, 0.07791923522949219, 0.07783116912841796, 0.07799705505371093, 0.07772774505615235, 0.07799091339111328, 0.07786393737792968, 0.07784754943847656, 0.0778967056274414, 0.07811180877685547, 0.07756998443603516, 0.07754444885253907, 0.08015360260009766, 0.08101785278320313, 0.08032051086425782, 0.07865446472167968, 0.08153907012939453, 0.078727294921875, 0.07774912261962891, 0.0782182388305664, 0.07816499328613281, 0.07796940612792969, 0.07738470458984376, 0.08143769836425781, 0.0808058853149414, 0.07757004547119141, 0.07835356903076172, 0.07809827423095703, 0.08068608093261719, 0.08145305633544922, 0.08136617279052734, 0.08127062225341797, 0.08130252838134766, 0.08150527954101562, 0.08302489471435547, 0.08381439971923828, 0.0822824935913086, 0.08181145477294922, 0.08138137817382812, 0.08140493011474609, 0.07823257446289063, 0.07797555541992188, 0.07817132568359375, 0.07769683074951172, 0.07833497619628907, 0.0778946533203125, 0.07803699493408203, 0.07789977264404296, 0.07788646697998047, 0.07793459320068359, 0.07824793243408203, 0.07824281311035156, 0.0779276123046875, 0.0780040283203125, 0.0779112319946289, 0.07785964965820312, 0.07826227569580078, 0.08185651397705078, 0.08159334564208984, 0.0780400619506836, 0.08042201232910157, 0.08159526062011718, 0.08133427429199219, 0.6656102294921875, 0.08164864349365235, 0.0825928955078125, 0.08414399719238282, 0.0821022720336914, 0.08235110473632813, 0.08180735778808594, 0.08164659118652344, 0.08218931579589844, 0.08201522827148437, 0.08214733123779297, 0.08248115539550781, 0.08194764709472656, 0.08146550750732422, 0.08154096221923827, 0.08179609680175781, 0.08177458953857422, 0.08144710540771484, 0.08145798492431641, 0.08194252777099609, 0.08160050964355468, 0.08173776245117187, 0.08153596496582032, 0.08315904235839844, 0.08191795349121093, 0.08120028686523438, 0.0805478744506836, 0.08130847930908203, 0.08132608032226563, 0.08154521942138672, 0.08176640319824219, 0.08173670196533203, 0.08173977661132813, 0.08151961517333985, 0.08127180480957032, 0.08171113586425781, 0.08177046203613281, 0.08155136108398438, 0.08085298919677734, 0.0810250244140625, 0.08155136108398438, 0.08091238403320312, 0.08146125030517579, 0.08133939361572265, 0.08176844787597656, 0.0814163818359375, 0.081653564453125, 0.07750962829589844, 0.08073420715332032, 0.08069222259521484, 0.0827852783203125, 0.08272077178955078, 0.08201522827148437, 0.08141619110107422, 0.08072089385986328, 0.08127897644042968, 0.08163430023193359, 0.08206028747558594, 0.08171417236328125, 0.08186675262451172, 0.08251699066162109, 0.08043315124511718, 0.08146841430664062, 0.6693641967773437, 0.08206448364257812, 0.08240019226074219, 0.08181552124023438, 0.08168246459960937, 0.08171209716796875, 0.08166194915771484, 0.08153395080566406, 0.08148172760009766, 0.08155852508544922, 0.08130662536621094, 0.08189030456542969, 0.08139059448242188, 0.08146125030517579, 0.08164864349365235, 0.08184524536132813, 0.08188518524169922, 0.08159347534179688, 0.08324492645263672, 0.08229170989990234, 0.08163737487792969, 0.08156877136230468, 0.07990271759033203, 0.08209613037109376, 0.0821391372680664, 0.08164157104492188, 0.0812943344116211, 0.08186460876464843, 0.08141926574707031, 0.08160972595214844, 0.08137734222412109, 0.08141407775878906, 0.08073932647705079, 0.08131584167480468, 0.08116035461425782, 0.08141398620605468, 0.08145613098144532, 0.08088166046142578, 0.08128614044189453, 0.08138444519042969, 0.08004198455810548, 0.0816015396118164, 0.08122777557373047, 0.08015257263183594, 0.08131292724609375, 0.08146927642822266, 0.08136402893066406, 0.08098502349853516, 0.08175923156738281, 0.08144179534912109, 0.08176032257080078, 0.08200287628173829, 0.08134451293945312, 0.08161705780029296, 0.08128189086914063, 0.08215449523925782, 0.08272402954101563, 0.08222496032714843, 0.08155750274658204, 0.08195481872558594, 0.0815288314819336, 0.08163430023193359, 0.08198860931396484, 0.6629324951171875, 0.07831449890136719, 0.08065638732910156, 0.08173772430419922, 0.08174591827392579, 0.07905894470214844, 0.08188313293457031, 0.08141414642333984, 0.08146841430664062, 0.0813311996459961, 0.08091033935546875, 0.08099737548828125, 0.08050994873046875, 0.08022630310058594, 0.08185958099365234, 0.08196198272705078, 0.08165171051025391, 0.08163021087646484, 0.08073420715332032, 0.0812976303100586, 0.08113027191162109, 0.08124313354492188, 0.08088690948486328, 0.08081804656982422, 0.08159232330322265, 0.08143154907226563, 0.08153190612792968, 0.08174387359619141, 0.0820101089477539, 0.08198143768310547, 0.08203059387207032, 0.08126873779296875, 0.08250367736816407, 0.08279347229003907, 0.08148274993896484, 0.08171417236328125, 0.08203059387207032, 0.08150118255615234, 0.08181247711181641, 0.08187904357910156, 0.07972761535644532, 0.08162303924560547, 0.08191590118408203, 0.08157081604003906, 0.08166912078857422, 0.08133222198486328, 0.08178482818603515, 0.08138137817382812, 0.08160870361328125, 0.08150527954101562, 0.08147154998779296, 0.08147245025634765, 0.08286310577392578, 0.08272697448730469, 0.08185030364990234, 0.08227021026611328, 0.08171212768554688, 0.0816732177734375, 0.08092060852050781, 0.0821472930908203, 0.08211270141601562, 0.08459347534179687, 0.0823920669555664, 0.6647736206054687, 0.08188006591796874, 0.08189043426513672, 0.08091123199462891, 0.08149298858642579, 0.08203263854980469, 0.0821760025024414, 0.08150540924072265, 0.08150310516357422, 0.08447590637207031, 0.08190982055664063, 0.08173868560791016, 0.08170297241210937, 0.08169261169433593, 0.08128717041015625, 0.08216063690185547, 0.08190668487548829, 0.0815841293334961, 0.08160870361328125, 0.08176850891113281, 0.08176428985595703, 0.08213196563720702, 0.08238285064697265, 0.08154112243652344, 0.08155648040771485, 0.08143257904052735, 0.08200601959228515, 0.08178892517089843, 0.08140499114990235, 0.08160454559326172, 0.0820101089477539, 0.08201420593261718, 0.07984333038330078, 0.08177356719970703, 0.08219545745849609, 0.08154112243652344, 0.0815472640991211, 0.081447998046875, 0.0830084457397461, 0.08281600189208985, 0.08126992034912109, 0.08128905487060546, 0.08192729949951172, 0.08113651275634766, 0.08141107177734375, 0.08161603546142578, 0.08171708679199219, 0.0810096664428711, 0.0804136962890625, 0.08092876434326172, 0.081438720703125, 0.08066867065429688, 0.0804290542602539, 0.07728230285644531, 0.07898214721679687, 0.08123699188232422, 0.08158822631835938, 0.08082125091552735, 0.08359321594238281, 0.08226713562011718, 0.08175206756591796, 0.08144793701171875, 0.08139673614501954]",tokens/s,11.280176669038072,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10347,7 +10347,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10469,7 +10469,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 28.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10528,7 +10528,7 @@ ValueError: CodeGenForCausalLM does not support an attention implementation thro ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1517.993984,9754.37824,0.0,9107.931136,8963.121152,s,10,6.474350524902343,0.6474350524902344,0.0071454832222554365,0.6491697692871093,0.65416982421875,0.6561875732421875,0.6578017724609375,"[0.6534266357421875, 0.64244091796875, 0.6496339721679687, 0.6359340209960938, 0.635143798828125, 0.6492542724609375, 0.6475048828125, 0.653721435546875, 0.6490852661132812, 0.658205322265625]",tokens/s,395.406456625024,kWh,7.807321080731022e-06,4.278093292271023e-06,1.6910239222626272e-05,2.8995653595628316e-05,tokens/kWh,8828909.448642233,MB,1517.993984,9754.37824,0.0,9107.931136,8963.123712,s,10,379.70457421875,37.970457421875,0.22482269222568638,37.894755859374996,38.21149375,38.36042265625,38.47956578125,"[37.74726171875, 37.91973828125, 37.79262890625, 38.0966875, 38.5093515625, 38.1783984375, 37.83587109375, 37.76275390625, 37.8697734375, 37.992109375]",tokens/s,1.65918464715954,kWh,0.00045022652441014847,0.00024676170439992117,0.0006422136491871809,0.0013392018779972506,tokens/kWh,47042.944783063795,,s,629,384.92266827392604,0.6119597269855735,0.07732522279345393,0.6043299560546875,0.6137816528320312,0.6157035034179688,1.2408375,"[0.591009765625, 0.5934766235351563, 0.5923543090820312, 0.5940879516601563, 0.60214990234375, 0.6043412475585938, 0.5927178344726562, 0.595937255859375, 0.5915699462890625, 0.6028973999023437, 0.592415771484375, 0.6051809692382812, 0.6002021484375, 0.60002099609375, 0.5911818237304688, 0.58952294921875, 0.5990758666992188, 0.5887201538085938, 0.5874544677734375, 0.5933403930664063, 0.608974853515625, 0.6007920532226563, 0.5878865966796875, 0.5978470458984375, 0.5910599975585937, 0.5960928955078125, 0.6032097778320312, 0.5888726196289062, 0.588521484375, 0.5878568725585938, 0.6083543090820313, 0.60647216796875, 0.6090127563476563, 0.6103470458984375, 0.6039091796875, 0.6121338500976562, 0.61477685546875, 0.6143733520507813, 0.6161305541992188, 0.5974599609375, 0.5948047485351563, 0.5932297973632813, 0.6146969604492187, 0.6005514526367187, 0.59856689453125, 0.5971763305664063, 0.5945743408203125, 0.5923471069335937, 0.5927260131835937, 0.5976678466796875, 0.5937889404296876, 0.608194580078125, 0.6015303955078125, 0.5988976440429687, 0.598729736328125, 0.59260009765625, 0.6162236938476563, 0.5990502319335937, 0.6065807495117187, 0.6045460205078125, 0.5923676147460938, 0.6035435791015625, 1.245834228515625, 0.5931939697265625, 0.5960591430664063, 0.6082078857421875, 0.6047938842773437, 0.6055885009765625, 0.6112420043945312, 0.6143733520507813, 0.6075443115234375, 0.609776611328125, 0.6117867431640625, 0.6004602661132813, 0.6088734741210937, 0.6127646484375, 0.5976627197265625, 0.5947258911132812, 0.5995858764648437, 0.6004089965820313, 0.6068223876953125, 0.6114129638671875, 0.600110107421875, 0.5907476196289062, 0.5962915649414062, 0.6039429321289063, 0.5897850952148438, 0.59139892578125, 0.5901127319335937, 0.5984378662109375, 0.6077962036132812, 0.6160588989257813, 0.60433203125, 0.5912842407226563, 0.5951201171875, 0.604832763671875, 0.6040432739257813, 0.605107177734375, 0.61203564453125, 0.6116607666015625, 0.591341552734375, 0.5916201171875, 0.5907077026367188, 0.5925325317382812, 0.6104186401367188, 0.606571533203125, 0.60406884765625, 0.598392822265625, 0.5964861450195312, 0.5926594848632812, 0.5983938598632812, 0.5913651123046875, 0.6097418212890625, 0.6015160522460937, 0.6009415893554687, 0.6038651123046875, 0.5905960693359374, 0.6054000854492188, 0.58945947265625, 0.5900738525390625, 0.6094213256835938, 0.6088775634765625, 0.6106152954101562, 0.6121236572265625, 0.6107811889648438, 1.240201171875, 0.60145458984375, 0.6063114013671875, 0.6000506591796875, 0.5938329467773438, 0.6147860717773438, 0.61549462890625, 0.6092933349609375, 0.6010162963867187, 0.6016266479492187, 0.60356201171875, 0.5935820922851562, 0.5922109375, 0.5997998046875, 0.5918760986328125, 0.5992703857421875, 0.5925755004882812, 0.5937223510742188, 0.6090741577148437, 0.6089686889648438, 0.5921444091796875, 0.5962076416015625, 0.6106767578125, 0.5926574096679688, 0.592884765625, 0.5916354370117187, 0.5904813842773438, 0.6025748291015625, 0.5909227294921875, 0.6148065185546875, 0.6080604248046875, 0.6043299560546875, 0.5950474243164062, 0.606476318359375, 0.6080932006835937, 0.6103142700195312, 0.604548095703125, 0.6048604125976562, 0.59738623046875, 0.594260986328125, 0.604579833984375, 0.5922979736328124, 0.5899970703125, 0.5934612426757813, 0.6071869506835937, 0.5968977661132813, 0.5957161254882812, 0.605191162109375, 0.592584716796875, 0.5992191772460937, 0.5921669311523438, 0.609533935546875, 0.6162237548828124, 0.5931397094726563, 0.59833447265625, 0.6118461303710937, 0.5926246948242188, 0.5976319580078125, 0.5976012573242188, 0.5918535766601563, 0.5936537475585938, 0.5920112915039063, 0.6014955444335938, 1.2383375244140624, 0.595831787109375, 0.6064168701171875, 0.6077655639648437, 0.59162109375, 0.6110136108398437, 0.6126735229492187, 0.6068971557617188, 0.6108283081054687, 0.6127401123046875, 0.5900902099609375, 0.6045153198242188, 0.5931345825195312, 0.6106470336914063, 0.6118164672851563, 0.6136740112304687, 0.6127462158203125, 0.6145474853515625, 0.6146611328125, 0.6126254272460937, 0.613064697265625, 0.6193981323242187, 0.5963991088867188, 0.5956423950195312, 0.597086181640625, 0.6127001342773437, 0.6043995971679688, 0.609976318359375, 0.6110525512695313, 0.5916866455078125, 0.591331298828125, 0.5906923217773438, 0.589781005859375, 0.5940776977539063, 0.61161474609375, 0.6000404663085938, 0.5916201171875, 0.6117590942382812, 0.611156982421875, 0.6106470336914063, 0.61089794921875, 0.6133575439453125, 0.6101012573242187, 0.5905131225585938, 0.6015027465820313, 0.6101964721679688, 0.61241650390625, 0.6102538452148437, 0.6083399658203125, 0.6069514770507812, 0.6010521240234376, 0.6009241333007812, 0.6015488891601563, 0.6098349609375, 0.611535888671875, 0.595936279296875, 0.58838427734375, 0.5979248657226562, 0.6079672241210937, 0.5946992797851562, 0.6041825561523437, 0.609533935546875, 0.6078883666992188, 1.2410849609375, 0.6061486206054687, 0.6111242065429687, 0.6117939453125, 0.6083594360351563, 0.6044241943359375, 0.6064056396484375, 0.6129326171875, 0.6117294311523438, 0.6129581909179688, 0.6110084838867188, 0.6145842895507813, 0.6136708984375, 0.6136156005859374, 0.6130155639648438, 0.6129642944335938, 0.6101432495117187, 0.611103759765625, 0.6098084716796875, 0.6117825927734375, 0.6123786010742187, 0.6084822998046875, 0.6103316650390626, 0.6132684936523437, 0.6116638793945313, 0.6128844604492187, 0.6167869262695312, 0.615857177734375, 0.615125, 0.6159298706054688, 0.6137988891601562, 0.615130126953125, 0.6149171142578125, 0.6128250732421875, 0.6101616821289062, 0.61254248046875, 0.6113843383789063, 0.6132254638671875, 0.6125025024414062, 0.6150317993164063, 0.6137426147460937, 0.6131988525390625, 0.6140170288085938, 0.5924607543945313, 0.60002099609375, 0.6001008911132812, 0.606307373046875, 0.6160578002929687, 0.6144020385742187, 0.6103572387695313, 0.6094633178710938, 0.6150062255859375, 0.6144573974609375, 0.6177146606445313, 0.614234130859375, 0.6144010009765625, 0.6169343872070312, 0.616237060546875, 0.6122782592773437, 0.613264404296875, 0.6082570190429688, 0.5977733764648437, 0.6009630126953125, 1.2645140380859374, 0.6144286499023438, 0.6149417114257812, 0.6130811157226562, 0.6085878295898437, 0.6012487182617188, 0.6021253051757812, 0.5963171997070312, 0.60879052734375, 0.6130841674804688, 0.606255126953125, 0.6084935913085937, 0.61300634765625, 0.6130288696289062, 0.6120908813476562, 0.6120724487304687, 0.6163815307617188, 0.6157157592773438, 0.6126489868164062, 0.6147338256835938, 0.6092134399414062, 0.5932584838867188, 0.601628662109375, 0.5970933837890625, 0.6128609008789062, 0.61948828125, 0.6102548217773438, 0.6038446044921875, 0.60258203125, 0.6102568969726563, 0.6136196899414063, 0.6131087646484376, 0.6116905517578125, 0.613224365234375, 0.6131722412109375, 0.6159820556640625, 0.6166394653320313, 0.6119905395507812, 0.6063062744140625, 0.6061199340820312, 0.6126755981445312, 0.6141430053710938, 0.6006558837890625, 0.6105211181640625, 0.609280029296875, 0.6156851196289063, 0.6158806762695312, 0.59983056640625, 0.5894092407226562, 0.5939415283203126, 0.5915299682617188, 0.5920184326171875, 0.5910589599609375, 0.5924464721679688, 0.5912391967773437, 0.5957437744140625, 0.59049267578125, 0.5946705932617188, 0.6059038696289063, 0.591810546875, 0.5919406127929687, 0.6015354614257813, 0.5913989868164062, 1.2456324462890624, 0.5966940307617188, 0.608090087890625, 0.6086410522460938, 0.6126356201171875, 0.6094448852539063, 0.5897902221679687, 0.6053949584960937, 0.6021570434570312, 0.5902796630859375, 0.6087976684570312, 0.6116290283203125, 0.5929953002929688, 0.6065018920898437, 0.5896099853515625, 0.6046760864257813, 0.5881272583007813, 0.5896714477539062, 0.590182373046875, 0.5894523315429687, 0.5892074584960938, 0.5899437866210937, 0.589571044921875, 0.594998291015625, 0.5976381225585937, 0.6034319458007813, 0.5973401489257812, 0.6032527465820312, 0.6075299682617188, 0.6092636108398437, 0.6044149780273438, 0.5933331909179688, 0.6064097290039062, 0.6066636962890625, 0.6075924682617188, 0.6104013061523438, 0.589043701171875, 0.59704931640625, 0.609122314453125, 0.6122700805664063, 0.6011904296875, 0.6019645385742187, 0.5958850708007812, 0.6043453369140624, 0.6080983276367188, 0.6118748779296875, 0.61377734375, 0.6149539794921876, 0.592195556640625, 0.5983784790039063, 0.5897216186523437, 0.6100153198242187, 0.5959157104492188, 0.5982883911132812, 0.6080347900390625, 0.5919672241210937, 0.590328857421875, 0.5909555053710938, 0.6135726318359375, 0.60111767578125, 0.6074511108398437, 0.592912353515625, 0.591572998046875, 1.25821337890625, 0.6047611083984376, 0.5907282104492187, 0.6005862426757812, 0.5998653564453125, 0.5920716552734375, 0.6011074829101563, 0.5977374877929688, 0.6010511474609375, 0.6155213012695312, 0.615593994140625, 0.6083256225585938, 0.611852294921875, 0.6105733032226562, 0.6102313842773438, 0.607099853515625, 0.6118441162109375, 0.6134220581054688, 0.6099793701171875, 0.5916702880859375, 0.5911357421875, 0.6134814453125, 0.6091110229492187, 0.5918709716796875, 0.5907415161132813, 0.5927587890625, 0.6047836303710937, 0.6023209228515625, 0.59105078125, 0.6023792724609375, 0.5943121948242187, 0.5900390625, 0.5901015014648437, 0.59013427734375, 0.588537841796875, 0.5901578369140625, 0.5890877685546875, 0.5939619750976562, 0.589792236328125, 0.590160888671875, 0.5906472778320313, 0.5900728149414063, 0.5947412719726562, 0.5907322998046876, 0.60094873046875, 0.6065571899414063, 0.6097581787109375, 0.6101790771484376, 0.6125066528320312, 0.6081013793945312, 0.590540771484375, 0.5914685668945312, 0.5975357055664062, 0.5930045166015625, 0.5997967529296875, 0.5985044555664063, 0.5937930297851562, 0.5949368286132812, 0.5983057861328125, 0.5917982788085937, 0.6037380981445313, 0.605043701171875, 0.5956433715820313, 1.250555908203125, 0.6096578369140625, 0.5920112915039063, 0.5929564208984375, 0.5903390502929687, 0.5918883666992187, 0.5892321166992187, 0.6002769775390625, 0.5913046875, 0.6059878540039062, 0.5945855712890625, 0.6044682006835937, 0.5910394897460938, 0.5906258544921875, 0.5948170166015625, 0.6045911254882812, 0.5937489624023438, 0.5900349731445312, 0.6011268920898437, 0.6021734619140625, 0.5980886840820312, 0.6113668823242188, 0.6105261840820313, 0.6123714599609374, 0.599636962890625, 0.5881242065429687, 0.6095267333984375, 0.593201171875, 0.5987359008789063, 0.6047682495117187, 0.6061424560546875, 0.6116505737304687, 0.6097725219726563, 0.6104166259765625, 0.6037554931640625, 0.6160834350585938, 0.6079334716796875, 0.5892044677734375, 0.610471923828125, 0.5885634765625, 0.6057564086914062, 0.5880125732421875, 0.5870551147460937, 0.5882880249023438, 0.5920297241210938, 0.6119669799804688, 0.6120171508789063, 0.6052720336914063, 0.598614013671875, 0.6120150756835937, 0.6081822509765625, 0.5933629150390625, 0.60659814453125, 0.6092308349609376, 0.5913395385742187, 0.5908449096679688, 0.6021325073242187, 0.608415771484375, 0.6057246704101562, 0.6092308349609376, 0.6085570678710938, 0.6119300537109374, 0.6105446166992188, 1.2620257568359374, 0.6103807983398437, 0.6009579467773437, 0.6055291137695312, 0.594575439453125, 0.6064035034179688, 0.6033684692382812, 0.6022789306640625, 0.5941166381835937, 0.6087178344726563, 0.5888788452148438, 0.6004879150390625, 0.5893048095703125, 0.59618408203125, 0.589707275390625, 0.5911101684570312, 0.6066534423828125, 0.606671875, 0.60813720703125, 0.5883770751953125, 0.5950556030273437, 0.606650390625, 0.61174169921875, 0.591035400390625, 0.5904373779296875, 0.5889259643554687, 0.6027018432617187, 0.5948764038085937, 0.606255126953125, 0.596262939453125, 0.5974906616210938, 0.588832763671875, 0.5915269165039062, 0.6116823120117187, 0.6113034057617187, 0.61197314453125, 0.6082130126953125, 0.6120345458984375, 0.611030029296875, 0.6130022583007813, 0.6009098510742188, 0.6030602416992188, 0.60984521484375, 0.6080450439453124, 0.5966366577148438, 0.6068182983398438, 0.6000220336914063, 0.6067886352539062, 0.5939312744140625, 0.6200504150390626, 0.6089615478515625, 0.6108784790039062, 0.6129920043945313, 0.6041405639648437, 0.617702392578125, 0.6168995971679687, 0.591740966796875, 0.5966581420898438, 0.5956751098632812, 0.60830615234375, 0.6165657348632813, 0.6145525512695312, 0.6142586669921875]",tokens/s,1.6340944606369074,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10557,7 +10557,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10765,7 +10765,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1267.675136,5310.513152,0.0,4664.066048,4388.327424,s,10,3.149306671142578,0.3149306671142578,0.0003786809210590018,0.3148073425292969,0.3151804443359375,0.31560160522460934,0.31593853393554683,"[0.31602276611328123, 0.31467379760742187, 0.31508685302734374, 0.3148222045898437, 0.3147432861328125, 0.3148201904296875, 0.3147587280273437, 0.31479449462890624, 0.31484408569335937, 0.3147402648925781]",tokens/s,812.8773305748672,kWh,3.725219916345345e-06,2.0412616345595324e-06,2.375377768356235e-05,2.9520259234467228e-05,tokens/kWh,8672010.566258844,MB,1267.675136,5310.513152,0.0,4664.066048,4388.329984,s,10,40.68903320312501,4.068903320312501,0.02568155718462094,4.07563818359375,4.0924989746093745,4.100485815429687,4.106875288085938,"[4.05038134765625, 4.01346337890625, 4.10847265625, 4.085849609375, 4.0824853515625, 4.08030810546875, 4.0488828125, 4.09072412109375, 4.05749755859375, 4.07096826171875]",tokens/s,15.483287520127528,kWh,4.6836518693922296e-05,2.566947359195978e-05,0.00011625844543723774,0.00018876443772311983,tokens/kWh,333749.30553608073,,s,629,43.45549528503415,0.0690866379730273,0.03736675855632884,0.06506396484375,0.06555607147216798,0.06619919128417968,0.3779529541015625,"[0.06755551910400391, 0.06591366577148437, 0.06367334365844726, 0.0689122543334961, 0.06633049774169922, 0.06534553527832031, 0.06497280120849609, 0.06484684753417969, 0.06473318481445313, 0.06487567901611328, 0.0648917465209961, 0.06490419006347656, 0.0657940444946289, 0.06514995574951171, 0.06512025451660156, 0.06227558517456055, 0.0642191390991211, 0.06501993560791015, 0.06741193389892577, 0.0656701431274414, 0.06505369567871094, 0.06504857635498047, 0.06490322875976562, 0.06515193939208984, 0.06524313354492188, 0.06535785675048827, 0.06593532562255859, 0.06548172760009766, 0.06506905364990234, 0.06224895858764649, 0.062458976745605466, 0.06490409851074219, 0.06507110595703125, 0.06520240020751954, 0.06493981170654296, 0.06505574035644532, 0.0651847686767578, 0.06512127685546874, 0.0653096923828125, 0.06527897644042968, 0.06522061157226562, 0.06269337463378906, 0.06252444839477539, 0.06254691314697265, 0.062478336334228515, 0.06245580673217774, 0.06243430328369141, 0.062456832885742185, 0.06250508880615234, 0.06233689498901367, 0.06238105773925781, 0.06246912002563477, 0.06251110458374023, 0.06247219085693359, 0.062304256439208984, 0.06254796981811524, 0.062174240112304685, 0.06279062271118165, 0.06267087936401367, 0.06258480072021484, 0.06256230545043945, 0.062484607696533204, 0.3779767150878906, 0.06256947326660156, 0.0647188491821289, 0.06288076782226562, 0.06262783813476562, 0.06261248016357422, 0.06259199905395507, 0.06249369430541992, 0.06279475021362305, 0.06264934539794922, 0.06278963088989258, 0.06262067031860352, 0.06258892822265626, 0.06252870559692383, 0.06262252807617187, 0.06223987197875976, 0.06274969482421874, 0.06253247833251953, 0.06261862564086915, 0.06270975875854493, 0.06467378997802735, 0.0649318389892578, 0.06497382354736328, 0.06492569732666016, 0.06501990509033204, 0.0649912338256836, 0.0650956802368164, 0.06503030395507813, 0.06507094573974609, 0.06510079956054687, 0.06250188827514648, 0.06255001449584961, 0.06241593551635742, 0.062491649627685546, 0.06260115051269531, 0.06265856170654296, 0.06257782363891602, 0.06247612762451172, 0.06255923080444335, 0.06237190246582031, 0.062416831970214845, 0.062378017425537106, 0.062496734619140626, 0.06300579071044922, 0.0651121597290039, 0.06495827484130859, 0.06519817352294922, 0.06501366424560547, 0.06509977722167969, 0.0651673583984375, 0.06430105590820312, 0.06438108825683594, 0.06513136291503906, 0.0652042236328125, 0.06466764831542969, 0.0648182373046875, 0.0652471694946289, 0.06463795471191407, 0.06535270690917969, 0.06472806549072266, 0.06492364501953125, 0.06434099578857422, 0.06462464141845703, 0.38250927734375, 0.06540060424804688, 0.06434006500244141, 0.06549903869628906, 0.06582067108154296, 0.06531788635253906, 0.065512451171875, 0.06530668640136719, 0.06484678649902344, 0.06522573089599609, 0.06520832061767579, 0.06513459014892578, 0.06517247772216797, 0.06514380645751953, 0.06566300964355469, 0.0655738525390625, 0.06528409576416015, 0.06539469146728516, 0.06510079956054687, 0.06280307388305664, 0.06266662216186523, 0.06264115142822266, 0.06500454711914062, 0.0650782699584961, 0.06692352294921874, 0.06611251068115234, 0.06553395080566406, 0.06558617401123047, 0.06529126739501953, 0.06523289489746094, 0.06502194976806641, 0.06531686401367187, 0.0651847686767578, 0.06523085021972656, 0.06536908721923829, 0.06602444458007813, 0.06519910430908203, 0.06543257904052735, 0.06541926574707031, 0.06592729949951172, 0.06681075286865235, 0.06539884948730469, 0.06508956909179688, 0.06490614318847657, 0.06526060485839844, 0.06566102600097656, 0.06539965057373047, 0.06648934173583984, 0.06473216247558594, 0.06533952331542969, 0.06518873596191406, 0.06527283477783204, 0.06523004913330079, 0.06529007720947265, 0.06512137603759766, 0.06488886260986328, 0.06517024230957032, 0.06512640380859375, 0.06485708618164063, 0.062481441497802735, 0.06458262634277344, 0.0654919662475586, 0.06536601257324219, 0.38089627075195315, 0.06518988800048828, 0.06530355072021485, 0.06517257690429687, 0.06545295715332031, 0.06555545806884766, 0.06555136108398438, 0.06549811553955077, 0.06560358428955078, 0.06535577392578125, 0.06535782623291016, 0.06521446228027344, 0.06530665588378906, 0.06529840087890625, 0.06531584167480468, 0.06410034942626953, 0.06526687622070312, 0.06571295928955079, 0.06529856109619141, 0.06531878662109375, 0.0653096923828125, 0.06534349060058593, 0.06537010955810547, 0.06535577392578125, 0.06526976013183594, 0.06529638671875, 0.06539059448242188, 0.06516019439697265, 0.06256435012817382, 0.06261862564086915, 0.062561279296875, 0.06540595245361328, 0.06517862701416016, 0.06510079956054687, 0.06531590270996093, 0.06521337890625, 0.06527385711669922, 0.06463999938964844, 0.06377369689941406, 0.06530764770507813, 0.06524006652832032, 0.06485708618164063, 0.06524928283691406, 0.06593228912353516, 0.06523699188232422, 0.06519827270507812, 0.06496339416503906, 0.06503116607666015, 0.06410240173339844, 0.06495645141601562, 0.06295446395874023, 0.06288703918457031, 0.06234201431274414, 0.06445894622802735, 0.06491731262207032, 0.06413107299804688, 0.06404198455810547, 0.06366310501098633, 0.06431145477294922, 0.06444220733642578, 0.06451318359375, 0.06470435333251953, 0.06535679626464844, 0.3778488464355469, 0.06316851043701172, 0.06511923217773438, 0.06508646392822266, 0.06487347412109375, 0.06503321838378906, 0.06483251190185547, 0.06532096099853515, 0.06524527740478515, 0.06453955078125, 0.06536601257324219, 0.06488678741455078, 0.06233190536499023, 0.06260838317871094, 0.06270156860351563, 0.06339379119873047, 0.06497280120849609, 0.06460006713867188, 0.06657433319091798, 0.06536704254150391, 0.06515507507324218, 0.0650782699584961, 0.06544076538085937, 0.0624189453125, 0.06262886428833007, 0.06534143829345704, 0.06503628540039062, 0.06469331359863281, 0.0651396484375, 0.06494924926757813, 0.06497689819335938, 0.06507110595703125, 0.06530048370361329, 0.06549606323242188, 0.06478848266601563, 0.06228889465332031, 0.0625172462463379, 0.06242521667480469, 0.06265945434570312, 0.06509977722167969, 0.06541007995605469, 0.06532195281982422, 0.06589952087402344, 0.06736998748779296, 0.06574908447265625, 0.06534339141845703, 0.06542642974853516, 0.06500863647460937, 0.06533017730712891, 0.06544080352783203, 0.06535062408447266, 0.06522476959228515, 0.06511507415771485, 0.06541926574707031, 0.06541619110107422, 0.06543769836425781, 0.06487145233154297, 0.06447113800048829, 0.06626290893554687, 0.06690099334716797, 0.06553395080566406, 0.06534041595458985, 0.06525132751464843, 0.37862808227539063, 0.06263603210449219, 0.062483455657958986, 0.06429901123046874, 0.0652943344116211, 0.06521855926513671, 0.06511411285400391, 0.06497792053222656, 0.06539679718017578, 0.06275680160522461, 0.06258380889892579, 0.06530560302734376, 0.06528614044189453, 0.06509670257568359, 0.06526464080810547, 0.06539469146728516, 0.0660684814453125, 0.0654694366455078, 0.06526054382324219, 0.06528323364257813, 0.0652132797241211, 0.06538137817382812, 0.06511830139160156, 0.06512937927246094, 0.0651704330444336, 0.06474240112304687, 0.06539878082275391, 0.06570611572265625, 0.06532902526855469, 0.06506508636474609, 0.06518160247802735, 0.0652154541015625, 0.06563021087646484, 0.06531788635253906, 0.06546758270263672, 0.06540780639648437, 0.06512127685546874, 0.06503731536865234, 0.06516738891601563, 0.06527177429199219, 0.065185791015625, 0.06272921752929687, 0.06525049591064454, 0.06545491027832032, 0.06454988861083985, 0.06471782684326172, 0.06474956512451172, 0.06520012664794922, 0.06517759704589844, 0.0652779541015625, 0.06481919860839844, 0.06451712036132813, 0.06510899353027344, 0.06409523010253906, 0.06458573150634765, 0.06457036590576172, 0.06437888336181641, 0.06520626831054688, 0.06500761413574219, 0.06277836990356446, 0.062053375244140625, 0.06241996765136719, 0.06438713836669922, 0.37796038818359373, 0.06235136032104492, 0.06301388931274414, 0.06501376342773438, 0.06514278411865235, 0.06538140869140625, 0.06288380813598633, 0.06259609603881836, 0.062475265502929686, 0.0625797119140625, 0.06240563201904297, 0.06547456359863281, 0.06501683044433594, 0.06498323059082031, 0.06503609466552734, 0.06473420715332032, 0.06501376342773438, 0.0651704330444336, 0.06520124816894532, 0.06506601715087891, 0.06543551635742187, 0.06529853057861328, 0.0624475212097168, 0.06249577713012695, 0.062387168884277346, 0.06229913711547851, 0.06252236938476563, 0.06237900924682617, 0.06391193771362305, 0.0624005126953125, 0.06242099380493164, 0.06583622741699219, 0.0651497573852539, 0.0650782699584961, 0.06493695831298828, 0.06500249481201172, 0.06555545806884766, 0.06658866882324219, 0.06554828643798828, 0.06492160034179688, 0.06500045013427734, 0.06515916442871093, 0.06501478576660157, 0.06511411285400391, 0.06500054168701172, 0.06491228485107423, 0.06493494415283203, 0.06524515533447266, 0.06497280120849609, 0.06482022094726562, 0.06477311706542968, 0.06489395141601563, 0.06476707458496093, 0.06491741180419922, 0.06496153259277344, 0.06261452865600586, 0.06229913711547851, 0.062077953338623044, 0.06206684875488281, 0.06376227188110352, 0.06483865356445312, 0.06509465789794922, 0.06623744201660156, 0.3784427490234375, 0.06271180725097657, 0.06354739379882812, 0.06501478576660157, 0.06600601959228515, 0.06535475158691406, 0.06518988800048828, 0.06522879791259766, 0.06529843139648438, 0.06503731536865234, 0.0671098861694336, 0.06607974243164062, 0.06312960052490234, 0.06237286376953125, 0.06323404693603515, 0.06615257263183594, 0.06567001342773438, 0.06650163269042969, 0.06588416290283203, 0.0651704330444336, 0.06506604766845703, 0.06490822601318359, 0.06518681335449218, 0.06547353363037109, 0.06521548461914063, 0.06331289672851563, 0.0650987548828125, 0.06526566314697266, 0.06501888275146485, 0.065006591796875, 0.06514995574951171, 0.06519602966308594, 0.06493901062011719, 0.06508243560791016, 0.06532089233398437, 0.06534662628173828, 0.06522259521484375, 0.06534041595458985, 0.06430105590820312, 0.06521548461914063, 0.06510499572753906, 0.06493174743652344, 0.06464217376708985, 0.06440742492675781, 0.06513356781005859, 0.0648622055053711, 0.06505779266357421, 0.06548480224609375, 0.06474649810791015, 0.06417721557617187, 0.06475052642822265, 0.064468994140625, 0.06240153503417969, 0.06427648162841797, 0.06534451293945312, 0.06511417388916016, 0.06516422271728516, 0.06483353424072266, 0.06506819152832032, 0.06571401977539063, 0.06503526306152344, 0.06576844787597656, 0.06538751983642578, 0.3793786926269531, 0.06513878631591796, 0.06499932861328125, 0.06506086730957031, 0.06596607971191407, 0.06498508453369141, 0.062427135467529295, 0.062066688537597656, 0.06369279861450196, 0.06546031951904296, 0.06506896209716796, 0.06504672241210938, 0.065010498046875, 0.06629580688476562, 0.06541007995605469, 0.06515708923339844, 0.06511615753173829, 0.06510079956054687, 0.06517350769042969, 0.06503014373779296, 0.06492057800292969, 0.06498406219482422, 0.06260540771484376, 0.06236662292480469, 0.06508338928222657, 0.06521654510498047, 0.06490723419189454, 0.06248755264282226, 0.06242099380493164, 0.06213836669921875, 0.06240665435791016, 0.06476493072509766, 0.0650465316772461, 0.06498508453369141, 0.06513478088378906, 0.06512313842773437, 0.06501376342773438, 0.06479257965087891, 0.06507520294189453, 0.06527078247070313, 0.06526976013183594, 0.06262067031860352, 0.06375219345092774, 0.06516121673583984, 0.06499635314941406, 0.06481817626953125, 0.06228684616088867, 0.06424473571777344, 0.062327808380126956, 0.06227251052856445, 0.06438706970214844, 0.06528102111816406, 0.06513561248779297, 0.06518991851806641, 0.06499734497070313, 0.06537935638427735, 0.0662988510131836, 0.06254796981811524, 0.062379199981689455, 0.06231849670410156, 0.06420982360839844, 0.0651151351928711, 0.06504243469238281, 0.377933837890625, 0.06211174392700195, 0.06514892578125, 0.0653834228515625, 0.06531807708740234, 0.06488966369628907, 0.06517453002929688, 0.06513356781005859, 0.06508953857421874, 0.06498406219482422, 0.06454886627197266, 0.06513983917236328, 0.06549696350097656, 0.06654668426513671, 0.06550630187988281, 0.06512742614746093, 0.06506396484375, 0.06512022399902344, 0.06520524597167969, 0.06295859146118164, 0.06256332778930664, 0.0624793586730957, 0.06250707244873047, 0.0627946891784668, 0.06482329559326172, 0.06497074890136718, 0.06502198028564453, 0.06499427032470703, 0.06490009307861327, 0.06841139221191406, 0.06654054260253907, 0.06527999877929687, 0.06453657531738281, 0.0652779541015625, 0.0652779541015625, 0.06667571258544921, 0.06675865936279297, 0.0650403823852539, 0.06536089324951172, 0.06430003356933593, 0.0648642578125, 0.06497996520996094, 0.06521855926513671, 0.06523596954345703, 0.06468720245361329, 0.064470947265625, 0.06477721405029296, 0.06553292846679687, 0.0654233627319336, 0.06623027038574218, 0.06555852508544922, 0.06566194915771484, 0.06392012786865234, 0.062088191986083986, 0.06292889785766602, 0.0625602569580078, 0.06329651260375976, 0.06531584167480468, 0.06246303939819336, 0.06242297744750976, 0.06270566558837891, 0.0625684471130371, 0.062459903717041014]",tokens/s,14.474579011797019,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10978,7 +10978,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11177,7 +11177,7 @@ OSError: M is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11597,7 +11597,7 @@ Traceback (most recent call last): OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11726,7 +11726,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1258.205184,20246.429696,0.0,19599.982592,18883.392512,s,10,14.637272094726562,1.463727209472656,0.0011422170241880813,1.4633424682617187,1.4647964965820313,1.4656775451660156,1.4663823840332033,"[1.46655859375, 1.4646007080078125, 1.464500732421875, 1.462626708984375, 1.4627454833984375, 1.4629149169921876, 1.4630015869140625, 1.463301513671875, 1.463638427734375, 1.4633834228515625]",tokens/s,174.8959767525469,kWh,1.7319040844837825e-05,9.490735810250043e-06,0.000115426231229801,0.00014223600788488885,tokens/kWh,1799825.542117155,MB,1258.205184,20246.429696,0.0,19599.982592,18883.395072,s,10,74.99658349609375,7.4996583496093745,0.07745567856721523,7.51322607421875,7.594755517578125,7.617662670898437,7.635988393554688,"[7.40775830078125, 7.40062255859375, 7.5155556640625, 7.510896484375, 7.64056982421875, 7.3875166015625, 7.5222734375, 7.5896650390625, 7.516587890625, 7.5051376953125]",tokens/s,8.400382665869227,kWh,8.816561998592486e-05,4.832256516634515e-05,0.0003498333076441984,0.0004863214927964684,tokens/kWh,129543.9353044721,,s,629,88.04938673400879,0.13998312676313002,0.17380070114491072,0.11748863983154297,0.12289228973388672,0.1233991683959961,1.580889833984375,"[0.11697049713134766, 0.11677391815185546, 0.11645235443115234, 0.11661513519287109, 0.11661516571044922, 0.11714457702636719, 0.11668685150146485, 0.11664895629882813, 0.1166376953125, 0.116853759765625, 0.11657318115234375, 0.11655987548828126, 0.12227174377441406, 0.11783372497558593, 0.11668582153320313, 0.1219788818359375, 0.11895091247558594, 0.1225185317993164, 0.12198502349853516, 0.1193482208251953, 0.12163788604736328, 0.12191027069091796, 0.12210892486572265, 0.11984384155273438, 0.1190113296508789, 0.11865395355224609, 0.11642476654052734, 0.11650041961669921, 0.1170873260498047, 0.11642668914794922, 0.11743126678466796, 0.11731046295166016, 0.11665408325195313, 0.11662643432617187, 0.11647078704833984, 0.11681996917724609, 0.11640115356445313, 0.11671244812011719, 0.11657113647460937, 0.11655577850341797, 0.11771084594726562, 0.11678822326660156, 0.11621587371826173, 0.11655264282226563, 0.11678822326660156, 0.11654962921142578, 0.1166192626953125, 0.11692339324951172, 0.11663565063476562, 0.11656816101074219, 0.11652188873291015, 0.11677696228027344, 0.11702783966064453, 0.11717839813232422, 0.116938720703125, 0.11687731170654297, 0.11710566711425781, 0.11706163024902344, 0.11662540435791016, 0.11696332550048828, 0.11677490997314453, 0.11707801818847656, 1.5830650634765624, 0.11702988433837891, 0.1173780517578125, 0.11717648315429688, 0.1172980194091797, 0.11702783966064453, 0.11782246398925782, 0.11699712371826172, 0.11643289947509766, 0.12108799743652343, 0.11891609954833984, 0.1175541763305664, 0.11761561584472656, 0.11693363189697266, 0.11722035217285157, 0.11695308685302734, 0.11759107208251954, 0.11680252838134765, 0.11728179168701172, 0.11716812896728515, 0.11738214111328125, 0.11714457702636719, 0.11859455871582031, 0.11908505249023438, 0.11773133087158204, 0.11740271759033204, 0.11747216033935547, 0.11695308685302734, 0.11939020538330078, 0.11750605010986329, 0.1169623031616211, 0.11706368255615235, 0.11740467071533203, 0.11661721801757813, 0.11719996643066406, 0.11918940734863281, 0.11963187408447265, 0.11795763397216796, 0.11755007934570312, 0.1173565444946289, 0.11739238739013672, 0.11735040283203126, 0.11774163055419921, 0.11692230224609375, 0.11704528045654297, 0.11618096160888672, 0.11564646148681641, 0.11656294250488282, 0.11666944122314453, 0.11695820617675781, 0.11738317108154298, 0.11707392120361328, 0.11839590454101563, 0.11699404907226563, 0.11711385345458984, 0.1172295684814453, 0.11738829040527343, 0.11713951873779296, 0.1174896011352539, 0.11723382568359375, 0.11751817321777344, 0.11703807830810548, 0.11779481506347657, 1.584041015625, 0.11698687744140625, 0.11732582092285156, 0.11827916717529297, 0.11817779541015624, 0.11919564819335937, 0.11756646728515625, 0.11749581146240234, 0.11880550384521485, 0.1167615966796875, 0.11686911773681641, 0.11654450988769531, 0.1168384017944336, 0.11704013061523437, 0.11718553924560547, 0.11697062683105469, 0.12026764678955078, 0.11739552307128906, 0.11875526428222656, 0.1180200958251953, 0.117032958984375, 0.11903807830810546, 0.11744038391113282, 0.11820134735107422, 0.12284416198730469, 0.12328038024902344, 0.12200447845458984, 0.1163325424194336, 0.11634188842773438, 0.11696217346191407, 0.11718450927734375, 0.11699404907226563, 0.12314214324951171, 0.12321289825439453, 0.12345651245117187, 0.12300176239013672, 0.12299878692626953, 0.11723583984375, 0.11821043395996093, 0.11717017364501953, 0.11701145935058593, 0.11683328247070313, 0.11808870697021484, 0.12274585723876953, 0.12400640106201172, 0.12288102722167969, 0.11721113586425781, 0.12258099365234375, 0.1177528305053711, 0.12230655670166016, 0.11694502258300782, 0.11877772521972656, 0.12266496276855468, 0.12000972747802735, 0.12285234832763672, 0.12257081604003907, 0.11730540466308594, 0.12036799621582031, 0.12296205139160156, 0.12274674987792969, 0.1223178253173828, 0.11648307037353516, 0.12008448028564453, 1.580580810546875, 0.11685379028320313, 0.1168148193359375, 0.11668185424804688, 0.11680867004394531, 0.11675126647949219, 0.11655270385742188, 0.12148121643066406, 0.11710566711425781, 0.11658649444580078, 0.11620774078369141, 0.11638976287841797, 0.11647590637207031, 0.12168105316162109, 0.12297097778320312, 0.11673497772216797, 0.11654246520996094, 0.1167237091064453, 0.12039065551757812, 0.11676876831054687, 0.12154268646240235, 0.12241506958007813, 0.12281958770751954, 0.11727462768554688, 0.12269977569580078, 0.1240731201171875, 0.11954364776611329, 0.1231800308227539, 0.11741081237792969, 0.11680870056152344, 0.11664998626708985, 0.11674111938476563, 0.116890625, 0.11674626922607421, 0.11686294555664062, 0.12165325164794921, 0.12328656005859374, 0.1196748504638672, 0.12251554870605469, 0.1228737564086914, 0.12453478240966796, 0.12351999664306641, 0.1228422088623047, 0.12079606628417969, 0.12249394989013672, 0.11961138916015625, 0.11766067504882813, 0.11614208221435547, 0.11674111938476563, 0.11675238037109376, 0.1169090576171875, 0.11782144165039063, 0.11671858978271485, 0.12218582153320312, 0.1226454086303711, 0.11664281463623047, 0.12307660675048829, 0.12240076446533203, 0.11695616149902344, 0.11721932983398438, 0.12115455627441406, 0.12194713592529297, 0.11803955078125, 1.58009033203125, 0.11817696380615235, 0.1164071044921875, 0.11654656219482422, 0.11913932800292969, 0.12317183685302735, 0.12245401763916015, 0.1231431655883789, 0.12262911987304688, 0.12266700744628906, 0.12289842987060547, 0.12256156921386718, 0.12007421112060547, 0.12125389099121094, 0.12253081512451172, 0.12239052581787109, 0.12257177734375, 0.12305305480957031, 0.12214169311523437, 0.12233113861083984, 0.12244172668457032, 0.12314217376708984, 0.12242121887207032, 0.12271001434326172, 0.12220928192138672, 0.12257894134521484, 0.12289638519287109, 0.12243762969970703, 0.12312796783447266, 0.12344099426269531, 0.11931852722167968, 0.12063846588134766, 0.12275199890136719, 0.12561510467529297, 0.12448255920410156, 0.12362547302246094, 0.12231475067138672, 0.12190729522705078, 0.118868896484375, 0.12236799621582031, 0.12029132843017579, 0.12211920166015625, 0.12320355224609375, 0.12243257904052735, 0.12240377807617188, 0.12116889953613282, 0.12278886413574219, 0.12295270538330078, 0.12516864013671875, 0.12338995361328126, 0.12247555541992187, 0.11693670654296875, 0.11603657531738282, 0.11643289947509766, 0.11661315155029296, 0.1166878433227539, 0.11654246520996094, 0.11693363189697266, 0.1169532470703125, 0.11660784149169921, 0.12330290985107421, 0.12353330993652344, 0.1228062744140625, 1.581010009765625, 0.117834716796875, 0.11690393829345704, 0.11656089782714844, 0.11591372680664062, 0.11627315521240235, 0.11710975646972656, 0.11639398193359375, 0.11659468841552735, 0.11548365020751954, 0.11614927673339843, 0.11595568084716797, 0.11622809600830078, 0.11592499542236329, 0.11508531188964843, 0.1160939483642578, 0.11654246520996094, 0.11663053131103515, 0.11678521728515626, 0.11654354858398437, 0.11675942230224609, 0.11667865753173828, 0.11667049407958985, 0.1167052459716797, 0.11615763092041016, 0.11552851104736328, 0.11577958679199218, 0.11648716735839844, 0.11680870056152344, 0.1165322265625, 0.11690393829345704, 0.11679436492919922, 0.12105830383300781, 0.12254720306396484, 0.11724800109863281, 0.11636224365234375, 0.11613798522949219, 0.11670425415039062, 0.11677798461914063, 0.11677286529541016, 0.11707801818847656, 0.116384765625, 0.11954585266113281, 0.12351590728759766, 0.12475814056396485, 0.11717926025390625, 0.11675545501708984, 0.11686093139648437, 0.11688550567626953, 0.12201062774658203, 0.1220495376586914, 0.11716710662841796, 0.11716095733642579, 0.11677305603027344, 0.11648697662353516, 0.11677388763427735, 0.11667565155029297, 0.11801696014404296, 0.1178449935913086, 0.11684249877929688, 0.11700326538085938, 0.11652301025390625, 0.11673907470703125, 1.581275146484375, 0.11748863983154297, 0.11664486694335938, 0.11649126434326172, 0.11785948944091797, 0.11713929748535157, 0.11679641723632812, 0.11788390350341797, 0.12206182098388672, 0.12239564514160156, 0.12237312316894532, 0.1168384017944336, 0.11705548858642578, 0.1168384017944336, 0.1169100799560547, 0.11635916900634766, 0.11654962921142578, 0.11633049774169922, 0.11702185821533204, 0.11704815673828126, 0.11633561706542969, 0.11652521514892578, 0.12138684844970703, 0.12240076446533203, 0.12277555084228516, 0.12353536224365234, 0.11729312133789062, 0.1222174072265625, 0.12251763153076171, 0.12197670745849609, 0.12041852569580078, 0.1223309097290039, 0.12219391632080077, 0.12216934204101562, 0.12186422729492187, 0.12212735748291016, 0.12243247985839843, 0.11990118408203125, 0.12330086517333984, 0.11856690979003906, 0.12289126586914062, 0.12039987182617187, 0.12219391632080077, 0.11851884460449219, 0.12242221069335937, 0.12012339019775391, 0.12225331115722657, 0.1183888931274414, 0.11949142456054687, 0.1223526382446289, 0.12256358337402344, 0.12382937622070313, 0.12346969604492188, 0.11700531005859376, 0.11652915191650391, 0.11690598297119141, 0.11651583862304687, 0.11644108581542968, 0.11620556640625, 0.11591168212890625, 0.11649852752685547, 0.11619625854492187, 0.11659673309326171, 1.5818782958984374, 0.11682300567626953, 0.12141670227050781, 0.12372480010986328, 0.11739443206787109, 0.1169459228515625, 0.11882086181640625, 0.12272745513916015, 0.12225532531738281, 0.12241613006591796, 0.12245503997802734, 0.12226866912841797, 0.11698175811767578, 0.11699302673339844, 0.11690496063232422, 0.11671756744384766, 0.11661824035644532, 0.11654978942871094, 0.11654025268554688, 0.11698995208740234, 0.1224796142578125, 0.12271820831298828, 0.12233318328857422, 0.12256870269775391, 0.11703398132324219, 0.116853759765625, 0.11730944061279297, 0.12038143920898438, 0.11666124725341796, 0.11692342376708985, 0.1169735336303711, 0.11704319763183593, 0.11704013061523437, 0.12262911987304688, 0.12276428985595703, 0.12273267364501952, 0.12245491027832031, 0.12226150512695312, 0.1190983657836914, 0.12055859375, 0.1224089584350586, 0.12299059295654297, 0.12269782257080078, 0.12255836486816406, 0.12242431640625, 0.11794329833984375, 0.11944454193115234, 0.12315251159667968, 0.12314604949951172, 0.12050640106201171, 0.12275606536865234, 0.12213555145263671, 0.12255948638916016, 0.1188362274169922, 0.12251238250732421, 0.12267622375488281, 0.12289740753173828, 0.12249702453613281, 0.122608642578125, 0.12282588958740234, 0.1251654052734375, 0.12357939147949219, 0.12244172668457032, 1.587788818359375, 0.11697459411621093, 0.11695104217529297, 0.11671244812011719, 0.11685478210449218, 0.11689472198486328, 0.11674931335449219, 0.12031999969482422, 0.12235884857177734, 0.11755104064941406, 0.12134400177001953, 0.12108799743652343, 0.12294143676757813, 0.11628441619873046, 0.12067635345458984, 0.1228062744140625, 0.11708534240722657, 0.11753865814208984, 0.12036300659179687, 0.12275917053222657, 0.11877785491943359, 0.12369622039794922, 0.11707382202148438, 0.1166407699584961, 0.12219404602050782, 0.11680857849121094, 0.11952243041992187, 0.12274368286132813, 0.12147404479980468, 0.12212223815917969, 0.12273458862304687, 0.1187041244506836, 0.11961241912841797, 0.11963289642333984, 0.12265676879882813, 0.12340531158447265, 0.12238848114013672, 0.11632230377197265, 0.11668172454833985, 0.11670118713378906, 0.1182218246459961, 0.11650969696044922, 0.11663475036621093, 0.11924578857421875, 0.12250819396972656, 0.12217241668701172, 0.12290150451660156, 0.1168875503540039, 0.11657939147949219, 0.11703084564208985, 0.11662438201904297, 0.11717632293701172, 0.11672582244873046, 0.11667449951171875, 0.11703501129150391, 0.1167462387084961, 0.11695410919189453, 0.11826892852783204, 0.1223034896850586, 0.12212531280517579, 0.12264345550537109, 0.116600830078125, 0.11975373077392579, 1.5813775634765626, 0.11741388702392579, 0.1229496307373047, 0.12204544067382812, 0.12242534637451172, 0.1226424331665039, 0.1226168670654297, 0.12079004669189453, 0.11764422607421875, 0.1168752670288086, 0.11698790740966797, 0.11660697937011719, 0.11682303619384765, 0.11655987548828126, 0.11689266967773437, 0.11718656158447266, 0.11672166442871093, 0.11654144287109375, 0.1182382049560547, 0.12277359771728516, 0.12621097564697265, 0.11681279754638672, 0.12052992248535156, 0.12342988586425781, 0.12130713653564452, 0.12307456207275391, 0.1173719711303711, 0.1226321258544922, 0.11656294250488282, 0.11661721801757813, 0.11708211517333984, 0.11675238037109376, 0.1190759048461914, 0.11658131408691406, 0.1166192626953125, 0.1170022430419922, 0.11695718383789062, 0.11674419403076172, 0.11668582153320313, 0.11692031860351562, 0.1163387222290039, 0.12035785675048828, 0.1218519058227539, 0.11692765045166016, 0.12120252990722656, 0.12251974487304687, 0.12275794982910156, 0.12311551666259765, 0.1164738540649414, 0.11763510131835937, 0.11669808197021485, 0.12166963195800781, 0.12263014221191407, 0.12276428985595703, 0.11694284820556641, 0.11680563354492188, 0.12139826965332032, 0.12269680023193359, 0.1225030746459961, 0.11705753326416016, 0.11745791625976562, 0.11709849548339844, 0.11657437133789063]",tokens/s,7.143718126057667,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11854,7 +11854,7 @@ OSError: google/recurrentgemma-7b is not a local folder and is not a valid model If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12199,7 +12199,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12301,7 +12301,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 96.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12603,7 +12603,7 @@ OSError: 0 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-bnb-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-bnb-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,bnb,0.0,True,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12843,7 +12843,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h raise RuntimeError(f""Isolated process exited with non-zero code {isolated_process.exitcode}"") RuntimeError: Isolated process exited with non-zero code -9 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13126,7 +13126,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1329.041408,14877.720576,0.0,14222.884864,13889.381376,s,10,1.4499510650634766,0.14499510650634767,0.003622698731783719,0.14306078338623046,0.14787741088867187,0.15126964569091797,0.15398343353271485,"[0.15466188049316407, 0.14258636474609376, 0.14280589294433593, 0.1425797119140625, 0.14328611755371093, 0.1424152069091797, 0.14660438537597656, 0.14505247497558593, 0.14712358093261718, 0.14283544921875]",tokens/s,1765.5768264758142,kWh,1.7007446265409864e-06,9.318517474743462e-07,5.344702688457207e-06,7.97729906247254e-06,tokens/kWh,32091062.149631076,MB,1329.041408,14877.720576,0.0,14222.884864,14149.18144,s,10,90.27044531250002,9.02704453125,0.16841050656631779,8.950941894531251,9.26629443359375,9.266525146484375,9.266709716796875,"[8.866875, 8.87135546875, 8.85364453125, 8.8795283203125, 8.9002509765625, 9.1472412109375, 9.21691796875, 9.266755859375, 9.2662431640625, 9.0016328125]",tokens/s,6.979028383199546,kWh,0.00010646788979806598,5.8352572228785514e-05,0.00021878811550654116,0.00038360857753339265,tokens/kWh,164229.90436003983,,s,629,91.42366313171381,0.1453476361394497,0.017392421234384187,0.14162124633789064,0.147710791015625,0.14806041870117187,0.2835976916503906,"[0.1402603454589844, 0.14037298583984376, 0.1464268798828125, 0.14179327392578125, 0.14173081970214843, 0.13927529907226563, 0.13979029846191407, 0.13987635803222656, 0.14051840209960936, 0.1409372100830078, 0.14039141845703124, 0.14034739685058595, 0.14041622924804686, 0.14138140869140625, 0.1421332550048828, 0.14084095764160157, 0.14070169067382812, 0.14022041320800782, 0.13921279907226564, 0.13935842895507813, 0.13890538024902344, 0.13885235595703124, 0.1391708221435547, 0.13949337768554687, 0.14031881713867186, 0.14153001403808593, 0.1435125732421875, 0.1402275848388672, 0.13964598083496094, 0.14154339599609375, 0.14011085510253907, 0.14128536987304688, 0.14071824645996095, 0.14032791137695313, 0.14039744567871093, 0.14089436340332032, 0.14175830078125, 0.14113381958007812, 0.14111347961425783, 0.14121868896484374, 0.14064332580566405, 0.1406771240234375, 0.1398835144042969, 0.14068028259277343, 0.14064834594726563, 0.1407836151123047, 0.14071090698242186, 0.14064956665039063, 0.14063104248046876, 0.140978271484375, 0.1408725128173828, 0.14080921936035157, 0.14458674621582032, 0.1409556427001953, 0.13938175964355468, 0.1406228485107422, 0.14070271301269532, 0.1407098846435547, 0.13964083862304688, 0.1407764434814453, 0.14051327514648437, 0.14049996948242188, 0.28369918823242185, 0.14024703979492187, 0.139863037109375, 0.1404549102783203, 0.14012416076660156, 0.140704833984375, 0.14019065856933594, 0.14082354736328126, 0.14052659606933593, 0.14179750061035157, 0.14145526123046875, 0.14110614013671874, 0.14052351379394531, 0.14059429931640624, 0.14019891357421874, 0.1423225555419922, 0.141011962890625, 0.13974528503417968, 0.14011289978027344, 0.13898045349121094, 0.13951171875, 0.14035763549804686, 0.14147584533691407, 0.1399603271484375, 0.13958758544921876, 0.14093927001953124, 0.13986405944824218, 0.14072525024414062, 0.1402767333984375, 0.14047952270507813, 0.14075389099121094, 0.14076722717285156, 0.14113075256347657, 0.14133657836914063, 0.14340914916992187, 0.14082456970214843, 0.14061158752441405, 0.14063002014160156, 0.14129356384277345, 0.14074674987792968, 0.1408737335205078, 0.14188954162597656, 0.14042434692382813, 0.14083363342285157, 0.14294117736816406, 0.14097817993164063, 0.1424803771972656, 0.14126182556152345, 0.14050816345214845, 0.1407068176269531, 0.14032281494140625, 0.14141337585449218, 0.139720703125, 0.14003404235839845, 0.14095977783203126, 0.14106416320800783, 0.1404067840576172, 0.14025021362304688, 0.14061967468261719, 0.1403822021484375, 0.1444444122314453, 0.14167961120605468, 0.1405839385986328, 0.28396035766601563, 0.14065968322753905, 0.14101402282714845, 0.14088499450683595, 0.1407406005859375, 0.14055322265625, 0.14124237060546874, 0.14063002014160156, 0.14043341064453124, 0.14068223571777344, 0.14037709045410157, 0.14000230407714845, 0.1405072021484375, 0.14406649780273437, 0.13911961364746095, 0.13865061950683594, 0.13842329406738282, 0.13882266235351562, 0.13917695617675782, 0.13875302124023436, 0.14064646911621093, 0.13859117126464843, 0.138392578125, 0.13921586608886719, 0.14080613708496093, 0.1405286407470703, 0.14030642700195312, 0.1406947479248047, 0.14104246520996094, 0.14394163513183594, 0.14096383666992188, 0.14137957763671874, 0.1405716552734375, 0.14055632019042968, 0.140348388671875, 0.14031564331054688, 0.14044058227539064, 0.14087986755371093, 0.14068019104003907, 0.14076643371582032, 0.14059703063964843, 0.14274765014648438, 0.14095584106445314, 0.14040249633789062, 0.14102732849121094, 0.1401282501220703, 0.1405173797607422, 0.14013235473632812, 0.14007705688476563, 0.14030029296875, 0.14083599853515624, 0.1426287078857422, 0.1406023712158203, 0.14042930603027343, 0.14100787353515626, 0.14083174133300783, 0.13972787475585938, 0.14043545532226562, 0.14062591552734374, 0.14066175842285156, 0.14050015258789061, 0.14091961669921876, 0.14061056518554688, 0.2832723388671875, 0.14092784118652343, 0.1407283172607422, 0.1418577880859375, 0.14189158630371093, 0.14050521850585937, 0.14065664672851563, 0.14066169738769532, 0.14011386108398438, 0.13987020874023437, 0.13993778991699218, 0.14005964660644532, 0.13910630798339843, 0.1395353546142578, 0.13887590026855468, 0.13946266174316407, 0.13874790954589844, 0.13999411010742188, 0.14009855651855468, 0.14021221923828125, 0.14184243774414063, 0.14108364868164064, 0.1406750793457031, 0.14061978149414062, 0.14074266052246093, 0.1408155517578125, 0.14050096130371093, 0.14061039733886718, 0.14055526733398438, 0.14050611877441407, 0.14036480712890625, 0.14058189392089843, 0.1401231689453125, 0.14053578186035157, 0.14038426208496094, 0.1406156768798828, 0.14324237060546874, 0.1437664031982422, 0.14272921752929688, 0.1440809020996094, 0.14227558898925782, 0.14239347839355468, 0.1416538848876953, 0.14050405883789063, 0.13961932373046876, 0.13987327575683595, 0.1411112976074219, 0.14117376708984375, 0.14199296569824219, 0.14025932312011719, 0.1406156768798828, 0.14192025756835938, 0.14207180786132811, 0.14102015686035158, 0.141085693359375, 0.1406894073486328, 0.14104066467285156, 0.14095254516601563, 0.14383110046386718, 0.1420389709472656, 0.14088088989257813, 0.14321868896484374, 0.1410519104003906, 0.2849310607910156, 0.14295654296875, 0.14093331909179688, 0.1415084228515625, 0.1410867156982422, 0.14080307006835938, 0.14064639282226563, 0.13962547302246095, 0.14005043029785155, 0.13997669982910158, 0.14062797546386718, 0.14273023986816405, 0.1395916748046875, 0.13970431518554688, 0.1417872314453125, 0.1409064025878906, 0.1406750793457031, 0.14018663024902345, 0.14106521606445313, 0.14078157043457032, 0.14062693786621094, 0.14070169067382812, 0.1402890625, 0.14045590209960937, 0.1403822021484375, 0.14092288208007814, 0.1405849609375, 0.14085939025878907, 0.14295347595214844, 0.14229299926757813, 0.1414082489013672, 0.14331904602050782, 0.14096092224121093, 0.140540771484375, 0.14029618835449217, 0.14162124633789064, 0.14073651123046876, 0.14152703857421875, 0.14072627258300782, 0.14313471984863282, 0.14023680114746093, 0.14266879272460936, 0.14233088684082032, 0.14444851684570312, 0.14382080078125, 0.14268214416503905, 0.1423380126953125, 0.14065664672851563, 0.13995213317871094, 0.14267401123046874, 0.1408408660888672, 0.14078668212890624, 0.14072525024414062, 0.1423493194580078, 0.14353526306152345, 0.14239010620117187, 0.14404197692871093, 0.14009957885742189, 0.14021939086914062, 0.14058189392089843, 0.14086553955078124, 0.13991322326660155, 0.14053683471679687, 0.28333670043945314, 0.1473904571533203, 0.1476444091796875, 0.14804071044921874, 0.14739967346191407, 0.14656614685058594, 0.14412185668945313, 0.14544793701171874, 0.14030540466308594, 0.14021836853027345, 0.1409986572265625, 0.13920477294921876, 0.1398055419921875, 0.14363442993164063, 0.14538240051269533, 0.1462775421142578, 0.147297119140625, 0.14977030944824218, 0.1476392364501953, 0.1465529022216797, 0.14685279846191407, 0.14831103515625, 0.1473822784423828, 0.14729728698730468, 0.14736793518066407, 0.14796389770507812, 0.147704833984375, 0.14720822143554688, 0.14720407104492186, 0.14435430908203126, 0.14047442626953124, 0.14115423583984374, 0.14079898071289063, 0.1477734375, 0.14729318237304687, 0.14761882019042968, 0.1476444091796875, 0.14776626586914063, 0.14800384521484375, 0.1468211212158203, 0.14527693176269532, 0.14556474304199218, 0.14009866333007812, 0.14488150024414062, 0.14618009948730468, 0.1407139892578125, 0.1406750793457031, 0.14068736267089843, 0.14438092041015624, 0.14784614562988282, 0.14692352294921876, 0.146204833984375, 0.14070664978027345, 0.14088703918457032, 0.14037298583984376, 0.1425070037841797, 0.147061767578125, 0.14701266479492187, 0.14766893005371093, 0.14769970703125, 0.14805926513671874, 0.14724185180664062, 0.1470013427734375, 0.2936524658203125, 0.14070579528808594, 0.13933363342285157, 0.13916876220703125, 0.139683837890625, 0.14026138305664063, 0.1463889923095703, 0.14655897521972655, 0.14566502380371094, 0.14515814208984376, 0.14460928344726562, 0.14752153015136718, 0.1466542053222656, 0.14700440979003906, 0.147162109375, 0.14745805358886718, 0.14741606140136718, 0.146988037109375, 0.14716534423828126, 0.1464891815185547, 0.14759027099609376, 0.14738937377929687, 0.1490984344482422, 0.14782566833496094, 0.14733106994628906, 0.1439805450439453, 0.14775296020507814, 0.14289407348632813, 0.14773861694335938, 0.14791372680664064, 0.1474088897705078, 0.14718995666503906, 0.1458051452636719, 0.14718777465820312, 0.14729820251464842, 0.14723481750488282, 0.14710592651367188, 0.14465939331054686, 0.14665516662597655, 0.14572441101074218, 0.146693115234375, 0.14650880432128907, 0.1468098602294922, 0.1472379913330078, 0.1469715576171875, 0.1470556182861328, 0.14690509033203125, 0.1473454132080078, 0.14380134582519533, 0.14746931457519533, 0.1472481231689453, 0.1477570495605469, 0.14733824157714845, 0.14953273010253906, 0.147463134765625, 0.14791183471679686, 0.14748550415039063, 0.14737619018554687, 0.1475911102294922, 0.14600607299804688, 0.1467012481689453, 0.14695423889160156, 0.14652723693847655, 0.2940283508300781, 0.14675961303710938, 0.143463623046875, 0.1435482177734375, 0.1456670684814453, 0.14573362731933595, 0.1467535400390625, 0.14677520751953124, 0.14599868774414063, 0.14887834167480468, 0.1497733154296875, 0.14764236450195312, 0.14775193786621094, 0.14767308044433594, 0.1472296905517578, 0.14759730529785156, 0.1476495361328125, 0.14722149658203126, 0.1465232391357422, 0.147669921875, 0.143727783203125, 0.14686703491210937, 0.14619853210449218, 0.14686822509765626, 0.14794752502441405, 0.14725234985351562, 0.15463104248046874, 0.14785740661621094, 0.1473228759765625, 0.14746112060546876, 0.14710476684570312, 0.14583711242675781, 0.14731666564941406, 0.14712013244628908, 0.14708224487304689, 0.14478131103515626, 0.14523699951171876, 0.14730752563476562, 0.14505369567871093, 0.1487974395751953, 0.14810931396484375, 0.14776115417480468, 0.1470269470214844, 0.1475205078125, 0.14665113830566406, 0.14677197265625, 0.14784921264648437, 0.1472163848876953, 0.14769151306152345, 0.14728396606445313, 0.14720921325683595, 0.1477007293701172, 0.147736572265625, 0.14749388122558593, 0.14801612854003907, 0.14691123962402344, 0.14745497131347657, 0.14928179931640626, 0.1444323272705078, 0.1465536651611328, 0.1473976287841797, 0.14587699890136718, 0.14578175354003906, 0.29205606079101565, 0.1445918731689453, 0.14497587585449218, 0.14696563720703126, 0.14890687561035157, 0.14824960327148437, 0.147239013671875, 0.14738832092285156, 0.14819949340820313, 0.14741804504394532, 0.14798643493652344, 0.1476136932373047, 0.14797824096679688, 0.14783180236816407, 0.1499351043701172, 0.14828662109375, 0.14801187133789062, 0.14734950256347656, 0.14749900817871095, 0.1477591094970703, 0.14761164855957032, 0.14745805358886718, 0.14760140991210938, 0.1470023651123047, 0.146974853515625, 0.14734425354003905, 0.14674943542480468, 0.14688461303710937, 0.1467054138183594, 0.146008056640625, 0.14830490112304687, 0.14895936584472655, 0.14728076171875, 0.1471631317138672, 0.14741299438476563, 0.1476168975830078, 0.1477089538574219, 0.14737802124023439, 0.14726553344726562, 0.1465149383544922, 0.14608589172363282, 0.14401434326171875, 0.14744371032714843, 0.14752359008789062, 0.14806118774414062, 0.14913638305664062, 0.1477181396484375, 0.14762086486816406, 0.14745497131347657, 0.147842041015625, 0.14800282287597658, 0.147230712890625, 0.1467485809326172, 0.14647279357910156, 0.14609408569335938, 0.1456486358642578, 0.1465927734375, 0.14605722045898437, 0.1461862335205078, 0.14512025451660157, 0.14509671020507814, 0.14835302734375, 0.14072627258300782, 0.285275146484375, 0.1478481903076172, 0.14727474975585939, 0.14664089965820312, 0.14112051391601563, 0.1407406005859375, 0.14566604614257814, 0.14792807006835937, 0.148136962890625, 0.1481862030029297, 0.1471753234863281, 0.14738438415527344, 0.14729209899902343, 0.1475440673828125, 0.14769561767578124, 0.14695321655273438, 0.1469501495361328, 0.1472368621826172, 0.1485281219482422, 0.14786151123046876, 0.14744677734375, 0.1410549774169922, 0.14070169067382812, 0.14729624938964844, 0.14191416931152342, 0.14044256591796875, 0.14091673278808595, 0.14107749938964845, 0.14222438049316405, 0.14221720886230468, 0.14153114318847657, 0.14100582885742188, 0.14189260864257813, 0.14279986572265624, 0.1414307861328125, 0.1413949432373047, 0.14117068481445313, 0.14074777221679688, 0.14066790771484375, 0.14058189392089843, 0.14069760131835937, 0.1412024383544922, 0.14101298522949218, 0.14064332580566405, 0.14066278076171876, 0.14039859008789063, 0.14069760131835937, 0.14059519958496094, 0.140653564453125, 0.1409290313720703, 0.14143283081054686, 0.14213740539550782, 0.1396796875, 0.13964799499511718, 0.14026751708984375, 0.14111351013183593, 0.14037590026855468, 0.14000128173828125, 0.13970431518554688, 0.13942477416992188, 0.140653564453125, 0.14047660827636718, 0.14014341735839844]",tokens/s,6.880056852390624,,,True -8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13155,7 +13155,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True -8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13475,7 +13475,7 @@ RuntimeError: Isolated process exited with non-zero code -9 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5247.500288,4037.541888,0.0,3382.706176,3158.445056,s,10,0.8721447677612305,0.08721447677612304,0.0031421768280667146,0.08589214324951172,0.08872400512695312,0.09241056365966796,0.09535981048583984,"[0.09609712219238281, 0.08523616027832032, 0.08789775848388671, 0.08790476989746093, 0.08559571075439452, 0.08514822387695313, 0.0860749740600586, 0.08738550567626953, 0.08570931243896485, 0.08509523010253907]",tokens/s,2935.2925049031064,kWh,1.027848125943815e-06,5.632060641927755e-07,2.0028373510609353e-06,3.593891541197526e-06,tokens/kWh,71231977.11044386,MB,5247.500288,4037.541888,0.0,3382.706176,3158.447616,s,10,53.1214912109375,5.31214912109375,0.04650171531989755,5.3025014648437505,5.369855761718751,5.380759912109375,5.389483232421875,"[5.3674326171875, 5.36521484375, 5.32823681640625, 5.3916640625, 5.31485498046875, 5.275693359375, 5.26392041015625, 5.29014794921875, 5.26582421875, 5.258501953125]",tokens/s,11.859606830282008,kWh,6.296429655474165e-05,3.450854355784585e-05,0.00010277185757973723,0.0002002446976923247,tokens/kWh,314615.0720894457,,s,629,53.80693609619144,0.0855436185948989,0.010343391704824992,0.08387481689453125,0.08595374145507813,0.08656732177734375,0.16944504699707033,"[0.09003008270263672, 0.09052467346191406, 0.08755712127685547, 0.08822579193115235, 0.08802098846435547, 0.08344882965087891, 0.08568934631347656, 0.08573951721191406, 0.08564019012451173, 0.08323276519775391, 0.08730009460449219, 0.08577228546142578, 0.08292044830322266, 0.08304230499267579, 0.08383385467529297, 0.08587776184082031, 0.08593510437011719, 0.08330239868164062, 0.08336589050292968, 0.08521625518798828, 0.0858265609741211, 0.08580915069580078, 0.08608870697021484, 0.08389939117431641, 0.0838440933227539, 0.085718017578125, 0.08592800140380859, 0.08485472106933593, 0.08554291534423829, 0.08562278747558594, 0.0849612808227539, 0.08280268859863281, 0.08484556579589844, 0.08304434967041016, 0.08329933166503906, 0.08316620635986328, 0.08303718566894532, 0.0832890853881836, 0.08735539245605468, 0.08369049835205078, 0.08321644592285156, 0.08325523376464844, 0.08472882843017578, 0.08545382690429687, 0.08546099090576172, 0.08719155120849609, 0.08454144287109375, 0.08275865936279297, 0.08294092559814453, 0.08260300445556641, 0.08464179229736328, 0.08566783905029297, 0.08569241333007813, 0.08571084594726562, 0.08593920135498047, 0.08589107513427735, 0.08562687683105469, 0.08576409912109376, 0.08587366485595703, 0.0847984619140625, 0.08558386993408203, 0.08587161254882812, 0.17014682006835938, 0.08343551635742187, 0.08329933166503906, 0.08327474975585937, 0.08375193786621093, 0.08574259185791015, 0.08607539367675782, 0.08603955078125, 0.08405811309814454, 0.08570572662353515, 0.08592797088623047, 0.08638050842285157, 0.08658329772949219, 0.08621056365966796, 0.08608665466308593, 0.08525619506835938, 0.08453119659423829, 0.08554806518554688, 0.08405705261230469, 0.08304640197753907, 0.0842209243774414, 0.08543231964111328, 0.0857548828125, 0.08591462707519532, 0.08556134033203125, 0.0831457290649414, 0.08488448333740234, 0.08547430419921875, 0.08563507080078125, 0.08556543731689453, 0.08581017303466797, 0.08562790679931641, 0.08578559875488281, 0.08557977294921874, 0.08584703826904297, 0.08391168212890625, 0.08575078582763672, 0.08550099182128906, 0.0855059814453125, 0.08304025268554688, 0.08516403198242188, 0.08283033752441406, 0.08511897277832031, 0.0839178237915039, 0.0865433578491211, 0.08618701171875, 0.08642867279052735, 0.08407142639160156, 0.08583577728271484, 0.08580812835693359, 0.08586342620849609, 0.08569344329833985, 0.08550707244873047, 0.08564121246337891, 0.08578867340087891, 0.08559513854980469, 0.08580812835693359, 0.0847267837524414, 0.08318566131591797, 0.08408780670166016, 0.0857681884765625, 0.08585727691650391, 0.08593817901611328, 0.17426329040527344, 0.08697548675537109, 0.08563404846191407, 0.08592076873779297, 0.08471858978271485, 0.08524492645263672, 0.08488652801513671, 0.08525414276123047, 0.08332697296142579, 0.08319181060791016, 0.083093505859375, 0.08534732818603516, 0.08466432189941406, 0.08506880187988282, 0.08295423889160156, 0.08362496185302734, 0.08389119720458985, 0.0857518081665039, 0.08302285003662109, 0.08377753448486328, 0.08289997100830078, 0.08441548919677734, 0.08568422698974609, 0.08546611022949219, 0.08261939239501953, 0.08512409973144532, 0.08593612670898437, 0.08389017486572266, 0.08483026885986328, 0.08756320190429688, 0.08599040222167968, 0.0851630096435547, 0.08313753509521485, 0.08482303619384765, 0.08604774475097657, 0.08573951721191406, 0.08692431640625, 0.0835522232055664, 0.08365158081054687, 0.08568220520019532, 0.08447894287109375, 0.08314265441894532, 0.08392499542236329, 0.08497151947021485, 0.08538931274414062, 0.08357273864746094, 0.08299008178710937, 0.0827166748046875, 0.08306278228759766, 0.08348159790039063, 0.0840847396850586, 0.08587059020996093, 0.08575078582763672, 0.08571596527099609, 0.08593714904785156, 0.08536064147949218, 0.08350003051757812, 0.08355020904541016, 0.08463565063476562, 0.08456499481201171, 0.08303001403808594, 0.083198974609375, 0.0834867172241211, 0.1739008026123047, 0.08579891204833984, 0.08635187530517578, 0.085718017578125, 0.08566783905029297, 0.0833812484741211, 0.08508006286621093, 0.08582553863525391, 0.08582963562011718, 0.08525004577636719, 0.08693862152099609, 0.08582867431640626, 0.08577222442626953, 0.0857702407836914, 0.08600985717773438, 0.08582246398925782, 0.0847984619140625, 0.08450559997558593, 0.08597708892822266, 0.08559923553466797, 0.08593920135498047, 0.08573849487304687, 0.08601702117919922, 0.08586239624023438, 0.08572927856445313, 0.08574668884277344, 0.08595455932617188, 0.08594226837158203, 0.08567193603515624, 0.08596377563476562, 0.08611328125, 0.08596377563476562, 0.08586752319335937, 0.08632627105712891, 0.08665497589111328, 0.08358092498779297, 0.08633958435058593, 0.08401203155517578, 0.08223436737060547, 0.08387174224853515, 0.08427008056640625, 0.08511590576171875, 0.08579788970947265, 0.08599756622314453, 0.08602931213378906, 0.08438886260986328, 0.08223129272460937, 0.08431820678710937, 0.08585215759277344, 0.0856258544921875, 0.08616754913330078, 0.08551526641845703, 0.08328806304931641, 0.08506060791015625, 0.08577126312255859, 0.09385676574707032, 0.0867799072265625, 0.08598118591308594, 0.08595865631103515, 0.08583372497558593, 0.08590438079833984, 0.08524492645263672, 0.08528793334960938, 0.16948941040039062, 0.08370073699951172, 0.08343551635742187, 0.08495616149902344, 0.08540672302246094, 0.08462131500244141, 0.0844554214477539, 0.08563404846191407, 0.08586854553222656, 0.08576716613769532, 0.0854549102783203, 0.08528582763671876, 0.08585113525390625, 0.0854097900390625, 0.08558796691894531, 0.08586342620849609, 0.08547020721435547, 0.08567910766601562, 0.08478617858886718, 0.0847984619140625, 0.08551526641845703, 0.0855910415649414, 0.08553266906738281, 0.08559001922607422, 0.08561766052246093, 0.08698675537109375, 0.08482099151611328, 0.0828375015258789, 0.08300339508056641, 0.08316108703613281, 0.08487117004394532, 0.08570572662353515, 0.08608870697021484, 0.08535756683349609, 0.08339968109130859, 0.08323276519775391, 0.08315187072753906, 0.08345804595947266, 0.08327474975585937, 0.08335564422607422, 0.08313855743408204, 0.08319385528564453, 0.08344271850585938, 0.08523158264160156, 0.08416973114013672, 0.08317951965332031, 0.08319487762451172, 0.08326246643066407, 0.08492851257324219, 0.08470732879638672, 0.0832174072265625, 0.08330854034423828, 0.08325536346435547, 0.0836033935546875, 0.08312525177001953, 0.08348467254638672, 0.0834549789428711, 0.08345394897460938, 0.08302694702148437, 0.08530022430419922, 0.08327986907958984, 0.08292864227294922, 0.08248524475097656, 0.16946482849121094, 0.08301465606689454, 0.08296141052246093, 0.08293888092041016, 0.08296141052246093, 0.08530944061279297, 0.08531251525878907, 0.08570777893066406, 0.08339456176757812, 0.08311500549316406, 0.08301363372802735, 0.08328192138671875, 0.0827146224975586, 0.08361881256103515, 0.083093505859375, 0.08353385925292969, 0.08388912200927734, 0.0831283187866211, 0.08429158020019531, 0.08508927917480469, 0.08310374450683594, 0.08285593414306641, 0.08308223724365234, 0.08316313934326172, 0.08497663879394532, 0.082229248046875, 0.0830382080078125, 0.08300032043457031, 0.08277811431884766, 0.08388198089599609, 0.08334233856201172, 0.08317337799072265, 0.08295219421386718, 0.08301773071289062, 0.08406425476074218, 0.0853934097290039, 0.08444518280029296, 0.08305868530273437, 0.08334233856201172, 0.08330239868164062, 0.08454656219482422, 0.08342733001708984, 0.08342221069335938, 0.08492543792724609, 0.08518450927734375, 0.08653517150878906, 0.084600830078125, 0.0840284194946289, 0.08309862518310547, 0.08302694702148437, 0.08317337799072265, 0.0828927993774414, 0.08419123077392578, 0.08512716674804688, 0.08524396514892578, 0.08484758758544922, 0.08496431732177734, 0.08352767944335937, 0.08530022430419922, 0.08345088195800782, 0.08268287658691406, 0.08209919738769532, 0.08258252716064453, 0.16952627563476563, 0.08308019256591796, 0.08312934112548828, 0.08258150482177734, 0.08366182708740234, 0.08349491119384765, 0.08257331085205079, 0.08242278289794921, 0.08185753631591797, 0.08244429016113282, 0.0849991683959961, 0.08375091552734375, 0.0834549789428711, 0.08324403381347656, 0.08316928100585938, 0.08371609497070312, 0.0831272964477539, 0.08728883361816406, 0.08347545623779297, 0.08325631713867188, 0.08328192138671875, 0.08307199859619141, 0.08253030395507813, 0.08320921325683593, 0.08423117065429687, 0.0848957748413086, 0.0832501449584961, 0.08318975830078125, 0.08393830108642578, 0.0833986587524414, 0.08343756866455078, 0.08331372833251953, 0.08341804504394532, 0.08331059265136719, 0.08328806304931641, 0.08297471618652344, 0.08410726165771484, 0.08523571014404296, 0.08347238159179687, 0.08317337799072265, 0.08333516693115234, 0.08376627349853516, 0.08334438323974609, 0.0833095703125, 0.08334848022460938, 0.08325939178466797, 0.08293580627441406, 0.08432742309570312, 0.08447795104980468, 0.08309248352050781, 0.083525634765625, 0.08322354888916016, 0.08326246643066407, 0.083272705078125, 0.08355328369140624, 0.08311398315429687, 0.08351538848876953, 0.0833597412109375, 0.08303411102294922, 0.08349900817871093, 0.08534835052490235, 0.0855050277709961, 0.08562073516845703, 0.17103666687011718, 0.08302489471435547, 0.08521625518798828, 0.08401715087890625, 0.08225791931152343, 0.08271871948242188, 0.08310272216796875, 0.0847667236328125, 0.08585011291503906, 0.08474214172363281, 0.08309555053710938, 0.08301567840576173, 0.08329730987548828, 0.08300233459472656, 0.08324198150634765, 0.0859535369873047, 0.08576204681396485, 0.08367826843261719, 0.08261113739013672, 0.08418816375732421, 0.08275762939453125, 0.0831805419921875, 0.08316108703613281, 0.0830997085571289, 0.08339756774902343, 0.08295219421386718, 0.08297881317138672, 0.082946044921875, 0.08302285003662109, 0.08294809722900391, 0.08310886383056641, 0.08326348876953125, 0.08322252655029297, 0.0831805419921875, 0.08310886383056641, 0.08308633422851562, 0.08476467132568359, 0.08618803405761719, 0.08345702362060547, 0.08549068450927734, 0.0856258544921875, 0.08553478240966797, 0.08431609344482421, 0.08624332427978515, 0.08562483215332031, 0.08598834991455079, 0.08559001922607422, 0.08474009704589844, 0.08330035400390624, 0.08471858978271485, 0.08536576080322265, 0.08325836944580078, 0.08315392303466797, 0.08301465606689454, 0.0850913314819336, 0.08487117004394532, 0.08412262725830078, 0.08338022613525391, 0.08320515441894531, 0.083283935546875, 0.08316416168212891, 0.08440841674804687, 0.08565955352783203, 0.16939418029785155, 0.08440729522705079, 0.08702464294433594, 0.0858419189453125, 0.08364646148681641, 0.08301875305175781, 0.08324403381347656, 0.08340172576904296, 0.08715980529785156, 0.0837734375, 0.0832542724609375, 0.08338534545898438, 0.0833597412109375, 0.08333824157714843, 0.08325631713867188, 0.08421478271484376, 0.08312525177001953, 0.08310578918457032, 0.083198974609375, 0.08310681915283204, 0.08331059265136719, 0.08320716857910156, 0.08322560119628906, 0.08318669128417969, 0.08280166625976562, 0.08332185363769531, 0.08412876892089843, 0.08343244934082031, 0.08313958740234376, 0.08361984252929687, 0.08561869049072265, 0.08557670593261718, 0.0850544662475586, 0.08538317108154297, 0.0842209243774414, 0.08301465606689454, 0.08322150421142578, 0.08332083129882813, 0.08333618927001953, 0.08346828460693359, 0.08309760284423828, 0.08334745788574219, 0.08331571197509766, 0.08290509033203125, 0.08312012481689453, 0.08305458831787109, 0.08314060974121094, 0.08323891448974609, 0.08319385528564453, 0.08313139343261719, 0.08344576263427735, 0.08324198150634765, 0.08304537963867188, 0.08294092559814453, 0.08300543975830078, 0.0828927993774414, 0.08246886444091797, 0.08296550750732422, 0.08301055908203125, 0.0830218276977539, 0.08244223785400391, 0.08466329956054687, 0.08301773071289062, 0.16836819458007812, 0.08288864135742187, 0.08313241577148438, 0.08298905944824218, 0.08301055908203125, 0.08326041412353516, 0.08387481689453125, 0.08673689270019531, 0.08330652618408203, 0.08312828826904296, 0.08289997100830078, 0.08339148712158204, 0.0836495361328125, 0.08304025268554688, 0.08298188781738282, 0.08328294372558594, 0.08245862579345703, 0.08297267150878906, 0.08290099334716797, 0.082840576171875, 0.0828958740234375, 0.08289177703857421, 0.08293170928955078, 0.08315699005126953, 0.08276377868652343, 0.08302489471435547, 0.08323583984375, 0.08304537963867188, 0.08671949005126953, 0.08339250946044922, 0.08312422180175781, 0.08302796936035156, 0.08298188781738282, 0.08302899169921875, 0.08331366729736328, 0.08554393768310548, 0.08422911834716797, 0.08373862457275391, 0.08298086547851563, 0.0833280029296875, 0.0831488037109375, 0.08490290832519531, 0.08303308868408203, 0.08296345520019531, 0.08327065277099609, 0.08311910247802734, 0.08332390594482422, 0.08383897399902343, 0.08313549041748047, 0.08300032043457031, 0.083093505859375, 0.08306483459472656, 0.0828958740234375, 0.08329933166503906, 0.08313760375976563, 0.08340985870361328, 0.08315904235839844, 0.08326553344726563, 0.0868485107421875, 0.08454656219482422, 0.08522752380371093, 0.08517529296875, 0.08232243347167968]",tokens/s,11.68994270321447,,,True 8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,3764.559872,2282.225664,0.0,1627.389952,1558.84032,s,10,0.5292642555236816,0.052926425552368164,0.0026179030576980763,0.051947040557861326,0.05383301353454589,0.057238715171813954,0.05996327648162842,"[0.06064441680908203, 0.052558433532714846, 0.05158838272094726, 0.053076190948486326, 0.05195430374145508, 0.05161686325073242, 0.0525863037109375, 0.05171996688842773, 0.051939777374267575, 0.05157961654663086]",tokens/s,4836.9032544376205,kWh,6.215678173626072e-07,3.4047576992200954e-07,1.3812881905445604e-06,2.343331777829177e-06,tokens/kWh,109246160.7110343,MB,3768.479744,2282.225664,0.0,1627.389952,1603.517952,s,10,32.604803222656244,3.260480322265625,0.03325579430264878,3.2471553955078125,3.3117323242187497,3.3129788330078123,3.313976040039062,"[3.302928955078125, 3.314225341796875, 3.2516142578125, 3.311455322265625, 3.253620361328125, 3.2339873046875, 3.22545263671875, 3.2389208984375, 3.242696533203125, 3.229901611328125]",tokens/s,19.322306461958004,kWh,3.8606225664252255e-05,2.115701204683752e-05,6.402673635805206e-05,0.00012378997406914185,tokens/kWh,508926.51423298527,,s,629,33.01790412521366,0.052492693362819746,0.006296198022439321,0.05143046569824219,0.05343252487182617,0.054154866790771486,0.10345484130859375,"[0.05613158416748047, 0.05537484741210937, 0.05493657684326172, 0.05605376052856445, 0.05522227096557617, 0.05535027313232422, 0.0545904655456543, 0.05481574249267578, 0.054336544036865234, 0.05377532958984375, 0.051763198852539063, 0.05151641464233398, 0.05159628677368164, 0.05245859146118164, 0.05146716690063476, 0.051476478576660156, 0.0525926399230957, 0.05208883285522461, 0.052555774688720705, 0.05151846313476562, 0.0519628791809082, 0.05162496185302735, 0.05159628677368164, 0.05167103958129883, 0.05156249618530274, 0.051484672546386716, 0.05182668685913086, 0.05145910263061523, 0.05194851303100586, 0.05160038375854492, 0.05133107376098633, 0.0522608642578125, 0.0516126708984375, 0.05142118453979492, 0.0516434555053711, 0.0516187858581543, 0.05181539154052734, 0.05159833526611328, 0.05191987228393555, 0.05137510299682617, 0.051434497833251956, 0.05171916961669922, 0.05160140609741211, 0.05216563034057617, 0.05157580947875977, 0.05239603042602539, 0.05157580947875977, 0.05185331344604492, 0.05166387176513672, 0.051468288421630856, 0.051530750274658206, 0.05259881591796875, 0.05157475280761719, 0.05168332672119141, 0.052759552001953126, 0.05161180877685547, 0.05163504028320313, 0.05283225631713867, 0.053937152862548826, 0.05299507141113281, 0.05208575820922851, 0.05148979187011719, 0.10594815826416015, 0.052482048034667966, 0.05304217529296875, 0.05149798583984375, 0.05145087814331055, 0.05130547332763672, 0.05135564804077149, 0.0514600944519043, 0.05150207901000976, 0.05134131240844726, 0.05147238540649414, 0.05144166564941406, 0.051525630950927735, 0.051830783843994144, 0.05231718444824219, 0.05178265762329102, 0.05135564804077149, 0.05319680023193359, 0.05314252853393555, 0.05313024139404297, 0.05280767822265625, 0.05328486251831055, 0.05319680023193359, 0.05328076934814453, 0.05329510498046875, 0.053157886505126956, 0.05390137481689453, 0.05398214340209961, 0.05312716674804688, 0.05326028823852539, 0.05348147201538086, 0.05330944061279297, 0.053594112396240234, 0.053433345794677733, 0.05338828659057617, 0.05332889556884766, 0.05303705596923828, 0.05334220886230469, 0.053154815673828126, 0.053817344665527345, 0.05297663879394531, 0.05365350341796875, 0.05345382308959961, 0.053989376068115234, 0.05367910385131836, 0.05296537780761719, 0.05322956848144531, 0.05337702560424805, 0.053321727752685545, 0.054010879516601565, 0.05175603103637695, 0.051917823791503906, 0.05153484725952148, 0.051138561248779295, 0.05137715148925781, 0.052125694274902344, 0.05241856002807617, 0.05342924880981445, 0.051609600067138675, 0.051323902130126955, 0.0512624626159668, 0.051248126983642575, 0.05127372741699219, 0.10346086120605469, 0.051504127502441405, 0.051514366149902346, 0.051506175994873046, 0.05153798294067383, 0.05151532745361328, 0.051607551574707033, 0.05157888031005859, 0.05159423828125, 0.05141708755493164, 0.05144985580444336, 0.05147340774536133, 0.051515392303466793, 0.0517498893737793, 0.051607551574707033, 0.05166284942626953, 0.05153792190551758, 0.051253246307373046, 0.05149388885498047, 0.051643390655517575, 0.051397632598876954, 0.052574207305908206, 0.052877311706542966, 0.05160857772827149, 0.051422206878662106, 0.05127987289428711, 0.05168025588989258, 0.051514366149902346, 0.05145702362060547, 0.05135059356689453, 0.05161568069458008, 0.051955711364746096, 0.05164236831665039, 0.05155635070800781, 0.05123788833618164, 0.05145395278930664, 0.05185945510864258, 0.05144063949584961, 0.05143558502197266, 0.051633087158203125, 0.05228646469116211, 0.05158195114135742, 0.051350528717041016, 0.05157580947875977, 0.051452926635742184, 0.05173350524902344, 0.05144678497314453, 0.051514366149902346, 0.05153996658325195, 0.05178275299072266, 0.0514804801940918, 0.05143040084838867, 0.05167513656616211, 0.05138022232055664, 0.05166387176513672, 0.05161164855957031, 0.05148876953125, 0.051361793518066405, 0.052768768310546874, 0.05111603164672852, 0.05144063949584961, 0.05148262405395508, 0.05243596649169922, 0.10517196655273438, 0.05126348876953125, 0.05126144027709961, 0.051135486602783206, 0.05122662353515625, 0.05139868927001953, 0.05129724884033203, 0.05227417755126953, 0.052501502990722655, 0.0514068489074707, 0.0513259506225586, 0.05139865493774414, 0.051149822235107424, 0.051181568145751956, 0.05121638488769531, 0.0513331184387207, 0.05125836944580078, 0.05129011154174805, 0.051280895233154294, 0.05144985580444336, 0.051119102478027346, 0.05124915313720703, 0.05403238296508789, 0.05393203353881836, 0.05403647994995117, 0.054148128509521484, 0.054257633209228516, 0.05415935897827148, 0.054365184783935545, 0.054079486846923826, 0.05390028762817383, 0.05394944000244141, 0.05112422561645508, 0.051101760864257814, 0.05111084747314453, 0.05119180679321289, 0.051149822235107424, 0.051187713623046874, 0.051176448822021485, 0.051267742156982425, 0.0511445426940918, 0.051192832946777345, 0.05107712173461914, 0.051146785736083986, 0.051292129516601566, 0.05168844985961914, 0.054280193328857425, 0.05429862213134766, 0.053955646514892576, 0.05405587387084961, 0.054022144317626954, 0.05403443145751953, 0.05412761688232422, 0.054024192810058595, 0.05413273620605469, 0.054441982269287106, 0.0542371826171875, 0.054168575286865236, 0.05461708831787109, 0.05568000030517578, 0.05409996795654297, 0.05422182464599609, 0.05424025726318359, 0.10375475311279297, 0.051418113708496097, 0.05132287979125977, 0.05136383819580078, 0.051329025268554686, 0.05131980895996094, 0.05217484664916992, 0.05141708755493164, 0.051253246307373046, 0.05110271835327149, 0.05156147384643555, 0.051192832946777345, 0.051714046478271485, 0.0513966064453125, 0.05136076736450195, 0.05130649566650391, 0.05130649566650391, 0.05136588668823242, 0.05136383819580078, 0.05101772689819336, 0.05118361663818359, 0.051264511108398435, 0.05131980895996094, 0.05133414459228516, 0.05132799911499023, 0.051367935180664064, 0.05120307159423828, 0.0539504623413086, 0.05328179168701172, 0.05343231964111328, 0.0526561279296875, 0.05136383819580078, 0.05182156753540039, 0.05296844863891602, 0.05135871887207031, 0.052523006439208986, 0.051743743896484375, 0.051560447692871096, 0.05120512008666992, 0.052741119384765625, 0.05170278549194336, 0.052519935607910156, 0.05112934494018555, 0.05369241714477539, 0.05340774536132813, 0.051454975128173826, 0.051158016204833984, 0.05123993682861328, 0.05113139343261719, 0.05192294311523438, 0.05221171188354492, 0.05174476623535156, 0.05126758575439453, 0.05100339126586914, 0.051120128631591794, 0.05127987289428711, 0.05112934494018555, 0.051165184020996096, 0.05128812789916992, 0.05119071960449219, 0.05113555145263672, 0.05114156723022461, 0.050953216552734375, 0.10343936157226563, 0.05124710464477539, 0.05126860809326172, 0.05121331024169922, 0.05136076736450195, 0.05122969436645508, 0.05108531188964844, 0.051676158905029294, 0.05126553726196289, 0.05129318237304688, 0.05135257720947266, 0.05125328063964844, 0.051724254608154295, 0.051585025787353515, 0.051958782196044925, 0.05148262405395508, 0.05132185745239258, 0.05275852966308594, 0.052911136627197264, 0.051404769897460935, 0.05132185745239258, 0.05132185745239258, 0.0512911376953125, 0.051487743377685545, 0.05132287979125977, 0.051095550537109374, 0.0515676155090332, 0.05120512008666992, 0.05137823867797851, 0.05132281494140625, 0.052326400756835936, 0.05145804977416992, 0.0513004150390625, 0.05145596694946289, 0.05110678482055664, 0.05122355270385742, 0.05137612915039062, 0.05143046569824219, 0.05147846221923828, 0.05129830551147461, 0.051198974609375, 0.05133004760742187, 0.05161574554443359, 0.05048012924194336, 0.05059993743896484, 0.051143680572509766, 0.05138025665283203, 0.05134435272216797, 0.05135974502563476, 0.05123993682861328, 0.05142022323608399, 0.050812862396240235, 0.05127167892456055, 0.05130649566650391, 0.05130752182006836, 0.05125120162963867, 0.05123072052001953, 0.05116108703613281, 0.051318782806396485, 0.051310592651367185, 0.05021798324584961, 0.050383872985839843, 0.05035417556762695, 0.10463948822021485, 0.05177139282226562, 0.051364864349365234, 0.05175910568237305, 0.05104742431640625, 0.051079166412353515, 0.051092479705810545, 0.05106175994873047, 0.05140991973876953, 0.051366912841796876, 0.05130035018920898, 0.05129011154174805, 0.05122457504272461, 0.05134131240844726, 0.05168025588989258, 0.05074124908447265, 0.05059993743896484, 0.050557952880859375, 0.051212287902832034, 0.05132089614868164, 0.05133100891113281, 0.05169561767578125, 0.051560447692871096, 0.05127782440185547, 0.05040643310546875, 0.05045142364501953, 0.050836479187011716, 0.0511016960144043, 0.05123891067504883, 0.051130367279052735, 0.051151870727539066, 0.051694591522216796, 0.050661376953125, 0.050487297058105465, 0.05048934555053711, 0.05045043182373047, 0.05095116806030273, 0.05120409774780273, 0.05122150421142578, 0.05139558410644531, 0.051722240447998044, 0.05042489624023438, 0.05042784118652344, 0.05044019317626953, 0.05104742431640625, 0.05130342483520508, 0.05157990264892578, 0.051335166931152344, 0.05172326278686523, 0.05088665771484375, 0.050506752014160154, 0.05268070220947266, 0.05133824157714844, 0.051266559600830076, 0.05135564804077149, 0.05142937469482422, 0.05126860809326172, 0.051781726837158204, 0.05061110305786133, 0.05089791870117188, 0.051367935180664064, 0.052203521728515626, 0.05187689590454102, 0.10347212982177735, 0.05145600128173828, 0.05122659301757813, 0.05121843338012695, 0.05136076736450195, 0.050506752014160154, 0.05052620697021484, 0.050574337005615234, 0.05122252655029297, 0.051318782806396485, 0.051146751403808595, 0.05124505615234375, 0.051536895751953124, 0.05079142379760742, 0.05071974563598633, 0.050912254333496096, 0.05077503967285156, 0.05123072052001953, 0.051171329498291014, 0.05125734329223633, 0.051266559600830076, 0.05180108642578125, 0.05043199920654297, 0.05062451171875, 0.050500606536865236, 0.05110784149169922, 0.05132185745239258, 0.05118566513061523, 0.05122662353515625, 0.05127987289428711, 0.0528988151550293, 0.05310976028442383, 0.05298175811767578, 0.05278105545043945, 0.05121535873413086, 0.051607551574707033, 0.051264511108398435, 0.05147238540649414, 0.051591167449951174, 0.05035417556762695, 0.050697216033935545, 0.05126144027709961, 0.0516864013671875, 0.05128806304931641, 0.05132799911499023, 0.05126860809326172, 0.05420646286010742, 0.05229875183105469, 0.05127884674072265, 0.050710529327392576, 0.05124505615234375, 0.05130854415893555, 0.05228441619873047, 0.05409791946411133, 0.05227315139770508, 0.051699710845947267, 0.050401344299316406, 0.0507022705078125, 0.05117030334472656, 0.05121023941040039, 0.05147238540649414, 0.05173452758789063, 0.051353599548339846, 0.10390937805175782, 0.051585025787353515, 0.05145087814331055, 0.05133926391601563, 0.05153177642822265, 0.051192832946777345, 0.05131980895996094, 0.05130137634277344, 0.051227649688720706, 0.051576831817626956, 0.05170995330810547, 0.050724864959716794, 0.050697216033935545, 0.050786304473876956, 0.05134745788574219, 0.05147545623779297, 0.051490814208984374, 0.051468288421630856, 0.05175603103637695, 0.05168844985961914, 0.050678783416748044, 0.051356670379638675, 0.05161779022216797, 0.05158092880249023, 0.05135871887207031, 0.05122355270385742, 0.05164134216308594, 0.05222092819213867, 0.05083955383300781, 0.05141401672363281, 0.05432320022583008, 0.05159423828125, 0.05125734329223633, 0.051282943725585936, 0.05126041412353516, 0.05148979187011719, 0.05164646530151367, 0.05055180740356445, 0.051154945373535154, 0.05143040084838867, 0.0513546257019043, 0.05149798583984375, 0.05306982421875, 0.05144166564941406, 0.051661823272705076, 0.050928672790527346, 0.05167919921875, 0.05124607849121094, 0.0517498893737793, 0.05123072052001953, 0.051294208526611325, 0.051225601196289064, 0.051394561767578124, 0.05137919998168945, 0.05098905563354492, 0.05156556701660156, 0.05203353500366211, 0.051568641662597656, 0.05131468963623047, 0.05125120162963867, 0.05154099273681641, 0.05268479919433594, 0.05105152130126953, 0.10318441772460937, 0.05048118209838867, 0.05074425506591797, 0.051231742858886715, 0.05163622283935547, 0.05126553726196289, 0.05163520050048828, 0.05135871887207031, 0.05174272155761719, 0.05089177703857422, 0.05143142318725586, 0.05176115036010742, 0.05131673431396484, 0.051264511108398435, 0.05128499221801758, 0.05135257720947266, 0.051476478576660156, 0.05102284622192383, 0.05038694381713867, 0.05041254425048828, 0.05074943923950195, 0.05119385528564453, 0.05118054580688477, 0.05168844985961914, 0.05116825485229492, 0.051343360900878904, 0.050495487213134765, 0.05041561508178711, 0.050423809051513675, 0.05101055908203125, 0.051568641662597656, 0.05301862335205078, 0.05307699203491211, 0.0517283821105957, 0.05130444717407227, 0.05036236953735351, 0.050493438720703124, 0.050569217681884764, 0.05095219039916992, 0.05138227081298828, 0.05127679824829102, 0.05142732620239258, 0.05142425537109375, 0.051231742858886715, 0.05054361724853516, 0.05068185424804687, 0.051176448822021485, 0.051416065216064455, 0.05121638488769531, 0.05126553726196289, 0.05188406372070312, 0.05152463912963867, 0.050936767578125, 0.051806209564208984, 0.05132185745239258, 0.051329025268554686, 0.05110784149169922, 0.05123276901245117, 0.05108531188964844, 0.05152153778076172, 0.05257113647460938, 0.052294654846191405, 0.051197952270507815]",tokens/s,19.05027035073597,,,True -8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13719,7 +13719,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4252.811264,10769.399808,0.0,10114.564096,9989.950464,s,10,1.169741928100586,0.11697419281005861,0.0028193884080607253,0.11599164962768554,0.11867264633178712,0.12153163566589355,0.12381882713317871,"[0.124390625, 0.11801824188232422, 0.115059326171875, 0.1146569595336914, 0.11692076873779297, 0.114829345703125, 0.11506253051757813, 0.11803731536865235, 0.11492160034179688, 0.11784521484375]",tokens/s,2188.5169185624554,kWh,1.3863510594195483e-06,7.594575223117508e-07,3.760406755092725e-06,5.906215336824023e-06,tokens/kWh,43344169.72640555,MB,4252.811264,10769.399808,0.0,10114.564096,10046.339584,s,10,70.53508935546876,7.053508935546875,0.011930270007887878,7.0534931640625,7.061798779296875,7.070816137695312,7.078030024414063,"[7.04911083984375, 7.05199560546875, 7.059794921875, 7.05392041015625, 7.03281396484375, 7.07983349609375, 7.039171875, 7.05306591796875, 7.0576513671875, 7.05773095703125]",tokens/s,8.931724702651909,kWh,8.481260620500382e-05,4.648280134391486e-05,0.0001672769522969082,0.0002985723598458269,tokens/kWh,211004.12654584355,,s,629,71.46780875396729,0.11362131757387486,0.014146979172250683,0.11153202819824219,0.11394601135253905,0.11509207153320312,0.22721429077148436,"[0.11163136291503906, 0.11125657653808593, 0.11266867065429688, 0.11582771301269532, 0.11417497253417969, 0.11043532562255859, 0.10988953399658204, 0.10950144195556641, 0.11115007781982422, 0.11157196807861328, 0.11138969421386719, 0.11159661102294922, 0.11170297241210937, 0.1115863037109375, 0.1139261474609375, 0.11125965118408203, 0.1111562271118164, 0.11148390197753906, 0.1112279052734375, 0.11114905548095703, 0.11125657653808593, 0.111678466796875, 0.1124136962890625, 0.11196518707275391, 0.11283763122558593, 0.11112448120117188, 0.11147673797607421, 0.11142451477050781, 0.11116646575927734, 0.11118386840820313, 0.11162528228759766, 0.11166099548339843, 0.11230207824707031, 0.11133952331542969, 0.11121356964111329, 0.11130982208251954, 0.11161395263671875, 0.1114071044921875, 0.11034111785888671, 0.11141529846191406, 0.11126067352294922, 0.1111541748046875, 0.11003596496582031, 0.11004927825927735, 0.110740478515625, 0.11129862213134765, 0.11163231658935546, 0.1112125473022461, 0.11137843322753906, 0.11293081665039062, 0.11521126556396484, 0.11229695892333984, 0.11184025573730469, 0.11143679809570313, 0.111388671875, 0.11119206237792968, 0.11444735717773437, 0.11509452819824219, 0.11429478454589843, 0.11328717041015625, 0.11398348999023437, 0.1145384979248047, 0.23350682067871092, 0.11467775726318359, 0.11260723114013672, 0.11255398559570312, 0.11328102111816406, 0.11172659301757812, 0.11148086547851563, 0.1108469467163086, 0.11113062286376953, 0.11125965118408203, 0.11147878265380859, 0.11138764953613281, 0.11164466857910156, 0.11487334442138672, 0.11126783752441406, 0.11139276885986328, 0.11132927703857422, 0.11122380828857421, 0.11145011138916015, 0.11114803314208985, 0.11104051208496094, 0.11187097930908203, 0.11187200164794922, 0.1101127700805664, 0.11159244537353516, 0.11144601440429687, 0.11143679809570313, 0.11146649932861329, 0.11119411468505859, 0.1155041275024414, 0.11162214660644532, 0.11146240234375, 0.11119513702392578, 0.11085721588134766, 0.11109990692138672, 0.11148697662353516, 0.11151769256591797, 0.11141836547851562, 0.11126271820068359, 0.11158732604980469, 0.11134770965576171, 0.1113917465209961, 0.11165900421142579, 0.11152998352050782, 0.11142144012451172, 0.11150950622558593, 0.11223551940917968, 0.11329331207275391, 0.11110502624511719, 0.11184947204589844, 0.1114419174194336, 0.11255296325683593, 0.113691650390625, 0.11337318420410156, 0.11179724884033203, 0.11493888092041016, 0.11153510284423829, 0.11192524719238281, 0.11143782043457032, 0.11161702728271485, 0.11310694122314453, 0.11165900421142579, 0.11236255645751952, 0.22743341064453124, 0.11400498962402343, 0.11199183654785157, 0.11182895660400391, 0.11122278594970703, 0.1135257568359375, 0.11508838653564453, 0.1114286117553711, 0.11138764953613281, 0.11050086212158203, 0.11081932830810547, 0.11403673553466796, 0.11240140533447265, 0.11359846496582031, 0.11460300445556641, 0.11090534210205077, 0.11149209594726563, 0.11310899353027344, 0.11220070648193359, 0.11176038360595703, 0.11098521423339844, 0.11132006072998046, 0.11066675567626953, 0.11155353546142578, 0.11115724945068359, 0.11148185729980468, 0.11140096282958985, 0.11345101165771485, 0.11437055969238281, 0.1115514907836914, 0.11137535858154297, 0.11146752166748047, 0.11149107360839844, 0.11155763244628907, 0.11158732604980469, 0.11150643157958984, 0.11164876556396484, 0.11234508514404297, 0.11171123504638672, 0.11263590240478516, 0.11160371398925781, 0.11153817749023437, 0.11617075347900391, 0.11177881622314453, 0.11168358612060547, 0.11156172943115235, 0.11166617584228515, 0.11136102294921875, 0.11173887634277344, 0.11356364440917968, 0.11313664245605469, 0.11155763244628907, 0.11163954925537109, 0.11147878265380859, 0.11348480224609375, 0.11219558715820313, 0.11139584350585938, 0.11150745391845703, 0.11136819458007813, 0.11121971130371094, 0.11152588653564453, 0.11139788818359375, 0.1117501449584961, 0.2267176971435547, 0.11183001708984375, 0.11144499206542968, 0.11041075134277344, 0.1134991683959961, 0.111078369140625, 0.1115893783569336, 0.11136614227294922, 0.1146060791015625, 0.11244748687744141, 0.11343360137939452, 0.11103846740722656, 0.11426918029785156, 0.11156684875488282, 0.11133849334716797, 0.11143679809570313, 0.11147161865234376, 0.11134976196289062, 0.11197030639648438, 0.11189247894287109, 0.11101900482177735, 0.11159142303466797, 0.1113169937133789, 0.11158834838867188, 0.11150131225585938, 0.11172457885742187, 0.11087356567382813, 0.11115724945068359, 0.11131295776367188, 0.11138758087158203, 0.11138355255126953, 0.11171942138671875, 0.11138873291015625, 0.1115770263671875, 0.11158834838867188, 0.11144300842285157, 0.1111428451538086, 0.11547545623779297, 0.11177677154541016, 0.11515699005126953, 0.11158118438720703, 0.11141017913818359, 0.11154739379882812, 0.11154841613769531, 0.11136204528808594, 0.11524403381347656, 0.1115525131225586, 0.11258367919921874, 0.11138355255126953, 0.1116385269165039, 0.11127808380126954, 0.11113062286376953, 0.11560147094726562, 0.11459577941894532, 0.11151974487304687, 0.11179622650146484, 0.11128832244873046, 0.11148697662353516, 0.11077017974853516, 0.11118080139160157, 0.11110912322998047, 0.11147366333007812, 0.11467263793945312, 0.22850253295898437, 0.11147264099121093, 0.11148287963867187, 0.11142041778564453, 0.11153510284423829, 0.11151564788818359, 0.11083980560302735, 0.11140914916992188, 0.11137638092041016, 0.111531005859375, 0.1113733139038086, 0.11222118377685547, 0.11133849334716797, 0.11069849395751953, 0.1110149154663086, 0.11413401794433593, 0.11116544342041015, 0.1114746856689453, 0.11063910675048828, 0.11185049438476563, 0.11096678161621094, 0.11122176361083984, 0.11166413116455078, 0.1114746856689453, 0.11132927703857422, 0.1111562271118164, 0.1114439697265625, 0.11113577270507813, 0.11146749114990234, 0.11127705383300782, 0.11007801818847657, 0.11124114990234375, 0.1111900177001953, 0.11165593719482422, 0.11190271759033203, 0.11211980438232422, 0.11169382476806641, 0.11454668426513671, 0.11180134582519531, 0.1118760986328125, 0.11161804962158203, 0.1118545913696289, 0.11185561370849609, 0.11168768310546875, 0.11344390106201171, 0.11193132781982422, 0.11116544342041015, 0.11160883331298828, 0.11130470275878906, 0.11145932769775391, 0.11212287902832031, 0.11106098937988282, 0.11146956634521485, 0.11160781097412109, 0.11160781097412109, 0.11235533142089844, 0.1107558364868164, 0.11150335693359376, 0.11144089508056641, 0.1116211166381836, 0.11132109069824218, 0.11141939544677734, 0.11519795227050782, 0.22721945190429688, 0.1157918701171875, 0.1119815673828125, 0.11125759887695312, 0.11179110717773437, 0.11158425903320313, 0.11183103942871093, 0.11509862518310547, 0.11521842956542969, 0.11402035522460938, 0.11353292846679687, 0.11532697296142579, 0.11502591705322265, 0.11505458831787109, 0.1169070053100586, 0.11631410980224609, 0.11330252838134766, 0.1113169937133789, 0.11262361907958984, 0.11127705383300782, 0.11295645141601562, 0.11117462158203124, 0.11514575958251953, 0.11158627319335937, 0.11346739196777343, 0.11146444702148438, 0.11149209594726563, 0.11155865478515625, 0.11167027282714843, 0.11089920043945313, 0.11089510345458985, 0.11158425903320313, 0.11173887634277344, 0.11192626953125, 0.11137741088867187, 0.11117670440673828, 0.1113702392578125, 0.11202969360351563, 0.1119078369140625, 0.11150028991699219, 0.11306803131103516, 0.11231436920166016, 0.11151769256591797, 0.1116231689453125, 0.11165388488769531, 0.11136921691894532, 0.11139993286132813, 0.1113016357421875, 0.11156377410888672, 0.11141120147705078, 0.11132621002197265, 0.11212902069091797, 0.11247923278808594, 0.11166310119628907, 0.1117992935180664, 0.11162931060791016, 0.11161599731445312, 0.111351806640625, 0.11184441375732422, 0.11185145568847656, 0.11174809265136719, 0.11186176300048828, 0.11284992218017578, 0.22720101928710937, 0.11156070709228516, 0.11150438690185546, 0.11143373107910157, 0.11157913970947266, 0.11216998291015624, 0.11128729248046874, 0.11312844848632812, 0.11200614166259766, 0.11152896118164063, 0.11150540924072265, 0.11122892761230468, 0.11133952331542969, 0.11127193450927735, 0.1112647705078125, 0.11145625305175781, 0.11106816101074218, 0.11088179016113281, 0.1108449249267578, 0.11109478759765624, 0.11117362976074219, 0.11130879974365235, 0.11137126159667969, 0.11362201690673829, 0.11233586883544922, 0.11096268463134766, 0.11514982604980468, 0.1118197784423828, 0.1133834228515625, 0.11147264099121093, 0.1112995834350586, 0.1112965087890625, 0.11320937347412109, 0.11162825775146484, 0.11116441345214843, 0.11117056274414062, 0.11063603210449219, 0.11095654296875, 0.11073638153076172, 0.11460915374755859, 0.11170816040039062, 0.11131187438964844, 0.1120030746459961, 0.11132927703857422, 0.11140914916992188, 0.11164057922363281, 0.11139481353759766, 0.11135897827148437, 0.11144499206542968, 0.11492147064208984, 0.11197750091552734, 0.11159248352050781, 0.11067488098144532, 0.11129344177246094, 0.11203379058837891, 0.11154329681396484, 0.11149209594726563, 0.11241881561279297, 0.11272601318359375, 0.11172665405273438, 0.11140806579589843, 0.11118390655517578, 0.11105072021484375, 0.24237260437011718, 0.1119969253540039, 0.11517849731445312, 0.11248947143554687, 0.11148799896240234, 0.11207679748535156, 0.11062169647216796, 0.1116354522705078, 0.11175116729736329, 0.11309977722167969, 0.11192934417724609, 0.11211775970458984, 0.11250176239013672, 0.11150028991699219, 0.11219455718994141, 0.11314482879638672, 0.11243417358398437, 0.11195085144042968, 0.11206348419189453, 0.11343462371826171, 0.11198566436767578, 0.11114189147949219, 0.11141426849365234, 0.11133747100830078, 0.1113333740234375, 0.11143885040283204, 0.11134054565429688, 0.11132927703857422, 0.11140608215332032, 0.11128323364257812, 0.11124425506591797, 0.1114081268310547, 0.11106201934814453, 0.11103743743896484, 0.11122688293457031, 0.11113164520263671, 0.11153305816650391, 0.11168256378173828, 0.11110912322998047, 0.11135692596435547, 0.11143577575683594, 0.11124940490722657, 0.11208191680908203, 0.11177983856201172, 0.11069951629638672, 0.11118592071533204, 0.11136614227294922, 0.11173478698730469, 0.1124659194946289, 0.1139415054321289, 0.11130777740478516, 0.11220070648193359, 0.11446476745605469, 0.1112248306274414, 0.11138253021240234, 0.11161702728271485, 0.11135794830322265, 0.1112616958618164, 0.11089817810058594, 0.11102719879150391, 0.11114189147949219, 0.11130982208251954, 0.11121971130371094, 0.22831309509277345, 0.1122877426147461, 0.11148697662353516, 0.11074457550048829, 0.1117501449584961, 0.11198566436767578, 0.11136307525634766, 0.11136511993408203, 0.11225907135009766, 0.1111009292602539, 0.11131187438964844, 0.11180850982666016, 0.11202867126464844, 0.11240550231933594, 0.11235533142089844, 0.11108966064453125, 0.11133958435058594, 0.11168863677978516, 0.11141324615478515, 0.11143987274169923, 0.11137843322753906, 0.11138355255126953, 0.11132723236083984, 0.11143993377685547, 0.11138553619384765, 0.11149311828613281, 0.11166310119628907, 0.11473919677734375, 0.11161910247802734, 0.11157705688476563, 0.11136511993408203, 0.11109375762939454, 0.1103421401977539, 0.11104972839355469, 0.11120127868652344, 0.11195085144042968, 0.11222528076171875, 0.11130879974365235, 0.11083468627929688, 0.111425537109375, 0.11132518768310547, 0.11183001708984375, 0.11359744262695312, 0.11132211303710937, 0.11134054565429688, 0.1136711654663086, 0.1134694366455078, 0.11500339508056641, 0.11352166748046875, 0.11387391662597657, 0.11296153259277343, 0.11332915496826172, 0.1151488037109375, 0.11401318359375, 0.11297996520996094, 0.11247821044921875, 0.11190374755859375, 0.11123712158203125, 0.11146546936035157, 0.1116211166381836, 0.11118796539306641, 0.11190886688232422, 0.11300761413574219, 0.23373414611816407, 0.11295948791503906, 0.11331385803222656, 0.11327584075927734, 0.11353804779052734, 0.11412786865234376, 0.11396403503417969, 0.11282125091552735, 0.11217100524902343, 0.11378790283203125, 0.1168179168701172, 0.11518669128417969, 0.11457846069335938, 0.1151764144897461, 0.11155661010742188, 0.11139689636230468, 0.11085308837890626, 0.11119821166992187, 0.11161804962158203, 0.11173887634277344, 0.11172147369384766, 0.11142963409423828, 0.11166207885742188, 0.11141120147705078, 0.11153202819824219, 0.1130557403564453, 0.11204198455810546, 0.11153510284423829, 0.11187200164794922, 0.11033497619628906, 0.11098521423339844, 0.1102735366821289, 0.11123609924316406, 0.11167641448974609, 0.11165286254882813, 0.11119206237792968, 0.11164262390136719, 0.11137945556640624, 0.11077836608886718, 0.11162214660644532, 0.11144499206542968, 0.11129344177246094, 0.11182592010498046, 0.11139379119873047, 0.1113169937133789, 0.1113733139038086, 0.11132825469970703, 0.11151360321044922, 0.11164672088623047, 0.1113855972290039, 0.1111910400390625, 0.1113631362915039, 0.11366802978515625, 0.111963134765625, 0.11058175659179688, 0.11269324493408203, 0.11076403045654297, 0.11029094696044922, 0.11109990692138672, 0.11058688354492187, 0.1113855972290039, 0.11144908905029297, 0.1112811508178711]",tokens/s,8.801165321373915,,,True 8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1231.405056,8273.788928,0.0,7618.953216,7296.841728,s,10,0.7893702774047852,0.0789370277404785,0.0020570642836913,0.07802062225341796,0.0798980842590332,0.08239622764587402,0.08439474235534668,"[0.08489437103271484, 0.07804547119140624, 0.07791593933105469, 0.07806009674072266, 0.07792639923095702, 0.0793429412841797, 0.07931648254394531, 0.07799577331542969, 0.07792915344238281, 0.07794364929199218]",tokens/s,3243.091453121999,kWh,9.306917170735691e-07,5.098247244694809e-07,2.802530675968085e-06,4.243047117511135e-06,tokens/kWh,60333998.87158529,MB,1231.405056,8273.788928,0.0,7618.953216,7296.844288,s,10,46.61180224609375,4.661180224609375,0.0252475975361493,4.655324462890626,4.693600781250001,4.69970126953125,4.70458166015625,"[4.68605029296875, 4.6922451171875, 4.66919482421875, 4.65787646484375, 4.62487451171875, 4.6527724609375, 4.64473974609375, 4.64592626953125, 4.7058017578125, 4.63232080078125]",tokens/s,13.515890174634825,kWh,5.556471532450261e-05,3.045302852353244e-05,0.00011160875529743558,0.0001976264991454706,tokens/kWh,318783.16051951324,,s,629,47.24290555572502,0.0751079579582275,0.009359550209562799,0.07319961547851563,0.07603097534179687,0.07639427032470703,0.15113872009277343,"[0.0755445785522461, 0.07345868682861328, 0.07376588439941406, 0.07603097534179687, 0.07583026885986328, 0.0746956787109375, 0.07736319732666015, 0.07364198303222656, 0.07356006622314454, 0.073238525390625, 0.073312255859375, 0.07319142150878906, 0.07639244842529297, 0.07653580474853515, 0.07678668975830077, 0.07592038726806641, 0.07607193756103516, 0.07599411010742188, 0.07615078735351563, 0.07370649719238281, 0.07326617431640625, 0.073169921875, 0.07311154937744141, 0.07311974334716796, 0.07290060424804687, 0.07549951934814453, 0.07590399932861328, 0.07278797149658203, 0.07450418853759766, 0.0742635498046875, 0.07593062591552735, 0.07296409606933593, 0.07310233306884766, 0.07301222229003906, 0.07260979461669922, 0.07291494750976563, 0.07320063781738281, 0.07301222229003906, 0.07431884765625, 0.07570636749267579, 0.07611698913574219, 0.07526604461669922, 0.07310848236083985, 0.07400141143798829, 0.0761712646484375, 0.07585689544677734, 0.07602278137207032, 0.0761343994140625, 0.07583744049072266, 0.07554354858398438, 0.07591939544677734, 0.07557424163818359, 0.07312281799316406, 0.07294361877441406, 0.07309516906738281, 0.07281459045410156, 0.07327436828613282, 0.07302143859863282, 0.07316172790527344, 0.0729692153930664, 0.07294566345214844, 0.07296409606933593, 0.1515161590576172, 0.07281254577636719, 0.07321804809570312, 0.07467427062988281, 0.07300495910644532, 0.07353241729736328, 0.07590399932861328, 0.0759152603149414, 0.07549235534667968, 0.07323750305175782, 0.07309721374511718, 0.07339212799072266, 0.07320575714111328, 0.07323340606689453, 0.07302556610107422, 0.0727838363647461, 0.07294361877441406, 0.07576985931396485, 0.07581491088867187, 0.07607603454589844, 0.07598899078369141, 0.07581388854980468, 0.07628697967529297, 0.07595315551757813, 0.07584358215332031, 0.07592960357666016, 0.07578419494628906, 0.07469261169433594, 0.07275827026367188, 0.07436185455322265, 0.07631462097167968, 0.07615283203125, 0.07599718475341796, 0.07607091522216797, 0.07579033660888672, 0.07603097534179687, 0.07581593322753906, 0.07605145263671875, 0.07566233825683594, 0.07297638702392578, 0.07290265655517578, 0.07326822662353516, 0.07305010986328125, 0.07330303955078125, 0.07601152038574219, 0.07319961547851563, 0.0748779525756836, 0.07560908508300782, 0.07649897766113281, 0.07614157104492188, 0.07600022125244141, 0.07726182556152343, 0.07230361938476562, 0.07330815887451173, 0.073059326171875, 0.0732774429321289, 0.07311360168457032, 0.07318220520019532, 0.07252377319335937, 0.07320883178710938, 0.07310131072998047, 0.0730982437133789, 0.07309619140625, 0.15160115051269532, 0.07576473236083985, 0.07348735809326172, 0.07327232360839844, 0.07298662567138672, 0.07572480010986328, 0.07300096130371093, 0.07346790313720702, 0.07330508422851563, 0.07340646362304687, 0.07520256042480469, 0.07330918121337891, 0.07330099487304688, 0.07271833801269531, 0.07318732452392578, 0.07316377258300781, 0.07323238372802734, 0.0732968978881836, 0.073059326171875, 0.07313100433349609, 0.07644057464599609, 0.07676927947998047, 0.07342797088623047, 0.07323033905029297, 0.07322112274169922, 0.07545343780517579, 0.07647539520263671, 0.07323033905029297, 0.07547801971435547, 0.07333273315429688, 0.07301837158203126, 0.07287398529052734, 0.07455744171142578, 0.07422054290771485, 0.07331942749023437, 0.07609241485595702, 0.07585382080078125, 0.07593984222412109, 0.07591117095947265, 0.07318220520019532, 0.07297126770019531, 0.07443865966796875, 0.07292313385009766, 0.0723978271484375, 0.07261491394042968, 0.07323442840576172, 0.07308393859863281, 0.0763954849243164, 0.07633817291259766, 0.07554560089111328, 0.07279513549804688, 0.07568077087402343, 0.0757739486694336, 0.07595417785644532, 0.07307981109619141, 0.07631769561767578, 0.0729692153930664, 0.07308697509765626, 0.07287296295166015, 0.07312284851074219, 0.07442838287353516, 0.07387852478027344, 0.07557939147949219, 0.15159706115722657, 0.07404236602783203, 0.07334809875488281, 0.07327436828613282, 0.07293644714355468, 0.07324569702148437, 0.07321600341796874, 0.07318220520019532, 0.07301734161376953, 0.07334706878662109, 0.07293952178955078, 0.07316582489013672, 0.07306854248046875, 0.07289548492431641, 0.07312691497802734, 0.07326207733154297, 0.07356211090087891, 0.07309107208251953, 0.0753407974243164, 0.07334400177001953, 0.07299993896484375, 0.07289142608642578, 0.07531724548339844, 0.07303778839111329, 0.07307469177246094, 0.07579647827148438, 0.07300198364257812, 0.07305010986328125, 0.07351500701904297, 0.07552716827392578, 0.07574323272705077, 0.07593369293212891, 0.07307571411132813, 0.07308287811279297, 0.07311666870117188, 0.07280435180664062, 0.07482982635498046, 0.07594802856445312, 0.0759183349609375, 0.07585177612304687, 0.07571558380126953, 0.0755384292602539, 0.07309414672851562, 0.07314022064208985, 0.07296511840820312, 0.073169921875, 0.07296729278564452, 0.07288511657714844, 0.07288015747070313, 0.07307568359375, 0.07476838684082031, 0.07379148864746093, 0.07558963012695312, 0.07576371002197266, 0.07378431701660157, 0.07639142608642578, 0.07585075378417969, 0.07590297698974609, 0.07675596618652344, 0.07330508422851563, 0.07296409606933593, 0.073133056640625, 0.07303987121582031, 0.15102873229980468, 0.07319551849365234, 0.07309209442138671, 0.07304294586181641, 0.07256678771972656, 0.07306444549560547, 0.07253708648681641, 0.07310745239257813, 0.07280435180664062, 0.07321600341796874, 0.07299072265625, 0.07320269012451172, 0.07370956420898438, 0.07317708587646485, 0.07267635345458984, 0.07315660858154296, 0.0728985595703125, 0.07292723083496094, 0.07305113220214844, 0.07316275024414062, 0.07275929260253906, 0.0730040283203125, 0.07312486267089843, 0.07343103790283204, 0.07317708587646485, 0.07305522918701172, 0.07260057830810547, 0.07282278442382813, 0.07714201354980468, 0.07391334533691406, 0.07291801452636719, 0.07313510131835937, 0.07281561279296875, 0.07718399810791016, 0.07579647827148438, 0.07586099243164063, 0.0756490249633789, 0.07552819061279296, 0.07454105377197266, 0.07292723083496094, 0.07275724792480469, 0.07294668579101563, 0.07279513549804688, 0.07289958190917968, 0.07298252868652344, 0.07305010986328125, 0.0727193603515625, 0.07310745239257813, 0.0728309783935547, 0.07297843170166016, 0.07271833801269531, 0.07224217224121093, 0.07278694152832031, 0.07278797149658203, 0.07299788665771484, 0.07332249450683594, 0.07290367889404296, 0.07284019470214843, 0.0741242904663086, 0.07310540771484375, 0.07278284454345703, 0.07531008148193359, 0.07580985260009766, 0.15569818115234374, 0.07325484466552734, 0.07306342315673828, 0.07312691497802734, 0.07297433471679687, 0.07300300598144531, 0.0730583038330078, 0.07312486267089843, 0.0755230712890625, 0.07598284912109375, 0.07572684478759766, 0.07567257690429688, 0.075720703125, 0.07279001617431641, 0.0730408935546875, 0.07312486267089843, 0.07281670379638672, 0.07280019378662109, 0.07297638702392578, 0.07320883178710938, 0.07256063842773437, 0.07415296173095703, 0.07416217803955077, 0.07563263702392578, 0.0757544937133789, 0.07627468872070313, 0.07312588500976562, 0.07302349090576171, 0.07244086456298827, 0.07299478149414063, 0.07283100891113281, 0.07330300903320312, 0.073312255859375, 0.07314739227294922, 0.07302758026123046, 0.07557427215576172, 0.07507762908935547, 0.074281982421875, 0.07310540771484375, 0.07593984222412109, 0.07586713409423829, 0.07613849639892578, 0.07600128173828125, 0.07603711700439453, 0.07613337707519531, 0.07313203430175781, 0.0730224609375, 0.07309619140625, 0.07404646301269531, 0.07319142150878906, 0.0729354248046875, 0.07320575714111328, 0.07308287811279297, 0.07316070556640625, 0.07301119995117188, 0.07307263946533203, 0.07310848236083985, 0.07297945404052734, 0.07294668579101563, 0.07294054412841797, 0.0724858856201172, 0.07309004974365234, 0.07307161712646484, 0.1527859191894531, 0.0737044448852539, 0.07706931304931641, 0.07324262237548829, 0.07301119995117188, 0.07327027130126954, 0.07272755432128906, 0.07289958190917968, 0.0730931167602539, 0.07241011047363281, 0.07289548492431641, 0.07333478546142579, 0.07326412963867188, 0.07310745239257813, 0.07281356811523437, 0.07283404541015626, 0.0729692153930664, 0.07528652954101563, 0.07494143676757813, 0.07375872039794922, 0.07302758026123046, 0.07273369598388672, 0.07280947113037109, 0.07290879821777344, 0.07290879821777344, 0.07317094421386719, 0.07293746948242187, 0.07284019470214843, 0.07299584197998046, 0.0731299819946289, 0.07301939392089844, 0.07360205078125, 0.0730931167602539, 0.07296307373046874, 0.0729139175415039, 0.07339520263671875, 0.07823667144775391, 0.07623168182373047, 0.07582310485839844, 0.07591321563720703, 0.0766402587890625, 0.07625318145751953, 0.07287296295166015, 0.0731935043334961, 0.07291286468505859, 0.07533567810058593, 0.0731514892578125, 0.07323340606689453, 0.07287091064453124, 0.07586201477050782, 0.07581388854980468, 0.07601766204833985, 0.07548416137695313, 0.07329792022705078, 0.07302041625976563, 0.07318118286132813, 0.07291289520263672, 0.07304710388183594, 0.07277664184570312, 0.07291808319091797, 0.07283808135986328, 0.07321600341796874, 0.07315763092041015, 0.1511731262207031, 0.07347609710693359, 0.07603404998779296, 0.07340953826904296, 0.0736890869140625, 0.07718605041503906, 0.07603199768066406, 0.07488921356201172, 0.07574937438964843, 0.07324364471435547, 0.07557734680175782, 0.07407103729248046, 0.07534489440917969, 0.07309721374511718, 0.07253298950195312, 0.07329177856445312, 0.07296511840820312, 0.07288527679443359, 0.07352726745605469, 0.07601971435546875, 0.07565824127197265, 0.07308595275878907, 0.07289759826660157, 0.07285958099365235, 0.07312691497802734, 0.07303679656982422, 0.07266918182373047, 0.0729722900390625, 0.07286067199707032, 0.0730224609375, 0.07309721374511718, 0.07305625915527343, 0.07309107208251953, 0.07335116577148437, 0.07298560333251954, 0.07305625915527343, 0.0729200668334961, 0.07550669097900391, 0.07300300598144531, 0.07319961547851563, 0.07294464111328125, 0.07308902740478515, 0.07374028778076172, 0.07326310729980469, 0.07590502166748046, 0.0759183349609375, 0.0758292465209961, 0.0757985610961914, 0.07546466827392578, 0.07296819305419922, 0.0728985595703125, 0.07331942749023437, 0.07303782653808594, 0.07312793731689453, 0.07287296295166015, 0.07295590209960938, 0.07329177856445312, 0.07311360168457032, 0.07294163513183594, 0.07298758697509766, 0.07293644714355468, 0.07294873809814453, 0.07297331237792969, 0.15123974609375, 0.07318726348876953, 0.07289958190917968, 0.07319757080078125, 0.07291187286376953, 0.07325491333007812, 0.07614873504638672, 0.076474365234375, 0.07479090881347657, 0.07814860534667968, 0.07577088165283204, 0.07610163116455078, 0.07617024230957031, 0.075936767578125, 0.07581081390380859, 0.07594710540771485, 0.07557520294189453, 0.07309107208251953, 0.07292518615722657, 0.07543603515625, 0.07567462158203125, 0.07564185333251953, 0.07563878631591797, 0.07566438293457031, 0.0732856674194336, 0.07495369720458984, 0.0755384292602539, 0.07294873809814453, 0.0750192642211914, 0.07620403289794922, 0.0731176986694336, 0.07587225341796874, 0.07608013153076172, 0.0765818862915039, 0.07283609771728515, 0.0742318115234375, 0.07550873565673828, 0.0753602523803711, 0.07581388854980468, 0.07458303833007812, 0.07516876983642579, 0.07688909149169922, 0.0779130859375, 0.07725055694580078, 0.07574630737304687, 0.07588556671142578, 0.0757022705078125, 0.07749734497070312, 0.07561113739013672, 0.0732221450805664, 0.0727715835571289, 0.0726118392944336, 0.07262822723388672, 0.07316172790527344, 0.07312895965576172, 0.07298764801025391, 0.07275520324707031, 0.07292313385009766, 0.07259033966064453, 0.0731678695678711, 0.07282176208496094, 0.07297945404052734, 0.0727930908203125, 0.1510502471923828, 0.07308697509765626, 0.07303782653808594, 0.07307571411132813, 0.07315968322753906, 0.07319961547851563, 0.07596953582763671, 0.07580671691894532, 0.07573709106445313, 0.07601971435546875, 0.07300300598144531, 0.07301837158203126, 0.07287808227539062, 0.07301535797119141, 0.07286573028564453, 0.07302352142333984, 0.07294358062744141, 0.07305010986328125, 0.07270611572265626, 0.0729578857421875, 0.07297740936279297, 0.07294156646728515, 0.07304707336425781, 0.07298454284667968, 0.07270809936523437, 0.07288729858398438, 0.07278694152832031, 0.07288114929199219, 0.07503359985351563, 0.07571561431884766, 0.07457276916503906, 0.07489638519287109, 0.07317298889160156, 0.07312793731689453, 0.07306041717529296, 0.07305209350585938, 0.07304499053955078, 0.07297126770019531, 0.0728770523071289, 0.07314534759521485, 0.07293440246582031, 0.07278797149658203, 0.07293644714355468, 0.07279206085205078, 0.07592243194580078, 0.07301427459716797, 0.07297945404052734, 0.07306240081787109, 0.07294873809814453, 0.072953857421875, 0.07344230651855468, 0.07307571411132813, 0.07297433471679687, 0.07304908752441407, 0.0730091552734375, 0.07554662322998047, 0.07547289276123047, 0.07562649536132812, 0.0739399642944336, 0.07577088165283204, 0.07293952178955078, 0.07277568054199218, 0.07285555267333985]",tokens/s,13.31416839419554,,,True -8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14144,7 +14144,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,True 8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1901.572096,1430.781952,0.0,775.94624,728.000512,s,10,0.5779317779541016,0.057793177795410156,0.002725987252288835,0.056286560058593754,0.059226506805419916,0.062293604660034176,0.06474728294372559,"[0.06536070251464844, 0.05848086547851562, 0.056102657318115236, 0.05854492950439453, 0.056225631713867186, 0.05626652908325196, 0.05614499282836914, 0.056306591033935545, 0.05844131088256836, 0.05605756759643555]",tokens/s,4429.588573693747,kWh,6.783094917024885e-07,3.7168158849713755e-07,1.1718326834970443e-06,2.2218237636966706e-06,tokens/kWh,115220659.79439664,MB,1901.572096,1430.781952,0.0,775.94624,769.381888,s,10,35.94489428710938,3.5944894287109377,0.04122920498905532,3.587615600585938,3.639309399414062,3.6603683227539063,3.6772154614257815,"[3.68142724609375, 3.634629638671875, 3.620576904296875, 3.57650537109375, 3.548501953125, 3.55525048828125, 3.54324951171875, 3.577020263671875, 3.60952197265625, 3.5982109375]",tokens/s,17.526828566190325,kWh,4.1868221404060483e-05,2.294285302515813e-05,6.349696905150463e-05,0.00012830804348072325,tokens/kWh,491005.85038119607,,s,629,36.39814343261718,0.057866682722761825,0.006863055416139851,0.05711155319213867,0.058331340026855465,0.05874790420532227,0.11222326965332032,"[0.06004019165039062, 0.057990142822265625, 0.059404289245605466, 0.058840065002441405, 0.058071041107177736, 0.05943091201782227, 0.056787967681884766, 0.06513766479492188, 0.05888614273071289, 0.05831372833251953, 0.05803724670410156, 0.05813043212890625, 0.057998336791992185, 0.05825126266479492, 0.058738689422607425, 0.05879500961303711, 0.05910015869140625, 0.058831871032714846, 0.05819596862792969, 0.05806387329101562, 0.05800755310058594, 0.05813462448120117, 0.05825526428222656, 0.058297344207763675, 0.05799116897583008, 0.0581662712097168, 0.058031105041503904, 0.05804646301269531, 0.058123264312744144, 0.05806489562988281, 0.058047489166259764, 0.05814476776123047, 0.058087425231933595, 0.05831167984008789, 0.05842432022094726, 0.05840281677246094, 0.058033153533935546, 0.05821542358398438, 0.0559370231628418, 0.05797990417480469, 0.058205184936523435, 0.058428417205810546, 0.05836288070678711, 0.058363903045654295, 0.058437664031982424, 0.058289119720458984, 0.05839974212646484, 0.058413055419921874, 0.05833523178100586, 0.058515457153320315, 0.058444801330566405, 0.05870796966552734, 0.05835468673706055, 0.058485759735107425, 0.05817036819458008, 0.058498046875, 0.0584796142578125, 0.058383392333984374, 0.05877657699584961, 0.05781705474853516, 0.05838643264770508, 0.05779865646362305, 0.11724697875976563, 0.05817449569702148, 0.05805871963500977, 0.05833216094970703, 0.059649024963378906, 0.05852364730834961, 0.056648704528808595, 0.058797054290771485, 0.05806694412231445, 0.05811609649658203, 0.058000385284423826, 0.058055679321289064, 0.05833011245727539, 0.05818777465820312, 0.0558837776184082, 0.0558766098022461, 0.055725055694580077, 0.055820289611816405, 0.055700481414794924, 0.05569331359863281, 0.055806976318359375, 0.05576704025268555, 0.05711155319213867, 0.05820211029052735, 0.05801881790161133, 0.05781913757324219, 0.05788876724243164, 0.057915393829345706, 0.05799423980712891, 0.0579502067565918, 0.05793689727783203, 0.058159103393554686, 0.05795328140258789, 0.05795635223388672, 0.05799731063842774, 0.058033153533935546, 0.05795942306518555, 0.058008575439453126, 0.057990142822265625, 0.058188800811767576, 0.05812838363647461, 0.0583004150390625, 0.059055103302001956, 0.05830758285522461, 0.0581130256652832, 0.05814169692993164, 0.05793689727783203, 0.057965568542480465, 0.058262527465820314, 0.058031105041503904, 0.05830348968505859, 0.05811916732788086, 0.057821182250976565, 0.057925697326660155, 0.057936832427978514, 0.05788467025756836, 0.05790105438232422, 0.05830656051635742, 0.05801062393188477, 0.05795942306518555, 0.055812095642089846, 0.055752704620361325, 0.055651329040527345, 0.11214950561523437, 0.05627084732055664, 0.0579051513671875, 0.05797990417480469, 0.05793689727783203, 0.058031105041503904, 0.05791027069091797, 0.057896961212158204, 0.05872844696044922, 0.05839155197143555, 0.05769113540649414, 0.05805055999755859, 0.05590016174316406, 0.05577523040771484, 0.05571891021728516, 0.05584896087646484, 0.055695358276367186, 0.05596057510375976, 0.055785472869873044, 0.05590220642089844, 0.0557209587097168, 0.05574553680419922, 0.055779327392578126, 0.0558919677734375, 0.055744510650634765, 0.05712691116333008, 0.05807923126220703, 0.05804544067382812, 0.058100799560546874, 0.05807404708862305, 0.058085376739501954, 0.05798092651367188, 0.05834035110473633, 0.058421279907226564, 0.058084320068359375, 0.05793689727783203, 0.058172416687011716, 0.058003456115722656, 0.05582131195068359, 0.05576806259155274, 0.056235008239746094, 0.058003456115722656, 0.05813248062133789, 0.057995262145996096, 0.058194942474365234, 0.05815091323852539, 0.058123264312744144, 0.05812838363647461, 0.058024959564208986, 0.058052608489990234, 0.05812531280517578, 0.05802700805664063, 0.058087425231933595, 0.05811097717285156, 0.057988094329833983, 0.05806489562988281, 0.05802188873291016, 0.05812428665161133, 0.05813248062133789, 0.05805977630615235, 0.05875404739379883, 0.05936640167236328, 0.058331134796142575, 0.11702886199951172, 0.058085376739501954, 0.057981952667236325, 0.05807820892333984, 0.058120193481445315, 0.05813862228393555, 0.05809151840209961, 0.058103809356689455, 0.057970687866210936, 0.058173439025878904, 0.05818572616577149, 0.0581478385925293, 0.055771137237548826, 0.055771137237548826, 0.05555712127685547, 0.055653377532958986, 0.05596876907348633, 0.0570695686340332, 0.0559370231628418, 0.056005630493164066, 0.0558551025390625, 0.05581721496582031, 0.05586739349365234, 0.05679718399047851, 0.05598310470581055, 0.05588582229614258, 0.05578035354614258, 0.05767987060546875, 0.058123264312744144, 0.058142719268798826, 0.05825228881835937, 0.05588787078857422, 0.05777612686157227, 0.05832089614868164, 0.05788876724243164, 0.05814374542236328, 0.058246143341064455, 0.058261566162109375, 0.05610899353027344, 0.05596979141235352, 0.055820289611816405, 0.056097793579101565, 0.055992321014404295, 0.05596979141235352, 0.055979007720947264, 0.05584691238403321, 0.05592473602294922, 0.05586739349365234, 0.05589404678344727, 0.05583561706542969, 0.05722316741943359, 0.058047489166259764, 0.05607424163818359, 0.05604044723510742, 0.05591555023193359, 0.05593801498413086, 0.056019966125488284, 0.05596876907348633, 0.05613260650634765, 0.05608652877807617, 0.055962623596191405, 0.05597398376464844, 0.05585712051391602, 0.1125805435180664, 0.05603430557250977, 0.05604249572753906, 0.055929855346679686, 0.05602406311035156, 0.055994369506835937, 0.05609676742553711, 0.05613260650634765, 0.05601484680175781, 0.05618380737304687, 0.056136703491210936, 0.055731201171875, 0.055890945434570315, 0.05611929702758789, 0.05656371307373047, 0.05599334335327148, 0.05603430557250977, 0.056169471740722655, 0.055981056213378906, 0.0560711669921875, 0.05607833480834961, 0.05602201461791992, 0.056005630493164066, 0.05616025543212891, 0.05610604858398437, 0.058043327331542965, 0.05733171081542969, 0.05605478286743164, 0.0560445442199707, 0.05726617431640625, 0.05755904006958008, 0.05617868804931641, 0.05611212921142578, 0.05601484680175781, 0.05625241470336914, 0.055989246368408206, 0.05592268753051758, 0.05610393524169922, 0.05606195068359375, 0.055994369506835937, 0.055981056213378906, 0.0561162223815918, 0.05616332626342774, 0.05613260650634765, 0.05607424163818359, 0.05591654586791992, 0.05608454513549805, 0.05609772872924805, 0.05603942489624023, 0.056118270874023435, 0.05638041687011719, 0.058452991485595705, 0.05657088088989258, 0.05627801513671875, 0.05605990219116211, 0.05607731246948242, 0.057839614868164066, 0.05867929458618164, 0.05654118347167969, 0.05691187286376953, 0.05704092788696289, 0.05597798538208008, 0.05619504165649414, 0.11253350067138672, 0.05620940780639649, 0.05641318511962891, 0.05596371078491211, 0.05575775909423828, 0.05598720169067383, 0.05637324905395508, 0.05719347381591797, 0.05790105438232422, 0.05735424041748047, 0.05588172912597656, 0.057796607971191405, 0.05588991928100586, 0.05638041687011719, 0.057078784942626956, 0.056374271392822264, 0.05770345687866211, 0.057278430938720704, 0.057629695892333986, 0.05825024032592773, 0.05569228744506836, 0.055567359924316405, 0.05579980850219726, 0.05586636734008789, 0.05641625595092774, 0.05565849685668945, 0.05581619262695312, 0.05580492782592773, 0.05580287933349609, 0.0568350715637207, 0.05705321502685547, 0.05733679962158203, 0.05608761596679687, 0.05629433441162109, 0.056787967681884766, 0.0564029426574707, 0.057518081665039064, 0.057659393310546876, 0.05798912048339844, 0.05586431884765625, 0.055839744567871094, 0.056622081756591794, 0.056976383209228515, 0.05614080047607422, 0.05781196975708008, 0.05928243255615234, 0.05699686431884766, 0.05576499176025391, 0.05546188735961914, 0.05536870574951172, 0.055803905487060546, 0.05569638442993164, 0.055462913513183595, 0.055478271484375, 0.05592063903808594, 0.05581619262695312, 0.055836673736572265, 0.05566156768798828, 0.05587353515625, 0.055907329559326174, 0.0557762565612793, 0.05598310470581055, 0.05588787078857422, 0.1122324447631836, 0.05631590270996094, 0.05595443344116211, 0.05598822402954102, 0.055766014099121096, 0.05570460891723633, 0.055920608520507814, 0.05544345474243164, 0.05686374282836914, 0.05620940780639649, 0.05592268753051758, 0.05579673767089844, 0.05570150375366211, 0.057382911682128904, 0.055947265625, 0.055478271484375, 0.0559554557800293, 0.05573836898803711, 0.05545062255859375, 0.05589913558959961, 0.05589811325073242, 0.05598822402954102, 0.05587968063354492, 0.05556326293945312, 0.056275966644287106, 0.05612236785888672, 0.05595238494873047, 0.05615923309326172, 0.05589503860473633, 0.055897087097167966, 0.05583257675170898, 0.057462814331054685, 0.05774332809448242, 0.05599129486083984, 0.05584793472290039, 0.05587353515625, 0.058039329528808595, 0.05993366241455078, 0.057166847229003906, 0.057599998474121096, 0.05694976043701172, 0.05626163101196289, 0.05697740936279297, 0.05767679977416992, 0.05807001495361328, 0.05718534469604492, 0.05664147186279297, 0.05595647811889649, 0.055932926177978515, 0.05580595016479492, 0.0559370231628418, 0.055757823944091796, 0.0558837776184082, 0.055731201171875, 0.05574041748046875, 0.055790592193603515, 0.055944190979003904, 0.05588787078857422, 0.05582131195068359, 0.05511990356445313, 0.05580489730834961, 0.05583462524414062, 0.05589503860473633, 0.11259603118896484, 0.05576492691040039, 0.05574348831176758, 0.05575987243652344, 0.055926784515380856, 0.05582745742797852, 0.055782398223876956, 0.05584691238403321, 0.055839744567871094, 0.05590016174316406, 0.056323070526123044, 0.05777305603027344, 0.05587968063354492, 0.055298046112060545, 0.05588787078857422, 0.05575884628295898, 0.05542911911010742, 0.055949310302734374, 0.055785472869873044, 0.056118270874023435, 0.056443904876708986, 0.05585919952392578, 0.05586022567749024, 0.05584588623046875, 0.05584076690673828, 0.055766014099121096, 0.056139774322509765, 0.05710540771484375, 0.05816831970214844, 0.05807308959960938, 0.05796352005004883, 0.058262527465820314, 0.05813862228393555, 0.05806387329101562, 0.05820828628540039, 0.05820412826538086, 0.0580761604309082, 0.058038272857666016, 0.05579264068603516, 0.05584691238403321, 0.05575987243652344, 0.05593804931640625, 0.05876326370239258, 0.05823385620117188, 0.05807513427734375, 0.05791027069091797, 0.058024959564208986, 0.05811404800415039, 0.05589811325073242, 0.0557916145324707, 0.05582131195068359, 0.057818111419677735, 0.05807308959960938, 0.05788467025756836, 0.05813043212890625, 0.05792870330810547, 0.057924606323242187, 0.05807001495361328, 0.05830963134765625, 0.055911422729492184, 0.05595750427246094, 0.055897087097167966, 0.0562083854675293, 0.11693977355957032, 0.058156032562255856, 0.05791436767578125, 0.05803519821166992, 0.05813145446777344, 0.0581069450378418, 0.05796857452392578, 0.058006526947021485, 0.058006526947021485, 0.058052608489990234, 0.05799935913085937, 0.05808639907836914, 0.057945152282714844, 0.05808838272094727, 0.05795942306518555, 0.05803724670410156, 0.05846527862548828, 0.05814476776123047, 0.05797990417480469, 0.05802188873291016, 0.055947265625, 0.05751603317260742, 0.058103809356689455, 0.057841728210449216, 0.05802284622192383, 0.0579788818359375, 0.05784883117675781, 0.0580208625793457, 0.05794508743286133, 0.05975142288208008, 0.05937152099609375, 0.05811711883544922, 0.058090496063232425, 0.058314750671386716, 0.05809868621826172, 0.058003456115722656, 0.0576286735534668, 0.05805158233642578, 0.05776793670654297, 0.057883647918701174, 0.05789388656616211, 0.05584896087646484, 0.05586329650878906, 0.055913471221923826, 0.05580083084106445, 0.0557496337890625, 0.05579980850219726, 0.05578137588500977, 0.05577830505371094, 0.055841793060302736, 0.05584281539916992, 0.05635583877563476, 0.05595238494873047, 0.05585203170776367, 0.055760894775390625, 0.055877632141113284, 0.05580492782592773, 0.05600153732299805, 0.05584793472290039, 0.05583359909057617, 0.055831550598144535, 0.05582032012939453, 0.05661692810058594, 0.11219967651367188, 0.05580799865722656, 0.05563699340820313, 0.05566873550415039, 0.05575584030151367, 0.05771155166625976, 0.058019840240478515, 0.056458240509033204, 0.05607526397705078, 0.05592473602294922, 0.05661084747314453, 0.05880521774291992, 0.05757747268676758, 0.05812223815917969, 0.05797478485107422, 0.05746790313720703, 0.057995262145996096, 0.05818777465820312, 0.05807513427734375, 0.05587353515625, 0.057148414611816405, 0.060609535217285154, 0.05641731262207031, 0.05592879867553711, 0.05582748794555664, 0.0559318733215332, 0.05581619262695312, 0.055946239471435545, 0.057985023498535154, 0.058241024017333984, 0.058164222717285156, 0.05822982406616211, 0.05803513717651367, 0.05816729736328125, 0.058000446319580075, 0.05585913467407227, 0.05583052825927735, 0.05590220642089844, 0.05554687881469727, 0.05571686553955078, 0.05598822402954102, 0.055874561309814455, 0.05583052825927735, 0.055774208068847655, 0.055934974670410156, 0.05581926345825195, 0.05595238494873047, 0.055809024810791016, 0.0563507194519043, 0.05797478485107422, 0.058126335144042966, 0.05808844757080078, 0.05775872039794922, 0.057998336791992185, 0.058234878540039066, 0.05836697769165039, 0.05817753601074219, 0.05861478424072265, 0.059649024963378906, 0.058420223236083986, 0.058019840240478515, 0.058203166961669925, 0.058082271575927734]",tokens/s,17.281101195846684,,,True -8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14308,7 +14308,7 @@ If this is a private repository, make sure to pass a token having permission to 8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1995.53024,1726.480384,0.0,1071.644672,1001.713664,s,10,0.7049398422241212,0.07049398422241213,0.0024606827235681547,0.06945233535766601,0.07317600326538086,0.07519488182067871,0.076809984664917,"[0.07721376037597656, 0.07272736358642579, 0.06931212615966797, 0.06912992095947265, 0.06970499420166015, 0.06964585876464843, 0.06941340637207032, 0.06949126434326172, 0.06904025268554688, 0.06926089477539063]",tokens/s,3631.515551629298,kWh,8.375258650997664e-07,4.5892308306916823e-07,1.4185689047515943e-06,2.715017852920529e-06,tokens/kWh,94290356.03748325,MB,1995.53024,1726.480384,0.0,1071.644672,1024.136704,s,10,43.077471679687505,4.307747167968751,0.0442409680485182,4.2932119140625,4.360292626953125,4.3745864990234375,4.386021596679687,"[4.3571162109375, 4.38888037109375, 4.34852685546875, 4.29621923828125, 4.259873046875, 4.25883251953125, 4.26343896484375, 4.27718408203125, 4.33719580078125, 4.29020458984375]",tokens/s,14.62481374683525,kWh,5.143175892631682e-05,2.818536529607685e-05,7.837496715224743e-05,0.00015799209137464106,tokens/kWh,398754.13669036335,,s,629,43.637266403198275,0.06937562226263631,0.008448929213692532,0.06757068634033203,0.0703541259765625,0.07072071380615234,0.13712687561035156,"[0.06743049621582031, 0.06641756439208985, 0.06657023620605469, 0.06655590057373047, 0.06754815673828125, 0.06761369323730469, 0.068168701171875, 0.07222067260742188, 0.07090892791748046, 0.06737612915039062, 0.0675072021484375, 0.06748774719238282, 0.06764543914794922, 0.067346435546875, 0.06740172576904296, 0.06768025970458984, 0.06747545623779297, 0.06741401672363281, 0.06748057556152344, 0.06726656341552735, 0.06832434844970703, 0.07105843353271485, 0.06786150360107422, 0.06988390350341797, 0.0709191665649414, 0.07030681610107421, 0.0695736312866211, 0.07029145812988281, 0.07231590270996094, 0.07028018951416015, 0.07027200317382812, 0.07035903930664063, 0.0697548828125, 0.07009689331054687, 0.07025872039794921, 0.07024124908447266, 0.07021568298339843, 0.07005184173583984, 0.07003347015380859, 0.06811334228515625, 0.06995865631103515, 0.06940160369873047, 0.070002685546875, 0.07031705474853515, 0.06862438201904297, 0.0694692153930664, 0.06968112182617188, 0.06975794982910156, 0.07021465301513671, 0.06939852905273437, 0.07020748901367188, 0.07068883514404296, 0.06897657775878906, 0.0695951690673828, 0.06884041595458984, 0.07022694396972656, 0.06972927856445313, 0.07013683319091797, 0.06931558227539063, 0.06981836700439453, 0.06857727813720703, 0.07020543670654297, 0.14347161865234376, 0.0704000015258789, 0.07029248046875, 0.07024947357177734, 0.07041433715820312, 0.06891315460205077, 0.07033344268798829, 0.07019519805908203, 0.07010918426513672, 0.07032422637939453, 0.06962175750732422, 0.06962073516845703, 0.071225341796875, 0.0705433578491211, 0.07032627105712891, 0.06988800048828125, 0.07026483154296875, 0.06874521636962891, 0.0697548828125, 0.07011737823486328, 0.07013478088378906, 0.06828134155273438, 0.0701317138671875, 0.06990643310546875, 0.07017574310302735, 0.0703846435546875, 0.06922956848144532, 0.06984607696533203, 0.07022176361083984, 0.07008460998535156, 0.06941494750976562, 0.06964733123779297, 0.07022592163085938, 0.0710983657836914, 0.07038771057128906, 0.06999142456054687, 0.07039282989501953, 0.07020441436767579, 0.0693237762451172, 0.07022284698486328, 0.07020134735107422, 0.06978457641601563, 0.06865408325195313, 0.06965760040283203, 0.07016960144042969, 0.0682977294921875, 0.07028018951416015, 0.07000985717773438, 0.06735871887207032, 0.06748159790039063, 0.06758911895751953, 0.06750617980957031, 0.06752460479736327, 0.06908415985107422, 0.07013887786865235, 0.07035289764404297, 0.06866944122314453, 0.07013887786865235, 0.06987980651855469, 0.07023616027832032, 0.06918758392333985, 0.0678042221069336, 0.06753785705566406, 0.13878886413574218, 0.06763520050048828, 0.06754611206054688, 0.07073177337646484, 0.07045734405517579, 0.067451904296875, 0.06804377746582031, 0.07022188568115234, 0.06730335998535156, 0.06743859100341797, 0.06733721923828125, 0.06729625701904297, 0.06735871887207032, 0.06729523468017579, 0.06739046478271485, 0.06721638488769531, 0.06737612915039062, 0.06736691284179687, 0.07039590454101563, 0.06981529235839844, 0.07017984008789062, 0.07010918426513672, 0.07007027435302735, 0.06803763580322265, 0.07009184265136718, 0.07207008361816407, 0.07051366424560547, 0.0700794906616211, 0.07030271911621094, 0.07009279632568359, 0.068927490234375, 0.06632755279541015, 0.06746931457519531, 0.06824755096435547, 0.06765875244140625, 0.06756658935546875, 0.0673628158569336, 0.06796595001220704, 0.07062118530273437, 0.07026073455810547, 0.07157555389404296, 0.07031193542480468, 0.07020543670654297, 0.06981324768066406, 0.07054137420654297, 0.07186220550537109, 0.07062118530273437, 0.07063660430908203, 0.06995142364501954, 0.06877388763427734, 0.06950911712646485, 0.06939443206787109, 0.07045734405517579, 0.07033859252929688, 0.07044297790527344, 0.07006208038330078, 0.06987264251708984, 0.06725529479980469, 0.067346435546875, 0.0680284194946289, 0.06730445098876953, 0.06736486053466798, 0.0677427215576172, 0.1370951690673828, 0.06738636779785157, 0.06746521759033203, 0.06756864166259766, 0.06744166564941406, 0.06722969818115235, 0.06735564422607422, 0.06744064331054687, 0.06907392120361328, 0.07031910705566406, 0.06906777954101563, 0.06806221008300781, 0.06722662353515625, 0.06994124603271484, 0.07015731048583984, 0.07010918426513672, 0.06880255889892578, 0.07000678253173828, 0.06957263946533203, 0.07037436676025391, 0.06743142700195312, 0.06735462188720703, 0.0674150390625, 0.06752870178222656, 0.06720921325683593, 0.06740275573730468, 0.06759935760498047, 0.06755225372314454, 0.06730854034423828, 0.06749081420898438, 0.07013683319091797, 0.06936576080322265, 0.07008255767822266, 0.07028736114501953, 0.07038566589355469, 0.07017369842529297, 0.06959513854980469, 0.07015731048583984, 0.07019417572021484, 0.06882508850097656, 0.0673802261352539, 0.06905036926269531, 0.06771302032470704, 0.06747238159179687, 0.06732902526855469, 0.06704640197753907, 0.0671272964477539, 0.06733004760742188, 0.06743961334228515, 0.0670904312133789, 0.06670130920410157, 0.06742221069335938, 0.06702079772949218, 0.06737407684326172, 0.06731263732910156, 0.06773043060302734, 0.06748467254638672, 0.0675225601196289, 0.06684774780273438, 0.06723583984375, 0.06727680206298828, 0.06736793518066406, 0.06794445037841797, 0.13754060363769532, 0.06699827575683594, 0.06684159851074219, 0.06734130859375, 0.06724095916748046, 0.06727680206298828, 0.06753485107421875, 0.06713855743408204, 0.06749081420898438, 0.06713855743408204, 0.06649756622314452, 0.06733615875244141, 0.06746112060546874, 0.06731263732910156, 0.06726451110839844, 0.06704434967041016, 0.06736589050292968, 0.06724403381347656, 0.06742937469482421, 0.07004364776611328, 0.06723788452148438, 0.06995353698730469, 0.07074406433105469, 0.06757273864746094, 0.06735564422607422, 0.06732697296142579, 0.06733926391601562, 0.06775193786621093, 0.06763520050048828, 0.06753382110595703, 0.06746828460693359, 0.06739250946044922, 0.06742835235595702, 0.06737612915039062, 0.06711004638671875, 0.06775382232666016, 0.06726656341552735, 0.0676864013671875, 0.0674672622680664, 0.06746316528320312, 0.06745906829833985, 0.06693888092041016, 0.06735359954833985, 0.06803763580322265, 0.06830694580078125, 0.06761779022216798, 0.06732697296142579, 0.06699622344970703, 0.06767411041259766, 0.06760140991210938, 0.06779596710205078, 0.06751334381103516, 0.06753279876708984, 0.06744371032714844, 0.0696780776977539, 0.06771097564697266, 0.06809600067138671, 0.06781747436523437, 0.06764236450195313, 0.06857933044433594, 0.06740684509277343, 0.06741401672363281, 0.06726348876953125, 0.13843455505371094, 0.06815744018554687, 0.06736589050292968, 0.06758611297607423, 0.06722758483886719, 0.06736589050292968, 0.06744780731201172, 0.0673433609008789, 0.06746009826660156, 0.06744882965087891, 0.06785945892333985, 0.06740889739990234, 0.06752051544189454, 0.06737203216552734, 0.06750208282470703, 0.06727577972412109, 0.06799667358398437, 0.06718566131591797, 0.06747443389892578, 0.06731059265136718, 0.06748159790039063, 0.0673986587524414, 0.0675440673828125, 0.06750310516357422, 0.06736179351806641, 0.0670535659790039, 0.06757478332519531, 0.06749183654785157, 0.0675225601196289, 0.06746316528320312, 0.06772838592529297, 0.06738227081298828, 0.06734233856201172, 0.06742425537109376, 0.0686192626953125, 0.06744678497314453, 0.0674151382446289, 0.06733712005615235, 0.06803250885009765, 0.06742015838623047, 0.06749286651611328, 0.06743449401855468, 0.06751641845703125, 0.06624050903320312, 0.0676485137939453, 0.06727884674072265, 0.06757478332519531, 0.0672706527709961, 0.06748883056640625, 0.06746412658691406, 0.06743452453613281, 0.06719280242919921, 0.0669491195678711, 0.06755840301513671, 0.0675594253540039, 0.06860185241699218, 0.06992588806152343, 0.07098368072509766, 0.06745906829833985, 0.06732492828369141, 0.06783795166015626, 0.06735564422607422, 0.06761369323730469, 0.1371392059326172, 0.06737100982666015, 0.06738636779785157, 0.06773043060302734, 0.06790860748291015, 0.06763929748535157, 0.06747853088378906, 0.06725225830078126, 0.06746006774902344, 0.06740172576904296, 0.06743654632568359, 0.0679188461303711, 0.06745702362060547, 0.06692454528808593, 0.06735359954833985, 0.06734848022460938, 0.0674672622680664, 0.06737920379638672, 0.06987366485595703, 0.06749081420898438, 0.06731468963623047, 0.06746214294433593, 0.0674734115600586, 0.06750208282470703, 0.06762086486816406, 0.06789529418945313, 0.06751436614990235, 0.06727474975585937, 0.06737100982666015, 0.06724403381347656, 0.06809907531738281, 0.07002726745605468, 0.06752870178222656, 0.06763724517822266, 0.06680780792236328, 0.06763008117675781, 0.06759426879882813, 0.06765155029296875, 0.06745702362060547, 0.0674775390625, 0.06785430145263673, 0.06886605072021484, 0.0706662368774414, 0.07001190185546875, 0.06753485107421875, 0.06751846313476563, 0.06739456176757813, 0.06731673431396484, 0.06764339447021485, 0.06748467254638672, 0.06751334381103516, 0.06752460479736327, 0.06757068634033203, 0.06709145355224609, 0.06695116424560547, 0.06790758514404296, 0.06752665710449218, 0.06749491119384765, 0.06747853088378906, 0.06743654632568359, 0.06722560119628906, 0.06742323303222657, 0.06748467254638672, 0.13912474060058594, 0.0679178237915039, 0.06786764526367188, 0.06751026916503906, 0.06756864166259766, 0.06750514984130859, 0.06994226837158203, 0.0707041244506836, 0.07019110107421875, 0.069823486328125, 0.06723583984375, 0.06712012481689453, 0.0677201919555664, 0.06726963043212891, 0.06998528289794922, 0.07064985656738282, 0.0693729248046875, 0.06735359954833985, 0.0673433609008789, 0.07043891143798828, 0.07088435363769531, 0.06736899566650391, 0.06737814331054688, 0.06719078063964844, 0.06727891540527343, 0.06741190338134766, 0.0678123550415039, 0.07006515502929687, 0.06998528289794922, 0.06990541076660156, 0.07003852844238281, 0.06716928100585938, 0.06733106994628907, 0.06704537963867188, 0.06707814025878907, 0.06724198150634765, 0.06734130859375, 0.06756454467773437, 0.0674334716796875, 0.06743859100341797, 0.06752358245849609, 0.06730547332763671, 0.0672103042602539, 0.06745184326171876, 0.06748057556152344, 0.06746419525146484, 0.06744374084472657, 0.0674662094116211, 0.06709657287597656, 0.06692044830322266, 0.06734745788574219, 0.06733209228515626, 0.06756249237060546, 0.06746623992919921, 0.06757990264892579, 0.06722560119628906, 0.06654566192626953, 0.0661944351196289, 0.0664463348388672, 0.066482177734375, 0.06660403442382813, 0.06645555114746093, 0.06649858856201171, 0.1386700439453125, 0.06732902526855469, 0.06740684509277343, 0.07023820495605469, 0.06744678497314453, 0.06748159790039063, 0.06712217712402344, 0.0675953598022461, 0.06757469177246093, 0.06737715148925781, 0.067378173828125, 0.06710578918457032, 0.06729011535644532, 0.06744985961914063, 0.06741709136962891, 0.06734438323974609, 0.06720921325683593, 0.06826496124267578, 0.07156940460205079, 0.06848111724853516, 0.06911888122558593, 0.07064780426025391, 0.0690145263671875, 0.06772121429443359, 0.06744678497314453, 0.06804275512695312, 0.070181884765625, 0.07029452514648438, 0.06997299194335938, 0.07016448211669922, 0.0672573471069336, 0.06961049652099609, 0.0704000015258789, 0.07023513793945313, 0.07003238677978516, 0.0674181137084961, 0.06744371032714844, 0.06774374389648438, 0.06725017547607422, 0.06717440032958985, 0.06726963043212891, 0.06885068511962891, 0.07013785552978516, 0.07033036804199219, 0.07023107147216796, 0.0678399658203125, 0.07103794860839843, 0.0703477783203125, 0.07032217407226563, 0.0701839370727539, 0.07003651428222656, 0.06756246185302735, 0.06736383819580079, 0.067451904296875, 0.0688721923828125, 0.06854860687255859, 0.07121920013427735, 0.07080348968505859, 0.0700016326904297, 0.07014911651611327, 0.07204354858398437, 0.07053411102294922, 0.07017984008789062, 0.136953857421875, 0.06727884674072265, 0.06774988555908203, 0.06740172576904296, 0.06736179351806641, 0.06728498840332031, 0.06735871887207032, 0.06704640197753907, 0.06692864227294922, 0.0673966064453125, 0.06758297729492188, 0.07042765045166016, 0.067346435546875, 0.06728604888916015, 0.06719996643066406, 0.06728396606445312, 0.06736383819580079, 0.06747443389892578, 0.06817279815673828, 0.07137586975097657, 0.07003955078125, 0.06764543914794922, 0.0672911376953125, 0.06722457885742188, 0.0673433609008789, 0.06727577972412109, 0.06753997039794922, 0.0673986587524414, 0.06727174377441406, 0.06864991760253907, 0.06758809661865234, 0.0704901123046875, 0.06749183654785157, 0.07135948944091797, 0.07054236602783204, 0.07037641906738282, 0.07059353637695312, 0.07142809295654297, 0.0712273941040039, 0.06735769653320313, 0.06720409393310547, 0.06732697296142579, 0.06732905578613281, 0.06735356903076171, 0.06733830261230468, 0.067295166015625, 0.06732288360595703, 0.06716722869873047, 0.06738227081298828, 0.06736077117919922, 0.06743551635742187, 0.06729017639160156, 0.06730850982666016, 0.06715494537353515, 0.06742115020751953, 0.0674703369140625, 0.0675072021484375, 0.06733414459228515, 0.06761062622070313, 0.07013069152832031, 0.06913536071777343, 0.07012966156005859, 0.07008358764648437]",tokens/s,14.414285124741445,,,True 8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1812.873216,1174.929408,0.0,520.093696,421.38624,s,10,0.6843496627807618,0.06843496627807619,0.0027651970796180084,0.0673604965209961,0.06967556838989258,0.07312517585754394,0.07588486183166504,"[0.07657478332519531, 0.06733692932128907, 0.06738406372070313, 0.0672044448852539, 0.06740876770019531, 0.0680913314819336, 0.0671644515991211, 0.06890898895263672, 0.06719996643066406, 0.06707593536376953]",tokens/s,3740.777762054836,kWh,8.095983807082591e-07,4.436213311760495e-07,1.3120330130959507e-06,2.5652527249802592e-06,tokens/kWh,99795235.57546169,MB,1812.873216,1174.929408,0.0,520.093696,446.103552,s,10,42.014822753906245,4.201482275390625,0.017125822995660088,4.19976708984375,4.22275654296875,4.224974462890625,4.226748798828125,"[4.2271923828125, 4.1861474609375, 4.20341064453125, 4.18505908203125, 4.222263671875, 4.21279541015625, 4.18522412109375, 4.19612353515625, 4.17720556640625, 4.21940087890625]",tokens/s,14.994708026976667,kWh,4.980282585052958e-05,2.729481079746844e-05,7.46637222119036e-05,0.00015176135885990158,tokens/kWh,415125.4342560178,,s,629,42.5558221054078,0.06765631495295343,0.008103380321689464,0.06647090911865235,0.06769684448242187,0.06875012969970702,0.13383589904785156,"[0.06650470733642579, 0.06613504028320312, 0.0663193588256836, 0.06626713562011718, 0.06561177825927735, 0.06649552154541015, 0.06632137298583984, 0.07682252502441406, 0.06930329895019531, 0.06741709136962891, 0.06680678558349609, 0.0669839324951172, 0.06668704223632813, 0.06750611114501953, 0.06876569366455078, 0.06785024261474609, 0.06798745727539063, 0.06866329956054687, 0.06728806304931641, 0.07015321350097656, 0.06926847839355468, 0.06691020965576172, 0.06660403442382813, 0.0670013427734375, 0.06664192199707031, 0.06665523529052735, 0.06687334442138672, 0.06697062683105469, 0.0666081314086914, 0.06665318298339844, 0.06660300445556641, 0.06833561706542969, 0.06880255889892578, 0.06652518463134766, 0.06659174346923828, 0.0667177276611328, 0.06656918334960937, 0.06668185424804687, 0.06682726287841798, 0.06671564483642578, 0.06595174407958984, 0.06646790313720703, 0.06678419494628907, 0.06667263793945312, 0.06658252716064453, 0.06664297485351563, 0.06646985626220703, 0.06658150482177734, 0.06660710144042968, 0.0665159683227539, 0.06637670135498047, 0.06615449523925782, 0.06655795288085938, 0.06647193908691407, 0.06654367828369141, 0.06663263702392579, 0.06674022674560547, 0.06677606201171875, 0.06663270568847657, 0.0664606704711914, 0.06650163269042969, 0.06652722930908203, 0.1343231964111328, 0.06652931213378906, 0.06645552062988282, 0.0665528335571289, 0.06543360137939454, 0.06651699066162109, 0.06630809783935547, 0.06551449584960937, 0.06542745971679688, 0.06670342254638671, 0.06657939147949218, 0.06644429016113282, 0.06634188842773438, 0.06641458892822266, 0.06550323486328125, 0.06625382232666016, 0.0665159683227539, 0.06637875366210938, 0.06623436737060547, 0.06598963165283203, 0.06640230560302735, 0.06661222076416015, 0.06661222076416015, 0.06635724639892578, 0.06655487823486328, 0.06641868591308593, 0.06646272277832031, 0.06652825927734375, 0.06700953674316407, 0.0666961898803711, 0.06618521881103516, 0.06530764770507813, 0.06615039825439453, 0.06655795288085938, 0.06667059326171874, 0.06658252716064453, 0.06652006530761718, 0.06654463958740234, 0.0667628173828125, 0.06703814697265625, 0.06660403442382813, 0.06642278289794921, 0.06637055969238281, 0.06755328369140624, 0.06618726348876953, 0.0666275863647461, 0.06656716918945313, 0.06657638549804687, 0.0665528335571289, 0.06636236572265625, 0.06640025329589844, 0.06667775726318359, 0.06652722930908203, 0.06644121551513672, 0.0665692138671875, 0.06706893157958985, 0.06656614685058594, 0.0665692138671875, 0.06651699066162109, 0.06636339569091797, 0.06645145416259765, 0.06661427307128906, 0.06653440093994141, 0.134108154296875, 0.0664453125, 0.06595993804931641, 0.0656209945678711, 0.06547456359863281, 0.06662655639648438, 0.06631526184082032, 0.06629068756103515, 0.06687744140625, 0.06665216064453125, 0.06636134338378906, 0.06882918548583984, 0.06843596649169922, 0.0665384979248047, 0.06639718627929687, 0.06648524475097656, 0.0664453125, 0.06638285064697266, 0.0664985580444336, 0.06783999633789063, 0.06658048248291015, 0.06622720336914062, 0.06652722930908203, 0.06664704132080078, 0.06634393310546875, 0.06574285125732422, 0.0663552017211914, 0.06754099273681641, 0.06646169281005859, 0.06644223785400391, 0.06645974731445313, 0.06643190765380859, 0.06634393310546875, 0.06632038116455079, 0.06803865814208984, 0.06646578979492188, 0.06750521850585937, 0.0664677734375, 0.06654566192626953, 0.0669665298461914, 0.06615654754638672, 0.0662999038696289, 0.06655795288085938, 0.06867046356201172, 0.06678937530517579, 0.06591180419921874, 0.06617906951904297, 0.06677094268798828, 0.06634905242919922, 0.06888038635253907, 0.06691532897949219, 0.06648934173583984, 0.06629785919189453, 0.06634700775146485, 0.06836428833007813, 0.06864588928222656, 0.06689689636230468, 0.06642585754394531, 0.06750822448730469, 0.06629682922363281, 0.06644019317626954, 0.06629478454589843, 0.0663367691040039, 0.13442457580566405, 0.0663746566772461, 0.06666035461425782, 0.06651187133789062, 0.06649651336669922, 0.06645862579345703, 0.0663552017211914, 0.06701465606689454, 0.067557373046875, 0.0666204833984375, 0.0665630111694336, 0.06640025329589844, 0.06633881378173828, 0.06692454528808593, 0.06653337860107422, 0.06796083068847657, 0.06694297790527344, 0.06639103698730468, 0.0659230728149414, 0.0665354232788086, 0.06639615631103515, 0.06634291076660156, 0.06649343872070312, 0.06642585754394531, 0.06655385589599609, 0.06636038208007812, 0.06630188751220703, 0.06634086608886719, 0.06635929870605468, 0.06632959747314453, 0.06652210998535156, 0.06716006469726563, 0.06658048248291015, 0.06627021026611328, 0.06635008239746094, 0.06623747253417969, 0.06623433685302735, 0.06609919738769532, 0.06637977600097657, 0.06661017608642578, 0.06636032104492187, 0.06623337554931641, 0.0663674545288086, 0.06639718627929687, 0.06650367736816407, 0.06639308929443359, 0.06623948669433594, 0.06634598541259766, 0.06624460601806641, 0.06633369445800781, 0.06631833648681641, 0.06623027038574218, 0.06618624114990235, 0.06616883087158203, 0.06636748504638672, 0.06621286773681641, 0.06631629180908204, 0.06615654754638672, 0.06630912017822266, 0.06630400085449219, 0.06477721405029296, 0.0658892822265625, 0.06630707550048828, 0.13415731811523438, 0.06667059326171874, 0.06632860565185547, 0.06637052917480468, 0.06608793640136719, 0.06854246520996093, 0.06768230438232421, 0.06865203094482422, 0.06664399719238281, 0.06849635314941406, 0.06644735717773438, 0.06643917083740235, 0.06655999755859375, 0.06750105285644531, 0.06708633422851562, 0.06686822509765625, 0.06753997039794922, 0.06640742492675782, 0.06689485168457031, 0.06709145355224609, 0.06747647857666016, 0.0666982421875, 0.067162109375, 0.06637158203125, 0.06683545684814453, 0.0673986587524414, 0.0663531494140625, 0.06653132629394531, 0.06637773132324219, 0.06623232269287109, 0.06644735717773438, 0.06636544036865234, 0.06650572967529297, 0.06955213165283203, 0.06782361602783203, 0.06751334381103516, 0.06656719970703125, 0.06663990020751953, 0.06626604461669922, 0.06670642852783203, 0.0664238052368164, 0.06777855682373046, 0.0674897918701172, 0.06728396606445312, 0.06645247650146484, 0.06682316589355469, 0.06786969757080077, 0.0659752960205078, 0.06847078704833984, 0.06627430725097656, 0.0691230697631836, 0.06647090911865235, 0.06799366760253907, 0.06640531158447266, 0.06817382049560547, 0.06768844604492187, 0.06595276641845703, 0.0667136001586914, 0.06713862609863282, 0.06644525146484374, 0.06705049896240234, 0.06686412811279296, 0.06655078125, 0.13611622619628908, 0.07056998443603515, 0.06917324829101562, 0.06794854736328125, 0.06668595123291016, 0.06652210998535156, 0.06653132629394531, 0.07042457580566407, 0.06774578857421874, 0.06652518463134766, 0.06661017608642578, 0.0666398696899414, 0.06940774536132813, 0.06636032104492187, 0.06636851501464844, 0.06650367736816407, 0.0666982421875, 0.0665159683227539, 0.066482177734375, 0.06641664123535156, 0.06645967864990235, 0.06650262451171875, 0.06670540618896484, 0.06677811431884766, 0.06670130920410157, 0.06645145416259765, 0.06652620697021484, 0.0663900146484375, 0.06650265502929688, 0.06645043182373046, 0.06659481811523438, 0.06679859161376953, 0.066482177734375, 0.06553907012939453, 0.06548377227783203, 0.06635929870605468, 0.06668697357177734, 0.06652518463134766, 0.06660710144042968, 0.06654873657226562, 0.06659686279296875, 0.06671257781982422, 0.06620365142822265, 0.06652108764648437, 0.06648941040039062, 0.06691734313964844, 0.06694806671142578, 0.06660198211669922, 0.06635417938232421, 0.06640128326416016, 0.06742425537109376, 0.06839398193359375, 0.0672368621826172, 0.06713241577148438, 0.0665374755859375, 0.06651084899902343, 0.06669721221923829, 0.06633369445800781, 0.06733106994628907, 0.06648729705810547, 0.06676480102539062, 0.0665354232788086, 0.06643609619140625, 0.13382041931152344, 0.06618112182617188, 0.06658150482177734, 0.06657132720947266, 0.06658758544921875, 0.06603775787353515, 0.06631219482421875, 0.06646784210205078, 0.06547353363037109, 0.06636444854736329, 0.06653846740722656, 0.06676070404052735, 0.06665318298339844, 0.06649446105957031, 0.06582886505126953, 0.06637574768066407, 0.06642374420166015, 0.06561382293701172, 0.06546125030517579, 0.06586166381835938, 0.0665692138671875, 0.06643094635009765, 0.0664668197631836, 0.0675440673828125, 0.06666854095458985, 0.06640640258789063, 0.06630502319335937, 0.06612480163574219, 0.0659415054321289, 0.066334716796875, 0.06634188842773438, 0.06767616271972657, 0.06686412811279296, 0.06645657348632812, 0.06647602844238282, 0.0667658233642578, 0.06632857513427734, 0.06625177764892579, 0.06642073822021484, 0.06636953735351563, 0.06594662475585937, 0.06601216125488281, 0.06793113708496094, 0.06635529327392578, 0.06632028961181641, 0.06631526184082032, 0.06635826873779296, 0.06651904296875, 0.06628761291503907, 0.0664637451171875, 0.06650879669189454, 0.06630604553222656, 0.06646685028076171, 0.06639612579345704, 0.066337890625, 0.06784812927246094, 0.06673097229003906, 0.06635622406005859, 0.06623542022705078, 0.06633980560302734, 0.06623744201660156, 0.0666060791015625, 0.06640537261962891, 0.13640704345703125, 0.06601628875732422, 0.06654972839355469, 0.06644429016113282, 0.06559539031982421, 0.06644739532470703, 0.06631318664550781, 0.06618931579589844, 0.06641049957275391, 0.06650675201416016, 0.06629068756103515, 0.06636032104492187, 0.06620467376708984, 0.06633984375, 0.06621907043457032, 0.06632953643798828, 0.06634291076660156, 0.06617401885986328, 0.06627117156982422, 0.06572134399414062, 0.06618112182617188, 0.06620569610595703, 0.06638188934326172, 0.06625580596923829, 0.06950707244873047, 0.06655487823486328, 0.06633881378173828, 0.06654061126708985, 0.06651283264160156, 0.06921318054199219, 0.06728294372558594, 0.06643711853027344, 0.06644429016113282, 0.06649753570556641, 0.06938214111328125, 0.0687267837524414, 0.06678118133544922, 0.06627839660644531, 0.06636441802978515, 0.06641049957275391, 0.06624460601806641, 0.06616678619384765, 0.06570598602294922, 0.06629277038574219, 0.06611043548583985, 0.06651392364501953, 0.06651392364501953, 0.06640128326416016, 0.06634393310546875, 0.06730751800537109, 0.066766845703125, 0.066408447265625, 0.06633164978027344, 0.06650265502929688, 0.06649958038330078, 0.06630604553222656, 0.06895513916015625, 0.06670642852783203, 0.06653644561767579, 0.06616780853271484, 0.06627327728271484, 0.0662845458984375, 0.06631116485595703, 0.13373234558105468, 0.06632243347167968, 0.0662138900756836, 0.06649343872070312, 0.06643199920654297, 0.06639615631103515, 0.0661585922241211, 0.06606540679931641, 0.06623436737060547, 0.06681088256835938, 0.06625484466552735, 0.06642176055908203, 0.0663746566772461, 0.06614940643310546, 0.06621897888183594, 0.06626815795898437, 0.06623641967773437, 0.0662138900756836, 0.06632959747314453, 0.06634803009033204, 0.06642585754394531, 0.06620262145996093, 0.06614527893066406, 0.06625791931152344, 0.06537625885009765, 0.06640435028076172, 0.06638489532470702, 0.06592716979980469, 0.06627430725097656, 0.06615245056152344, 0.06628761291503907, 0.06624358367919922, 0.06631731414794922, 0.06630092620849609, 0.06642793273925782, 0.06627222442626954, 0.06638182067871094, 0.06643711853027344, 0.06650163269042969, 0.06548480224609375, 0.06562611389160156, 0.06633478546142578, 0.06667565155029297, 0.06656819152832032, 0.06645145416259765, 0.0664668197631836, 0.0675758056640625, 0.06663788604736329, 0.06633465576171875, 0.06616268920898437, 0.06645452880859375, 0.06623846435546875, 0.06627225494384766, 0.06649549102783203, 0.06626406097412109, 0.06617510223388672, 0.06556249237060546, 0.06671052551269531, 0.06643199920654297, 0.06642892456054687, 0.06631423950195313, 0.06610841369628906, 0.06626611328125, 0.1338419189453125, 0.06657331085205079, 0.06636646270751953, 0.06645452880859375, 0.0663900146484375, 0.06627635192871094, 0.06639513397216797, 0.06668185424804687, 0.06627327728271484, 0.06637264251708984, 0.06891824340820313, 0.06696550750732422, 0.06633984375, 0.06797926330566406, 0.0687831039428711, 0.06694809722900391, 0.06647602844238282, 0.06626918029785156, 0.06629580688476562, 0.06661734771728516, 0.06656614685058594, 0.06644429016113282, 0.06629376220703125, 0.06638489532470702, 0.0664668197631836, 0.0666081314086914, 0.06669725036621094, 0.06644322967529297, 0.06620262145996093, 0.06643302154541016, 0.06729625701904297, 0.06910975646972656, 0.06757478332519531, 0.06641766357421874, 0.067557373046875, 0.06629478454589843, 0.06769766235351563, 0.06783590698242188, 0.06699827575683594, 0.0677027816772461, 0.06629888153076172, 0.06682418823242188, 0.06694707489013672, 0.06769664001464844, 0.06708531188964843, 0.06848921966552735, 0.07025564575195313, 0.06716413116455078, 0.06607667541503906, 0.0672204818725586, 0.06623538970947265, 0.06721638488769531, 0.06643917083740235, 0.0663050537109375, 0.06612886047363281, 0.06697267150878906, 0.06905753326416016, 0.06652729797363281, 0.06728800201416016, 0.06760655975341796, 0.0676761245727539, 0.06641254425048829, 0.06634393310546875]",tokens/s,14.780586271885719,,,True 8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5078.016,7927.758848,0.0,7272.923136,7060.974592,s,10,0.9193428421020508,0.09193428421020508,0.0020703056821796606,0.09164439773559571,0.09345917282104492,0.0955564182281494,0.097234214553833,"[0.0976536636352539, 0.09299311828613281, 0.09171622467041016, 0.09169945526123047, 0.09158934020996094, 0.09077107238769531, 0.09051526641845703, 0.09031244659423829, 0.09183126068115234, 0.09026099395751953]",tokens/s,2784.5977395621358,kWh,1.0966354896747527e-06,6.009029849186238e-07,2.9771345833521222e-06,4.674673057945498e-06,tokens/kWh,54763188.10464813,MB,5078.016,7927.758848,0.0,7272.923136,7223.648256,s,10,56.4374970703125,5.64374970703125,0.025440362452633102,5.641560302734375,5.662229150390624,5.685672436523437,5.704427065429687,"[5.70911572265625, 5.6436025390625, 5.64651611328125, 5.6130107421875, 5.6518125, 5.62650537109375, 5.63951806640625, 5.62623779296875, 5.62415869140625, 5.65701953125]",tokens/s,11.16279127713825,kWh,6.724247649741273e-05,3.685338217553787e-05,0.00013783575584464715,0.00024193161451759775,tokens/kWh,260404.16472902708,,s,629,57.16333369445806,0.09087970380676948,0.010943706355990221,0.08912079620361328,0.09142804260253906,0.09189273681640625,0.1802723162841797,"[0.09198796844482422, 0.09496985626220703, 0.09575321960449219, 0.09295257568359375, 0.09338681793212891, 0.09012012481689453, 0.0918466567993164, 0.0885401611328125, 0.08884941101074219, 0.08878797149658203, 0.08882176208496094, 0.08894258880615234, 0.08863744354248047, 0.08898252868652344, 0.08875724792480469, 0.09054515075683593, 0.08888217926025391, 0.08969318389892578, 0.08897535705566406, 0.0889139175415039, 0.08916582489013672, 0.09334067535400391, 0.09186406707763672, 0.09162751770019531, 0.08921087646484376, 0.09045606231689453, 0.09031680297851563, 0.09157529449462891, 0.08901222229003906, 0.09192243194580078, 0.09189990234375, 0.09120460510253907, 0.09186918640136718, 0.08868153381347656, 0.08903263854980469, 0.08896409606933593, 0.0913623046875, 0.09012633514404297, 0.08905522918701173, 0.08991232299804687, 0.09252146911621094, 0.09055129241943359, 0.08913932800292969, 0.09095155334472656, 0.09164492797851563, 0.09026764678955078, 0.08923750305175782, 0.08868556976318359, 0.09111961364746093, 0.09178316497802734, 0.09191423797607422, 0.08963174438476562, 0.09307449340820312, 0.09228492736816406, 0.09179641723632813, 0.09145958709716796, 0.09172991943359375, 0.0898887710571289, 0.08863641357421875, 0.08982733154296875, 0.08900198364257812, 0.08885453033447266, 0.18209791564941405, 0.0888770523071289, 0.09123532867431641, 0.08944947052001953, 0.09113702392578125, 0.0894228515625, 0.09140735626220703, 0.09189683532714844, 0.08926617431640625, 0.09138687896728516, 0.09014784240722656, 0.08850739288330078, 0.08847052764892578, 0.09091584014892579, 0.09125273895263672, 0.09069261169433594, 0.08918220520019532, 0.08879718780517579, 0.08883916473388671, 0.08862924957275391, 0.08829132843017579, 0.08785919952392578, 0.08833229064941406, 0.0886824951171875, 0.08833843231201172, 0.08874393463134765, 0.08817664337158203, 0.0887531509399414, 0.0885749740600586, 0.08883814239501953, 0.09111449432373046, 0.0897791976928711, 0.08860774230957032, 0.0887388153076172, 0.08859648132324219, 0.08963174438476562, 0.09132236480712891, 0.09133465576171874, 0.09088719940185547, 0.09043555450439453, 0.09143705749511719, 0.09079500579833985, 0.08854937744140624, 0.0886845474243164, 0.08852787017822265, 0.08857190704345703, 0.09133465576171874, 0.09138790130615235, 0.08885657501220703, 0.08873062133789063, 0.08912076568603515, 0.0912721939086914, 0.09131520080566406, 0.08866918182373047, 0.09001369476318359, 0.088855712890625, 0.08914518737792969, 0.08802822113037109, 0.08970950317382813, 0.09133875274658203, 0.0885749740600586, 0.08872550201416016, 0.08834150695800781, 0.1812183074951172, 0.08886067199707032, 0.08862310028076172, 0.08888326263427734, 0.088723388671875, 0.09115853118896484, 0.0885063705444336, 0.08847257232666016, 0.0888084487915039, 0.08876134490966797, 0.0889815673828125, 0.08888211059570313, 0.08853401947021484, 0.08874803161621093, 0.08883712005615234, 0.08895692443847657, 0.08852070617675781, 0.08846540832519531, 0.08868863677978515, 0.08894258880615234, 0.08861497497558593, 0.0886578598022461, 0.08844287872314453, 0.08891494750976563, 0.09112371063232422, 0.09148313903808594, 0.08861695861816406, 0.0882155532836914, 0.08842137908935546, 0.08869171142578125, 0.0887224349975586, 0.08894361877441406, 0.08870195007324219, 0.09153638458251953, 0.09185689544677735, 0.09003110504150391, 0.091863037109375, 0.09367961883544922, 0.08965119934082032, 0.0902850570678711, 0.08868863677978515, 0.09048883056640625, 0.09169715118408203, 0.08863744354248047, 0.08868556976318359, 0.09163980865478516, 0.0887562255859375, 0.09192243194580078, 0.09132236480712891, 0.08898355102539063, 0.09142578887939454, 0.08880332946777343, 0.09134899139404297, 0.08894361877441406, 0.09103052520751953, 0.09174425506591796, 0.09169203186035156, 0.08860470581054687, 0.09179952239990234, 0.0892907485961914, 0.08899890899658203, 0.08922726440429687, 0.0885002212524414, 0.18290585327148437, 0.08871731567382812, 0.08859750366210938, 0.08881561279296875, 0.08818585968017578, 0.08871218872070312, 0.08843366241455078, 0.08894566345214844, 0.0885770263671875, 0.0884316177368164, 0.08833843231201172, 0.08866611480712891, 0.0884295654296875, 0.08862207794189453, 0.08825138854980469, 0.08867123413085938, 0.08853094482421875, 0.08870604705810547, 0.088342529296875, 0.09014476776123047, 0.08992768096923828, 0.08852377319335937, 0.08966860961914062, 0.08963686370849609, 0.08971366119384766, 0.09131110382080078, 0.08860671997070313, 0.08868045043945312, 0.08916582489013672, 0.08872652435302734, 0.08975263977050782, 0.08944633483886719, 0.08919039916992187, 0.08873471832275391, 0.08829542541503907, 0.08879206085205078, 0.09014476776123047, 0.09077145385742187, 0.08850125122070313, 0.08900096130371093, 0.08844083404541016, 0.08911571502685547, 0.08895891571044921, 0.08888114929199219, 0.08886988830566406, 0.0888647689819336, 0.0883773422241211, 0.09087078094482422, 0.08854732513427735, 0.09121279907226562, 0.08916275024414062, 0.08863641357421875, 0.09062400054931641, 0.09226854705810547, 0.08870502471923829, 0.08844902038574219, 0.08850431823730469, 0.08874803161621093, 0.0887162857055664, 0.08922726440429687, 0.08883507537841796, 0.08859449768066406, 0.0884940185546875, 0.18186956787109376, 0.08917094421386719, 0.0889722900390625, 0.0888985595703125, 0.08845619201660156, 0.09133465576171874, 0.09011199951171875, 0.08851660919189454, 0.08873062133789063, 0.08874495697021484, 0.0907489242553711, 0.089753662109375, 0.08873567962646485, 0.08992460632324219, 0.09075609588623047, 0.08961331176757813, 0.08971366119384766, 0.09375129699707031, 0.08930099487304688, 0.08881049346923828, 0.09114729309082031, 0.091304931640625, 0.09007820892333984, 0.09216716766357422, 0.08855449676513671, 0.09081958770751954, 0.08827597045898437, 0.08852992248535156, 0.08911257934570313, 0.08848793792724609, 0.0886845474243164, 0.08855142211914062, 0.08939622497558594, 0.08994509124755859, 0.09036697387695312, 0.08867533111572265, 0.08849100494384765, 0.08860364532470703, 0.08846233367919921, 0.09105510711669922, 0.09121382141113281, 0.09108889770507812, 0.09103059387207031, 0.08937567901611328, 0.08927232360839844, 0.09018265533447266, 0.08851046752929688, 0.09191731262207031, 0.09071923065185547, 0.08875110626220703, 0.08994303894042968, 0.08982425689697265, 0.08953855895996093, 0.09102848052978516, 0.08899075317382812, 0.09025737762451172, 0.08975769805908203, 0.08875212860107422, 0.09052877044677735, 0.08891187286376953, 0.0885432357788086, 0.08876748657226563, 0.0899092788696289, 0.18033251953125, 0.0892293472290039, 0.0900679702758789, 0.09088508605957031, 0.08860262298583985, 0.08922624206542969, 0.0898314208984375, 0.08868761444091797, 0.08870809936523437, 0.08953343963623046, 0.08943411254882813, 0.08938086700439453, 0.08875520324707031, 0.09063219451904297, 0.08944230651855468, 0.08909516906738281, 0.08976691436767578, 0.09101824188232421, 0.08952114868164063, 0.08896511840820312, 0.08868147277832031, 0.08865280151367187, 0.08877062225341797, 0.08915347290039062, 0.08851967620849609, 0.08905216217041016, 0.08950886535644531, 0.08966349029541015, 0.08893030548095703, 0.08888832092285157, 0.0891883544921875, 0.09128959655761719, 0.08851046752929688, 0.090102783203125, 0.08989491271972656, 0.09148825836181641, 0.09119436645507813, 0.09125785827636719, 0.08842034912109376, 0.08864358520507812, 0.08853708648681641, 0.08863436889648438, 0.08954879760742188, 0.08819916534423829, 0.08834355163574219, 0.08851763153076173, 0.08974131011962891, 0.089206787109375, 0.08885759735107422, 0.08846444702148437, 0.0907754898071289, 0.08876544189453126, 0.08820121765136718, 0.09035775756835937, 0.08839167785644532, 0.08948223876953125, 0.0887930908203125, 0.08916275024414062, 0.08850534057617188, 0.08922624206542969, 0.0886087646484375, 0.089491455078125, 0.08858009338378907, 0.18126336669921875, 0.08926207733154297, 0.08861901092529297, 0.09108889770507812, 0.09087283325195313, 0.09055948638916016, 0.09032601928710937, 0.08985088348388671, 0.09031884765625, 0.08912079620361328, 0.08957231903076172, 0.08956928253173828, 0.08912281799316406, 0.08976076507568359, 0.09171865844726562, 0.08874086761474609, 0.08974540710449219, 0.09024307250976563, 0.08861901092529297, 0.08898662567138672, 0.08911769866943359, 0.08884326171875, 0.08913100433349609, 0.08951910400390625, 0.08884633636474609, 0.0885186538696289, 0.08868966674804687, 0.08906240081787109, 0.08936140441894531, 0.08888832092285157, 0.08861798095703124, 0.08989798736572266, 0.0914554901123047, 0.08895897674560548, 0.0905697250366211, 0.08873267364501954, 0.08852787017822265, 0.09112786865234375, 0.09076531219482421, 0.0896234893798828, 0.09100902557373047, 0.08863129425048828, 0.08971981048583984, 0.08889958190917968, 0.08851455688476563, 0.08974642944335938, 0.08853708648681641, 0.08896514892578125, 0.08871113586425781, 0.08902143859863282, 0.08970649719238281, 0.08960717010498047, 0.08840499114990234, 0.09012838745117188, 0.08880025482177735, 0.08862515258789062, 0.08975667572021484, 0.09105101013183593, 0.08866611480712891, 0.08865894317626953, 0.08859033966064453, 0.09094246673583985, 0.08978125, 0.18011750793457032, 0.08888527679443359, 0.08881353759765626, 0.08953036499023438, 0.09079705810546874, 0.08851763153076173, 0.08862310028076172, 0.08965837097167968, 0.0883978271484375, 0.08874188995361328, 0.08847257232666016, 0.089312255859375, 0.08846438598632812, 0.08858214569091796, 0.088774658203125, 0.0894361572265625, 0.08843878173828125, 0.08861491394042968, 0.09013145446777343, 0.08854630279541016, 0.09029837036132812, 0.08869171142578125, 0.08823705291748046, 0.08870604705810547, 0.08849305725097656, 0.08915558624267578, 0.09091891479492188, 0.08824832153320313, 0.0883220443725586, 0.08855757141113281, 0.08895692443847657, 0.08924671936035156, 0.08842752075195312, 0.08820633697509765, 0.08964915466308594, 0.08885043334960938, 0.08939315032958985, 0.08979558563232422, 0.08822886657714844, 0.08856781005859375, 0.08878797149658203, 0.08914227294921875, 0.08863539123535157, 0.08976383972167969, 0.08820941162109375, 0.08963174438476562, 0.09521049499511719, 0.09198182678222656, 0.0916941146850586, 0.09183843231201172, 0.09042431640625, 0.08955187225341797, 0.09125682830810547, 0.08954879760742188, 0.08979251098632812, 0.08875007629394531, 0.08952320098876954, 0.08875724792480469, 0.08847666931152344, 0.08991948699951172, 0.08863744354248047, 0.08862617492675781, 0.08855654144287109, 0.18134425354003905, 0.09056256103515625, 0.09024102020263672, 0.08992153930664062, 0.08949964904785156, 0.09115033721923828, 0.08940338897705079, 0.08934809875488281, 0.08932454681396484, 0.0891514892578125, 0.0887562255859375, 0.08910131072998047, 0.09059123229980469, 0.08881664276123047, 0.089059326171875, 0.08913715362548828, 0.09188658905029297, 0.08903782653808594, 0.08873887634277344, 0.0887664031982422, 0.09054822540283203, 0.09168895721435547, 0.08874495697021484, 0.0899788818359375, 0.09113600158691407, 0.08919245147705078, 0.0884510726928711, 0.08905420684814454, 0.08889548492431641, 0.08882380676269531, 0.08773529815673828, 0.08978739166259765, 0.08876544189453126, 0.08896614074707031, 0.08842752075195312, 0.08874803161621093, 0.0887357406616211, 0.08949247741699219, 0.08863334655761719, 0.088880126953125, 0.08948633575439453, 0.0885923843383789, 0.08877875518798828, 0.08879513549804688, 0.08993689727783204, 0.08867225646972657, 0.08891494750976563, 0.08962355041503907, 0.09070591735839843, 0.0887224349975586, 0.08883612823486328, 0.08884732818603516, 0.08841318511962891, 0.08881970977783203, 0.08894464111328125, 0.08959900665283203, 0.0885882568359375, 0.08975462341308593, 0.08872857666015625, 0.08892108917236329, 0.08840611267089844, 0.08888822174072265, 0.08857292938232422, 0.17932902526855468, 0.08851455688476563, 0.09112064361572265, 0.08872345733642578, 0.09154764556884766, 0.08945970916748047, 0.08955391693115235, 0.08845209503173829, 0.08918323516845703, 0.09133773040771484, 0.08935321807861328, 0.08874803161621093, 0.08824524688720703, 0.08866815948486328, 0.08867533111572265, 0.08857292938232422, 0.08832614135742188, 0.08981094360351563, 0.08865996551513672, 0.08876547241210937, 0.0885544662475586, 0.09043456268310547, 0.08831394958496094, 0.08908585357666016, 0.09061689758300781, 0.09059014129638672, 0.08951398468017578, 0.08875110626220703, 0.08992460632324219, 0.09138380432128906, 0.08841318511962891, 0.09146470642089843, 0.09209241485595702, 0.08882278442382813, 0.09044889831542968, 0.08977101135253907, 0.09302220916748047, 0.0907479019165039, 0.08930508422851563, 0.08930918121337891, 0.08904908752441407, 0.08915968322753906, 0.08841932678222657, 0.08981708526611328, 0.0899205093383789, 0.09108582305908203, 0.08898560333251954, 0.09154867553710938, 0.09122819519042968, 0.09126601409912109, 0.088595458984375, 0.0896153564453125, 0.08889036560058594, 0.090355712890625, 0.09144115447998047, 0.09134899139404297, 0.09115545654296875, 0.09033625793457031, 0.09124454498291015, 0.09016217803955077, 0.08989593505859375, 0.08864256286621094, 0.08954265594482422]",tokens/s,11.003556989206556,,,True -8bit-bnb-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +8bit-bnb-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,bnb,0.0,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch