diff --git "a/perf-df-unquantized-1xA10.csv" "b/perf-df-unquantized-1xA10.csv" --- "a/perf-df-unquantized-1xA10.csv" +++ "b/perf-df-unquantized-1xA10.csv" @@ -201,7 +201,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1276.993536,4276.617216,0.0,3632.267264,3408.132096,s,10,0.23940703964233398,0.0239407039642334,0.0003155813142295948,0.023847408294677734,0.02398629741668701,0.024434684658050536,0.02479339445114136,"[0.024883071899414064, 0.023886655807495116, 0.023772607803344725, 0.023813919067382814, 0.02384752082824707, 0.02381644821166992, 0.023857664108276368, 0.023819103240966796, 0.02386275291442871, 0.0238472957611084]",tokens/s,10693.085733086853,kWh,2.839145672499624e-07,1.5557130726576287e-07,1.4168026825865747e-06,1.8562885571023002e-06,tokens/kWh,137909593.32293713,MB,1276.993536,4278.714368,0.0,3632.267264,3474.14784,s,10,12.64006494140625,1.264006494140625,0.012221827649893114,1.2609061279296876,1.2754159545898438,1.2825172424316407,1.2881982727050783,"[1.2896185302734375, 1.2725411376953124, 1.2540771484375, 1.255328369140625, 1.2603609619140625, 1.2468157958984376, 1.273837890625, 1.25360791015625, 1.2614512939453124, 1.2724259033203125]",tokens/s,49.8415160776785,kWh,1.4844918173666186e-05,8.134760645182703e-06,3.992906372741369e-05,6.290874254626257e-05,tokens/kWh,1001450.6322975749,,s,629,12.83316420936584,0.020402486819341568,0.002910507754959045,0.01984000015258789,0.020758118057250977,0.020815462493896486,0.043985878601074224,"[0.022249343872070313, 0.020015104293823242, 0.019563520431518554, 0.020341760635375978, 0.02031718444824219, 0.020344831466674804, 0.01965260887145996, 0.019603456497192383, 0.019529727935791014, 0.019575807571411134, 0.019569664001464843, 0.01967001533508301, 0.019629056930541993, 0.01966592025756836, 0.0198656005859375, 0.01978982353210449, 0.020781055450439453, 0.020579328536987306, 0.020864000320434572, 0.020596736907958983, 0.019772415161132813, 0.02064588737487793, 0.01966592025756836, 0.01960960006713867, 0.01984921646118164, 0.019788799285888673, 0.019784704208374023, 0.020091903686523437, 0.020955135345458984, 0.02109644889831543, 0.020981760025024415, 0.020800512313842775, 0.020802560806274413, 0.020745216369628908, 0.020780031204223632, 0.02069811248779297, 0.02072166442871094, 0.020743200302124024, 0.020793312072753905, 0.020742143630981445, 0.020668415069580077, 0.020701183319091796, 0.020806655883789063, 0.020725759506225586, 0.020715520858764647, 0.020736000061035157, 0.020770816802978515, 0.02070425605773926, 0.020690944671630858, 0.020757503509521484, 0.020763647079467772, 0.020786176681518553, 0.020700159072875975, 0.020742143630981445, 0.02104012870788574, 0.020815872192382814, 0.020814847946166993, 0.020690944671630858, 0.020790271759033203, 0.020765695571899414, 0.02070528030395508, 0.020719615936279297, 0.045159423828125, 0.020746240615844725, 0.02069606399536133, 0.020760576248168947, 0.02088140869140625, 0.020649984359741212, 0.020766719818115235, 0.020776960372924806, 0.02083635139465332, 0.020743167877197266, 0.020767744064331056, 0.02075257682800293, 0.02082899284362793, 0.020764671325683593, 0.02082713508605957, 0.020784128189086915, 0.02081279945373535, 0.02082918357849121, 0.02044108772277832, 0.020757503509521484, 0.020771839141845702, 0.020753408432006838, 0.020695039749145508, 0.020798528671264648, 0.02076460838317871, 0.020791296005249024, 0.019911680221557617, 0.019793920516967774, 0.019819520950317384, 0.019708927154541016, 0.0198604793548584, 0.020007936477661133, 0.01967513656616211, 0.01958822441101074, 0.019557247161865233, 0.019571712493896484, 0.019773439407348634, 0.019877887725830077, 0.01982054328918457, 0.019959808349609375, 0.019786752700805665, 0.0198656005859375, 0.019986431121826173, 0.01989836883544922, 0.019817472457885742, 0.01984819221496582, 0.019805183410644533, 0.01986355209350586, 0.019812416076660157, 0.01974060821533203, 0.019687423706054686, 0.019771392822265626, 0.019572736740112305, 0.019579904556274414, 0.01965363121032715, 0.019728384017944335, 0.019862527847290038, 0.019963903427124022, 0.01982054328918457, 0.019800064086914062, 0.01985740852355957, 0.01984000015258789, 0.01984511947631836, 0.04400537490844727, 0.0198154239654541, 0.01982464027404785, 0.019786752700805665, 0.019811328887939454, 0.01980723190307617, 0.019862527847290038, 0.01985843276977539, 0.019908607482910155, 0.01984921646118164, 0.01982361602783203, 0.01977446365356445, 0.01985945510864258, 0.019891199111938478, 0.01987276840209961, 0.019899391174316407, 0.019770368576049805, 0.01982156753540039, 0.019805183410644533, 0.019834880828857423, 0.019784704208374023, 0.01988403129577637, 0.0198656005859375, 0.019901439666748046, 0.019812351226806642, 0.01984000015258789, 0.01984819221496582, 0.019813375473022463, 0.01984000015258789, 0.01983692741394043, 0.019786752700805665, 0.019970048904418947, 0.019812351226806642, 0.01986662483215332, 0.01986355209350586, 0.019931135177612306, 0.019812351226806642, 0.01982156753540039, 0.019843072891235353, 0.02002841567993164, 0.01993123245239258, 0.01985424041748047, 0.01985228729248047, 0.01983078384399414, 0.01983795166015625, 0.01982259178161621, 0.01983078384399414, 0.019914751052856446, 0.019779584884643556, 0.01983283233642578, 0.01985945510864258, 0.0197857608795166, 0.019861503601074217, 0.019742687225341796, 0.01984921646118164, 0.020208639144897463, 0.02145484733581543, 0.02071347236633301, 0.020761600494384767, 0.019785728454589844, 0.019737600326538086, 0.0198175048828125, 0.019766239166259764, 0.04402179336547852, 0.019805152893066405, 0.01984921646118164, 0.01983692741394043, 0.019772415161132813, 0.019819520950317384, 0.019757055282592775, 0.019852352142333985, 0.019805152893066405, 0.019734495162963866, 0.01985024070739746, 0.019764223098754884, 0.019696640014648437, 0.01979801559448242, 0.019714048385620117, 0.019794944763183595, 0.019885055541992186, 0.019758079528808595, 0.019788799285888673, 0.019794944763183595, 0.019801088333129883, 0.019805183410644533, 0.019766271591186522, 0.019842048645019532, 0.019788799285888673, 0.019826688766479493, 0.01982259178161621, 0.01983180809020996, 0.019755008697509766, 0.019811391830444336, 0.01973040008544922, 0.019833824157714845, 0.019803136825561524, 0.019776512145996093, 0.019736576080322265, 0.01981955146789551, 0.019859424591064455, 0.019834880828857423, 0.01979804801940918, 0.019857376098632813, 0.01979084777832031, 0.01982975959777832, 0.019781631469726564, 0.01983283233642578, 0.01964339256286621, 0.01979084777832031, 0.022689792633056642, 0.020799488067626954, 0.020588544845581053, 0.020601856231689454, 0.020365312576293947, 0.020551776885986327, 0.01959516716003418, 0.0195666561126709, 0.01953068733215332, 0.01961369514465332, 0.01965670394897461, 0.02033158493041992, 0.020458431243896485, 0.020420608520507814, 0.020403200149536133, 0.019582975387573243, 0.019564544677734375, 0.043911167144775394, 0.019620895385742188, 0.019941343307495116, 0.020537343978881836, 0.020418560028076172, 0.020332544326782227, 0.020331520080566406, 0.02084556770324707, 0.021608448028564452, 0.020546560287475587, 0.020493312835693358, 0.020426752090454102, 0.02029363250732422, 0.020287488937377928, 0.020379648208618165, 0.019677343368530272, 0.019605472564697267, 0.019594112396240235, 0.019537919998168944, 0.019654687881469728, 0.019588064193725586, 0.019498016357421873, 0.019522560119628905, 0.019624927520751952, 0.019579904556274414, 0.01979084777832031, 0.019851264953613282, 0.02057318305969238, 0.020373504638671876, 0.020539392471313478, 0.02059878349304199, 0.020582399368286132, 0.02066739273071289, 0.02064793586730957, 0.019825664520263672, 0.019756032943725587, 0.019766271591186522, 0.019744768142700195, 0.01978265571594238, 0.019847263336181642, 0.019767200469970703, 0.01983078384399414, 0.01983283233642578, 0.01982464027404785, 0.01982054328918457, 0.019787776947021486, 0.01987174415588379, 0.019927040100097656, 0.019817472457885742, 0.019725311279296876, 0.01985740852355957, 0.01981439971923828, 0.019811328887939454, 0.019812416076660157, 0.019889087677001954, 0.019825664520263672, 0.019861503601074217, 0.01982361602783203, 0.019802112579345704, 0.019766271591186522, 0.019804159164428712, 0.019800064086914062, 0.019944448471069336, 0.04413951873779297, 0.0198287353515625, 0.01985536003112793, 0.01982464027404785, 0.019894271850585937, 0.01990656089782715, 0.01986867141723633, 0.019862527847290038, 0.019851264953613282, 0.019772415161132813, 0.019817472457885742, 0.0198154239654541, 0.01982259178161621, 0.020009983062744142, 0.0198604793548584, 0.01986662483215332, 0.019750911712646483, 0.01979497528076172, 0.01986147117614746, 0.019714048385620117, 0.0196495361328125, 0.019610624313354492, 0.019483648300170898, 0.01962188720703125, 0.019525632858276368, 0.019851264953613282, 0.019862560272216796, 0.019749919891357423, 0.01975391960144043, 0.019927040100097656, 0.01982054328918457, 0.019801088333129883, 0.019776512145996093, 0.019765247344970704, 0.019876863479614256, 0.019769344329833984, 0.019768320083618163, 0.019588096618652344, 0.01967513656616211, 0.019644416809082032, 0.019477664947509767, 0.01954595184326172, 0.019803136825561524, 0.019951616287231445, 0.019761152267456054, 0.019745792388916016, 0.019893247604370116, 0.01980723190307617, 0.019903488159179687, 0.019723264694213868, 0.019779584884643556, 0.019783679962158202, 0.019759103775024413, 0.019817472457885742, 0.01982156753540039, 0.01984102439880371, 0.019804159164428712, 0.01979801559448242, 0.01984716796875, 0.019795967102050782, 0.01982054328918457, 0.019805183410644533, 0.019748895645141602, 0.04502729415893555, 0.021325824737548828, 0.01983283233642578, 0.020133888244628906, 0.020566015243530272, 0.020590591430664062, 0.020588544845581053, 0.020586496353149415, 0.02062848091125488, 0.020549631118774413, 0.019776512145996093, 0.02006425666809082, 0.02060697555541992, 0.020547584533691408, 0.020593664169311524, 0.019853311538696287, 0.019723264694213868, 0.019799072265625, 0.019860448837280272, 0.019984384536743165, 0.019766271591186522, 0.019819520950317384, 0.01982361602783203, 0.019769344329833984, 0.019810304641723633, 0.019779584884643556, 0.01979084777832031, 0.019793920516967774, 0.020478975296020507, 0.02063564872741699, 0.020567039489746093, 0.020756479263305663, 0.02060697555541992, 0.020619264602661135, 0.02061311912536621, 0.020594688415527345, 0.020682752609252928, 0.020731903076171874, 0.01982464027404785, 0.01985638427734375, 0.01980723190307617, 0.0198154239654541, 0.0198154239654541, 0.019862527847290038, 0.020271104812622072, 0.02064896011352539, 0.019817472457885742, 0.019869695663452147, 0.019742719650268553, 0.02017791938781738, 0.019908607482910155, 0.01989017677307129, 0.019908607482910155, 0.020599807739257812, 0.020782079696655274, 0.02072166442871094, 0.02062950325012207, 0.020734975814819336, 0.02065407943725586, 0.019766271591186522, 0.019779584884643556, 0.019804159164428712, 0.019747840881347657, 0.04393574523925781, 0.01978985595703125, 0.019811296463012697, 0.019843072891235353, 0.019775487899780272, 0.020611072540283205, 0.020874271392822264, 0.021965791702270508, 0.020808704376220705, 0.020626432418823244, 0.01979903984069824, 0.019760128021240234, 0.019763200759887696, 0.020376575469970702, 0.02089779281616211, 0.019968000411987305, 0.019762176513671875, 0.019777536392211914, 0.019748863220214845, 0.019779584884643556, 0.01964646339416504, 0.020568063735961914, 0.020274175643920898, 0.019663871765136717, 0.01945497512817383, 0.01962598419189453, 0.01960857582092285, 0.01944166374206543, 0.019507200241088866, 0.019579904556274414, 0.019503103256225587, 0.019562496185302734, 0.019533824920654298, 0.019501056671142578, 0.019543039321899415, 0.020529151916503906, 0.020356096267700196, 0.020539392471313478, 0.020371456146240235, 0.020392959594726562, 0.02025164794921875, 0.020478975296020507, 0.02026188850402832, 0.019555328369140625, 0.01945907211303711, 0.01946419143676758, 0.019521535873413084, 0.019578880310058593, 0.019582975387573243, 0.019764320373535156, 0.019569568634033203, 0.01946531105041504, 0.019436447143554688, 0.019516416549682617, 0.019520511627197267, 0.019545087814331053, 0.019494911193847657, 0.019677183151245118, 0.019548160552978516, 0.019747840881347657, 0.01948569679260254, 0.01979084777832031, 0.019869695663452147, 0.044025856018066405, 0.019795007705688476, 0.01993516731262207, 0.01983795166015625, 0.02005504035949707, 0.020685823440551757, 0.020585472106933594, 0.020668415069580077, 0.020575231552124023, 0.02068070411682129, 0.02066329574584961, 0.02059775924682617, 0.02051584053039551, 0.02081279945373535, 0.02035916709899902, 0.02065100860595703, 0.020549631118774413, 0.020612096786499022, 0.020593664169311524, 0.020736000061035157, 0.020728832244873048, 0.01984716796875, 0.019811328887939454, 0.0197806396484375, 0.019838943481445312, 0.019773439407348634, 0.019714048385620117, 0.019803136825561524, 0.019752960205078125, 0.019732479095458985, 0.01978265571594238, 0.0196997127532959, 0.019718143463134767, 0.01979903984069824, 0.0198154239654541, 0.019779584884643556, 0.019765247344970704, 0.019768320083618163, 0.01978265571594238, 0.019875904083251954, 0.019820480346679686, 0.01983692741394043, 0.01981644821166992, 0.01982771110534668, 0.019808256149291992, 0.01982464027404785, 0.019739648818969727, 0.019901439666748046, 0.01977446365356445, 0.020001792907714845, 0.019922943115234376, 0.019770368576049805, 0.01980723190307617, 0.019795967102050782, 0.019776512145996093, 0.01984819221496582, 0.0198287353515625, 0.019784704208374023, 0.01985843276977539, 0.019765247344970704, 0.019794944763183595, 0.019701759338378907, 0.019833856582641602, 0.04410367965698242, 0.01983590316772461, 0.019777536392211914, 0.019811328887939454, 0.01982054328918457, 0.019720191955566405, 0.01982156753540039, 0.02002841567993164, 0.019811328887939454, 0.01982156753540039, 0.019927040100097656, 0.01990656089782715, 0.019902463912963866, 0.02149990463256836, 0.020798463821411133, 0.020840448379516603, 0.020807680130004884, 0.020690944671630858, 0.019755008697509766, 0.019883007049560548, 0.02003660774230957, 0.019705856323242187, 0.019738624572753907, 0.019751935958862304, 0.019610624313354492, 0.020142080307006836, 0.02047091293334961, 0.020378496170043944, 0.01985228729248047, 0.019851264953613282, 0.02017791938781738, 0.020739072799682616, 0.01980723190307617, 0.019818496704101563, 0.019795967102050782, 0.020386816024780274, 0.02063155174255371, 0.020718591690063477, 0.02122854423522949, 0.020789247512817383, 0.020694015502929687, 0.02069811248779297, 0.020584447860717774, 0.020677631378173827, 0.02068377685546875, 0.020736000061035157, 0.020658176422119142, 0.01978265571594238, 0.019891231536865235, 0.019780576705932616, 0.019938304901123048, 0.019636224746704102, 0.01984409523010254, 0.019760128021240234, 0.019738624572753907, 0.01965977668762207, 0.02002339172363281, 0.020392864227294923, 0.020510719299316405, 0.020593664169311524, 0.02062745666503906, 0.020585535049438476, 0.020603872299194335]",tokens/s,49.01363293870625,,,,, -float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -238,7 +238,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 768.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -585,7 +585,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -612,7 +612,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -768,7 +768,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 216.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -811,7 +811,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GP ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1308.086272,3848.798208,0.0,3202.351104,3105.82784,s,10,0.2721215019226074,0.027212150192260743,0.0005506786455814327,0.02701697540283203,0.027374995040893555,0.02811365737915039,0.02870458724975586,"[0.028852319717407225, 0.027018911361694337, 0.027015039443969727, 0.02704364776611328, 0.02697238349914551, 0.027210847854614258, 0.026960575103759765, 0.027014848709106445, 0.027043231964111326, 0.026989696502685546]",tokens/s,9407.562364285624,kWh,3.2189005781151185e-07,1.7636263548326822e-07,1.3816753408828266e-06,1.8799280341776065e-06,tokens/kWh,136175425.51940814,MB,1308.086272,3848.798208,0.0,3202.351104,3188.152832,s,10,12.426950073242189,1.242695007324219,0.008562259006632902,1.2417922973632813,1.2499968627929687,1.2562812072753906,1.261308682861328,"[1.2625655517578125, 1.2339195556640625, 1.240325927734375, 1.2399776611328126, 1.2432586669921875, 1.2456578369140625, 1.248600341796875, 1.2466455078125, 1.2324085693359375, 1.2335904541015625]",tokens/s,50.69626869721809,kWh,1.458196097775694e-05,7.990656032580708e-06,3.975949646091645e-05,6.233211347125409e-05,tokens/kWh,1010714.9668373417,,s,629,12.648431613922098,0.020108794298763305,0.003280659308354322,0.01960038375854492,0.02013347854614258,0.02032803840637207,0.04704813217163086,"[0.0214783992767334, 0.021115903854370118, 0.020468736648559572, 0.020332544326782227, 0.020254720687866212, 0.020347904205322266, 0.020740095138549804, 0.020281343460083007, 0.020090879440307616, 0.020147199630737304, 0.020297727584838866, 0.020521984100341797, 0.020248575210571287, 0.020165632247924805, 0.020307968139648438, 0.020371456146240235, 0.020257791519165038, 0.020256767272949217, 0.02067148780822754, 0.020296703338623046, 0.020556800842285155, 0.02008883285522461, 0.020115455627441405, 0.02023219108581543, 0.020761600494384767, 0.020948991775512696, 0.020332544326782227, 0.020207616806030275, 0.020082687377929686, 0.019530752182006835, 0.020118528366088868, 0.020188159942626953, 0.019778560638427735, 0.019520511627197267, 0.019552255630493166, 0.019744768142700195, 0.01953388786315918, 0.019320768356323244, 0.01942118453979492, 0.019566591262817384, 0.019548160552978516, 0.020168703079223634, 0.02017791938781738, 0.01982464027404785, 0.019555328369140625, 0.019459104537963866, 0.019784671783447266, 0.019486719131469727, 0.019564544677734375, 0.019489791870117186, 0.019513343811035155, 0.020002815246582033, 0.019563520431518554, 0.019555360794067382, 0.019584991455078124, 0.019551231384277345, 0.019553279876708983, 0.019561471939086913, 0.019582975387573243, 0.019479551315307618, 0.019614719390869142, 0.019607551574707033, 0.04705587387084961, 0.019513343811035155, 0.019746816635131836, 0.01989836883544922, 0.019544063568115236, 0.019530752182006835, 0.019528703689575197, 0.019704832077026366, 0.019781631469726564, 0.019590143203735352, 0.01949286460876465, 0.019535871505737306, 0.019504127502441407, 0.019942399978637695, 0.02003660774230957, 0.01942937660217285, 0.019337215423583985, 0.019526655197143555, 0.019591167449951173, 0.019478527069091797, 0.019520511627197267, 0.019573759078979493, 0.019545087814331053, 0.019629056930541993, 0.019511295318603517, 0.019347455978393553, 0.019390464782714844, 0.019573759078979493, 0.019532800674438477, 0.019591167449951173, 0.019467264175415038, 0.01963212776184082, 0.019481599807739256, 0.019610624313354492, 0.019531776428222656, 0.019518463134765626, 0.019507200241088866, 0.019497983932495116, 0.020135936737060548, 0.019662912368774415, 0.019652544021606447, 0.01960038375854492, 0.019510271072387696, 0.019577856063842772, 0.019582975387573243, 0.019571712493896484, 0.019501056671142578, 0.01966182327270508, 0.019482624053955077, 0.01964339256286621, 0.01950931167602539, 0.019533760070800783, 0.019546112060546874, 0.01946009635925293, 0.0196177921295166, 0.01960140800476074, 0.019551231384277345, 0.019531776428222656, 0.019500032424926757, 0.019696640014648437, 0.019522560119628905, 0.019571712493896484, 0.019483648300170898, 0.04720435333251953, 0.019708927154541016, 0.019531776428222656, 0.019674144744873046, 0.01991881561279297, 0.019949567794799804, 0.019680255889892577, 0.01961881637573242, 0.019574783325195313, 0.019575807571411134, 0.019936256408691407, 0.019579904556274414, 0.019325952529907226, 0.01958502388000488, 0.019467264175415038, 0.019826688766479493, 0.01947750473022461, 0.019554304122924804, 0.01972844886779785, 0.019658687591552735, 0.0196495361328125, 0.019556352615356445, 0.019562496185302734, 0.0196495361328125, 0.019896320343017578, 0.01962495994567871, 0.019536895751953123, 0.019524608612060547, 0.019512351989746095, 0.019564512252807618, 0.019567615509033204, 0.0196495361328125, 0.01961676788330078, 0.01964339256286621, 0.019512319564819337, 0.01965465545654297, 0.019556352615356445, 0.01964646339416504, 0.019570688247680663, 0.01967411231994629, 0.019721216201782226, 0.020162559509277343, 0.021012479782104493, 0.020463615417480468, 0.01987379264831543, 0.019793920516967774, 0.019688447952270507, 0.019631103515625, 0.019312639236450196, 0.01960038375854492, 0.019574783325195313, 0.01959321594238281, 0.019612672805786133, 0.02006425666809082, 0.019579904556274414, 0.019673088073730468, 0.01968435287475586, 0.01980620765686035, 0.01949286460876465, 0.019517440795898438, 0.01959321594238281, 0.01968639945983887, 0.019525632858276368, 0.047126529693603515, 0.019548160552978516, 0.019597312927246095, 0.01957072067260742, 0.01950716781616211, 0.019537919998168944, 0.01959321594238281, 0.01959017562866211, 0.019514335632324218, 0.019490816116333007, 0.019470367431640625, 0.019619808197021485, 0.020369407653808593, 0.019561471939086913, 0.019414016723632813, 0.01965977668762207, 0.019550207138061524, 0.019851264953613282, 0.019603456497192383, 0.01944780731201172, 0.01958502388000488, 0.01949286460876465, 0.019631103515625, 0.019580928802490235, 0.019559423446655275, 0.019532800674438477, 0.019587072372436523, 0.019602432250976562, 0.019506175994873046, 0.019540992736816407, 0.019739648818969727, 0.020460544586181642, 0.020523008346557618, 0.020206592559814454, 0.02006937599182129, 0.019494911193847657, 0.019673088073730468, 0.019796991348266603, 0.019545087814331053, 0.019687423706054686, 0.01957683181762695, 0.01946316719055176, 0.020048896789550782, 0.019998720169067383, 0.019566591262817384, 0.019553279876708983, 0.019595264434814453, 0.019981311798095702, 0.019703807830810546, 0.019510271072387696, 0.019519487380981446, 0.019604480743408204, 0.020039680480957032, 0.01948569679260254, 0.0196177921295166, 0.019603456497192383, 0.01960038375854492, 0.019706880569458008, 0.01963417625427246, 0.019569664001464843, 0.01965772819519043, 0.019547136306762695, 0.02001203155517578, 0.04705791854858398, 0.019501056671142578, 0.019579904556274414, 0.019899391174316407, 0.02085273551940918, 0.019495935440063478, 0.019583999633789064, 0.019533824920654298, 0.019936256408691407, 0.019564544677734375, 0.019487808227539063, 0.01955833625793457, 0.019543039321899415, 0.019687423706054686, 0.019579904556274414, 0.019589120864868165, 0.01966592025756836, 0.0195020809173584, 0.019769344329833984, 0.019532800674438477, 0.019834911346435547, 0.01963209533691406, 0.019553279876708983, 0.019591167449951173, 0.019927040100097656, 0.019802112579345704, 0.019354623794555666, 0.019812351226806642, 0.020262975692749024, 0.020265920639038086, 0.019543039321899415, 0.01965056037902832, 0.019544095993041993, 0.020062175750732422, 0.02002022361755371, 0.019474431991577147, 0.019793920516967774, 0.02002739143371582, 0.02007859230041504, 0.019887104034423828, 0.019458047866821288, 0.019372032165527343, 0.01948467254638672, 0.019753984451293945, 0.019979263305664064, 0.019772415161132813, 0.019478527069091797, 0.01990656089782715, 0.019960832595825196, 0.01984102439880371, 0.020082687377929686, 0.01945907211303711, 0.0196177921295166, 0.019561471939086913, 0.019969024658203126, 0.019900480270385743, 0.01972115135192871, 0.01957683181762695, 0.020057088851928712, 0.019987455368041994, 0.019350528717041016, 0.01943756866455078, 0.019497983932495116, 0.04791398239135742, 0.019306495666503908, 0.019522560119628905, 0.019486783981323242, 0.01995871925354004, 0.020290559768676757, 0.019602432250976562, 0.019720191955566405, 0.019519487380981446, 0.019732479095458985, 0.02008576011657715, 0.01984716796875, 0.019508224487304687, 0.01968332862854004, 0.019586048126220702, 0.019945472717285157, 0.019948575973510744, 0.019479520797729494, 0.01984409523010254, 0.019537919998168944, 0.020123647689819335, 0.01984716796875, 0.019532800674438477, 0.019693567276000978, 0.01948876762390137, 0.019705856323242187, 0.01940787124633789, 0.01933523178100586, 0.019267520904541015, 0.019405824661254883, 0.019562496185302734, 0.019513343811035155, 0.01984819221496582, 0.01963417625427246, 0.02021683120727539, 0.0195020809173584, 0.019573759078979493, 0.019467264175415038, 0.019751935958862304, 0.019628032684326172, 0.01999564743041992, 0.02027008056640625, 0.0198154239654541, 0.019915775299072267, 0.019567615509033204, 0.019978239059448243, 0.02005401611328125, 0.019697664260864257, 0.019737600326538086, 0.02006118392944336, 0.020262912750244142, 0.020132863998413086, 0.01986764717102051, 0.01957683181762695, 0.020105215072631837, 0.020214784622192384, 0.020076543807983398, 0.020174848556518556, 0.019588096618652344, 0.01966182327270508, 0.019596288681030274, 0.019786752700805665, 0.019753984451293945, 0.04746649551391602, 0.020074495315551756, 0.02006630325317383, 0.01997209548950195, 0.01970278358459473, 0.019775487899780272, 0.019787776947021486, 0.020006912231445313, 0.019499008178710937, 0.02007142448425293, 0.02012057685852051, 0.019869695663452147, 0.01969152069091797, 0.019536895751953123, 0.019599359512329103, 0.019495935440063478, 0.019990528106689453, 0.019756032943725587, 0.02000486373901367, 0.019885055541992186, 0.020008960723876954, 0.02048204803466797, 0.020092927932739257, 0.01979084777832031, 0.01943756866455078, 0.019885055541992186, 0.019917823791503905, 0.01946419143676758, 0.01971609687805176, 0.01966182327270508, 0.020093952178955078, 0.01986867141723633, 0.019903488159179687, 0.01948057556152344, 0.019508224487304687, 0.01964134407043457, 0.019869695663452147, 0.02001408004760742, 0.01947750473022461, 0.019923967361450197, 0.020040704727172853, 0.02007142448425293, 0.02009702491760254, 0.019556352615356445, 0.019701759338378907, 0.019831840515136718, 0.020070367813110352, 0.019932159423828123, 0.019677183151245118, 0.0194334716796875, 0.019739648818969727, 0.02003558349609375, 0.019894271850585937, 0.019978239059448243, 0.019526655197143555, 0.019579904556274414, 0.01940377616882324, 0.01946316719055176, 0.019610624313354492, 0.019721216201782226, 0.019676191329956055, 0.019994592666625975, 0.019943424224853516, 0.04702822494506836, 0.019796991348266603, 0.019961856842041017, 0.020136959075927736, 0.02007244873046875, 0.019705856323242187, 0.019548160552978516, 0.020307968139648438, 0.019979263305664064, 0.01965875244140625, 0.01986867141723633, 0.019701759338378907, 0.020057151794433594, 0.01975904083251953, 0.020101119995117187, 0.020321279525756835, 0.02008064079284668, 0.020084768295288085, 0.0200980167388916, 0.020139007568359374, 0.01984819221496582, 0.01945088005065918, 0.01942527961730957, 0.01963932800292969, 0.01988400077819824, 0.021565439224243164, 0.020529151916503906, 0.019919872283935547, 0.019610624313354492, 0.019523584365844726, 0.019543039321899415, 0.01961881637573242, 0.01945497512817383, 0.020152320861816408, 0.019743743896484374, 0.019430400848388672, 0.019594240188598632, 0.019571712493896484, 0.019615743637084963, 0.019588096618652344, 0.019580928802490235, 0.019532800674438477, 0.019619903564453124, 0.019587007522583008, 0.01967411231994629, 0.019563520431518554, 0.01967513656616211, 0.019580928802490235, 0.019591167449951173, 0.019664896011352538, 0.019513343811035155, 0.019558399200439454, 0.019499008178710937, 0.019779584884643556, 0.019554304122924804, 0.019771392822265626, 0.019704832077026366, 0.020051967620849608, 0.019916799545288084, 0.019560447692871095, 0.01957379150390625, 0.01950716781616211, 0.019520511627197267, 0.04714393615722656, 0.019533824920654298, 0.01955027198791504, 0.01954400062561035, 0.01960140800476074, 0.01944576072692871, 0.019504127502441407, 0.019555328369140625, 0.01958502388000488, 0.019537919998168944, 0.019573759078979493, 0.01966592025756836, 0.0194467830657959, 0.019559423446655275, 0.019527679443359376, 0.019587072372436523, 0.01967616081237793, 0.019588096618652344, 0.019497983932495116, 0.01963724708557129, 0.01984511947631836, 0.019544063568115236, 0.019521535873413084, 0.01960960006713867, 0.019751935958862304, 0.0196628475189209, 0.019539968490600586, 0.01947238349914551, 0.019570688247680663, 0.019517440795898438, 0.019538944244384765, 0.019566591262817384, 0.019479551315307618, 0.01943961524963379, 0.019491840362548828, 0.01957683181762695, 0.01946112060546875, 0.019505151748657225, 0.0194969596862793, 0.019350528717041016, 0.01939455986022949, 0.01940377616882324, 0.019507200241088866, 0.019475456237792968, 0.019910655975341796, 0.019482624053955077, 0.019573759078979493, 0.019516416549682617, 0.019495935440063478, 0.019611648559570313, 0.019534847259521485, 0.019558399200439454, 0.019574783325195313, 0.019513343811035155, 0.01945702362060547, 0.019781631469726564, 0.019611648559570313, 0.019568639755249022, 0.019534847259521485, 0.01946828842163086, 0.019672063827514647, 0.019598335266113282, 0.019536895751953123, 0.047026176452636716, 0.01962188720703125, 0.019534847259521485, 0.019543039321899415, 0.019545087814331053, 0.019581951141357423, 0.019606592178344727, 0.019630016326904295, 0.019583999633789064, 0.01945292854309082, 0.01960550308227539, 0.01944371223449707, 0.019479551315307618, 0.019487743377685548, 0.01960857582092285, 0.019479551315307618, 0.019523584365844726, 0.019555328369140625, 0.019311616897583008, 0.019518463134765626, 0.01929113578796387, 0.01948876762390137, 0.019575807571411134, 0.019664896011352538, 0.019534847259521485, 0.019474431991577147, 0.019438592910766602, 0.01958502388000488, 0.019550207138061524, 0.01960044860839844, 0.01957267189025879, 0.019570688247680663, 0.02003046417236328, 0.01971609687805176, 0.0195645751953125, 0.019586015701293945, 0.019544095993041993, 0.019584991455078124, 0.019524608612060547, 0.01963827133178711, 0.0194652156829834, 0.01962188720703125, 0.019876863479614256, 0.019526655197143555, 0.019575807571411134, 0.019375104904174805, 0.019583999633789064, 0.01969152069091797, 0.019614719390869142, 0.01960038375854492, 0.019529727935791014, 0.019521535873413084, 0.019422208786010742, 0.01945088005065918, 0.019358720779418945, 0.019322879791259767, 0.020797439575195312, 0.01995676803588867, 0.01929417610168457, 0.019313663482666017, 0.019559423446655275, 0.019591167449951173, 0.01986662483215332]",tokens/s,49.729485773371316,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1315.725312,6797.39392,0.0,6150.946816,5989.292032,s,10,0.4270291862487793,0.04270291862487793,0.0014589085893857718,0.042228033065795895,0.0427962459564209,0.044935611915588376,0.046647104682922366,"[0.04707497787475586, 0.04216880035400391, 0.042320831298828125, 0.04214396667480469, 0.04211011123657227, 0.04223136138916016, 0.04222470474243164, 0.04216451263427735, 0.04228224182128906, 0.04230767822265625]",tokens/s,5994.90639618387,kWh,5.012936111392275e-07,2.746845990513667e-07,2.6209189988051623e-06,3.3968972089957568e-06,tokens/kWh,75362892.73695235,MB,1315.725312,6797.39392,0.0,6150.946816,6115.763712,s,10,16.56944055175781,1.6569440551757815,0.022761192137948933,1.65922216796875,1.6744492065429688,1.6892416931152343,1.7010756823730468,"[1.6711619873046875, 1.652427978515625, 1.64447705078125, 1.6204678955078125, 1.641165771484375, 1.666016357421875, 1.631757080078125, 1.6691287841796876, 1.7040341796875, 1.668803466796875]",tokens/s,38.02180272967422,kWh,1.9981832101997237e-05,1.095027865999512e-05,6.577088571459451e-05,9.670299647658683e-05,tokens/kWh,651479.2953210421,,s,629,16.920467454910273,0.026900584189046544,0.005115355429982433,0.02596659278869629,0.027069645309448245,0.02723287010192871,0.0685291537475586,"[0.02694655990600586, 0.026814464569091798, 0.02693120002746582, 0.02692915153503418, 0.026862592697143556, 0.026812416076660156, 0.02688102340698242, 0.02592563247680664, 0.02592870330810547, 0.025809919357299805, 0.02593484878540039, 0.025873407363891602, 0.025997312545776367, 0.02588057518005371, 0.0272936954498291, 0.026875904083251953, 0.02679193687438965, 0.027023359298706053, 0.027052032470703126, 0.026602495193481446, 0.027470848083496095, 0.02771251106262207, 0.02629631996154785, 0.0259420166015625, 0.025964544296264647, 0.02590105628967285, 0.02587648010253906, 0.025852928161621092, 0.02630451202392578, 0.0271779842376709, 0.027049983978271484, 0.026813440322875977, 0.02689023971557617, 0.026818559646606444, 0.025903104782104492, 0.025821184158325194, 0.026123296737670897, 0.02581808090209961, 0.025996288299560546, 0.02593791961669922, 0.02590105628967285, 0.025847808837890625, 0.02595737648010254, 0.025977855682373048, 0.02593280029296875, 0.025874431610107423, 0.025995264053344725, 0.02590822410583496, 0.026630144119262695, 0.026810367584228514, 0.026786815643310546, 0.026857471466064452, 0.026884096145629883, 0.026830848693847657, 0.0269434871673584, 0.026879999160766603, 0.02689023971557617, 0.02688921546936035, 0.026798080444335938, 0.02688719940185547, 0.02674787139892578, 0.026824703216552736, 0.06859264373779297, 0.02592153549194336, 0.02589388847351074, 0.02591334342956543, 0.025802751541137696, 0.025866239547729493, 0.025819135665893556, 0.025935871124267578, 0.025808895111083984, 0.025874431610107423, 0.025775104522705077, 0.025842687606811524, 0.025791488647460937, 0.025793535232543945, 0.025857023239135742, 0.026072063446044923, 0.02570751953125, 0.025594879150390625, 0.025646080017089845, 0.02594713592529297, 0.025881599426269532, 0.02592153549194336, 0.02588979148864746, 0.026398719787597655, 0.028299264907836914, 0.027207679748535156, 0.026867712020874023, 0.02592153549194336, 0.02590822410583496, 0.025850879669189454, 0.02594099235534668, 0.025985023498535157, 0.025841663360595703, 0.02588057518005371, 0.02592870330810547, 0.025830400466918944, 0.025862144470214843, 0.02595840072631836, 0.026690559387207033, 0.02690559959411621, 0.02676531219482422, 0.026825727462768553, 0.026770431518554686, 0.02589286422729492, 0.025810943603515626, 0.025931776046752928, 0.026738687515258788, 0.026828800201416016, 0.02674073600769043, 0.027128896713256835, 0.02589689636230469, 0.025825279235839844, 0.025780223846435548, 0.025834495544433594, 0.025785343170166015, 0.026878976821899415, 0.026855424880981447, 0.02687286376953125, 0.02719024085998535, 0.026784767150878908, 0.026712064743041993, 0.027131904602050783, 0.02714112091064453, 0.06956755065917969, 0.026673088073730467, 0.026673152923583986, 0.026679296493530274, 0.02636595153808594, 0.025854976654052734, 0.02572083282470703, 0.02594713592529297, 0.025746431350708008, 0.025843711853027345, 0.025828351974487306, 0.025785375595092773, 0.026262496948242186, 0.026654720306396484, 0.0267007999420166, 0.02692095947265625, 0.02674892807006836, 0.026465343475341796, 0.0262225284576416, 0.026558464050292968, 0.02673459243774414, 0.026690559387207033, 0.026746879577636717, 0.026636287689208983, 0.025838592529296874, 0.025783296585083007, 0.025797632217407225, 0.025766912460327147, 0.025956352233886718, 0.026884096145629883, 0.026847232818603517, 0.025993215560913087, 0.025831424713134765, 0.025790464401245116, 0.025757696151733397, 0.025540607452392578, 0.025511936187744142, 0.02573107147216797, 0.027313152313232423, 0.027067392349243165, 0.026814464569091798, 0.02681548881530762, 0.0266844482421875, 0.02558051109313965, 0.025776128768920898, 0.025883647918701173, 0.025811967849731447, 0.0259368953704834, 0.02578124809265137, 0.02555084800720215, 0.025475072860717773, 0.025480192184448244, 0.02549247932434082, 0.025638912200927736, 0.025520128250122072, 0.025618431091308593, 0.02549558448791504, 0.025612255096435547, 0.025795583724975587, 0.025830400466918944, 0.02572800064086914, 0.02557542419433594, 0.025623552322387694, 0.06838988494873047, 0.02593894386291504, 0.025807872772216797, 0.02592153549194336, 0.025564159393310547, 0.02567475128173828, 0.025779199600219727, 0.025879552841186523, 0.025808895111083984, 0.025842687606811524, 0.026496000289916992, 0.02671615982055664, 0.02671513557434082, 0.026298368453979492, 0.025491455078125, 0.02565020751953125, 0.02555695915222168, 0.025552896499633788, 0.02549247932434082, 0.02555801582336426, 0.025520128250122072, 0.025539583206176757, 0.02557542419433594, 0.025554943084716796, 0.02548121643066406, 0.02549862480163574, 0.025531391143798828, 0.025579519271850586, 0.025758720397949218, 0.02576383972167969, 0.02571878433227539, 0.025899007797241212, 0.02573516845703125, 0.025779232025146485, 0.02552009582519531, 0.026063871383666993, 0.025487360000610353, 0.025463808059692384, 0.025524255752563476, 0.025710559844970702, 0.025479167938232423, 0.02551910400390625, 0.02553036880493164, 0.025615392684936525, 0.025540576934814454, 0.025581567764282227, 0.025540607452392578, 0.02589286422729492, 0.025589759826660157, 0.025580543518066406, 0.025620479583740235, 0.025840639114379883, 0.025511936187744142, 0.025604095458984375, 0.0255416316986084, 0.025734144210815428, 0.025619455337524414, 0.025705472946166992, 0.025825279235839844, 0.0259051513671875, 0.025787391662597657, 0.02570035171508789, 0.02551807975769043, 0.0685506591796875, 0.02550476837158203, 0.025664512634277343, 0.025569280624389647, 0.02552115249633789, 0.025617408752441406, 0.02568806457519531, 0.02564409637451172, 0.02559993553161621, 0.025769983291625977, 0.025787391662597657, 0.025817087173461914, 0.026792959213256837, 0.026770431518554686, 0.025584640502929686, 0.026158079147338868, 0.025973760604858398, 0.025651199340820312, 0.025845760345458983, 0.025684991836547853, 0.025614336013793947, 0.025851903915405275, 0.025858112335205078, 0.025695167541503906, 0.02548940849304199, 0.02575667190551758, 0.02579769515991211, 0.02584876823425293, 0.025842687606811524, 0.02593075180053711, 0.026785791397094725, 0.026895360946655275, 0.026755071640014647, 0.02675916862487793, 0.026730495452880858, 0.026785791397094725, 0.026771455764770507, 0.02673459243774414, 0.026764287948608398, 0.02597068786621094, 0.025878528594970703, 0.025866239547729493, 0.025971712112426756, 0.025826303482055665, 0.025846784591674804, 0.025832447052001953, 0.025819135665893556, 0.025928768157958984, 0.025840576171875, 0.025788415908813478, 0.025597951889038087, 0.025581567764282227, 0.02565225601196289, 0.025778144836425782, 0.025803775787353517, 0.02593791961669922, 0.025857023239135742, 0.02697216033935547, 0.026828800201416016, 0.026635263442993166, 0.026833919525146483, 0.026702848434448243, 0.0267007999420166, 0.06968013000488281, 0.02679193687438965, 0.026746879577636717, 0.0267007999420166, 0.02667830467224121, 0.026762208938598632, 0.026888191223144533, 0.025874431610107423, 0.026598400115966796, 0.026867712020874023, 0.02682368087768555, 0.026849279403686522, 0.026829824447631836, 0.026720256805419923, 0.026764287948608398, 0.026878976821899415, 0.02668339157104492, 0.026630144119262695, 0.026707008361816408, 0.025767871856689453, 0.025794559478759766, 0.025866239547729493, 0.025852928161621092, 0.026612735748291014, 0.02651238441467285, 0.02636288070678711, 0.02571161651611328, 0.02650931167602539, 0.02667519950866699, 0.027024383544921874, 0.027020288467407227, 0.02590822410583496, 0.025855039596557616, 0.026547136306762694, 0.02659744071960449, 0.02680415916442871, 0.02674176025390625, 0.025858047485351563, 0.025808895111083984, 0.02613043212890625, 0.026385408401489258, 0.025630720138549806, 0.025635839462280274, 0.025806848526000976, 0.025865215301513672, 0.025806848526000976, 0.026462207794189452, 0.02707967948913574, 0.0267007999420166, 0.02590412712097168, 0.025796607971191408, 0.025805824279785155, 0.026727424621582032, 0.0267509765625, 0.02648575973510742, 0.025935871124267578, 0.025750528335571288, 0.027829248428344725, 0.028071935653686524, 0.02731007957458496, 0.026452991485595705, 0.02589388847351074, 0.02574950408935547, 0.068663330078125, 0.02576380729675293, 0.02587552070617676, 0.027017152786254883, 0.02676736068725586, 0.025822208404541015, 0.025842687606811524, 0.025911296844482422, 0.025780223846435548, 0.026096639633178712, 0.02676940727233887, 0.025817087173461914, 0.025810943603515626, 0.025772064208984376, 0.02580271911621094, 0.025891839981079103, 0.025775104522705077, 0.025798688888549803, 0.026255392074584962, 0.026112960815429687, 0.025795583724975587, 0.025827327728271485, 0.025834495544433594, 0.02593075180053711, 0.02571878433227539, 0.02592051124572754, 0.02574950408935547, 0.025787391662597657, 0.025769983291625977, 0.025845760345458983, 0.025784320831298828, 0.025859071731567384, 0.025789440155029295, 0.025832447052001953, 0.025784320831298828, 0.025821184158325194, 0.025965568542480468, 0.025835519790649415, 0.025863168716430664, 0.026294271469116212, 0.025779199600219727, 0.025794559478759766, 0.025886720657348632, 0.025823232650756835, 0.025765888214111327, 0.025849855422973633, 0.02588057518005371, 0.025879552841186523, 0.02572697639465332, 0.025853952407836913, 0.02576486396789551, 0.025772031784057618, 0.0257259521484375, 0.025819135665893556, 0.025766912460327147, 0.02591948890686035, 0.025829376220703124, 0.025890815734863282, 0.025848831176757812, 0.025783296585083007, 0.025791488647460937, 0.02573516845703125, 0.025790464401245116, 0.0684738540649414, 0.025795583724975587, 0.025782272338867186, 0.025835519790649415, 0.025785343170166015, 0.025772031784057618, 0.025669631958007814, 0.025894912719726562, 0.02574336051940918, 0.02572083282470703, 0.025800703048706054, 0.026085376739501953, 0.02576076889038086, 0.025757696151733397, 0.025823232650756835, 0.025960447311401368, 0.025867263793945314, 0.02608332824707031, 0.025776128768920898, 0.02590105628967285, 0.02678169631958008, 0.027848703384399414, 0.026252288818359375, 0.026055679321289063, 0.025830432891845702, 0.025797599792480468, 0.026068992614746093, 0.026242048263549804, 0.02652774429321289, 0.026717216491699218, 0.02640176010131836, 0.026194976806640624, 0.025761760711669923, 0.02657587242126465, 0.02678374481201172, 0.02670796775817871, 0.026853376388549805, 0.02675814437866211, 0.026731552124023436, 0.027158496856689453, 0.02679091262817383, 0.02673971176147461, 0.026711040496826172, 0.026714111328125, 0.02750464057922363, 0.027048959732055664, 0.02673766326904297, 0.02688921546936035, 0.02673971176147461, 0.026814464569091798, 0.02717184066772461, 0.027020288467407227, 0.027054079055786134, 0.027064319610595702, 0.027115520477294923, 0.027223039627075195, 0.02699776077270508, 0.027555871963500976, 0.027730911254882813, 0.027246591567993163, 0.027041791915893554, 0.02712166404724121, 0.026928127288818358, 0.06986956787109375, 0.027034624099731445, 0.026528799057006835, 0.02694243240356445, 0.027089920043945313, 0.02709503936767578, 0.027173887252807616, 0.027380735397338866, 0.027069440841674806, 0.026999807357788085, 0.02698956871032715, 0.027031551361083983, 0.026945600509643553, 0.027084735870361327, 0.027026432037353516, 0.02714521598815918, 0.02696499252319336, 0.026987520217895508, 0.02695475196838379, 0.026640384674072266, 0.026605567932128905, 0.027001855850219726, 0.027014144897460936, 0.027193344116210938, 0.027025407791137695, 0.02654719924926758, 0.02729471969604492, 0.02718617630004883, 0.027045888900756834, 0.027026432037353516, 0.027021312713623048, 0.0272936954498291, 0.028247039794921876, 0.027448320388793947, 0.02716876792907715, 0.0269752311706543, 0.027014144897460936, 0.027099136352539063, 0.02699776077270508, 0.027193344116210938, 0.027032575607299804, 0.027158527374267577, 0.02700492858886719, 0.027066368103027344, 0.027025407791137695, 0.026944511413574217, 0.027060224533081056, 0.027262975692749023, 0.027056127548217773, 0.026944511413574217, 0.02697216033935547, 0.02697932815551758, 0.02657177543640137, 0.026603519439697267, 0.02707046318054199, 0.027052032470703126, 0.02695577621459961, 0.027059200286865235, 0.027107328414916993, 0.0265533447265625, 0.027289600372314454, 0.027166719436645507, 0.027062271118164064, 0.06987881469726563, 0.02695881652832031, 0.026427391052246094, 0.027094079971313478, 0.026931135177612305, 0.02692915153503418, 0.026409984588623047, 0.025798656463623046, 0.025858047485351563, 0.025911296844482422, 0.02611712074279785, 0.02697113609313965, 0.02696601676940918, 0.026587135314941408, 0.026426368713378907, 0.027239423751831054, 0.02711347198486328, 0.026993663787841796, 0.026918912887573244, 0.02691379165649414, 0.026990591049194337, 0.0269117431640625, 0.02646937561035156, 0.026529792785644532, 0.0263505916595459, 0.026673152923583986, 0.026933248519897462, 0.027206655502319335, 0.02697318458557129, 0.02696294403076172, 0.027057151794433593, 0.027284511566162108, 0.02696700859069824, 0.027061248779296877, 0.026994688034057617, 0.026945535659790038, 0.02734182357788086, 0.02709708786010742, 0.0268984317779541, 0.027123712539672853, 0.02696396827697754, 0.02597068786621094, 0.025853952407836913, 0.025862144470214843, 0.025827327728271485, 0.025882623672485353, 0.02590105628967285, 0.026035200119018553, 0.025842687606811524, 0.025894912719726562, 0.025816064834594726, 0.025865215301513672, 0.025833471298217774, 0.025818111419677735, 0.025818111419677735, 0.02596659278869629, 0.025874431610107423, 0.025923583984375, 0.02567475128173828, 0.02553036880493164, 0.025878528594970703, 0.025968639373779297, 0.025935871124267578]",tokens/s,37.173913881289714,,,,, -float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1255,7 +1255,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1299.443712,5222.432768,0.0,4575.985664,4408.404992,s,10,0.2515818881988526,0.02515818881988526,0.0004316832491674847,0.02500271987915039,0.025320387840270995,0.025869905948638916,0.02630952043533325,"[0.026419424057006837, 0.02496672058105469, 0.025182367324829102, 0.02489097595214844, 0.024919551849365236, 0.024940799713134766, 0.024999679565429686, 0.025005760192871093, 0.02505833625793457, 0.025198272705078125]",tokens/s,10175.613269809603,kWh,2.9668963177892634e-07,1.6257227317216066e-07,1.3936045679599308e-06,1.8528664729110178e-06,tokens/kWh,138164300.41923165,MB,1299.443712,5222.432768,0.0,4575.985664,4408.407552,s,10,13.746141967773436,1.3746141967773435,0.012068877601918208,1.3695853881835938,1.3918320556640624,1.3927660034179687,1.3935131616210938,"[1.39162451171875, 1.3634466552734374, 1.368049560546875, 1.3711212158203125, 1.361060302734375, 1.363426025390625, 1.3650748291015624, 1.380155029296875, 1.38848388671875, 1.393699951171875]",tokens/s,45.83104128249054,kWh,1.6011149473747676e-05,8.771077954546157e-06,4.0418401103841396e-05,6.520062853213523e-05,tokens/kWh,966248.3540162406,,s,629,13.947795446395862,0.022174555558657986,0.0030496775723597627,0.02167193603515625,0.022245785522460936,0.02251919288635254,0.046950276489257814,"[0.02288640022277832, 0.02246246337890625, 0.022433792114257813, 0.02268671989440918, 0.022524927139282228, 0.022582271575927734, 0.0224849910736084, 0.02244095993041992, 0.022395904541015626, 0.022575103759765625, 0.023621631622314454, 0.02288332748413086, 0.02253107261657715, 0.022656000137329102, 0.022479936599731444, 0.022508480072021483, 0.02260383987426758, 0.02281465530395508, 0.02229862403869629, 0.022204416275024414, 0.02211123275756836, 0.022552576065063477, 0.02287820816040039, 0.02267852783203125, 0.022592512130737305, 0.022412288665771486, 0.022122495651245116, 0.021751808166503905, 0.021548032760620117, 0.021643264770507813, 0.021566463470458985, 0.022872064590454103, 0.022042623519897463, 0.021644351959228515, 0.02156332778930664, 0.021745664596557617, 0.02162483215332031, 0.021583871841430666, 0.02162380790710449, 0.021588991165161133, 0.021570592880249023, 0.021568479537963866, 0.021530624389648437, 0.021544960021972655, 0.021576704025268553, 0.02151628875732422, 0.021704704284667968, 0.021578752517700195, 0.021510143280029297, 0.021531648635864258, 0.021539840698242187, 0.02209587287902832, 0.022054912567138672, 0.021630975723266603, 0.021630975723266603, 0.021567487716674806, 0.021604352951049805, 0.021615615844726564, 0.021495872497558594, 0.021494720458984377, 0.021606399536132814, 0.021622783660888673, 0.04694732666015625, 0.02149990463256836, 0.021506080627441405, 0.02151420783996582, 0.021542911529541017, 0.021522432327270507, 0.02147020721435547, 0.021599231719970705, 0.02152448081970215, 0.02144972801208496, 0.021313535690307618, 0.021501951217651367, 0.02156550407409668, 0.021486528396606447, 0.021595136642456055, 0.021566463470458985, 0.021576704025268553, 0.021612543106079102, 0.021534719467163087, 0.021538816452026367, 0.021534719467163087, 0.021542911529541017, 0.02156342315673828, 0.021603296279907227, 0.021554176330566405, 0.021531648635864258, 0.021561344146728514, 0.021571584701538086, 0.021606399536132814, 0.021547008514404296, 0.021507072448730468, 0.02161465644836426, 0.021548992156982423, 0.021581823348999024, 0.021605375289916993, 0.021571584701538086, 0.02164121627807617, 0.021559295654296876, 0.021578752517700195, 0.02168320083618164, 0.02164531135559082, 0.02163711929321289, 0.021549055099487305, 0.021629951477050782, 0.021558271408081055, 0.021584896087646483, 0.021611520767211914, 0.02163302421569824, 0.021555200576782226, 0.021561344146728514, 0.02150092887878418, 0.021549055099487305, 0.021548032760620117, 0.021817407608032226, 0.023067583084106447, 0.02248806381225586, 0.022123519897460937, 0.02169241523742676, 0.021929983139038087, 0.021817344665527344, 0.02186444854736328, 0.021950464248657226, 0.021619712829589844, 0.04780441665649414, 0.022130687713623046, 0.02162073516845703, 0.02165452766418457, 0.021505023956298826, 0.021571584701538086, 0.021646335601806642, 0.02231500816345215, 0.021991424560546875, 0.021728256225585937, 0.021742591857910155, 0.021569536209106444, 0.021737472534179687, 0.02166169548034668, 0.021529600143432616, 0.021643264770507813, 0.021597183227539063, 0.02187980842590332, 0.022001663208007814, 0.022384639739990234, 0.02250649642944336, 0.021992448806762696, 0.022120447158813478, 0.02169241523742676, 0.021916671752929686, 0.022411264419555665, 0.021980159759521483, 0.02147327995300293, 0.021531648635864258, 0.021511167526245118, 0.021463071823120117, 0.021717983245849608, 0.021606399536132814, 0.021571584701538086, 0.02150092887878418, 0.021519359588623048, 0.021781503677368166, 0.02147532844543457, 0.02169036865234375, 0.02165555191040039, 0.0214968318939209, 0.02148249626159668, 0.02149068832397461, 0.021522432327270507, 0.021711872100830077, 0.021832704544067383, 0.021578752517700195, 0.02165452766418457, 0.02149171257019043, 0.02152448081970215, 0.021578752517700195, 0.021590015411376954, 0.021534719467163087, 0.021545984268188476, 0.021597183227539063, 0.02167807960510254, 0.021617664337158202, 0.021510143280029297, 0.02161664009094238, 0.021584896087646483, 0.02164838409423828, 0.021595136642456055, 0.02168012809753418, 0.046889984130859375, 0.021547008514404296, 0.021567487716674806, 0.021604352951049805, 0.021572608947753907, 0.021583871841430666, 0.021702655792236326, 0.02168524742126465, 0.022210559844970702, 0.022600704193115235, 0.021905439376831055, 0.021652448654174806, 0.0217262077331543, 0.0218787841796875, 0.021635072708129883, 0.02169343948364258, 0.02164531135559082, 0.02168832015991211, 0.022055936813354493, 0.021537792205810546, 0.021582847595214845, 0.021642240524291992, 0.02153267288208008, 0.021587968826293946, 0.021598207473754884, 0.021595136642456055, 0.021614591598510743, 0.021954559326171876, 0.021597183227539063, 0.021543935775756837, 0.02152448081970215, 0.02185215950012207, 0.021642240524291992, 0.021595136642456055, 0.021554176330566405, 0.021547008514404296, 0.021773311614990236, 0.021793792724609375, 0.023221248626708983, 0.0222873592376709, 0.021884927749633788, 0.021941247940063476, 0.021567487716674806, 0.021533695220947266, 0.021602304458618164, 0.02167091178894043, 0.02150912094116211, 0.021547008514404296, 0.022218751907348632, 0.02185215950012207, 0.021635072708129883, 0.021598207473754884, 0.021592063903808592, 0.02242355155944824, 0.02231603240966797, 0.021988351821899413, 0.021639167785644533, 0.022130687713623046, 0.02162073516845703, 0.021561344146728514, 0.02144256019592285, 0.021734399795532225, 0.021617664337158202, 0.04705587387084961, 0.021587968826293946, 0.021554176330566405, 0.021530624389648437, 0.021789695739746092, 0.0215285758972168, 0.021600255966186522, 0.02147123146057129, 0.021745664596557617, 0.02168832015991211, 0.02162483215332031, 0.021579776763916016, 0.021394432067871092, 0.021385215759277345, 0.021514240264892577, 0.021553152084350585, 0.021517311096191406, 0.021525503158569336, 0.021779455184936524, 0.02148249626159668, 0.021490751266479494, 0.02153571128845215, 0.0215316162109375, 0.021610496520996093, 0.021584896087646483, 0.021557247161865235, 0.021909503936767577, 0.021564416885375977, 0.02143129539489746, 0.021749759674072267, 0.021559295654296876, 0.021574655532836915, 0.02169139289855957, 0.02170163154602051, 0.02168627166748047, 0.021600255966186522, 0.02169036865234375, 0.02149171257019043, 0.02167091178894043, 0.021646335601806642, 0.02166579246520996, 0.021570592880249023, 0.02155926322937012, 0.021577728271484374, 0.021585920333862304, 0.02165350341796875, 0.021742591857910155, 0.02158697509765625, 0.021603296279907227, 0.021541887283325196, 0.021615615844726564, 0.021603328704833984, 0.021566463470458985, 0.0216494083404541, 0.021555200576782226, 0.02189004707336426, 0.021558271408081055, 0.021566463470458985, 0.021549055099487305, 0.021561344146728514, 0.021567487716674806, 0.021535743713378908, 0.021533695220947266, 0.04710297775268555, 0.02148249626159668, 0.02148761558532715, 0.021515264511108398, 0.021536767959594725, 0.021536767959594725, 0.021521408081054686, 0.021559295654296876, 0.02148044776916504, 0.021742591857910155, 0.021617664337158202, 0.021578752517700195, 0.021394432067871092, 0.021424127578735352, 0.021557247161865235, 0.02168627166748047, 0.021574655532836915, 0.021572608947753907, 0.021819391250610352, 0.02167193603515625, 0.021550079345703126, 0.021364736557006835, 0.021547040939331054, 0.02171286392211914, 0.022006784439086914, 0.021372928619384765, 0.021578752517700195, 0.02162380790710449, 0.021766143798828123, 0.021506080627441405, 0.021520351409912108, 0.021539840698242187, 0.02165247917175293, 0.021548032760620117, 0.021588991165161133, 0.021544960021972655, 0.021587968826293946, 0.02186956787109375, 0.021591039657592775, 0.021573631286621094, 0.021603328704833984, 0.021506048202514647, 0.021579776763916016, 0.021536767959594725, 0.021521408081054686, 0.021617664337158202, 0.02187468719482422, 0.02211020851135254, 0.02210406494140625, 0.021610496520996093, 0.021733375549316408, 0.02207027244567871, 0.02162892723083496, 0.021630975723266603, 0.021613567352294923, 0.021758975982666014, 0.021992448806762696, 0.021586944580078125, 0.021575679779052736, 0.02166169548034668, 0.021985279083251954, 0.021586944580078125, 0.021592063903808592, 0.04695142364501953, 0.021402624130249022, 0.021393407821655275, 0.021329919815063478, 0.0216494083404541, 0.021761024475097656, 0.021736448287963867, 0.021651456832885742, 0.02166374397277832, 0.02167091178894043, 0.021582847595214845, 0.021535743713378908, 0.02161664009094238, 0.021603328704833984, 0.0216627197265625, 0.021794815063476563, 0.021598207473754884, 0.021742591857910155, 0.021610496520996093, 0.021943296432495117, 0.021743616104125976, 0.021569536209106444, 0.021557247161865235, 0.02165452766418457, 0.021906431198120118, 0.021873664855957032, 0.02147942352294922, 0.0214517765045166, 0.021725183486938478, 0.02150399971008301, 0.02148249626159668, 0.021580799102783203, 0.02147532844543457, 0.021807104110717773, 0.02186751937866211, 0.02149990463256836, 0.021493759155273438, 0.021574655532836915, 0.022009855270385743, 0.021703680038452147, 0.021573631286621094, 0.021517311096191406, 0.021741567611694337, 0.021621759414672852, 0.021556224822998047, 0.021541887283325196, 0.021651456832885742, 0.021849088668823242, 0.021514240264892577, 0.021526527404785157, 0.02145484733581543, 0.021402624130249022, 0.021586944580078125, 0.021543935775756837, 0.02186956787109375, 0.021578752517700195, 0.02182963180541992, 0.021800960540771484, 0.02211123275756836, 0.02230067253112793, 0.02210406494140625, 0.021929983139038087, 0.021629951477050782, 0.04704051208496094, 0.021749759674072267, 0.02152448081970215, 0.021741567611694337, 0.021511167526245118, 0.02151219177246094, 0.021717023849487305, 0.021791711807250976, 0.02191257667541504, 0.02146611213684082, 0.021710847854614256, 0.021521408081054686, 0.021536767959594725, 0.021530624389648437, 0.02163199996948242, 0.021901311874389647, 0.021593088150024413, 0.021679103851318358, 0.021976064682006836, 0.02185318374633789, 0.02164121627807617, 0.021597183227539063, 0.021615615844726564, 0.021634048461914062, 0.02171392059326172, 0.021596160888671875, 0.021541887283325196, 0.021608448028564452, 0.02173651123046875, 0.021968832015991212, 0.022009855270385743, 0.022152191162109376, 0.02209280014038086, 0.02209791946411133, 0.022026239395141603, 0.022037504196166992, 0.02208665657043457, 0.02208051109313965, 0.02211840057373047, 0.022226943969726562, 0.022162431716918944, 0.02204876708984375, 0.022108160018920898, 0.02208153533935547, 0.022076416015625, 0.022510591506958007, 0.022375423431396483, 0.02227507209777832, 0.02227097511291504, 0.022240255355834963, 0.022125568389892578, 0.022125600814819336, 0.022187999725341797, 0.02207539176940918, 0.022187007904052734, 0.022016000747680665, 0.021898239135742188, 0.02171801567077637, 0.021828607559204103, 0.022218751907348632, 0.022039583206176758, 0.022118368148803712, 0.022023231506347656, 0.047406017303466795, 0.022054912567138672, 0.021940223693847655, 0.021799936294555664, 0.02190336036682129, 0.02205286407470703, 0.021950464248657226, 0.022010879516601564, 0.02186956787109375, 0.022033407211303712, 0.02226380729675293, 0.021755903244018555, 0.022030336380004883, 0.022012992858886717, 0.022056896209716795, 0.02184806442260742, 0.021938175201416017, 0.02186342430114746, 0.02147532844543457, 0.02188287925720215, 0.021884927749633788, 0.02186854362487793, 0.02208051109313965, 0.02209382438659668, 0.022071296691894532, 0.022041599273681642, 0.022137855529785155, 0.022200319290161134, 0.022192127227783204, 0.02202726364135742, 0.02212761688232422, 0.02222489547729492, 0.02206515121459961, 0.02211737632751465, 0.022133760452270508, 0.021832704544067383, 0.021775360107421874, 0.021948415756225585, 0.02209689521789551, 0.021999616622924805, 0.022034431457519533, 0.022161407470703123, 0.022040576934814454, 0.022032384872436524, 0.022171648025512695, 0.022165504455566407, 0.022204416275024414, 0.02225971221923828, 0.022090751647949217, 0.021938175201416017, 0.022208511352539064, 0.022153280258178712, 0.02202412796020508, 0.021946367263793946, 0.02211123275756836, 0.02222390365600586, 0.021888032913208007, 0.022071231842041017, 0.022125568389892578, 0.022154239654541014, 0.022191104888916017, 0.02211942481994629, 0.022167552947998048, 0.04878131103515625, 0.022565887451171874, 0.022185983657836913, 0.02208358383178711, 0.022173696517944336, 0.022063104629516602, 0.022143999099731446, 0.02189004707336426, 0.022107135772705077, 0.02220953559875488, 0.02229350471496582, 0.022143999099731446, 0.022038528442382813, 0.022071296691894532, 0.02209382438659668, 0.02218297576904297, 0.022261695861816408, 0.02203545570373535, 0.022129663467407225, 0.02188287925720215, 0.021986303329467775, 0.02162892723083496, 0.021578752517700195, 0.021572608947753907, 0.02169753646850586, 0.02171801567077637, 0.021803007125854493, 0.02169343948364258, 0.021685312271118164, 0.023793600082397462, 0.022803455352783202, 0.02231500816345215, 0.021643264770507813, 0.02168320083618164, 0.021646335601806642, 0.02204569625854492, 0.02231091117858887, 0.022163455963134765, 0.02205081558227539, 0.022231039047241212, 0.021991424560546875, 0.0222423038482666, 0.022098943710327147, 0.022153215408325197, 0.02210918426513672, 0.022151199340820313, 0.02233955192565918, 0.022163455963134765, 0.022586368560791017, 0.0220579833984375, 0.022339584350585938, 0.0220446720123291, 0.022211584091186523, 0.02209791946411133, 0.022107135772705077, 0.02234060859680176, 0.02211327934265137, 0.022185983657836913, 0.021766143798828123, 0.02215116882324219, 0.0220897274017334, 0.022131711959838866, 0.02209280014038086]",tokens/s,45.09673248488415,,,main,False,False float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1213.095936,3538.419712,0.0,2891.972608,2759.918592,s,10,0.22366297340393068,0.022366297340393065,0.00031044612423922424,0.02226214408874512,0.022430630111694334,0.022861218643188474,0.023205689468383788,"[0.023291807174682617, 0.022239007949829102, 0.022261728286743165, 0.022259519577026366, 0.022193151473999022, 0.022251327514648436, 0.022297216415405274, 0.02226255989074707, 0.022334943771362303, 0.022271711349487303]",tokens/s,11445.792573707287,kWh,2.643789069112664e-07,1.448672797228913e-07,1.3312234114854498e-06,1.7404695981196073e-06,tokens/kWh,147086740.42717025,MB,1213.095936,3538.419712,0.0,2891.972608,2796.953088,s,10,10.460348999023438,1.0460348999023439,0.010496409539496829,1.0462456665039062,1.0595597045898437,1.0617674621582032,1.0635336682128906,"[1.0639752197265624, 1.032802490234375, 1.0487801513671875, 1.0443074951171876, 1.048183837890625, 1.055190185546875, 1.0375479736328126, 1.031301025390625, 1.059069091796875, 1.0391915283203126]",tokens/s,60.227436011821,kWh,1.2212047820330672e-05,6.691703820415936e-06,3.374263798091576e-05,5.264638962166234e-05,tokens/kWh,1196663.2555953555,,s,629,10.642643985748302,0.01691994274363798,0.00268828401669231,0.016693248748779296,0.016881459045410156,0.017036256408691407,0.03882958724975586,"[0.019746816635131836, 0.018382848739624022, 0.01701273536682129, 0.016849983215332032, 0.01669932746887207, 0.016692224502563476, 0.01680691146850586, 0.016694271087646484, 0.016714752197265623, 0.01717043113708496, 0.016660480499267577, 0.016712703704833985, 0.016696319580078126, 0.016719871520996094, 0.01678643226623535, 0.016929792404174804, 0.016880640029907225, 0.016780288696289062, 0.016722976684570314, 0.016667680740356447, 0.01679865646362305, 0.016866336822509764, 0.016791519165039064, 0.016763904571533202, 0.0167587833404541, 0.01677926445007324, 0.017565696716308594, 0.01683046340942383, 0.016694271087646484, 0.016649215698242188, 0.01680076789855957, 0.016748544692993163, 0.016890880584716796, 0.01702195167541504, 0.017115135192871094, 0.017147903442382813, 0.016978944778442383, 0.016745471954345705, 0.01680588722229004, 0.016702463150024414, 0.016724992752075195, 0.016688127517700196, 0.016785408020019533, 0.016694271087646484, 0.016761856079101564, 0.01659596824645996, 0.016747520446777343, 0.016703487396240235, 0.016726015090942382, 0.016715776443481444, 0.016660480499267577, 0.016672767639160157, 0.016738304138183592, 0.016698400497436525, 0.016764896392822266, 0.016734207153320312, 0.016719871520996094, 0.016715776443481444, 0.016722944259643553, 0.01683967971801758, 0.016669696807861328, 0.016727039337158203, 0.03879935836791992, 0.016088064193725587, 0.016125951766967773, 0.016087039947509766, 0.016087039947509766, 0.0161474552154541, 0.016137216567993166, 0.01622425651550293, 0.016097280502319337, 0.016143360137939454, 0.016110591888427735, 0.01618227195739746, 0.016087039947509766, 0.016135168075561524, 0.01609939193725586, 0.01618118476867676, 0.016144384384155275, 0.016116735458374023, 0.01604812812805176, 0.016072704315185548, 0.016060415267944335, 0.016150527954101563, 0.0160501766204834, 0.016152576446533205, 0.01623142433166504, 0.016093183517456054, 0.01621504020690918, 0.016142335891723633, 0.016142335891723633, 0.01616486358642578, 0.016107519149780272, 0.016161792755126952, 0.01607372856140137, 0.016130048751831053, 0.016067647933959962, 0.016158655166625978, 0.016133119583129883, 0.016957439422607423, 0.017089536666870117, 0.01683558464050293, 0.016681983947753908, 0.016679935455322266, 0.016657407760620118, 0.016873472213745116, 0.016767999649047852, 0.016739391326904298, 0.016672704696655275, 0.016738304138183592, 0.016660480499267577, 0.016710655212402344, 0.016677888870239257, 0.01679769515991211, 0.016688127517700196, 0.016759807586669923, 0.016949247360229493, 0.016829439163208008, 0.016728063583374024, 0.016753664016723634, 0.016701440811157226, 0.01678745651245117, 0.016648191452026367, 0.016714752197265623, 0.0166246395111084, 0.03943526458740235, 0.016693311691284178, 0.0166246395111084, 0.0167576961517334, 0.016649215698242188, 0.016711679458618164, 0.016668672561645507, 0.01620684814453125, 0.01604096031188965, 0.016103424072265626, 0.016026624679565428, 0.01613107109069824, 0.016123903274536132, 0.016136192321777345, 0.016090112686157225, 0.01617715263366699, 0.016080896377563478, 0.01618227195739746, 0.01616383934020996, 0.01639628791809082, 0.01618534469604492, 0.016620607376098634, 0.01668396759033203, 0.016914432525634765, 0.016708608627319335, 0.016821279525756835, 0.016722911834716796, 0.016842752456665038, 0.017936384201049805, 0.016990207672119142, 0.016742399215698242, 0.016995328903198242, 0.01679974365234375, 0.016899072647094726, 0.016739328384399413, 0.016884735107421875, 0.016718847274780273, 0.016689151763916017, 0.01685196876525879, 0.01682636833190918, 0.016773120880126953, 0.01680486488342285, 0.016684032440185546, 0.016734207153320312, 0.016665599822998048, 0.016699392318725585, 0.016716800689697265, 0.01679363250732422, 0.01685091209411621, 0.016731136322021483, 0.016702463150024414, 0.016702463150024414, 0.016687103271484375, 0.016688127517700196, 0.016746496200561522, 0.016740352630615234, 0.016687103271484375, 0.0168724479675293, 0.01676697540283203, 0.016708608627319335, 0.016713727951049806, 0.016679935455322266, 0.016884735107421875, 0.039583744049072264, 0.016735231399536133, 0.016842752456665038, 0.01686835289001465, 0.01699635124206543, 0.016905216217041014, 0.016713727951049806, 0.01683046340942383, 0.016661504745483398, 0.016703487396240235, 0.016673791885375978, 0.016793600082397463, 0.016720895767211915, 0.01682841682434082, 0.01681920051574707, 0.01701785659790039, 0.01680486488342285, 0.016732160568237304, 0.016656383514404297, 0.016712703704833985, 0.016737279891967775, 0.016724992752075195, 0.017552383422851564, 0.01803468894958496, 0.017076223373413087, 0.01682636833190918, 0.016740352630615234, 0.01677516746520996, 0.016680959701538087, 0.016748544692993163, 0.01665433692932129, 0.016776191711425782, 0.016656383514404297, 0.016677888870239257, 0.016722944259643553, 0.016694271087646484, 0.016719871520996094, 0.016635904312133788, 0.01658880043029785, 0.016753664016723634, 0.016187423706054686, 0.016183263778686525, 0.016101375579833984, 0.016130048751831053, 0.016136192321777345, 0.01617100715637207, 0.016133119583129883, 0.01617817687988281, 0.016125951766967773, 0.016105472564697267, 0.016135168075561524, 0.016138240814208983, 0.016153600692749022, 0.016149503707885742, 0.016149503707885742, 0.016191488265991212, 0.016100351333618163, 0.016106496810913085, 0.01607891273498535, 0.016128959655761718, 0.016160768508911134, 0.016234495162963866, 0.01611782455444336, 0.038784961700439456, 0.016092159271240233, 0.016078847885131836, 0.016096256256103517, 0.016038911819458008, 0.01607372856140137, 0.016079967498779296, 0.016116640090942384, 0.01607372856140137, 0.01614847946166992, 0.016490495681762696, 0.016752639770507814, 0.016740352630615234, 0.016748544692993163, 0.01664204788208008, 0.016764928817749023, 0.016713727951049806, 0.016732160568237304, 0.016655359268188476, 0.016690176010131837, 0.01661337661743164, 0.016695295333862305, 0.016664575576782227, 0.016708608627319335, 0.016668672561645507, 0.016719871520996094, 0.016652288436889647, 0.016652288436889647, 0.016694271087646484, 0.016711679458618164, 0.016660480499267577, 0.016717824935913086, 0.0166430721282959, 0.016740352630615234, 0.016675840377807616, 0.01665433692932129, 0.016657407760620118, 0.016653312683105468, 0.01661030387878418, 0.016744447708129884, 0.0166430721282959, 0.016686080932617187, 0.01665843200683594, 0.016695295333862305, 0.016761856079101564, 0.016723968505859374, 0.016695295333862305, 0.01679974365234375, 0.016723968505859374, 0.016702463150024414, 0.016692224502563476, 0.01675775909423828, 0.016685056686401366, 0.016703487396240235, 0.016943103790283204, 0.017343488693237305, 0.017329151153564454, 0.017098751068115235, 0.01680793571472168, 0.016756736755371093, 0.016683008193969725, 0.01676697540283203, 0.016693248748779296, 0.03884134292602539, 0.016709632873535156, 0.016669696807861328, 0.016719871520996094, 0.016689184188842774, 0.016759775161743165, 0.01664204788208008, 0.016987136840820313, 0.01676288032531738, 0.016747520446777343, 0.016665599822998048, 0.016711679458618164, 0.01663488006591797, 0.016769023895263673, 0.016825344085693358, 0.01676288032531738, 0.016709632873535156, 0.01677004814147949, 0.01663488006591797, 0.016739328384399413, 0.016669696807861328, 0.01682022476196289, 0.016744447708129884, 0.01712646484375, 0.01674131202697754, 0.016681983947753908, 0.016688127517700196, 0.016685056686401366, 0.016688127517700196, 0.016731136322021483, 0.016680959701538087, 0.016761856079101564, 0.016635904312133788, 0.016827423095703124, 0.017042400360107422, 0.01685196876525879, 0.01666662406921387, 0.016714752197265623, 0.016942079544067384, 0.016849920272827147, 0.016755712509155272, 0.01680281639099121, 0.016699392318725585, 0.016907264709472656, 0.016764928817749023, 0.01675775909423828, 0.016653312683105468, 0.016780288696289062, 0.016729087829589845, 0.016679935455322266, 0.016691200256347655, 0.016729087829589845, 0.016729087829589845, 0.016898048400878905, 0.01681510353088379, 0.016793600082397463, 0.016667648315429686, 0.016705535888671876, 0.016717887878417968, 0.016744384765625, 0.016676864624023437, 0.016750591278076172, 0.016693248748779296, 0.038989822387695314, 0.01615564727783203, 0.01643212890625, 0.01620889663696289, 0.016089088439941408, 0.01613209533691406, 0.01621299171447754, 0.01618534469604492, 0.016108543395996093, 0.016183296203613282, 0.016108543395996093, 0.016142400741577148, 0.016252864837646486, 0.016739328384399413, 0.01678643226623535, 0.016846847534179688, 0.01665126419067383, 0.016749568939208984, 0.016655359268188476, 0.016711679458618164, 0.016707584381103514, 0.016752639770507814, 0.016715776443481444, 0.016785408020019533, 0.016759807586669923, 0.016711679458618164, 0.016638975143432617, 0.01672915267944336, 0.016703424453735353, 0.016715776443481444, 0.016692224502563476, 0.016737279891967775, 0.016662527084350585, 0.016824352264404298, 0.016655328750610352, 0.016709632873535156, 0.016706560134887697, 0.016710655212402344, 0.016649215698242188, 0.01681203269958496, 0.016696319580078126, 0.016676864624023437, 0.0166297607421875, 0.016670719146728515, 0.01662054443359375, 0.01697996711730957, 0.016746528625488283, 0.016205791473388673, 0.016093183517456054, 0.016110591888427735, 0.016071680068969727, 0.016118783950805664, 0.016091136932373046, 0.016118783950805664, 0.016125951766967773, 0.01620377540588379, 0.016045055389404296, 0.016102399826049805, 0.01616486358642578, 0.01623040008544922, 0.01615667152404785, 0.016145408630371092, 0.016202751159667968, 0.03892736053466797, 0.016202848434448244, 0.01609718322753906, 0.01623664093017578, 0.016144287109375, 0.01615875244140625, 0.016097248077392576, 0.016191488265991212, 0.01601126480102539, 0.016101375579833984, 0.016119808197021485, 0.016145408630371092, 0.0160501766204834, 0.016121856689453123, 0.016079872131347657, 0.016150527954101563, 0.016052223205566405, 0.016144384384155275, 0.016013311386108398, 0.016270336151123048, 0.01619353675842285, 0.01613107109069824, 0.01620684814453125, 0.016146432876586913, 0.016064512252807618, 0.01617715263366699, 0.016037887573242187, 0.016080896377563478, 0.015911935806274414, 0.016020479202270507, 0.016078847885131836, 0.016141311645507812, 0.01617407989501953, 0.016072704315185548, 0.016047103881835938, 0.0163450870513916, 0.01659391975402832, 0.016726079940795897, 0.0165467529296875, 0.016746496200561522, 0.016579584121704103, 0.01661952018737793, 0.01660620880126953, 0.016683008193969725, 0.01662566375732422, 0.016697343826293946, 0.01657241630554199, 0.016726015090942382, 0.016737279891967775, 0.016675840377807616, 0.016628768920898436, 0.01666044807434082, 0.01662566375732422, 0.01663692855834961, 0.016673791885375978, 0.016777215957641603, 0.01665023994445801, 0.016698368072509767, 0.016705535888671876, 0.016711679458618164, 0.016918527603149415, 0.016889856338500975, 0.016701440811157226, 0.03908198547363281, 0.016928768157958983, 0.016750591278076172, 0.016894975662231446, 0.016755712509155272, 0.016773120880126953, 0.016668672561645507, 0.01684377670288086, 0.016655359268188476, 0.01742336082458496, 0.01676697540283203, 0.016841728210449217, 0.0166748161315918, 0.016662527084350585, 0.016638975143432617, 0.016793600082397463, 0.016696319580078126, 0.016792575836181642, 0.0166748161315918, 0.016683008193969725, 0.016738304138183592, 0.016737279891967775, 0.016659456253051756, 0.016973823547363282, 0.016729087829589845, 0.016684032440185546, 0.016661504745483398, 0.016670719146728515, 0.01782374382019043, 0.017117183685302736, 0.017903615951538086, 0.017076223373413087, 0.01677824020385742, 0.016805919647216797, 0.01674339294433594, 0.016678911209106445, 0.016683008193969725, 0.016695327758789062, 0.017027040481567383, 0.016885759353637696, 0.016695295333862305, 0.016716800689697265, 0.01668614387512207, 0.01672390365600586, 0.01665433692932129, 0.016684032440185546, 0.016631807327270508, 0.016915456771850586, 0.016738304138183592, 0.016769023895263673, 0.016723968505859374, 0.016729120254516602, 0.016629728317260742, 0.016749568939208984, 0.01681305694580078, 0.017292287826538084, 0.016854015350341797, 0.01683558464050293, 0.01681305694580078, 0.016695295333862305, 0.016717824935913086, 0.016825344085693358, 0.016498687744140626, 0.03892326354980469, 0.016114688873291014, 0.016193599700927735, 0.016104383468627928, 0.016054271697998047, 0.01611065673828125, 0.01612691116333008, 0.016154624938964843, 0.016071680068969727, 0.016145408630371092, 0.016097280502319337, 0.016134143829345703, 0.01620070457458496, 0.01621811294555664, 0.016194559097290038, 0.016272384643554686, 0.016123903274536132, 0.016147520065307615, 0.016159679412841795, 0.016142335891723633, 0.016067583084106444, 0.016115711212158202, 0.016098335266113283, 0.016190431594848634, 0.016086015701293945, 0.01610655975341797, 0.01611974334716797, 0.017861631393432616, 0.017024063110351564, 0.016791488647460936, 0.01675468826293945, 0.01680179214477539, 0.016688127517700196, 0.016660480499267577, 0.016681983947753908, 0.016655359268188476, 0.016639999389648438, 0.016705535888671876, 0.016690176010131837, 0.016734207153320312, 0.0166748161315918, 0.016631807327270508, 0.016660480499267577, 0.01665843200683594, 0.016662527084350585, 0.016676864624023437, 0.016685056686401366, 0.01680793571472168, 0.016745471954345705, 0.016673791885375978, 0.016702463150024414, 0.016773120880126953, 0.016793600082397463, 0.016708608627319335, 0.016667648315429686, 0.016785408020019533, 0.016761856079101564, 0.016710655212402344, 0.01679871940612793, 0.016683008193969725, 0.016667648315429686, 0.016684032440185546, 0.016709632873535156]",tokens/s,59.10185484380597,,,,, -float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1496,7 +1496,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1318.805504,17102.798848,0.0,16456.351744,16359.853056,s,10,1.0681468887329102,0.10681468887329101,0.0002945769989060828,0.1067933464050293,0.10693519973754882,0.10728094291687011,0.10755753746032715,"[0.1076266860961914, 0.10685836791992187, 0.1065077133178711, 0.10670118713378907, 0.10657039642333985, 0.10680992126464844, 0.10684003448486327, 0.10677677154541015, 0.10662172698974609, 0.10683408355712891]",tokens/s,2396.6741157078136,kWh,1.2634387024046402e-06,6.922380217552114e-07,6.126260811170152e-06,8.081937535330004e-06,tokens/kWh,31675572.70529522,MB,1318.805504,17102.798848,0.0,16456.351744,16416.242176,s,10,23.64958349609375,2.364958349609375,0.003937901517589722,2.363560546875,2.3683912841796873,2.3719263549804688,2.3747544116210935,"[2.37546142578125, 2.362560546875, 2.36191064453125, 2.363364013671875, 2.367605712890625, 2.361707275390625, 2.363757080078125, 2.3654501953125, 2.365556884765625, 2.362209716796875]",tokens/s,26.638946943994107,kWh,2.784717529272357e-05,1.5261355628182618e-05,0.00011678417085042906,0.00015989270177133526,tokens/kWh,394014.23143188335,,s,629,24.57144319915771,0.039064297613923236,0.012748992410659883,0.03751424026489258,0.03768427658081054,0.03780403289794922,0.1447934802246094,"[0.038095870971679685, 0.03825155258178711, 0.03809481430053711, 0.0381736946105957, 0.038158336639404294, 0.03813683319091797, 0.03821670532226563, 0.038354942321777344, 0.037966846466064456, 0.03798732757568359, 0.03798015975952149, 0.038345729827880856, 0.038324222564697266, 0.03758899307250976, 0.037379070281982424, 0.037443584442138675, 0.037397502899169925, 0.03744153594970703, 0.03743648147583008, 0.037551040649414065, 0.03747225570678711, 0.03751935958862305, 0.03739852905273437, 0.03750502395629883, 0.037424129486083986, 0.03749990463256836, 0.03742313766479492, 0.0376923828125, 0.03776204681396484, 0.037766143798828124, 0.037585918426513674, 0.03765760040283203, 0.03743231964111328, 0.0375552978515625, 0.03743939208984375, 0.037541889190673826, 0.037485633850097656, 0.03756844711303711, 0.037634048461914066, 0.0376360969543457, 0.0374835205078125, 0.037571582794189456, 0.03747123336791992, 0.037561344146728515, 0.03755212783813477, 0.03765555191040039, 0.037612545013427735, 0.03768326568603515, 0.03752544021606445, 0.03762278366088867, 0.03750912094116211, 0.03759206390380859, 0.037556224822998044, 0.037664768218994144, 0.0379576301574707, 0.03766579055786133, 0.0375470085144043, 0.03761663818359375, 0.03750912094116211, 0.037703678131103514, 0.0375551986694336, 0.03763302230834961, 0.14494720458984375, 0.037305343627929685, 0.03734527969360352, 0.03726438522338867, 0.03735039901733399, 0.03739648056030274, 0.037375999450683595, 0.037367809295654295, 0.037424129486083986, 0.03736576080322265, 0.03743641662597656, 0.03754598236083984, 0.03752959823608398, 0.03738726425170898, 0.0374323501586914, 0.03739542388916016, 0.03747020721435547, 0.037384193420410154, 0.037410911560058595, 0.03741891098022461, 0.03746201705932617, 0.0373831672668457, 0.03747532653808594, 0.0374015998840332, 0.037466110229492186, 0.03746099090576172, 0.03749785614013672, 0.03743334579467773, 0.03754086303710937, 0.03743948745727539, 0.03751628875732422, 0.03746918487548828, 0.03750297546386719, 0.03745280075073242, 0.03750502395629883, 0.037424129486083986, 0.037577728271484374, 0.037479423522949216, 0.03751628875732422, 0.03751935958862305, 0.03753676986694336, 0.0374835205078125, 0.03752447891235351, 0.03750707244873047, 0.03759718322753906, 0.03747430419921875, 0.03759001541137695, 0.0375203857421875, 0.03757670211791992, 0.037541889190673826, 0.03758182525634766, 0.03755011367797852, 0.03762684631347656, 0.03752345657348633, 0.037634048461914066, 0.03758489608764649, 0.03765555191040039, 0.03760947036743164, 0.03763814544677734, 0.03760025787353516, 0.037769248962402344, 0.037604320526123045, 0.037644287109375, 0.14463282775878905, 0.037379070281982424, 0.037353473663330077, 0.03727155303955078, 0.03738214492797851, 0.03729715347290039, 0.03735039901733399, 0.037372928619384765, 0.037408767700195314, 0.037305343627929685, 0.037400577545166014, 0.03733196640014649, 0.037401630401611326, 0.037364734649658206, 0.0374425277709961, 0.03733401489257813, 0.0374200325012207, 0.03735756683349609, 0.037450782775878905, 0.03737091064453125, 0.03742822265625, 0.037382080078125, 0.03748761749267578, 0.03747225570678711, 0.037498878479003905, 0.03740979385375977, 0.03748454284667969, 0.03738726425170898, 0.03754291152954101, 0.03748044967651367, 0.03753676986694336, 0.037425151824951174, 0.0374835205078125, 0.03745075225830078, 0.037510143280029294, 0.03745792007446289, 0.0374835205078125, 0.03743436813354492, 0.037533695220947266, 0.03742924880981445, 0.0374835205078125, 0.03747430419921875, 0.03753267288208008, 0.037730304718017575, 0.037953536987304685, 0.03752140808105469, 0.03761356735229492, 0.03755929565429687, 0.037601280212402347, 0.03749785614013672, 0.03758694458007812, 0.037515262603759765, 0.03754291152954101, 0.037525505065917966, 0.03761459350585938, 0.03750502395629883, 0.03761459350585938, 0.03753676986694336, 0.03768832015991211, 0.03752447891235351, 0.03761356735229492, 0.0375623664855957, 0.03770880126953125, 0.1448079376220703, 0.037343231201171875, 0.03733708953857422, 0.037303295135498044, 0.037370880126953124, 0.037321727752685545, 0.037369857788085936, 0.03735551834106445, 0.03737702560424805, 0.037364734649658206, 0.03744255828857422, 0.03732992172241211, 0.03738828659057617, 0.037395454406738284, 0.03743027114868164, 0.037362686157226564, 0.037465087890625, 0.03734630584716797, 0.03745280075073242, 0.037354496002197264, 0.03742310333251953, 0.03736678314208984, 0.03743948745727539, 0.037375999450683595, 0.03750092697143555, 0.03741388702392578, 0.03745075225830078, 0.03741593551635742, 0.03750502395629883, 0.03741798400878906, 0.03754291152954101, 0.03745177459716797, 0.037485633850097656, 0.03755513763427734, 0.037797889709472655, 0.03751321411132812, 0.03752959823608398, 0.037477375030517575, 0.037539840698242184, 0.03748761749267578, 0.03757670211791992, 0.03751116943359375, 0.037566463470458986, 0.03752243041992188, 0.037610496520996094, 0.03751424026489258, 0.03761151885986328, 0.03751424026489258, 0.03762073516845703, 0.03749683380126953, 0.03765657424926758, 0.0378419189453125, 0.03779072189331055, 0.03754393768310547, 0.03767603302001953, 0.0376484489440918, 0.03766368103027344, 0.03761971282958984, 0.03770675277709961, 0.03770675277709961, 0.037703678131103514, 0.03761663818359375, 0.037754878997802735, 0.14479055786132813, 0.03738723373413086, 0.03736166381835938, 0.03741798400878906, 0.03748659133911133, 0.037438465118408204, 0.037370880126953124, 0.03737497711181641, 0.03752140808105469, 0.037365825653076175, 0.037401535034179687, 0.03757056045532227, 0.03783168029785156, 0.03757670211791992, 0.03743436813354492, 0.037473281860351565, 0.03784806442260742, 0.037663745880126956, 0.03759212875366211, 0.03739744186401367, 0.03747430419921875, 0.037599231719970705, 0.03748969650268555, 0.03757360076904297, 0.03764223861694336, 0.037623809814453124, 0.037612545013427735, 0.03752140808105469, 0.03767295837402344, 0.03762790298461914, 0.037666912078857424, 0.037580703735351564, 0.03761356735229492, 0.03750707244873047, 0.03750912094116211, 0.03747225570678711, 0.03760639953613281, 0.037574657440185545, 0.03763302230834961, 0.03752755355834961, 0.037610496520996094, 0.037550079345703126, 0.03766886520385742, 0.03763097763061524, 0.03761459350585938, 0.03766681671142578, 0.03771596908569336, 0.037570625305175784, 0.037676990509033205, 0.03757056045532227, 0.03759206390380859, 0.03754393768310547, 0.037598270416259766, 0.03759199905395508, 0.03758899307250976, 0.03756851196289063, 0.037644287109375, 0.03763916778564453, 0.037661697387695314, 0.03754598236083984, 0.03763814544677734, 0.03757363128662109, 0.03784499359130859, 0.14500352478027342, 0.037343231201171875, 0.037340160369873046, 0.03729715347290039, 0.037394432067871096, 0.037303359985351565, 0.03741484832763672, 0.03736678314208984, 0.03734630584716797, 0.037378047943115236, 0.03741593551635742, 0.03733196640014649, 0.03742105484008789, 0.037354496002197264, 0.0374128646850586, 0.037365791320800784, 0.037456863403320315, 0.03732992172241211, 0.037427200317382815, 0.037389312744140625, 0.03742211151123047, 0.03739337539672852, 0.037459968566894535, 0.037359615325927735, 0.0374917106628418, 0.03741183853149414, 0.0375203857421875, 0.03741491317749023, 0.037563392639160156, 0.037438465118408204, 0.03751839828491211, 0.03743020629882812, 0.03750502395629883, 0.03744255828857422, 0.03749273681640625, 0.037459968566894535, 0.037510143280029294, 0.037495807647705076, 0.03765555191040039, 0.03747635269165039, 0.03757056045532227, 0.037493759155273435, 0.037602302551269534, 0.03748863983154297, 0.03753881454467774, 0.03748966217041016, 0.03759308624267578, 0.03751424026489258, 0.03760332870483398, 0.03750707244873047, 0.03762176132202148, 0.03750604629516602, 0.03757875061035156, 0.037539840698242184, 0.037601280212402347, 0.03756854248046875, 0.037622753143310546, 0.03753676986694336, 0.03764121627807617, 0.037548030853271484, 0.03767295837402344, 0.03758694458007812, 0.03764019012451172, 0.14501785278320312, 0.03732992172241211, 0.0373309440612793, 0.037343231201171875, 0.037482494354248046, 0.037324798583984374, 0.03737702560424805, 0.037369857788085936, 0.037424129486083986, 0.03738828659057617, 0.03738623809814453, 0.0373831672668457, 0.03748147201538086, 0.03740262222290039, 0.037501953125, 0.0373831672668457, 0.03745894241333008, 0.03745280075073242, 0.03758182525634766, 0.03767295837402344, 0.03750912094116211, 0.03741593551635742, 0.03750400161743164, 0.037459968566894535, 0.03748044967651367, 0.0374835205078125, 0.0374917106628418, 0.03743129730224609, 0.03756851196289063, 0.03746918487548828, 0.03751935958862305, 0.03753267288208008, 0.03767295837402344, 0.03746406555175781, 0.03748659133911133, 0.03746099090576172, 0.03758899307250976, 0.03743948745727539, 0.037550079345703126, 0.03750809478759766, 0.03751424026489258, 0.03754707336425781, 0.03754387283325195, 0.037498878479003905, 0.037596160888671876, 0.03750400161743164, 0.03758796691894531, 0.037539840698242184, 0.037574657440185545, 0.0375470085144043, 0.03758694458007812, 0.03761459350585938, 0.0376360969543457, 0.037526527404785154, 0.03759513473510742, 0.03760537719726562, 0.03776716613769531, 0.03758694458007812, 0.037628929138183595, 0.03757875061035156, 0.0376627197265625, 0.037607425689697264, 0.03765350341796875, 0.14492160034179688, 0.03733401489257813, 0.037359615325927735, 0.037303359985351565, 0.03747116851806641, 0.037351425170898435, 0.03737395095825195, 0.03734527969360352, 0.03736783981323242, 0.03732988739013672, 0.037407745361328126, 0.03735968017578125, 0.037417919158935546, 0.03739347076416016, 0.03740972900390625, 0.03734527969360352, 0.03746713638305664, 0.03735551834106445, 0.03744768142700195, 0.03741491317749023, 0.037449726104736326, 0.0373831672668457, 0.03743641662597656, 0.03739136123657227, 0.03746918487548828, 0.03740979385375977, 0.03746201705932617, 0.03745180892944336, 0.037537761688232425, 0.03741388702392578, 0.03755417633056641, 0.03745587158203125, 0.037515262603759765, 0.037410816192626956, 0.03749273681640625, 0.03742524719238281, 0.03769232177734375, 0.03791360092163086, 0.03763097763061524, 0.037645313262939455, 0.03770265579223633, 0.03759206390380859, 0.03773235321044922, 0.03763916778564453, 0.037718017578125, 0.03768115234375, 0.03758694458007812, 0.03753574371337891, 0.03767193603515625, 0.038105152130126954, 0.03770566558837891, 0.03755215835571289, 0.037645278930664064, 0.03762483215332031, 0.03791872024536133, 0.037781505584716796, 0.03769139099121094, 0.037610496520996094, 0.037763072967529294, 0.037667839050292966, 0.03772723388671875, 0.03760639953613281, 0.03769036865234375, 0.14479461669921875, 0.03740979385375977, 0.037531646728515625, 0.03756032180786133, 0.03778355026245117, 0.037533695220947266, 0.037512191772460936, 0.037477375030517575, 0.0376258544921875, 0.03739136123657227, 0.03742105484008789, 0.03744153594970703, 0.03750707244873047, 0.037394432067871096, 0.03746918487548828, 0.03748761749267578, 0.037749759674072264, 0.03750912094116211, 0.03751628875732422, 0.0378081283569336, 0.037618686676025394, 0.03741593551635742, 0.03748966217041016, 0.037400577545166014, 0.037459968566894535, 0.03745587158203125, 0.03751935958862305, 0.03745177459716797, 0.03753779220581055, 0.03746406555175781, 0.03750912094116211, 0.03746815872192383, 0.03759308624267578, 0.03749068832397461, 0.03748966217041016, 0.03746713638305664, 0.03755427169799805, 0.03747113418579102, 0.037525505065917966, 0.03748767852783203, 0.03752544021606445, 0.03752447891235351, 0.037550079345703126, 0.037544960021972655, 0.037569534301757815, 0.03752447891235351, 0.03762688064575195, 0.037541889190673826, 0.037703678131103514, 0.03753779220581055, 0.03757567977905273, 0.03753676986694336, 0.037602302551269534, 0.037574657440185545, 0.03757567977905273, 0.037544960021972655, 0.03774566268920899, 0.037569534301757815, 0.03763916778564453, 0.037553150177001955, 0.03763302230834961, 0.0375880012512207, 0.03763504028320312, 0.14493907165527345, 0.037376960754394534, 0.03738521575927734, 0.03729100799560547, 0.037364734649658206, 0.0373309440612793, 0.037369857788085936, 0.03739648056030274, 0.03737395095825195, 0.037364734649658206, 0.03746815872192383, 0.03733299255371094, 0.0374466552734375, 0.03735756683349609, 0.03740979385375977, 0.03738009643554688, 0.03743948745727539, 0.03733708953857422, 0.03743641662597656, 0.03735859298706055, 0.037425151824951174, 0.037375999450683595, 0.03744768142700195, 0.037372928619384765, 0.03748454284667969, 0.03742822265625, 0.03752553558349609, 0.0374425277709961, 0.03751321411132812, 0.03744255828857422, 0.037531646728515625, 0.03743948745727539, 0.03752345657348633, 0.03743027114868164, 0.03746406555175781, 0.03750400161743164, 0.03753881454467774, 0.03749990463256836, 0.037526527404785154, 0.03747532653808594, 0.037563392639160156, 0.037550079345703126, 0.037571582794189456, 0.037533695220947266, 0.037580799102783204, 0.03751321411132812, 0.03763097763061524, 0.03750297546386719, 0.03764121627807617, 0.037493759155273435, 0.037596160888671876, 0.03752246475219727, 0.037606369018554686, 0.037526527404785154, 0.037594112396240234, 0.03760537719726562, 0.037644287109375, 0.037544960021972655, 0.0376545295715332, 0.037556224822998044, 0.03766681671142578, 0.037571582794189456, 0.03776921463012695]",tokens/s,25.598821969950933,,,main,False,False float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1278.64832,15196.48768,0.0,14550.040576,13944.957952,s,10,0.8168599700927734,0.08168599700927734,0.0017125434954674181,0.08090859222412108,0.08301425933837892,0.0847530174255371,0.08614402389526367,"[0.08649177551269531, 0.08102448272705078, 0.081532958984375, 0.08075113677978515, 0.08064419555664062, 0.08055107116699219, 0.08077008056640625, 0.0807927017211914, 0.08262786865234376, 0.08167369842529297]",tokens/s,3133.9520771341663,kWh,9.578823829637206e-07,5.248717686505011e-07,5.411986856467704e-06,6.894741008081926e-06,tokens/kWh,37129748.557620965,MB,1278.976,15196.48768,0.0,14550.040576,13945.037824,s,10,29.099600341796872,2.9099600341796874,0.007177875359054941,2.9103421630859376,2.9147032958984376,2.9201170288085936,2.924448015136719,"[2.92553076171875, 2.91293896484375, 2.90659423828125, 2.89784130859375, 2.900981689453125, 2.913500244140625, 2.913074462890625, 2.910731201171875, 2.908454345703125, 2.909953125]",tokens/s,21.649781873296273,kWh,3.440782479594686e-05,1.8857076072979383e-05,0.00012732428599913426,0.00018058918686806048,tokens/kWh,348858.0966147667,,s,629,29.782428653717,0.04734885318555968,0.009705406038681034,0.04612710571289062,0.04654249038696289,0.046845132446289066,0.12721811279296877,"[0.04724531173706055, 0.04674355316162109, 0.04698316955566406, 0.047265792846679686, 0.04716134262084961, 0.04681830215454102, 0.04755251312255859, 0.04667596817016602, 0.047184928894042966, 0.04760367965698242, 0.047064064025878906, 0.04638924789428711, 0.04663097763061524, 0.04603897476196289, 0.046258174896240234, 0.04610559844970703, 0.046148609161376954, 0.04613324737548828, 0.04612300872802735, 0.046173183441162106, 0.046808063507080076, 0.04640153503417969, 0.0463636474609375, 0.046004222869873046, 0.046124031066894534, 0.046413822174072264, 0.04611481475830078, 0.04610047912597656, 0.04619878387451172, 0.046088191986083986, 0.04631552124023437, 0.04662681579589844, 0.046306304931640625, 0.046071807861328126, 0.04640358352661133, 0.04619673538208008, 0.04623155212402344, 0.046061569213867185, 0.04623360061645508, 0.046031871795654294, 0.04622444915771484, 0.04634515380859375, 0.046496768951416016, 0.04604108810424805, 0.04610662460327149, 0.04598988723754883, 0.04610559844970703, 0.046086143493652344, 0.04684799957275391, 0.047483905792236325, 0.04658380889892578, 0.04656742477416992, 0.046712833404541014, 0.046077953338623044, 0.046235649108886716, 0.04609638214111328, 0.04611276626586914, 0.04597555160522461, 0.046074878692626955, 0.04595916748046875, 0.04613119888305664, 0.046042110443115236, 0.12811776733398436, 0.04605952072143555, 0.04600115203857422, 0.04608204650878906, 0.0464087028503418, 0.04651212692260742, 0.0463636474609375, 0.046663681030273435, 0.04601036834716797, 0.046137344360351565, 0.04633804702758789, 0.04612710571289062, 0.04618137741088867, 0.04613017654418945, 0.04610662460327149, 0.04616499328613281, 0.04618137741088867, 0.04732723236083984, 0.04656639862060547, 0.046723072052001956, 0.04606054306030274, 0.04642406463623047, 0.0460871696472168, 0.04612812805175781, 0.046025726318359376, 0.046159870147705076, 0.04615679931640625, 0.04614144134521484, 0.046077953338623044, 0.046176254272460936, 0.04607078552246094, 0.04613119888305664, 0.04616499328613281, 0.046139392852783206, 0.0460882568359375, 0.04608607864379883, 0.04655001449584961, 0.04639641571044922, 0.04612198257446289, 0.046192638397216795, 0.046088191986083986, 0.04611481475830078, 0.0464640007019043, 0.04693196868896484, 0.04614656066894531, 0.04637286376953125, 0.0463267822265625, 0.04611379241943359, 0.04614348983764648, 0.046134273529052736, 0.046066688537597655, 0.046129150390625, 0.046033920288085936, 0.046412799835205076, 0.04630835342407227, 0.046104576110839846, 0.0459950065612793, 0.046139392852783206, 0.046080001831054686, 0.046091262817382815, 0.046063617706298826, 0.046069793701171875, 0.04605436706542969, 0.12855807495117189, 0.04613017654418945, 0.04614553451538086, 0.046115840911865234, 0.04598476791381836, 0.046150657653808595, 0.04603596878051758, 0.04613020706176758, 0.046008289337158205, 0.046139392852783206, 0.046064640045166014, 0.046167041778564455, 0.045859840393066405, 0.0459315185546875, 0.04600115203857422, 0.046088191986083986, 0.046102527618408204, 0.046164031982421874, 0.04594899368286133, 0.04687449645996094, 0.04614553451538086, 0.046265342712402346, 0.04603801727294922, 0.04614553451538086, 0.04600729751586914, 0.046104576110839846, 0.04601036834716797, 0.04622441482543945, 0.046024673461914065, 0.046072830200195314, 0.04602982330322265, 0.046050304412841796, 0.04598169708251953, 0.046182430267333985, 0.04615779113769531, 0.04609740829467773, 0.04604108810424805, 0.04616908645629883, 0.04602982330322265, 0.04614348983764648, 0.04624076843261719, 0.046134273529052736, 0.04596121597290039, 0.04598886489868164, 0.04605132675170898, 0.046124031066894534, 0.04602880096435547, 0.046102527618408204, 0.04608204650878906, 0.04615167999267578, 0.04605235290527344, 0.046077953338623044, 0.04600627136230469, 0.04659609603881836, 0.04694937515258789, 0.046682113647460936, 0.04607078552246094, 0.04613328170776367, 0.045957088470458984, 0.04610969543457031, 0.046020606994628906, 0.04612300872802735, 0.04600729751586914, 0.12716134643554688, 0.04579328155517578, 0.04561920166015625, 0.04581683349609375, 0.04567244720458984, 0.04595609664916992, 0.045908992767333984, 0.045808639526367184, 0.04564889526367188, 0.04589567947387695, 0.04581990432739258, 0.046002174377441404, 0.0459315185546875, 0.04613324737548828, 0.0460134391784668, 0.04604313659667969, 0.046031871795654294, 0.04607897567749023, 0.045985790252685545, 0.046080001831054686, 0.04597555160522461, 0.045770751953125, 0.0458526725769043, 0.046034942626953124, 0.04603289413452148, 0.04609228897094726, 0.04603596878051758, 0.04611481475830078, 0.0459950065612793, 0.04586822509765625, 0.045786945343017575, 0.046077953338623044, 0.04607385635375977, 0.04610969543457031, 0.04605542373657227, 0.046069759368896485, 0.0460840950012207, 0.04621414566040039, 0.04606054306030274, 0.04615270233154297, 0.04593356704711914, 0.046043167114257814, 0.046147552490234375, 0.046104576110839846, 0.04616396713256836, 0.04612505722045898, 0.04605952072143555, 0.0460840950012207, 0.0459048957824707, 0.04605440139770508, 0.045695999145507815, 0.04575033569335937, 0.04584646224975586, 0.04615475082397461, 0.04641996765136719, 0.046148609161376954, 0.04630220794677734, 0.04613017654418945, 0.04597555160522461, 0.04587417602539062, 0.04580966567993164, 0.04609740829467773, 0.045960193634033204, 0.12724018859863281, 0.04617523193359375, 0.04598169708251953, 0.046050304412841796, 0.046004222869873046, 0.04606771087646484, 0.04600729751586914, 0.046840831756591796, 0.046383102416992186, 0.04615884780883789, 0.04669440078735351, 0.046377983093261715, 0.046058494567871096, 0.04604528045654297, 0.04600412750244141, 0.04619673538208008, 0.0461415023803711, 0.046054336547851564, 0.0458076171875, 0.0458526725769043, 0.04584960174560547, 0.04589056015014648, 0.046167041778564455, 0.045965312957763675, 0.04626432037353516, 0.04598896026611328, 0.0458064956665039, 0.04604108810424805, 0.045841407775878903, 0.045843616485595706, 0.04580438232421875, 0.045888671875, 0.0458238410949707, 0.0459417610168457, 0.04587519836425781, 0.04611686325073242, 0.046020606994628906, 0.04628582382202148, 0.046341121673583986, 0.04613539123535156, 0.04677519989013672, 0.04654188919067383, 0.04608710479736328, 0.04647116851806641, 0.0460134391784668, 0.04598988723754883, 0.04575539016723633, 0.04590591812133789, 0.045795326232910154, 0.04592435073852539, 0.04572985458374024, 0.04600620651245117, 0.045720577239990234, 0.04581587219238281, 0.046367679595947266, 0.04591939163208008, 0.04584534454345703, 0.04603903961181641, 0.045674495697021485, 0.04581478500366211, 0.0458721923828125, 0.04580550384521485, 0.04575743865966797, 0.12702105712890624, 0.04650393676757812, 0.045930496215820314, 0.046371841430664064, 0.04579328155517578, 0.046322689056396485, 0.04591206359863281, 0.04608512115478516, 0.04635443115234375, 0.04627763366699219, 0.046440448760986325, 0.046430206298828124, 0.04608822250366211, 0.04622025680541992, 0.04643430328369141, 0.04610559844970703, 0.04599603271484375, 0.04643430328369141, 0.046266368865966793, 0.04613631820678711, 0.046031871795654294, 0.046734336853027345, 0.04647628784179687, 0.04646912002563477, 0.04619468688964844, 0.046102527618408204, 0.046042110443115236, 0.0460840950012207, 0.04624486541748047, 0.046120960235595705, 0.04608102416992187, 0.046139392852783206, 0.046304256439208984, 0.046502910614013675, 0.04657254409790039, 0.046514175415039063, 0.046148609161376954, 0.046238719940185545, 0.046150657653808595, 0.046440448760986325, 0.046170112609863284, 0.047333377838134766, 0.04649574279785156, 0.04621823883056641, 0.04622335815429687, 0.04635238265991211, 0.046048255920410154, 0.04623360061645508, 0.046167041778564455, 0.04617113494873047, 0.04607590484619141, 0.04604927825927734, 0.046024768829345704, 0.046093246459960935, 0.04625612640380859, 0.04633599853515625, 0.04598067092895508, 0.046601215362548826, 0.046064640045166014, 0.046053375244140625, 0.046148609161376954, 0.04622335815429687, 0.04604006576538086, 0.12749721527099608, 0.04660736083984375, 0.046058494567871096, 0.04618239974975586, 0.04592127990722656, 0.04613529586791992, 0.04612505722045898, 0.04609228897094726, 0.046050304412841796, 0.04617113494873047, 0.046072830200195314, 0.047718399047851565, 0.04685619354248047, 0.04643430328369141, 0.04601446533203125, 0.04626227188110352, 0.046099456787109375, 0.04617932891845703, 0.04600320053100586, 0.046104576110839846, 0.046074878692626955, 0.04616908645629883, 0.04643840026855469, 0.046415870666503906, 0.046235649108886716, 0.046120960235595705, 0.04616089630126953, 0.04636774444580078, 0.04645273590087891, 0.04626739120483398, 0.046140415191650394, 0.0462213134765625, 0.046020606994628906, 0.046548992156982424, 0.04602675247192383, 0.04628070449829102, 0.04603084945678711, 0.04619878387451172, 0.04623462295532227, 0.04618751907348633, 0.046170112609863284, 0.04618547058105469, 0.046066688537597655, 0.0461475830078125, 0.04641689682006836, 0.046388225555419924, 0.04610047912597656, 0.04614451217651367, 0.04603289413452148, 0.04613324737548828, 0.04603903961181641, 0.04620185470581055, 0.04618854522705078, 0.04628684616088867, 0.04658585739135742, 0.046611457824707034, 0.04618137741088867, 0.046388225555419924, 0.04597862243652344, 0.046069759368896485, 0.04617216110229492, 0.04619571304321289, 0.046015487670898435, 0.12743577575683593, 0.046284801483154295, 0.04600627136230469, 0.0461578254699707, 0.04592127990722656, 0.046231582641601564, 0.046134239196777345, 0.04617523193359375, 0.046069759368896485, 0.04610047912597656, 0.046061569213867185, 0.04607078552246094, 0.046209022521972655, 0.04585881423950195, 0.0459048957824707, 0.04615167999267578, 0.046268417358398435, 0.04612812805175781, 0.04615167999267578, 0.046074878692626955, 0.046023681640625, 0.046192638397216795, 0.046382080078125, 0.04645171356201172, 0.04590796661376953, 0.046355457305908204, 0.04589363098144531, 0.04599091339111328, 0.04620492935180664, 0.04676403045654297, 0.04693401718139648, 0.04628684616088867, 0.04616908645629883, 0.046324798583984375, 0.046093246459960935, 0.046148609161376954, 0.046023681640625, 0.04611072158813476, 0.04615884780883789, 0.04627872085571289, 0.046179264068603516, 0.046271488189697264, 0.04598076629638672, 0.046097312927246094, 0.04636876678466797, 0.04610047912597656, 0.04605132675170898, 0.0460134391784668, 0.04602880096435547, 0.04654489517211914, 0.04620185470581055, 0.04636774444580078, 0.04655411148071289, 0.046339073181152345, 0.04632371139526367, 0.04628889465332031, 0.04616089630126953, 0.04612607955932617, 0.046063617706298826, 0.046666751861572264, 0.04620800018310547, 0.04634316635131836, 0.04616294479370117, 0.1293834228515625, 0.046120960235595705, 0.04658585739135742, 0.04615679931640625, 0.046004222869873046, 0.04636159896850586, 0.046257152557373046, 0.04631449508666992, 0.04600320053100586, 0.04629401779174805, 0.046344192504882815, 0.04694527816772461, 0.04612812805175781, 0.04591820907592774, 0.04593868637084961, 0.04597555160522461, 0.04588236618041992, 0.04599091339111328, 0.045930496215820314, 0.04607590484619141, 0.04600627136230469, 0.04606771087646484, 0.046252033233642575, 0.0460840950012207, 0.046083072662353515, 0.04614963150024414, 0.047263744354248044, 0.046074878692626955, 0.04591734313964844, 0.04601327896118164, 0.046183425903320315, 0.04595404815673828, 0.04585779190063476, 0.046061569213867185, 0.045813758850097655, 0.04598067092895508, 0.045753345489501954, 0.04597862243652344, 0.045870079040527346, 0.04591513442993164, 0.046205951690673826, 0.04625920104980469, 0.04597760009765625, 0.04642201614379883, 0.04693196868896484, 0.04663808059692383, 0.04605542373657227, 0.04632985687255859, 0.046134273529052736, 0.04627251052856445, 0.046268417358398435, 0.04616294479370117, 0.04601139068603516, 0.046170112609863284, 0.046265342712402346, 0.04632783889770508, 0.04625616073608398, 0.04617824172973633, 0.04601036834716797, 0.04608102416992187, 0.04599705505371094, 0.0462366714477539, 0.04605235290527344, 0.12838911437988282, 0.046166015625, 0.04619059371948242, 0.04621721649169922, 0.04603289413452148, 0.04617113494873047, 0.04603903961181641, 0.046266368865966793, 0.04611174392700195, 0.046150657653808595, 0.04604108810424805, 0.04611072158813476, 0.04637081527709961, 0.046137344360351565, 0.04603084945678711, 0.04611481475830078, 0.04602880096435547, 0.046238719940185545, 0.04616294479370117, 0.04616396713256836, 0.046203903198242184, 0.0461844482421875, 0.046137344360351565, 0.04612716674804687, 0.04622739028930664, 0.046271488189697264, 0.046120960235595705, 0.046089214324951173, 0.04615577697753906, 0.046219264984130856, 0.04609024047851563, 0.046140415191650394, 0.046089214324951173, 0.04733747100830078, 0.047145984649658204, 0.046473217010498044, 0.04608102416992187, 0.04627251052856445, 0.04632473754882813, 0.04624998474121094, 0.046091262817382815, 0.04611072158813476, 0.04615679931640625, 0.046117889404296876, 0.0464087028503418, 0.04618137741088867, 0.046063617706298826, 0.0460871696472168, 0.04605952072143555, 0.04614451217651367, 0.04609228897094726, 0.04626739120483398, 0.046050304412841796, 0.04609638214111328, 0.046034942626953124, 0.046148609161376954, 0.04610559844970703, 0.04605033493041992, 0.046024673461914065, 0.04611481475830078, 0.04597452926635742, 0.04616089630126953, 0.046031871795654294]",tokens/s,21.1198357029052,,,,, -float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2012,7 +2012,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1295.101952,1711.80032,0.0,1065.353216,1018.226688,s,10,0.2013744029998779,0.02013744029998779,0.0008730184027724601,0.01980724811553955,0.020394617843627928,0.021564012908935542,0.02249952896118164,"[0.022733407974243162, 0.01984217643737793, 0.02013475227355957, 0.01979795265197754, 0.01981654357910156, 0.019989408493041993, 0.01977084732055664, 0.019749568939208983, 0.01976016044616699, 0.019779584884643556]",tokens/s,12712.638557153423,kWh,2.3687972911772287e-07,1.297966301216819e-07,6.353857231823793e-07,1.0020620824217839e-06,tokens/kWh,255473193.21902606,MB,1295.101952,1711.80032,0.0,1065.353216,1058.88,s,10,11.73953161621094,1.1739531616210939,0.011730416071497548,1.1694104003906252,1.1823773559570314,1.1941164123535157,1.203507657470703,"[1.20585546875, 1.1772835693359376, 1.1797686767578126, 1.1686829833984376, 1.169909912109375, 1.164593017578125, 1.165302490234375, 1.1649390869140626, 1.168910888671875, 1.1742855224609374]",tokens/s,53.66483268634352,kWh,1.3735118035257752e-05,7.526501696206713e-06,2.3843189415217027e-05,4.5104809146681495e-05,tokens/kWh,1396746.8478831844,,s,629,11.897618431091306,0.018915132640844688,0.002428042337562234,0.01850160026550293,0.018991923141479492,0.019597497177124022,0.038714585723876954,"[0.019733535766601563, 0.019600351333618163, 0.01957689666748047, 0.01968841552734375, 0.01974064064025879, 0.019740671157836915, 0.020624383926391602, 0.01965260887145996, 0.01963417625427246, 0.01960038375854492, 0.019558399200439454, 0.019720191955566405, 0.01971004867553711, 0.01987164878845215, 0.01964339256286621, 0.01983180809020996, 0.01967616081237793, 0.0196495361328125, 0.019817472457885742, 0.02000588798522949, 0.019535871505737306, 0.01959321594238281, 0.01948057556152344, 0.019458047866821288, 0.019532800674438477, 0.01983897590637207, 0.02004377555847168, 0.020068351745605468, 0.018985984802246093, 0.018689023971557618, 0.01847091293334961, 0.01883955192565918, 0.018696191787719727, 0.018479103088378905, 0.01863580894470215, 0.018444255828857423, 0.018516992568969725, 0.01856716728210449, 0.018562047958374024, 0.01842278480529785, 0.018962432861328125, 0.01863065528869629, 0.019122175216674805, 0.01839923286437988, 0.018315263748168945, 0.01861222457885742, 0.018464767456054687, 0.018489343643188477, 0.018540544509887694, 0.018570240020751954, 0.01864499282836914, 0.018524160385131837, 0.01840640068054199, 0.018461696624755858, 0.018490367889404297, 0.018449407577514648, 0.01869004821777344, 0.018420736312866212, 0.01848320007324219, 0.019054592132568358, 0.01873209571838379, 0.01857632064819336, 0.038790145874023435, 0.018527231216430663, 0.01846784019470215, 0.018348031997680665, 0.018406463623046876, 0.01850060844421387, 0.018520000457763672, 0.01841152000427246, 0.018647039413452148, 0.01882521629333496, 0.01848320007324219, 0.018485248565673826, 0.018715648651123046, 0.018749439239501953, 0.01869004821777344, 0.01842790412902832, 0.018449407577514648, 0.01839923286437988, 0.01840742492675781, 0.018485248565673826, 0.01887539291381836, 0.01845043182373047, 0.018482175827026368, 0.018689023971557618, 0.0187064323425293, 0.01866547203063965, 0.01880268859863281, 0.018959360122680666, 0.018705408096313478, 0.018324480056762696, 0.018431999206542968, 0.018770944595336913, 0.018502656936645507, 0.018391040802001952, 0.018770944595336913, 0.018840576171875, 0.018780160903930664, 0.01887027168273926, 0.01884876823425293, 0.01878118324279785, 0.018873344421386717, 0.01902284812927246, 0.01881395149230957, 0.018721792221069337, 0.018758655548095703, 0.018882560729980468, 0.018717695236206054, 0.018635839462280274, 0.018702272415161134, 0.019133440017700197, 0.018856960296630858, 0.018766847610473633, 0.018860031127929687, 0.018561023712158203, 0.01861734390258789, 0.01884160041809082, 0.019008575439453126, 0.018829248428344728, 0.01879347229003906, 0.018949119567871094, 0.018926591873168946, 0.018877439498901367, 0.018648128509521484, 0.03926111984252929, 0.01864396858215332, 0.01859584045410156, 0.0188221435546875, 0.01885593605041504, 0.018736127853393555, 0.018884607315063476, 0.018957311630249024, 0.01899622344970703, 0.019158016204833983, 0.01900441551208496, 0.018893823623657227, 0.018694143295288086, 0.018671615600585938, 0.018991104125976564, 0.01882624053955078, 0.018612287521362306, 0.018602943420410155, 0.018579456329345705, 0.01902796745300293, 0.018968576431274413, 0.018755584716796874, 0.018506752014160157, 0.01843814468383789, 0.018508800506591795, 0.018542591094970702, 0.01847603225708008, 0.018506784439086914, 0.018529247283935547, 0.018777088165283205, 0.01879347229003906, 0.018507776260375978, 0.018521087646484375, 0.018545663833618165, 0.018545663833618165, 0.018413568496704103, 0.018512895584106445, 0.018485248565673826, 0.019091455459594727, 0.019005439758300782, 0.01860095977783203, 0.018436159133911133, 0.01873094367980957, 0.01905971145629883, 0.019960832595825196, 0.019219455718994142, 0.01899519920349121, 0.01901875114440918, 0.018957311630249024, 0.018531328201293946, 0.018790399551391602, 0.01903923225402832, 0.018548736572265623, 0.01843302345275879, 0.01847603225708008, 0.01843612861633301, 0.018780128479003906, 0.0186429443359375, 0.018315263748168945, 0.018357248306274415, 0.01841766357421875, 0.018472959518432617, 0.01846784019470215, 0.038719486236572266, 0.018576383590698242, 0.018502656936645507, 0.018482175827026368, 0.018485248565673826, 0.018332735061645507, 0.018309055328369142, 0.018417728424072265, 0.01851897621154785, 0.018501632690429686, 0.018529279708862305, 0.018502656936645507, 0.0184453125, 0.018511871337890624, 0.018479103088378905, 0.018515968322753908, 0.01845452880859375, 0.01848422431945801, 0.018456575393676757, 0.018542591094970702, 0.018295808792114256, 0.018457599639892578, 0.018528255462646484, 0.018367488861083983, 0.01828556823730469, 0.018336767196655272, 0.01845248031616211, 0.018487295150756835, 0.018921472549438476, 0.018703359603881836, 0.018774015426635742, 0.018947071075439453, 0.018489343643188477, 0.018540544509887694, 0.01845145606994629, 0.018495487213134765, 0.018534400939941405, 0.01864089584350586, 0.01840947151184082, 0.01845248031616211, 0.018779136657714843, 0.019087360382080077, 0.018893823623657227, 0.01939967918395996, 0.019009536743164062, 0.018570240020751954, 0.01887129592895508, 0.018463743209838866, 0.01846272087097168, 0.018461696624755858, 0.01838591957092285, 0.01847500801086426, 0.018489343643188477, 0.018489343643188477, 0.018463743209838866, 0.018916351318359375, 0.018341888427734376, 0.01844326400756836, 0.018463743209838866, 0.018440191268920898, 0.01843097686767578, 0.018356224060058594, 0.01848320007324219, 0.03881881713867188, 0.018539520263671876, 0.018508800506591795, 0.018479103088378905, 0.018525184631347655, 0.018542591094970702, 0.01855897521972656, 0.018515968322753908, 0.01880575942993164, 0.018932735443115235, 0.018707456588745116, 0.018548736572265623, 0.018448383331298827, 0.018548736572265623, 0.01862553596496582, 0.018714624404907225, 0.01848422431945801, 0.018563072204589845, 0.018626560211181642, 0.018496511459350586, 0.018405376434326173, 0.018447359085083007, 0.01843302345275879, 0.018391040802001952, 0.018488319396972656, 0.018447359085083007, 0.018513919830322266, 0.018576448440551757, 0.018470848083496094, 0.018478080749511717, 0.018491392135620118, 0.018525184631347655, 0.018518016815185546, 0.018533376693725585, 0.018585599899291993, 0.018440191268920898, 0.018487295150756835, 0.018603008270263673, 0.01844326400756836, 0.018392127990722658, 0.018435039520263673, 0.018387935638427735, 0.01847603225708008, 0.01841459274291992, 0.018441247940063476, 0.01850160026550293, 0.01901158332824707, 0.018525215148925783, 0.01865212821960449, 0.018457599639892578, 0.01840233612060547, 0.01844528007507324, 0.018463775634765624, 0.01849545669555664, 0.018371583938598633, 0.018448383331298827, 0.018485248565673826, 0.020388864517211915, 0.019298303604125978, 0.018562047958374024, 0.01846380805969238, 0.01849337577819824, 0.0184770565032959, 0.039277568817138675, 0.018464767456054687, 0.01844633674621582, 0.018480127334594726, 0.018456607818603515, 0.01846985626220703, 0.018428928375244142, 0.018431999206542968, 0.018909183502197266, 0.01845452880859375, 0.01840025520324707, 0.01847500801086426, 0.01843302345275879, 0.018447359085083007, 0.01845248031616211, 0.018421760559082033, 0.01845248031616211, 0.01845248031616211, 0.01841049575805664, 0.018455551147460936, 0.018391103744506837, 0.018453439712524413, 0.018481151580810547, 0.01843507194519043, 0.01846272087097168, 0.018471935272216796, 0.01852012825012207, 0.018537408828735353, 0.01840332794189453, 0.01842995262145996, 0.01842585563659668, 0.018507776260375978, 0.018486272811889647, 0.018449407577514648, 0.018455551147460936, 0.018579456329345705, 0.01843507194519043, 0.01846067237854004, 0.01843507194519043, 0.018449407577514648, 0.018431999206542968, 0.018437120437622072, 0.01843609619140625, 0.018428928375244142, 0.018663423538208008, 0.018478080749511717, 0.018453504562377928, 0.018449407577514648, 0.018479103088378905, 0.01848320007324219, 0.018426944732666015, 0.01841862487792969, 0.0184586238861084, 0.018548736572265623, 0.01842483139038086, 0.018523136138916017, 0.01847500801086426, 0.018473983764648438, 0.018465887069702147, 0.018414495468139648, 0.018473983764648438, 0.01845452880859375, 0.018778112411499022, 0.03877996826171875, 0.01850054359436035, 0.018472959518432617, 0.01847091293334961, 0.01841049575805664, 0.018469951629638673, 0.018467775344848635, 0.018440191268920898, 0.01843404769897461, 0.01843404769897461, 0.01859993553161621, 0.01864499282836914, 0.01846886444091797, 0.01906380844116211, 0.01860095977783203, 0.0184770565032959, 0.018586624145507814, 0.018533376693725585, 0.01843302345275879, 0.01843814468383789, 0.0182794246673584, 0.01837065505981445, 0.018466720581054686, 0.01844326400756836, 0.018535423278808593, 0.018521087646484375, 0.01843097686767578, 0.018496511459350586, 0.0184770565032959, 0.018731008529663085, 0.018544639587402344, 0.018488319396972656, 0.018480127334594726, 0.018737152099609376, 0.018786304473876952, 0.01843507194519043, 0.01843916893005371, 0.018413568496704103, 0.018478080749511717, 0.01840127944946289, 0.01842995262145996, 0.01839923286437988, 0.01841263961791992, 0.018468767166137694, 0.018487295150756835, 0.018437120437622072, 0.01849241638183594, 0.01841766357421875, 0.01844223976135254, 0.018423807144165038, 0.018428928375244142, 0.01841971206665039, 0.01887948799133301, 0.01847603225708008, 0.018481151580810547, 0.018397184371948243, 0.018431999206542968, 0.01839411163330078, 0.01839820861816406, 0.01843302345275879, 0.01843814468383789, 0.018465791702270508, 0.0184116153717041, 0.03870198440551758, 0.018505727767944336, 0.018463743209838866, 0.01842483139038086, 0.01846067237854004, 0.01840742492675781, 0.018502656936645507, 0.018465791702270508, 0.018471935272216796, 0.018499584197998048, 0.018421760559082033, 0.018455551147460936, 0.018415615081787108, 0.01843507194519043, 0.01841971206665039, 0.01850060844421387, 0.018417728424072265, 0.018436031341552736, 0.0184453125, 0.018503679275512695, 0.01841663932800293, 0.01841971206665039, 0.018510847091674804, 0.018480127334594726, 0.01864396858215332, 0.018501632690429686, 0.01845043182373047, 0.018503679275512695, 0.018226175308227538, 0.018330623626708984, 0.01848320007324219, 0.01846067237854004, 0.018522111892700196, 0.018927616119384767, 0.018724863052368163, 0.018456575393676757, 0.01838591957092285, 0.018397216796875, 0.01844220733642578, 0.01840025520324707, 0.018312192916870116, 0.018486272811889647, 0.018696256637573242, 0.01859987258911133, 0.018705440521240235, 0.018421728134155272, 0.018325504302978517, 0.01825382423400879, 0.018249727249145507, 0.018318336486816408, 0.01820467185974121, 0.018321407318115233, 0.018272256851196288, 0.01844428825378418, 0.018939903259277344, 0.018840576171875, 0.018480127334594726, 0.018486272811889647, 0.018461696624755858, 0.018874368667602538, 0.018914304733276367, 0.01856716728210449, 0.018502656936645507, 0.03866316986083984, 0.018490367889404297, 0.018498559951782227, 0.01862246322631836, 0.018948095321655273, 0.01875663948059082, 0.018512863159179688, 0.018510847091674804, 0.01840640068054199, 0.018463743209838866, 0.018428928375244142, 0.018463743209838866, 0.018885696411132812, 0.018785215377807616, 0.0184453125, 0.018440191268920898, 0.01844326400756836, 0.01843916893005371, 0.018464767456054687, 0.01841049575805664, 0.018899967193603515, 0.01897881507873535, 0.018525184631347655, 0.01840640068054199, 0.018463743209838866, 0.01841663932800293, 0.018395135879516602, 0.01843404769897461, 0.018739200592041014, 0.01866854476928711, 0.018449407577514648, 0.018440191268920898, 0.01884671974182129, 0.01884671974182129, 0.0186296329498291, 0.01845452880859375, 0.018465791702270508, 0.01845145606994629, 0.018478080749511717, 0.018554880142211915, 0.018903039932250978, 0.018525184631347655, 0.018480127334594726, 0.018397184371948243, 0.018494464874267577, 0.0188242244720459, 0.018550752639770508, 0.018479103088378905, 0.018447359085083007, 0.018521087646484375, 0.018530303955078126, 0.018420736312866212, 0.01846272087097168, 0.018373632431030275, 0.018495487213134765, 0.018415615081787108, 0.018606111526489257, 0.018444255828857423, 0.018486272811889647, 0.018553855895996094, 0.01847500801086426, 0.01865011215209961, 0.018542591094970702, 0.039008255004882815, 0.01843916893005371, 0.018572320938110353, 0.018645984649658203, 0.018721792221069337, 0.018486272811889647, 0.018408512115478514, 0.018450368881225587, 0.01859993553161621, 0.018506752014160157, 0.01842995262145996, 0.01848320007324219, 0.018528255462646484, 0.01882931137084961, 0.018415615081787108, 0.018464767456054687, 0.01842995262145996, 0.018480127334594726, 0.018689023971557618, 0.01867366409301758, 0.01841766357421875, 0.018509824752807616, 0.01886207962036133, 0.018561023712158203, 0.01927065658569336, 0.018735103607177735, 0.018537471771240235, 0.018486272811889647, 0.018457599639892578, 0.018547712326049806, 0.01882316780090332, 0.018619392395019533, 0.018562047958374024, 0.01861529541015625, 0.018619392395019533, 0.0187064323425293, 0.018515968322753908, 0.018487295150756835, 0.018381824493408205, 0.01860403251647949, 0.018646015167236327, 0.018466880798339844, 0.018506687164306642, 0.01840947151184082, 0.018453504562377928, 0.01860915184020996, 0.01865216064453125, 0.018463743209838866, 0.018577407836914063, 0.01884262466430664, 0.018671615600585938, 0.018929664611816405, 0.018790399551391602, 0.01878118324279785, 0.018528255462646484, 0.018726911544799805, 0.018951168060302736, 0.018900991439819336, 0.01901875114440918, 0.018754560470581053, 0.018983936309814452, 0.018982912063598634, 0.01886617660522461]",tokens/s,52.867723371954284,,,main,False,False -float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2232,7 +2232,7 @@ float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyT float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1409.2288,2403.86048,0.0,1757.413376,1301.782528,s,12,0.20092431831359867,0.016743693192799888,0.0005845947355924244,0.016412447929382325,0.017577545166015625,0.01788786821365356,0.01813288827896118,"[0.018194143295288085, 0.01703993606567383, 0.017637279510498045, 0.01698348808288574, 0.016280256271362304, 0.016399551391601562, 0.016408575057983397, 0.016375999450683593, 0.01641632080078125, 0.01635270309448242, 0.01640096092224121, 0.016435104370117186]",tokens/s,15289.338920166369,kWh,1.972021156705348e-07,1.0805755274807403e-07,6.430013491252678e-07,9.482610175438766e-07,tokens/kWh,269967862.5016922,MB,1409.2288,2403.86048,0.0,1757.413376,1324.7616,s,12,10.47739764404297,0.8731164703369142,0.012463386110935145,0.871078643798828,0.8861030456542969,0.8963321441650391,0.9051717022705079,"[0.907381591796875, 0.8711703491210937, 0.8872916870117188, 0.8711393432617187, 0.8620211791992187, 0.8750968017578125, 0.8642526245117188, 0.8656286010742188, 0.8710179443359375, 0.8599682006835937, 0.8670240478515625, 0.8754052734375]",tokens/s,72.15532193052073,kWh,1.0350986601054533e-05,5.671367858698863e-06,2.157481735104288e-05,3.759717181079627e-05,tokens/kWh,1675657.9541950852,,s,755,10.645324805259706,0.014099767953986363,0.002072898476441421,0.013693951606750488,0.014268006706237794,0.014471187019348145,0.03049243476867676,"[0.014705663681030273, 0.014632960319519044, 0.014416959762573242, 0.014392255783081054, 0.014154751777648926, 0.014309375762939454, 0.014519295692443847, 0.01438822364807129, 0.01427353572845459, 0.014445568084716797, 0.014393343925476074, 0.014351360321044922, 0.014326815605163573, 0.014907423973083495, 0.014781375885009766, 0.014867456436157226, 0.014431232452392579, 0.014438400268554688, 0.0142673921585083, 0.01437183952331543, 0.014529536247253418, 0.01427455997467041, 0.014268416404724121, 0.014256128311157227, 0.014230527877807618, 0.014222335815429688, 0.014404607772827148, 0.014523391723632812, 0.01447116756439209, 0.014431232452392579, 0.014433279991149902, 0.014166015625, 0.014306303977966308, 0.014012415885925293, 0.01410051155090332, 0.014724063873291015, 0.014698495864868164, 0.014879743576049804, 0.014593024253845215, 0.014564352035522461, 0.014632960319519044, 0.014234623908996581, 0.014155776023864745, 0.01419878387451172, 0.014146559715270996, 0.014079999923706055, 0.014119935989379882, 0.014319616317749024, 0.014210047721862793, 0.014202879905700684, 0.01417728042602539, 0.014112768173217773, 0.014146559715270996, 0.014128128051757812, 0.014600192070007324, 0.015132736206054688, 0.014814144134521484, 0.014343168258666992, 0.014254079818725587, 0.014228480339050293, 0.014252032279968262, 0.014241791725158692, 0.03164057540893555, 0.014221311569213867, 0.014102527618408203, 0.014171135902404786, 0.014193663597106934, 0.0141844482421875, 0.014132224082946777, 0.014152704238891601, 0.013701120376586913, 0.013706239700317382, 0.01367244815826416, 0.013661184310913087, 0.013692928314208984, 0.013634559631347656, 0.0136878080368042, 0.01365503978729248, 0.013643775939941406, 0.013627391815185547, 0.013661184310913087, 0.013625344276428223, 0.013674495697021484, 0.013752320289611816, 0.013720576286315917, 0.01368883228302002, 0.013664256095886231, 0.013664256095886231, 0.013619199752807617, 0.013683712005615235, 0.013740032196044923, 0.01365503978729248, 0.013674495697021484, 0.013652992248535157, 0.013678591728210449, 0.013606911659240722, 0.013622271537780761, 0.01363046360015869, 0.013688896179199219, 0.013662143707275391, 0.01367961597442627, 0.013623295783996582, 0.013685759544372558, 0.013658111572265624, 0.013594623565673827, 0.013684736251831055, 0.013670399665832519, 0.013578240394592284, 0.013699071884155273, 0.013613056182861329, 0.013648896217346192, 0.01365503978729248, 0.013649920463562011, 0.013689855575561523, 0.013637632369995116, 0.013644800186157227, 0.013636608123779297, 0.013639679908752441, 0.013623295783996582, 0.013640704154968262, 0.013664256095886231, 0.013670399665832519, 0.013669376373291015, 0.014028800010681153, 0.019300352096557616, 0.032573440551757815, 0.014261247634887696, 0.014099519729614257, 0.014151616096496582, 0.01417420768737793, 0.014179327964782714, 0.014352383613586426, 0.014143487930297852, 0.014120960235595703, 0.014172160148620605, 0.014102527618408203, 0.01439641571044922, 0.014232576370239258, 0.014088191986083985, 0.014140416145324707, 0.014141440391540527, 0.01417523193359375, 0.014153727531433105, 0.014125056266784668, 0.01417523193359375, 0.014141440391540527, 0.014104576110839843, 0.014156831741333007, 0.014214112281799316, 0.014169088363647461, 0.014213120460510254, 0.014171135902404786, 0.01419878387451172, 0.014145536422729492, 0.014308416366577148, 0.014143424034118652, 0.014193663597106934, 0.014216192245483398, 0.014167039871215821, 0.014216192245483398, 0.013705216407775878, 0.014040063858032227, 0.013678591728210449, 0.013702143669128418, 0.013699071884155273, 0.013637632369995116, 0.013656064033508301, 0.013709312438964843, 0.013692928314208984, 0.01430016040802002, 0.01415782356262207, 0.013697024345397948, 0.013647871971130371, 0.013656064033508301, 0.013745152473449707, 0.013841407775878906, 0.01417728042602539, 0.01426028823852539, 0.014168000221252442, 0.014231552124023437, 0.014161919593811035, 0.014187520027160644, 0.014138400077819824, 0.014204895973205567, 0.014178303718566895, 0.014166015625, 0.014193663597106934, 0.013710335731506347, 0.03161087989807129, 0.014070783615112305, 0.013612031936645508, 0.013648896217346192, 0.013717503547668456, 0.013637632369995116, 0.014560256004333496, 0.014256128311157227, 0.014317567825317384, 0.014187520027160644, 0.013700160026550293, 0.01366419219970703, 0.013799424171447755, 0.014566495895385742, 0.013682592391967773, 0.013919232368469238, 0.014119935989379882, 0.014135295867919923, 0.014183423995971679, 0.014183423995971679, 0.01415171241760254, 0.014193632125854493, 0.014161919593811035, 0.014127103805541993, 0.014155776023864745, 0.01417625617980957, 0.013743103981018067, 0.01367244815826416, 0.013675519943237305, 0.013661184310913087, 0.01376153564453125, 0.013639679908752441, 0.013647871971130371, 0.013667327880859375, 0.013669376373291015, 0.013652992248535157, 0.01366220760345459, 0.013651968002319336, 0.013676544189453126, 0.013700096130371094, 0.013608960151672364, 0.013692928314208984, 0.01368064022064209, 0.013691904067993164, 0.013652992248535157, 0.013620223999023438, 0.013623295783996582, 0.013647871971130371, 0.01365503978729248, 0.013683712005615235, 0.01370419216156006, 0.013656064033508301, 0.013607935905456543, 0.013596672058105469, 0.013624320030212403, 0.013619199752807617, 0.013692928314208984, 0.013656064033508301, 0.013703167915344238, 0.013726719856262207, 0.013697024345397948, 0.013637632369995116, 0.013667327880859375, 0.030373888015747072, 0.013656064033508301, 0.013631487846374512, 0.013651968002319336, 0.01368064022064209, 0.013697024345397948, 0.013669376373291015, 0.013636608123779297, 0.013661184310913087, 0.01363865566253662, 0.013664256095886231, 0.01364684772491455, 0.01363046360015869, 0.013675519943237305, 0.01367347240447998, 0.013789183616638183, 0.013643775939941406, 0.013700096130371094, 0.013686783790588379, 0.01368064022064209, 0.013659135818481445, 0.01367347240447998, 0.01367961597442627, 0.013668352127075196, 0.013643775939941406, 0.013624320030212403, 0.013664256095886231, 0.013708288192749024, 0.013807616233825683, 0.013690879821777344, 0.013698047637939453, 0.013656064033508301, 0.013675583839416503, 0.013654975891113282, 0.013692928314208984, 0.013677599906921387, 0.013669343948364257, 0.013613056182861329, 0.013727744102478028, 0.013675519943237305, 0.013705216407775878, 0.013718527793884277, 0.01376255989074707, 0.013684736251831055, 0.013721599578857421, 0.013749247550964355, 0.013674495697021484, 0.013710335731506347, 0.013679679870605468, 0.01367750358581543, 0.013644800186157227, 0.013690912246704101, 0.013679583549499512, 0.013713408470153808, 0.013660160064697266, 0.013658111572265624, 0.013635583877563476, 0.013644800186157227, 0.013644800186157227, 0.013685759544372558, 0.01367347240447998, 0.013633536338806153, 0.013651968002319336, 0.030510080337524413, 0.01359769630432129, 0.013637632369995116, 0.01368064022064209, 0.013633536338806153, 0.013670399665832519, 0.013633536338806153, 0.013642751693725585, 0.013647871971130371, 0.013639679908752441, 0.013618176460266113, 0.013666303634643554, 0.013957119941711426, 0.014775296211242676, 0.014713855743408203, 0.014245887756347657, 0.014179327964782714, 0.014214143753051758, 0.014208000183105468, 0.014145536422729492, 0.014137344360351562, 0.014107647895812989, 0.014128128051757812, 0.014220288276672363, 0.014130175590515137, 0.014222335815429688, 0.01420083236694336, 0.014121983528137207, 0.014140416145324707, 0.013974528312683105, 0.013905920028686524, 0.014087167739868164, 0.014088191986083985, 0.014150655746459961, 0.014471232414245605, 0.014548928260803222, 0.014204928398132324, 0.01415987205505371, 0.013767680168151856, 0.013620287895202636, 0.013629376411437988, 0.013637632369995116, 0.013721599578857421, 0.013661184310913087, 0.01370419216156006, 0.013661184310913087, 0.013692928314208984, 0.013658111572265624, 0.013641728401184081, 0.01366220760345459, 0.013623295783996582, 0.013637632369995116, 0.013644800186157227, 0.01363046360015869, 0.013632512092590332, 0.013693951606750488, 0.013657088279724122, 0.01373084831237793, 0.013635552406311036, 0.013610048294067383, 0.013654975891113282, 0.013621248245239258, 0.01368166446685791, 0.030512128829956055, 0.013631520271301269, 0.013653984069824219, 0.013642751693725585, 0.013632512092590332, 0.013700096130371094, 0.013603839874267578, 0.01366220760345459, 0.013629440307617188, 0.013697024345397948, 0.013683712005615235, 0.013813759803771973, 0.013686783790588379, 0.013682687759399414, 0.013627391815185547, 0.013643775939941406, 0.013648896217346192, 0.013674495697021484, 0.013598719596862792, 0.013697024345397948, 0.013709312438964843, 0.013669376373291015, 0.013647871971130371, 0.013661184310913087, 0.01366323184967041, 0.013621248245239258, 0.01368166446685791, 0.013708288192749024, 0.013648896217346192, 0.01367142391204834, 0.013691935539245605, 0.013931488037109375, 0.01375334358215332, 0.013651968002319336, 0.013669376373291015, 0.013744128227233888, 0.013641728401184081, 0.013642751693725585, 0.01358950424194336, 0.013714431762695312, 0.013628416061401367, 0.01363871955871582, 0.01360377597808838, 0.013645824432373046, 0.014258175849914552, 0.014231552124023437, 0.014106623649597168, 0.014112768173217773, 0.014113792419433594, 0.014199808120727539, 0.013643775939941406, 0.013726719856262207, 0.013635583877563476, 0.013625344276428223, 0.013649920463562011, 0.0136878080368042, 0.013631487846374512, 0.01359769630432129, 0.013755423545837402, 0.013459424018859863, 0.013518912315368652, 0.013725631713867188, 0.013658111572265624, 0.030483455657958985, 0.013664256095886231, 0.013645824432373046, 0.013864959716796875, 0.013658111572265624, 0.013682687759399414, 0.013627391815185547, 0.01363257598876953, 0.013691840171813964, 0.013911040306091308, 0.014115839958190919, 0.014071807861328126, 0.014092288017272948, 0.01429094409942627, 0.01417420768737793, 0.014114815711975098, 0.013666303634643554, 0.01364684772491455, 0.013710335731506347, 0.013643775939941406, 0.013657088279724122, 0.013640704154968262, 0.013684736251831055, 0.013631487846374512, 0.013724672317504882, 0.013611007690429687, 0.013648896217346192, 0.013643775939941406, 0.013708288192749024, 0.013719552040100098, 0.013682687759399414, 0.013616127967834473, 0.013668352127075196, 0.013629440307617188, 0.013669376373291015, 0.013660160064697266, 0.01363046360015869, 0.013675519943237305, 0.013636608123779297, 0.013518879890441895, 0.013538271903991698, 0.01365401554107666, 0.013634559631347656, 0.013656064033508301, 0.013670399665832519, 0.013604864120483399, 0.013684736251831055, 0.013747200012207032, 0.013735936164855958, 0.01365401554107666, 0.013676544189453126, 0.013658111572265624, 0.01377996826171875, 0.01427455997467041, 0.014128128051757812, 0.014007295608520508, 0.013503487586975099, 0.013629440307617188, 0.013716480255126954, 0.013656064033508301, 0.013667327880859375, 0.013644800186157227, 0.01366323184967041, 0.030502975463867186, 0.01367955207824707, 0.013595647811889648, 0.013651968002319336, 0.013606911659240722, 0.01367244815826416, 0.013631487846374512, 0.013664256095886231, 0.01363865566253662, 0.013643775939941406, 0.013598719596862792, 0.013618176460266113, 0.013642751693725585, 0.013689855575561523, 0.013645824432373046, 0.013623295783996582, 0.013737983703613281, 0.013752320289611816, 0.013656064033508301, 0.013596672058105469, 0.01364684772491455, 0.013616127967834473, 0.013684736251831055, 0.013656064033508301, 0.014016511917114258, 0.013742079734802246, 0.015156224250793457, 0.014486528396606446, 0.014292991638183594, 0.014235648155212402, 0.014156800270080566, 0.01360588836669922, 0.013707263946533203, 0.01367142391204834, 0.013633536338806153, 0.013642751693725585, 0.013598719596862792, 0.013723648071289063, 0.013882368087768555, 0.014111743927001954, 0.014193663597106934, 0.014116864204406738, 0.014128128051757812, 0.014109760284423829, 0.013642687797546387, 0.013706239700317382, 0.013608960151672364, 0.013640704154968262, 0.013864959716796875, 0.013759488105773926, 0.013676544189453126, 0.013721599578857421, 0.013670399665832519, 0.01368883228302002, 0.013652992248535157, 0.014693375587463378, 0.014191616058349609, 0.014254079818725587, 0.014138367652893067, 0.013615103721618652, 0.013707263946533203, 0.013793279647827148, 0.013750271797180176, 0.03043328094482422, 0.013585408210754395, 0.013641728401184081, 0.013623295783996582, 0.013744128227233888, 0.013713408470153808, 0.01372265625, 0.013668319702148438, 0.013683712005615235, 0.01364684772491455, 0.013624320030212403, 0.013634559631347656, 0.013637632369995116, 0.013637632369995116, 0.013632512092590332, 0.013657088279724122, 0.013640704154968262, 0.01363865566253662, 0.013776896476745605, 0.013639679908752441, 0.01365503978729248, 0.013621248245239258, 0.01365401554107666, 0.013637696266174316, 0.013674431800842286, 0.013685759544372558, 0.01367347240447998, 0.013616127967834473, 0.013478912353515626, 0.01344819164276123, 0.013637696266174316, 0.013530048370361328, 0.013593600273132325, 0.013678591728210449, 0.013645824432373046, 0.013633536338806153, 0.013621248245239258, 0.013647871971130371, 0.01365401554107666, 0.013643775939941406, 0.013617152214050294, 0.013726719856262207, 0.013623295783996582, 0.013659135818481445, 0.013664256095886231, 0.013709312438964843, 0.013676544189453126, 0.013590527534484862, 0.01366323184967041, 0.013670399665832519, 0.013714431762695312, 0.013631487846374512, 0.01363865566253662, 0.013693951606750488, 0.01377996826171875, 0.013677568435668945, 0.013625344276428223, 0.013642751693725585, 0.013664256095886231, 0.013628416061401367, 0.013642751693725585, 0.013484031677246093, 0.013549568176269532, 0.03052448081970215, 0.013649855613708496, 0.013633536338806153, 0.013502464294433594, 0.013503487586975099, 0.013628416061401367, 0.013903871536254882, 0.013701120376586913, 0.013594623565673827, 0.013512767791748048, 0.01355667209625244, 0.01366220760345459, 0.01377791976928711, 0.013627391815185547, 0.013657088279724122, 0.0141844482421875, 0.014118911743164063, 0.014150655746459961, 0.014105600357055664, 0.014138367652893067, 0.014142463684082032, 0.01417728042602539, 0.014208000183105468, 0.014123007774353028, 0.01367961597442627, 0.01364684772491455, 0.013851648330688476, 0.013729855537414551, 0.013717439651489258, 0.01369808006286621, 0.013657055854797363, 0.01368064022064209, 0.013719552040100098, 0.013684736251831055, 0.01368064022064209, 0.013650943756103515, 0.013739007949829102, 0.013710335731506347, 0.013690879821777344, 0.013708288192749024, 0.013667327880859375, 0.013632512092590332, 0.01370419216156006, 0.013832192420959472, 0.013745152473449707, 0.013731840133666993, 0.013822976112365723, 0.013692928314208984, 0.013713408470153808, 0.013742079734802246, 0.013706239700317382, 0.01376255989074707, 0.013720576286315917, 0.01367244815826416, 0.013732864379882812, 0.013719552040100098, 0.013750271797180176, 0.013684736251831055, 0.013742143630981446, 0.013717439651489258, 0.013693984031677247, 0.013608927726745606, 0.01360588836669922, 0.03060736083984375, 0.013993984222412109, 0.013774847984313965, 0.013719552040100098, 0.013713408470153808, 0.013724672317504882, 0.013661184310913087, 0.01376153564453125, 0.013720576286315917, 0.013742079734802246, 0.013689888000488282, 0.013784031867980957, 0.013717503547668456, 0.013714431762695312, 0.014386176109313965, 0.01429100799560547, 0.013773759841918946, 0.013821951866149903, 0.013708288192749024, 0.013699071884155273, 0.013805567741394043, 0.013757439613342285, 0.01375436782836914, 0.01429094409942627, 0.014359552383422852, 0.014256128311157227, 0.014210047721862793, 0.014270463943481445, 0.013602815628051757, 0.013836288452148437, 0.014110719680786133, 0.014208000183105468, 0.014219264030456542, 0.014232576370239258, 0.01437286376953125, 0.014454848289489746, 0.014202816009521484, 0.014205951690673829, 0.01409331226348877, 0.014435327529907227, 0.014436351776123046, 0.013774847984313965, 0.013735936164855958, 0.013751296043395997, 0.013711359977722168, 0.013729791641235351, 0.013737983703613281, 0.01374617576599121, 0.013714431762695312, 0.013716480255126954, 0.01367244815826416, 0.013723648071289063, 0.013692928314208984, 0.013703167915344238, 0.013792320251464844, 0.013847488403320313, 0.013676544189453126, 0.0136079683303833, 0.013741024017333984, 0.013666303634643554, 0.013679648399353027, 0.013644767761230468, 0.013708288192749024]",tokens/s,70.92315300956953,,,main,False,False float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1188.442112,1455.947776,0.0,809.500672,723.533824,s,13,0.19395561599731448,0.014919662769024188,0.0006263890092848103,0.014573216438293457,0.015982310676574706,0.016239801406860353,0.01632139320373535,"[0.016341791152954102, 0.014573216438293457, 0.014939104080200195, 0.014454079627990722, 0.01617180824279785, 0.014889568328857421, 0.015062560081481933, 0.014445055961608886, 0.015224320411682129, 0.014482432365417481, 0.014442303657531737, 0.014444767951965332, 0.014484607696533203]",tokens/s,17158.56477208724,kWh,1.7314691897115033e-07,9.487634185087318e-08,4.882458761990672e-07,7.562691370210907e-07,tokens/kWh,338503830.80337274,MB,1188.442112,1455.947776,0.0,809.500672,749.21984,s,13,10.502904846191404,0.8079157573993387,0.009505664839521193,0.80391015625,0.823531689453125,0.8280160888671875,0.8301153271484375,"[0.8262667236328125, 0.812591552734375, 0.7997078247070313, 0.8014157104492188, 0.8047113037109375, 0.8048790283203126, 0.8026774291992187, 0.80391015625, 0.83064013671875, 0.799624755859375, 0.8023453369140625, 0.8034739990234375, 0.810660888671875]",tokens/s,77.97842711076149,kWh,9.53687434708987e-06,5.225731663599511e-06,1.7475379605797372e-05,3.223798561648675e-05,tokens/kWh,1954216.3939604627,,s,818,10.666156065940855,0.013039310594059727,0.0018623556299793502,0.01268838405609131,0.013345097541809082,0.013433292531967162,0.02762204216003418,"[0.014240768432617188, 0.014169088363647461, 0.01459404754638672, 0.013119487762451172, 0.012732416152954102, 0.012678144454956054, 0.01267404842376709, 0.012804096221923827, 0.012634112358093261, 0.012650495529174804, 0.012620800018310547, 0.01265664005279541, 0.012689408302307128, 0.012697600364685058, 0.012653568267822265, 0.012718079566955566, 0.012681216239929198, 0.01273036766052246, 0.01273855972290039, 0.01277132797241211, 0.01287065601348877, 0.012635135650634765, 0.012709888458251953, 0.012792832374572754, 0.012743680000305176, 0.012625920295715331, 0.012667903900146485, 0.013364224433898926, 0.013499423980712891, 0.013488096237182617, 0.01343283176422119, 0.013444095611572266, 0.013674495697021484, 0.013442048072814941, 0.013431808471679688, 0.013444095611572266, 0.01338265609741211, 0.013436927795410156, 0.013435903549194337, 0.013356032371520997, 0.013319168090820312, 0.01337446403503418, 0.013344767570495606, 0.013794303894042969, 0.013689855575561523, 0.013816831588745117, 0.013475839614868163, 0.013178879737854005, 0.013119487762451172, 0.01305292797088623, 0.0130764799118042, 0.013071359634399414, 0.012957695960998536, 0.01304371166229248, 0.012981247901916505, 0.012682239532470703, 0.012643327713012695, 0.012678144454956054, 0.01263923168182373, 0.012892160415649414, 0.012887040138244628, 0.012646431922912598, 0.027743200302124023, 0.012618816375732422, 0.012636096000671386, 0.012667936325073242, 0.012675040245056153, 0.01268019199371338, 0.012786687850952149, 0.013175807952880859, 0.013033472061157226, 0.013027327537536621, 0.01287168025970459, 0.012644351959228516, 0.012681216239929198, 0.012724224090576173, 0.012677120208740235, 0.012673055648803711, 0.013089759826660156, 0.013005824089050292, 0.01303756809234619, 0.013096960067749023, 0.01306828784942627, 0.01285529613494873, 0.013398015975952148, 0.013198335647583008, 0.013189120292663574, 0.013041664123535156, 0.012875776290893554, 0.012718079566955566, 0.012957695960998536, 0.012887040138244628, 0.013853759765625, 0.013312959671020507, 0.01313587188720703, 0.012958720207214355, 0.012652544021606446, 0.012717056274414062, 0.012744704246520995, 0.012703743934631348, 0.012854304313659668, 0.012755935668945312, 0.012660736083984376, 0.012723199844360352, 0.012751872062683106, 0.013008895874023438, 0.013055999755859376, 0.013071359634399414, 0.013191167831420898, 0.012981247901916505, 0.0129617919921875, 0.012751872062683106, 0.012683327674865724, 0.01267807960510254, 0.012695551872253418, 0.012652544021606446, 0.012698623657226562, 0.012679167747497559, 0.01267199993133545, 0.012724224090576173, 0.01276211166381836, 0.01316864013671875, 0.013163552284240723, 0.013011936187744141, 0.013031423568725586, 0.02811494445800781, 0.012653568267822265, 0.012625920295715331, 0.012712960243225097, 0.01267199993133545, 0.012660736083984376, 0.012636159896850586, 0.012699647903442383, 0.012702719688415527, 0.012731424331665039, 0.012691424369812012, 0.012652544021606446, 0.012620800018310547, 0.012661760330200195, 0.012686335563659668, 0.012689408302307128, 0.01266483211517334, 0.012604415893554688, 0.01264742374420166, 0.012640255928039551, 0.012633088111877442, 0.013162495613098145, 0.012964863777160645, 0.01267404842376709, 0.012651519775390625, 0.012662783622741699, 0.012654591560363769, 0.012649472236633302, 0.012652544021606446, 0.01264844799041748, 0.012696576118469239, 0.012709888458251953, 0.012782591819763184, 0.01273036766052246, 0.012650495529174804, 0.012634112358093261, 0.012666879653930664, 0.012620800018310547, 0.012634112358093261, 0.012636159896850586, 0.012592127799987793, 0.012606464385986327, 0.012710911750793457, 0.012608511924743653, 0.012685312271118163, 0.012636159896850586, 0.01267404842376709, 0.012951552391052246, 0.013063167572021485, 0.012712960243225097, 0.012618751525878906, 0.012733440399169921, 0.012619775772094726, 0.012630016326904296, 0.012682239532470703, 0.012653568267822265, 0.012666879653930664, 0.012626943588256835, 0.012678144454956054, 0.012660736083984376, 0.012755968093872071, 0.012668928146362305, 0.012577856063842773, 0.027557823181152345, 0.01268019199371338, 0.012681247711181641, 0.012670944213867188, 0.012637184143066407, 0.012636159896850586, 0.012627967834472656, 0.012699647903442383, 0.012670975685119629, 0.012638208389282226, 0.01275699234008789, 0.012728320121765137, 0.012866559982299805, 0.012702719688415527, 0.012619775772094726, 0.01268838405609131, 0.012700672149658204, 0.012640255928039551, 0.012675071716308594, 0.01264742374420166, 0.012584959983825684, 0.012597248077392579, 0.012613632202148438, 0.012653568267822265, 0.012667936325073242, 0.012670944213867188, 0.012726271629333496, 0.01266585636138916, 0.012860416412353515, 0.012805120468139648, 0.01266483211517334, 0.012630016326904296, 0.01263923168182373, 0.012644351959228516, 0.012660736083984376, 0.012700703620910645, 0.012637151718139648, 0.012637184143066407, 0.012658687591552734, 0.012681216239929198, 0.012719103813171387, 0.01268838405609131, 0.01268838405609131, 0.012645376205444337, 0.012675071716308594, 0.012662783622741699, 0.0126310396194458, 0.012697600364685058, 0.012628000259399414, 0.012675104141235351, 0.012633024215698242, 0.012652544021606446, 0.01267199993133545, 0.012668928146362305, 0.012744704246520995, 0.012694527626037597, 0.012626943588256835, 0.01277337646484375, 0.01297920036315918, 0.013008895874023438, 0.013016063690185547, 0.013484031677246093, 0.013275135993957519, 0.029944831848144532, 0.013188096046447753, 0.013150208473205567, 0.013015040397644043, 0.012989439964294434, 0.012986368179321289, 0.013121536254882812, 0.013156352043151855, 0.013131775856018067, 0.013018112182617187, 0.013028351783752442, 0.01276313591003418, 0.012759039878845215, 0.012679200172424316, 0.01259830379486084, 0.012623807907104493, 0.012736512184143066, 0.01276313591003418, 0.013088768005371093, 0.012916735649108887, 0.012654591560363769, 0.012597248077392579, 0.012613632202148438, 0.012823552131652831, 0.013117440223693847, 0.012744704246520995, 0.012619775772094726, 0.012611583709716797, 0.012609536170959473, 0.012619775772094726, 0.012646400451660156, 0.012701696395874023, 0.012603391647338867, 0.012619775772094726, 0.012678144454956054, 0.012645376205444337, 0.01266380786895752, 0.012632063865661621, 0.01264742374420166, 0.012626943588256835, 0.012624896049499512, 0.012678144454956054, 0.012712960243225097, 0.012628992080688477, 0.012992544174194336, 0.012634079933166504, 0.012583935737609863, 0.012695551872253418, 0.012606528282165528, 0.012789759635925293, 0.012874688148498535, 0.01263315200805664, 0.012698559761047364, 0.01266380786895752, 0.01265664005279541, 0.012699647903442383, 0.012698623657226562, 0.012641280174255372, 0.012625920295715331, 0.012706815719604492, 0.012696576118469239, 0.01258188819885254, 0.012677120208740235, 0.02831667137145996, 0.012606464385986327, 0.01265664005279541, 0.012679167747497559, 0.012706815719604492, 0.01266585636138916, 0.012889087677001953, 0.012800000190734863, 0.012645376205444337, 0.012679167747497559, 0.012709888458251953, 0.012957695960998536, 0.013225983619689942, 0.012997695922851563, 0.013011903762817382, 0.012888064384460449, 0.012697600364685058, 0.012643327713012695, 0.012949503898620606, 0.013002816200256348, 0.012976063728332519, 0.012841983795166016, 0.012986368179321289, 0.012749823570251465, 0.013048831939697265, 0.012907520294189453, 0.01277337646484375, 0.012610560417175292, 0.012690431594848632, 0.01277132797241211, 0.01277952003479004, 0.012600383758544921, 0.012631999969482421, 0.012697600364685058, 0.012607487678527832, 0.012606464385986327, 0.012785663604736328, 0.012637184143066407, 0.012717056274414062, 0.012916735649108887, 0.012967935562133789, 0.012820480346679687, 0.01265664005279541, 0.012654623985290527, 0.012633055686950683, 0.012536831855773926, 0.012596223831176758, 0.012686335563659668, 0.012594176292419433, 0.012640255928039551, 0.01259008026123047, 0.012613632202148438, 0.012716032028198243, 0.012677120208740235, 0.013054976463317871, 0.01303756809234619, 0.013161503791809082, 0.01261257553100586, 0.012543999671936035, 0.012634112358093261, 0.01265664005279541, 0.01267302417755127, 0.012973055839538575, 0.02857369613647461, 0.012920831680297852, 0.01265664005279541, 0.01267199993133545, 0.012651519775390625, 0.012661760330200195, 0.012716032028198243, 0.012653568267822265, 0.012934176445007325, 0.01292899227142334, 0.012678144454956054, 0.012637184143066407, 0.01266585636138916, 0.012812288284301757, 0.012645376205444337, 0.012654591560363769, 0.012593152046203614, 0.01264844799041748, 0.01265766429901123, 0.012711935997009278, 0.012685312271118163, 0.012659711837768555, 0.01264134407043457, 0.01267091178894043, 0.012679167747497559, 0.01275494384765625, 0.012920831680297852, 0.012644351959228516, 0.01265766429901123, 0.012677120208740235, 0.012643327713012695, 0.012636159896850586, 0.012632063865661621, 0.012723199844360352, 0.012698623657226562, 0.012840959548950195, 0.013023232460021973, 0.012902400016784669, 0.012978176116943359, 0.012928000450134277, 0.01265766429901123, 0.012686335563659668, 0.013231103897094726, 0.013021183967590331, 0.01307033634185791, 0.01287065601348877, 0.01267302417755127, 0.012659711837768555, 0.012668928146362305, 0.012627967834472656, 0.012677120208740235, 0.012650495529174804, 0.01285324764251709, 0.0126310396194458, 0.012684288024902344, 0.012651583671569825, 0.012615615844726563, 0.012626943588256835, 0.012667903900146485, 0.012568575859069824, 0.012666879653930664, 0.012611583709716797, 0.012628992080688477, 0.02757529640197754, 0.012661760330200195, 0.0126310396194458, 0.012653568267822265, 0.012684288024902344, 0.01265664005279541, 0.012895232200622558, 0.012717056274414062, 0.012638208389282226, 0.012760064125061036, 0.012695551872253418, 0.012587008476257324, 0.012679167747497559, 0.012686335563659668, 0.013087743759155274, 0.012921855926513673, 0.012700672149658204, 0.012634112358093261, 0.01266585636138916, 0.012699647903442383, 0.012696576118469239, 0.012741632461547851, 0.012981247901916505, 0.012956671714782715, 0.012623871803283691, 0.012635135650634765, 0.012576767921447754, 0.012645376205444337, 0.012579839706420898, 0.01263923168182373, 0.012615679740905762, 0.012609536170959473, 0.012608511924743653, 0.012613632202148438, 0.012637184143066407, 0.012741632461547851, 0.012616703987121582, 0.012633088111877442, 0.012609536170959473, 0.012736512184143066, 0.01306828784942627, 0.01283993625640869, 0.012678144454956054, 0.012614656448364257, 0.012616703987121582, 0.01255731201171875, 0.012610560417175292, 0.012644351959228516, 0.01273036766052246, 0.012637184143066407, 0.01265664005279541, 0.012678144454956054, 0.012640255928039551, 0.012692480087280274, 0.012614656448364257, 0.01265664005279541, 0.012595199584960937, 0.013418496131896973, 0.013426688194274903, 0.013331487655639648, 0.013219807624816894, 0.0133570556640625, 0.013368320465087891, 0.029088768005371093, 0.013415424346923829, 0.013373439788818359, 0.013381631851196289, 0.013415424346923829, 0.01336627197265625, 0.013334527969360351, 0.013522944450378417, 0.013331456184387207, 0.01347379207611084, 0.01335910415649414, 0.013396991729736327, 0.013362175941467285, 0.013454336166381836, 0.013419520378112794, 0.013431808471679688, 0.013356032371520997, 0.013361151695251466, 0.013379584312438965, 0.013342720031738281, 0.013344799995422364, 0.01334169578552246, 0.013369312286376953, 0.013345791816711425, 0.013394944190979004, 0.01335807991027832, 0.01334988784790039, 0.01346457576751709, 0.013472767829895019, 0.013459456443786622, 0.013431808471679688, 0.013305855751037597, 0.013364224433898926, 0.013399040222167969, 0.013304832458496094, 0.013323264122009277, 0.013394944190979004, 0.013298687934875488, 0.01339187240600586, 0.013467647552490235, 0.013634559631347656, 0.012940287590026855, 0.013379584312438965, 0.013088768005371093, 0.013044735908508302, 0.012978176116943359, 0.013116415977478027, 0.01293824005126953, 0.012650495529174804, 0.012600319862365723, 0.012641280174255372, 0.01263923168182373, 0.012625920295715331, 0.012577823638916016, 0.012571616172790527, 0.012713983535766601, 0.012654591560363769, 0.012658687591552734, 0.012650495529174804, 0.012661760330200195, 0.01267302417755127, 0.012633088111877442, 0.012692480087280274, 0.02755379295349121, 0.012614656448364257, 0.012627967834472656, 0.01264742374420166, 0.012670975685119629, 0.012587039947509766, 0.012592096328735352, 0.012668928146362305, 0.012605440139770508, 0.012703743934631348, 0.012628992080688477, 0.012661760330200195, 0.01264742374420166, 0.012667903900146485, 0.012621824264526366, 0.012642304420471191, 0.012634112358093261, 0.012627967834472656, 0.012626943588256835, 0.012633119583129883, 0.012674015998840332, 0.012614656448364257, 0.012769344329833985, 0.012698559761047364, 0.012705792427062988, 0.012649472236633302, 0.012717056274414062, 0.01264844799041748, 0.012660736083984376, 0.012628992080688477, 0.012652607917785644, 0.012702655792236329, 0.012614656448364257, 0.012883968353271484, 0.012732416152954102, 0.01266585636138916, 0.01266585636138916, 0.01264742374420166, 0.012682239532470703, 0.01264844799041748, 0.012615679740905762, 0.012621824264526366, 0.012618751525878906, 0.012709888458251953, 0.012693504333496093, 0.012662783622741699, 0.012726271629333496, 0.012676159858703613, 0.012638143539428711, 0.012684288024902344, 0.012635135650634765, 0.012579839706420898, 0.012618751525878906, 0.012637184143066407, 0.012933119773864746, 0.012983296394348144, 0.01298739242553711, 0.012948479652404785, 0.012741632461547851, 0.01287782382965088, 0.0127293758392334, 0.01274672031402588, 0.01276416015625, 0.02775551986694336, 0.013016063690185547, 0.013085696220397949, 0.013007871627807617, 0.012955648422241211, 0.012884991645812988, 0.013025279998779296, 0.013073408126831054, 0.012821503639221191, 0.012677120208740235, 0.012650495529174804, 0.01266380786895752, 0.012933152198791504, 0.013019136428833008, 0.012878815650939941, 0.0127457275390625, 0.012684288024902344, 0.012620800018310547, 0.012676095962524414, 0.012650495529174804, 0.012615679740905762, 0.012611583709716797, 0.012623871803283691, 0.012766207695007324, 0.012654591560363769, 0.012570624351501464, 0.012652544021606446, 0.01262284755706787, 0.012633088111877442, 0.012670975685119629, 0.012620800018310547, 0.012747776031494141, 0.012667903900146485, 0.0126310396194458, 0.01264742374420166, 0.012791808128356934, 0.01293619155883789, 0.01277132797241211, 0.012660799980163574, 0.012624832153320312, 0.012646400451660156, 0.012686335563659668, 0.012658687591552734, 0.012612607955932617, 0.012638208389282226, 0.012646400451660156, 0.012668928146362305, 0.012695551872253418, 0.012690431594848632, 0.01266585636138916, 0.012695551872253418, 0.012638208389282226, 0.012676159858703613, 0.01266374397277832, 0.012600319862365723, 0.012750847816467285, 0.012659711837768555, 0.01269148826599121, 0.012680159568786622, 0.012645376205444337, 0.01285529613494873, 0.012615679740905762, 0.012691455841064453, 0.027667455673217774, 0.012662783622741699, 0.012801024436950683, 0.012718079566955566, 0.012702719688415527, 0.012652544021606446, 0.012624896049499512, 0.012679167747497559, 0.01263923168182373, 0.01264844799041748, 0.012722175598144531, 0.012602368354797363, 0.012630016326904296, 0.01266585636138916, 0.012653568267822265, 0.012707839965820313, 0.012662783622741699, 0.012679167747497559, 0.012661760330200195, 0.012652544021606446, 0.012644351959228516, 0.01265766429901123, 0.012623871803283691, 0.012640255928039551, 0.012630016326904296, 0.01267404842376709, 0.012741632461547851, 0.012731391906738282, 0.013028351783752442, 0.012931072235107421, 0.013342720031738281, 0.01307852840423584, 0.012836864471435547, 0.012995583534240723, 0.012844032287597656, 0.012902400016784669, 0.012878848075866698, 0.01296895980834961, 0.012806143760681152, 0.012687359809875488, 0.012687359809875488, 0.012642304420471191, 0.012614656448364257, 0.012822527885437012, 0.012659711837768555, 0.012650495529174804, 0.01287168025970459, 0.012810239791870117, 0.012850175857543946, 0.012685312271118163, 0.012826623916625977, 0.012894207954406739, 0.01266380786895752, 0.012678144454956054, 0.012677120208740235, 0.012659711837768555, 0.01275494384765625, 0.012603391647338867, 0.012624896049499512, 0.012687359809875488, 0.012644351959228516, 0.01284505558013916, 0.013010944366455078, 0.027631616592407225, 0.012638208389282226, 0.01266585636138916, 0.012693504333496093, 0.01266483211517334, 0.012918784141540527, 0.013204480171203613, 0.013760512351989745, 0.013348863601684571, 0.013119487762451172, 0.01306220817565918, 0.01278048038482666, 0.012691455841064453, 0.012751872062683106, 0.012723199844360352, 0.012685312271118163, 0.012743680000305176, 0.012720128059387208, 0.01276211166381836, 0.012649472236633302, 0.012650495529174804, 0.012619839668273927, 0.012644288063049316, 0.012620832443237305, 0.012706784248352051, 0.012622912406921387, 0.012762047767639161, 0.012717056274414062, 0.01264844799041748, 0.012726271629333496, 0.012751872062683106, 0.012689408302307128, 0.012744704246520995, 0.012652544021606446, 0.012651519775390625, 0.012691455841064453, 0.012668928146362305, 0.01267199993133545, 0.012662783622741699, 0.012638208389282226, 0.012719103813171387, 0.01287782382965088, 0.013023232460021973, 0.012935168266296386, 0.013083647727966309, 0.013044735908508302, 0.012875776290893554, 0.012675104141235351, 0.012684255599975587, 0.012780544281005859, 0.01268838405609131, 0.01266380786895752, 0.012666879653930664, 0.01264844799041748, 0.01266380786895752, 0.013351936340332032, 0.01343283176422119, 0.013469696044921875, 0.0134717435836792, 0.013395968437194825, 0.013404159545898438, 0.013455360412597657, 0.013386783599853515]",tokens/s,76.69117111571578,,,main,False,False float16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1210.916864,14382.792704,0.0,13736.3456,13487.531008,s,10,0.9088285980224609,0.0908828598022461,0.0011188802829312761,0.0904892463684082,0.091262548828125,0.09272698516845702,0.09389853424072266,"[0.09419142150878906, 0.09067654418945313, 0.09067558288574219, 0.09093711853027343, 0.09034416198730469, 0.09030604553222656, 0.09044735717773437, 0.09036310577392578, 0.09035612487792968, 0.09053113555908203]",tokens/s,2816.812769283842,kWh,1.0696121238850738e-06,5.861005212913417e-07,5.323087341549429e-06,6.978799986725844e-06,tokens/kWh,36682524.2859703,MB,1210.916864,14382.792704,0.0,13736.3456,13661.259776,s,10,22.005575439453125,2.2005575439453127,0.0067534240568617905,2.197949462890625,2.2105153076171873,2.210736901855469,2.2109141772460936,"[2.2097392578125, 2.21095849609375, 2.210466064453125, 2.192959716796875, 2.19519482421875, 2.199596923828125, 2.1960966796875, 2.196302001953125, 2.194256591796875, 2.2000048828125]",tokens/s,28.629108188213618,kWh,2.591680861017606e-05,1.4203239006414692e-05,0.00011372264678385089,0.00015384269440044164,tokens/kWh,409509.2083866879,,s,629,22.785184753417955,0.036224459067437154,0.010793842260097662,0.034890750885009765,0.0351723518371582,0.035259391784667964,0.12553335479736327,"[0.034729984283447264, 0.03492147064208984, 0.03466239929199219, 0.034726913452148435, 0.03465427017211914, 0.03471558380126953, 0.034783233642578126, 0.034738208770751955, 0.03466134262084961, 0.034802688598632815, 0.034715648651123046, 0.03478227233886719, 0.03479750442504883, 0.03497983932495117, 0.035151870727539065, 0.0347955207824707, 0.034890750885009765, 0.03502592086791992, 0.034872318267822264, 0.03528396987915039, 0.0354334716796875, 0.03521331024169922, 0.03520000076293945, 0.03530342483520508, 0.03505254364013672, 0.0351907844543457, 0.03506892776489258, 0.035160064697265625, 0.035105792999267575, 0.0351016960144043, 0.035253246307373046, 0.03513139343261719, 0.03510476684570313, 0.03511500930786133, 0.03514572906494141, 0.03513651275634765, 0.035149822235107424, 0.035151870727539065, 0.035266559600830076, 0.03520512008666992, 0.035113983154296875, 0.03518463897705078, 0.03514470291137695, 0.03523379135131836, 0.035110912322998046, 0.03525222396850586, 0.03516211318969727, 0.03523075103759766, 0.035122142791748044, 0.03521129608154297, 0.03513238525390625, 0.035250175476074216, 0.03515903854370117, 0.035326976776123044, 0.035266559600830076, 0.035236862182617186, 0.03511603164672852, 0.03519385528564453, 0.035176448822021485, 0.035269630432128905, 0.035053569793701174, 0.035266559600830076, 0.1260953598022461, 0.0350013427734375, 0.034955265045166016, 0.03492454528808594, 0.0350013427734375, 0.034988033294677735, 0.034985984802246094, 0.03500236892700195, 0.03505254364013672, 0.03496345520019531, 0.03504435348510742, 0.03492454528808594, 0.0350832633972168, 0.035020801544189455, 0.03505459213256836, 0.03502489471435547, 0.03504540634155273, 0.035012577056884764, 0.03504230499267578, 0.03498393630981445, 0.035026943206787106, 0.03497574234008789, 0.03504230499267578, 0.035053569793701174, 0.03501670455932617, 0.03504947280883789, 0.0351099853515625, 0.03498179244995117, 0.03516108703613281, 0.0350013427734375, 0.03510988616943359, 0.03512627029418945, 0.03508736038208008, 0.03502489471435547, 0.03522364807128906, 0.035313568115234374, 0.03527065658569336, 0.03509968185424805, 0.03512931060791016, 0.03513753509521484, 0.0351016960144043, 0.035138561248779294, 0.03509145736694336, 0.03508633422851563, 0.03506175994873047, 0.03514572906494141, 0.035197952270507815, 0.03511296081542969, 0.035166206359863283, 0.035122177124023435, 0.0351723518371582, 0.03512934494018555, 0.03513446426391602, 0.03510067367553711, 0.035195903778076174, 0.03519385528564453, 0.03521023941040039, 0.03526348876953125, 0.035323936462402346, 0.03511088180541992, 0.035181568145751956, 0.03508531188964844, 0.03519180679321289, 0.12617318725585938, 0.0349378547668457, 0.035079166412353514, 0.03498700714111328, 0.03501772689819336, 0.03491027069091797, 0.035101631164550784, 0.035138561248779294, 0.035040256500244144, 0.034871295928955076, 0.03506995010375977, 0.03498092651367188, 0.035047359466552734, 0.03500032043457031, 0.035026943206787106, 0.03508531188964844, 0.035053569793701174, 0.03485184097290039, 0.03495935821533203, 0.03502182388305664, 0.03527167892456055, 0.03511500930786133, 0.03508633422851563, 0.03509862518310547, 0.0350382080078125, 0.034917377471923826, 0.035062782287597655, 0.0351723518371582, 0.03571814346313477, 0.03512831878662109, 0.035160064697265625, 0.035110912322998046, 0.034852897644042966, 0.034930656433105466, 0.03508736038208008, 0.035148799896240236, 0.035178497314453126, 0.035171329498291014, 0.03512115097045899, 0.035113983154296875, 0.03506687927246094, 0.03499212646484375, 0.03513241577148438, 0.03503615951538086, 0.035076095581054685, 0.034988033294677735, 0.0351201286315918, 0.035138561248779294, 0.0351836166381836, 0.03504844665527344, 0.03506790542602539, 0.03512627029418945, 0.03512319946289062, 0.035056640625, 0.035110912322998046, 0.0351539192199707, 0.03511603164672852, 0.0350750732421875, 0.03517542266845703, 0.0351016960144043, 0.03518054580688477, 0.035093505859375, 0.03516211318969727, 0.12640870666503906, 0.03479040145874023, 0.03474534225463867, 0.03468492889404297, 0.034678783416748044, 0.03475558471679688, 0.034835456848144535, 0.03487539291381836, 0.03471462249755859, 0.03468601608276367, 0.034755519866943356, 0.034654239654541015, 0.034728927612304686, 0.034715648651123046, 0.03477196884155274, 0.03472281646728516, 0.03476377487182617, 0.034738174438476564, 0.03479142379760742, 0.03471257781982422, 0.034767871856689454, 0.03470848083496094, 0.03476070404052734, 0.034732032775878906, 0.034702335357666016, 0.03470336151123047, 0.034902015686035154, 0.03489996719360351, 0.034746368408203124, 0.03481190490722656, 0.03477913665771484, 0.03483135986328125, 0.03479142379760742, 0.034801727294921876, 0.034834369659423825, 0.03482726287841797, 0.03479244613647461, 0.03474943923950195, 0.03474943923950195, 0.03482112121582031, 0.03481292724609375, 0.03482726287841797, 0.03480473709106445, 0.034723838806152346, 0.03484569549560547, 0.03482624053955078, 0.03486412811279297, 0.0348487663269043, 0.03485696029663086, 0.03477401733398437, 0.0348590087890625, 0.03476070404052734, 0.034816001892089846, 0.03485184097290039, 0.03486207962036133, 0.03488467025756836, 0.034990047454833986, 0.03485795211791992, 0.03491020965576172, 0.034831390380859376, 0.03486921691894531, 0.034841598510742186, 0.035127296447753906, 0.12554137420654296, 0.03465625762939453, 0.03468492889404297, 0.03463065719604492, 0.0346879997253418, 0.03465830230712891, 0.03465727996826172, 0.034764801025390625, 0.034718719482421875, 0.0346429443359375, 0.03470438385009766, 0.03467270278930664, 0.03479955291748047, 0.03465830230712891, 0.034729984283447264, 0.034802688598632815, 0.03467673492431641, 0.03465830230712891, 0.034753536224365236, 0.03491839981079101, 0.034797569274902344, 0.03468288040161133, 0.034767871856689454, 0.034812992095947265, 0.03476883316040039, 0.03477612686157226, 0.035131328582763674, 0.03489177703857422, 0.03481705474853516, 0.034819038391113284, 0.03482729721069336, 0.0349071044921875, 0.034783233642578126, 0.034710529327392575, 0.03479142379760742, 0.03476172637939453, 0.03481087875366211, 0.03478220748901367, 0.03484364700317383, 0.03486310577392578, 0.03484467315673828, 0.03479654312133789, 0.035165184020996096, 0.03510784149169922, 0.03555430221557617, 0.035160064697265625, 0.034917377471923826, 0.03521023941040039, 0.03502191925048828, 0.034874271392822266, 0.03485696029663086, 0.03482009506225586, 0.03487334442138672, 0.034885631561279294, 0.0349716796875, 0.035009502410888675, 0.03505766296386719, 0.0348487663269043, 0.03487539291381836, 0.034825214385986326, 0.034931713104248044, 0.03481804656982422, 0.034852863311767575, 0.12547481536865235, 0.03471462249755859, 0.03468492889404297, 0.034648063659667966, 0.034735103607177735, 0.03467264175415039, 0.034767871856689454, 0.0350013427734375, 0.03484672164916992, 0.03473612976074219, 0.03473920059204102, 0.03488256072998047, 0.03484979248046875, 0.03474534225463867, 0.034825214385986326, 0.03477913665771484, 0.03472793579101562, 0.03482828903198242, 0.034985984802246094, 0.03472895812988281, 0.03475558471679688, 0.034700286865234374, 0.03489689636230469, 0.03499318313598633, 0.03486102294921875, 0.03475046539306641, 0.034786304473876956, 0.03472281646728516, 0.03479142379760742, 0.03477196884155274, 0.03481292724609375, 0.03480473709106445, 0.03489177703857422, 0.03481497573852539, 0.03482316970825195, 0.03474534225463867, 0.03493478393554687, 0.03477407836914063, 0.03486611175537109, 0.03500032043457031, 0.03512627029418945, 0.035095550537109374, 0.035043327331542966, 0.035023872375488284, 0.035068958282470704, 0.03504431915283203, 0.03506073760986328, 0.035148799896240236, 0.0350750732421875, 0.035071041107177736, 0.03517331314086914, 0.03506892776489258, 0.03521535873413086, 0.03503615951538086, 0.034939903259277344, 0.03499724960327148, 0.03509664154052734, 0.034831295013427736, 0.035296257019042966, 0.03536076736450195, 0.035304447174072266, 0.035092479705810545, 0.03486617660522461, 0.12567040252685546, 0.03473100662231445, 0.03498086547851562, 0.03486105728149414, 0.03467468643188477, 0.03471155166625976, 0.03473100662231445, 0.03474227142333984, 0.034700286865234374, 0.03468185424804687, 0.03470336151123047, 0.03465523147583008, 0.03474534225463867, 0.034783233642578126, 0.03473100662231445, 0.03484979248046875, 0.03470848083496094, 0.03466854476928711, 0.03470131301879883, 0.03468185424804687, 0.0347586555480957, 0.03467161560058594, 0.034756607055664065, 0.034799617767333986, 0.0348109130859375, 0.034726879119873044, 0.0347955207824707, 0.034723838806152346, 0.03481702423095703, 0.03479040145874023, 0.03481497573852539, 0.03491328048706055, 0.03482726287841797, 0.03472588729858399, 0.03487744140625, 0.03488051223754883, 0.03499212646484375, 0.034769920349121096, 0.03505254364013672, 0.035093505859375, 0.034988033294677735, 0.03479449462890625, 0.03488051223754883, 0.034783233642578126, 0.035079166412353514, 0.03507814407348633, 0.03502284622192383, 0.034885631561279294, 0.03491635131835937, 0.03481087875366211, 0.03486207962036133, 0.0347852783203125, 0.03503513717651367, 0.03540383911132813, 0.035403713226318356, 0.03496550369262695, 0.034907135009765625, 0.03499929428100586, 0.03503308868408203, 0.03493686294555664, 0.03493475341796875, 0.03485184097290039, 0.034941951751708986, 0.12551273345947267, 0.03473507308959961, 0.034715648651123046, 0.034612224578857424, 0.0346798095703125, 0.034661376953125, 0.03468185424804687, 0.03479347229003906, 0.034713600158691404, 0.03471772766113281, 0.034717758178710936, 0.03465001678466797, 0.03472895812988281, 0.03467779159545899, 0.03470230484008789, 0.03476377487182617, 0.03481087875366211, 0.034699264526367186, 0.03474431991577148, 0.03471974563598633, 0.03479040145874023, 0.03469311904907227, 0.03477913665771484, 0.03542937469482422, 0.03522662353515625, 0.03483443069458008, 0.03482726287841797, 0.03500646209716797, 0.03478937530517578, 0.034909183502197266, 0.03505152130126953, 0.035033153533935546, 0.035034046173095704, 0.03484979248046875, 0.03483443069458008, 0.034800640106201174, 0.03502592086791992, 0.0347883529663086, 0.03484262466430664, 0.03506073760986328, 0.03504230499267578, 0.035076095581054685, 0.03481708908081055, 0.034758590698242185, 0.035225601196289064, 0.03481190490722656, 0.0349532470703125, 0.03495113754272461, 0.03489177703857422, 0.03486207962036133, 0.034945022583007815, 0.03486105728149414, 0.035056640625, 0.03486207962036133, 0.03486515045166016, 0.034887680053710936, 0.03490816116333008, 0.03487334442138672, 0.034933761596679686, 0.03485084915161133, 0.03488355255126953, 0.034852863311767575, 0.03487948989868164, 0.1255731201171875, 0.03466243362426758, 0.034718753814697266, 0.03467257690429688, 0.0346951675415039, 0.03467366409301758, 0.034677761077880856, 0.03476172637939453, 0.03475251388549805, 0.034625537872314455, 0.034683902740478514, 0.03469107055664063, 0.034726913452148435, 0.03467366409301758, 0.0347770881652832, 0.03484467315673828, 0.03479142379760742, 0.034813953399658204, 0.03479449462890625, 0.03477401733398437, 0.03475251388549805, 0.0351016960144043, 0.03500236892700195, 0.03482419204711914, 0.034718719482421875, 0.03472793579101562, 0.03482316970825195, 0.03473408126831055, 0.03477811050415039, 0.03469209671020508, 0.0348221435546875, 0.03484467315673828, 0.034772991180419925, 0.03480575942993164, 0.03481804656982422, 0.03478940963745117, 0.03485897445678711, 0.03477503967285156, 0.03490508651733398, 0.03496857452392578, 0.034887680053710936, 0.034799678802490235, 0.03488147354125976, 0.03484569549560547, 0.03491635131835937, 0.03486822509765625, 0.0349306869506836, 0.03492659378051758, 0.034895870208740236, 0.03488358306884766, 0.0349409294128418, 0.03492147064208984, 0.034890750885009765, 0.03487539291381836, 0.03492761611938477, 0.03498086547851562, 0.03492454528808594, 0.034909183502197266, 0.03498700714111328, 0.034857982635498046, 0.034923519134521484, 0.034841598510742186, 0.034928638458251955, 0.1257523193359375, 0.035064830780029296, 0.03537510299682617, 0.035378177642822264, 0.03510784149169922, 0.03473408126831055, 0.034738174438476564, 0.03501567840576172, 0.03498188781738281, 0.03481702423095703, 0.03481190490722656, 0.034732032775878906, 0.03487539291381836, 0.03477503967285156, 0.03500236892700195, 0.034977790832519534, 0.034956287384033204, 0.03479654312133789, 0.03480889511108398, 0.03498284912109375, 0.03504127883911133, 0.03495935821533203, 0.034874366760253905, 0.03476889419555664, 0.03484672164916992, 0.03483238220214844, 0.03480780792236328, 0.03479347229003906, 0.034842655181884764, 0.03487535858154297, 0.03479449462890625, 0.03484467315673828, 0.03486310577392578, 0.034890750885009765, 0.034974720001220705, 0.03501260757446289, 0.034955265045166016, 0.0349194221496582, 0.03486105728149414, 0.034885631561279294, 0.034950145721435545, 0.03498092651367188, 0.034893760681152346, 0.03479244613647461, 0.034854911804199216, 0.03503513717651367, 0.0349306869506836, 0.034939903259277344, 0.03488256072998047, 0.03484467315673828, 0.03496550369262695, 0.034854911804199216, 0.0349306869506836, 0.0348671989440918, 0.034931713104248044, 0.034953216552734374, 0.034941951751708986, 0.034923519134521484, 0.035007488250732424, 0.034948097229003904, 0.03496044921875, 0.034936767578125, 0.034907135009765625]",tokens/s,27.605657220121714,,,main,False,False -float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2900,7 +2900,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1270.910976,4299.685888,0.0,3653.238784,3408.132096,s,10,0.2341802215576172,0.02341802215576172,0.0002974755822864154,0.023323872566223147,0.02352355136871338,0.0239109112739563,0.024220799198150633,"[0.024298271179199218, 0.023331424713134766, 0.023350719451904298, 0.023249759674072265, 0.02326576042175293, 0.02343747138977051, 0.023316320419311524, 0.023334911346435547, 0.023300479888916016, 0.02329510307312012]",tokens/s,10931.751550034907,kWh,2.777987985201782e-07,1.5216764054777846e-07,1.3773817748309843e-06,1.8073482138989412e-06,tokens/kWh,141643983.17451978,MB,1270.910976,4299.685888,0.0,3653.238784,3474.14784,s,10,12.8284580078125,1.28284580078125,0.02374972515115311,1.28130615234375,1.3215483520507811,1.3223690856933594,1.3230256726074219,"[1.2873050537109374, 1.2818013916015625, 1.2574949951171874, 1.2532855224609376, 1.2692486572265624, 1.3231898193359375, 1.2808109130859375, 1.2578243408203125, 1.29613134765625, 1.321365966796875]",tokens/s,49.109565593645904,kWh,1.4904885215720724e-05,8.16770102005328e-06,3.9106511723368935e-05,6.217909795914294e-05,tokens/kWh,1013202.2185557671,,s,629,13.015847944259642,0.02069292200995174,0.002854091788221458,0.020067327499389647,0.021038489532470703,0.021207449340820313,0.04363136993408203,"[0.021745664596557617, 0.020215808868408205, 0.020060159683227538, 0.02001408004760742, 0.019984384536743165, 0.01991372871398926, 0.019983360290527344, 0.019928064346313477, 0.01987379264831543, 0.019949567794799804, 0.01998028755187988, 0.019973119735717772, 0.01993011283874512, 0.019812351226806642, 0.020019264221191407, 0.019854272842407227, 0.019893280029296877, 0.01991983985900879, 0.019961856842041017, 0.019776512145996093, 0.019936256408691407, 0.019756032943725587, 0.019734527587890623, 0.019952640533447266, 0.01979804801940918, 0.019649503707885742, 0.019636224746704102, 0.021346303939819337, 0.020783103942871094, 0.020749311447143554, 0.02089472007751465, 0.020787200927734374, 0.02067865562438965, 0.020617216110229493, 0.020555776596069338, 0.02063052749633789, 0.020785152435302736, 0.020738048553466795, 0.020744192123413087, 0.020651103973388672, 0.02063043212890625, 0.020768768310546876, 0.021725183486938478, 0.021118976593017577, 0.020783103942871094, 0.020553728103637696, 0.020529312133789064, 0.02051875114440918, 0.020531200408935548, 0.02065407943725586, 0.020583423614501953, 0.02067865562438965, 0.020592639923095703, 0.020633600234985353, 0.020558847427368163, 0.02064691162109375, 0.020653055191040038, 0.02059775924682617, 0.020684799194335936, 0.02063564872741699, 0.020549631118774413, 0.02067353630065918, 0.044382209777832034, 0.020682752609252928, 0.02066022491455078, 0.020503551483154296, 0.02061516761779785, 0.02062950325012207, 0.019954687118530275, 0.019971103668212892, 0.01992188835144043, 0.019942399978637695, 0.0198922233581543, 0.019920896530151368, 0.019803136825561524, 0.019735551834106444, 0.019708927154541016, 0.0198973445892334, 0.019952640533447266, 0.02002739143371582, 0.01987276840209961, 0.01987583923339844, 0.019942399978637695, 0.02000588798522949, 0.01986662483215332, 0.01998028755187988, 0.019954687118530275, 0.020093952178955078, 0.020123647689819335, 0.01988812828063965, 0.020187135696411132, 0.0206561279296875, 0.02066739273071289, 0.01986662483215332, 0.019742719650268553, 0.01983897590637207, 0.02004991912841797, 0.019983360290527344, 0.0198287353515625, 0.019923967361450197, 0.019941375732421874, 0.019999744415283204, 0.01982361602783203, 0.020876287460327148, 0.021183488845825195, 0.020563968658447264, 0.020563968658447264, 0.020624383926391602, 0.020797439575195312, 0.020732927322387695, 0.020773887634277344, 0.020715520858764647, 0.020717567443847656, 0.02079641532897949, 0.02068070411682129, 0.020709375381469726, 0.02068070411682129, 0.020802560806274413, 0.020727807998657227, 0.02145280075073242, 0.021909503936767577, 0.021165056228637694, 0.020758527755737305, 0.020809728622436522, 0.020805631637573242, 0.043633663177490234, 0.01987379264831543, 0.019883007049560548, 0.020031488418579102, 0.019974208831787108, 0.02038470458984375, 0.020025344848632814, 0.020137983322143553, 0.020101119995117187, 0.02005504035949707, 0.02002022361755371, 0.020048896789550782, 0.020025407791137696, 0.020079551696777345, 0.02007040023803711, 0.020115455627441405, 0.02000588798522949, 0.020074495315551756, 0.020127840042114258, 0.0200467529296875, 0.02003660774230957, 0.02002739143371582, 0.019998720169067383, 0.020212736129760742, 0.020157440185546875, 0.019935232162475586, 0.019979263305664064, 0.02002124786376953, 0.020039680480957032, 0.019886079788208007, 0.019989503860473632, 0.019861503601074217, 0.01979903984069824, 0.02000694465637207, 0.019782623291015624, 0.019973119735717772, 0.01985228729248047, 0.019896320343017578, 0.019900415420532228, 0.01983795166015625, 0.02001817512512207, 0.01984716796875, 0.019940351486206053, 0.019818496704101563, 0.01985638427734375, 0.01985443115234375, 0.019885984420776368, 0.01987993621826172, 0.019988576889038087, 0.01981328010559082, 0.01985228729248047, 0.019987455368041994, 0.01984614372253418, 0.019924991607666014, 0.019721216201782226, 0.020283456802368163, 0.019814336776733398, 0.019855392456054687, 0.019803104400634767, 0.01969049644470215, 0.019789920806884766, 0.019769247055053712, 0.019746816635131836, 0.04333158493041992, 0.01997209548950195, 0.02001817512512207, 0.0196945915222168, 0.01969254493713379, 0.01988198471069336, 0.01986457633972168, 0.019869695663452147, 0.019653728485107422, 0.019724191665649413, 0.019720191955566405, 0.019738624572753907, 0.01969049644470215, 0.019704832077026366, 0.019787776947021486, 0.01990656089782715, 0.01964236831665039, 0.019680416107177735, 0.01964531135559082, 0.019972063064575194, 0.019923967361450197, 0.019942399978637695, 0.019932159423828123, 0.02003455924987793, 0.019851264953613282, 0.019920896530151368, 0.019927040100097656, 0.020001792907714845, 0.01997209548950195, 0.020033536911010744, 0.019953664779663087, 0.019792896270751953, 0.019681280136108398, 0.019945472717285157, 0.01986867141723633, 0.019681280136108398, 0.019722240447998047, 0.019770368576049805, 0.019920896530151368, 0.02003455924987793, 0.019948543548583983, 0.019940351486206053, 0.02044927978515625, 0.019939327239990236, 0.019949567794799804, 0.019940383911132814, 0.0199270076751709, 0.02003660774230957, 0.019962879180908204, 0.019979263305664064, 0.01985843276977539, 0.019983360290527344, 0.019940351486206053, 0.019904512405395508, 0.019933183670043944, 0.020065280914306642, 0.01984000015258789, 0.019982336044311523, 0.019997695922851562, 0.019952640533447266, 0.019926015853881835, 0.020039680480957032, 0.019946495056152345, 0.04366233444213867, 0.020079679489135742, 0.020071359634399415, 0.019999744415283204, 0.020051008224487306, 0.020023231506347657, 0.02002943992614746, 0.02003763198852539, 0.019997695922851562, 0.020009983062744142, 0.02006630325317383, 0.02007040023803711, 0.01986867141723633, 0.019826688766479493, 0.019978239059448243, 0.02006630325317383, 0.019811328887939454, 0.01984614372253418, 0.019983360290527344, 0.019974143981933593, 0.020033567428588868, 0.020037599563598633, 0.02003660774230957, 0.02002943992614746, 0.0200898551940918, 0.02004787254333496, 0.020033536911010744, 0.01999667167663574, 0.020282367706298828, 0.019911680221557617, 0.02000387191772461, 0.020131807327270507, 0.02005299186706543, 0.020023359298706054, 0.020051904678344726, 0.01998236846923828, 0.020031455993652345, 0.02003046417236328, 0.02003660774230957, 0.020063232421875, 0.020135936737060548, 0.020084735870361328, 0.019990528106689453, 0.019984384536743165, 0.020059135437011717, 0.020060159683227538, 0.020011007308959963, 0.02012876892089844, 0.02001203155517578, 0.019947519302368166, 0.02005299186706543, 0.019979263305664064, 0.02009702491760254, 0.019969024658203126, 0.02004275131225586, 0.020065280914306642, 0.020376575469970702, 0.02172211265563965, 0.021393407821655275, 0.02124595260620117, 0.020942848205566408, 0.020891647338867187, 0.020988927841186524, 0.04510105514526367, 0.020916223526000977, 0.020926528930664063, 0.020947904586791993, 0.020888576507568358, 0.020970495223999023, 0.021120000839233398, 0.020977664947509765, 0.021032960891723632, 0.020953088760375976, 0.020884479522705078, 0.020926464080810548, 0.021393407821655275, 0.020914176940917968, 0.020915199279785156, 0.02107904052734375, 0.020968479156494142, 0.02096227264404297, 0.020883455276489257, 0.020953088760375976, 0.020938751220703124, 0.020899839401245117, 0.02090291213989258, 0.021133312225341795, 0.021147680282592774, 0.02110870361328125, 0.020958208084106447, 0.020939775466918945, 0.020932607650756836, 0.020970495223999023, 0.021127168655395507, 0.02112512016296387, 0.020987903594970703, 0.020958208084106447, 0.021003263473510742, 0.020988927841186524, 0.020931583404541015, 0.0206059513092041, 0.020609024047851563, 0.020945920944213867, 0.02104729652404785, 0.021313535690307618, 0.021115903854370118, 0.02106368064880371, 0.02098080062866211, 0.02092844772338867, 0.02101043128967285, 0.02104422378540039, 0.0210565128326416, 0.020954111099243163, 0.020989952087402345, 0.021073919296264648, 0.02084249687194824, 0.020896768569946288, 0.02110771179199219, 0.021196800231933592, 0.020963327407836914, 0.02091423988342285, 0.020849599838256835, 0.020975616455078124, 0.021371936798095702, 0.020995040893554688, 0.02101862335205078, 0.0438548469543457, 0.020142080307006836, 0.02010419273376465, 0.01984409523010254, 0.01984409523010254, 0.019984384536743165, 0.020109312057495117, 0.02012876892089844, 0.020737024307250978, 0.021004287719726563, 0.021003328323364257, 0.020971456527709962, 0.020948991775512696, 0.020928512573242186, 0.020600831985473633, 0.02068070411682129, 0.020880384445190428, 0.020953088760375976, 0.02088755226135254, 0.02005504035949707, 0.019975168228149414, 0.020067327499389647, 0.020024320602416993, 0.020115455627441405, 0.020032512664794923, 0.02000486373901367, 0.02004787254333496, 0.019998720169067383, 0.020039680480957032, 0.020050943374633787, 0.01987481689453125, 0.020271104812622072, 0.02029465675354004, 0.02002841567993164, 0.019802112579345704, 0.019793920516967774, 0.020607999801635742, 0.02085478401184082, 0.020904960632324218, 0.020916223526000977, 0.02006227111816406, 0.01998329544067383, 0.02008678436279297, 0.020065280914306642, 0.019986431121826173, 0.0200130558013916, 0.02045337677001953, 0.02003558349609375, 0.019869695663452147, 0.019973119735717772, 0.019720191955566405, 0.019794944763183595, 0.019565568923950196, 0.019907583236694337, 0.01999667167663574, 0.02007756805419922, 0.02106982421875, 0.02084249687194824, 0.02110771179199219, 0.02090598487854004, 0.02102681541442871, 0.021204992294311522, 0.02104422378540039, 0.043676673889160154, 0.02001919937133789, 0.02003868865966797, 0.019924959182739257, 0.019914751052856446, 0.01990656089782715, 0.019909664154052733, 0.020010976791381835, 0.01984614372253418, 0.019778560638427735, 0.019825664520263672, 0.019948543548583983, 0.019908607482910155, 0.01999564743041992, 0.019959808349609375, 0.020083711624145507, 0.020039680480957032, 0.019927040100097656, 0.019912704467773438, 0.019953664779663087, 0.019912704467773438, 0.019956735610961913, 0.019983360290527344, 0.019971071243286134, 0.01987993621826172, 0.02001817512512207, 0.019924991607666014, 0.020002815246582033, 0.019941375732421874, 0.019976192474365235, 0.01986764717102051, 0.020370431900024414, 0.019978239059448243, 0.019935232162475586, 0.019893247604370116, 0.019936256408691407, 0.019908607482910155, 0.019953664779663087, 0.019964927673339843, 0.02002124786376953, 0.019948543548583983, 0.019935232162475586, 0.01982464027404785, 0.019915775299072267, 0.019951616287231445, 0.019944448471069336, 0.019893247604370116, 0.019989503860473632, 0.019998720169067383, 0.020221952438354493, 0.02003455924987793, 0.019917823791503905, 0.01988812828063965, 0.019946495056152345, 0.019891199111938478, 0.020016128540039063, 0.020343807220458983, 0.019942399978637695, 0.019942399978637695, 0.019912704467773438, 0.019932159423828123, 0.019947519302368166, 0.019910655975341796, 0.04362547302246094, 0.02008166313171387, 0.020084735870361328, 0.020109312057495117, 0.02009600067138672, 0.02003046417236328, 0.019975168228149414, 0.02001817512512207, 0.020333568572998048, 0.02005504035949707, 0.019964927673339843, 0.019992576599121094, 0.019933183670043944, 0.019965951919555663, 0.019959808349609375, 0.02007244873046875, 0.020385791778564453, 0.02003046417236328, 0.019973184585571287, 0.01994643211364746, 0.01999564743041992, 0.02004684829711914, 0.019998720169067383, 0.021311487197875977, 0.02208768081665039, 0.021323776245117186, 0.02102783966064453, 0.021223424911499023, 0.021209087371826172, 0.02150297546386719, 0.020987903594970703, 0.02087116813659668, 0.020920320510864256, 0.020726783752441406, 0.02070425605773926, 0.020586496353149415, 0.020773887634277344, 0.02084556770324707, 0.020586496353149415, 0.020749311447143554, 0.020813823699951172, 0.020794368743896483, 0.020611072540283205, 0.020718591690063477, 0.02083020782470703, 0.020853887557983397, 0.02089561653137207, 0.020938751220703124, 0.020946943283081054, 0.02065100860595703, 0.020713632583618163, 0.020960159301757812, 0.02082195281982422, 0.020625408172607423, 0.02061414337158203, 0.020739072799682616, 0.020855808258056642, 0.020703231811523438, 0.02050048065185547, 0.020596736907958983, 0.020686847686767578, 0.02079641532897949, 0.020648128509521486, 0.0447393913269043, 0.02084659194946289, 0.02069708824157715, 0.020653055191040038, 0.02060492706298828, 0.020818944931030273, 0.02071244812011719, 0.020807680130004884, 0.02086809539794922, 0.020943872451782225, 0.020899839401245117, 0.021009408950805664, 0.020891647338867187, 0.02115897560119629, 0.020947904586791993, 0.020979711532592774, 0.020956159591674805, 0.02090598487854004, 0.02084659194946289, 0.020831232070922853, 0.02108723258972168, 0.022380544662475587, 0.02164838409423828, 0.021226495742797852, 0.021082111358642578, 0.021053440093994142, 0.02100223922729492, 0.020875263214111327, 0.020892671585083008, 0.020954111099243163, 0.021037055969238282, 0.020916223526000977, 0.020916223526000977, 0.02081996726989746, 0.02085990333557129, 0.021155839920043946, 0.021579776763916016, 0.021963775634765623, 0.021143552780151367, 0.021143552780151367, 0.021067808151245117, 0.02090595245361328, 0.020690944671630858, 0.020789247512817383, 0.020644863128662108, 0.020374528884887694, 0.020557823181152343, 0.02068889617919922, 0.02086297607421875, 0.02123673629760742, 0.02090598487854004, 0.02091007995605469, 0.02091007995605469, 0.021028863906860353, 0.02088652801513672, 0.020967424392700194, 0.020964351654052735, 0.021045248031616212, 0.020952064514160155, 0.020765695571899414, 0.020781055450439453, 0.020942848205566408, 0.020982784271240236]",tokens/s,48.32570284269545,,,,, -bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2937,7 +2937,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 768.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3284,7 +3284,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3311,7 +3311,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3467,7 +3467,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 216.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3510,7 +3510,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GP ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1286.262784,3848.798208,0.0,3202.351104,3105.82784,s,10,0.24081798553466796,0.024081798553466795,0.00038659848964031586,0.02391201686859131,0.024304146957397462,0.024740841484069825,0.025090197105407715,"[0.02517753601074219, 0.024207103729248048, 0.023898752212524414, 0.023821407318115235, 0.024194112777709963, 0.023959680557250975, 0.02386591911315918, 0.023917184829711915, 0.023906848907470704, 0.02386944007873535]",tokens/s,10630.43524060816,kWh,2.84827513946906e-07,1.5607165578747143e-07,1.2483898875999952e-06,1.6892890573343727e-06,tokens/kWh,151543040.48115796,MB,1286.262784,3848.798208,0.0,3202.351104,3188.152832,s,10,12.569388427734376,1.2569388427734376,0.014872383514358697,1.2568008422851562,1.2705527954101563,1.2805973205566406,1.2886329406738282,"[1.290641845703125, 1.2683206787109376, 1.259688232421875, 1.2572567138671875, 1.2520152587890625, 1.2313369140625, 1.26155517578125, 1.2448743896484376, 1.247354248046875, 1.256344970703125]",tokens/s,50.12177033290689,kWh,1.4532639801382539e-05,7.963009467016875e-06,3.82537250474e-05,6.074937431579941e-05,tokens/kWh,1037047.7179320555,,s,629,12.76368588638305,0.020292028436221077,0.002905302237704868,0.020099071502685546,0.020469760513305665,0.020771033477783203,0.04389322860717773,"[0.021151807785034178, 0.019891136169433592, 0.020847616195678712, 0.020914176940917968, 0.020915199279785156, 0.02111692810058594, 0.02111692810058594, 0.020792320251464845, 0.020775936126708985, 0.02079641532897949, 0.02068172836303711, 0.020815872192382814, 0.020719615936279297, 0.020724735260009765, 0.02085171127319336, 0.020757503509521484, 0.02087731170654297, 0.02070528030395508, 0.020705312728881837, 0.020793312072753905, 0.020719615936279297, 0.020801536560058592, 0.020790271759033203, 0.02076367950439453, 0.02073904037475586, 0.02065100860595703, 0.02084147262573242, 0.02065203285217285, 0.020731903076171874, 0.020502559661865233, 0.020415519714355467, 0.02015225601196289, 0.020214784622192384, 0.02018611145019531, 0.02021785545349121, 0.02026393508911133, 0.020148223876953125, 0.02012876892089844, 0.02026393508911133, 0.02018611145019531, 0.020222976684570314, 0.019801088333129883, 0.019523584365844726, 0.019513343811035155, 0.019523584365844726, 0.01947238349914551, 0.020706304550170897, 0.02047385597229004, 0.02043801689147949, 0.020192256927490236, 0.020212736129760742, 0.020229120254516602, 0.020196352005004883, 0.020288511276245116, 0.02051584053039551, 0.020315135955810547, 0.02010419273376465, 0.02039193534851074, 0.02030182456970215, 0.020184064865112306, 0.020719680786132812, 0.020320192337036132, 0.04494438552856445, 0.02020351982116699, 0.02027827262878418, 0.020206592559814454, 0.02023423957824707, 0.020159488677978517, 0.020176895141601564, 0.02027008056640625, 0.02019327926635742, 0.020350976943969725, 0.020204608917236327, 0.020264896392822265, 0.020212736129760742, 0.020196352005004883, 0.020686847686767578, 0.020505599975585938, 0.020063232421875, 0.020143104553222657, 0.020145151138305666, 0.0194652156829834, 0.01945395278930664, 0.019491840362548828, 0.019962879180908204, 0.021910528182983398, 0.020692991256713866, 0.020377599716186523, 0.020299776077270508, 0.02028441619873047, 0.020185087203979494, 0.02021171188354492, 0.02027827262878418, 0.02027827262878418, 0.02022604751586914, 0.02021785545349121, 0.020237312316894532, 0.02026905632019043, 0.020231168746948244, 0.020274175643920898, 0.020180992126464844, 0.02027212715148926, 0.020329471588134765, 0.020324352264404297, 0.020385791778564453, 0.020288511276245116, 0.02024345588684082, 0.020207679748535157, 0.020246463775634764, 0.02050662422180176, 0.020231168746948244, 0.020379648208618165, 0.02028339195251465, 0.020220928192138672, 0.019507200241088866, 0.019505151748657225, 0.01946316719055176, 0.019615743637084963, 0.019482624053955077, 0.019563520431518554, 0.019511295318603517, 0.019519487380981446, 0.019551231384277345, 0.01946828842163086, 0.019519487380981446, 0.04389580917358398, 0.019528703689575197, 0.01942835235595703, 0.019365888595581054, 0.019504127502441407, 0.019555328369140625, 0.019558399200439454, 0.019569664001464843, 0.019843072891235353, 0.02023526382446289, 0.02001203155517578, 0.020419584274291993, 0.020810752868652343, 0.01988198471069336, 0.019573759078979493, 0.01967411231994629, 0.019490816116333007, 0.0196177921295166, 0.01961676788330078, 0.01965977668762207, 0.019656736373901366, 0.019595232009887695, 0.019581951141357423, 0.01962393569946289, 0.01944268798828125, 0.019603456497192383, 0.019483648300170898, 0.01983795166015625, 0.0195020809173584, 0.019555328369140625, 0.01968332862854004, 0.02049843215942383, 0.020368383407592772, 0.020337663650512695, 0.020238336563110353, 0.02022400093078613, 0.020126720428466797, 0.020195392608642577, 0.02016249656677246, 0.02007040023803711, 0.02028339195251465, 0.020200447082519533, 0.02022400093078613, 0.02020351982116699, 0.02026905632019043, 0.020372512817382813, 0.020278240203857423, 0.020368383407592772, 0.020162559509277343, 0.0204769287109375, 0.021015552520751952, 0.020327423095703127, 0.020190208435058594, 0.020032512664794923, 0.019955711364746095, 0.020406272888183592, 0.02030899238586426, 0.020246528625488282, 0.02024550437927246, 0.020190208435058594, 0.020315135955810547, 0.02025881576538086, 0.02020966339111328, 0.04391526412963867, 0.019556352615356445, 0.0194703369140625, 0.01962291145324707, 0.01949388885498047, 0.01964339256286621, 0.02002739143371582, 0.01942323112487793, 0.019357696533203125, 0.019296255111694336, 0.019309568405151366, 0.019731456756591798, 0.020345855712890625, 0.020303871154785155, 0.019999744415283204, 0.02030899238586426, 0.02022809600830078, 0.02022809600830078, 0.02028646469116211, 0.020157440185546875, 0.020344831466674804, 0.020315135955810547, 0.020353023529052734, 0.020099071502685546, 0.019663871765136717, 0.019560447692871095, 0.019535871505737306, 0.019458047866821288, 0.019479551315307618, 0.019519487380981446, 0.019553279876708983, 0.019589120864868165, 0.019518463134765626, 0.019500032424926757, 0.01945088005065918, 0.019501056671142578, 0.019952640533447266, 0.020179967880249023, 0.020221952438354493, 0.02024550437927246, 0.020157440185546875, 0.02009600067138672, 0.02022604751586914, 0.020164608001708984, 0.020365312576293947, 0.020338687896728515, 0.020237312316894532, 0.02027622413635254, 0.02022604751586914, 0.020199424743652345, 0.020343807220458983, 0.020328447341918944, 0.020205568313598633, 0.020162559509277343, 0.020285472869873047, 0.02021993637084961, 0.020214719772338866, 0.020336639404296874, 0.020273151397705077, 0.020164608001708984, 0.02021171188354492, 0.01944473648071289, 0.019750911712646483, 0.044450817108154295, 0.019489791870117186, 0.019546112060546874, 0.01944166374206543, 0.019371007919311522, 0.019351551055908203, 0.019405824661254883, 0.019426368713378907, 0.019482559204101562, 0.01941196823120117, 0.019332096099853514, 0.019368959426879884, 0.01946214485168457, 0.0194969596862793, 0.019611648559570313, 0.019578943252563475, 0.01955526351928711, 0.02004377555847168, 0.02020966339111328, 0.02025062370300293, 0.020132863998413086, 0.020262912750244142, 0.020411392211914063, 0.020392959594726562, 0.020173824310302735, 0.020156415939331054, 0.020149248123168945, 0.020167680740356447, 0.020146175384521483, 0.02025881576538086, 0.020208639144897463, 0.02031001663208008, 0.020305919647216796, 0.020187135696411132, 0.02024038314819336, 0.020195327758789062, 0.020199424743652345, 0.02018611145019531, 0.02023423957824707, 0.020281343460083007, 0.020161535263061522, 0.020141056060791016, 0.020146175384521483, 0.02021785545349121, 0.02018611145019531, 0.02023526382446289, 0.0202106876373291, 0.020241472244262697, 0.020245439529418947, 0.019703807830810546, 0.019560447692871095, 0.019599359512329103, 0.019527679443359376, 0.019488800048828126, 0.019586015701293945, 0.019599359512329103, 0.019535871505737306, 0.01948057556152344, 0.01942118453979492, 0.019373056411743163, 0.019479551315307618, 0.019543039321899415, 0.01982259178161621, 0.044063743591308595, 0.019581951141357423, 0.019532800674438477, 0.019489791870117186, 0.019562496185302734, 0.019557376861572266, 0.019491840362548828, 0.019546112060546874, 0.01942937660217285, 0.019339263916015623, 0.019567615509033204, 0.019539968490600586, 0.019520511627197267, 0.019345407485961915, 0.01927884864807129, 0.01940275192260742, 0.019366912841796875, 0.019557376861572266, 0.019628032684326172, 0.019509248733520508, 0.019334144592285156, 0.01939148712158203, 0.019334144592285156, 0.019377151489257814, 0.019342336654663086, 0.019333120346069335, 0.019319807052612305, 0.019325952529907226, 0.019383359909057617, 0.019322816848754882, 0.019386367797851564, 0.01943244743347168, 0.019507200241088866, 0.019522560119628905, 0.019523584365844726, 0.019558399200439454, 0.019721216201782226, 0.01954921531677246, 0.019512319564819337, 0.019634143829345703, 0.019499008178710937, 0.019516416549682617, 0.019524608612060547, 0.019520511627197267, 0.019526655197143555, 0.019475456237792968, 0.019582975387573243, 0.019564544677734375, 0.019538944244384765, 0.019509248733520508, 0.019536895751953123, 0.019569664001464843, 0.019538944244384765, 0.019531839370727538, 0.01954502487182617, 0.019569664001464843, 0.0194969596862793, 0.01945088005065918, 0.01951955223083496, 0.02041644859313965, 0.02039708709716797, 0.020190176010131837, 0.020192256927490236, 0.04383846282958984, 0.019486719131469727, 0.019507200241088866, 0.01964134407043457, 0.01968332862854004, 0.019500032424926757, 0.019562496185302734, 0.01949286460876465, 0.01942323112487793, 0.01948467254638672, 0.019577856063842772, 0.019374080657958984, 0.019354623794555666, 0.019550207138061524, 0.021597183227539063, 0.02074118423461914, 0.020300735473632814, 0.020200447082519533, 0.02021990394592285, 0.02025369644165039, 0.02019327926635742, 0.020200447082519533, 0.020587520599365236, 0.020340736389160157, 0.02024860763549805, 0.02024239921569824, 0.020220928192138672, 0.02025984001159668, 0.020347904205322266, 0.02005299186706543, 0.019965951919555663, 0.020171775817871093, 0.02023526382446289, 0.020162559509277343, 0.02062848091125488, 0.020468736648559572, 0.021319679260253906, 0.020706304550170897, 0.020364288330078126, 0.02044108772277832, 0.020363264083862305, 0.020254720687866212, 0.020195327758789062, 0.020158464431762696, 0.02012876892089844, 0.019971071243286134, 0.019389440536499023, 0.019570688247680663, 0.019463199615478516, 0.019528671264648436, 0.019507200241088866, 0.019552255630493166, 0.019551231384277345, 0.019548160552978516, 0.019647487640380858, 0.019537919998168944, 0.019539968490600586, 0.020363264083862305, 0.02025881576538086, 0.02023526382446289, 0.02023321533203125, 0.020281343460083007, 0.020231168746948244, 0.044044288635253906, 0.019486719131469727, 0.019364864349365234, 0.019422208786010742, 0.019577856063842772, 0.019527679443359376, 0.019518463134765626, 0.01961881637573242, 0.019942399978637695, 0.01960960006713867, 0.019505151748657225, 0.019591167449951173, 0.019475456237792968, 0.019515392303466796, 0.019458047866821288, 0.019529727935791014, 0.019581951141357423, 0.019540992736816407, 0.019551231384277345, 0.01946009635925293, 0.01945599937438965, 0.019479551315307618, 0.01948876762390137, 0.019594240188598632, 0.019524608612060547, 0.019514368057250975, 0.019438592910766602, 0.019551231384277345, 0.019341312408447265, 0.019386367797851564, 0.019343360900878907, 0.019834880828857423, 0.019714048385620117, 0.019581951141357423, 0.019551231384277345, 0.019507200241088866, 0.019675167083740234, 0.019625951766967773, 0.01946828842163086, 0.019575807571411134, 0.01963827133178711, 0.019748863220214845, 0.01966592025756836, 0.01940787124633789, 0.0193832950592041, 0.0202926082611084, 0.020148223876953125, 0.020154367446899413, 0.020404224395751954, 0.020257791519165038, 0.02016972732543945, 0.020231168746948244, 0.02024448013305664, 0.02031827163696289, 0.020296640396118164, 0.02006425666809082, 0.02012473678588867, 0.020514751434326173, 0.02024140739440918, 0.020642847061157227, 0.020385759353637695, 0.02025881576538086, 0.020265983581542968, 0.04407910537719727, 0.01964134407043457, 0.019686431884765626, 0.019511264801025392, 0.019408895492553712, 0.019358720779418945, 0.019491840362548828, 0.019555328369140625, 0.019561471939086913, 0.01942732810974121, 0.019518463134765626, 0.019543039321899415, 0.019501056671142578, 0.019561471939086913, 0.01958502388000488, 0.019539968490600586, 0.01943142318725586, 0.019536895751953123, 0.019709951400756837, 0.019595264434814453, 0.019430400848388672, 0.019536928176879884, 0.01944265556335449, 0.019369983673095705, 0.01947750473022461, 0.019561471939086913, 0.019528703689575197, 0.019544095993041993, 0.019529695510864257, 0.019568639755249022, 0.019598335266113282, 0.019752960205078125, 0.01973151969909668, 0.019793920516967774, 0.02035500717163086, 0.019987455368041994, 0.020392959594726562, 0.021120000839233398, 0.020279296875, 0.020282367706298828, 0.02023526382446289, 0.020214784622192384, 0.02025267219543457, 0.02017791938781738, 0.020206592559814454, 0.020418560028076172, 0.02027519989013672, 0.020183040618896485, 0.020255775451660157, 0.02030281639099121, 0.020139007568359374, 0.020289535522460937, 0.02018611145019531, 0.019962879180908204, 0.020106239318847655, 0.020050943374633787, 0.019778560638427735, 0.019348480224609374, 0.019486719131469727, 0.01940377616882324, 0.019476512908935546, 0.01951024055480957, 0.019513343811035155, 0.043886592864990234, 0.019506175994873046, 0.019521535873413084, 0.019764223098754884, 0.019554304122924804, 0.019523584365844726, 0.0195020809173584, 0.019595264434814453, 0.019559423446655275, 0.01946316719055176, 0.019500032424926757, 0.01943552017211914, 0.019620864868164063, 0.02089369583129883, 0.02066534423828125, 0.0204400634765625, 0.020248575210571287, 0.02024038314819336, 0.02019327926635742, 0.020279296875, 0.020205568313598633, 0.020212736129760742, 0.02023526382446289, 0.020057088851928712, 0.02004275131225586, 0.019920896530151368, 0.020113407135009767, 0.020142080307006836, 0.02011955261230469, 0.02027724838256836, 0.02026188850402832, 0.020530176162719727, 0.02041548728942871, 0.020206592559814454, 0.02025881576538086, 0.020222976684570314, 0.020190208435058594, 0.020314111709594726, 0.020159488677978517, 0.02045747184753418, 0.020254720687866212, 0.02024140739440918, 0.020153343200683595, 0.020123647689819335, 0.020176895141601564, 0.02032537651062012, 0.02016972732543945, 0.02028441619873047, 0.019542015075683594, 0.019518463134765626, 0.019560447692871095, 0.0194969596862793, 0.01948876762390137, 0.019557376861572266, 0.019524608612060547, 0.019534847259521485, 0.019568639755249022, 0.019487743377685548, 0.0194969596862793, 0.019486719131469727, 0.019589120864868165, 0.019422208786010742, 0.019499008178710937]",tokens/s,49.28043557316377,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1279.287296,6797.39392,0.0,6150.946816,5989.292032,s,10,0.3887236557006836,0.03887236557006836,0.0014476602822727281,0.038411041259765626,0.03893620071411132,0.041074644851684564,0.04278540016174316,"[0.04321308898925781, 0.038460990905761716, 0.038429759979248045, 0.038342273712158204, 0.03834566497802734, 0.03831264114379883, 0.038414592742919924, 0.03835456085205078, 0.03840748977661133, 0.03844259262084961]",tokens/s,6585.655291251929,kWh,4.5476279133914886e-07,2.4918756812649705e-07,2.3326225925923065e-06,3.0365729520579523e-06,tokens/kWh,84305565.53120293,MB,1279.287296,6797.39392,0.0,6150.946816,6115.763712,s,10,16.95939392089844,1.6959393920898438,0.01658036035440633,1.700268310546875,1.7105757202148437,1.7186706115722656,1.7251465246582032,"[1.7267655029296876, 1.6696607666015626, 1.7003759765625, 1.69278466796875, 1.674107421875, 1.6788770751953126, 1.70877685546875, 1.7042218017578126, 1.7036632080078125, 1.70016064453125]",tokens/s,37.14755391250593,kWh,2.0089898790177118e-05,1.1009501097148761e-05,6.44621530654086e-05,9.556155295273448e-05,tokens/kWh,659260.9480840093,,s,629,17.276188758850115,0.027466118853497774,0.004633908549784592,0.02698851203918457,0.02728058891296387,0.027619328308105467,0.06496825531005859,"[0.027757568359375, 0.027805696487426756, 0.027840511322021484, 0.027918336868286132, 0.027936767578125, 0.027876352310180662, 0.027886592864990234, 0.028139551162719725, 0.02875491142272949, 0.027892736434936522, 0.026928127288818358, 0.0267509765625, 0.02698240089416504, 0.027613183975219727, 0.027894784927368164, 0.027662336349487306, 0.02750771141052246, 0.02796134376525879, 0.03131596755981445, 0.029711360931396483, 0.0285614070892334, 0.027473920822143554, 0.027185152053833008, 0.027123807907104492, 0.027047840118408203, 0.027011072158813477, 0.026999807357788085, 0.02692915153503418, 0.026806272506713868, 0.02655539131164551, 0.02688928031921387, 0.026999744415283203, 0.027049983978271484, 0.02696396827697754, 0.02695782470703125, 0.02696294403076172, 0.026995712280273438, 0.026854400634765626, 0.027240447998046875, 0.026995712280273438, 0.027012096405029298, 0.026842111587524413, 0.026944511413574217, 0.026937408447265623, 0.02707961654663086, 0.02693836784362793, 0.02699673652648926, 0.02690559959411621, 0.02698137664794922, 0.026934272766113283, 0.026894336700439454, 0.026945535659790038, 0.0269117431640625, 0.026960895538330077, 0.027157503128051756, 0.026959871292114256, 0.02691276741027832, 0.027456512451171877, 0.027404287338256835, 0.028463104248046874, 0.02730291175842285, 0.027026432037353516, 0.06489600372314454, 0.025996288299560546, 0.02595020866394043, 0.025981952667236328, 0.02590924835205078, 0.02591436767578125, 0.025839616775512695, 0.02598297691345215, 0.0259368953704834, 0.025993215560913087, 0.02614476776123047, 0.025964544296264647, 0.025977855682373048, 0.02595737648010254, 0.025948160171508788, 0.026062847137451172, 0.025965568542480468, 0.026005504608154296, 0.026010623931884767, 0.025994239807128908, 0.025935871124267578, 0.026876928329467774, 0.02693222427368164, 0.027124736785888674, 0.027088895797729492, 0.027019264221191407, 0.027026432037353516, 0.026951679229736326, 0.027002880096435547, 0.027088895797729492, 0.026909696578979493, 0.027230207443237304, 0.027076608657836915, 0.026960895538330077, 0.02696294403076172, 0.027002880096435547, 0.026960895538330077, 0.02695680046081543, 0.026948640823364258, 0.027046880722045898, 0.027125759124755858, 0.02707148742675781, 0.02635468864440918, 0.02588569641113281, 0.026051584243774413, 0.026060800552368164, 0.026020927429199217, 0.02623891258239746, 0.026065919876098635, 0.026041343688964845, 0.026073087692260744, 0.026056703567504884, 0.026022911071777344, 0.02688204765319824, 0.02697113609313965, 0.027099136352539063, 0.02690662384033203, 0.026892288208007813, 0.027035648345947266, 0.027023359298706053, 0.026884096145629883, 0.026928127288818358, 0.02692915153503418, 0.06595686340332031, 0.026848255157470705, 0.02690355110168457, 0.027062271118164064, 0.027304960250854493, 0.027249664306640626, 0.02709199905395508, 0.027053024291992186, 0.02736025619506836, 0.02719436836242676, 0.02696703910827637, 0.027068416595458986, 0.026876928329467774, 0.026070016860961914, 0.026005504608154296, 0.0261212158203125, 0.02615603256225586, 0.026998783111572267, 0.027281408309936524, 0.027034624099731445, 0.02705311965942383, 0.027073471069335938, 0.027032575607299804, 0.02713907241821289, 0.02696601676940918, 0.027000831604003905, 0.0265850887298584, 0.02697113609313965, 0.027052032470703126, 0.026936319351196288, 0.026845184326171875, 0.027288576126098633, 0.026901504516601563, 0.026901504516601563, 0.027543552398681642, 0.027382783889770508, 0.027064319610595702, 0.02697222328186035, 0.02688505554199219, 0.02705820846557617, 0.02699977684020996, 0.027274240493774415, 0.02700595283508301, 0.027044864654541017, 0.027072511672973632, 0.027026432037353516, 0.02698240089416504, 0.027081727981567383, 0.02712883186340332, 0.02756096076965332, 0.026704896926879884, 0.026908672332763672, 0.02693836784362793, 0.02698956871032715, 0.026874879837036132, 0.026993663787841796, 0.02699673652648926, 0.02710527992248535, 0.026969120025634764, 0.027040735244750976, 0.02698240089416504, 0.026942464828491212, 0.026953727722167968, 0.06499635314941406, 0.026013696670532226, 0.026012672424316406, 0.026011648178100585, 0.025959423065185547, 0.026016767501831056, 0.025971712112426756, 0.02608742332458496, 0.02591744041442871, 0.02608332824707031, 0.02695782470703125, 0.02701312065124512, 0.0273756160736084, 0.02715443229675293, 0.027220991134643553, 0.0271329288482666, 0.027045919418334962, 0.02695369529724121, 0.026945535659790038, 0.026999807357788085, 0.0269434871673584, 0.026984447479248046, 0.027085823059082033, 0.027030527114868166, 0.026842111587524413, 0.027010047912597656, 0.026959871292114256, 0.02695884895324707, 0.02690559959411621, 0.026885120391845704, 0.027057151794433593, 0.027222015380859374, 0.026969087600708007, 0.026955808639526367, 0.026957792282104494, 0.02694758415222168, 0.026993663787841796, 0.02694041633605957, 0.026885120391845704, 0.027072511672973632, 0.026987520217895508, 0.026918912887573244, 0.02691481590270996, 0.026952703475952147, 0.027241472244262696, 0.02697318458557129, 0.026908735275268554, 0.027116479873657225, 0.027019264221191407, 0.02715443229675293, 0.02705516815185547, 0.0272608642578125, 0.027087936401367186, 0.026987455368041993, 0.027001855850219726, 0.027142143249511717, 0.0269752311706543, 0.026951679229736326, 0.02697420883178711, 0.027038719177246092, 0.027018239974975586, 0.027067392349243165, 0.02691993522644043, 0.06595174407958984, 0.027057151794433593, 0.027023359298706053, 0.027166719436645507, 0.026986495971679687, 0.02707967948913574, 0.02693836784362793, 0.027037727355957032, 0.02694243240356445, 0.0271646728515625, 0.027600896835327147, 0.02740838432312012, 0.027089920043945313, 0.027032575607299804, 0.02694655990600586, 0.027088895797729492, 0.02698137664794922, 0.026933248519897462, 0.027059200286865235, 0.02700492858886719, 0.027124736785888674, 0.027060224533081056, 0.027140096664428712, 0.027157503128051756, 0.027017215728759765, 0.026070016860961914, 0.026064895629882814, 0.025956352233886718, 0.025959423065185547, 0.025968639373779297, 0.026014720916748047, 0.026077184677124023, 0.02592665672302246, 0.0259368953704834, 0.026001407623291017, 0.02598297691345215, 0.02595327949523926, 0.026063871383666993, 0.02593382453918457, 0.02611404800415039, 0.026009599685668947, 0.026070016860961914, 0.026034175872802736, 0.025990144729614258, 0.025981952667236328, 0.026011648178100585, 0.02607923126220703, 0.02615910339355469, 0.02596249580383301, 0.02595840072631836, 0.025903104782104492, 0.025989120483398437, 0.025960447311401368, 0.02597068786621094, 0.026014720916748047, 0.02726911926269531, 0.02705311965942383, 0.02698851203918457, 0.027003936767578125, 0.027001792907714844, 0.02695884895324707, 0.02697420883178711, 0.02713599967956543, 0.06481613159179687, 0.025976831436157227, 0.026065919876098635, 0.025956352233886718, 0.025976863861083985, 0.025912288665771485, 0.025742336273193358, 0.025738239288330078, 0.02571980857849121, 0.025683967590332032, 0.0259420166015625, 0.02596659278869629, 0.02596767997741699, 0.025823232650756835, 0.025933759689331055, 0.02612531280517578, 0.026023935317993165, 0.025999359130859375, 0.026004480361938476, 0.02614784049987793, 0.025992191314697266, 0.026029056549072265, 0.026171392440795898, 0.026100736618041992, 0.02592870330810547, 0.026853376388549805, 0.02718003273010254, 0.026991615295410155, 0.027104352951049803, 0.02730793571472168, 0.027288576126098633, 0.027250688552856447, 0.02710937690734863, 0.027083776473999024, 0.027062271118164064, 0.027378688812255858, 0.027240447998046875, 0.027054079055786134, 0.026983423233032225, 0.027189279556274416, 0.02697929573059082, 0.027081727981567383, 0.02687283134460449, 0.02730086326599121, 0.028120063781738282, 0.02719340705871582, 0.026993600845336915, 0.027061248779296877, 0.027297792434692384, 0.027084800720214845, 0.027064384460449217, 0.026728384017944334, 0.027010047912597656, 0.02698137664794922, 0.02695577621459961, 0.027065343856811523, 0.026885120391845704, 0.02692198371887207, 0.02696703910827637, 0.026927167892456055, 0.026924991607666017, 0.02699673652648926, 0.026983488082885743, 0.0659486083984375, 0.027015167236328123, 0.02710937690734863, 0.027196416854858397, 0.02693939208984375, 0.02695577621459961, 0.02694758415222168, 0.027455488204956056, 0.02853068733215332, 0.027623424530029295, 0.02710527992248535, 0.027217920303344727, 0.02715545654296875, 0.0269803524017334, 0.02695680046081543, 0.027021312713623048, 0.027044864654541017, 0.027027456283569336, 0.02714112091064453, 0.027085823059082033, 0.027046911239624022, 0.026953727722167968, 0.027108352661132814, 0.0271329288482666, 0.027089920043945313, 0.02710425567626953, 0.02698137664794922, 0.027035680770874024, 0.027056095123291015, 0.02707155227661133, 0.02708883285522461, 0.027195423126220704, 0.02705200004577637, 0.026961919784545898, 0.02696703910827637, 0.027030527114868166, 0.02693529510498047, 0.0269434871673584, 0.026990591049194337, 0.027045888900756834, 0.026895360946655275, 0.026999807357788085, 0.02698137664794922, 0.027048959732055664, 0.02694963264465332, 0.02733465576171875, 0.027691007614135742, 0.02839347267150879, 0.02731520080566406, 0.027185152053833008, 0.0269803524017334, 0.026992639541625976, 0.027035648345947266, 0.027030527114868166, 0.02698240089416504, 0.027072511672973632, 0.027027456283569336, 0.027001855850219726, 0.027053056716918947, 0.026984447479248046, 0.027047935485839843, 0.026951679229736326, 0.027015167236328123, 0.06589030456542969, 0.026983423233032225, 0.027019264221191407, 0.026968063354492186, 0.026977279663085937, 0.02730700874328613, 0.02697318458557129, 0.027048959732055664, 0.02712678337097168, 0.027051008224487305, 0.02731110382080078, 0.027174911499023437, 0.026953727722167968, 0.027000831604003905, 0.02689945602416992, 0.02705510330200195, 0.027011072158813477, 0.027107328414916993, 0.02698854446411133, 0.027014144897460936, 0.026961919784545898, 0.027059200286865235, 0.027106304168701172, 0.02715340805053711, 0.026968063354492186, 0.027051008224487305, 0.026736640930175783, 0.027594751358032226, 0.027266048431396486, 0.02698956871032715, 0.02697318458557129, 0.027159551620483398, 0.02696499252319336, 0.02697420883178711, 0.027280384063720704, 0.027048959732055664, 0.027031551361083983, 0.02696703910827637, 0.02696396827697754, 0.02707865524291992, 0.02705510330200195, 0.026945535659790038, 0.026918912887573244, 0.026953727722167968, 0.027000831604003905, 0.027018239974975586, 0.02693939208984375, 0.027100160598754884, 0.02695884895324707, 0.027007999420166014, 0.02700595283508301, 0.027022367477416993, 0.027075551986694337, 0.027001855850219726, 0.02714521598815918, 0.027258880615234377, 0.027001855850219726, 0.02711756706237793, 0.02711961555480957, 0.026942464828491212, 0.026952703475952147, 0.02694963264465332, 0.026950656890869142, 0.06597734069824218, 0.02693734359741211, 0.026918975830078126, 0.026946495056152344, 0.02690457534790039, 0.026993663787841796, 0.027066368103027344, 0.02732748794555664, 0.027100160598754884, 0.027014144897460936, 0.026986495971679687, 0.02709401512145996, 0.027072511672973632, 0.02720256042480469, 0.026878976821899415, 0.027148288726806642, 0.027073535919189453, 0.027064319610595702, 0.027024383544921874, 0.027012096405029298, 0.027732992172241212, 0.027082752227783204, 0.027009023666381835, 0.027107328414916993, 0.027035648345947266, 0.027151359558105468, 0.02692403221130371, 0.02736639976501465, 0.02715545654296875, 0.02713497543334961, 0.02694963264465332, 0.02708684730529785, 0.02692300796508789, 0.0269803524017334, 0.027059200286865235, 0.027048959732055664, 0.026961919784545898, 0.02698956871032715, 0.027046911239624022, 0.02731724739074707, 0.026680320739746095, 0.026995712280273438, 0.02692095947265625, 0.027060224533081056, 0.026994688034057617, 0.02700595283508301, 0.026944511413574217, 0.027204608917236327, 0.02691276741027832, 0.026984447479248046, 0.026936384201049806, 0.026974143981933593, 0.0269752311706543, 0.026978303909301758, 0.02693939208984375, 0.027051008224487305, 0.02686566352844238, 0.02696703910827637, 0.02697420883178711, 0.027019264221191407, 0.026950656890869142, 0.026977279663085937, 0.02696601676940918, 0.06687026977539062, 0.027146240234375, 0.026933248519897462, 0.02695475196838379, 0.02692403221130371, 0.026852352142333984, 0.0269486083984375, 0.027090944290161133, 0.026853376388549805, 0.027206655502319335, 0.027046943664550783, 0.02696291160583496, 0.026818559646606444, 0.027083776473999024, 0.02691584014892578, 0.027059200286865235, 0.027009023666381835, 0.027001855850219726, 0.02693836784362793, 0.026978303909301758, 0.02692403221130371, 0.02694144058227539, 0.027006975173950197, 0.027058176040649414, 0.02693734359741211, 0.02692198371887207, 0.026976255416870116, 0.026933248519897462, 0.026986495971679687, 0.027115520477294923, 0.026999807357788085, 0.027028480529785157, 0.02694764709472656, 0.026938304901123047, 0.02696499252319336, 0.026902528762817384, 0.0270960636138916, 0.027151359558105468, 0.02697113609313965, 0.027124799728393555, 0.02686867141723633, 0.026969087600708007, 0.02690355110168457, 0.02699673652648926, 0.026851327896118164, 0.026612735748291014, 0.026868736267089844, 0.027020288467407227, 0.026901504516601563, 0.027037696838378908, 0.026894336700439454, 0.0269752311706543, 0.02696396827697754, 0.027007999420166014, 0.026968063354492186, 0.027192319869995117, 0.026908672332763672, 0.027029535293579102, 0.02698031997680664, 0.027067392349243165, 0.027448320388793947, 0.02666700744628906, 0.0260250244140625]",tokens/s,36.408493145097246,,,,, -bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3954,7 +3954,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 448.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1277.599744,5222.432768,0.0,4575.985664,4408.404992,s,10,0.25400146865844725,0.025400146865844724,0.0011341561227013026,0.02571254348754883,0.026372268295288086,0.02708008651733398,0.027646341094970703,"[0.027787904739379882, 0.025948863983154297, 0.025637727737426758, 0.024192192077636718, 0.024207199096679687, 0.0257873592376709, 0.025839487075805665, 0.026214975357055664, 0.024190399169921876, 0.02419536018371582]",tokens/s,10078.681881333536,kWh,2.890672709785768e-07,1.5835740973770973e-07,1.3255090745330683e-06,1.7729337552493548e-06,tokens/kWh,144393437.8495686,MB,1277.599744,5222.432768,0.0,4575.985664,4408.407552,s,10,14.477070800781249,1.4477070800781249,0.0393811811920201,1.469327880859375,1.4762752685546874,1.4882212524414062,1.4977780395507814,"[1.500167236328125, 1.47362060546875, 1.4732127685546874, 1.4460653076171874, 1.3849686279296876, 1.4007744140625, 1.4701929931640625, 1.472905517578125, 1.3867005615234376, 1.4684627685546876]",tokens/s,43.51709048532127,kWh,1.7152702484570325e-05,9.398499110037754e-06,4.147702516746769e-05,6.802822676207577e-05,tokens/kWh,926086.1703824552,,s,629,14.677949451446525,0.02333537273679894,0.0030918058258084533,0.02332979202270508,0.023565312576293945,0.024161074829101563,0.0465539079284668,"[0.02447667121887207, 0.024321023941040038, 0.024489984512329102, 0.024466432571411133, 0.02455449676513672, 0.02447667121887207, 0.024414207458496092, 0.024225791931152343, 0.024343551635742186, 0.024534015655517577, 0.02448896026611328, 0.025266176223754884, 0.024423423767089843, 0.02433126449584961, 0.02464460754394531, 0.02484230422973633, 0.024510400772094727, 0.024212480545043946, 0.024246271133422852, 0.02432614326477051, 0.024137727737426756, 0.02411520004272461, 0.0240762882232666, 0.023802879333496094, 0.023525375366210938, 0.023545856475830077, 0.023785472869873047, 0.023442432403564452, 0.023411712646484374, 0.023390207290649414, 0.023533567428588868, 0.023368703842163087, 0.023343103408813477, 0.023393280029296876, 0.023318527221679687, 0.023384063720703126, 0.023366655349731445, 0.02333695983886719, 0.023515167236328124, 0.02356118392944336, 0.023359487533569336, 0.023322656631469728, 0.023494623184204103, 0.023362560272216795, 0.023364608764648437, 0.02345471954345703, 0.023444480895996093, 0.023385087966918947, 0.02328166389465332, 0.023339008331298827, 0.0233175048828125, 0.02329599952697754, 0.02326118469238281, 0.02332262420654297, 0.023351295471191406, 0.02327142333984375, 0.02330828857421875, 0.02329599952697754, 0.023344127655029297, 0.02331648063659668, 0.023401472091674806, 0.02328883171081543, 0.049775615692138675, 0.023232511520385742, 0.023362560272216795, 0.02326323127746582, 0.023357440948486328, 0.023326719284057617, 0.023739391326904297, 0.02349158477783203, 0.02331443214416504, 0.02325606346130371, 0.023326719284057617, 0.023348224639892577, 0.023294015884399413, 0.023399360656738283, 0.023000064849853515, 0.023372800827026367, 0.02332876777648926, 0.023459840774536132, 0.023368703842163087, 0.023368703842163087, 0.02327142333984375, 0.023347200393676756, 0.023371776580810546, 0.023367679595947266, 0.023267328262329103, 0.023342079162597656, 0.023376895904541017, 0.023392255783081056, 0.023330816268920897, 0.02331648063659668, 0.023311424255371093, 0.02346284866333008, 0.023378944396972655, 0.02342911911010742, 0.023407648086547852, 0.02336764717102051, 0.023372800827026367, 0.023384063720703126, 0.023370752334594725, 0.02352025604248047, 0.023463935852050782, 0.023369728088378908, 0.023355392456054686, 0.023436288833618164, 0.023401472091674806, 0.023582719802856447, 0.023446527481079102, 0.023497791290283204, 0.0233686408996582, 0.02371788787841797, 0.02332569694519043, 0.023311359405517578, 0.023310335159301757, 0.023382015228271484, 0.023443456649780273, 0.02346905517578125, 0.023377920150756838, 0.023430143356323242, 0.023335935592651368, 0.023427072525024413, 0.023382015228271484, 0.023638015747070314, 0.023356416702270507, 0.04945612716674805, 0.023371776580810546, 0.02288332748413086, 0.022931455612182617, 0.023343103408813477, 0.023410688400268553, 0.023391231536865235, 0.023394304275512694, 0.023285791397094725, 0.02334511947631836, 0.023353343963623048, 0.02345577621459961, 0.0233287353515625, 0.023373823165893554, 0.02330828857421875, 0.023167999267578124, 0.02294988822937012, 0.023584768295288085, 0.02352639961242676, 0.023370752334594725, 0.023351295471191406, 0.023391231536865235, 0.023330816268920897, 0.02350387191772461, 0.023357440948486328, 0.0232806396484375, 0.023353343963623048, 0.023665664672851562, 0.023292959213256834, 0.02336867141723633, 0.023363584518432616, 0.02346905517578125, 0.02348134422302246, 0.02332876777648926, 0.02332979202270508, 0.023380992889404296, 0.023434240341186522, 0.023362560272216795, 0.023406591415405274, 0.02354380798339844, 0.023455743789672853, 0.023426048278808592, 0.023359487533569336, 0.023416831970214845, 0.023427072525024413, 0.02351206398010254, 0.023417856216430662, 0.02353049659729004, 0.023412736892700195, 0.023426048278808592, 0.023446527481079102, 0.023378944396972655, 0.02352639961242676, 0.023415807723999024, 0.023421951293945312, 0.023446527481079102, 0.02329395294189453, 0.023426048278808592, 0.023406591415405274, 0.023362560272216795, 0.023357440948486328, 0.023335935592651368, 0.023455743789672853, 0.046565376281738284, 0.021884927749633788, 0.021916671752929686, 0.021942272186279296, 0.02190336036682129, 0.0218920955657959, 0.02230886459350586, 0.021970943450927736, 0.02189619255065918, 0.0219289608001709, 0.02200371170043945, 0.022008832931518556, 0.02182143974304199, 0.021789695739746092, 0.0218470401763916, 0.021946367263793946, 0.02185523223876953, 0.02183782386779785, 0.021854240417480467, 0.022307807922363282, 0.02395238494873047, 0.02511769676208496, 0.02366771125793457, 0.023417856216430662, 0.02329910469055176, 0.023410655975341796, 0.023536640167236327, 0.0234967041015625, 0.023414783477783203, 0.023388160705566406, 0.023240703582763672, 0.023409664154052736, 0.023385087966918947, 0.023340032577514647, 0.02327347183227539, 0.02330419158935547, 0.02332057571411133, 0.022975488662719725, 0.02330521583557129, 0.023426048278808592, 0.023577600479125976, 0.02333798408508301, 0.023370752334594725, 0.023376895904541017, 0.02339638328552246, 0.023255008697509766, 0.023418880462646483, 0.023599103927612306, 0.022979583740234375, 0.02332569694519043, 0.023723007202148438, 0.023369728088378908, 0.02333695983886719, 0.02326118469238281, 0.023065631866455077, 0.023078880310058593, 0.023218175888061524, 0.02335436820983887, 0.02333798408508301, 0.023550975799560548, 0.023391231536865235, 0.023400447845458985, 0.023398399353027344, 0.04666470336914062, 0.02182963180541992, 0.021924863815307616, 0.021833728790283204, 0.021873664855957032, 0.021828607559204103, 0.02191155242919922, 0.02188800048828125, 0.021737472534179687, 0.021612543106079102, 0.021813247680664064, 0.021980159759521483, 0.022039552688598633, 0.02167807960510254, 0.021634048461914062, 0.021769216537475586, 0.0218787841796875, 0.021970943450927736, 0.021805055618286134, 0.021801984786987305, 0.021866559982299805, 0.021868480682373046, 0.021864511489868163, 0.0219299201965332, 0.02184499168395996, 0.02180201530456543, 0.021610464096069336, 0.02186649513244629, 0.02186140823364258, 0.02192278480529785, 0.02242355155944824, 0.022611967086791994, 0.022054912567138672, 0.021972991943359374, 0.021813247680664064, 0.02186956787109375, 0.02186649513244629, 0.021843967437744142, 0.021840896606445313, 0.0219289608001709, 0.0218603515625, 0.021803007125854493, 0.021843967437744142, 0.021865472793579102, 0.02185830307006836, 0.02186137580871582, 0.021849088668823242, 0.02189004707336426, 0.02191974449157715, 0.02183782386779785, 0.021820415496826173, 0.021811199188232423, 0.02184601593017578, 0.02184499168395996, 0.021910528182983398, 0.02192793655395508, 0.021783552169799804, 0.02183782386779785, 0.02210406494140625, 0.022245376586914063, 0.024010751724243166, 0.023772159576416017, 0.02352332878112793, 0.04823859024047852, 0.022140928268432617, 0.021952512741088868, 0.021931007385253908, 0.021873664855957032, 0.021944320678710938, 0.02192793655395508, 0.02205183982849121, 0.02203651237487793, 0.02211734390258789, 0.021949472427368163, 0.021790687561035155, 0.02185625648498535, 0.02189619255065918, 0.021811199188232423, 0.021939199447631837, 0.021946367263793946, 0.021799936294555664, 0.021850112915039063, 0.02185420799255371, 0.02184499168395996, 0.02186342430114746, 0.021836799621582033, 0.021944320678710938, 0.0219238395690918, 0.0218787841796875, 0.02190745544433594, 0.02188390350341797, 0.02165862464904785, 0.021594112396240234, 0.021590015411376954, 0.021778432846069336, 0.0218787841796875, 0.02187059211730957, 0.0218920955657959, 0.021918720245361328, 0.021906431198120118, 0.021917695999145507, 0.021638143539428712, 0.02188083267211914, 0.02190540885925293, 0.021850112915039063, 0.02188902473449707, 0.021811199188232423, 0.021818368911743165, 0.021917695999145507, 0.02184806442260742, 0.022122495651245116, 0.025406463623046875, 0.024176639556884767, 0.023573503494262696, 0.022031423568725585, 0.021895103454589844, 0.02208768081665039, 0.02344960021972656, 0.023444480895996093, 0.023790592193603514, 0.023463935852050782, 0.023374847412109375, 0.023293983459472655, 0.023298015594482423, 0.023365631103515624, 0.02331340789794922, 0.04960768127441406, 0.02347110366821289, 0.02325299263000488, 0.022972415924072266, 0.02291097640991211, 0.023639039993286134, 0.02306764793395996, 0.023257087707519532, 0.023323648452758788, 0.02327039909362793, 0.02353561592102051, 0.02330009651184082, 0.023311359405517578, 0.02306662368774414, 0.023012351989746094, 0.02332569694519043, 0.02325606346130371, 0.02329497528076172, 0.02353049659729004, 0.023403520584106444, 0.02347417640686035, 0.02330419158935547, 0.02325503921508789, 0.023347200393676756, 0.02323865509033203, 0.023563264846801758, 0.02330316734313965, 0.023342079162597656, 0.02330316734313965, 0.023249919891357423, 0.023331840515136718, 0.023557119369506836, 0.02345062446594238, 0.02333286476135254, 0.023408639907836915, 0.023401472091674806, 0.023712799072265624, 0.0233338565826416, 0.023393280029296876, 0.023537664413452147, 0.023341056823730468, 0.02333286476135254, 0.023266304016113282, 0.02330828857421875, 0.023387136459350585, 0.023393280029296876, 0.02327654457092285, 0.023404544830322265, 0.023556095123291015, 0.02332569694519043, 0.02328678321838379, 0.02352332878112793, 0.023340032577514647, 0.023266304016113282, 0.02330009651184082, 0.02331340789794922, 0.023361568450927735, 0.02331644821166992, 0.023174144744873046, 0.022986751556396484, 0.02326937675476074, 0.023345151901245118, 0.023339008331298827, 0.050165760040283204, 0.023383039474487305, 0.023378944396972655, 0.023378944396972655, 0.023404544830322265, 0.023276575088500978, 0.023394271850585936, 0.02388479995727539, 0.02341996765136719, 0.023425983428955077, 0.023451648712158202, 0.023438335418701172, 0.023404544830322265, 0.023399423599243165, 0.023740415573120118, 0.023417856216430662, 0.023403520584106444, 0.023441408157348635, 0.023444480895996093, 0.02332467269897461, 0.02328883171081543, 0.02335436820983887, 0.02328883171081543, 0.023378944396972655, 0.02342300796508789, 0.023326688766479493, 0.023351295471191406, 0.02330419158935547, 0.023304224014282227, 0.02335331153869629, 0.02330009651184082, 0.023431167602539063, 0.023299072265625, 0.02327244758605957, 0.023376928329467774, 0.023258079528808595, 0.02328883171081543, 0.02330624008178711, 0.023213056564331053, 0.023398399353027344, 0.02325299263000488, 0.023318527221679687, 0.023224384307861327, 0.023303104400634767, 0.02329804801940918, 0.02329292869567871, 0.023352319717407227, 0.023337024688720703, 0.02335532760620117, 0.023352319717407227, 0.02330931282043457, 0.023388160705566406, 0.0232857608795166, 0.02333695983886719, 0.023387136459350585, 0.023377920150756838, 0.02327859115600586, 0.0237260799407959, 0.023356416702270507, 0.023431167602539063, 0.023385087966918947, 0.02332467269897461, 0.02338102340698242, 0.04647727966308594, 0.02190342330932617, 0.021847999572753907, 0.02185215950012207, 0.02184601593017578, 0.021879840850830078, 0.02191663932800293, 0.021952512741088868, 0.021849088668823242, 0.021779455184936524, 0.021884927749633788, 0.021772287368774415, 0.021669952392578126, 0.02179372787475586, 0.021574655532836915, 0.021803007125854493, 0.021801984786987305, 0.02185523223876953, 0.021818368911743165, 0.02181228828430176, 0.021816255569458008, 0.021786624908447266, 0.021769216537475586, 0.021960704803466798, 0.021794815063476563, 0.02163609504699707, 0.02163609504699707, 0.02165657615661621, 0.0218603515625, 0.021793792724609375, 0.021816320419311523, 0.021959680557250977, 0.021901311874389647, 0.021914623260498048, 0.02184294319152832, 0.021933055877685546, 0.021770240783691407, 0.021763071060180664, 0.022226943969726562, 0.021901311874389647, 0.021803007125854493, 0.021792768478393554, 0.02187571144104004, 0.021816320419311523, 0.021815296173095702, 0.02183065605163574, 0.0218603515625, 0.021608448028564452, 0.02171801567077637, 0.02186342430114746, 0.02187980842590332, 0.02183475112915039, 0.02188083267211914, 0.021850112915039063, 0.021857280731201172, 0.022622207641601562, 0.023740415573120118, 0.023199743270874023, 0.023341056823730468, 0.02297145652770996, 0.023253952026367188, 0.023359487533569336, 0.02332467269897461, 0.046524417877197265, 0.021865472793579102, 0.02186444854736328, 0.022347776412963868, 0.023578624725341796, 0.023421951293945312, 0.023407615661621094, 0.02346086311340332, 0.023472127914428712, 0.023517183303833008, 0.023386112213134767, 0.023340032577514647, 0.023165952682495116, 0.023371776580810546, 0.02330112075805664, 0.02349567985534668, 0.023341056823730468, 0.02332979202270508, 0.02333695983886719, 0.023961599349975587, 0.023372800827026367, 0.023352319717407227, 0.02353971290588379, 0.023423999786376954, 0.023411712646484374, 0.023373823165893554, 0.023416831970214845, 0.02333695983886719, 0.023384063720703126, 0.023319551467895508, 0.023377920150756838, 0.023401472091674806, 0.023443456649780273, 0.023398399353027344, 0.02332262420654297, 0.023348224639892577, 0.023180288314819338, 0.02289254379272461, 0.02342911911010742, 0.023622655868530275, 0.023385087966918947, 0.023380992889404296, 0.02290790367126465, 0.023373823165893554, 0.023331840515136718, 0.023384063720703126, 0.023384063720703126, 0.023433216094970705, 0.023366655349731445, 0.023374847412109375, 0.023327743530273438, 0.023344127655029297, 0.023349248886108398, 0.023369728088378908, 0.023315456390380858, 0.023380992889404296, 0.02332057571411133, 0.023335935592651368, 0.02330931282043457, 0.023560192108154295, 0.023612415313720703, 0.02391756820678711, 0.023368703842163087]",tokens/s,42.85339734141209,,,main,False,False bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1197.768704,3538.419712,0.0,2891.972608,2759.918592,s,10,0.22341420936584472,0.02234142093658447,0.00029111464340745864,0.022243712425231932,0.022395789527893065,0.02280235834121704,0.02312761339187622,"[0.023208927154541016, 0.022286048889160158, 0.02230544090270996, 0.022193695068359377, 0.02222368049621582, 0.022235776901245115, 0.02225164794921875, 0.02222073554992676, 0.02227689552307129, 0.022211360931396484]",tokens/s,11458.537070074872,kWh,2.6478590429809244e-07,1.4508975474799803e-07,1.2966988450313906e-06,1.706574504077481e-06,tokens/kWh,150008100.66501334,MB,1197.768704,3538.419712,0.0,2891.972608,2796.953088,s,10,10.40897344970703,1.040897344970703,0.013836335344277185,1.0432322387695312,1.0601688232421875,1.0615470581054687,1.062649645996094,"[1.0455521240234376, 1.0386646728515625, 1.0411461181640624, 1.045318359375, 1.0273031005859374, 1.0465562744140624, 1.059862548828125, 1.0212946166992187, 1.020350341796875, 1.06292529296875]",tokens/s,60.524700446587445,kWh,1.2491260500143105e-05,6.8447319731998554e-06,3.38512181621686e-05,5.3187210635511567e-05,tokens/kWh,1184495.2808624394,,s,629,10.591118341445927,0.016838025980041214,0.0027084463618811923,0.016556032180786134,0.016903577423095704,0.017006591796875,0.0388757894897461,"[0.017183744430541992, 0.017547264099121093, 0.0164454402923584, 0.01615564727783203, 0.01619046401977539, 0.016202783584594725, 0.016288736343383788, 0.01615667152404785, 0.01618636894226074, 0.015978495597839357, 0.016028671264648436, 0.01618636894226074, 0.01621196746826172, 0.016108543395996093, 0.01623961639404297, 0.016228351593017578, 0.016339967727661133, 0.01621196746826172, 0.01615667152404785, 0.016339967727661133, 0.01622425651550293, 0.01619558334350586, 0.016256032943725587, 0.016133087158203125, 0.01658880043029785, 0.01682329559326172, 0.01617715263366699, 0.0160819206237793, 0.01596723175048828, 0.016323583602905273, 0.01685196876525879, 0.016556032180786134, 0.0164771842956543, 0.016656383514404297, 0.016920576095581053, 0.016724992752075195, 0.01681305694580078, 0.01678643226623535, 0.016760831832885743, 0.016701440811157226, 0.01619865608215332, 0.01679462432861328, 0.01677107238769531, 0.01682022476196289, 0.01684787178039551, 0.016781312942504883, 0.01680076789855957, 0.016755712509155272, 0.016915456771850586, 0.01681407928466797, 0.016854015350341797, 0.016706560134887697, 0.01702195167541504, 0.017006591796875, 0.01817190361022949, 0.017079296112060546, 0.016927743911743166, 0.016397312164306642, 0.016886783599853517, 0.01675775909423828, 0.016715776443481444, 0.016750591278076172, 0.03967488098144531, 0.016723968505859374, 0.01683148765563965, 0.01675468826293945, 0.016748544692993163, 0.01680691146850586, 0.016763904571533202, 0.01634611129760742, 0.01618636894226074, 0.01624575996398926, 0.016157695770263672, 0.016228351593017578, 0.016247840881347658, 0.016202720642089843, 0.01614028739929199, 0.016332799911499024, 0.01618943977355957, 0.01623142433166504, 0.01623859214782715, 0.01620889663696289, 0.016129024505615236, 0.016438272476196288, 0.01619660758972168, 0.016285696029663087, 0.01617715263366699, 0.016285696029663087, 0.016397312164306642, 0.01671785545349121, 0.016700384140014648, 0.016703487396240235, 0.016709632873535156, 0.016746496200561522, 0.016728063583374024, 0.016755712509155272, 0.016706560134887697, 0.016745471954345705, 0.016733184814453125, 0.016503807067871093, 0.016092159271240233, 0.01623961639404297, 0.016265216827392577, 0.01618022346496582, 0.01617203140258789, 0.01619660758972168, 0.01620070457458496, 0.016243711471557617, 0.016240640640258788, 0.016275455474853515, 0.01617407989501953, 0.016175104141235352, 0.01617616081237793, 0.016330720901489258, 0.016235519409179687, 0.016309247970581055, 0.016776191711425782, 0.01681100845336914, 0.016750591278076172, 0.01665433692932129, 0.016192512512207033, 0.016539648056030275, 0.01661644744873047, 0.018372608184814454, 0.017739776611328126, 0.03963187026977539, 0.016698368072509767, 0.016691200256347655, 0.016886783599853517, 0.016745471954345705, 0.016942079544067384, 0.016694271087646484, 0.016748544692993163, 0.016698368072509767, 0.016744447708129884, 0.01618534469604492, 0.01680793571472168, 0.01674448013305664, 0.01625289535522461, 0.016152576446533205, 0.01619558334350586, 0.01617817687988281, 0.01638297653198242, 0.016021503448486327, 0.016123903274536132, 0.01622425651550293, 0.01621299171447754, 0.016350208282470705, 0.016863231658935548, 0.016655359268188476, 0.016781312942504883, 0.016699392318725585, 0.016690176010131837, 0.016614463806152342, 0.01670035171508789, 0.016703487396240235, 0.01613830375671387, 0.015962047576904295, 0.01658060836791992, 0.016659456253051756, 0.016687103271484375, 0.01673936080932617, 0.016813024520874024, 0.016729087829589845, 0.016312320709228514, 0.01619660758972168, 0.01618534469604492, 0.016153600692749022, 0.0164003849029541, 0.016694271087646484, 0.016714752197265623, 0.016680959701538087, 0.01678643226623535, 0.01618943977355957, 0.01615564727783203, 0.016157695770263672, 0.016157695770263672, 0.016149503707885742, 0.016265216827392577, 0.016666656494140626, 0.016766944885253907, 0.017048576354980468, 0.0167956485748291, 0.016656383514404297, 0.01665126419067383, 0.016630783081054687, 0.016694271087646484, 0.016151552200317384, 0.03896627044677734, 0.01616383934020996, 0.016137216567993166, 0.01620172882080078, 0.016252927780151367, 0.016313343048095702, 0.016250879287719726, 0.016242687225341796, 0.01622323226928711, 0.01658470344543457, 0.016138240814208983, 0.01617100715637207, 0.016175104141235352, 0.01619558334350586, 0.016760831832885743, 0.016792575836181642, 0.016777215957641603, 0.016708608627319335, 0.01620070457458496, 0.016202751159667968, 0.016244735717773438, 0.01617305564880371, 0.016194559097290038, 0.016558080673217773, 0.016731136322021483, 0.016686080932617187, 0.016731199264526368, 0.016703424453735353, 0.01676595115661621, 0.016717824935913086, 0.016752639770507814, 0.01679667282104492, 0.01679667282104492, 0.016694271087646484, 0.016748544692993163, 0.016714752197265623, 0.016358400344848634, 0.016690176010131837, 0.016773120880126953, 0.01678643226623535, 0.01678233528137207, 0.016738304138183592, 0.016834592819213866, 0.016740320205688476, 0.01676288032531738, 0.016677888870239257, 0.01679052734375, 0.01676697540283203, 0.016874496459960937, 0.01643724822998047, 0.016631807327270508, 0.016709632873535156, 0.016792575836181642, 0.016755712509155272, 0.0167956485748291, 0.016747520446777343, 0.016744447708129884, 0.016737279891967775, 0.016867328643798828, 0.016730112075805666, 0.0168724479675293, 0.01678335952758789, 0.016903167724609376, 0.03890892791748047, 0.01617919921875, 0.016270368576049805, 0.01621705627441406, 0.01660108757019043, 0.016745471954345705, 0.01677926445007324, 0.01689910316467285, 0.016774112701416016, 0.016761856079101564, 0.01762303924560547, 0.016952352523803713, 0.016784351348876955, 0.016491519927978517, 0.017101823806762697, 0.01685606384277344, 0.01678643226623535, 0.016763904571533202, 0.016490495681762696, 0.016471040725708007, 0.016533504486083983, 0.016028671264648436, 0.016012287139892577, 0.016111616134643555, 0.016047103881835938, 0.01603276824951172, 0.016052223205566405, 0.0160184326171875, 0.016077823638916015, 0.016087039947509766, 0.016053247451782226, 0.016058368682861326, 0.016029695510864257, 0.016048160552978516, 0.01603171157836914, 0.016046079635620117, 0.016062463760375977, 0.015996928215026854, 0.016051200866699217, 0.016035839080810545, 0.016314367294311523, 0.016440319061279296, 0.016181280136108398, 0.016033760070800783, 0.01599078369140625, 0.016002048492431642, 0.01615564727783203, 0.016094207763671875, 0.016301088333129883, 0.016063488006591797, 0.01625904083251953, 0.016184320449829103, 0.016283647537231445, 0.016259071350097656, 0.016260095596313476, 0.01620070457458496, 0.01618124771118164, 0.01619152069091797, 0.016125919342041016, 0.01602764892578125, 0.016030719757080078, 0.016119808197021485, 0.016004095077514647, 0.03882086563110351, 0.01680384063720703, 0.016724992752075195, 0.016715776443481444, 0.016729087829589845, 0.016759807586669923, 0.016257024765014647, 0.016846847534179688, 0.016252927780151367, 0.017087488174438475, 0.01804902458190918, 0.0170700798034668, 0.01677824020385742, 0.01679769515991211, 0.016745471954345705, 0.016805919647216797, 0.01670345687866211, 0.016192512512207033, 0.016159744262695314, 0.01616383934020996, 0.016281600952148437, 0.016755712509155272, 0.016730112075805666, 0.016744447708129884, 0.016729087829589845, 0.01617100715637207, 0.016146432876586913, 0.016167936325073243, 0.016194559097290038, 0.016540672302246092, 0.016738304138183592, 0.016863264083862305, 0.016797664642333985, 0.01616383934020996, 0.01623142433166504, 0.016175104141235352, 0.016187456130981444, 0.016136127471923827, 0.01621401596069336, 0.01625497627258301, 0.016740352630615234, 0.016631807327270508, 0.016356351852416993, 0.016733184814453125, 0.016703487396240235, 0.0168222713470459, 0.01686425590515137, 0.016891904830932617, 0.016915456771850586, 0.01681305694580078, 0.016349184036254884, 0.016730112075805666, 0.01680998420715332, 0.01681612777709961, 0.016928768157958983, 0.016270336151123048, 0.016333824157714845, 0.01637990379333496, 0.01679871940612793, 0.016734207153320312, 0.01683456039428711, 0.016777215957641603, 0.016881664276123046, 0.03941273498535156, 0.0168724479675293, 0.01681920051574707, 0.016965631484985352, 0.01679052734375, 0.016915456771850586, 0.016748544692993163, 0.016870399475097657, 0.016870399475097657, 0.01677414321899414, 0.01679052734375, 0.01685196876525879, 0.017006591796875, 0.016744447708129884, 0.01679155158996582, 0.01679155158996582, 0.016763904571533202, 0.01681305694580078, 0.016742399215698242, 0.01676595115661621, 0.016721920013427736, 0.016703519821166992, 0.016768991470336916, 0.016973888397216797, 0.01676691246032715, 0.016865280151367186, 0.016743423461914063, 0.016777215957641603, 0.016712703704833985, 0.01681407928466797, 0.016751615524291993, 0.017123327255249024, 0.01678950309753418, 0.016793600082397463, 0.01682329559326172, 0.016788480758666992, 0.016756736755371093, 0.016747520446777343, 0.01683148765563965, 0.016825344085693358, 0.017082368850708008, 0.01761484718322754, 0.01760665512084961, 0.017183744430541992, 0.016899072647094726, 0.016842752456665038, 0.01705779266357422, 0.016877567291259766, 0.016780288696289062, 0.01683660888671875, 0.016742399215698242, 0.016196640014648437, 0.01628976058959961, 0.01679667282104492, 0.01677824020385742, 0.016895999908447267, 0.01663692855834961, 0.016849920272827147, 0.01682636833190918, 0.016424959182739257, 0.017123327255249024, 0.01637068748474121, 0.016251903533935547, 0.0388935661315918, 0.016280576705932616, 0.016232448577880858, 0.016199680328369142, 0.01618227195739746, 0.016151552200317384, 0.016165887832641602, 0.016235519409179687, 0.01617919921875, 0.016289791107177733, 0.016252927780151367, 0.01620070457458496, 0.016154624938964843, 0.016280576705932616, 0.016151552200317384, 0.016194559097290038, 0.016161792755126952, 0.01616281509399414, 0.016157695770263672, 0.0161474552154541, 0.01618022346496582, 0.01620172882080078, 0.01617919921875, 0.016434175491333008, 0.016219135284423827, 0.01623859214782715, 0.016227327346801757, 0.01619046401977539, 0.01626316833496094, 0.016150527954101563, 0.01622425651550293, 0.01620479965209961, 0.01621299171447754, 0.01620172882080078, 0.016302080154418946, 0.016331775665283203, 0.01620582389831543, 0.016152576446533205, 0.01622118377685547, 0.016234495162963866, 0.01618739128112793, 0.01614847946166992, 0.01621401596069336, 0.016160768508911134, 0.01622118377685547, 0.01619558334350586, 0.016201759338378908, 0.016211936950683594, 0.01618739128112793, 0.016160768508911134, 0.01618227195739746, 0.016120832443237306, 0.01618227195739746, 0.016210943222045898, 0.01622528076171875, 0.01620070457458496, 0.01617715263366699, 0.016159744262695314, 0.016261119842529297, 0.016158720016479493, 0.016199680328369142, 0.01620070457458496, 0.016149503707885742, 0.038989822387695314, 0.016088064193725587, 0.016088064193725587, 0.01604198455810547, 0.01618227195739746, 0.01617203140258789, 0.01619660758972168, 0.016232448577880858, 0.01620070457458496, 0.01618124771118164, 0.016335872650146483, 0.01617203140258789, 0.01618227195739746, 0.01619660758972168, 0.01618022346496582, 0.016269311904907227, 0.016133119583129883, 0.01619660758972168, 0.01615564727783203, 0.016141311645507812, 0.01637580871582031, 0.016159744262695314, 0.016194559097290038, 0.01623347282409668, 0.01623142433166504, 0.01618124771118164, 0.016167936325073243, 0.01621504020690918, 0.016248832702636717, 0.016149503707885742, 0.016264192581176756, 0.01620172882080078, 0.016153600692749022, 0.01613926315307617, 0.01617715263366699, 0.016167936325073243, 0.016184320449829103, 0.01616281509399414, 0.01619865608215332, 0.016232511520385743, 0.016261056900024416, 0.016134143829345703, 0.016142335891723633, 0.016199680328369142, 0.01616383934020996, 0.01617100715637207, 0.01617919921875, 0.0161976318359375, 0.01619558334350586, 0.016144384384155275, 0.01619865608215332, 0.016191488265991212, 0.01620684814453125, 0.01616998481750488, 0.01619046401977539, 0.016282623291015624, 0.016149503707885742, 0.01618227195739746, 0.0162293758392334, 0.016165887832641602, 0.016191488265991212, 0.016157695770263672, 0.016175104141235352, 0.038830078125, 0.016130048751831053, 0.016566272735595702, 0.016849920272827147, 0.01683148765563965, 0.01686841583251953, 0.016824256896972655, 0.016954368591308593, 0.016969728469848632, 0.016861183166503906, 0.016913408279418944, 0.01683046340942383, 0.016915456771850586, 0.016867328643798828, 0.01682022476196289, 0.016928768157958983, 0.016930816650390625, 0.016937984466552734, 0.016951295852661134, 0.016946176528930663, 0.016891904830932617, 0.016957439422607423, 0.016866304397583007, 0.016889856338500975, 0.016884735107421875, 0.01683660888671875, 0.016837631225585938, 0.01683046340942383, 0.016837631225585938, 0.016888832092285155, 0.016913408279418944, 0.016894975662231446, 0.016833536148071288, 0.01680076789855957, 0.016859136581420898, 0.016973823547363282, 0.016882688522338866, 0.01681407928466797, 0.01683660888671875, 0.016869375228881836, 0.01683456039428711, 0.016849920272827147, 0.016921600341796874, 0.01683558464050293, 0.01681715202331543, 0.016905216217041014, 0.016888832092285155, 0.016909311294555664, 0.016899072647094726, 0.0168407039642334, 0.01680691146850586, 0.0168222713470459, 0.016913408279418944, 0.016915456771850586, 0.01686425590515137, 0.01754832077026367, 0.017020896911621095, 0.016879615783691407, 0.016910335540771485, 0.016885759353637696, 0.016912384033203123, 0.01698921585083008, 0.016853984832763673]",tokens/s,59.38938455050135,,,,, -bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4195,7 +4195,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1303.093248,17102.798848,0.0,16456.351744,16359.853056,s,10,1.0194314193725587,0.10194314193725587,0.0002529759528872485,0.10188860702514649,0.10205652389526366,0.10236344413757324,0.1026089803314209,"[0.1026703643798828, 0.10180748748779297, 0.10198831939697266, 0.10172902679443359, 0.10191126251220703, 0.10188739013671876, 0.1017694091796875, 0.10191958618164063, 0.10188982391357422, 0.10185874938964844]",tokens/s,2511.203746864731,kWh,1.2064162498158392e-06,6.610585252520669e-07,5.900701999469618e-06,7.768176774537524e-06,tokens/kWh,32954965.80859424,MB,1303.420928,17102.798848,0.0,16456.351744,16416.242176,s,10,23.7518544921875,2.3751854492187503,0.0033105326535980578,2.3739964599609373,2.3780450195312497,2.38121611328125,2.38375298828125,"[2.38438720703125, 2.373756591796875, 2.37312353515625, 2.374914306640625, 2.37734033203125, 2.373713134765625, 2.3730146484375, 2.374236328125, 2.374585693359375, 2.37278271484375]",tokens/s,26.524244673493623,kWh,2.7978138962840816e-05,1.5333065776871895e-05,0.00011402269949472894,0.00015733390423444167,tokens/kWh,400422.27583778976,,s,629,24.629738498687743,0.03915697694544951,0.012174665303252123,0.03768115234375,0.03781857223510742,0.03790479431152344,0.14013506042480467,"[0.038335487365722655, 0.03846758270263672, 0.03829350280761719, 0.038324222564697266, 0.038228992462158204, 0.03825766372680664, 0.038346752166748044, 0.0383559684753418, 0.0382105598449707, 0.03811123275756836, 0.03820032119750977, 0.038307838439941407, 0.03793817520141601, 0.03768729782104492, 0.03760748672485351, 0.037632961273193356, 0.0377149429321289, 0.037819393157958986, 0.037556224822998044, 0.03764223861694336, 0.037634048461914066, 0.03768220901489258, 0.037555168151855466, 0.03766886520385742, 0.03762278366088867, 0.03767910385131836, 0.037694465637207034, 0.037678081512451174, 0.0376258544921875, 0.03778252792358398, 0.037634048461914066, 0.03767193603515625, 0.037579776763916016, 0.037705726623535156, 0.03761459350585938, 0.037787647247314454, 0.03773952102661133, 0.03771907043457031, 0.037696479797363285, 0.03768627166748047, 0.037663745880126956, 0.03773747253417969, 0.037759998321533206, 0.037766143798828124, 0.03769651031494141, 0.0377784309387207, 0.03773440170288086, 0.037765121459960936, 0.03769036865234375, 0.037776382446289065, 0.037722110748291016, 0.037771263122558595, 0.03773235321044922, 0.0378081283569336, 0.037767230987548826, 0.037794750213623045, 0.03775385665893555, 0.03780819320678711, 0.037739456176757814, 0.037852161407470705, 0.037789695739746096, 0.03779993438720703, 0.1402255401611328, 0.03753676986694336, 0.037548030853271484, 0.03746201705932617, 0.03756032180786133, 0.037498878479003905, 0.03753779220581055, 0.03756748962402344, 0.037614654541015625, 0.037557182312011717, 0.03754291152954101, 0.03756032180786133, 0.037612545013427735, 0.03757875061035156, 0.03768524932861328, 0.03760844802856445, 0.03763507080078125, 0.03763097763061524, 0.037610496520996094, 0.037563392639160156, 0.0376545295715332, 0.03755110549926758, 0.03762995147705078, 0.03764121627807617, 0.03769036865234375, 0.037663745880126956, 0.03767091369628906, 0.0376360969543457, 0.0377077751159668, 0.03763507080078125, 0.037631999969482424, 0.03768729782104492, 0.03768320083618164, 0.03760435104370117, 0.03767609786987305, 0.03761452865600586, 0.03770982360839844, 0.03763206481933594, 0.03764012908935547, 0.03773747253417969, 0.03769343948364258, 0.03768217468261719, 0.03765862274169922, 0.03772108840942383, 0.03775385665893555, 0.037661697387695314, 0.03785318374633789, 0.037705726623535156, 0.037787647247314454, 0.03769343948364258, 0.03775078582763672, 0.03791155242919922, 0.03785318374633789, 0.03769753646850586, 0.03773235321044922, 0.0377784309387207, 0.03779072189331055, 0.03773952102661133, 0.037789695739746096, 0.03771289443969727, 0.03783679962158203, 0.037773311614990236, 0.037784576416015625, 0.1403187255859375, 0.037648384094238284, 0.037585918426513674, 0.03748044967651367, 0.03755929565429687, 0.03746406555175781, 0.03754905700683594, 0.03770470428466797, 0.037564449310302735, 0.03755414581298828, 0.03756851196289063, 0.037561344146728515, 0.03768217468261719, 0.037553150177001955, 0.03762688064575195, 0.0375551986694336, 0.03757567977905273, 0.03752959823608398, 0.03759513473510742, 0.037528575897216795, 0.03759513473510742, 0.037532737731933594, 0.03764729690551758, 0.0376545295715332, 0.03761459350585938, 0.03759206390380859, 0.03766579055786133, 0.037577728271484374, 0.03764223861694336, 0.037634048461914066, 0.03769036865234375, 0.03760639953613281, 0.03764025497436523, 0.03760838317871094, 0.03769036865234375, 0.037639232635498045, 0.037698497772216795, 0.037645313262939455, 0.03769343948364258, 0.03773747253417969, 0.03770880126953125, 0.03768012619018555, 0.03773952102661133, 0.037610496520996094, 0.037720062255859374, 0.03766681671142578, 0.037817344665527344, 0.037743614196777346, 0.03775897598266602, 0.03767398452758789, 0.037771263122558595, 0.03769452667236328, 0.03772927856445313, 0.03769747161865234, 0.038121536254882814, 0.03773843383789063, 0.037748737335205076, 0.03772927856445313, 0.037822463989257815, 0.03765760040283203, 0.037789695739746096, 0.03767705535888672, 0.03782860946655273, 0.14010780334472656, 0.037492702484130856, 0.03754086303710937, 0.037456897735595705, 0.03749683380126953, 0.03748863983154297, 0.03758803176879883, 0.0377138557434082, 0.03757567977905273, 0.03749785614013672, 0.03758694458007812, 0.0375203857421875, 0.03761151885986328, 0.03754598236083984, 0.0375551986694336, 0.037670944213867186, 0.03758281707763672, 0.037466110229492186, 0.03760537719726562, 0.03752959823608398, 0.037765121459960936, 0.038100990295410156, 0.03772825622558594, 0.03760639953613281, 0.03764019012451172, 0.03767603302001953, 0.03773952102661133, 0.037694465637207034, 0.0377149429321289, 0.03757056045532227, 0.03771187210083008, 0.03764019012451172, 0.03767091369628906, 0.037582847595214845, 0.03769651031494141, 0.037610496520996094, 0.03769139099121094, 0.037749759674072264, 0.03801599884033203, 0.03783679962158203, 0.03787673568725586, 0.037751808166503906, 0.03780915069580078, 0.03771289443969727, 0.03777740859985351, 0.037639198303222654, 0.03775484848022461, 0.03770473480224609, 0.03778355026245117, 0.03766268920898438, 0.03778252792358398, 0.03770470428466797, 0.03780198287963867, 0.037852161407470705, 0.037751808166503906, 0.037735424041748046, 0.03782860946655273, 0.03772723388671875, 0.0378081283569336, 0.03774259185791016, 0.0378152961730957, 0.037731361389160153, 0.03784086227416992, 0.14026956176757813, 0.03751020812988281, 0.037644222259521486, 0.03782451248168945, 0.03781017684936523, 0.0376514892578125, 0.040468448638916014, 0.03775795364379883, 0.03759513473510742, 0.037544960021972655, 0.037579872131347655, 0.03750595092773437, 0.037634048461914066, 0.037574657440185545, 0.037594112396240234, 0.03759308624267578, 0.03759513473510742, 0.03757875061035156, 0.037596160888671876, 0.03754598236083984, 0.03765350341796875, 0.03757670211791992, 0.03763507080078125, 0.03760947036743164, 0.03764940643310547, 0.03763715362548828, 0.03764118576049805, 0.03759209442138672, 0.037691360473632814, 0.037670944213867186, 0.03768931198120117, 0.037645313262939455, 0.03767398452758789, 0.0376258544921875, 0.037689342498779296, 0.03764121627807617, 0.03769651031494141, 0.037667839050292966, 0.0376627197265625, 0.03771289443969727, 0.037733375549316404, 0.03769548797607422, 0.037703678131103514, 0.037663745880126956, 0.037736446380615234, 0.03767705535888672, 0.03774566268920899, 0.037730304718017575, 0.03775388717651367, 0.03769443130493164, 0.037771297454833985, 0.03775484848022461, 0.03778559875488281, 0.03772108840942383, 0.03774771118164062, 0.0377784309387207, 0.03778358459472656, 0.03788694381713867, 0.037806079864501956, 0.037706783294677734, 0.03785929489135742, 0.03772108840942383, 0.03784089660644531, 0.14026956176757813, 0.037569534301757815, 0.03760025787353516, 0.03746918487548828, 0.037526527404785154, 0.03751321411132812, 0.03753267288208008, 0.037533695220947266, 0.037550079345703126, 0.03749990463256836, 0.03758489608764649, 0.03752243041992188, 0.037553150177001955, 0.037531646728515625, 0.037563392639160156, 0.037722110748291016, 0.03790643310546875, 0.03756137466430664, 0.03763504028320312, 0.03752243041992188, 0.0376360969543457, 0.03751935958862305, 0.03760537719726562, 0.037634048461914066, 0.03763302230834961, 0.03759308624267578, 0.03767295837402344, 0.037579776763916016, 0.037765121459960936, 0.037582847595214845, 0.03768217468261719, 0.037645313262939455, 0.03769036865234375, 0.03757875061035156, 0.037645313262939455, 0.03762790298461914, 0.03769753646850586, 0.037631999969482424, 0.037838848114013675, 0.03771088027954102, 0.03775382232666016, 0.0376627197265625, 0.03774156951904297, 0.037684223175048825, 0.037773311614990236, 0.03765657424926758, 0.03774566268920899, 0.03770265579223633, 0.037754878997802735, 0.037703678131103514, 0.03773132705688476, 0.03770880126953125, 0.03779379272460937, 0.03784294509887695, 0.03805184173583984, 0.03776716613769531, 0.03781324768066406, 0.03769241714477539, 0.03783679962158203, 0.03772313690185547, 0.03783168029785156, 0.037749759674072264, 0.03782656097412109, 0.14007501220703125, 0.037479423522949216, 0.037495807647705076, 0.037482494354248046, 0.0375623664855957, 0.03752345657348633, 0.037528575897216795, 0.037553150177001955, 0.03758694458007812, 0.037528575897216795, 0.03752140808105469, 0.03754086303710937, 0.03759718322753906, 0.03755929565429687, 0.037617664337158206, 0.03762790298461914, 0.03763916778564453, 0.03753676986694336, 0.03760748672485351, 0.03757868957519531, 0.03762790298461914, 0.037617664337158206, 0.03759206390380859, 0.03768217468261719, 0.03763814544677734, 0.037651454925537106, 0.03765862274169922, 0.03760435104370117, 0.03773747253417969, 0.037623809814453124, 0.03769139099121094, 0.03765964889526367, 0.0376627197265625, 0.03766681671142578, 0.037645313262939455, 0.037647361755371096, 0.03772415924072266, 0.037645313262939455, 0.03768115234375, 0.03770163345336914, 0.037716991424560545, 0.03767910385131836, 0.03774566268920899, 0.03769343948364258, 0.03773132705688476, 0.037718017578125, 0.03771596908569336, 0.03774979019165039, 0.0377476806640625, 0.03773132705688476, 0.03773545455932617, 0.03771184158325196, 0.037784576416015625, 0.0377077751159668, 0.037748737335205076, 0.037717025756835935, 0.03779580688476562, 0.03773235321044922, 0.037787647247314454, 0.037719039916992186, 0.03779993438720703, 0.03776204681396484, 0.03779072189331055, 0.1401692199707031, 0.03748147201538086, 0.03748454284667969, 0.03749478530883789, 0.03753267288208008, 0.03746303939819336, 0.03753574371337891, 0.037533695220947266, 0.03750502395629883, 0.03750400161743164, 0.03753676986694336, 0.03753267288208008, 0.037541889190673826, 0.03750400161743164, 0.037563392639160156, 0.03753779220581055, 0.037759998321533206, 0.037498878479003905, 0.03756748962402344, 0.03752447891235351, 0.037585918426513674, 0.03750707244873047, 0.0377077751159668, 0.037746688842773435, 0.0378081283569336, 0.03756851196289063, 0.037664768218994144, 0.037563392639160156, 0.03767500686645508, 0.03760332870483398, 0.037667873382568356, 0.03766883087158203, 0.037664768218994144, 0.03758694458007812, 0.03769651031494141, 0.03760947036743164, 0.03769343948364258, 0.037631999969482424, 0.03772927856445313, 0.037820449829101564, 0.038062049865722654, 0.03822387313842773, 0.037868545532226565, 0.03786649703979492, 0.03790848159790039, 0.037722110748291016, 0.03774156951904297, 0.03774054336547852, 0.037768192291259765, 0.03770880126953125, 0.037735424041748046, 0.037664768218994144, 0.03779891204833984, 0.037678081512451174, 0.03775692749023438, 0.03786137771606445, 0.0378869743347168, 0.03781119918823242, 0.03782656097412109, 0.03774771118164062, 0.037803009033203126, 0.037748737335205076, 0.037781505584716796, 0.14017945861816405, 0.0375654411315918, 0.03756748962402344, 0.03750912094116211, 0.03756032180786133, 0.037618686676025394, 0.03790233612060547, 0.037596160888671876, 0.03760332870483398, 0.03765964889526367, 0.03783782577514649, 0.03760947036743164, 0.037615615844726565, 0.03753676986694336, 0.03757670211791992, 0.0376360969543457, 0.03759718322753906, 0.03777536010742188, 0.03771599960327148, 0.03756745529174805, 0.03767910385131836, 0.03758796691894531, 0.03759929656982422, 0.03766470336914062, 0.0376258544921875, 0.03762790298461914, 0.03765043258666992, 0.037602302551269534, 0.037725215911865236, 0.037620704650878904, 0.03767910385131836, 0.03770675277709961, 0.03775897598266602, 0.03765657424926758, 0.037668895721435544, 0.037663711547851565, 0.03768832015991211, 0.03767507171630859, 0.037687232971191406, 0.037694465637207034, 0.0377149429321289, 0.03763203048706055, 0.037721057891845704, 0.037678081512451174, 0.03777536010742188, 0.03770163345336914, 0.0377446403503418, 0.03771289443969727, 0.03772528076171875, 0.037711776733398435, 0.03773132705688476, 0.03770675277709961, 0.03779180908203125, 0.037733375549316404, 0.037736385345458985, 0.03779379272460937, 0.03784294509887695, 0.03772415924072266, 0.03779174423217774, 0.03771084976196289, 0.037819393157958986, 0.037730304718017575, 0.03779891204833984, 0.14014566040039061, 0.037498878479003905, 0.03746815872192383, 0.03746310424804687, 0.0375428466796875, 0.03753676986694336, 0.037634048461914066, 0.037684223175048825, 0.03755417633056641, 0.03752140808105469, 0.03756032180786133, 0.03753267288208008, 0.037544960021972655, 0.037539840698242184, 0.03755212783813477, 0.03754905700683594, 0.03763814544677734, 0.03749990463256836, 0.03762790298461914, 0.03752243041992188, 0.037618686676025394, 0.03753676986694336, 0.03765248107910156, 0.037678081512451174, 0.03766886520385742, 0.037580799102783204, 0.03766579055786133, 0.03757875061035156, 0.037719039916992186, 0.03756851196289063, 0.03767295837402344, 0.0377077751159668, 0.03765248107910156, 0.03765248107910156, 0.03765248107910156, 0.03757875061035156, 0.03770982360839844, 0.037634048461914066, 0.0377149429321289, 0.03766681671142578, 0.037719039916992186, 0.03769651031494141, 0.0377446403503418, 0.03769139099121094, 0.03775078582763672, 0.0377149429321289, 0.0377446403503418, 0.037733375549316404, 0.03775283050537109, 0.03765043258666992, 0.03779174423217774, 0.037689342498779296, 0.03779174423217774, 0.03768627166748047, 0.03778867340087891, 0.037733375549316404, 0.037817344665527344, 0.03772825622558594, 0.037781505584716796, 0.037763072967529294, 0.037800960540771485, 0.037759998321533206, 0.03781836700439453]",tokens/s,25.53823297935188,,,main,False,False bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1272.496128,15196.48768,0.0,14550.040576,13944.957952,s,10,0.8173146514892577,0.08173146514892578,0.0015967359572456455,0.08116712188720704,0.08230683975219726,0.08438847770690917,0.0860537880706787,"[0.08647011566162109, 0.08184425354003906, 0.08111856079101562, 0.08103363037109375, 0.08127327728271484, 0.08121568298339844, 0.08125302124023437, 0.08110214233398437, 0.08101407623291015, 0.08098989105224609]",tokens/s,3132.2086241025218,kWh,9.611658960607748e-07,5.266747113459931e-07,5.150008636715389e-06,6.637849244122157e-06,tokens/kWh,38566708.97229084,MB,1272.496128,15196.48768,0.0,14550.040576,13945.037824,s,10,28.97916967773438,2.8979169677734378,0.00738291206369429,2.895778686523437,2.909371606445313,2.9102209594726562,2.9109004418945315,"[2.909182861328125, 2.896923828125, 2.8941103515625, 2.88953759765625, 2.894633544921875, 2.9110703125, 2.88955517578125, 2.901176025390625, 2.89083056640625, 2.9021494140625]",tokens/s,21.73975331267166,kWh,3.413491141034228e-05,1.870739413134244e-05,0.00012447001172588484,0.00017731231726756956,tokens/kWh,355305.2657076898,,s,629,29.66327807617189,0.047159424604406794,0.009708392241723048,0.04591308975219727,0.04647259597778321,0.04670177307128906,0.1273804360961914,"[0.04682137680053711, 0.04638412857055664, 0.04673843383789063, 0.04697292709350586, 0.046768192291259766, 0.046653377532958985, 0.04726067352294922, 0.046543872833251954, 0.04674252700805664, 0.047372287750244144, 0.04690022277832031, 0.04581171035766601, 0.04586598587036133, 0.04589363098144531, 0.04590591812133789, 0.04575027084350586, 0.045911041259765625, 0.04595199966430664, 0.04596121597290039, 0.04587212753295898, 0.045949951171875, 0.04596633529663086, 0.04597555160522461, 0.04585062408447266, 0.047710208892822265, 0.046963710784912106, 0.046769153594970705, 0.04592435073852539, 0.04597248077392578, 0.045843456268310545, 0.04584447860717773, 0.04589670562744141, 0.04594483184814453, 0.04594585418701172, 0.04620800018310547, 0.04636467361450195, 0.04595404815673828, 0.04583321762084961, 0.04589567947387695, 0.045927425384521485, 0.04598169708251953, 0.04584550476074219, 0.04599091339111328, 0.04595302581787109, 0.04592127990722656, 0.04583116912841797, 0.04643532943725586, 0.04587212753295898, 0.04591513442993164, 0.04578303909301758, 0.04583833694458008, 0.045794303894042966, 0.04605644989013672, 0.045843456268310545, 0.04585574340820313, 0.04589363098144531, 0.04593561553955078, 0.045830142974853515, 0.04591513442993164, 0.04581785583496094, 0.04637081527709961, 0.04662374496459961, 0.12934144592285157, 0.04629401779174805, 0.04579942321777344, 0.0459315185546875, 0.04580454254150391, 0.0459683837890625, 0.04575129699707031, 0.04602272033691406, 0.04591609573364258, 0.04598169708251953, 0.04579225540161133, 0.04597350311279297, 0.04591308975219727, 0.04593561553955078, 0.04588236618041992, 0.04597862243652344, 0.04598476791381836, 0.04589158248901367, 0.04582195281982422, 0.04597145462036133, 0.045848575592041016, 0.04598681640625, 0.045758464813232425, 0.045965312957763675, 0.04584447860717773, 0.04596223831176758, 0.045911041259765625, 0.045889537811279295, 0.045917182922363284, 0.04585779190063476, 0.04589875030517578, 0.045861888885498046, 0.04595507049560547, 0.045946880340576174, 0.04588339233398438, 0.04593766403198242, 0.0459233283996582, 0.04605644989013672, 0.04585574340820313, 0.045948928833007815, 0.045832191467285156, 0.046091262817382815, 0.045825023651123044, 0.04604620742797851, 0.04591923141479492, 0.045894657135009766, 0.04587519836425781, 0.04575948715209961, 0.04576870346069336, 0.04592844772338867, 0.04591308975219727, 0.045897727966308595, 0.045810688018798826, 0.04724838256835937, 0.04599603271484375, 0.045889537811279295, 0.04587116622924805, 0.04591609573364258, 0.04579020690917969, 0.046358528137207033, 0.04714700698852539, 0.04601241683959961, 0.04584755325317383, 0.12747264099121094, 0.04591513442993164, 0.045795326232910154, 0.04611379241943359, 0.04577280044555664, 0.04605132675170898, 0.04592639923095703, 0.045916160583496096, 0.04581785583496094, 0.04595814514160156, 0.04578303909301758, 0.04593664169311523, 0.045846527099609374, 0.04594790267944336, 0.04589158248901367, 0.045916160583496096, 0.045871105194091794, 0.04597350311279297, 0.045878273010253906, 0.0459950065612793, 0.045886463165283206, 0.04598988723754883, 0.04585574340820313, 0.04583119964599609, 0.045899742126464846, 0.04592230224609375, 0.045884414672851564, 0.04597043228149414, 0.04597145462036133, 0.046061569213867185, 0.0466319351196289, 0.04672512054443359, 0.04631654357910156, 0.0459233283996582, 0.04589977645874024, 0.045908992767333984, 0.045871105194091794, 0.0460052490234375, 0.045827072143554685, 0.04597452926635742, 0.04578508758544922, 0.04595814514160156, 0.045819969177246095, 0.045950912475585935, 0.04593971252441406, 0.04642611312866211, 0.046355457305908204, 0.04606771087646484, 0.045538303375244144, 0.04585779190063476, 0.04582809448242187, 0.04591513442993164, 0.04585062408447266, 0.04589056015014648, 0.04574310302734375, 0.04592236709594726, 0.045746112823486326, 0.04590591812133789, 0.04563353729248047, 0.045574176788330076, 0.04564169692993164, 0.045965312957763675, 0.045791233062744144, 0.1273671646118164, 0.04584755325317383, 0.04608102416992187, 0.04597555160522461, 0.045797374725341795, 0.045911041259765625, 0.04575743865966797, 0.045886463165283206, 0.04590387344360351, 0.04605440139770508, 0.045835262298583986, 0.045960193634033204, 0.04588236618041992, 0.04586905670166016, 0.04580044937133789, 0.045889537811279295, 0.04591001510620117, 0.04598067092895508, 0.04579840087890625, 0.04594073486328125, 0.045917182922363284, 0.04589158248901367, 0.045840385437011716, 0.04596223831176758, 0.04595199966430664, 0.04588748931884765, 0.045808639526367184, 0.045927425384521485, 0.0459048957824707, 0.045927425384521485, 0.045785152435302734, 0.045915073394775394, 0.045825023651123044, 0.04597350311279297, 0.0459048957824707, 0.04599705505371094, 0.04586393737792969, 0.04597248077392578, 0.04584243011474609, 0.045861888885498046, 0.04584550476074219, 0.045723648071289064, 0.04561417770385742, 0.045661087036132815, 0.045709312438964846, 0.04571852874755859, 0.04557926559448242, 0.04562944030761719, 0.04623769760131836, 0.04664831924438476, 0.04611388778686523, 0.04563753509521484, 0.045505535125732424, 0.04557619094848633, 0.04550656127929688, 0.04557209777832031, 0.045682689666748044, 0.04585472106933594, 0.04581990432739258, 0.04599808120727539, 0.045824001312255856, 0.04594278335571289, 0.045818878173828126, 0.12761497497558594, 0.04581171035766601, 0.04579635238647461, 0.045897727966308595, 0.04575027084350586, 0.04585881423950195, 0.04575129699707031, 0.04596736145019531, 0.04592947387695313, 0.04590182495117188, 0.04582195281982422, 0.045884414672851564, 0.045908992767333984, 0.04602265548706055, 0.045871105194091794, 0.045985790252685545, 0.045824001312255856, 0.04594585418701172, 0.04580556869506836, 0.04594585418701172, 0.04584755325317383, 0.04591206359863281, 0.04584447860717773, 0.046666751861572264, 0.046594112396240235, 0.04621615982055664, 0.046870494842529295, 0.04604006576538086, 0.045861888885498046, 0.04594278335571289, 0.045965312957763675, 0.04604620742797851, 0.04583935928344727, 0.045948928833007815, 0.045846527099609374, 0.04588544082641602, 0.04583935928344727, 0.046033920288085936, 0.04595609664916992, 0.04598169708251953, 0.04588339233398438, 0.04596121597290039, 0.04583321762084961, 0.045843456268310545, 0.04587519836425781, 0.04594585418701172, 0.04591513442993164, 0.04589977645874024, 0.04580044937133789, 0.04595814514160156, 0.04589363098144531, 0.045987838745117186, 0.045840385437011716, 0.045927425384521485, 0.04579635238647461, 0.04605644989013672, 0.04578303909301758, 0.04599193572998047, 0.04579945755004883, 0.045904865264892576, 0.04574521636962891, 0.04589459228515625, 0.04589363098144531, 0.12774092864990233, 0.04582809448242187, 0.045840385437011716, 0.04590796661376953, 0.04581990432739258, 0.04596121597290039, 0.04597248077392578, 0.045932544708251956, 0.045778976440429685, 0.04593967819213867, 0.0458158073425293, 0.04615884780883789, 0.046483455657958986, 0.04620800018310547, 0.0459048957824707, 0.04564377593994141, 0.04579328155517578, 0.04597350311279297, 0.04696780776977539, 0.0468592643737793, 0.046534656524658206, 0.04662988662719727, 0.04646092987060547, 0.04659609603881836, 0.04646604919433594, 0.046430206298828124, 0.0462110710144043, 0.04655411148071289, 0.04655411148071289, 0.04659404754638672, 0.045871105194091794, 0.04631039810180664, 0.04579635238647461, 0.04592639923095703, 0.04636979293823242, 0.04615577697753906, 0.04600012969970703, 0.04610153579711914, 0.046289886474609375, 0.04588339233398438, 0.045864959716796876, 0.04593664169311523, 0.046513153076171876, 0.04652134323120117, 0.046344192504882815, 0.046534656524658206, 0.04644454574584961, 0.04661862564086914, 0.046568447113037106, 0.046195743560791015, 0.04608918380737305, 0.04642201614379883, 0.046486526489257815, 0.0470384635925293, 0.04657561492919922, 0.04647116851806641, 0.04620083236694336, 0.04595609664916992, 0.045764606475830076, 0.04592639923095703, 0.04581785583496094, 0.04600012969970703, 0.045830142974853515, 0.12767129516601564, 0.04593971252441406, 0.045791233062744144, 0.04592230224609375, 0.04584447860717773, 0.045902847290039066, 0.04581273651123047, 0.04571033477783203, 0.04556288146972656, 0.0457237434387207, 0.04601551818847656, 0.04579110336303711, 0.0455813102722168, 0.045677566528320314, 0.04554035186767578, 0.0455731201171875, 0.04547788619995117, 0.045682815551757815, 0.04567027282714844, 0.0458158073425293, 0.04585472106933594, 0.04588339233398438, 0.045813758850097655, 0.04600012969970703, 0.045879295349121094, 0.045960193634033204, 0.04583731079101563, 0.04586700820922852, 0.04589977645874024, 0.04593971252441406, 0.04610969543457031, 0.04624076843261719, 0.045884414672851564, 0.04592639923095703, 0.045927425384521485, 0.045868030548095705, 0.045813758850097655, 0.04591308975219727, 0.04575539016723633, 0.04598886489868164, 0.04585062408447266, 0.045982719421386715, 0.045936672210693356, 0.04592227172851562, 0.045859840393066405, 0.04584755325317383, 0.045917182922363284, 0.04605542373657227, 0.04585881423950195, 0.04590694427490234, 0.04576972961425781, 0.04583833694458008, 0.0457523193359375, 0.04590182495117188, 0.04589875030517578, 0.04591820907592774, 0.04576665496826172, 0.04642816162109375, 0.045965312957763675, 0.045843456268310545, 0.045764606475830076, 0.0459048957824707, 0.045897727966308595, 0.12745216369628906, 0.04623360061645508, 0.04640671920776367, 0.04590380859375, 0.046063617706298826, 0.04595609664916992, 0.04629094314575195, 0.04687257766723633, 0.04696579360961914, 0.04659199905395508, 0.04647830581665039, 0.046445568084716796, 0.046437374114990236, 0.04647116851806641, 0.04650393676757812, 0.046496768951416016, 0.04641996765136719, 0.04637388610839844, 0.04581273651123047, 0.04592127990722656, 0.04594790267944336, 0.046042110443115236, 0.045868030548095705, 0.04598681640625, 0.045871135711669925, 0.04592022323608398, 0.04592127990722656, 0.04595507049560547, 0.04597353744506836, 0.04590895843505859, 0.04588544082641602, 0.04590182495117188, 0.04579840087890625, 0.045946880340576174, 0.04591206359863281, 0.046034942626953124, 0.04595814514160156, 0.04599808120727539, 0.045868030548095705, 0.04595404815673828, 0.04590387344360351, 0.04602675247192383, 0.045856769561767576, 0.04585369491577149, 0.045553665161132816, 0.04580556869506836, 0.04577382278442383, 0.04682963180541992, 0.04621408081054688, 0.04593971252441406, 0.04582092666625977, 0.04603289413452148, 0.045830142974853515, 0.045889537811279295, 0.045655040740966796, 0.045636768341064456, 0.045765472412109376, 0.04582092666625977, 0.0458158073425293, 0.04582915115356445, 0.04575948715209961, 0.045902816772460935, 0.046017536163330076, 0.12738559722900392, 0.045841407775878903, 0.04583321762084961, 0.045876224517822264, 0.04573081588745117, 0.04592537689208984, 0.04580147171020508, 0.045876224517822264, 0.045770751953125, 0.045879295349121094, 0.045884414672851564, 0.045930496215820314, 0.04581683349609375, 0.04592435073852539, 0.04575129699707031, 0.045884414672851564, 0.04586700820922852, 0.045949951171875, 0.04582195281982422, 0.045832191467285156, 0.045813758850097655, 0.04602880096435547, 0.04581273651123047, 0.04590591812133789, 0.04575743865966797, 0.04590182495117188, 0.04595404815673828, 0.04596633529663086, 0.0458260498046875, 0.04591513442993164, 0.045780990600585936, 0.04580556869506836, 0.0458260498046875, 0.04590591812133789, 0.04604313659667969, 0.04648038482666016, 0.04625612640380859, 0.0459683837890625, 0.04553731155395508, 0.04566422271728516, 0.04549836730957031, 0.046521438598632815, 0.0463553581237793, 0.04629401779174805, 0.045916160583496096, 0.04616294479370117, 0.045608959197998046, 0.045620223999023435, 0.04558233642578125, 0.045777919769287106, 0.04582809448242187, 0.045873153686523435, 0.04581785583496094, 0.0459048957824707, 0.04575539016723633, 0.04593664169311523, 0.04578611373901367, 0.04582809448242187, 0.04576972961425781, 0.045916160583496096, 0.04582092666625977, 0.04592947387695313, 0.04565404891967773, 0.12725142669677733, 0.04589363098144531, 0.04581683349609375, 0.04588851165771484, 0.045690879821777344, 0.04591001510620117, 0.04573081588745117, 0.04592639923095703, 0.046202880859375, 0.04597248077392578, 0.04585369491577149, 0.04687155151367187, 0.047094783782958984, 0.046371841430664064, 0.045861888885498046, 0.04590796661376953, 0.046317569732666014, 0.046048255920410154, 0.04583935928344727, 0.04591923141479492, 0.046502944946289065, 0.045970401763916015, 0.04623974227905273, 0.04628070449829102, 0.04585062408447266, 0.0459683837890625, 0.04579942321777344, 0.04652032089233398, 0.046025726318359376, 0.04589977645874024, 0.04600831985473633, 0.046266368865966793, 0.0459233283996582, 0.046018558502197264, 0.04636671829223633, 0.04596121597290039, 0.0458526725769043, 0.046053375244140625, 0.04625408172607422, 0.045827072143554685, 0.04593459320068359, 0.046311424255371096, 0.04611276626586914, 0.045871105194091794, 0.04621721649169922, 0.04650905609130859, 0.04627046585083008, 0.04590694427490234, 0.04616908645629883, 0.04654079818725586, 0.04582195281982422, 0.045835262298583986, 0.046162975311279296, 0.04636361694335937, 0.0463636474609375, 0.04615884780883789, 0.046102527618408204, 0.045548545837402345, 0.04548710250854492, 0.04593068695068359, 0.04588729476928711, 0.046152862548828125, 0.04556579208374024]",tokens/s,21.204669233953194,,,,, -bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4711,7 +4711,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1277.681664,1711.80032,0.0,1065.353216,1018.226688,s,10,0.19355452919006347,0.01935545291900635,0.0005977656089848024,0.019076640129089356,0.019849494552612305,0.020449354934692382,0.020929243240356447,"[0.021049215316772462, 0.019716192245483398, 0.019075616836547852, 0.019038528442382813, 0.019128671646118166, 0.01907747268676758, 0.019044479370117186, 0.019075807571411134, 0.019059680938720704, 0.019288864135742188]",tokens/s,13226.246942979946,kWh,2.2911734102513646e-07,1.2554492998809343e-07,6.189835916550509e-07,9.736458626682808e-07,tokens/kWh,262929274.20081758,MB,1277.681664,1711.80032,0.0,1065.353216,1058.88,s,10,11.631632934570314,1.1631632934570313,0.005588300337279548,1.1620383911132812,1.1683673583984375,1.1721897216796875,1.1752476123046876,"[1.1760120849609375, 1.16357421875, 1.1675179443359376, 1.167354736328125, 1.163515869140625, 1.1561873779296874, 1.1590982666015626, 1.1599454345703124, 1.1605609130859376, 1.1578660888671874]",tokens/s,54.16264453528106,kWh,1.3755305475145452e-05,7.537566972796201e-06,2.363948025954468e-05,4.4932352707486335e-05,tokens/kWh,1402107.750959218,,s,629,11.783421957015998,0.018733580217831468,0.0023476004409559346,0.018391040802001952,0.018720768737792968,0.018957926177978517,0.03788746841430664,"[0.020666368484497072, 0.019337215423583985, 0.018974720001220705, 0.01846681594848633, 0.018428928375244142, 0.018456575393676757, 0.018391040802001952, 0.018716672897338867, 0.0189040641784668, 0.018515968322753908, 0.01840332794189453, 0.01841459274291992, 0.01838387107849121, 0.01842995262145996, 0.01841155242919922, 0.01869513511657715, 0.018958335876464845, 0.018621471405029295, 0.018400224685668945, 0.01839923286437988, 0.01845248031616211, 0.018415615081787108, 0.01840025520324707, 0.01846886444091797, 0.019115007400512696, 0.01882931137084961, 0.018440191268920898, 0.01841459274291992, 0.018364416122436524, 0.018894847869873048, 0.019141632080078123, 0.019273727416992188, 0.019169279098510742, 0.018382848739624022, 0.01841459274291992, 0.01836953544616699, 0.018516992568969725, 0.018944000244140623, 0.018655231475830078, 0.01836031913757324, 0.018921472549438476, 0.018924543380737305, 0.018390016555786134, 0.01841049575805664, 0.018343936920166014, 0.018396160125732423, 0.018471935272216796, 0.01842995262145996, 0.018815999984741212, 0.018938880920410156, 0.018832384109497072, 0.01849760055541992, 0.018367424011230468, 0.01840127944946289, 0.01840332794189453, 0.01904332733154297, 0.018621503829956056, 0.018237375259399415, 0.018311168670654295, 0.018421760559082033, 0.01846784019470215, 0.018340864181518556, 0.03871744155883789, 0.01836031913757324, 0.018267135620117187, 0.018919424057006837, 0.01848320007324219, 0.0184268798828125, 0.018367488861083983, 0.018347007751464844, 0.018320383071899413, 0.018354175567626953, 0.01889795112609863, 0.01828758430480957, 0.018380800247192384, 0.018378751754760742, 0.018334720611572267, 0.018375680923461913, 0.018345983505249023, 0.018317312240600587, 0.018667520523071288, 0.01882931137084961, 0.01860403251647949, 0.018379776000976563, 0.01835212707519531, 0.01866444778442383, 0.01840127944946289, 0.018357248306274415, 0.01842790412902832, 0.018942975997924806, 0.018743295669555664, 0.018339839935302735, 0.018367488861083983, 0.018413568496704103, 0.01847500801086426, 0.018348031997680665, 0.018350080490112306, 0.018852863311767578, 0.018791423797607423, 0.018315263748168945, 0.018386943817138672, 0.018404352188110353, 0.018449407577514648, 0.01846067237854004, 0.018361343383789062, 0.018751487731933594, 0.018917375564575196, 0.018548736572265623, 0.018711551666259766, 0.018421760559082033, 0.018359296798706053, 0.018330623626708984, 0.018354175567626953, 0.018386943817138672, 0.01836953544616699, 0.01836755180358887, 0.018406335830688476, 0.018372608184814454, 0.018322431564331054, 0.018304000854492186, 0.018320383071899413, 0.018397184371948243, 0.018408447265625, 0.018328575134277342, 0.018378751754760742, 0.03801702499389648, 0.01879449653625488, 0.020534271240234374, 0.018957311630249024, 0.018502656936645507, 0.018412544250488282, 0.01841663932800293, 0.018371583938598633, 0.018337791442871093, 0.018380800247192384, 0.018341888427734376, 0.01842278480529785, 0.018380800247192384, 0.018437120437622072, 0.01843609619140625, 0.018374719619750977, 0.018373567581176756, 0.018325504302978517, 0.018397184371948243, 0.018341888427734376, 0.018367488861083983, 0.018355199813842774, 0.018437120437622072, 0.018358272552490236, 0.01840640068054199, 0.018378751754760742, 0.018362367630004883, 0.018336767196655272, 0.018359296798706053, 0.018316287994384766, 0.018377792358398436, 0.018415552139282226, 0.018379776000976563, 0.018356224060058594, 0.01838489532470703, 0.01842790412902832, 0.019042303085327148, 0.020187135696411132, 0.019148799896240236, 0.018917375564575196, 0.018967552185058592, 0.01882316780090332, 0.018400287628173827, 0.01870025634765625, 0.018549760818481444, 0.018413568496704103, 0.018461696624755858, 0.01839820861816406, 0.018404352188110353, 0.018565120697021483, 0.018318336486816408, 0.018363391876220703, 0.018405376434326173, 0.018313215255737304, 0.01839923286437988, 0.018370559692382812, 0.018371583938598633, 0.018365503311157227, 0.01871455955505371, 0.0183767032623291, 0.018340864181518556, 0.018355199813842774, 0.018359296798706053, 0.03789004898071289, 0.018387968063354493, 0.018348031997680665, 0.018518016815185546, 0.018372608184814454, 0.018311168670654295, 0.018344959259033202, 0.018324480056762696, 0.01839411163330078, 0.01840742492675781, 0.018326528549194337, 0.01840332794189453, 0.01841049575805664, 0.018396160125732423, 0.018306047439575195, 0.018396160125732423, 0.018379776000976563, 0.01841049575805664, 0.018387968063354493, 0.01836751937866211, 0.018369504928588867, 0.018465791702270508, 0.018501632690429686, 0.018471935272216796, 0.019573759078979493, 0.0190699520111084, 0.01881804847717285, 0.018549760818481444, 0.018374656677246092, 0.018405439376831055, 0.018399168014526367, 0.018448383331298827, 0.0184268798828125, 0.01840230369567871, 0.018626560211181642, 0.020023296356201172, 0.01943142318725586, 0.018778112411499022, 0.018354175567626953, 0.018780160903930664, 0.01843507194519043, 0.01843507194519043, 0.018449440002441406, 0.018387935638427735, 0.018370559692382812, 0.01843404769897461, 0.018685951232910156, 0.018421760559082033, 0.018737152099609376, 0.018447359085083007, 0.01847500801086426, 0.018397184371948243, 0.01843609619140625, 0.01879756736755371, 0.018545663833618165, 0.018463743209838866, 0.01842790412902832, 0.018508800506591795, 0.018516992568969725, 0.01837772750854492, 0.018557952880859374, 0.018397184371948243, 0.01838591957092285, 0.037897216796875, 0.01840947151184082, 0.018512895584106445, 0.01836031913757324, 0.018345983505249023, 0.018440191268920898, 0.018361343383789062, 0.018412544250488282, 0.018349056243896485, 0.018333696365356447, 0.01817190361022949, 0.018283519744873047, 0.018472959518432617, 0.01840332794189453, 0.018364416122436524, 0.018378751754760742, 0.018601984024047852, 0.018457599639892578, 0.01839411163330078, 0.018345983505249023, 0.018353151321411132, 0.018349056243896485, 0.018362367630004883, 0.01846886444091797, 0.01844223976135254, 0.01839820861816406, 0.018387968063354493, 0.018295808792114256, 0.019204095840454103, 0.018714624404907225, 0.01886412811279297, 0.018185216903686522, 0.018256895065307616, 0.01836953544616699, 0.018557952880859374, 0.01883545684814453, 0.018428928375244142, 0.018330623626708984, 0.0184586238861084, 0.018321407318115233, 0.018367488861083983, 0.01862246322631836, 0.018340864181518556, 0.01838489532470703, 0.018405376434326173, 0.018378751754760742, 0.018540544509887694, 0.01841971206665039, 0.018456575393676757, 0.018359296798706053, 0.018363391876220703, 0.018513952255249023, 0.01849648094177246, 0.018397184371948243, 0.018189311981201172, 0.018159616470336915, 0.01846272087097168, 0.01844326400756836, 0.018176000595092775, 0.018991104125976564, 0.01977446365356445, 0.019125247955322267, 0.018304000854492186, 0.038005760192871094, 0.01841868782043457, 0.018449472427368163, 0.01818822479248047, 0.01818009567260742, 0.018164735794067383, 0.018323455810546875, 0.01838591957092285, 0.018372608184814454, 0.01839923286437988, 0.018306047439575195, 0.018387968063354493, 0.018319360733032225, 0.01840127944946289, 0.01836751937866211, 0.01832547187805176, 0.018375680923461913, 0.018378751754760742, 0.018372608184814454, 0.018413568496704103, 0.018334720611572267, 0.01843612861633301, 0.018380767822265626, 0.01838489532470703, 0.018345983505249023, 0.018373632431030275, 0.018313215255737304, 0.0181790714263916, 0.018370559692382812, 0.018416704177856444, 0.018422719955444335, 0.018412544250488282, 0.018388992309570314, 0.018366464614868162, 0.018329599380493163, 0.01839308738708496, 0.018280448913574218, 0.01841152000427246, 0.018317312240600587, 0.018428928375244142, 0.018359296798706053, 0.018356224060058594, 0.018345983505249023, 0.018356224060058594, 0.018268159866333008, 0.018351104736328124, 0.01846886444091797, 0.01836031913757324, 0.018342912673950194, 0.018311168670654295, 0.018359296798706053, 0.018275327682495117, 0.018371583938598633, 0.018350080490112306, 0.01827840042114258, 0.01837772750854492, 0.018408447265625, 0.01844633674621582, 0.01818623924255371, 0.018137088775634767, 0.018155519485473632, 0.018361343383789062, 0.01836953544616699, 0.03784601593017578, 0.01839411163330078, 0.018324480056762696, 0.018343936920166014, 0.018461696624755858, 0.018326528549194337, 0.018395135879516602, 0.018334720611572267, 0.018340864181518556, 0.018316287994384766, 0.018320383071899413, 0.01902284812927246, 0.018523136138916017, 0.018367488861083983, 0.018404352188110353, 0.018282495498657226, 0.018347007751464844, 0.018365440368652345, 0.01833990478515625, 0.018400192260742188, 0.018323455810546875, 0.018327552795410155, 0.01835113525390625, 0.0184616641998291, 0.018373632431030275, 0.018349056243896485, 0.018358272552490236, 0.018322431564331054, 0.018481151580810547, 0.018315263748168945, 0.018453504562377928, 0.018345983505249023, 0.018354175567626953, 0.018355199813842774, 0.018362367630004883, 0.0182794246673584, 0.01845145606994629, 0.018647071838378906, 0.01848111915588379, 0.018379776000976563, 0.01842278480529785, 0.01835212707519531, 0.01829478454589844, 0.018513919830322266, 0.01844121551513672, 0.01836851119995117, 0.018469888687133788, 0.01840230369567871, 0.018457599639892578, 0.018456575393676757, 0.01837772750854492, 0.018397184371948243, 0.018397184371948243, 0.01836953544616699, 0.018307104110717773, 0.018311136245727538, 0.018408447265625, 0.018336767196655272, 0.01836031913757324, 0.018338815689086914, 0.018357248306274415, 0.018362367630004883, 0.018363391876220703, 0.03804569625854492, 0.018342912673950194, 0.01838591957092285, 0.018374656677246092, 0.01836031913757324, 0.018349056243896485, 0.018324480056762696, 0.01836953544616699, 0.01842790412902832, 0.018472959518432617, 0.018428928375244142, 0.01842790412902832, 0.018370559692382812, 0.018381824493408205, 0.018333696365356447, 0.01843404769897461, 0.018423807144165038, 0.01841766357421875, 0.01843507194519043, 0.018361343383789062, 0.018357248306274415, 0.018339839935302735, 0.01841868782043457, 0.01839923286437988, 0.018396160125732423, 0.018437183380126954, 0.018374591827392577, 0.018437120437622072, 0.018371583938598633, 0.01838595199584961, 0.018382816314697265, 0.018455551147460936, 0.01840742492675781, 0.01837772750854492, 0.018330623626708984, 0.01841152000427246, 0.018374656677246092, 0.01842995262145996, 0.018295839309692384, 0.018826208114624023, 0.01867366409301758, 0.018693119049072265, 0.018490367889404297, 0.018447359085083007, 0.018340864181518556, 0.01843814468383789, 0.01841766357421875, 0.018356224060058594, 0.01838489532470703, 0.01841868782043457, 0.01839308738708496, 0.018362367630004883, 0.018373632431030275, 0.018592767715454102, 0.01841152000427246, 0.018440191268920898, 0.0184770565032959, 0.01842585563659668, 0.01823846435546875, 0.018206720352172853, 0.01825279998779297, 0.01819545555114746, 0.01842585563659668, 0.03788083267211914, 0.01844633674621582, 0.01842995262145996, 0.018349056243896485, 0.018354175567626953, 0.01844326400756836, 0.018382848739624022, 0.018482175827026368, 0.018416671752929686, 0.018434015274047852, 0.018391040802001952, 0.018388992309570314, 0.01844633674621582, 0.01840947151184082, 0.018707456588745116, 0.018741247177124023, 0.018423807144165038, 0.01846886444091797, 0.01840742492675781, 0.01841049575805664, 0.018453504562377928, 0.01838387107849121, 0.0184770565032959, 0.01841152000427246, 0.01843609619140625, 0.018446367263793947, 0.018409439086914062, 0.01847603225708008, 0.018461696624755858, 0.018404352188110353, 0.01843916893005371, 0.018373632431030275, 0.01842995262145996, 0.018375680923461913, 0.018448383331298827, 0.018372608184814454, 0.01845043182373047, 0.01841663932800293, 0.01841663932800293, 0.018336767196655272, 0.01841766357421875, 0.018338815689086914, 0.018373632431030275, 0.018333696365356447, 0.01840742492675781, 0.01840025520324707, 0.01841766357421875, 0.018378751754760742, 0.01843404769897461, 0.018382848739624022, 0.018372608184814454, 0.018336767196655272, 0.01841766357421875, 0.018356224060058594, 0.01844326400756836, 0.01840332794189453, 0.018351104736328124, 0.018397184371948243, 0.018349056243896485, 0.018363391876220703, 0.018350080490112306, 0.018353151321411132, 0.018438207626342774, 0.038718399047851564, 0.018365440368652345, 0.018356224060058594, 0.018354175567626953, 0.018331680297851562, 0.01830601692199707, 0.018322431564331054, 0.018358272552490236, 0.018327552795410155, 0.018372608184814454, 0.018341888427734376, 0.018340864181518556, 0.01835212707519531, 0.018366464614868162, 0.018292736053466797, 0.018318336486816408, 0.018347007751464844, 0.018359296798706053, 0.018376768112182616, 0.018367424011230468, 0.018275327682495117, 0.018296831130981444, 0.018347007751464844, 0.018363391876220703, 0.018343936920166014, 0.018367488861083983, 0.018405376434326173, 0.018439231872558595, 0.01833568000793457, 0.018343936920166014, 0.018341888427734376, 0.01839411163330078, 0.018386943817138672, 0.018473983764648438, 0.01843507194519043, 0.01841766357421875, 0.018336767196655272, 0.018299903869628906, 0.018306047439575195, 0.01837772750854492, 0.0183767032623291, 0.01841459274291992, 0.01832863998413086, 0.01837664031982422, 0.018342912673950194, 0.018481151580810547, 0.01840233612060547, 0.018336736679077148, 0.01836031913757324, 0.018495487213134765, 0.018322431564331054, 0.018353151321411132, 0.018367488861083983, 0.018397184371948243, 0.018336767196655272, 0.0184268798828125, 0.018344959259033202, 0.018361343383789062, 0.018388992309570314, 0.018326559066772462, 0.018369504928588867, 0.018367488861083983, 0.018348031997680665]",tokens/s,53.38007942807189,,,main,False,False -bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4931,7 +4931,7 @@ bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.Py bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1388.036096,2403.86048,0.0,1757.413376,1301.782528,s,12,0.20144134140014652,0.016786778450012207,0.0018214050335381856,0.016153103828430174,0.016703455734252932,0.019434478092193598,0.02210406023025513,"[0.022771455764770507, 0.016551008224487306, 0.016201311111450196, 0.0167042236328125, 0.016696544647216798, 0.016167776107788086, 0.016049343109130858, 0.016016319274902345, 0.016138431549072265, 0.01601785659790039, 0.01609654426574707, 0.016030527114868163]",tokens/s,15250.097019050958,kWh,1.921774259858446e-07,1.0530401582644383e-07,6.138130475090691e-07,9.112944893213575e-07,tokens/kWh,280919069.5212517,MB,1388.036096,2403.86048,0.0,1757.413376,1324.7616,s,12,10.558549926757813,0.8798791605631511,0.010784227070228428,0.8786428222656251,0.8903626342773437,0.8984366760253907,0.9055257141113281,"[0.9072979736328125, 0.881484375, 0.8828522338867187, 0.8911865234375, 0.8829476318359375, 0.882841552734375, 0.8734462890625, 0.8682412719726562, 0.8711069946289063, 0.875488037109375, 0.8658557739257813, 0.87580126953125]",tokens/s,71.60074112867723,kWh,1.0095053256607961e-05,5.531636060229464e-06,2.1067384852988094e-05,3.669407416982551e-05,tokens/kWh,1716898.4754439325,,s,755,10.721403898239132,0.014200534964555147,0.0020255489759926147,0.013850624084472657,0.014336614418029785,0.014493081760406494,0.030158765869140625,"[0.01530777645111084, 0.01447935962677002, 0.015130656242370605, 0.014796768188476563, 0.0142807035446167, 0.01449062442779541, 0.014376959800720214, 0.014313471794128419, 0.014429183959960937, 0.014726143836975097, 0.014943231582641601, 0.015244288444519043, 0.014939135551452636, 0.014836735725402832, 0.015038463592529297, 0.015326208114624023, 0.014498815536499024, 0.014320639610290528, 0.01439948844909668, 0.014312447547912598, 0.014470144271850586, 0.01447219181060791, 0.014910464286804198, 0.014941184043884278, 0.014276608467102051, 0.01438822364807129, 0.014529536247253418, 0.014299136161804199, 0.01425715160369873, 0.01417728042602539, 0.014098431587219238, 0.014183423995971679, 0.01399507236480713, 0.013927359580993653, 0.014554112434387208, 0.014814208030700684, 0.014774271965026856, 0.014413824081420898, 0.014178303718566895, 0.014191679954528809, 0.014287808418273925, 0.014317567825317384, 0.014322688102722168, 0.014212096214294433, 0.014258175849914552, 0.014268416404724121, 0.01360588836669922, 0.013647871971130371, 0.013599743843078613, 0.013652992248535157, 0.013706303596496583, 0.013746111869812012, 0.013850624084472657, 0.013767680168151856, 0.015051775932312012, 0.014500864028930664, 0.014235679626464843, 0.01431446361541748, 0.014199808120727539, 0.014208000183105468, 0.013628416061401367, 0.01368166446685791, 0.031204351425170897, 0.014247936248779297, 0.013598719596862792, 0.013565952301025391, 0.013685759544372558, 0.013692928314208984, 0.013604928016662597, 0.013604800224304199, 0.013643775939941406, 0.014348287582397461, 0.014272512435913086, 0.013604864120483399, 0.014402591705322265, 0.014265312194824218, 0.014206975936889648, 0.014162943840026856, 0.014297087669372559, 0.014167039871215821, 0.01365503978729248, 0.013606911659240722, 0.014262271881103515, 0.014229503631591797, 0.014305279731750489, 0.014249983787536622, 0.014173184394836426, 0.01420083236694336, 0.014223360061645507, 0.014201855659484864, 0.014272512435913086, 0.014284799575805664, 0.014422016143798828, 0.01416806411743164, 0.014263296127319336, 0.0135731201171875, 0.01365503978729248, 0.013587455749511718, 0.013649920463562011, 0.013632512092590332, 0.01364684772491455, 0.013639679908752441, 0.013790207862854004, 0.014238719940185546, 0.01425715160369873, 0.014249983787536622, 0.014197759628295899, 0.014195712089538574, 0.01376972770690918, 0.013666303634643554, 0.013640704154968262, 0.013647871971130371, 0.013629440307617188, 0.014131199836730958, 0.014272512435913086, 0.014150655746459961, 0.01417728042602539, 0.014129152297973633, 0.013595647811889648, 0.013619199752807617, 0.013720576286315917, 0.014217215538024902, 0.01416806411743164, 0.014461952209472656, 0.013668352127075196, 0.030331903457641602, 0.013631487846374512, 0.013629440307617188, 0.013670399665832519, 0.013591551780700683, 0.014339072227478027, 0.014220288276672363, 0.014690303802490234, 0.015069184303283691, 0.014460927963256836, 0.014255104064941406, 0.014277631759643555, 0.01417625617980957, 0.014187520027160644, 0.013637696266174316, 0.013637568473815918, 0.014225407600402832, 0.014327839851379394, 0.01365500831604004, 0.013674495697021484, 0.013644800186157227, 0.013644800186157227, 0.013643775939941406, 0.014126079559326172, 0.01430835247039795, 0.01399295997619629, 0.014062591552734375, 0.013521920204162598, 0.013940735816955567, 0.013748224258422852, 0.013991968154907226, 0.01404412841796875, 0.014261247634887696, 0.014225407600402832, 0.014316543579101563, 0.014173215866088867, 0.014258144378662109, 0.014189567565917969, 0.014412799835205077, 0.014314496040344238, 0.014173184394836426, 0.01419878387451172, 0.014213120460510254, 0.014160896301269531, 0.01365401554107666, 0.013683712005615235, 0.013677568435668945, 0.013714431762695312, 0.013792256355285644, 0.014287872314453125, 0.014243840217590332, 0.01421827220916748, 0.01370518398284912, 0.01368883228302002, 0.013612031936645508, 0.013674495697021484, 0.013647871971130371, 0.013618176460266113, 0.013612031936645508, 0.014309375762939454, 0.014212096214294433, 0.014179327964782714, 0.014226431846618653, 0.0314204158782959, 0.014199808120727539, 0.014220288276672363, 0.014219264030456542, 0.014256128311157227, 0.01428889560699463, 0.014225407600402832, 0.01435750389099121, 0.014276608467102051, 0.014249983787536622, 0.014249983787536622, 0.014217215538024902, 0.014322688102722168, 0.014245887756347657, 0.014218239784240723, 0.01438310432434082, 0.01417420768737793, 0.013648896217346192, 0.014229503631591797, 0.014286848068237304, 0.014244864463806153, 0.013723648071289063, 0.01367142391204834, 0.014216192245483398, 0.014236672401428223, 0.014240768432617188, 0.014216192245483398, 0.013658111572265624, 0.013657088279724122, 0.013592576026916504, 0.013749247550964355, 0.013691904067993164, 0.01364684772491455, 0.013683712005615235, 0.013666303634643554, 0.013718591690063477, 0.0141659517288208, 0.014210047721862793, 0.014279680252075195, 0.014277631759643555, 0.014208000183105468, 0.014228480339050293, 0.01425920009613037, 0.014231552124023437, 0.014268416404724121, 0.014223360061645507, 0.014389247894287109, 0.014271488189697265, 0.014260224342346191, 0.014178303718566895, 0.014152704238891601, 0.014183423995971679, 0.01417625617980957, 0.014181376457214356, 0.014199808120727539, 0.014248959541320801, 0.014325759887695312, 0.014284799575805664, 0.014260224342346191, 0.014287872314453125, 0.014338047981262207, 0.014202879905700684, 0.014384127616882325, 0.031308799743652346, 0.014306303977966308, 0.01426534366607666, 0.013439999580383302, 0.013478912353515626, 0.013599743843078613, 0.013618176460266113, 0.013650943756103515, 0.013716480255126954, 0.013647871971130371, 0.013629440307617188, 0.014221311569213867, 0.01417420768737793, 0.01406156826019287, 0.014043168067932129, 0.013950943946838379, 0.014063648223876953, 0.013977567672729492, 0.013945856094360352, 0.01419878387451172, 0.013645824432373046, 0.013700096130371094, 0.013670399665832519, 0.014195712089538574, 0.014187520027160644, 0.014510080337524415, 0.014345215797424317, 0.014197759628295899, 0.01366220760345459, 0.013649920463562011, 0.014160896301269531, 0.014219264030456542, 0.014231552124023437, 0.014477312088012695, 0.013919232368469238, 0.014194687843322755, 0.014226431846618653, 0.014301183700561524, 0.014233599662780762, 0.014213120460510254, 0.014244864463806153, 0.014254079818725587, 0.014317567825317384, 0.014449664115905762, 0.014379008293151856, 0.014263296127319336, 0.0142807035446167, 0.014256128311157227, 0.013820927619934082, 0.0136878080368042, 0.014114815711975098, 0.01409945583343506, 0.013516799926757812, 0.013850624084472657, 0.014088191986083985, 0.014215167999267577, 0.01427558422088623, 0.014180352210998535, 0.013632512092590332, 0.013623295783996582, 0.013634559631347656, 0.013656064033508301, 0.013607935905456543, 0.030935039520263673, 0.014208000183105468, 0.014180352210998535, 0.014253055572509766, 0.01417420768737793, 0.013635583877563476, 0.013657088279724122, 0.015430656433105469, 0.015126527786254883, 0.014407679557800293, 0.01425920009613037, 0.014338047981262207, 0.014346240043640136, 0.014218239784240723, 0.01425920009613037, 0.014226495742797851, 0.014233535766601562, 0.013649920463562011, 0.013616127967834473, 0.01363865566253662, 0.014310400009155273, 0.014270463943481445, 0.014194687843322755, 0.014149632453918457, 0.014183423995971679, 0.013806591987609864, 0.013742079734802246, 0.013614080429077148, 0.013703167915344238, 0.013717503547668456, 0.014044159889221192, 0.014831616401672363, 0.014458880424499512, 0.014185471534729004, 0.014240768432617188, 0.013615103721618652, 0.013802495956420899, 0.014393343925476074, 0.01428377628326416, 0.013690879821777344, 0.013684736251831055, 0.013622271537780761, 0.013670399665832519, 0.01368172836303711, 0.013761471748352051, 0.013664256095886231, 0.013685759544372558, 0.013701120376586913, 0.01365401554107666, 0.01363865566253662, 0.013682687759399414, 0.013670399665832519, 0.01360588836669922, 0.013599743843078613, 0.013617183685302734, 0.013633503913879394, 0.013599743843078613, 0.013612031936645508, 0.014169088363647461, 0.014277631759643555, 0.014226431846618653, 0.014316543579101563, 0.014204928398132324, 0.03015782356262207, 0.01360588836669922, 0.013601792335510255, 0.013599743843078613, 0.013565952301025391, 0.013614080429077148, 0.013577280044555664, 0.013657024383544922, 0.01358233642578125, 0.013652992248535157, 0.01375641632080078, 0.013717503547668456, 0.01368064022064209, 0.013667327880859375, 0.013674495697021484, 0.013629440307617188, 0.013698047637939453, 0.013602815628051757, 0.014050304412841797, 0.01427558422088623, 0.014402560234069824, 0.014203904151916504, 0.014337023735046387, 0.014303232192993164, 0.014318592071533203, 0.014615551948547363, 0.014318592071533203, 0.014310400009155273, 0.014237728118896485, 0.014213088035583497, 0.013505536079406738, 0.013540351867675781, 0.013527039527893067, 0.013573151588439941, 0.013600735664367676, 0.013645824432373046, 0.013622271537780761, 0.013652992248535157, 0.013608960151672364, 0.014220319747924804, 0.014303199768066406, 0.014193663597106934, 0.014064640045166015, 0.013477888107299805, 0.013487104415893555, 0.013794303894042969, 0.013661184310913087, 0.013634559631347656, 0.01364684772491455, 0.013668352127075196, 0.013629440307617188, 0.01362435245513916, 0.014179295539855957, 0.014215167999267577, 0.014258175849914552, 0.014204928398132324, 0.014169088363647461, 0.014298111915588378, 0.013817855834960938, 0.01368883228302002, 0.013667327880859375, 0.013576191902160644, 0.01416806411743164, 0.030083072662353515, 0.013622271537780761, 0.014060544013977052, 0.014247936248779297, 0.014435327529907227, 0.014224384307861328, 0.014233599662780762, 0.014221343994140625, 0.014329824447631836, 0.013616127967834473, 0.01363865566253662, 0.013666303634643554, 0.013587455749511718, 0.013658111572265624, 0.013622271537780761, 0.013698047637939453, 0.01386291217803955, 0.013703167915344238, 0.01375334358215332, 0.01366220760345459, 0.013701120376586913, 0.01359769630432129, 0.013666303634643554, 0.01359984016418457, 0.013728672027587891, 0.013637632369995116, 0.013912128448486329, 0.014220224380493165, 0.013942784309387207, 0.014271488189697265, 0.014307328224182129, 0.014325759887695312, 0.014211071968078613, 0.01364684772491455, 0.013686783790588379, 0.013742079734802246, 0.013689855575561523, 0.013633536338806153, 0.013645824432373046, 0.013659135818481445, 0.013619199752807617, 0.013651968002319336, 0.013618176460266113, 0.013648896217346192, 0.013620223999023438, 0.013603839874267578, 0.013676544189453126, 0.01364684772491455, 0.013727744102478028, 0.013616127967834473, 0.013603839874267578, 0.01356390380859375, 0.013496319770812988, 0.013639712333679199, 0.013682656288146973, 0.013608991622924805, 0.01363145637512207, 0.013659135818481445, 0.013632512092590332, 0.013635583877563476, 0.013735936164855958, 0.013604864120483399, 0.013619199752807617, 0.030159872055053712, 0.013609984397888183, 0.013615103721618652, 0.013629440307617188, 0.013584383964538574, 0.013620223999023438, 0.013579263687133789, 0.013433856010437012, 0.013420543670654296, 0.013407232284545899, 0.0135731201171875, 0.013987839698791504, 0.01365503978729248, 0.013612031936645508, 0.013619263648986817, 0.013640640258789062, 0.013607935905456543, 0.01387724781036377, 0.01372879981994629, 0.01372873592376709, 0.014499839782714843, 0.014344191551208496, 0.014252032279968262, 0.01425312042236328, 0.014243776321411133, 0.014188544273376465, 0.014292991638183594, 0.014255104064941406, 0.013705216407775878, 0.013643775939941406, 0.013640704154968262, 0.013613056182861329, 0.013643839836120605, 0.013640640258789062, 0.013670399665832519, 0.01387007999420166, 0.01420803165435791, 0.014275551795959473, 0.014336000442504883, 0.01428889560699463, 0.01426534366607666, 0.014190655708312988, 0.014327743530273437, 0.01419878387451172, 0.013657088279724122, 0.013631487846374512, 0.013620223999023438, 0.013670399665832519, 0.013636608123779297, 0.013714431762695312, 0.01366323184967041, 0.013694975852966309, 0.013689855575561523, 0.013676544189453126, 0.013649951934814454, 0.013602784156799316, 0.01375436782836914, 0.013659135818481445, 0.01363865566253662, 0.013717503547668456, 0.014311488151550294, 0.01368057632446289, 0.013658111572265624, 0.030123008728027343, 0.01368166446685791, 0.013651968002319336, 0.013615103721618652, 0.013749247550964355, 0.013619199752807617, 0.013647871971130371, 0.013658111572265624, 0.013707263946533203, 0.013661184310913087, 0.01365401554107666, 0.013629440307617188, 0.013656064033508301, 0.013661184310913087, 0.013643775939941406, 0.013637632369995116, 0.013649920463562011, 0.01367347240447998, 0.013657088279724122, 0.01370419216156006, 0.013734911918640137, 0.01367244815826416, 0.01368064022064209, 0.013628416061401367, 0.013707263946533203, 0.01366323184967041, 0.01366528034210205, 0.013628416061401367, 0.014185471534729004, 0.014222335815429688, 0.014215167999267577, 0.014255104064941406, 0.01417420768737793, 0.014271488189697265, 0.013685759544372558, 0.013706239700317382, 0.013648896217346192, 0.013668352127075196, 0.013624320030212403, 0.013608960151672364, 0.013578240394592284, 0.013747200012207032, 0.01365503978729248, 0.013722623825073242, 0.013987839698791504, 0.014213120460510254, 0.01425100803375244, 0.014137344360351562, 0.014299136161804199, 0.014337023735046387, 0.01425920009613037, 0.014270463943481445, 0.014346240043640136, 0.014315520286560059, 0.014189567565917969, 0.014203935623168946, 0.014187487602233886, 0.014215167999267577, 0.01425920009613037, 0.014254079818725587, 0.014341119766235352, 0.01415167999267578, 0.01430016040802002, 0.030241792678833007, 0.013626367568969726, 0.01365401554107666, 0.013703167915344238, 0.014215167999267577, 0.014222335815429688, 0.01425715160369873, 0.01445683193206787, 0.013668352127075196, 0.013657088279724122, 0.013627391815185547, 0.01367244815826416, 0.013740032196044923, 0.013615103721618652, 0.013641728401184081, 0.013612031936645508, 0.013742079734802246, 0.013636608123779297, 0.013692928314208984, 0.014064640045166015, 0.01364684772491455, 0.013651968002319336, 0.013629440307617188, 0.01363865566253662, 0.013596672058105469, 0.01359769630432129, 0.013606911659240722, 0.01358028793334961, 0.01387110424041748, 0.014248959541320801, 0.014155776023864745, 0.013717503547668456, 0.013712384223937989, 0.013628416061401367, 0.01363046360015869, 0.013611007690429687, 0.013658176422119141, 0.013610943794250489, 0.01359769630432129, 0.013645824432373046, 0.013634559631347656, 0.013583359718322753, 0.013633536338806153, 0.013720576286315917, 0.013658111572265624, 0.013586432456970214, 0.013724672317504882, 0.01365401554107666, 0.013603839874267578, 0.013724672317504882, 0.013636608123779297, 0.01368064022064209, 0.013641728401184081, 0.01367244815826416, 0.013617152214050294, 0.013708288192749024, 0.01363046360015869, 0.013660160064697266, 0.01417625617980957, 0.014216192245483398, 0.013643775939941406, 0.013763584136962891, 0.013627391815185547, 0.030679040908813477, 0.013649920463562011, 0.013645824432373046, 0.013636608123779297, 0.013667360305786132, 0.013649888038635254, 0.013593600273132325, 0.013640704154968262, 0.013733887672424316, 0.014215167999267577, 0.014227456092834472, 0.014219264030456542, 0.013622271537780761, 0.013590527534484862, 0.013858816146850587, 0.01365503978729248, 0.013595647811889648, 0.01359769630432129, 0.013619199752807617, 0.013619199752807617, 0.013639679908752441, 0.01426636791229248, 0.014191616058349609, 0.014276608467102051, 0.014207039833068848, 0.01432262420654297, 0.014256128311157227, 0.014217215538024902, 0.014189567565917969, 0.014227456092834472, 0.014179327964782714, 0.014205951690673829, 0.014188544273376465, 0.014179327964782714, 0.013637632369995116, 0.013684736251831055, 0.013616127967834473, 0.013583359718322753, 0.013660160064697266, 0.013599743843078613, 0.01363046360015869, 0.013608960151672364, 0.013620223999023438, 0.013674495697021484, 0.014307328224182129, 0.014224448204040528, 0.014213055610656737, 0.014164992332458496, 0.014185471534729004, 0.014183423995971679, 0.01419878387451172, 0.014219264030456542, 0.014228480339050293, 0.01416806411743164, 0.014319616317749024, 0.013636608123779297, 0.013627391815185547, 0.013645888328552246, 0.013614015579223633, 0.013584383964538574, 0.013618176460266113, 0.013590527534484862, 0.013558783531188966]",tokens/s,70.41988224359311,,,main,False,False bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1174.433792,1455.947776,0.0,809.500672,723.533824,s,13,0.18676848125457762,0.014366806250352126,0.0005655927627597173,0.01455446434020996,0.014721689414978027,0.01518812141418457,0.01572413932800293,"[0.01585814380645752, 0.014642687797546386, 0.014741439819335938, 0.014609312057495117, 0.013956352233886719, 0.013896032333374024, 0.013839424133300781, 0.013809280395507812, 0.01460409641265869, 0.014568863868713379, 0.013815808296203613, 0.01455446434020996, 0.013872575759887696]",tokens/s,17818.852397604063,kWh,1.672100464858931e-07,9.160921237509845e-08,4.6943853365206314e-07,7.282577925130547e-07,tokens/kWh,351523873.32046425,MB,1174.433792,1455.947776,0.0,809.500672,749.21984,s,13,10.515557861328123,0.8088890662560095,0.01598588227034598,0.8103314819335937,0.8233277099609375,0.831214794921875,0.840103408203125,"[0.8423255615234375, 0.8213235473632813, 0.8238076171875, 0.8103518676757813, 0.782132568359375, 0.803363037109375, 0.7837216186523438, 0.7996719360351563, 0.8214080810546875, 0.81534326171875, 0.8048592529296875, 0.8103314819335937, 0.7969180297851562]",tokens/s,77.88459830666174,kWh,9.496202768999352e-06,5.203489430023765e-06,1.7370255225268167e-05,3.206994742429129e-05,tokens/kWh,1964455.980126766,,s,818,10.672301052093502,0.0130468228020703,0.0018048516405457077,0.01297974395751953,0.013116723251342773,0.013315583896636961,0.026699346485137943,"[0.014223360061645507, 0.014007295608520508, 0.015222784042358398, 0.014003199577331543, 0.013818880081176758, 0.015129599571228027, 0.014413824081420898, 0.014791680335998534, 0.013922304153442382, 0.01409126377105713, 0.014049280166625976, 0.013950976371765136, 0.01466476821899414, 0.01347372817993164, 0.013149184226989746, 0.013238271713256837, 0.013000703811645508, 0.013118464469909668, 0.013089792251586914, 0.013172736167907715, 0.013143039703369141, 0.01305292797088623, 0.013071359634399414, 0.01306214427947998, 0.013190143585205078, 0.013019136428833008, 0.013095935821533204, 0.013046784400939941, 0.01307033634185791, 0.013423616409301758, 0.01305292797088623, 0.013020159721374512, 0.01305395221710205, 0.013109248161315918, 0.013044735908508302, 0.013036543846130372, 0.013006848335266113, 0.013083647727966309, 0.01297715187072754, 0.012917759895324708, 0.012932095527648926, 0.013006848335266113, 0.013328384399414063, 0.013268992424011231, 0.013270015716552735, 0.013496319770812988, 0.013110272407531738, 0.01318297576904297, 0.013117440223693847, 0.01297926425933838, 0.012971967697143554, 0.013061152458190919, 0.01304368019104004, 0.013007871627807617, 0.012980223655700684, 0.013025279998779296, 0.012989439964294434, 0.01298534393310547, 0.01296895980834961, 0.013030400276184082, 0.013002752304077148, 0.01297715187072754, 0.028115968704223632, 0.013014016151428222, 0.013172736167907715, 0.013099007606506348, 0.013074432373046875, 0.013047807693481446, 0.013002752304077148, 0.013033472061157226, 0.012983296394348144, 0.013022208213806152, 0.01303756809234619, 0.012976127624511719, 0.012957695960998536, 0.012974080085754394, 0.012986368179321289, 0.012958720207214355, 0.012991488456726074, 0.012995648384094238, 0.013151167869567871, 0.013116415977478027, 0.013014080047607422, 0.013004735946655273, 0.012995583534240723, 0.013029376029968261, 0.013083647727966309, 0.013011967658996582, 0.013022208213806152, 0.012980223655700684, 0.012967935562133789, 0.013273088455200196, 0.01306214427947998, 0.012969984054565429, 0.01296281623840332, 0.012981247901916505, 0.01296895980834961, 0.013031423568725586, 0.012980223655700684, 0.013023232460021973, 0.013040639877319337, 0.012999679565429688, 0.01298739242553711, 0.013038592338562012, 0.012984319686889649, 0.013030400276184082, 0.013000703811645508, 0.012994560241699218, 0.01303756809234619, 0.013058048248291015, 0.013004799842834473, 0.013073408126831054, 0.013034496307373047, 0.012991488456726074, 0.012975104331970215, 0.013081600189208984, 0.012945407867431641, 0.01296281623840332, 0.013157376289367676, 0.013058048248291015, 0.012981247901916505, 0.013083647727966309, 0.013314047813415527, 0.012993535995483398, 0.013092864036560058, 0.028242944717407226, 0.013030400276184082, 0.013048831939697265, 0.01305395221710205, 0.013033472061157226, 0.013058048248291015, 0.013022208213806152, 0.013064255714416504, 0.01308255958557129, 0.013124608039855956, 0.013118464469909668, 0.012998656272888183, 0.013011967658996582, 0.013036543846130372, 0.013022208213806152, 0.013128704071044921, 0.013054976463317871, 0.013126655578613282, 0.013066240310668945, 0.013033503532409669, 0.012939231872558594, 0.013085696220397949, 0.013234175682067872, 0.013249600410461426, 0.013135807991027832, 0.013026304244995117, 0.01307033634185791, 0.013116415977478027, 0.013030400276184082, 0.013033503532409669, 0.013102047920227051, 0.013033472061157226, 0.01303756809234619, 0.013109248161315918, 0.012986368179321289, 0.013240384101867677, 0.013071359634399414, 0.013055935859680175, 0.01305292797088623, 0.013175807952880859, 0.013057024002075195, 0.013056063652038574, 0.013079487800598144, 0.01305292797088623, 0.01307545566558838, 0.013128704071044921, 0.013109248161315918, 0.013074463844299316, 0.013017120361328125, 0.012999615669250487, 0.013179903984069824, 0.013029376029968261, 0.013040639877319337, 0.013011967658996582, 0.013112319946289062, 0.01306726360321045, 0.01305907154083252, 0.012999679565429688, 0.013097984313964844, 0.013029376029968261, 0.013017087936401368, 0.01306828784942627, 0.013087743759155274, 0.02812313652038574, 0.01306726360321045, 0.013007871627807617, 0.01305907154083252, 0.013093888282775879, 0.01296281623840332, 0.013049856185913086, 0.013153280258178711, 0.013021183967590331, 0.013083647727966309, 0.013009920120239257, 0.013063167572021485, 0.01313587188720703, 0.013107199668884278, 0.01307033634185791, 0.013054976463317871, 0.013090815544128418, 0.012999679565429688, 0.013054976463317871, 0.01305395221710205, 0.013029376029968261, 0.012959744453430176, 0.012958720207214355, 0.012980223655700684, 0.013061152458190919, 0.013029343605041504, 0.013011967658996582, 0.013049856185913086, 0.01354854393005371, 0.01326694393157959, 0.0132925443649292, 0.01343283176422119, 0.013219840049743652, 0.013042688369750977, 0.013019136428833008, 0.013055999755859376, 0.013085696220397949, 0.012998656272888183, 0.013035519599914551, 0.013041664123535156, 0.013005824089050292, 0.013023232460021973, 0.012781567573547363, 0.012198944091796875, 0.012355551719665528, 0.012399616241455079, 0.01232588768005371, 0.012409855842590332, 0.012333056449890138, 0.012379136085510254, 0.012327936172485352, 0.012355584144592285, 0.012368895530700684, 0.012318719863891601, 0.012306431770324706, 0.012410880088806153, 0.012956671714782715, 0.012585984230041505, 0.012337151527404786, 0.012339200019836426, 0.012453887939453125, 0.012298239707946777, 0.012349439620971679, 0.02674892807006836, 0.012324864387512208, 0.012374015808105468, 0.012386303901672363, 0.012370944023132324, 0.012207103729248046, 0.01232588768005371, 0.012340224266052247, 0.012311552047729492, 0.012361791610717774, 0.012392383575439453, 0.01236684799194336, 0.012313599586486817, 0.012305407524108887, 0.012309503555297852, 0.012368895530700684, 0.012305472373962402, 0.012352448463439941, 0.012357631683349609, 0.012285951614379884, 0.012339200019836426, 0.012306431770324706, 0.012374015808105468, 0.012315648078918457, 0.012355615615844726, 0.012303327560424805, 0.012316672325134278, 0.012331007957458496, 0.012309503555297852, 0.012275712013244629, 0.01233407974243164, 0.012346367835998535, 0.012338175773620605, 0.012332032203674317, 0.012326911926269531, 0.012314623832702636, 0.01238425636291504, 0.01233510398864746, 0.012309503555297852, 0.012302335739135742, 0.012167231559753418, 0.012144607543945312, 0.012191712379455566, 0.012248064041137695, 0.012620800018310547, 0.012620800018310547, 0.012785663604736328, 0.012713024139404297, 0.012696512222290039, 0.012643327713012695, 0.012707839965820313, 0.012731391906738282, 0.012685312271118163, 0.012676095962524414, 0.012711935997009278, 0.012380160331726075, 0.012308480262756348, 0.012718079566955566, 0.012755968093872071, 0.012724224090576173, 0.012666879653930664, 0.012277759552001954, 0.012290047645568849, 0.027765760421752928, 0.012851200103759765, 0.0127457275390625, 0.012516351699829101, 0.012830719947814942, 0.012534784317016602, 0.012319744110107422, 0.012383232116699219, 0.01275494384765625, 0.012527615547180175, 0.012541952133178711, 0.012452863693237304, 0.012560383796691894, 0.012712960243225097, 0.012547072410583495, 0.012734527587890626, 0.01278048038482666, 0.01325062370300293, 0.013735872268676758, 0.012928000450134277, 0.012864543914794921, 0.012786656379699708, 0.012743680000305176, 0.012713983535766601, 0.012776448249816894, 0.012483584403991698, 0.012587008476257324, 0.012990464210510254, 0.012790847778320312, 0.012723135948181152, 0.012750847816467285, 0.012669952392578124, 0.012699647903442383, 0.012701696395874023, 0.012693504333496093, 0.012744704246520995, 0.012510208129882813, 0.012728320121765137, 0.012812288284301757, 0.012718079566955566, 0.012695551872253418, 0.012702752113342285, 0.012743712425231934, 0.013584320068359374, 0.012843008041381837, 0.012808192253112792, 0.01274880027770996, 0.012492799758911132, 0.0123023681640625, 0.012484607696533203, 0.012739551544189454, 0.012728320121765137, 0.012811264038085938, 0.012743680000305176, 0.012733440399169921, 0.01267302417755127, 0.01266585636138916, 0.012747776031494141, 0.01294438362121582, 0.012803071975708008, 0.012718079566955566, 0.012802047729492188, 0.012767231941223145, 0.026677248001098632, 0.01233510398864746, 0.012399616241455079, 0.012340224266052247, 0.01234227180480957, 0.012347392082214356, 0.012338175773620605, 0.012318719863891601, 0.012597248077392579, 0.01276416015625, 0.012697600364685058, 0.012747776031494141, 0.012697600364685058, 0.012760064125061036, 0.012676095962524414, 0.012666879653930664, 0.012536895751953126, 0.012476351737976074, 0.012791808128356934, 0.012881919860839844, 0.012755968093872071, 0.012694527626037597, 0.01273036766052246, 0.012691455841064453, 0.012704768180847169, 0.012584959983825684, 0.012279808044433594, 0.012285951614379884, 0.012317695617675782, 0.012290047645568849, 0.012299263954162597, 0.012317695617675782, 0.012283904075622559, 0.012333056449890138, 0.012694527626037597, 0.012721152305603027, 0.012320768356323243, 0.012329983711242675, 0.012291071891784668, 0.012314623832702636, 0.012403712272644044, 0.012311552047729492, 0.012311552047729492, 0.012315648078918457, 0.012322815895080566, 0.012337151527404786, 0.012316672325134278, 0.012316672325134278, 0.012309503555297852, 0.012301312446594239, 0.012277759552001954, 0.012327936172485352, 0.012253184318542481, 0.012318719863891601, 0.012386303901672363, 0.012343359947204589, 0.012355520248413085, 0.012326911926269531, 0.012302335739135742, 0.012288000106811523, 0.012263423919677734, 0.01216102409362793, 0.01215385627746582, 0.02656153678894043, 0.012348447799682617, 0.012369888305664063, 0.01235148811340332, 0.012339200019836426, 0.012312576293945313, 0.012298239707946777, 0.012348416328430176, 0.012303359985351562, 0.012333056449890138, 0.012321791648864745, 0.012340224266052247, 0.012338208198547364, 0.012369888305664063, 0.012312576293945313, 0.012354559898376465, 0.012323840141296387, 0.012315648078918457, 0.012307456016540527, 0.012324864387512208, 0.012324864387512208, 0.012316672325134278, 0.012356608390808106, 0.012294143676757812, 0.012301312446594239, 0.012297216415405274, 0.012285951614379884, 0.012377087593078612, 0.012300288200378418, 0.01228492832183838, 0.012285951614379884, 0.012311552047729492, 0.01226956844329834, 0.01248256015777588, 0.01307545566558838, 0.01325158405303955, 0.013030400276184082, 0.01297715187072754, 0.013220864295959473, 0.015228927612304688, 0.013443072319030762, 0.013102080345153809, 0.01306009578704834, 0.012984319686889649, 0.013091839790344239, 0.012957695960998536, 0.012953599929809571, 0.013028351783752442, 0.01303660774230957, 0.012982208251953125, 0.013002752304077148, 0.013020159721374512, 0.012996607780456543, 0.012988415718078614, 0.012908543586730957, 0.013014016151428222, 0.012984319686889649, 0.012951552391052246, 0.012943360328674316, 0.012975104331970215, 0.012999744415283203, 0.013018048286437989, 0.012899328231811523, 0.028044288635253906, 0.012953599929809571, 0.013091839790344239, 0.01304371166229248, 0.012972031593322754, 0.013033472061157226, 0.013017087936401368, 0.013030400276184082, 0.012959744453430176, 0.01297107219696045, 0.013031359672546386, 0.012981247901916505, 0.012950528144836425, 0.013038592338562012, 0.013040639877319337, 0.013007871627807617, 0.013009920120239257, 0.013024255752563477, 0.013074432373046875, 0.012963839530944824, 0.012983296394348144, 0.01298739242553711, 0.013008895874023438, 0.0129617919921875, 0.013113344192504883, 0.013194239616394043, 0.013031423568725586, 0.013039615631103516, 0.01303756809234619, 0.013022208213806152, 0.013096960067749023, 0.013009920120239257, 0.013005824089050292, 0.012964863777160645, 0.013024255752563477, 0.013145088195800781, 0.013058048248291015, 0.012984319686889649, 0.013088768005371093, 0.012981247901916505, 0.013007871627807617, 0.012974080085754394, 0.012992511749267579, 0.012940287590026855, 0.013108223915100097, 0.013048831939697265, 0.013044735908508302, 0.013011967658996582, 0.012949503898620606, 0.013010944366455078, 0.013050880432128906, 0.013435903549194337, 0.013109248161315918, 0.013085696220397949, 0.012995583534240723, 0.013046784400939941, 0.013028351783752442, 0.013064191818237305, 0.013047807693481446, 0.012988415718078614, 0.013049856185913086, 0.013050880432128906, 0.013023232460021973, 0.027986944198608397, 0.013011967658996582, 0.01308572769165039, 0.013063136100769044, 0.013027327537536621, 0.01305907154083252, 0.01305907154083252, 0.013063167572021485, 0.012978240013122559, 0.013069248199462891, 0.013006848335266113, 0.012994560241699218, 0.01303756809234619, 0.013050880432128906, 0.0129617919921875, 0.012963839530944824, 0.012956671714782715, 0.01297920036315918, 0.013063167572021485, 0.013036543846130372, 0.013066240310668945, 0.01298534393310547, 0.013001728057861327, 0.012942336082458495, 0.013001728057861327, 0.012949503898620606, 0.013008895874023438, 0.012924927711486817, 0.012967935562133789, 0.012945407867431641, 0.013012991905212403, 0.013001728057861327, 0.012991488456726074, 0.01305292797088623, 0.013029376029968261, 0.013027327537536621, 0.01298739242553711, 0.013308927536010743, 0.012399616241455079, 0.012290047645568849, 0.01226854419708252, 0.012279808044433594, 0.01236684799194336, 0.012346367835998535, 0.012294143676757812, 0.013014016151428222, 0.013021183967590331, 0.012988415718078614, 0.012942367553710938, 0.012983263969421387, 0.012990464210510254, 0.013075519561767578, 0.012958656311035157, 0.013017087936401368, 0.01298739242553711, 0.012951552391052246, 0.012986368179321289, 0.013030400276184082, 0.013031423568725586, 0.013029376029968261, 0.013004799842834473, 0.012983296394348144, 0.013027327537536621, 0.026658815383911134, 0.012326911926269531, 0.012370944023132324, 0.012357631683349609, 0.012401663780212402, 0.012301312446594239, 0.012289024353027344, 0.012345343589782716, 0.01234124755859375, 0.012340224266052247, 0.01235865592956543, 0.012290047645568849, 0.012332032203674317, 0.012308480262756348, 0.012302335739135742, 0.012155903816223144, 0.01256447982788086, 0.012445695877075195, 0.012326911926269531, 0.012345343589782716, 0.012348416328430176, 0.012308480262756348, 0.012354559898376465, 0.01266380786895752, 0.013042688369750977, 0.012975104331970215, 0.013011967658996582, 0.013011967658996582, 0.013005824089050292, 0.01305907154083252, 0.01297920036315918, 0.012951552391052246, 0.013012991905212403, 0.012963839530944824, 0.013010944366455078, 0.012988415718078614, 0.01308672046661377, 0.012988415718078614, 0.013019136428833008, 0.013014016151428222, 0.013021183967590331, 0.013032447814941407, 0.01298739242553711, 0.013016063690185547, 0.013002752304077148, 0.012993535995483398, 0.01305907154083252, 0.013041664123535156, 0.013065216064453124, 0.01303756809234619, 0.01298739242553711, 0.01297100830078125, 0.013324288368225098, 0.013092864036560058, 0.013030400276184082, 0.013009920120239257, 0.013025279998779296, 0.013011967658996582, 0.01298534393310547, 0.01306726360321045, 0.01307545566558838, 0.01297920036315918, 0.01294438362121582, 0.028052480697631835, 0.012974080085754394, 0.013095935821533204, 0.013039615631103516, 0.012986368179321289, 0.01298534393310547, 0.012939264297485351, 0.013031423568725586, 0.013142016410827637, 0.013030400276184082, 0.013020159721374512, 0.012976160049438476, 0.012978143692016601, 0.013171711921691894, 0.013093888282775879, 0.013016063690185547, 0.013008895874023438, 0.012967935562133789, 0.013001728057861327, 0.013021183967590331, 0.012945407867431641, 0.013028351783752442, 0.012909567832946778, 0.013049856185913086, 0.013009984016418457, 0.013238207817077637, 0.01295257568359375, 0.0129617919921875, 0.012969984054565429, 0.012990464210510254, 0.012967935562133789, 0.012988479614257812, 0.012958656311035157, 0.013046784400939941, 0.013035519599914551, 0.012996607780456543, 0.013031423568725586, 0.013010944366455078, 0.01296281623840332, 0.012992511749267579, 0.012935168266296386, 0.013081664085388184, 0.013012928009033203, 0.012912639617919922, 0.012860416412353515, 0.012777471542358398, 0.01286348819732666, 0.012805120468139648, 0.012726271629333496, 0.012781567573547363, 0.01276518440246582, 0.01227673625946045, 0.012314623832702636, 0.012329983711242675, 0.012315648078918457, 0.012312576293945313, 0.012346367835998535, 0.012312576293945313, 0.012282879829406738, 0.012322815895080566, 0.012355648040771484, 0.01234937572479248, 0.012306431770324706, 0.026703872680664063, 0.012537856101989747, 0.012406784057617188, 0.012346367835998535, 0.013423616409301758, 0.013212672233581543, 0.013048831939697265, 0.013064191818237305, 0.01306828784942627, 0.012991488456726074, 0.01305190372467041, 0.012982272148132324, 0.012990464210510254, 0.013031423568725586, 0.013014016151428222, 0.013667327880859375, 0.013643775939941406, 0.01397555160522461, 0.012969984054565429, 0.01287782382965088, 0.01286143970489502, 0.012811264038085938, 0.012800000190734863, 0.012703743934631348, 0.012715007781982422, 0.012696576118469239, 0.01266585636138916, 0.01237606430053711, 0.012340224266052247, 0.012355584144592285, 0.012337151527404786, 0.012338239669799805, 0.012320704460144043, 0.012310527801513671, 0.012346400260925292, 0.01231663990020752, 0.01237299156188965, 0.012289088249206543, 0.01235654354095459, 0.012388352394104005, 0.012314623832702636, 0.012311552047729492, 0.01227673625946045, 0.012338175773620605, 0.012363776206970215, 0.0127457275390625, 0.012875776290893554, 0.01274880027770996, 0.012710911750793457, 0.012705792427062988, 0.012588031768798828, 0.012382207870483398, 0.012307456016540527, 0.012422207832336427, 0.01236780834197998, 0.012340224266052247, 0.012315648078918457, 0.012404735565185548, 0.012314623832702636, 0.012368895530700684, 0.012321791648864745, 0.012285951614379884, 0.012305407524108887]",tokens/s,76.64701323615108,,,main,False,False bfloat16-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1188.245504,14382.792704,0.0,13736.3456,13487.531008,s,10,0.7251225891113281,0.0725122589111328,0.0011519276516436144,0.0721160774230957,0.07271428680419922,0.07433507919311523,0.07563171310424804,"[0.07595587158203125, 0.07203961944580078, 0.07235411071777344, 0.07218275451660157, 0.07221212768554687, 0.0720672607421875, 0.07216489410400391, 0.07205974578857421, 0.07203350067138672, 0.07205270385742188]",tokens/s,3530.437526622086,kWh,8.531013457514018e-07,4.6746000959154755e-07,4.600003680000037e-06,5.9205650353429865e-06,tokens/kWh,43239116.27890252,MB,1188.573184,14382.792704,0.0,13736.3456,13661.259776,s,10,21.688784423828125,2.1688784423828125,0.0017743762365916294,2.1693094482421875,2.1706022705078127,2.1710042602539064,2.1713258520507814,"[2.170212890625, 2.167744873046875, 2.17008251953125, 2.17140625, 2.1680048828125, 2.170512939453125, 2.17001708984375, 2.165689453125, 2.16651171875, 2.168601806640625]",tokens/s,29.04727105442839,kWh,2.5559453352106564e-05,1.4007375430240984e-05,0.00011063022739299909,0.00015019705617534665,tokens/kWh,419448.96660591685,,s,629,22.303378452301015,0.035458471307314825,0.008606498420304457,0.03440435028076172,0.03462369384765625,0.03471974258422852,0.10675292083740234,"[0.03420979309082031, 0.03431935882568359, 0.03419136047363281, 0.034323455810546875, 0.03461017608642578, 0.0345425910949707, 0.03450163269042969, 0.03425689697265625, 0.0346879997253418, 0.03490304183959961, 0.03458047866821289, 0.03445452880859375, 0.034490367889404294, 0.034473983764648435, 0.034541568756103515, 0.034598911285400394, 0.03448115158081055, 0.034500606536865236, 0.0341923828125, 0.03441664123535156, 0.03428966522216797, 0.034293758392333985, 0.034418689727783204, 0.03435728073120117, 0.03423638534545898, 0.03431219100952149, 0.034301952362060545, 0.03432550430297852, 0.03427840042114258, 0.03441049575805664, 0.03431628799438476, 0.03431219100952149, 0.034320384979248046, 0.03451903915405274, 0.034587646484375, 0.03446476745605469, 0.03448320007324219, 0.03434188842773438, 0.034871295928955076, 0.03489689636230469, 0.03439616012573242, 0.03453440093994141, 0.034460670471191404, 0.03456719970703125, 0.03429475021362305, 0.034618366241455076, 0.034436161041259766, 0.03445344161987305, 0.034328575134277346, 0.03440435028076172, 0.03442073440551758, 0.0343900146484375, 0.03446476745605469, 0.034408447265625, 0.03440435028076172, 0.03451289749145508, 0.03434598541259765, 0.03449753570556641, 0.03434905624389648, 0.034557952880859374, 0.03438489532470703, 0.03449139022827148, 0.10669977569580077, 0.034229248046875, 0.03427635192871094, 0.034233345031738284, 0.034315265655517575, 0.03440435028076172, 0.03422822570800781, 0.03426611328125, 0.03429785537719727, 0.03426406478881836, 0.03435212707519531, 0.0344268798828125, 0.034348033905029295, 0.03422719955444336, 0.034260990142822266, 0.03437363052368164, 0.034372608184814454, 0.034721790313720705, 0.03434188842773438, 0.034388992309570314, 0.03448627090454102, 0.034362400054931644, 0.03438179016113281, 0.03437363052368164, 0.034374656677246096, 0.03439411163330078, 0.03429683303833008, 0.0342845458984375, 0.03431219100952149, 0.03420467376708984, 0.03433171081542969, 0.03441145706176758, 0.034288639068603514, 0.034219009399414066, 0.03433267211914062, 0.03425689697265625, 0.03434393692016602, 0.03440639877319336, 0.03459174346923828, 0.03462758255004883, 0.03438796615600586, 0.034372608184814454, 0.0343818244934082, 0.03482624053955078, 0.034543617248535156, 0.034530303955078126, 0.03443404769897461, 0.03444224166870117, 0.03436646270751953, 0.03455590438842773, 0.034484222412109376, 0.034429950714111326, 0.034462718963623046, 0.03441766357421875, 0.03449753570556641, 0.03469107055664063, 0.03464704132080078, 0.034586654663085935, 0.03443299102783203, 0.03443609619140625, 0.034423809051513675, 0.03468288040161133, 0.03446579360961914, 0.10744217681884766, 0.034310142517089845, 0.03446783828735352, 0.03440947341918945, 0.03449760055541992, 0.03435103988647461, 0.03426508712768555, 0.03425894546508789, 0.03429177474975586, 0.03420665740966797, 0.03419955062866211, 0.034149375915527344, 0.034298881530761716, 0.03424358367919922, 0.03424361419677734, 0.03426095962524414, 0.03437977600097656, 0.03428147125244141, 0.03446579360961914, 0.03429580688476563, 0.03462656021118164, 0.03503923034667969, 0.03473100662231445, 0.03474739074707031, 0.03449651336669922, 0.03450265502929688, 0.03454265594482422, 0.034509761810302735, 0.03430099105834961, 0.034675647735595706, 0.03442483139038086, 0.034385921478271485, 0.03446681594848633, 0.034513919830322266, 0.0345456657409668, 0.034560001373291016, 0.03430092620849609, 0.03426816177368164, 0.03445452880859375, 0.03434598541259765, 0.0343900146484375, 0.034351104736328124, 0.034377727508544925, 0.034369537353515625, 0.03440947341918945, 0.03434188842773438, 0.03449139022827148, 0.034710529327392575, 0.034597888946533206, 0.03457228851318359, 0.03456512069702149, 0.03438387298583984, 0.034411518096923825, 0.03448115158081055, 0.03442278289794922, 0.03449343872070312, 0.034479103088378905, 0.03433779144287109, 0.03449760055541992, 0.0343438720703125, 0.03448524856567383, 0.034505729675292966, 0.03469004821777344, 0.10690354919433594, 0.034181121826171876, 0.03421593475341797, 0.0341739501953125, 0.0342476806640625, 0.03423027038574219, 0.03422208023071289, 0.03424358367919922, 0.03428870391845703, 0.034237377166748045, 0.03433779144287109, 0.0344453125, 0.03448534393310547, 0.03440732955932617, 0.03457024002075195, 0.03455590438842773, 0.03450470352172851, 0.03446783828735352, 0.03448831939697266, 0.03445558547973633, 0.0344749755859375, 0.034552833557128904, 0.03450368118286133, 0.034522113800048826, 0.03448320007324219, 0.034511871337890625, 0.03450470352172851, 0.034482177734375, 0.03456204986572266, 0.03451289749145508, 0.03451903915405274, 0.03458150482177735, 0.03436032104492188, 0.03428659057617187, 0.034551807403564457, 0.03445248031616211, 0.03450982284545898, 0.034411518096923825, 0.03543961715698242, 0.0348671989440918, 0.034667518615722655, 0.03457331085205078, 0.03457436752319336, 0.034474014282226566, 0.0343807373046875, 0.034374656677246096, 0.03438387298583984, 0.03441971206665039, 0.034495487213134765, 0.034391040802001956, 0.03444121551513672, 0.03450470352172851, 0.03460710525512695, 0.03432755279541016, 0.03460812759399414, 0.03464396667480469, 0.03460710525512695, 0.034388992309570314, 0.034484222412109376, 0.03435520172119141, 0.03448425674438477, 0.03430089569091797, 0.03439616012573242, 0.1069834213256836, 0.034255870819091795, 0.0343818244934082, 0.03443814468383789, 0.03428659057617187, 0.03420979309082031, 0.03417292785644531, 0.03434393692016602, 0.034290687561035156, 0.03420467376708984, 0.03422515106201172, 0.034219009399414066, 0.034280479431152346, 0.03424457550048828, 0.034305057525634765, 0.03450467300415039, 0.034232383728027345, 0.03423839950561523, 0.034435073852539064, 0.03436544036865234, 0.03454880142211914, 0.03447187042236328, 0.034500606536865236, 0.03454054260253906, 0.03443404769897461, 0.034233345031738284, 0.03427020645141601, 0.034272254943847655, 0.034374656677246096, 0.0342927360534668, 0.03430604934692383, 0.03441561508178711, 0.034359294891357424, 0.03443711853027344, 0.03440947341918945, 0.034285568237304685, 0.03464704132080078, 0.034477054595947264, 0.03460300827026367, 0.034579456329345705, 0.034971649169921876, 0.03463372802734375, 0.03470336151123047, 0.0344453125, 0.03456921768188476, 0.034331649780273435, 0.03446169662475586, 0.03466649627685547, 0.034533374786376955, 0.03430604934692383, 0.03444838333129883, 0.034334720611572264, 0.034375679016113284, 0.034476032257080076, 0.034403358459472656, 0.03446780776977539, 0.03441971206665039, 0.034418689727783204, 0.034418689727783204, 0.03427942276000977, 0.03443814468383789, 0.03443302536010742, 0.03463782501220703, 0.10678374481201172, 0.03428966522216797, 0.03423231887817383, 0.03411455917358398, 0.03462348937988281, 0.03505868911743164, 0.03434086227416992, 0.03432550430297852, 0.03446169662475586, 0.03417190551757812, 0.03433881759643555, 0.03462451171875, 0.03439923095703125, 0.03423539352416992, 0.03427840042114258, 0.034277374267578126, 0.034305023193359374, 0.03430303955078125, 0.034270145416259765, 0.034200576782226565, 0.03432755279541016, 0.03424460983276367, 0.03428147125244141, 0.034342910766601564, 0.03434393692016602, 0.034310142517089845, 0.034523136138916014, 0.034477054595947264, 0.03455590438842773, 0.03432352066040039, 0.03450361633300781, 0.03452928161621094, 0.034590721130371094, 0.034462753295898436, 0.03460095977783203, 0.034369503021240234, 0.034356224060058595, 0.03446579360961914, 0.03438489532470703, 0.03445043182373047, 0.034541568756103515, 0.03445148849487305, 0.03433577728271484, 0.03437254333496094, 0.03438387298583984, 0.03434393692016602, 0.03438796615600586, 0.03441766357421875, 0.034590721130371094, 0.03496345520019531, 0.03461734390258789, 0.03465625762939453, 0.034786304473876956, 0.03452928161621094, 0.03464704132080078, 0.03445555114746094, 0.03449753570556641, 0.03448831939697266, 0.034716670989990234, 0.03457535934448242, 0.03458867263793945, 0.03456204986572266, 0.034609153747558595, 0.10681753540039063, 0.03440947341918945, 0.034353153228759765, 0.03425996780395508, 0.03448627090454102, 0.034207775115966794, 0.03421795272827149, 0.03425689697265625, 0.03425791931152344, 0.03428659057617187, 0.03438387298583984, 0.03441254425048828, 0.03426303863525391, 0.034184192657470705, 0.03433574295043945, 0.0342999038696289, 0.03426508712768555, 0.03417702484130859, 0.03438079833984375, 0.03442278289794922, 0.034476032257080076, 0.03443404769897461, 0.034282497406005856, 0.034423809051513675, 0.034321407318115234, 0.03422208023071289, 0.034405376434326174, 0.03430912017822266, 0.034358272552490236, 0.03432550430297852, 0.03446169662475586, 0.03441766357421875, 0.03434905624389648, 0.03451903915405274, 0.03460505676269531, 0.03421491241455078, 0.034522113800048826, 0.03439616012573242, 0.03458873748779297, 0.03447907257080078, 0.0343592643737793, 0.0343296012878418, 0.03453849411010742, 0.03445248031616211, 0.03442073440551758, 0.03454873657226563, 0.03458457565307617, 0.034699264526367186, 0.034601982116699216, 0.034358272552490236, 0.034597888946533206, 0.03484467315673828, 0.03479142379760742, 0.03462246322631836, 0.03464601516723633, 0.03461324691772461, 0.03464908981323242, 0.034759681701660154, 0.034631679534912106, 0.03459481430053711, 0.0347248649597168, 0.03444121551513672, 0.034609153747558595, 0.10678886413574219, 0.03424256134033203, 0.034212894439697265, 0.03421488189697266, 0.034162689208984375, 0.03419136047363281, 0.034318336486816405, 0.034299934387207034, 0.03422825622558594, 0.03414729690551758, 0.034276321411132814, 0.03429683303833008, 0.034477054595947264, 0.03433369445800781, 0.03431628799438476, 0.03452108764648437, 0.03461119842529297, 0.03444940948486328, 0.034334720611572264, 0.03433062362670898, 0.034402305603027344, 0.03428147125244141, 0.03428659057617187, 0.03450368118286133, 0.03431321716308594, 0.03421491241455078, 0.03437977600097656, 0.03425900650024414, 0.03428448104858398, 0.034334720611572264, 0.03436544036865234, 0.03433267211914062, 0.034476032257080076, 0.03435520172119141, 0.034361343383789066, 0.03432755279541016, 0.03435520172119141, 0.03432447814941406, 0.03438387298583984, 0.03441971206665039, 0.03445248031616211, 0.03428147125244141, 0.03431423950195313, 0.034356224060058595, 0.034365470886230466, 0.03432751846313477, 0.03450777435302734, 0.03462656021118164, 0.03443199920654297, 0.034293758392333985, 0.03455487823486328, 0.034374656677246096, 0.03443097686767578, 0.0343818244934082, 0.03443916702270508, 0.03469619369506836, 0.034716670989990234, 0.034425918579101565, 0.03452000045776367, 0.03436032104492188, 0.0344453125, 0.03435417556762695, 0.03442483139038086, 0.10674483489990234, 0.034162689208984375, 0.03439718246459961, 0.03475763320922851, 0.03462041473388672, 0.034200576782226565, 0.03426201629638672, 0.03427948760986328, 0.034311103820800784, 0.03417702484130859, 0.0343296012878418, 0.03426611328125, 0.03422822570800781, 0.03416883087158203, 0.03442483139038086, 0.03434188842773438, 0.03446681594848633, 0.0345241584777832, 0.03436032104492188, 0.034318336486816405, 0.034307071685791016, 0.034229248046875, 0.03439011383056641, 0.0343551025390625, 0.034253822326660154, 0.034255870819091795, 0.03427840042114258, 0.03420979309082031, 0.03428761672973633, 0.034595840454101565, 0.03462656021118164, 0.034479103088378905, 0.03430912017822266, 0.034405376434326174, 0.03462348937988281, 0.03444736099243164, 0.03433881759643555, 0.034323455810546875, 0.03452320098876953, 0.034376640319824216, 0.034374656677246096, 0.034288639068603514, 0.034328575134277346, 0.03437363052368164, 0.034362400054931644, 0.034394081115722654, 0.03445248031616211, 0.034375679016113284, 0.034531326293945314, 0.034338878631591796, 0.03454252624511719, 0.034477054595947264, 0.03448934555053711, 0.034388992309570314, 0.03443199920654297, 0.03444121551513672, 0.03439923095703125, 0.034377727508544925, 0.03453747177124023, 0.03433779144287109, 0.03443404769897461, 0.03438796615600586, 0.034533409118652346, 0.10675606536865234, 0.03417292785644531, 0.034283584594726565, 0.03416057586669922, 0.03428966522216797, 0.034522113800048826, 0.034925567626953126, 0.03478732681274414, 0.034490367889404294, 0.034407424926757815, 0.03430809783935547, 0.03421184158325195, 0.03434086227416992, 0.03420876693725586, 0.03429171371459961, 0.03427840042114258, 0.03431219100952149, 0.03421388626098633, 0.03424665451049805, 0.0343296012878418, 0.034313247680664065, 0.03422819137573242, 0.03426713562011719, 0.03434086227416992, 0.03426201629638672, 0.03422719955444336, 0.03427840042114258, 0.03436441421508789, 0.03439206314086914, 0.034301952362060545, 0.0343900146484375, 0.03433062362670898, 0.03437670516967774, 0.03425791931152344, 0.03453747177124023, 0.034784255981445314, 0.03486310577392578, 0.03429788970947266, 0.03442582321166992, 0.03439923095703125, 0.03456921768188476, 0.03443609619140625, 0.034587646484375, 0.03454771041870117, 0.03459686279296875, 0.034563072204589845, 0.03444838333129883, 0.03437977600097656, 0.03458150482177735, 0.034328575134277346, 0.034457599639892575, 0.03441459274291992, 0.03470745468139649, 0.03450982284545898, 0.03444224166870117, 0.03444326400756836, 0.03463577651977539, 0.034353153228759765, 0.0345846061706543, 0.03445756912231445, 0.03467161560058594, 0.03435417556762695, 0.03446579360961914]",tokens/s,28.20200542017465,,,main,False,False -bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5593,7 +5593,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5650,7 +5650,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1264.66048,4276.617216,0.0,3632.267264,3408.132096,s,10,0.213487455368042,0.0213487455368042,0.00038207157218038363,0.021232367515563965,0.021424921989440918,0.021956252574920652,0.02238131704330444,"[0.02248758316040039, 0.021306848526000975, 0.02123923110961914, 0.021238431930541993, 0.021165184020996094, 0.021216960906982423, 0.021226303100585937, 0.021224639892578126, 0.021244895935058593, 0.02113737678527832]",tokens/s,11991.336894182772,kWh,2.5255655744250117e-07,1.383878192229578e-07,1.2493534448760467e-06,1.6402978215415056e-06,tokens/kWh,156069219.03939274,MB,1264.66048,4278.714368,0.0,3632.267264,3474.14784,s,10,11.683104492187502,1.16831044921875,0.03179779513772458,1.1519906616210938,1.2127111938476562,1.2202495178222657,1.2262801770019531,"[1.227787841796875, 1.2032935791015624, 1.2110360107421876, 1.154431640625, 1.1466201171875, 1.14823046875, 1.16621044921875, 1.126555908203125, 1.1495496826171876, 1.1493887939453125]",tokens/s,53.92402339817139,kWh,1.3614571337912724e-05,7.45993332269954e-06,3.769223326392422e-05,5.876673792453647e-05,tokens/kWh,1072035.001855974,,s,629,11.854563329696665,0.018846682559136175,0.0026529975843851924,0.018518016815185546,0.019255309295654298,0.019856384277343753,0.039754261932373046,"[0.020084735870361328, 0.019014656066894533, 0.018783231735229493, 0.018775039672851563, 0.01883852767944336, 0.019876863479614256, 0.01988812828063965, 0.020040704727172853, 0.01999564743041992, 0.020165632247924805, 0.02001408004760742, 0.020239360809326173, 0.020168703079223634, 0.018760704040527345, 0.018883583068847656, 0.019992576599121094, 0.020246528625488282, 0.02003763198852539, 0.019959808349609375, 0.020493312835693358, 0.01999667167663574, 0.019877887725830077, 0.02018611145019531, 0.01987481689453125, 0.019722240447998047, 0.019971071243286134, 0.020051967620849608, 0.019685375213623048, 0.019695680618286134, 0.0195665283203125, 0.01925017547607422, 0.01929113578796387, 0.019106815338134766, 0.019129344940185547, 0.019131391525268555, 0.019169279098510742, 0.01924300765991211, 0.019117055892944337, 0.01924095916748047, 0.019198976516723632, 0.019117055892944337, 0.019273727416992188, 0.019224576950073242, 0.01922150421142578, 0.01925836753845215, 0.01920204734802246, 0.01918976020812988, 0.019185663223266602, 0.019135488510131835, 0.019102783203125, 0.019727359771728514, 0.019380159378051758, 0.019219455718994142, 0.019142656326293944, 0.019133440017700197, 0.01928294372558594, 0.01922764778137207, 0.019180543899536134, 0.019107839584350587, 0.019132415771484376, 0.019449855804443358, 0.019255359649658202, 0.04104185485839844, 0.019121152877807617, 0.017993728637695314, 0.01784934425354004, 0.019305471420288087, 0.019555328369140625, 0.020343807220458983, 0.01921023941040039, 0.019116031646728517, 0.019131391525268555, 0.01904947280883789, 0.018940927505493164, 0.019759103775024413, 0.01923583984375, 0.01903411293029785, 0.01919385528564453, 0.018884607315063476, 0.0193832950592041, 0.019002368927001953, 0.01907302474975586, 0.01902899169921875, 0.01909974479675293, 0.018896799087524414, 0.018964479446411133, 0.01901158332824707, 0.01903513526916504, 0.018955263137817382, 0.018966527938842775, 0.018876415252685547, 0.01902592086791992, 0.019036384582519533, 0.018917152404785156, 0.018874368667602538, 0.019066879272460938, 0.01904947280883789, 0.018949119567871094, 0.01890928077697754, 0.018877344131469728, 0.018852863311767578, 0.018979839324951172, 0.018905120849609373, 0.01912931251525879, 0.019134464263916014, 0.019132415771484376, 0.018922496795654296, 0.01922662353515625, 0.01927884864807129, 0.0198287353515625, 0.01924095916748047, 0.019117088317871095, 0.019179487228393556, 0.01923993682861328, 0.01919692802429199, 0.019152896881103516, 0.019204095840454103, 0.019125247955322267, 0.0192225284576416, 0.019186687469482423, 0.019117055892944337, 0.0191016960144043, 0.019135488510131835, 0.019110912322998046, 0.01915395164489746, 0.040965087890625, 0.019312671661376953, 0.019208223342895507, 0.019216320037841798, 0.019125247955322267, 0.01923891258239746, 0.019151872634887695, 0.019211263656616212, 0.01926553535461426, 0.01918976020812988, 0.01921843147277832, 0.019155967712402345, 0.01922662353515625, 0.019180576324462892, 0.019234783172607423, 0.019284000396728517, 0.0191231689453125, 0.01922559928894043, 0.01924403190612793, 0.019207168579101562, 0.019216384887695313, 0.019204095840454103, 0.01925529670715332, 0.019296255111694336, 0.01922559928894043, 0.01920512008666992, 0.01949388885498047, 0.019188735961914064, 0.019368959426879884, 0.01924812889099121, 0.019203071594238282, 0.01926348876953125, 0.019186687469482423, 0.019183616638183593, 0.01922969627380371, 0.019182592391967773, 0.01921843147277832, 0.01925734329223633, 0.019155967712402345, 0.01922662353515625, 0.01921846389770508, 0.019209184646606446, 0.019187711715698243, 0.019168256759643554, 0.019188735961914064, 0.01925529670715332, 0.019190784454345702, 0.019280895233154297, 0.01904025650024414, 0.019212287902832033, 0.019185663223266602, 0.019186687469482423, 0.01919692802429199, 0.019138559341430664, 0.019219455718994142, 0.01930342483520508, 0.019212287902832033, 0.019184640884399414, 0.019152896881103516, 0.019156991958618166, 0.019207168579101562, 0.019178495407104493, 0.01923788833618164, 0.04085760116577149, 0.01927884864807129, 0.019215360641479492, 0.019145727157592773, 0.019168256759643554, 0.019159040451049804, 0.01921331214904785, 0.01912633514404297, 0.0191026554107666, 0.019134464263916014, 0.019147775650024415, 0.01918156814575195, 0.019137535095214844, 0.019098623275756836, 0.01919900894165039, 0.01869308853149414, 0.01801113510131836, 0.0180316162109375, 0.017942527770996093, 0.017954816818237306, 0.017983488082885742, 0.017966079711914062, 0.01884160041809082, 0.01899212837219238, 0.01795686340332031, 0.01794047927856445, 0.01821696090698242, 0.01801318359375, 0.0180316162109375, 0.01799884796142578, 0.017928192138671875, 0.017938432693481447, 0.017985536575317384, 0.017963008880615236, 0.017935359954833984, 0.017999872207641602, 0.018159616470336915, 0.01804697608947754, 0.018000896453857423, 0.017970176696777345, 0.01802137565612793, 0.018000896453857423, 0.01795686340332031, 0.018025472640991212, 0.017991680145263672, 0.017946624755859376, 0.017977344512939454, 0.01798860740661621, 0.01802649688720703, 0.01799065589904785, 0.018050048828125, 0.017984512329101563, 0.017992704391479493, 0.017987583160400392, 0.017967103958129883, 0.018000959396362304, 0.018033599853515624, 0.01797324752807617, 0.018017280578613282, 0.01801215934753418, 0.01802239990234375, 0.018018367767333985, 0.018074560165405273, 0.03959500885009765, 0.01806540870666504, 0.01801215934753418, 0.01803980827331543, 0.018058240890502928, 0.018025472640991212, 0.018118656158447266, 0.017977344512939454, 0.01803059196472168, 0.01804287910461426, 0.017985536575317384, 0.01800704002380371, 0.01802137565612793, 0.01799884796142578, 0.017961984634399415, 0.017984512329101563, 0.017932287216186525, 0.01799065589904785, 0.018069503784179687, 0.01820364761352539, 0.01797324752807617, 0.017994752883911135, 0.01802444839477539, 0.017979391098022462, 0.01820467185974121, 0.01805516815185547, 0.018050048828125, 0.01800499153137207, 0.018027519226074217, 0.01800704002380371, 0.018070528030395508, 0.018022464752197265, 0.018017215728759764, 0.01795180892944336, 0.017980352401733398, 0.01802137565612793, 0.018052127838134764, 0.01805513572692871, 0.01802137565612793, 0.018542591094970702, 0.018745344161987306, 0.018569215774536133, 0.018529279708862305, 0.018553855895996094, 0.018511871337890624, 0.018541568756103514, 0.018538496017456055, 0.018495487213134765, 0.018338815689086914, 0.018391103744506837, 0.0183438720703125, 0.01828556823730469, 0.01839308738708496, 0.017986560821533205, 0.01802649688720703, 0.018028543472290038, 0.018205759048461913, 0.019146688461303712, 0.018747392654418944, 0.018565120697021483, 0.018513919830322266, 0.018601984024047852, 0.018572288513183592, 0.0403056640625, 0.018661376953125, 0.018519039154052733, 0.01843609619140625, 0.018516992568969725, 0.01840230369567871, 0.018286752700805663, 0.0183089599609375, 0.01839206314086914, 0.018584575653076172, 0.01864806365966797, 0.018548736572265623, 0.01858252716064453, 0.018531328201293946, 0.01860403251647949, 0.017958911895751953, 0.01804800033569336, 0.017993728637695314, 0.018000896453857423, 0.018076671600341796, 0.017968128204345703, 0.01802956771850586, 0.01798041534423828, 0.01803980827331543, 0.01800704002380371, 0.01802649688720703, 0.017976320266723633, 0.01804902458190918, 0.017938432693481447, 0.017992704391479493, 0.018074623107910158, 0.017937408447265626, 0.01799679946899414, 0.017945600509643556, 0.017992704391479493, 0.01802444839477539, 0.017994752883911135, 0.0179814395904541, 0.01800912094116211, 0.018014175415039062, 0.01799884796142578, 0.017999872207641602, 0.018085887908935547, 0.01798041534423828, 0.017942527770996093, 0.01795686340332031, 0.018217983245849608, 0.01799577522277832, 0.018714624404907225, 0.01862246322631836, 0.01858355140686035, 0.01858252716064453, 0.018586624145507814, 0.018572288513183592, 0.01859071922302246, 0.01861734390258789, 0.01860710334777832, 0.01855897521972656, 0.017985536575317384, 0.017977344512939454, 0.01799782371520996, 0.01795686340332031, 0.0179814395904541, 0.03977318572998047, 0.018085887908935547, 0.01803878402709961, 0.01802239990234375, 0.01801523208618164, 0.01802956771850586, 0.01799782371520996, 0.018009088516235353, 0.01802137565612793, 0.01802649688720703, 0.01799884796142578, 0.01798041534423828, 0.01799065589904785, 0.01803059196472168, 0.01797222328186035, 0.018104320526123048, 0.01846886444091797, 0.019963903427124022, 0.020136991500854493, 0.01948259162902832, 0.0188590087890625, 0.01863680076599121, 0.018510847091674804, 0.018619392395019533, 0.018541568756103514, 0.018537471771240235, 0.018573312759399413, 0.018570240020751954, 0.018671615600585938, 0.018581504821777343, 0.018563072204589845, 0.018610176086425782, 0.018597888946533202, 0.01859071922302246, 0.018700288772583007, 0.018580480575561522, 0.018521120071411133, 0.018568159103393555, 0.01858355140686035, 0.018562047958374024, 0.01859584045410156, 0.018538496017456055, 0.018646015167236327, 0.018518016815185546, 0.018536447525024414, 0.018564096450805666, 0.018542591094970702, 0.018538496017456055, 0.018534400939941405, 0.018552831649780274, 0.018530303955078126, 0.01859174346923828, 0.018586624145507814, 0.018677759170532226, 0.018572288513183592, 0.018903039932250978, 0.0186562557220459, 0.018512895584106445, 0.018532352447509767, 0.018535423278808593, 0.018515968322753908, 0.018544639587402344, 0.018597888946533202, 0.04017663955688477, 0.018563072204589845, 0.01850476837158203, 0.018381759643554686, 0.01809715270996094, 0.017863679885864257, 0.01781862449645996, 0.01784524726867676, 0.01786675262451172, 0.01784012794494629, 0.017686527252197267, 0.017757183074951173, 0.01770086479187012, 0.017725439071655275, 0.01784320068359375, 0.017703935623168944, 0.017732608795166017, 0.01777663993835449, 0.017819648742675782, 0.01792527961730957, 0.017791839599609376, 0.01773465538024902, 0.01777561569213867, 0.017767520904541017, 0.017779615402221678, 0.017728511810302734, 0.017803455352783205, 0.017824575424194335, 0.017731584548950196, 0.017750015258789064, 0.017686527252197267, 0.017780736923217775, 0.01780735969543457, 0.01801625633239746, 0.017744895935058593, 0.01780633544921875, 0.017756160736083985, 0.01775324821472168, 0.017786720275878905, 0.017720319747924804, 0.01781452751159668, 0.017756160736083985, 0.017701887130737306, 0.017757183074951173, 0.017794048309326172, 0.017775680541992186, 0.01771820831298828, 0.017878015518188475, 0.017714176177978515, 0.017934335708618163, 0.01790771293640137, 0.01800396728515625, 0.01797324752807617, 0.017838079452514647, 0.017943552017211914, 0.017921024322509766, 0.018066432952880858, 0.01804287910461426, 0.017954816818237306, 0.018028543472290038, 0.017991680145263672, 0.018086912155151368, 0.01804902458190918, 0.04018483352661133, 0.019185663223266602, 0.018629663467407225, 0.018362335205078125, 0.018516992568969725, 0.01823744010925293, 0.018296831130981444, 0.01842995262145996, 0.018309152603149415, 0.018204639434814453, 0.017959936141967774, 0.017930240631103517, 0.017975296020507812, 0.01799679946899414, 0.017985536575317384, 0.018001920700073244, 0.017960960388183594, 0.018035711288452147, 0.01801420783996582, 0.01799577522277832, 0.01799679946899414, 0.017920000076293945, 0.01798963165283203, 0.017941503524780272, 0.01805721664428711, 0.018167808532714845, 0.018661376953125, 0.018552831649780274, 0.018554880142211915, 0.01859993553161621, 0.018526208877563476, 0.018514944076538087, 0.01855897521972656, 0.0184770565032959, 0.018308095932006836, 0.01848320007324219, 0.018584575653076172, 0.018572288513183592, 0.018563072204589845, 0.018593791961669923, 0.01863167953491211, 0.018544639587402344, 0.01861631965637207, 0.018153472900390624, 0.018067455291748045, 0.01807155227661133, 0.018028543472290038, 0.018010112762451173, 0.017986623764038086, 0.017970111846923827, 0.018001920700073244, 0.0180316162109375, 0.01801625633239746, 0.01801215934753418, 0.01804083251953125, 0.01799884796142578, 0.01804902458190918, 0.01800704002380371, 0.018043903350830077, 0.01805516815185547, 0.01801215934753418, 0.018124799728393554, 0.018512895584106445, 0.03970560073852539, 0.018001920700073244, 0.01805516815185547, 0.01804697608947754, 0.018001920700073244, 0.018035711288452147, 0.01800396728515625, 0.01803264045715332, 0.01801420783996582, 0.01801318359375, 0.017960960388183594, 0.017991680145263672, 0.018027519226074217, 0.017933311462402343, 0.01801625633239746, 0.01802444839477539, 0.018421760559082033, 0.01859891128540039, 0.018514944076538087, 0.018568191528320312, 0.017986560821533205, 0.018066432952880858, 0.01799884796142578, 0.018027519226074217, 0.01802342414855957, 0.01806540870666504, 0.018018304824829103, 0.018076671600341796, 0.018268159866333008, 0.018523136138916017, 0.018300928115844727, 0.018545663833618165, 0.018568191528320312, 0.01903206443786621, 0.01922867202758789, 0.019533824920654298, 0.018744319915771485, 0.01863167953491211, 0.018572288513183592, 0.018276351928710938, 0.018582592010498045, 0.01851897621154785, 0.018724863052368163, 0.018562047958374024, 0.018515968322753908, 0.018545663833618165, 0.01825279998779297, 0.018027519226074217, 0.018018304824829103, 0.018113536834716795, 0.018009088516235353, 0.01804287910461426, 0.017991680145263672, 0.018044927597045898, 0.01799782371520996, 0.018043903350830077, 0.01802444839477539, 0.018052095413208007, 0.018061311721801757, 0.01803264045715332, 0.01800396728515625, 0.017963008880615236, 0.017966079711914062]",tokens/s,53.05973594356727,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5687,7 +5687,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 768.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6132,7 +6132,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6159,7 +6159,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6302,7 +6302,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6343,7 +6343,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6802,7 +6802,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7095,7 +7095,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7651,7 +7651,7 @@ OSError: google/recurrentgemma-7b is not a local folder and is not a valid model If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7678,7 +7678,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8018,7 +8018,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8490,7 +8490,7 @@ OSError: 0 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8745,7 +8745,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1272.25856,4276.617216,0.0,3632.267264,3408.132096,s,10,0.21382742500305174,0.021382742500305175,0.00040380614256645505,0.021255472183227538,0.02143521595001221,0.02201244020462036,0.022474219608306884,"[0.022589664459228515, 0.021306943893432618, 0.021212192535400392, 0.02128316879272461, 0.021218847274780274, 0.021263967514038085, 0.021209951400756835, 0.021289920806884764, 0.02120579147338867, 0.021246976852416992]",tokens/s,11972.271564152557,kWh,2.526427775706321e-07,1.3843550400918338e-07,1.2562159859785875e-06,1.6472942675584028e-06,tokens/kWh,155406356.3758039,MB,1272.25856,4278.714368,0.0,3632.267264,3474.14784,s,10,12.003229980468753,1.200322998046875,0.009540185117779846,1.2029205322265626,1.2098683227539062,1.2106496154785156,1.211274649658203,"[1.206054443359375, 1.1821016845703125, 1.1959776611328126, 1.20736474609375, 1.211430908203125, 1.19922900390625, 1.19978662109375, 1.206357177734375, 1.1852330322265625, 1.2096947021484374]",tokens/s,52.48587263804115,kWh,1.3756805777210114e-05,7.5383188242675455e-06,3.8057648798421595e-05,5.935277339989927e-05,tokens/kWh,1061449.9776704102,,s,629,12.17444862937927,0.019355244243846218,0.002585969156717962,0.019136512756347656,0.01934458808898926,0.019548159408569337,0.04027453338623047,"[0.019579967498779296, 0.01917024040222168, 0.019124223709106446, 0.01923072052001953, 0.019338239669799806, 0.019154943466186524, 0.01930342483520508, 0.019219455718994142, 0.01925119972229004, 0.019072000503540038, 0.019138559341430664, 0.019336191177368164, 0.019397632598876953, 0.01919385528564453, 0.019192832946777344, 0.019117055892944337, 0.01942118453979492, 0.01923583984375, 0.01924608039855957, 0.019359807968139648, 0.01912006378173828, 0.019132415771484376, 0.018818080902099608, 0.01891529655456543, 0.018658367156982422, 0.018852800369262696, 0.01965772819519043, 0.02002841567993164, 0.019401727676391603, 0.0196177921295166, 0.019132415771484376, 0.01921843147277832, 0.018778175354003907, 0.018218944549560547, 0.018255872726440428, 0.018367488861083983, 0.01835212707519531, 0.018981887817382814, 0.01927577590942383, 0.019293184280395507, 0.019075071334838867, 0.019713151931762696, 0.019009408950805665, 0.018917375564575196, 0.01899622344970703, 0.019100671768188478, 0.019843072891235353, 0.01965363121032715, 0.01923993682861328, 0.019124223709106446, 0.018924543380737305, 0.018934783935546876, 0.01882316780090332, 0.019082239151000976, 0.019115007400512696, 0.01903615951538086, 0.0188590087890625, 0.01902079963684082, 0.018898944854736328, 0.01880985641479492, 0.01906892776489258, 0.019330047607421876, 0.040850433349609375, 0.01903923225402832, 0.019086336135864256, 0.019113983154296875, 0.01855897521972656, 0.018543615341186523, 0.018523136138916017, 0.018457599639892578, 0.01848422431945801, 0.018448383331298827, 0.018490367889404297, 0.01844326400756836, 0.018428928375244142, 0.01845043182373047, 0.018496511459350586, 0.01846784019470215, 0.018446367263793947, 0.01846883201599121, 0.018951168060302736, 0.019107839584350587, 0.01906073570251465, 0.019155967712402345, 0.01908121681213379, 0.019148799896240236, 0.018694143295288086, 0.0184586238861084, 0.01859993553161621, 0.018546688079833985, 0.01846886444091797, 0.01863270378112793, 0.01841971206665039, 0.018463743209838866, 0.0184453125, 0.018408447265625, 0.01843404769897461, 0.01839923286437988, 0.0184586238861084, 0.019177471160888672, 0.019204095840454103, 0.01947238349914551, 0.019080192565917968, 0.01925017547607422, 0.01903308868408203, 0.01920921516418457, 0.019306495666503908, 0.018524160385131837, 0.018524160385131837, 0.018536447525024414, 0.018594816207885743, 0.019140607833862306, 0.019133440017700197, 0.019115007400512696, 0.019005439758300782, 0.01904435157775879, 0.019096576690673828, 0.018472959518432617, 0.01847091293334961, 0.018512895584106445, 0.01861529541015625, 0.018514944076538087, 0.01922764778137207, 0.018951168060302736, 0.018513919830322266, 0.04029747009277344, 0.019301376342773437, 0.019135488510131835, 0.01906380844116211, 0.019135488510131835, 0.01909663963317871, 0.019165119171142577, 0.019080192565917968, 0.019080255508422853, 0.01914566421508789, 0.019142656326293944, 0.019163135528564454, 0.018508895874023438, 0.01861008071899414, 0.018505727767944336, 0.018519039154052733, 0.018461696624755858, 0.019074047088623047, 0.01841868782043457, 0.018594816207885743, 0.018481151580810547, 0.018499584197998048, 0.018512895584106445, 0.018518016815185546, 0.018479135513305663, 0.01847702407836914, 0.018485248565673826, 0.01881395149230957, 0.019264511108398438, 0.018555904388427736, 0.018511903762817382, 0.018492383956909178, 0.01845452880859375, 0.019542015075683594, 0.019335168838500977, 0.019180543899536134, 0.019479551315307618, 0.019132415771484376, 0.01924198341369629, 0.01905971145629883, 0.019186752319335938, 0.01917740821838379, 0.019082239151000976, 0.019180543899536134, 0.019131391525268555, 0.019269632339477538, 0.019139583587646485, 0.01920204734802246, 0.019152896881103516, 0.019142656326293944, 0.019191808700561523, 0.019122175216674805, 0.019176448822021484, 0.019168256759643554, 0.019155967712402345, 0.019195903778076173, 0.019148799896240236, 0.019146751403808594, 0.019166208267211913, 0.01924300765991211, 0.01929523277282715, 0.019105791091918945, 0.01940787124633789, 0.04074905776977539, 0.01918060874938965, 0.01916102409362793, 0.018911231994628908, 0.018923519134521484, 0.02000588798522949, 0.019154943466186524, 0.019163135528564454, 0.01923481559753418, 0.019096576690673828, 0.019186687469482423, 0.019180543899536134, 0.01922150421142578, 0.019098623275756836, 0.018791423797607423, 0.018991104125976564, 0.019582975387573243, 0.01945497512817383, 0.01908940887451172, 0.018890752792358398, 0.018949119567871094, 0.019345407485961915, 0.019354623794555666, 0.019177471160888672, 0.019174400329589843, 0.01924710464477539, 0.01919692802429199, 0.01924608039855957, 0.0190382080078125, 0.01921433639526367, 0.019096576690673828, 0.019215360641479492, 0.01926144027709961, 0.019162176132202148, 0.0191343994140625, 0.01920512008666992, 0.01962393569946289, 0.01922150421142578, 0.019074047088623047, 0.019091455459594727, 0.01901158332824707, 0.019084287643432618, 0.019090431213378906, 0.01907302474975586, 0.01903001594543457, 0.01923276710510254, 0.019330047607421876, 0.019177471160888672, 0.019079168319702147, 0.019034143447875976, 0.01910780715942383, 0.019106815338134766, 0.019006464004516603, 0.019153919219970703, 0.019132415771484376, 0.019169279098510742, 0.019096576690673828, 0.01905971145629883, 0.019191808700561523, 0.019149824142456053, 0.019177471160888672, 0.01904025650024414, 0.01903923225402832, 0.04016230392456055, 0.018553855895996094, 0.019129344940185547, 0.01906483268737793, 0.019340288162231444, 0.01925734329223633, 0.019293184280395507, 0.019136512756347656, 0.0192174072265625, 0.019418111801147463, 0.019323904037475585, 0.01922047996520996, 0.019090431213378906, 0.019275808334350587, 0.019354591369628905, 0.019388416290283202, 0.019163135528564454, 0.019185663223266602, 0.019168256759643554, 0.01920204734802246, 0.019187711715698243, 0.019135488510131835, 0.01920102310180664, 0.019292160034179686, 0.01923788833618164, 0.019272703170776367, 0.019150848388671874, 0.01924710464477539, 0.01923993682861328, 0.019141632080078123, 0.019345439910888673, 0.019782623291015624, 0.019552255630493166, 0.019147775650024415, 0.019142816543579102, 0.019079008102416993, 0.01903308868408203, 0.019418111801147463, 0.0192174072265625, 0.019198976516723632, 0.019176448822021484, 0.019112960815429687, 0.01930342483520508, 0.01922662353515625, 0.019190784454345702, 0.01914681625366211, 0.019561439514160155, 0.019230688095092773, 0.01909350395202637, 0.019166208267211913, 0.019145727157592773, 0.01945907211303711, 0.019698688507080078, 0.019168256759643554, 0.01926144027709961, 0.019184640884399414, 0.019570688247680663, 0.01919385528564453, 0.019148799896240236, 0.01919385528564453, 0.019147775650024415, 0.019195903778076173, 0.019122175216674805, 0.04099071884155273, 0.01918976020812988, 0.019175424575805664, 0.019151872634887695, 0.019184640884399414, 0.01920921516418457, 0.01914886474609375, 0.019173311233520507, 0.019143680572509765, 0.01926348876953125, 0.019041471481323242, 0.019021631240844727, 0.01909350395202637, 0.01887846374511719, 0.018938880920410156, 0.018933759689331055, 0.01904128074645996, 0.01904025650024414, 0.018807872772216797, 0.0189736328125, 0.01885798454284668, 0.0190515193939209, 0.018919488906860352, 0.019041215896606446, 0.01901260757446289, 0.019072000503540038, 0.018918399810791017, 0.018931711196899414, 0.01896873664855957, 0.019012575149536134, 0.01906163215637207, 0.018932735443115235, 0.018932735443115235, 0.018969600677490234, 0.01916431999206543, 0.019007328033447266, 0.018977792739868164, 0.01900851249694824, 0.0190086727142334, 0.018988895416259765, 0.018936832427978514, 0.019110912322998046, 0.018998271942138673, 0.018966527938842775, 0.018942975997924806, 0.018918399810791017, 0.018948095321655273, 0.018959360122680666, 0.019208192825317383, 0.01921753692626953, 0.018999231338500976, 0.020249536514282226, 0.01944780731201172, 0.01926144027709961, 0.019272703170776367, 0.019332096099853514, 0.019170303344726563, 0.01906483268737793, 0.01841766357421875, 0.018499584197998048, 0.01841868782043457, 0.018511871337890624, 0.01847500801086426, 0.040215553283691405, 0.018527231216430663, 0.018514944076538087, 0.01849241638183594, 0.01842790412902832, 0.01846784019470215, 0.018454559326171877, 0.019949535369873046, 0.019783679962158202, 0.01948467254638672, 0.019177471160888672, 0.01936591911315918, 0.01913750457763672, 0.019105791091918945, 0.01844326400756836, 0.018493440628051756, 0.018453504562377928, 0.018516992568969725, 0.018497535705566406, 0.018963455200195312, 0.019112960815429687, 0.019363840103149413, 0.019133440017700197, 0.01922764778137207, 0.018481151580810547, 0.018530303955078126, 0.018566144943237304, 0.018463743209838866, 0.019121152877807617, 0.01922764778137207, 0.019156991958618166, 0.019095552444458007, 0.019149824142456053, 0.01943654441833496, 0.01918156814575195, 0.019109888076782225, 0.019160064697265625, 0.019172351837158205, 0.019143680572509765, 0.0192491512298584, 0.019308544158935546, 0.019148799896240236, 0.01921331214904785, 0.01920614433288574, 0.019192832946777344, 0.019138559341430664, 0.019216384887695313, 0.019144704818725586, 0.019216384887695313, 0.019173376083374022, 0.019117055892944337, 0.019161088943481445, 0.019106815338134766, 0.019149824142456053, 0.019188735961914064, 0.019182592391967773, 0.019319807052612305, 0.01918156814575195, 0.019169279098510742, 0.019153919219970703, 0.019122175216674805, 0.019344383239746094, 0.019308544158935546, 0.040787967681884765, 0.01918156814575195, 0.019183616638183593, 0.019006464004516603, 0.01922764778137207, 0.019215360641479492, 0.019294208526611328, 0.01925836753845215, 0.019186687469482423, 0.019216384887695313, 0.019138559341430664, 0.01923379135131836, 0.019155967712402345, 0.019203071594238282, 0.019191808700561523, 0.019105791091918945, 0.01923788833618164, 0.01926553535461426, 0.019198976516723632, 0.019203071594238282, 0.01921023941040039, 0.019508224487304687, 0.019418111801147463, 0.01926246452331543, 0.019134464263916014, 0.019494911193847657, 0.02004684829711914, 0.019304447174072266, 0.019109888076782225, 0.019138559341430664, 0.019304447174072266, 0.019184640884399414, 0.019105791091918945, 0.019187711715698243, 0.019095552444458007, 0.019382272720336914, 0.019153919219970703, 0.019195903778076173, 0.01924505615234375, 0.019096576690673828, 0.01903308868408203, 0.018541568756103514, 0.0184770565032959, 0.01864806365966797, 0.01918976020812988, 0.019088384628295898, 0.019109888076782225, 0.019489791870117186, 0.018503679275512695, 0.019104768753051758, 0.019088384628295898, 0.019169279098510742, 0.01905766487121582, 0.019123199462890626, 0.019313663482666017, 0.019078144073486326, 0.01926655960083008, 0.019108863830566408, 0.018490367889404297, 0.0189040641784668, 0.019147775650024415, 0.018555904388427736, 0.019115007400512696, 0.04064665603637695, 0.01918976020812988, 0.019009536743164062, 0.01927065658569336, 0.019288063049316406, 0.019151935577392577, 0.01928390312194824, 0.019131391525268555, 0.019277824401855468, 0.019309568405151366, 0.019096576690673828, 0.019070976257324217, 0.01906585693359375, 0.01906073570251465, 0.01900441551208496, 0.019084287643432618, 0.018963455200195312, 0.01843097686767578, 0.01942630386352539, 0.01925836753845215, 0.01903615951538086, 0.01841049575805664, 0.01844534492492676, 0.018623455047607422, 0.019499008178710937, 0.01909760093688965, 0.01904742431640625, 0.019096576690673828, 0.01903615951538086, 0.019006464004516603, 0.01909760093688965, 0.019084287643432618, 0.018531328201293946, 0.01843404769897461, 0.01836851119995117, 0.018381824493408205, 0.01836031913757324, 0.018480159759521483, 0.018365440368652345, 0.018444255828857423, 0.01842278480529785, 0.01844326400756836, 0.01846272087097168, 0.018533376693725585, 0.01845964813232422, 0.018404352188110353, 0.01836755180358887, 0.018269119262695314, 0.01826617622375488, 0.018185152053833007, 0.01820467185974121, 0.0183767032623291, 0.01844326400756836, 0.01845248031616211, 0.018513919830322266, 0.01845145606994629, 0.018456575393676757, 0.01848320007324219, 0.01852012825012207, 0.01892038345336914, 0.02003865623474121, 0.019587072372436523, 0.019283967971801756, 0.041128959655761715, 0.019110912322998046, 0.018928640365600585, 0.019177471160888672, 0.018961536407470704, 0.018986879348754884, 0.019083263397216797, 0.019188735961914064, 0.019755008697509766, 0.019857440948486328, 0.01946006393432617, 0.01930342483520508, 0.01927577590942383, 0.019155967712402345, 0.01923276710510254, 0.019219455718994142, 0.019170303344726563, 0.019318784713745117, 0.019143680572509765, 0.01927168083190918, 0.01924710464477539, 0.01922051239013672, 0.019190752029418945, 0.019117055892944337, 0.019200000762939453, 0.01924198341369629, 0.01940995216369629, 0.019125215530395506, 0.01920614433288574, 0.019207168579101562, 0.019102720260620116, 0.019139583587646485, 0.019140607833862306, 0.019056640625, 0.019198976516723632, 0.019171327590942384, 0.01921023941040039, 0.019190784454345702, 0.019149824142456053, 0.01925836753845215, 0.019140607833862306, 0.019187711715698243, 0.0191016960144043, 0.019195968627929688, 0.01909446334838867, 0.01915088081359863, 0.01912931251525879, 0.019137535095214844, 0.019191808700561523, 0.019115007400512696, 0.01920512008666992, 0.019166208267211913, 0.019176448822021484, 0.019146751403808594, 0.019144704818725586, 0.0192225284576416, 0.019155967712402345, 0.019211263656616212, 0.019098623275756836, 0.019151872634887695, 0.019084287643432618, 0.019200000762939453, 0.01909350395202637]",tokens/s,51.665584138414516,,,,, -float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8782,7 +8782,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 768.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9129,7 +9129,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9156,7 +9156,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9347,7 +9347,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GP ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1273.880576,3852.992512,0.0,3206.545408,3105.82784,s,10,0.22610268783569334,0.022610268783569337,0.0003043157777445479,0.022503392219543457,0.022695283699035645,0.023103321933746337,0.023429752521514892,"[0.02351136016845703, 0.022567775726318358, 0.022604608535766603, 0.022471967697143554, 0.02250752067565918, 0.022471071243286133, 0.022437023162841796, 0.022499263763427733, 0.022476512908935545, 0.02255558395385742]",tokens/s,11322.289109010182,kWh,2.67829261879742e-07,1.4675720045043093e-07,1.2030231342494404e-06,1.6176095965796133e-06,tokens/kWh,158258210.4738401,MB,1273.880576,3852.992512,0.0,3206.545408,3189.19168,s,10,10.90676721191406,1.0906767211914064,0.031225445779376405,1.0901775512695313,1.1228399291992188,1.1271606872558593,1.1306172937011718,"[1.1314814453125, 1.1189879150390625, 1.1182877197265626, 1.117508056640625, 1.1218797607421875, 1.0628470458984376, 1.06013671875, 1.0576103515625, 1.055681884765625, 1.0623463134765625]",tokens/s,57.76230369268506,kWh,1.248988201618795e-05,6.843971163420882e-06,3.5835700941150545e-05,5.516955412075939e-05,tokens/kWh,1141934.1882318049,,s,629,11.090413570404062,0.01763181807695398,0.002740089693097124,0.017496063232421876,0.01792020492553711,0.018042905044555663,0.039768857116699224,"[0.017466367721557616, 0.018114559173583983, 0.018130943298339842, 0.01821388816833496, 0.017960960388183594, 0.017921024322509766, 0.017947647094726564, 0.01789952087402344, 0.017757183074951173, 0.017733631134033204, 0.01826918411254883, 0.017854463577270507, 0.017765375137329103, 0.018148351669311523, 0.017920000076293945, 0.01775103950500488, 0.01787494468688965, 0.01783296012878418, 0.01784217643737793, 0.01778483200073242, 0.01794867134094238, 0.01788313674926758, 0.017745920181274414, 0.017741823196411134, 0.017960960388183594, 0.017898496627807618, 0.01773465538024902, 0.01809721565246582, 0.017894336700439453, 0.017936384201049805, 0.01816268730163574, 0.01802444839477539, 0.018181119918823242, 0.01823641586303711, 0.01800499153137207, 0.017970176696777345, 0.01785753631591797, 0.01773161506652832, 0.01773257637023926, 0.01782374382019043, 0.017904640197753906, 0.01801523208618164, 0.017771520614624024, 0.01805516815185547, 0.017889280319213868, 0.01785958480834961, 0.018250751495361327, 0.018108448028564452, 0.017756128311157228, 0.01787596893310547, 0.017762304306030274, 0.017770496368408203, 0.017821695327758787, 0.017872896194458008, 0.017944576263427735, 0.01802137565612793, 0.01820979118347168, 0.01887846374511719, 0.01822412872314453, 0.0178288631439209, 0.01777663993835449, 0.01786476707458496, 0.039884735107421875, 0.016793600082397463, 0.016763904571533202, 0.016755712509155272, 0.017069055557250978, 0.017730560302734375, 0.017739776611328126, 0.017765375137329103, 0.01774284744262695, 0.017736703872680663, 0.01776639938354492, 0.01781760025024414, 0.017744895935058593, 0.017698816299438477, 0.01780735969543457, 0.01769267272949219, 0.0177838077545166, 0.01768454360961914, 0.017737728118896484, 0.01772640037536621, 0.01784012794494629, 0.01783296012878418, 0.017693695068359376, 0.017786880493164063, 0.017667072296142578, 0.017745920181274414, 0.017744895935058593, 0.01776639938354492, 0.017777664184570312, 0.01776639938354492, 0.01778278350830078, 0.01775923156738281, 0.01777561569213867, 0.017789951324462892, 0.017779712677001954, 0.01776742362976074, 0.01782681655883789, 0.017829887390136717, 0.0178606071472168, 0.0178155517578125, 0.017803264617919923, 0.017862655639648437, 0.01785958480834961, 0.01784217643737793, 0.017946687698364258, 0.017865663528442384, 0.01780940818786621, 0.017900543212890627, 0.01784832000732422, 0.018039871215820312, 0.017830848693847656, 0.017778688430786133, 0.0184453125, 0.01901568031311035, 0.018107391357421874, 0.017942527770996093, 0.017804288864135744, 0.017833984375, 0.017772544860839845, 0.0178606071472168, 0.017768447875976562, 0.017780736923217775, 0.017821695327758787, 0.03989299011230469, 0.017688575744628905, 0.017720319747924804, 0.017738752365112305, 0.017715200424194336, 0.01779814338684082, 0.017622016906738282, 0.01779302406311035, 0.017725439071655275, 0.017999872207641602, 0.01779199981689453, 0.01780121612548828, 0.017941503524780272, 0.017957887649536132, 0.01776639938354492, 0.017789951324462892, 0.017803264617919923, 0.017773567199707033, 0.01785036849975586, 0.01777663993835449, 0.01763942337036133, 0.01774393653869629, 0.01773356819152832, 0.017704959869384765, 0.01770086479187012, 0.017743871688842772, 0.017913856506347657, 0.017752063751220702, 0.017836032867431642, 0.01780019187927246, 0.01787596893310547, 0.017915903091430666, 0.017729536056518554, 0.017819648742675782, 0.017758207321166994, 0.01806540870666504, 0.017719295501708983, 0.017896448135375977, 0.017596416473388672, 0.017488895416259767, 0.017496063232421876, 0.01759129524230957, 0.017819648742675782, 0.017992704391479493, 0.017768447875976562, 0.017698816299438477, 0.017746944427490235, 0.017732608795166017, 0.017715200424194336, 0.017770496368408203, 0.017735679626464843, 0.017773567199707033, 0.017863679885864257, 0.017510400772094727, 0.01742438316345215, 0.017501184463500977, 0.017483776092529296, 0.017735679626464843, 0.01789952087402344, 0.0176629753112793, 0.017881088256835938, 0.017917951583862304, 0.017812480926513673, 0.039839744567871094, 0.016785408020019533, 0.016777215957641603, 0.016749568939208984, 0.01696873664855957, 0.01807049560546875, 0.017889280319213868, 0.017779712677001954, 0.017724416732788087, 0.017715200424194336, 0.017777664184570312, 0.01781760025024414, 0.017763328552246094, 0.017902591705322265, 0.017829887390136717, 0.017699840545654297, 0.017752063751220702, 0.017687551498413084, 0.01776742362976074, 0.01781862449645996, 0.017731584548950196, 0.017753087997436523, 0.017999872207641602, 0.01774284744262695, 0.01770086479187012, 0.017740800857543947, 0.017785856246948242, 0.017755136489868165, 0.01780019187927246, 0.017721343994140625, 0.01789030456542969, 0.01781760025024414, 0.01779916763305664, 0.017745920181274414, 0.017777664184570312, 0.01777459144592285, 0.01781350326538086, 0.017770496368408203, 0.017836032867431642, 0.01800499153137207, 0.017757183074951173, 0.01776950454711914, 0.017792991638183594, 0.01781657600402832, 0.017844224929809572, 0.01780735969543457, 0.017812480926513673, 0.01807673645019531, 0.01772640037536621, 0.017780736923217775, 0.01789446449279785, 0.017663936614990234, 0.017945600509643556, 0.017923072814941408, 0.01785651206970215, 0.01764352035522461, 0.017904640197753906, 0.017718271255493166, 0.017789951324462892, 0.017757183074951173, 0.01779097557067871, 0.018036735534667968, 0.017819648742675782, 0.04085452651977539, 0.017826847076416016, 0.017741792678833006, 0.0178155517578125, 0.017718271255493166, 0.017940511703491212, 0.017727455139160155, 0.017779712677001954, 0.017723392486572266, 0.01784524726867676, 0.017955839157104494, 0.01773673629760742, 0.017733600616455077, 0.017861631393432616, 0.017730560302734375, 0.017888256072998047, 0.01784934425354004, 0.017887231826782226, 0.017836032867431642, 0.01777561569213867, 0.017702911376953127, 0.01749504089355469, 0.017521696090698244, 0.017771488189697267, 0.017819648742675782, 0.017758207321166994, 0.017675264358520508, 0.017546239852905272, 0.017547264099121093, 0.017768447875976562, 0.017771520614624024, 0.017918975830078124, 0.01783193588256836, 0.01779302406311035, 0.018044927597045898, 0.017740800857543947, 0.01779814338684082, 0.017821695327758787, 0.017514495849609374, 0.017582143783569336, 0.01771001625061035, 0.01784524726867676, 0.017880064010620117, 0.018009151458740234, 0.01780940818786621, 0.01795577621459961, 0.017804288864135744, 0.017821695327758787, 0.01783500862121582, 0.01779097557067871, 0.017797119140625, 0.01798041534423828, 0.017929216384887696, 0.018099264144897462, 0.017889215469360353, 0.017804288864135744, 0.01796505546569824, 0.017839103698730468, 0.01788313674926758, 0.017862655639648437, 0.01774185562133789, 0.017760223388671875, 0.017771520614624024, 0.039783489227294924, 0.016850879669189453, 0.016776191711425782, 0.016773120880126953, 0.017558528900146485, 0.01782681655883789, 0.017107967376708985, 0.017753087997436523, 0.017699840545654297, 0.017762304306030274, 0.017876991271972655, 0.016827392578125, 0.01680998420715332, 0.01684787178039551, 0.01679769515991211, 0.016760831832885743, 0.01681203269958496, 0.01678950309753418, 0.01680998420715332, 0.01680998420715332, 0.016784383773803712, 0.01677107238769531, 0.01678643226623535, 0.01677004814147949, 0.01681305694580078, 0.016959487915039064, 0.016897024154663084, 0.01680998420715332, 0.01678745651245117, 0.01680179214477539, 0.016718847274780273, 0.01660006332397461, 0.016630783081054687, 0.016703487396240235, 0.01681407928466797, 0.01685196876525879, 0.0167956485748291, 0.01679769515991211, 0.01677516746520996, 0.016698368072509767, 0.016590848922729492, 0.0166430721282959, 0.016652288436889647, 0.016694271087646484, 0.01664614486694336, 0.016673791885375978, 0.016683008193969725, 0.016744447708129884, 0.016753664016723634, 0.016662527084350585, 0.016730112075805666, 0.01662873649597168, 0.016816160202026368, 0.016820192337036133, 0.01682329559326172, 0.01682841682434082, 0.01680281639099121, 0.0168222713470459, 0.016740352630615234, 0.0166430721282959, 0.01678335952758789, 0.016785408020019533, 0.01681203269958496, 0.039726112365722654, 0.016792543411254884, 0.016727039337158203, 0.0167587833404541, 0.0167956485748291, 0.016753664016723634, 0.01680384063720703, 0.01679974365234375, 0.01677926445007324, 0.01685196876525879, 0.01676595115661621, 0.01681100845336914, 0.01679871940612793, 0.01684377670288086, 0.01680384063720703, 0.016877567291259766, 0.016745471954345705, 0.01680384063720703, 0.016777215957641603, 0.016887807846069337, 0.01681305694580078, 0.01677516746520996, 0.01681612777709961, 0.016741376876831054, 0.01677414321899414, 0.01679052734375, 0.016767999649047852, 0.0167445125579834, 0.016777151107788085, 0.01661235237121582, 0.01662566375732422, 0.016678911209106445, 0.016694271087646484, 0.01676595115661621, 0.016825344085693358, 0.01680793571472168, 0.01681715202331543, 0.01679462432861328, 0.016759807586669923, 0.017071104049682616, 0.01713667106628418, 0.01688675117492676, 0.01680793571472168, 0.016829439163208008, 0.016763904571533202, 0.01677004814147949, 0.016829439163208008, 0.016891904830932617, 0.016755712509155272, 0.01683660888671875, 0.01706188774108887, 0.01720832061767578, 0.016982015609741212, 0.016907264709472656, 0.016923648834228516, 0.016955392837524414, 0.016870399475097657, 0.01683046340942383, 0.016785408020019533, 0.01679155158996582, 0.016772096633911132, 0.01684480094909668, 0.01679974365234375, 0.03984076690673828, 0.01683660888671875, 0.016760831832885743, 0.016788480758666992, 0.016874496459960937, 0.016756736755371093, 0.016751615524291993, 0.016784383773803712, 0.016777215957641603, 0.01675775909423828, 0.016752639770507814, 0.016749568939208984, 0.016781312942504883, 0.01678643226623535, 0.016759807586669923, 0.016773120880126953, 0.01681305694580078, 0.01677516746520996, 0.01681715202331543, 0.016782400131225585, 0.016764863967895508, 0.01677824020385742, 0.0168089599609375, 0.016777215957641603, 0.016738304138183592, 0.016841728210449217, 0.01683865547180176, 0.01677107238769531, 0.01676288032531738, 0.01678335952758789, 0.016759807586669923, 0.016740352630615234, 0.01677414321899414, 0.016799808502197266, 0.016801727294921874, 0.016827392578125, 0.01685196876525879, 0.01682841682434082, 0.01679155158996582, 0.01683967971801758, 0.01678233528137207, 0.01680998420715332, 0.0168089599609375, 0.01681715202331543, 0.01676595115661621, 0.01675468826293945, 0.01684377670288086, 0.01681920051574707, 0.01677004814147949, 0.01678950309753418, 0.016736255645751954, 0.016710655212402344, 0.016780288696289062, 0.016730112075805666, 0.016735231399536133, 0.016617504119873047, 0.016613344192504882, 0.01681715202331543, 0.016788480758666992, 0.01676288032531738, 0.016741439819335936, 0.01675667190551758, 0.01679769515991211, 0.039731231689453125, 0.016744415283203126, 0.016673791885375978, 0.016702463150024414, 0.016752639770507814, 0.016697343826293946, 0.016676864624023437, 0.01654374313354492, 0.01657344055175781, 0.01662259292602539, 0.016718847274780273, 0.01680179214477539, 0.01678950309753418, 0.01680998420715332, 0.016718847274780273, 0.016773120880126953, 0.01721855926513672, 0.016749568939208984, 0.016728063583374024, 0.01675468826293945, 0.016736255645751954, 0.01675775909423828, 0.01678950309753418, 0.016745471954345705, 0.016746496200561522, 0.016781312942504883, 0.016750591278076172, 0.01681510353088379, 0.01676288032531738, 0.01675468826293945, 0.016745471954345705, 0.01680384063720703, 0.016631807327270508, 0.01660518455505371, 0.016673791885375978, 0.01658780860900879, 0.01662870407104492, 0.01676288032531738, 0.016717824935913086, 0.01675468826293945, 0.016649215698242188, 0.01658678436279297, 0.01661231994628906, 0.01666975975036621, 0.016779199600219726, 0.01677824020385742, 0.01680486488342285, 0.016769023895263673, 0.016780288696289062, 0.016793600082397463, 0.016750591278076172, 0.016764928817749023, 0.01681612777709961, 0.016781312942504883, 0.01677107238769531, 0.01681305694580078, 0.01679155158996582, 0.016947200775146484, 0.01701785659790039, 0.01677516746520996, 0.01680588722229004, 0.01682022476196289, 0.016761856079101564, 0.03988582229614258, 0.01681612777709961, 0.016760831832885743, 0.01678643226623535, 0.016775232315063476, 0.01672902488708496, 0.01677107238769531, 0.016784383773803712, 0.01677516746520996, 0.016825344085693358, 0.016780288696289062, 0.01683456039428711, 0.01677107238769531, 0.01681203269958496, 0.01678950309753418, 0.016729087829589845, 0.01679667282104492, 0.01678950309753418, 0.01680076789855957, 0.016685056686401366, 0.016631872177124023, 0.016600000381469728, 0.01662566375732422, 0.01678643226623535, 0.016718847274780273, 0.01681715202331543, 0.01680998420715332, 0.016793600082397463, 0.01676595115661621, 0.01677516746520996, 0.016729087829589845, 0.016892927169799805, 0.016745471954345705, 0.016792575836181642, 0.01681407928466797, 0.01680179214477539, 0.01679155158996582, 0.016784383773803712, 0.01680281639099121, 0.01678335952758789, 0.016911359786987306, 0.016740352630615234, 0.016846847534179688, 0.016947200775146484, 0.01683046340942383, 0.01686016082763672, 0.016776191711425782, 0.01681920051574707, 0.01676697540283203, 0.01680588722229004, 0.01678950309753418, 0.01700044822692871, 0.01720524787902832, 0.016745471954345705, 0.01678335952758789, 0.01675574493408203, 0.016754655838012694, 0.0167587833404541, 0.017737728118896484, 0.017736703872680663, 0.017539072036743163, 0.017468416213989257, 0.017398784637451172]",tokens/s,56.71564870029314,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1296.32256,6797.39392,0.0,6150.946816,5988.307968,s,10,0.3439705619812012,0.03439705619812013,0.00028379952317182644,0.0342990894317627,0.034425238037109375,0.03483617134094238,0.03516491798400879,"[0.03524710464477539, 0.03430201721191406, 0.034300289154052734, 0.03429308700561524, 0.034333919525146486, 0.03429167938232422, 0.03429788970947266, 0.03428041458129883, 0.034293888092041015, 0.03433027267456055]",tokens/s,7442.497361561742,kWh,4.077723770762769e-07,2.2344095077453947e-07,2.018158702648284e-06,2.6493720304991006e-06,tokens/kWh,96626671.1707429,MB,1296.65024,6797.39392,0.0,6150.946816,6117.056512,s,10,13.51379443359375,1.351379443359375,0.008485435861613749,1.3481452026367187,1.3610734985351562,1.364829229736328,1.3678338146972657,"[1.3685849609375, 1.3602388916015624, 1.357792724609375, 1.34811376953125, 1.3554085693359375, 1.346927978515625, 1.3481766357421876, 1.3412799072265624, 1.347203857421875, 1.340067138671875]",tokens/s,46.619030879579746,kWh,1.590264242727638e-05,8.714526017867747e-06,5.493133130135079e-05,7.95484997464949e-05,tokens/kWh,791969.6813990001,,s,629,13.799259147644053,0.021938408819783855,0.0041494085879454174,0.021323776245117186,0.021876121139526368,0.02208463363647461,0.05605527572631836,"[0.02187264060974121, 0.02148863983154297, 0.021340160369873046, 0.02130636787414551, 0.021355520248413085, 0.021312511444091797, 0.0212992000579834, 0.021531648635864258, 0.021313535690307618, 0.021332992553710937, 0.0212490234375, 0.021316608428955077, 0.0212674560546875, 0.02126131248474121, 0.0212541446685791, 0.021355520248413085, 0.021545984268188476, 0.02186137580871582, 0.02180607986450195, 0.021324800491333007, 0.02125721549987793, 0.02126848030090332, 0.02130636787414551, 0.021323776245117186, 0.021534719467163087, 0.026009599685668947, 0.02509926414489746, 0.022369279861450195, 0.021420032501220702, 0.021321727752685548, 0.021312511444091797, 0.02145280075073242, 0.021909503936767577, 0.021819391250610352, 0.02183475112915039, 0.021774335861206053, 0.02190438461303711, 0.021741567611694337, 0.022090751647949217, 0.02190438461303711, 0.021555200576782226, 0.02184294319152832, 0.0216494083404541, 0.022227968215942383, 0.021748735427856446, 0.02184294319152832, 0.021765119552612306, 0.021227519989013673, 0.021348352432250976, 0.021321727752685548, 0.021817344665527344, 0.021710847854614256, 0.021734399795532225, 0.021401599884033205, 0.021566463470458985, 0.02187059211730957, 0.02180607986450195, 0.021603328704833984, 0.021304319381713867, 0.021731327056884766, 0.021570560455322265, 0.021326847076416015, 0.056668159484863284, 0.02170675277709961, 0.021202943801879884, 0.021296127319335938, 0.021344255447387696, 0.021948415756225585, 0.021393407821655275, 0.021251071929931642, 0.02128486442565918, 0.021387264251708983, 0.021530624389648437, 0.02103091239929199, 0.02106675148010254, 0.02127667236328125, 0.021757951736450197, 0.021809152603149414, 0.021536800384521486, 0.02173641586303711, 0.021575679779052736, 0.02186956787109375, 0.021369855880737306, 0.022303743362426756, 0.02187468719482422, 0.021839872360229492, 0.02165043258666992, 0.021279743194580078, 0.021585920333862304, 0.02104319953918457, 0.021575679779052736, 0.0218603515625, 0.021320703506469727, 0.02148863983154297, 0.021336063385009766, 0.021746688842773438, 0.02167807960510254, 0.021291072845458985, 0.021342144012451172, 0.021585920333862304, 0.02186956787109375, 0.02186444854736328, 0.021569536209106444, 0.021316608428955077, 0.02173030471801758, 0.021602304458618164, 0.021315584182739256, 0.02132275199890137, 0.022208511352539064, 0.022006784439086914, 0.021720064163208007, 0.02106879997253418, 0.021312511444091797, 0.02168115234375, 0.0216760311126709, 0.021801984786987305, 0.021167104721069335, 0.021179391860961915, 0.021773311614990236, 0.021902336120605468, 0.02185625648498535, 0.02190336036682129, 0.021968896865844727, 0.022013952255249023, 0.021881856918334962, 0.056581119537353515, 0.021971967697143553, 0.02183888053894043, 0.021606367111206056, 0.02126335906982422, 0.021320703506469727, 0.021180416107177736, 0.021150720596313476, 0.02130227279663086, 0.021206016540527343, 0.021579776763916016, 0.02165760040283203, 0.02191155242919922, 0.02169856071472168, 0.021699583053588867, 0.021894144058227538, 0.021557247161865235, 0.02124595260620117, 0.02122444725036621, 0.02162483215332031, 0.021881856918334962, 0.021836799621582033, 0.021598207473754884, 0.02128281593322754, 0.021796863555908205, 0.021329919815063478, 0.021279743194580078, 0.021320703506469727, 0.02183884811401367, 0.021954559326171876, 0.021086208343505858, 0.02105753517150879, 0.021212160110473634, 0.02165350341796875, 0.021622783660888673, 0.021159936904907226, 0.021312511444091797, 0.021988351821899413, 0.023443456649780273, 0.022245376586914063, 0.021316608428955077, 0.021219327926635743, 0.021401599884033205, 0.021893119812011717, 0.021813247680664064, 0.02183475112915039, 0.021361663818359376, 0.021775360107421874, 0.02389299201965332, 0.02222489547729492, 0.02111187171936035, 0.021202943801879884, 0.02117318344116211, 0.021014528274536134, 0.021153791427612305, 0.02105753517150879, 0.021136383056640624, 0.02147225570678711, 0.021072895050048827, 0.02105855941772461, 0.02101759910583496, 0.021246976852416992, 0.021264415740966797, 0.05615100860595703, 0.021242879867553712, 0.02149580764770508, 0.02227097511291504, 0.022339584350585938, 0.02210304069519043, 0.02186751937866211, 0.021507072448730468, 0.021122047424316406, 0.02126438331604004, 0.021348352432250976, 0.021187583923339845, 0.02127769660949707, 0.021255168914794922, 0.02143027114868164, 0.02100223922729492, 0.02102681541442871, 0.021032960891723632, 0.021177343368530274, 0.02101862335205078, 0.02102374458312988, 0.02103091239929199, 0.021082111358642578, 0.021132287979125978, 0.02123472023010254, 0.021222368240356445, 0.021259263992309572, 0.021190656661987304, 0.021275648117065428, 0.021223424911499023, 0.02122444725036621, 0.021307392120361326, 0.0212807674407959, 0.021584896087646483, 0.02143129539489746, 0.021407743453979493, 0.021327871322631836, 0.021332992553710937, 0.021259263992309572, 0.021352447509765626, 0.021350400924682617, 0.02128895950317383, 0.02127872085571289, 0.02126540756225586, 0.02167296028137207, 0.021580799102783203, 0.021764095306396485, 0.021296127319335938, 0.021389312744140625, 0.021429248809814453, 0.021448703765869142, 0.021122047424316406, 0.021283840179443358, 0.02185113525390625, 0.02125619125366211, 0.02105548858642578, 0.02124185562133789, 0.02125312042236328, 0.023193599700927735, 0.022075456619262697, 0.021352415084838868, 0.021165023803710936, 0.021511167526245118, 0.05597491073608398, 0.02103910446166992, 0.021202943801879884, 0.021818368911743165, 0.021554176330566405, 0.021183488845825195, 0.021227519989013673, 0.02119987106323242, 0.02140880012512207, 0.021683168411254884, 0.021343231201171875, 0.021255168914794922, 0.02162073516845703, 0.021823488235473632, 0.021755903244018555, 0.021320703506469727, 0.021555200576782226, 0.021949440002441405, 0.021599231719970705, 0.02128998374938965, 0.02128998374938965, 0.021634143829345705, 0.021810079574584963, 0.02150297546386719, 0.0212674560546875, 0.021293119430541994, 0.021891008377075194, 0.021525503158569336, 0.021295103073120117, 0.021285888671875, 0.02142310333251953, 0.021819391250610352, 0.021799936294555664, 0.02143436813354492, 0.021356544494628905, 0.022255615234375, 0.022106111526489256, 0.021955583572387697, 0.02188800048828125, 0.02142207908630371, 0.02182246398925781, 0.021533695220947266, 0.02125721549987793, 0.02125209617614746, 0.021327871322631836, 0.02181427192687988, 0.02141187286376953, 0.021278688430786133, 0.02128998374938965, 0.021553152084350585, 0.021353471755981446, 0.021283840179443358, 0.02130227279663086, 0.02130636787414551, 0.02143129539489746, 0.021778432846069336, 0.02162380790710449, 0.02126950454711914, 0.021386240005493166, 0.021843967437744142, 0.021591039657592775, 0.021315584182739256, 0.02168012809753418, 0.056130561828613285, 0.021242879867553712, 0.021201919555664063, 0.021363712310791014, 0.021527551651000978, 0.021134336471557616, 0.02102681541442871, 0.02107187271118164, 0.021597183227539063, 0.0214783992767334, 0.021282880783081055, 0.021266368865966796, 0.02126848030090332, 0.0212490234375, 0.021234687805175782, 0.021230592727661132, 0.02127667236328125, 0.02130636787414551, 0.021352447509765626, 0.02127872085571289, 0.021390335083007812, 0.021444608688354492, 0.021352447509765626, 0.021309440612792968, 0.02128281593322754, 0.021323776245117186, 0.02130636787414551, 0.021325824737548828, 0.021337087631225587, 0.021346303939819337, 0.02126950454711914, 0.021326847076416015, 0.021340160369873046, 0.021313535690307618, 0.021296127319335938, 0.021259263992309572, 0.02131455993652344, 0.021983232498168945, 0.021578752517700195, 0.02125721549987793, 0.02128179168701172, 0.021349376678466796, 0.021382144927978516, 0.021325824737548828, 0.02131865692138672, 0.02128691291809082, 0.021331968307495116, 0.021326847076416015, 0.02127462387084961, 0.021308416366577147, 0.02127667236328125, 0.021349376678466796, 0.021326847076416015, 0.02122547149658203, 0.02126848030090332, 0.021954559326171876, 0.022030336380004883, 0.021574655532836915, 0.0213309440612793, 0.021441535949707033, 0.022102048873901367, 0.021575647354125975, 0.021755903244018555, 0.056215553283691405, 0.021843967437744142, 0.02228531265258789, 0.02168115234375, 0.021368831634521485, 0.021710847854614256, 0.021526527404785157, 0.021176319122314453, 0.021214208602905273, 0.02165452766418457, 0.021408767700195314, 0.02104319953918457, 0.02103193664550781, 0.021144575119018554, 0.021178367614746094, 0.020988927841186524, 0.02102681541442871, 0.02102579116821289, 0.021605375289916993, 0.021840896606445313, 0.02122444725036621, 0.021171199798583985, 0.02104729652404785, 0.021506048202514647, 0.021916671752929686, 0.021395456314086913, 0.02168217658996582, 0.021538816452026367, 0.02170163154602051, 0.021506048202514647, 0.020992000579833983, 0.02148044776916504, 0.021710847854614256, 0.021287935256958008, 0.021046272277832033, 0.02104422378540039, 0.020990976333618162, 0.02109644889831543, 0.021285888671875, 0.022010879516601564, 0.021547008514404296, 0.021948415756225585, 0.02198739242553711, 0.021816255569458008, 0.021598207473754884, 0.02126848030090332, 0.02145484733581543, 0.021703680038452147, 0.021735424041748046, 0.02148659133911133, 0.02112512016296387, 0.021238784790039062, 0.02146099281311035, 0.02104422378540039, 0.02122854423522949, 0.021242879867553712, 0.02130227279663086, 0.02105241584777832, 0.021078016281127928, 0.021000192642211913, 0.02109235191345215, 0.02122956848144531, 0.021246976852416992, 0.055943168640136716, 0.02109132766723633, 0.021082111358642578, 0.02102070426940918, 0.021032928466796875, 0.02104422378540039, 0.021506048202514647, 0.021737472534179687, 0.021202943801879884, 0.02109337615966797, 0.021355520248413085, 0.021815296173095702, 0.021771263122558594, 0.021189632415771483, 0.021731327056884766, 0.021550079345703126, 0.021564416885375977, 0.021155839920043946, 0.02106982421875, 0.02104729652404785, 0.02145996856689453, 0.021974016189575195, 0.021375999450683594, 0.021230592727661132, 0.021317632675170898, 0.021607456207275392, 0.021213151931762694, 0.021078016281127928, 0.02105036735534668, 0.02103603172302246, 0.021118976593017577, 0.021346303939819337, 0.021325824737548828, 0.021004287719726563, 0.021045248031616212, 0.02103193664550781, 0.02123161506652832, 0.02111692810058594, 0.021142528533935546, 0.02103500747680664, 0.021044288635253906, 0.021251007080078124, 0.021129215240478515, 0.021046272277832033, 0.021112831115722656, 0.021242879867553712, 0.021361663818359376, 0.02125312042236328, 0.02126540756225586, 0.0212490234375, 0.02128179168701172, 0.021308416366577147, 0.021304351806640625, 0.021446624755859376, 0.021368831634521485, 0.021235712051391603, 0.02127872085571289, 0.021576704025268553, 0.021835775375366212, 0.02129715156555176, 0.02123673629760742, 0.021315584182739256, 0.02140880012512207, 0.05616739273071289, 0.02126131248474121, 0.02145894432067871, 0.02130534362792969, 0.021348384857177733, 0.021228511810302733, 0.02123263931274414, 0.021809215545654296, 0.021246912002563477, 0.021209087371826172, 0.0212490234375, 0.02124595260620117, 0.021533695220947266, 0.021283840179443358, 0.02105036735534668, 0.02101759910583496, 0.02105753517150879, 0.021336063385009766, 0.02132275199890137, 0.02127257537841797, 0.02126540756225586, 0.02131046485900879, 0.021558271408081055, 0.02126848030090332, 0.021293119430541994, 0.021387199401855468, 0.021583871841430666, 0.02127462387084961, 0.02127667236328125, 0.02127359962463379, 0.021300224304199217, 0.02162892723083496, 0.021752832412719726, 0.021279743194580078, 0.021381120681762695, 0.021324800491333007, 0.022137855529785155, 0.02142720031738281, 0.021613567352294923, 0.02125619125366211, 0.021429248809814453, 0.02125823974609375, 0.02126335906982422, 0.021277727127075194, 0.02132374382019043, 0.021317663192749022, 0.021255136489868164, 0.021337087631225587, 0.021235712051391603, 0.021543935775756837, 0.02125823974609375, 0.0212490234375, 0.02146713638305664, 0.021336063385009766, 0.02163199996948242, 0.02172313690185547, 0.02125823974609375, 0.0212674560546875, 0.021343231201171875, 0.021354496002197267, 0.022286336898803712, 0.021815296173095702, 0.021078016281127928, 0.05608652877807617, 0.02123366355895996, 0.021243904113769533, 0.021344255447387696, 0.02128281593322754, 0.021098495483398438, 0.02103603172302246, 0.020993024826049804, 0.0210565128326416, 0.02123263931274414, 0.02106879997253418, 0.02112006378173828, 0.02122540855407715, 0.021175359725952147, 0.021110719680786132, 0.021013504028320314, 0.021016576766967773, 0.02104115104675293, 0.02142207908630371, 0.021167104721069335, 0.02106265640258789, 0.02101759910583496, 0.02105036735534668, 0.02105446434020996, 0.021016576766967773, 0.021012479782104493, 0.021038143157958985, 0.021285823822021484, 0.021600255966186522, 0.02124083137512207, 0.021130239486694336, 0.02106470489501953, 0.02148659133911133, 0.021191680908203125, 0.02129715156555176, 0.021136383056640624, 0.021172224044799806, 0.021572608947753907, 0.021213184356689452, 0.02109235191345215, 0.02107699203491211, 0.02128281593322754, 0.021521408081054686, 0.021321727752685548, 0.0210565128326416, 0.02106265640258789, 0.021334016799926758, 0.021308416366577147, 0.021106687545776368, 0.02109235191345215, 0.02101759910583496, 0.021126144409179686, 0.021061632156372072, 0.02307689666748047, 0.022600736618041992, 0.022022079467773438, 0.02145075225830078, 0.02128895950317383, 0.021338111877441408, 0.0212992000579834, 0.02165657615661621, 0.021303295135498047, 0.021315616607666017]",tokens/s,45.582157220910624,,,main,False,False -float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9817,7 +9817,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpoid6p7mm/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10110,7 +10110,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp9dwv_c8j/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10626,7 +10626,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1300.873216,1713.897472,0.0,1067.450368,1018.226688,s,10,0.16997203254699708,0.016997203254699708,0.00041507905486039146,0.016764400482177734,0.017378297042846682,0.017702124214172364,0.017961185951232913,"[0.01802595138549805, 0.01674083137512207, 0.016743040084838866, 0.017091136932373047, 0.01730633544921875, 0.016676704406738282, 0.017260576248168947, 0.016785760879516602, 0.01666521644592285, 0.01667647933959961]",tokens/s,15061.30133080666,kWh,2.0177843848050524e-07,1.1056505171908154e-07,5.087115370409696e-07,8.210550272405565e-07,tokens/kWh,311793962.0446364,MB,1300.873216,1713.897472,0.0,1067.450368,1059.39456,s,10,10.173291931152344,1.0173291931152344,0.011496006145209937,1.0139341125488281,1.0373599243164062,1.0384082946777344,1.0392469909667967,"[1.0083255615234374, 1.0059702758789062, 1.0048641357421875, 1.0131695556640625, 1.037126953125, 1.0214266357421875, 1.0394566650390624, 1.0166776733398437, 1.0146986694335938, 1.0115758056640625]",tokens/s,61.9268575268968,kWh,1.2068471352134998e-05,6.613018416964417e-06,2.1189972488558242e-05,3.987146225765766e-05,tokens/kWh,1580077.489831723,,s,629,10.308060156822195,0.016388012967920834,0.0020900941940995166,0.015986687660217287,0.01648599052429199,0.01668833236694336,0.03302572143554688,"[0.01744486427307129, 0.016336896896362304, 0.015946751594543458, 0.015913984298706055, 0.016058368682861326, 0.016199680328369142, 0.016114688873291014, 0.015892479896545412, 0.015848447799682617, 0.015821824073791502, 0.01581984043121338, 0.015839167594909668, 0.016021503448486327, 0.016135168075561524, 0.016210943222045898, 0.01589145565032959, 0.0158341121673584, 0.015820799827575685, 0.015887359619140624, 0.015841376304626464, 0.015886240005493164, 0.015796223640441896, 0.015760448455810545, 0.015928256034851075, 0.01587718391418457, 0.015871935844421387, 0.015812607765197755, 0.015836159706115724, 0.015807488441467284, 0.015812607765197755, 0.015892479896545412, 0.01581670379638672, 0.015814656257629393, 0.015789055824279787, 0.01578700828552246, 0.016380928039550782, 0.016539648056030275, 0.01658367919921875, 0.017051647186279297, 0.016524288177490236, 0.01601535987854004, 0.01586796760559082, 0.015919103622436523, 0.015910847663879395, 0.01580134391784668, 0.015847423553466796, 0.015970303535461427, 0.015848511695861815, 0.015943615913391114, 0.015871999740600586, 0.01589145565032959, 0.01584537601470947, 0.015911935806274414, 0.015930368423461915, 0.015904767990112305, 0.015805503845214845, 0.015861696243286132, 0.01581670379638672, 0.015857664108276368, 0.01582592010498047, 0.01593446445465088, 0.015844351768493654, 0.03305472183227539, 0.01598361587524414, 0.015930368423461915, 0.015981568336486816, 0.015898624420166017, 0.01592627239227295, 0.015904767990112305, 0.0158853120803833, 0.015904767990112305, 0.01590681552886963, 0.01589350414276123, 0.01586892795562744, 0.015882240295410157, 0.015831040382385253, 0.015969280242919923, 0.015846400260925292, 0.015888383865356445, 0.01587712001800537, 0.015851519584655763, 0.015840255737304687, 0.015890432357788087, 0.01590169620513916, 0.015863807678222656, 0.015861824035644533, 0.015911871910095216, 0.01587097644805908, 0.015922176361083985, 0.015895551681518554, 0.015884287834167482, 0.015824895858764648, 0.015873023986816406, 0.015949824333190917, 0.015952896118164063, 0.015876095771789552, 0.015887359619140624, 0.016089088439941408, 0.01638297653198242, 0.01621811294555664, 0.015873023986816406, 0.015981568336486816, 0.015871999740600586, 0.016227327346801757, 0.016313343048095702, 0.016475135803222657, 0.015997952461242675, 0.015935487747192383, 0.016076799392700195, 0.01590681552886963, 0.015913984298706055, 0.015866880416870118, 0.015882240295410157, 0.015873023986816406, 0.01587814426422119, 0.01590681552886963, 0.015932415962219237, 0.015930368423461915, 0.0159334716796875, 0.016533472061157226, 0.01647315216064453, 0.016004032135009765, 0.01597542381286621, 0.01590783977508545, 0.015903743743896484, 0.033363967895507815, 0.0163768310546875, 0.015852543830871584, 0.015987711906433104, 0.01642803192138672, 0.016021503448486327, 0.01583513641357422, 0.015953920364379884, 0.015805439949035646, 0.015904767990112305, 0.015859711647033693, 0.015904767990112305, 0.016062463760375977, 0.01621708869934082, 0.0158341121673584, 0.015898624420166017, 0.01588326358795166, 0.01589350414276123, 0.015833087921142578, 0.015866880416870118, 0.015824895858764648, 0.015881216049194336, 0.01590169620513916, 0.015905792236328126, 0.015905792236328126, 0.015827967643737794, 0.015866880416870118, 0.01589964771270752, 0.015977472305297852, 0.015865856170654297, 0.01586278438568115, 0.015864831924438477, 0.015900671958923338, 0.01590681552886963, 0.01618124771118164, 0.01636966323852539, 0.016339967727661133, 0.015910911560058593, 0.015905792236328126, 0.015904767990112305, 0.015900671958923338, 0.015924223899841307, 0.01594777584075928, 0.01594268798828125, 0.016163808822631836, 0.016119808197021485, 0.015917056083679198, 0.015946751594543458, 0.01590176010131836, 0.015846336364746094, 0.015889408111572266, 0.015865856170654297, 0.015879167556762695, 0.015838208198547362, 0.01585868835449219, 0.015836159706115724, 0.015851519584655763, 0.015856639862060547, 0.01586995220184326, 0.015879167556762695, 0.015859711647033693, 0.01586995220184326, 0.01588633632659912, 0.03381043243408203, 0.015919103622436523, 0.015846400260925292, 0.015859711647033693, 0.015852543830871584, 0.015849472045898438, 0.01581875228881836, 0.01585868835449219, 0.015823871612548827, 0.015781951904296876, 0.015829952239990234, 0.015831040382385253, 0.01594777584075928, 0.01582592010498047, 0.015856639862060547, 0.01582694435119629, 0.01582592010498047, 0.01581158447265625, 0.0158341121673584, 0.0158023681640625, 0.015873023986816406, 0.015848447799682617, 0.015857664108276368, 0.015849504470825195, 0.01609622383117676, 0.01661747169494629, 0.01682636833190918, 0.0162857608795166, 0.01607468795776367, 0.016063488006591797, 0.016004095077514647, 0.015944704055786133, 0.016004095077514647, 0.01597644805908203, 0.015856639862060547, 0.015807488441467284, 0.015742976188659667, 0.01567743968963623, 0.015900671958923338, 0.01599078369140625, 0.015953920364379884, 0.015949824333190917, 0.015921152114868165, 0.016261119842529297, 0.016463872909545898, 0.016734207153320312, 0.016343040466308592, 0.016358400344848634, 0.016352256774902343, 0.01640959930419922, 0.01637478446960449, 0.01639116859436035, 0.01647724723815918, 0.016352191925048828, 0.01637785530090332, 0.01636147117614746, 0.016453632354736326, 0.016390144348144533, 0.016257024765014647, 0.015841279983520508, 0.015847423553466796, 0.01684377670288086, 0.016446464538574217, 0.034802688598632815, 0.01747148895263672, 0.017386495590209963, 0.016472063064575194, 0.016440319061279296, 0.016278528213500978, 0.016363519668579102, 0.016508928298950197, 0.016909311294555664, 0.016881664276123046, 0.016772096633911132, 0.016570367813110352, 0.01643110466003418, 0.01637171173095703, 0.01660927963256836, 0.01641881561279297, 0.01638400077819824, 0.016381952285766603, 0.01640345573425293, 0.016406591415405274, 0.016439231872558593, 0.016451583862304688, 0.016689151763916017, 0.016525312423706053, 0.016455680847167968, 0.016348159790039063, 0.01639423942565918, 0.016338943481445312, 0.01637068748474121, 0.016360448837280273, 0.016442367553710938, 0.01640959930419922, 0.01640243148803711, 0.01638400077819824, 0.01640140724182129, 0.01641574478149414, 0.016481279373168945, 0.016356351852416993, 0.016336896896362304, 0.016670719146728515, 0.01656524848937988, 0.016373760223388673, 0.016347135543823242, 0.01637785530090332, 0.0164136962890625, 0.016408575057983397, 0.016492544174194337, 0.016446464538574217, 0.016372735977172852, 0.016473087310791015, 0.0163768310546875, 0.016366592407226564, 0.016350208282470705, 0.01637068748474121, 0.016326656341552736, 0.016408575057983397, 0.016105472564697267, 0.015841279983520508, 0.01703424072265625, 0.0163768310546875, 0.01599897575378418, 0.015867903709411622, 0.01584332847595215, 0.033031169891357424, 0.015859711647033693, 0.015847423553466796, 0.01587712001800537, 0.015896575927734375, 0.016067583084106444, 0.016114688873291014, 0.016885759353637696, 0.017666048049926757, 0.016687103271484375, 0.01639423942565918, 0.01591808032989502, 0.015927295684814453, 0.015881216049194336, 0.01590988826751709, 0.015823871612548827, 0.015864831924438477, 0.01584537601470947, 0.015863807678222656, 0.01581158447265625, 0.01639833641052246, 0.016430080413818358, 0.01615667152404785, 0.01642803192138672, 0.01642905616760254, 0.016336896896362304, 0.01636966323852539, 0.016368640899658202, 0.01636966323852539, 0.01637887954711914, 0.016314367294311523, 0.016397312164306642, 0.01638604736328125, 0.016434175491333008, 0.01640550422668457, 0.01641472053527832, 0.01637478446960449, 0.01660927963256836, 0.01644339179992676, 0.0164003849029541, 0.01640447998046875, 0.01641574478149414, 0.0158853120803833, 0.01583513641357422, 0.01588326358795166, 0.01599180793762207, 0.015815679550170898, 0.015814656257629393, 0.01581056022644043, 0.015908864021301268, 0.01590783977508545, 0.015796223640441896, 0.015884287834167482, 0.015856639862060547, 0.0158535680770874, 0.015968255996704102, 0.017472511291503907, 0.016491519927978517, 0.01640959930419922, 0.01640755271911621, 0.01641983985900879, 0.016455680847167968, 0.01642291259765625, 0.03405516815185547, 0.016364543914794923, 0.016424959182739257, 0.016668672561645507, 0.016679935455322266, 0.016359424591064452, 0.016392223358154295, 0.016350175857543944, 0.016372735977172852, 0.01659699249267578, 0.016424959182739257, 0.017356800079345702, 0.01657651138305664, 0.01641983985900879, 0.01640959930419922, 0.016387136459350585, 0.01634911918640137, 0.0164136962890625, 0.016464895248413085, 0.01636761665344238, 0.016377920150756835, 0.016424896240234375, 0.01653555107116699, 0.01640959930419922, 0.016472095489501952, 0.016417760848999024, 0.01639833641052246, 0.016373760223388673, 0.01637478446960449, 0.016448511123657226, 0.016454687118530275, 0.016370655059814453, 0.01645260810852051, 0.016501760482788085, 0.016488447189331054, 0.01638502311706543, 0.016447488784790038, 0.016438272476196288, 0.016342016220092775, 0.016408575057983397, 0.016338943481445312, 0.01641881561279297, 0.01643622398376465, 0.01643212890625, 0.016480255126953124, 0.01642291259765625, 0.01641062355041504, 0.016442367553710938, 0.01640652847290039, 0.016451583862304688, 0.016447488784790038, 0.016373760223388673, 0.016373760223388673, 0.017163263320922852, 0.016485376358032225, 0.016517120361328123, 0.01640550422668457, 0.016358400344848634, 0.01760972785949707, 0.01721139144897461, 0.016556032180786134, 0.016546815872192384, 0.016497663497924805, 0.033555454254150394, 0.016479232788085937, 0.0164136962890625, 0.016376895904541014, 0.016451520919799803, 0.016368640899658202, 0.01642393684387207, 0.016442367553710938, 0.0163656005859375, 0.016312288284301757, 0.016305152893066405, 0.01640345573425293, 0.01640243148803711, 0.01643929672241211, 0.01641062355041504, 0.01639833641052246, 0.01643622398376465, 0.01641164779663086, 0.01640345573425293, 0.016373760223388673, 0.01638502311706543, 0.01638707160949707, 0.01643622398376465, 0.01623859214782715, 0.015879167556762695, 0.015871999740600586, 0.01587814426422119, 0.015910911560058593, 0.01587814426422119, 0.015871999740600586, 0.01581670379638672, 0.01624678421020508, 0.016442367553710938, 0.016408575057983397, 0.015884287834167482, 0.015859711647033693, 0.015874048233032227, 0.015831040382385253, 0.015827967643737794, 0.01583619213104248, 0.015856608390808106, 0.015917087554931642, 0.015835103988647462, 0.015765503883361818, 0.015857664108276368, 0.015846400260925292, 0.015857664108276368, 0.01586899185180664, 0.01583199977874756, 0.015874048233032227, 0.016705535888671876, 0.016446464538574217, 0.01639321517944336, 0.01639219284057617, 0.016257024765014647, 0.01580339241027832, 0.015876095771789552, 0.015844351768493654, 0.01581772804260254, 0.01578598403930664, 0.015866880416870118, 0.015904767990112305, 0.01596723175048828, 0.03291648101806641, 0.015790143966674806, 0.015886272430419922, 0.01589452838897705, 0.015684608459472657, 0.015681568145751952, 0.015838175773620605, 0.01581670379638672, 0.01580851173400879, 0.015831040382385253, 0.015847423553466796, 0.015782912254333495, 0.015822848320007323, 0.015846400260925292, 0.01586892795562744, 0.01586176013946533, 0.015824895858764648, 0.015857664108276368, 0.01591808032989502, 0.015889408111572266, 0.015807488441467284, 0.015851519584655763, 0.015840288162231445, 0.01582588768005371, 0.016066560745239256, 0.016304128646850585, 0.0165980167388916, 0.01637171173095703, 0.016334848403930666, 0.016306175231933593, 0.01659187126159668, 0.016390144348144533, 0.016359487533569337, 0.016424896240234375, 0.016331775665283203, 0.016330751419067382, 0.01640345573425293, 0.016343040466308592, 0.016322559356689453, 0.016335872650146483, 0.016306175231933593, 0.016330751419067382, 0.016333856582641602, 0.01640547180175781, 0.01639936065673828, 0.016360448837280273, 0.01588633632659912, 0.01585459232330322, 0.015822848320007323, 0.017108991622924806, 0.01638707160949707, 0.016350208282470705, 0.0164003849029541, 0.016365568161010743, 0.016330751419067382, 0.015881216049194336, 0.01591603183746338, 0.015773695945739748, 0.015775744438171386, 0.01582694435119629, 0.015986687660217287, 0.016442367553710938, 0.016323583602905273, 0.03301171112060547, 0.015797247886657716, 0.015806464195251464, 0.015830016136169432, 0.015855615615844726, 0.015827967643737794, 0.015945728302001954, 0.01584537601470947, 0.015781888008117674, 0.015758336067199705, 0.01580953598022461, 0.016357376098632814, 0.01631545639038086, 0.01631430435180664, 0.016352256774902343, 0.016355327606201172, 0.01638502311706543, 0.016334848403930666, 0.016349184036254884, 0.016318464279174806, 0.016357376098632814, 0.015880191802978515, 0.015863807678222656, 0.01584332847595215, 0.01582592010498047, 0.015839232444763183, 0.01581670379638672, 0.015863807678222656, 0.015779840469360353, 0.015789055824279787, 0.01576140785217285, 0.015783935546875, 0.015783935546875, 0.015852543830871584, 0.015755264282226563, 0.016309247970581055, 0.016299007415771484, 0.0158023681640625, 0.015807488441467284, 0.015817791938781738, 0.01579206371307373, 0.015832063674926757, 0.01579520034790039, 0.015775744438171386, 0.01588633632659912, 0.015814656257629393, 0.01580851173400879, 0.016760831832885743, 0.016493568420410155, 0.01589145565032959, 0.0158341121673584, 0.017081344604492187, 0.016508928298950197, 0.016363519668579102, 0.016300031661987305, 0.016297983169555663, 0.016391231536865235, 0.01630611228942871, 0.016351264953613283, 0.01651606369018555, 0.016479232788085937, 0.01593958377838135, 0.01586892795562744]",tokens/s,61.02021044024541,,,main,False,False -float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10888,7 +10888,7 @@ ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please req ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1175.089152,1455.947776,0.0,809.500672,723.533824,s,15,0.17660559940338133,0.011773706626892089,0.0005060809352526505,0.01188105583190918,0.011992940521240234,0.012405353450775146,0.013093141155242918,"[0.013265088081359863, 0.011219743728637696, 0.011262335777282716, 0.011187264442443848, 0.011892576217651368, 0.01188105583190918, 0.011240384101867677, 0.012036895751953125, 0.011927007675170898, 0.011889760017395019, 0.011867008209228515, 0.011883199691772461, 0.011868415832519531, 0.011281344413757324, 0.01190351963043213]",tokens/s,21743.364949766583,kWh,1.3865128598706948e-07,7.597172755684907e-08,3.956406399342899e-07,6.102636534782083e-07,tokens/kWh,419490819.32198244,MB,1175.089152,1455.947776,0.0,809.500672,749.216256,s,15,10.484396118164064,0.6989597412109376,0.010546266198019412,0.6962911987304687,0.7025856811523438,0.7139740173339844,0.7317457800292969,"[0.736188720703125, 0.6892533569335938, 0.69624951171875, 0.6921040649414063, 0.6986005249023437, 0.698226806640625, 0.6946048583984376, 0.6997840576171875, 0.6971741943359375, 0.6962911987304687, 0.694924072265625, 0.6990907592773438, 0.6946989135742188, 0.6927516479492187, 0.7044534301757812]",tokens/s,90.13394661451233,kWh,7.764596984365612e-06,4.254614382124614e-06,1.4961130904734307e-05,2.6980342271224535e-05,tokens/kWh,2335033.3871483784,,s,944,10.635618295669536,0.011266544804734699,0.0014591324737404984,0.01107148838043213,0.011258982372283935,0.011755068683624267,0.02223625762939453,"[0.012138496398925782, 0.012057600021362304, 0.011867136001586913, 0.01165721607208252, 0.011790335655212402, 0.011788288116455077, 0.011667455673217773, 0.011686911582946777, 0.012061696052551269, 0.011962368011474609, 0.011684864044189454, 0.011777024269104003, 0.011752511978149414, 0.01197868824005127, 0.011803647994995118, 0.012000255584716797, 0.011875328063964843, 0.011687935829162598, 0.011875328063964843, 0.01174937629699707, 0.011942912101745605, 0.011983872413635254, 0.011774016380310058, 0.0118721923828125, 0.01233407974243164, 0.011925503730773926, 0.011716608047485352, 0.011866111755371094, 0.011828224182128906, 0.011709440231323242, 0.011905023574829102, 0.01174835205078125, 0.011743231773376465, 0.012090368270874024, 0.012113920211791992, 0.01176576042175293, 0.011701312065124512, 0.011774911880493164, 0.011737088203430175, 0.011653120040893555, 0.011552767753601074, 0.011234304428100587, 0.011653120040893555, 0.011614208221435546, 0.011580415725708008, 0.01183027172088623, 0.011583488464355468, 0.011491328239440919, 0.011707391738891602, 0.011467776298522948, 0.011268095970153809, 0.01115340805053711, 0.011082752227783203, 0.011146240234375, 0.011140095710754394, 0.011147263526916504, 0.011131903648376466, 0.01115340805053711, 0.011222016334533692, 0.011147263526916504, 0.011168800354003906, 0.01108784008026123, 0.02211942481994629, 0.010397695541381835, 0.010449919700622558, 0.010497023582458496, 0.010392576217651368, 0.010460160255432128, 0.010432512283325195, 0.01045094394683838, 0.010445823669433594, 0.010406911849975586, 0.010437631607055664, 0.010470399856567383, 0.01042636775970459, 0.010480640411376953, 0.010445823669433594, 0.010461183547973632, 0.010458111763000488, 0.010431488037109376, 0.010424320220947265, 0.01043455982208252, 0.010455039978027344, 0.01093017578125, 0.011171839714050292, 0.011189248085021973, 0.011116543769836425, 0.011207679748535156, 0.011160575866699218, 0.011148287773132324, 0.011110400199890137, 0.011133952140808106, 0.011155455589294434, 0.011123711585998536, 0.011090944290161133, 0.011088895797729491, 0.011189248085021973, 0.011101183891296386, 0.011117568016052246, 0.011094016075134277, 0.011119615554809571, 0.011193344116210938, 0.01113804817199707, 0.011147263526916504, 0.011320320129394532, 0.011206656455993653, 0.01112166404724121, 0.011126784324645997, 0.011096063613891602, 0.01107148838043213, 0.011132927894592285, 0.011144191741943359, 0.011135999679565429, 0.011383808135986329, 0.01175551986694336, 0.01153331184387207, 0.011256832122802735, 0.011196415901184082, 0.0111595516204834, 0.011420672416687011, 0.011188223838806152, 0.011104255676269532, 0.01116262435913086, 0.01116262435913086, 0.011131903648376466, 0.02213580894470215, 0.010478591918945313, 0.010484736442565918, 0.011142144203186035, 0.011163647651672364, 0.011189248085021973, 0.011179007530212403, 0.011090944290161133, 0.011057151794433593, 0.01115443229675293, 0.011074560165405273, 0.011171839714050292, 0.01111244773864746, 0.011139072418212891, 0.011106304168701172, 0.011200511932373047, 0.011057151794433593, 0.010911744117736816, 0.011289600372314454, 0.010860544204711914, 0.01103052806854248, 0.011191295623779298, 0.011142144203186035, 0.011233280181884766, 0.01112377643585205, 0.011103167533874511, 0.011147263526916504, 0.011070528030395509, 0.011084735870361328, 0.011485183715820312, 0.011078656196594238, 0.011074560165405273, 0.011172863960266113, 0.011123711585998536, 0.01103667163848877, 0.011100159645080567, 0.011033599853515624, 0.011095040321350098, 0.01107148838043213, 0.010899456024169921, 0.010878975868225099, 0.010903552055358886, 0.010842111587524414, 0.010879008293151856, 0.010818528175354005, 0.01083084774017334, 0.010789888381958008, 0.01083903980255127, 0.010860544204711914, 0.011130880355834961, 0.011092991828918456, 0.011107328414916993, 0.01112166404724121, 0.011177984237670899, 0.011076607704162598, 0.011125760078430176, 0.011097087860107421, 0.011111424446105958, 0.0111278076171875, 0.01112883186340332, 0.011160575866699218, 0.011094016075134277, 0.0110448637008667, 0.02204979133605957, 0.010391551971435547, 0.01081651210784912, 0.011019264221191406, 0.01102847957611084, 0.011114496231079102, 0.011080703735351562, 0.011111424446105958, 0.011027520179748534, 0.01104684829711914, 0.011084799766540527, 0.011072544097900391, 0.011072480201721191, 0.011060223579406739, 0.011164671897888183, 0.011094016075134277, 0.011081727981567382, 0.011073535919189453, 0.011040767669677735, 0.011068415641784669, 0.011085824012756347, 0.011061247825622558, 0.011024383544921875, 0.011081727981567382, 0.010852352142333984, 0.010819583892822266, 0.010806271553039551, 0.010788864135742187, 0.010807295799255372, 0.010853407859802246, 0.011113439559936523, 0.01116262435913086, 0.011104255676269532, 0.011017215728759766, 0.011003904342651367, 0.011041791915893554, 0.011025407791137695, 0.011066368103027344, 0.0110448637008667, 0.01103769588470459, 0.01102950382232666, 0.011049983978271484, 0.011041791915893554, 0.01104691219329834, 0.011077631950378418, 0.011032575607299805, 0.011010047912597656, 0.010844160079956054, 0.010813440322875977, 0.010787839889526368, 0.010875904083251953, 0.011091967582702637, 0.011053055763244628, 0.011053055763244628, 0.011111424446105958, 0.01101414394378662, 0.010912768363952637, 0.010781696319580078, 0.010753024101257324, 0.010797056198120117, 0.01083187198638916, 0.011004927635192872, 0.011039744377136231, 0.023410688400268553, 0.011052032470703126, 0.01105510425567627, 0.011003904342651367, 0.011023360252380371, 0.011087871551513672, 0.011041791915893554, 0.011075584411621094, 0.011066399574279785, 0.011058143615722656, 0.011043840408325196, 0.011074560165405273, 0.011161600112915039, 0.011259903907775879, 0.01107968044281006, 0.011074560165405273, 0.011040767669677735, 0.011058176040649414, 0.011141119956970215, 0.011090944290161133, 0.011088895797729491, 0.011089920043945312, 0.011066368103027344, 0.01115443229675293, 0.01107968044281006, 0.01114521598815918, 0.011061247825622558, 0.011072511672973634, 0.011019264221191406, 0.01107049560546875, 0.011074527740478516, 0.011058176040649414, 0.011052032470703126, 0.011066368103027344, 0.010970111846923827, 0.011059200286865235, 0.01102847957611084, 0.011056127548217774, 0.011078656196594238, 0.01115135955810547, 0.011070464134216309, 0.0111278076171875, 0.011064319610595704, 0.01104691219329834, 0.011062272071838379, 0.011057151794433593, 0.011061247825622558, 0.011083776473999024, 0.011062272071838379, 0.011045887947082519, 0.011051008224487305, 0.011097087860107421, 0.011019264221191406, 0.011114496231079102, 0.011066368103027344, 0.011139072418212891, 0.011181056022644043, 0.011150336265563965, 0.011089920043945312, 0.011372544288635255, 0.011087871551513672, 0.011062272071838379, 0.011053055763244628, 0.023417856216430662, 0.011049983978271484, 0.011092991828918456, 0.01105510425567627, 0.011049983978271484, 0.011139072418212891, 0.011018239974975585, 0.011054080009460449, 0.011051008224487305, 0.01103052806854248, 0.011158592224121093, 0.011132863998413086, 0.011143168449401856, 0.011181056022644043, 0.01115135955810547, 0.011072511672973634, 0.011059200286865235, 0.011190272331237794, 0.01112063980102539, 0.0110448637008667, 0.011105279922485351, 0.011070464134216309, 0.011113471984863281, 0.011107328414916993, 0.011034624099731445, 0.011074560165405273, 0.011105279922485351, 0.011081727981567382, 0.011089920043945312, 0.011096063613891602, 0.011076607704162598, 0.011095040321350098, 0.011051008224487305, 0.011147263526916504, 0.011059200286865235, 0.011070464134216309, 0.01103667163848877, 0.011065343856811523, 0.011130880355834961, 0.011054080009460449, 0.011077631950378418, 0.011048959732055665, 0.011065343856811523, 0.011052032470703126, 0.011113471984863281, 0.011072511672973634, 0.011115519523620606, 0.011100159645080567, 0.011015168190002441, 0.011144191741943359, 0.0110448637008667, 0.011076607704162598, 0.011132927894592285, 0.011057151794433593, 0.011021311759948731, 0.010954751968383789, 0.010874879837036134, 0.01084620761871338, 0.011087871551513672, 0.011068415641784669, 0.011119615554809571, 0.01111244773864746, 0.011054080009460449, 0.022202367782592772, 0.010402815818786621, 0.010449919700622558, 0.011032575607299805, 0.011088895797729491, 0.01104691219329834, 0.011006976127624512, 0.011085824012756347, 0.011072511672973634, 0.011025407791137695, 0.011027456283569336, 0.01103667163848877, 0.011070464134216309, 0.011086848258972168, 0.010819583892822266, 0.010833984375, 0.01081439971923828, 0.01095372772216797, 0.01083187198638916, 0.011095104217529296, 0.011084735870361328, 0.0110448637008667, 0.01112985610961914, 0.011302911758422851, 0.011109375953674316, 0.01101414394378662, 0.011039744377136231, 0.011113471984863281, 0.011100159645080567, 0.011259903907775879, 0.011041791915893554, 0.011059200286865235, 0.011110400199890137, 0.011118592262268067, 0.011064319610595704, 0.011016223907470703, 0.011090911865234376, 0.011254783630371093, 0.011076607704162598, 0.011010047912597656, 0.01116262435913086, 0.011091967582702637, 0.011072511672973634, 0.011025407791137695, 0.0110448637008667, 0.01107148838043213, 0.011069439888000488, 0.011080767631530762, 0.011090880393981933, 0.011061247825622558, 0.01092403221130371, 0.010891263961791992, 0.010870783805847169, 0.010827775955200195, 0.01084108829498291, 0.011094016075134277, 0.011086879730224609, 0.011106271743774414, 0.011158528327941895, 0.011142144203186035, 0.011083776473999024, 0.011052032470703126, 0.01102847957611084, 0.023323648452758788, 0.011080767631530762, 0.010983455657958984, 0.011079584121704102, 0.011096063613891602, 0.011982848167419433, 0.011134976387023926, 0.010999808311462403, 0.011023360252380371, 0.011230208396911622, 0.011049983978271484, 0.011078656196594238, 0.010974207878112792, 0.011049983978271484, 0.011017215728759766, 0.01213644790649414, 0.011421695709228515, 0.01214566421508789, 0.011182080268859864, 0.011249664306640626, 0.011106304168701172, 0.011099136352539063, 0.011016192436218262, 0.010991616249084473, 0.011004927635192872, 0.010983424186706543, 0.01100595188140869, 0.011000831604003907, 0.011051008224487305, 0.011083776473999024, 0.01101414394378662, 0.010982399940490722, 0.011083776473999024, 0.011109375953674316, 0.01105510425567627, 0.011027456283569336, 0.010991616249084473, 0.010999808311462403, 0.01102233600616455, 0.011200511932373047, 0.011125760078430176, 0.011000831604003907, 0.011080703735351562, 0.011008000373840332, 0.011061247825622558, 0.010998784065246582, 0.010796031951904296, 0.010761216163635253, 0.010984448432922364, 0.011100159645080567, 0.0111278076171875, 0.01108176040649414, 0.011106271743774414, 0.011104255676269532, 0.011049983978271484, 0.010997759819030761, 0.011039744377136231, 0.011043840408325196, 0.011081727981567382, 0.011049983978271484, 0.011004927635192872, 0.011086848258972168, 0.011081727981567382, 0.02349772834777832, 0.010971136093139648, 0.011004927635192872, 0.011001855850219726, 0.011040767669677735, 0.011041791915893554, 0.010982399940490722, 0.011059200286865235, 0.011111424446105958, 0.011010047912597656, 0.011004927635192872, 0.011035648345947266, 0.011043904304504394, 0.011024319648742675, 0.010991616249084473, 0.011039744377136231, 0.011060223579406739, 0.011073535919189453, 0.011016192436218262, 0.011066399574279785, 0.011066335678100586, 0.011020288467407227, 0.010982399940490722, 0.011085824012756347, 0.011023360252380371, 0.011041791915893554, 0.01103052806854248, 0.011102208137512207, 0.011045887947082519, 0.011048959732055665, 0.011117568016052246, 0.011016192436218262, 0.011064319610595704, 0.011072511672973634, 0.01103769588470459, 0.011317248344421387, 0.011150336265563965, 0.011147263526916504, 0.011049983978271484, 0.011048959732055665, 0.010994688034057617, 0.011033599853515624, 0.011058176040649414, 0.011073535919189453, 0.011021311759948731, 0.011062272071838379, 0.011096063613891602, 0.011086848258972168, 0.01101414394378662, 0.01102233600616455, 0.01104793643951416, 0.01103667163848877, 0.01103872013092041, 0.011034624099731445, 0.01122815990447998, 0.01126912021636963, 0.011086848258972168, 0.011060223579406739, 0.011107328414916993, 0.011076607704162598, 0.01103052806854248, 0.011092991828918456, 0.011011072158813476, 0.023358463287353515, 0.010990592002868652, 0.011009023666381837, 0.011004927635192872, 0.011003904342651367, 0.010980352401733399, 0.010906623840332032, 0.011048959732055665, 0.011114496231079102, 0.011043840408325196, 0.010972160339355469, 0.010981375694274903, 0.011039744377136231, 0.011023360252380371, 0.010992639541625977, 0.010978303909301757, 0.010966015815734862, 0.011045887947082519, 0.011019264221191406, 0.01096396827697754, 0.011002880096435547, 0.011033599853515624, 0.011007040023803712, 0.011027392387390136, 0.011003904342651367, 0.01103052806854248, 0.01103052806854248, 0.011103232383728028, 0.011019264221191406, 0.011002880096435547, 0.010986559867858887, 0.010992575645446777, 0.011026432037353515, 0.011149312019348144, 0.011064319610595704, 0.01103769588470459, 0.011196415901184082, 0.011034624099731445, 0.01100595188140869, 0.01100595188140869, 0.011006976127624512, 0.01102950382232666, 0.01103667163848877, 0.011004927635192872, 0.011056127548217774, 0.011084799766540527, 0.011019264221191406, 0.011095040321350098, 0.011042816162109375, 0.011066368103027344, 0.011100159645080567, 0.011077664375305175, 0.011090911865234376, 0.011150336265563965, 0.01111244773864746, 0.011092991828918456, 0.011107328414916993, 0.011163647651672364, 0.011117568016052246, 0.011105279922485351, 0.011048959732055665, 0.011288576126098633, 0.01112166404724121, 0.023423999786376954, 0.011000831604003907, 0.010979328155517578, 0.011001855850219726, 0.011026432037353515, 0.011033599853515624, 0.010976256370544434, 0.010976256370544434, 0.01102847957611084, 0.011041791915893554, 0.011056127548217774, 0.01105510425567627, 0.010995712280273438, 0.010989567756652833, 0.01100595188140869, 0.010988544464111329, 0.010991616249084473, 0.011126784324645997, 0.011016192436218262, 0.010951711654663085, 0.011007967948913575, 0.011049983978271484, 0.010992639541625977, 0.010916864395141602, 0.010827775955200195, 0.010815520286560059, 0.01099465560913086, 0.011035648345947266, 0.010984448432922364, 0.011013119697570802, 0.01095680046081543, 0.010994688034057617, 0.01102847957611084, 0.011065376281738282, 0.011089887619018555, 0.011064319610595704, 0.011024383544921875, 0.011042816162109375, 0.011025407791137695, 0.010976256370544434, 0.011019264221191406, 0.01104691219329834, 0.011012096405029297, 0.011039744377136231, 0.011031552314758301, 0.011027456283569336, 0.011076640129089356, 0.01110422420501709, 0.011056127548217774, 0.01102847957611084, 0.011065343856811523, 0.011042816162109375, 0.011115519523620606, 0.011028544425964355, 0.011113408088684083, 0.011076607704162598, 0.01105510425567627, 0.011061280250549317, 0.01109603214263916, 0.011019264221191406, 0.011023360252380371, 0.011065343856811523, 0.011039744377136231, 0.023407615661621094, 0.010978303909301757, 0.010981375694274903, 0.011058176040649414, 0.011073535919189453, 0.011033599853515624, 0.0109486083984375, 0.011002911567687988, 0.010992608070373534, 0.010993663787841796, 0.011011072158813476, 0.0110632963180542, 0.011070464134216309, 0.011092991828918456, 0.011156479835510253, 0.01144217586517334, 0.011598848342895507, 0.011254783630371093, 0.011147263526916504, 0.011066368103027344, 0.011118592262268067, 0.011103232383728028, 0.01112063980102539, 0.011068415641784669, 0.01103052806854248, 0.011039744377136231, 0.011169792175292969, 0.011075615882873535, 0.011022303581237793, 0.011060223579406739, 0.011160575866699218, 0.011082752227783203, 0.011065343856811523, 0.011053055763244628, 0.011033599853515624, 0.011076607704162598, 0.011049983978271484, 0.011026432037353515, 0.011016223907470703, 0.011038687705993653, 0.011024383544921875, 0.011085824012756347, 0.01111251163482666, 0.011107263565063476, 0.011247615814208984, 0.011100159645080567, 0.011042816162109375, 0.011062335968017579, 0.011111359596252441, 0.011152383804321288, 0.011080703735351562, 0.011119615554809571, 0.011057151794433593, 0.011068415641784669, 0.011021311759948731, 0.01104793643951416, 0.011161600112915039, 0.011126784324645997, 0.011076607704162598, 0.011054080009460449, 0.011202560424804688, 0.011080703735351562, 0.011062272071838379, 0.023365631103515624, 0.010972160339355469, 0.01102950382232666, 0.011017215728759766, 0.011016192436218262, 0.01102847957611084, 0.011011072158813476, 0.011003904342651367, 0.011016192436218262, 0.011031552314758301, 0.011053088188171387, 0.011050975799560547, 0.011039744377136231, 0.011001855850219726, 0.011066368103027344, 0.01112985610961914, 0.010855423927307128, 0.01092300796508789, 0.010872832298278809, 0.010899456024169921, 0.010778623580932617, 0.010829824447631836, 0.010994688034057617, 0.011018239974975585, 0.011035648345947266, 0.011034624099731445, 0.011143168449401856, 0.011077664375305175, 0.011027423858642579, 0.01105510425567627, 0.011064319610595704, 0.01108790397644043, 0.011046879768371582, 0.01101414394378662, 0.011060223579406739, 0.011064319610595704, 0.011098112106323242, 0.011034624099731445, 0.011096063613891602, 0.01104793643951416, 0.011021311759948731, 0.011065343856811523, 0.011182080268859864, 0.01112166404724121, 0.011107328414916993, 0.011080703735351562, 0.011080703735351562, 0.01113702392578125, 0.011103232383728028, 0.011082752227783203, 0.011018303871154786, 0.011081664085388184, 0.011163647651672364, 0.011080767631530762, 0.011104191780090332, 0.011074560165405273, 0.010855423927307128, 0.010910719871520995, 0.01083903980255127, 0.010877951622009278, 0.010870783805847169, 0.01080832004547119, 0.010946559906005859, 0.022261823654174805, 0.010372032165527343, 0.01041305637359619, 0.010391551971435547, 0.010406911849975586, 0.0104017915725708, 0.010338303565979003, 0.010470399856567383, 0.010395648002624512, 0.010359807968139649, 0.010357760429382324, 0.010380288124084473, 0.010528767585754394, 0.011116543769836425, 0.011110400199890137, 0.011115519523620606, 0.011082752227783203, 0.01115443229675293, 0.011058176040649414, 0.011031552314758301, 0.011048959732055665, 0.011095040321350098, 0.011101183891296386, 0.0110632963180542, 0.011089920043945312, 0.01107148838043213, 0.011062272071838379, 0.011088895797729491, 0.01105510425567627, 0.011023360252380371, 0.011111424446105958, 0.011047967910766602, 0.011032544136047363, 0.011105279922485351, 0.011359231948852539, 0.011072511672973634, 0.01105510425567627, 0.011084799766540527, 0.011080703735351562, 0.011078656196594238, 0.012894207954406739, 0.01135206413269043, 0.011204607963562012, 0.011106304168701172, 0.011083776473999024, 0.011175935745239257, 0.011152383804321288, 0.01111244773864746, 0.0110448637008667, 0.011191295623779298, 0.011116543769836425, 0.011083776473999024, 0.01107968044281006, 0.011076640129089356, 0.011076576232910156, 0.011108351707458495, 0.01112063980102539, 0.011068415641784669, 0.011074560165405273, 0.011169792175292969, 0.011084832191467286, 0.011101152420043945, 0.011115519523620606, 0.023419904708862304, 0.011096063613891602, 0.011101183891296386, 0.011091967582702637, 0.011140095710754394, 0.01112063980102539, 0.011058176040649414, 0.011149312019348144, 0.011003904342651367, 0.011060223579406739, 0.011060223579406739, 0.011188223838806152, 0.011062272071838379, 0.01140121555328369, 0.01126912021636963, 0.011206720352172852, 0.011110336303710937, 0.011169792175292969, 0.011081727981567382, 0.011107328414916993, 0.011073535919189453, 0.011057151794433593, 0.011122688293457032, 0.011110400199890137, 0.011061247825622558, 0.011032575607299805, 0.011084799766540527, 0.011122688293457032, 0.011084863662719727, 0.01113593578338623, 0.011147263526916504, 0.01306214427947998, 0.012184576034545898, 0.011279359817504882, 0.011052032470703126, 0.011067392349243164, 0.011080703735351562, 0.011072511672973634, 0.011158592224121093, 0.011073472023010254, 0.011085824012756347, 0.011311103820800781, 0.011406335830688476, 0.011143168449401856, 0.0111278076171875, 0.011105279922485351, 0.011146240234375, 0.011101183891296386, 0.011067392349243164, 0.01120358371734619, 0.011150336265563965, 0.011087871551513672, 0.011042816162109375, 0.011045887947082519, 0.011133952140808106, 0.011110400199890137, 0.011117568016052246, 0.011411456108093262, 0.011166720390319825, 0.011073535919189453, 0.011205632209777832, 0.011075584411621094, 0.011076607704162598]",tokens/s,88.75835647320693,,,main,False,False float16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1192.951808,14382.792704,0.0,13736.3456,13487.531008,s,10,0.8513474197387696,0.08513474197387695,0.00021429432968233625,0.08507120132446289,0.08527548065185546,0.08551233215332031,0.08570181335449219,"[0.08574918365478515, 0.08507942199707032, 0.08506537628173828, 0.08504169464111327, 0.0849843521118164, 0.084989501953125, 0.08503981018066406, 0.0850770263671875, 0.08522284698486328, 0.08509820556640625]",tokens/s,3006.9980135554056,kWh,1.0067580769403281e-06,5.516551041567917e-07,5.112296933525468e-06,6.670710114622587e-06,tokens/kWh,38376723.85715473,MB,1192.951808,14382.792704,0.0,13736.3456,13661.254656,s,10,19.876877197265628,1.9876877197265628,0.0030016916208529323,1.9866903076171876,1.9884474975585937,1.9924959045410158,1.9957346301269532,"[1.9965443115234376, 1.98677587890625, 1.9857510986328124, 1.9863314208984375, 1.986604736328125, 1.9861082763671876, 1.986561767578125, 1.9872159423828124, 1.9875478515625, 1.9874359130859376]",tokens/s,31.695119597894696,kWh,2.3404805134076856e-05,1.2826450611971834e-05,0.00010996662846287472,0.00014619788420892343,tokens/kWh,430922.7889370145,,s,629,20.61038695144655,0.0327669108926018,0.010142886768973333,0.031545343399047854,0.03165429801940918,0.03173130226135254,0.11686746459960938,"[0.03244851303100586, 0.03200102233886719, 0.032031742095947266, 0.031970304489135744, 0.03204710388183594, 0.03201638412475586, 0.03205120086669922, 0.03203583908081055, 0.0320634880065918, 0.03206047821044922, 0.03205318450927734, 0.03189145660400391, 0.031936511993408204, 0.03190169525146484, 0.03164159965515137, 0.03146956825256347, 0.03146444892883301, 0.031455232620239255, 0.03146342468261719, 0.03145219230651856, 0.03151049613952637, 0.03150643157958984, 0.0314654712677002, 0.03148595237731933, 0.031510528564453126, 0.0315043830871582, 0.031526912689208986, 0.03151155281066895, 0.0315729923248291, 0.03154022407531738, 0.03153510475158691, 0.03153408050537109, 0.03156480026245117, 0.031541248321533204, 0.03159654426574707, 0.03154947280883789, 0.031594463348388675, 0.03157811164855957, 0.03160272026062012, 0.03156374359130859, 0.031591424942016604, 0.03159654426574707, 0.03158425521850586, 0.03158732795715332, 0.03161497688293457, 0.03161907196044922, 0.03164262390136719, 0.03162623977661133, 0.03163039970397949, 0.03161087989807129, 0.03163030433654785, 0.031619104385375976, 0.03164255905151367, 0.03164057540893555, 0.03165593528747559, 0.03161292839050293, 0.03168364715576172, 0.031652799606323244, 0.03166924858093262, 0.031681535720825195, 0.031661056518554685, 0.03163545608520508, 0.11689984130859375, 0.03139583969116211, 0.03134576034545898, 0.03139676856994629, 0.031409151077270506, 0.03140403175354004, 0.03139379119873047, 0.03142758369445801, 0.03139993667602539, 0.03140505599975586, 0.031406080245971676, 0.03146956825256347, 0.03141119956970215, 0.03147776031494141, 0.031455232620239255, 0.03143680000305176, 0.031425535202026365, 0.03143270492553711, 0.031438848495483396, 0.03144704055786133, 0.031459327697753905, 0.031457279205322264, 0.03146444892883301, 0.031493120193481446, 0.031442943572998046, 0.03144806480407715, 0.031473663330078124, 0.0315167350769043, 0.03149100875854492, 0.03152076721191406, 0.03173785591125488, 0.03190169525146484, 0.0316682243347168, 0.03155148887634277, 0.03151974487304687, 0.03153510475158691, 0.031541248321533204, 0.03157401657104492, 0.031635488510131836, 0.03158012771606445, 0.03154944038391113, 0.03155353546142578, 0.03157094383239746, 0.03155558395385742, 0.03154841613769531, 0.03160063934326172, 0.03159449577331543, 0.03163443183898926, 0.03160985565185547, 0.031632383346557616, 0.03158527946472168, 0.03158425521850586, 0.03159347152709961, 0.031640640258789064, 0.031636415481567384, 0.03162521553039551, 0.03160063934326172, 0.03164672088623047, 0.03160479927062988, 0.03164358329772949, 0.031627264022827145, 0.03163340759277344, 0.03165388870239258, 0.11698585510253906, 0.03139174461364746, 0.03136511993408203, 0.03137740707397461, 0.031357952117919925, 0.03141427230834961, 0.03139993667602539, 0.03142348861694336, 0.03139379119873047, 0.031409151077270506, 0.03139891242980957, 0.0314337272644043, 0.03140812873840332, 0.03145113563537598, 0.03147980880737305, 0.0314521598815918, 0.03143276786804199, 0.03142444801330566, 0.031456256866455076, 0.03143577575683594, 0.03146240043640137, 0.031503360748291014, 0.031440895080566404, 0.03144908714294434, 0.031453184127807614, 0.03147980880737305, 0.0314839038848877, 0.03152179145812988, 0.031488000869750975, 0.0315156478881836, 0.031494144439697266, 0.031513599395751955, 0.03150643157958984, 0.031514623641967776, 0.03149516868591309, 0.03156070327758789, 0.03151155281066895, 0.03156073570251465, 0.031523807525634766, 0.03158323287963867, 0.03160371208190918, 0.031547391891479495, 0.0315729923248291, 0.03155763244628906, 0.03159449577331543, 0.0315863037109375, 0.03159859275817871, 0.031632383346557616, 0.03156172752380371, 0.03159552001953125, 0.03159859275817871, 0.03160063934326172, 0.03160678482055664, 0.03161087989807129, 0.03164672088623047, 0.031692800521850584, 0.03163750457763672, 0.03163955116271973, 0.03158937644958496, 0.03161087989807129, 0.03162521553039551, 0.03162521553039551, 0.03162524795532227, 0.1168803482055664, 0.03137638473510742, 0.03139795112609863, 0.031378368377685546, 0.03136102485656738, 0.03139993667602539, 0.03140403175354004, 0.03141119956970215, 0.03143065643310547, 0.031460351943969726, 0.031437824249267575, 0.03142758369445801, 0.031440895080566404, 0.0314204158782959, 0.03138662338256836, 0.03142860794067383, 0.03145011138916016, 0.031476736068725586, 0.03146240043640137, 0.03148595237731933, 0.031454208374023435, 0.031457279205322264, 0.031470592498779294, 0.03143475151062012, 0.03143577575683594, 0.031514623641967776, 0.03149004745483398, 0.03149619293212891, 0.03148185539245606, 0.031498239517211916, 0.03148902320861816, 0.031514656066894534, 0.031561695098876956, 0.03159040069580078, 0.031514623641967776, 0.03158425521850586, 0.031535167694091794, 0.03154118347167969, 0.03156172752380371, 0.03156172752380371, 0.031604736328125, 0.031558656692504884, 0.03156991958618164, 0.03159449577331543, 0.03158118438720703, 0.031644672393798826, 0.03164672088623047, 0.03157606315612793, 0.03158323287963867, 0.03160371208190918, 0.03162931251525879, 0.031615999221801756, 0.03163545608520508, 0.03162009620666504, 0.03165081596374512, 0.03165184020996094, 0.03163852882385254, 0.03162623977661133, 0.03167849540710449, 0.03163644790649414, 0.03162931251525879, 0.031677440643310545, 0.031631359100341795, 0.11683433532714843, 0.031388639450073245, 0.03139481544494629, 0.03139072036743164, 0.03138150405883789, 0.03142451286315918, 0.031543296813964845, 0.03148595237731933, 0.03141734313964844, 0.03141836738586426, 0.03145113563537598, 0.031440895080566404, 0.031389696121215824, 0.03143475151062012, 0.03141836738586426, 0.03141119956970215, 0.03143168067932129, 0.03145113563537598, 0.031456256866455076, 0.031426559448242186, 0.03143577575683594, 0.031456256866455076, 0.03151974487304687, 0.031444000244140624, 0.03147977638244629, 0.031486976623535154, 0.031541248321533204, 0.031475711822509765, 0.0315156478881836, 0.03160063934326172, 0.03160166358947754, 0.031525888442993165, 0.03153510475158691, 0.031541248321533204, 0.03153305625915527, 0.03153919982910156, 0.03152383995056152, 0.03160985565185547, 0.03159040069580078, 0.031541248321533204, 0.031546367645263675, 0.03157708740234375, 0.03153715133666992, 0.0315729923248291, 0.03159654426574707, 0.03159145545959473, 0.03159651184082031, 0.03160272026062012, 0.03160367965698242, 0.0316231689453125, 0.03159040069580078, 0.031631359100341795, 0.0316180477142334, 0.03165081596374512, 0.031661056518554685, 0.03163340759277344, 0.03160576057434082, 0.03165388870239258, 0.0315996150970459, 0.03161497688293457, 0.03169183921813965, 0.03169683265686035, 0.031659008026123044, 0.11681996917724609, 0.03136511993408203, 0.03139686393737793, 0.03136204719543457, 0.03136307144165039, 0.03138047981262207, 0.03142451286315918, 0.031425535202026365, 0.031389696121215824, 0.03141939163208008, 0.03141734313964844, 0.03141734313964844, 0.03140300750732422, 0.03141119956970215, 0.03140812873840332, 0.031474687576293944, 0.031454208374023435, 0.0314839038848877, 0.031454208374023435, 0.03144908714294434, 0.03147878456115723, 0.031508480072021484, 0.0314654712677002, 0.03150233650207519, 0.031453184127807614, 0.03155148887634277, 0.0314747200012207, 0.031484895706176755, 0.03148185539245606, 0.03151155281066895, 0.0314901123046875, 0.03150534439086914, 0.031579135894775394, 0.03163545608520508, 0.031528959274291994, 0.03157094383239746, 0.0315545597076416, 0.03156991958618164, 0.03156889533996582, 0.03156787109375, 0.03158937644958496, 0.03155251121520996, 0.031542272567749025, 0.031579135894775394, 0.031536191940307615, 0.031571903228759766, 0.03160063934326172, 0.03158835220336914, 0.03158118438720703, 0.031604736328125, 0.031615999221801756, 0.03164057540893555, 0.03163955116271973, 0.03162112045288086, 0.03161292839050293, 0.03164057540893555, 0.031615999221801756, 0.031611904144287106, 0.031643648147583005, 0.03161907196044922, 0.03162214469909668, 0.03164262390136719, 0.03165798377990723, 0.11688038635253906, 0.03139072036743164, 0.03137542343139649, 0.031378368377685546, 0.03137740707397461, 0.03141427230834961, 0.03137433624267578, 0.0314337272644043, 0.03139993667602539, 0.03142451286315918, 0.03138051223754883, 0.03139068794250488, 0.03142348861694336, 0.03143987274169922, 0.031455232620239255, 0.031421440124511715, 0.03142758369445801, 0.03141836738586426, 0.03141529655456543, 0.03164163208007813, 0.031812576293945315, 0.0315729923248291, 0.03155763244628906, 0.031497215270996096, 0.03148492813110351, 0.031458303451538085, 0.031470592498779294, 0.03148287963867188, 0.031470592498779294, 0.03151974487304687, 0.0315545597076416, 0.031494144439697266, 0.031525888442993165, 0.03155353546142578, 0.031527936935424807, 0.031531007766723636, 0.03154022407531738, 0.03156684875488281, 0.03155660820007324, 0.03156377601623535, 0.031541248321533204, 0.031559680938720705, 0.031545343399047854, 0.03155353546142578, 0.031559680938720705, 0.031631359100341795, 0.0316364803314209, 0.03159449577331543, 0.03161292839050293, 0.0315996150970459, 0.03161702346801758, 0.03161907196044922, 0.03161708831787109, 0.03163027191162109, 0.0316231689453125, 0.031632383346557616, 0.03162931251525879, 0.0316364803314209, 0.03160576057434082, 0.03163852882385254, 0.03160576057434082, 0.031673408508300784, 0.03167328071594238, 0.11695104217529297, 0.03136614418029785, 0.03140812873840332, 0.031341567993164066, 0.03134464073181152, 0.03138662338256836, 0.03139583969116211, 0.03141427230834961, 0.03141839981079102, 0.031402975082397463, 0.03142758369445801, 0.03140403175354004, 0.03143270492553711, 0.03143680000305176, 0.03141222381591797, 0.0314726390838623, 0.031498239517211916, 0.03145939254760742, 0.03145414352416992, 0.03146342468261719, 0.031473663330078124, 0.03148185539245606, 0.03149926376342774, 0.031480831146240236, 0.03144908714294434, 0.03149004745483398, 0.03147878456115723, 0.031512575149536134, 0.03147878456115723, 0.031486976623535154, 0.03150035285949707, 0.031532991409301755, 0.03156070327758789, 0.03155046463012695, 0.03151667213439941, 0.031545343399047854, 0.031528959274291994, 0.03162931251525879, 0.03155660820007324, 0.031591424942016604, 0.03156793594360351, 0.03155756759643555, 0.03153305625915527, 0.031719423294067385, 0.031751167297363284, 0.03162521553039551, 0.031562751770019534, 0.03163443183898926, 0.03162214469909668, 0.03162931251525879, 0.03162931251525879, 0.03163340759277344, 0.03163033676147461, 0.031662080764770506, 0.03163443183898926, 0.031635488510131836, 0.031657951354980465, 0.03162521553039551, 0.031645696640014646, 0.03162623977661133, 0.031665151596069335, 0.03172147178649903, 0.031932416915893554, 0.11726643371582031, 0.03140096092224121, 0.0313753604888916, 0.03136511993408203, 0.031369216918945314, 0.03141119956970215, 0.03139174461364746, 0.03143680000305176, 0.03141119956970215, 0.03142860794067383, 0.03138355255126953, 0.03143270492553711, 0.031388671875, 0.03146444892883301, 0.03147980880737305, 0.03142860794067383, 0.03144806480407715, 0.03144704055786133, 0.03144704055786133, 0.031438911437988284, 0.03145414352416992, 0.03148902320861816, 0.031471616744995115, 0.03142963218688965, 0.03143577575683594, 0.031475711822509765, 0.0314654712677002, 0.03149926376342774, 0.031473663330078124, 0.03168460845947266, 0.0315863037109375, 0.03153510475158691, 0.03156380844116211, 0.031540191650390625, 0.031558656692504884, 0.031547391891479495, 0.031525888442993165, 0.03157196807861328, 0.03155660820007324, 0.03160063934326172, 0.03164777565002441, 0.031593439102172854, 0.031543296813964845, 0.03155558395385742, 0.03156172752380371, 0.03161702346801758, 0.031600671768188475, 0.03158320045471191, 0.0316180477142334, 0.03159756851196289, 0.03162214469909668, 0.03169385528564453, 0.031658975601196286, 0.031666175842285156, 0.03163852882385254, 0.031644672393798826, 0.03165184020996094, 0.031628288269042966, 0.03160780715942383, 0.031698944091796875, 0.03174710464477539, 0.032173023223876954, 0.031937536239624024, 0.11699404907226563, 0.03139276885986328, 0.03143168067932129, 0.031413280487060546, 0.031399904251098634, 0.0313948802947998, 0.031426496505737304, 0.031406080245971676, 0.03138252830505371, 0.03141119956970215, 0.031442943572998046, 0.03140300750732422, 0.031409151077270506, 0.031412256240844726, 0.03142857551574707, 0.03144908714294434, 0.031456256866455076, 0.031719423294067385, 0.03181260871887207, 0.03155558395385742, 0.03146854400634766, 0.031473663330078124, 0.0314839038848877, 0.03146240043640137, 0.03145113563537598, 0.03151155281066895, 0.03166924858093262, 0.03152383995056152, 0.031513599395751955, 0.031493120193481446, 0.031486976623535154, 0.031525888442993165, 0.03151974487304687, 0.03158937644958496, 0.0315361270904541, 0.03159040069580078, 0.031562751770019534, 0.03158732795715332, 0.03156377601623535, 0.03157196807861328, 0.03156377601623535, 0.031562751770019534, 0.0315863037109375, 0.0316180477142334, 0.03159347152709961, 0.0316231689453125, 0.03157708740234375, 0.031579135894775394, 0.03163443183898926, 0.03165593528747559, 0.03164057540893555, 0.03163545608520508, 0.03159043121337891, 0.03164463996887207, 0.03160371208190918, 0.031665151596069335, 0.031628288269042966, 0.031643648147583005, 0.031678464889526366, 0.03165388870239258, 0.03162419128417969, 0.031648767471313476, 0.03162419128417969]",tokens/s,30.518592469019794,,,main,False,False -float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11584,7 +11584,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1262.53056,4276.617216,0.0,3632.267264,3408.132096,s,10,0.20172547531127932,0.020172547531127932,0.0006042842284314442,0.019867584228515624,0.020415465545654297,0.021160420989990235,0.021756385345458987,"[0.021905376434326173, 0.020249919891357424, 0.019857440948486328, 0.020231647491455076, 0.019795808792114258, 0.020247552871704103, 0.019860736846923827, 0.01987443161010742, 0.019846559524536133, 0.019856000900268556]",tokens/s,12690.514155684627,kWh,2.3836587425521627e-07,1.3061257829554786e-07,1.1820781186007993e-06,1.5510565711515636e-06,tokens/kWh,165048783.36573878,MB,1262.53056,4278.714368,0.0,3632.267264,3474.14784,s,10,11.84439318847656,1.1844393188476563,0.017149637998338907,1.1800570068359375,1.2016350952148436,1.2148355529785155,1.225395919189453,"[1.2280360107421875, 1.1738792724609375, 1.17937646484375, 1.1902847900390625, 1.1834696044921875, 1.19870166015625, 1.167730224609375, 1.1738626708984374, 1.180737548828125, 1.16831494140625]",tokens/s,53.189723608038314,kWh,1.4160408100587281e-05,7.759553078422177e-06,3.803503658839977e-05,5.995499776740923e-05,tokens/kWh,1050788.1301973127,,s,629,12.003487743377686,0.019083446332873906,0.002451252994608814,0.01860403251647949,0.019179769897460938,0.019804773712158205,0.038735421295166016,"[0.020542463302612304, 0.0214466552734375, 0.02125209617614746, 0.02061311912536621, 0.020566015243530272, 0.020291584014892578, 0.019958816528320312, 0.019805183410644533, 0.0193023681640625, 0.02062131118774414, 0.0198287353515625, 0.020178943634033202, 0.020082687377929686, 0.02007142448425293, 0.02020249557495117, 0.019955711364746095, 0.019833856582641602, 0.019764223098754884, 0.019976192474365235, 0.02011238479614258, 0.019804159164428712, 0.019768320083618163, 0.019677183151245118, 0.01968230438232422, 0.01901670455932617, 0.019979263305664064, 0.02011238479614258, 0.019560447692871095, 0.01923379135131836, 0.019111936569213867, 0.01924608039855957, 0.019086336135864256, 0.019124223709106446, 0.019074047088623047, 0.01901670455932617, 0.01906892776489258, 0.019103744506835937, 0.019129344940185547, 0.019164159774780275, 0.019133440017700197, 0.018712575912475587, 0.0190382080078125, 0.019080223083496092, 0.019180511474609373, 0.019127296447753905, 0.018946048736572265, 0.018535423278808593, 0.019070976257324217, 0.019120128631591796, 0.019128320693969726, 0.019109888076782225, 0.019177471160888672, 0.01882111930847168, 0.01920204734802246, 0.018685951232910156, 0.018507776260375978, 0.018527231216430663, 0.018521087646484375, 0.018503679275512695, 0.018556928634643553, 0.018671615600585938, 0.01925632095336914, 0.03919974517822265, 0.01861631965637207, 0.019177471160888672, 0.019099647521972657, 0.018858015060424806, 0.019159008026123046, 0.01917958450317383, 0.019191743850708008, 0.019079168319702147, 0.018308095932006836, 0.01856716728210449, 0.018927711486816406, 0.018954143524169922, 0.019115007400512696, 0.018861055374145508, 0.018431999206542968, 0.018561023712158203, 0.018520063400268554, 0.018456575393676757, 0.018655231475830078, 0.01902694320678711, 0.018495487213134765, 0.01842483139038086, 0.018486272811889647, 0.01847500801086426, 0.018513919830322266, 0.01843507194519043, 0.018495487213134765, 0.018380800247192384, 0.01845452880859375, 0.018482175827026368, 0.01843609619140625, 0.01846067237854004, 0.018529279708862305, 0.018449407577514648, 0.018412544250488282, 0.018395135879516602, 0.018659328460693358, 0.019100671768188478, 0.01879654312133789, 0.019056640625, 0.01904025650024414, 0.018675712585449217, 0.018489343643188477, 0.01820364761352539, 0.018440191268920898, 0.018295808792114256, 0.018214912414550782, 0.018335744857788085, 0.018545663833618165, 0.018498559951782227, 0.018548736572265623, 0.018561023712158203, 0.01850060844421387, 0.018526208877563476, 0.01884774398803711, 0.01880268859863281, 0.018512895584106445, 0.01844121551513672, 0.01847500801086426, 0.018453504562377928, 0.0184453125, 0.01846272087097168, 0.03910041427612305, 0.019066879272460938, 0.018532352447509767, 0.01863680076599121, 0.018544639587402344, 0.01841049575805664, 0.018513919830322266, 0.01845248031616211, 0.018541568756103514, 0.018669567108154296, 0.01903615951538086, 0.01903308868408203, 0.01908531188964844, 0.019106815338134766, 0.018551807403564453, 0.018509824752807616, 0.01843814468383789, 0.018505727767944336, 0.018486272811889647, 0.018520063400268554, 0.0184770565032959, 0.018503679275512695, 0.01846272087097168, 0.0186060791015625, 0.018523136138916017, 0.018553855895996094, 0.01862758445739746, 0.019130367279052735, 0.019154943466186524, 0.019125247955322267, 0.019168256759643554, 0.019204095840454103, 0.01881292724609375, 0.018530303955078126, 0.018546688079833985, 0.018530303955078126, 0.018527231216430663, 0.018471935272216796, 0.018535423278808593, 0.018525184631347655, 0.018541568756103514, 0.01859071922302246, 0.018577407836914063, 0.018478080749511717, 0.018478080749511717, 0.018824192047119142, 0.01904332733154297, 0.019099647521972657, 0.019079168319702147, 0.019099647521972657, 0.01864806365966797, 0.018506752014160157, 0.018541568756103514, 0.018465791702270508, 0.018586624145507814, 0.018512895584106445, 0.018491392135620118, 0.018508800506591795, 0.019112960815429687, 0.01902694320678711, 0.019100671768188478, 0.01906073570251465, 0.019109888076782225, 0.03967488098144531, 0.018980863571166993, 0.018959360122680666, 0.01905356788635254, 0.019014656066894533, 0.019109888076782225, 0.01842585563659668, 0.018521087646484375, 0.018512895584106445, 0.01840947151184082, 0.018479103088378905, 0.018529279708862305, 0.018541568756103514, 0.018449407577514648, 0.018489343643188477, 0.018473983764648438, 0.018508800506591795, 0.018973695755004884, 0.01903104019165039, 0.01884364891052246, 0.01903001594543457, 0.01906380844116211, 0.019094528198242186, 0.019133440017700197, 0.018968576431274413, 0.01906380844116211, 0.019105791091918945, 0.019078144073486326, 0.01905971145629883, 0.019108863830566408, 0.019158016204833983, 0.018573312759399413, 0.01858252716064453, 0.018556928634643553, 0.01900851249694824, 0.019284992218017577, 0.0185743350982666, 0.018363391876220703, 0.018496511459350586, 0.018543615341186523, 0.018938880920410156, 0.01904332733154297, 0.01881088066101074, 0.01902899169921875, 0.018911231994628908, 0.018923519134521484, 0.018784351348876953, 0.019183647155761718, 0.018815872192382812, 0.019014656066894533, 0.01904025650024414, 0.01876479911804199, 0.020634624481201173, 0.019451904296875, 0.019106815338134766, 0.019137535095214844, 0.018933792114257813, 0.018851808547973633, 0.018938880920410156, 0.018910207748413087, 0.018840576171875, 0.018852863311767578, 0.01882316780090332, 0.03844300842285156, 0.018851839065551757, 0.01883340835571289, 0.01881088066101074, 0.018976768493652343, 0.01887129592895508, 0.019036256790161132, 0.018843616485595703, 0.01892140769958496, 0.018868223190307617, 0.018856960296630858, 0.018538496017456055, 0.018280448913574218, 0.01825484848022461, 0.018266111373901366, 0.01823142433166504, 0.01822604751586914, 0.0182476806640625, 0.01845964813232422, 0.018541568756103514, 0.018534400939941405, 0.018465791702270508, 0.018488319396972656, 0.018486272811889647, 0.01839411163330078, 0.018455551147460936, 0.018525184631347655, 0.018516992568969725, 0.018450464248657227, 0.018618335723876955, 0.018513919830322266, 0.018513919830322266, 0.018619392395019533, 0.018453504562377928, 0.01861222457885742, 0.01840025520324707, 0.018787328720092773, 0.019181631088256837, 0.019106752395629884, 0.01908531188964844, 0.019148799896240236, 0.019058719635009765, 0.019175392150878906, 0.019174400329589843, 0.01904435157775879, 0.018679807662963867, 0.019136512756347656, 0.019075071334838867, 0.019108863830566408, 0.01903923225402832, 0.019118080139160155, 0.01923174476623535, 0.019125247955322267, 0.01907302474975586, 0.019110944747924803, 0.019160032272338867, 0.019133440017700197, 0.019054592132568358, 0.019722240447998047, 0.018531328201293946, 0.018663423538208008, 0.019140607833862306, 0.018979839324951172, 0.03991961669921875, 0.019216384887695313, 0.019150848388671874, 0.019146751403808594, 0.018997247695922852, 0.019105791091918945, 0.018881536483764647, 0.0184586238861084, 0.018907136917114258, 0.020163583755493163, 0.019698688507080078, 0.019131391525268555, 0.018559999465942383, 0.018554880142211915, 0.019185663223266602, 0.019090431213378906, 0.019098623275756836, 0.01905971145629883, 0.018977792739868164, 0.01920512008666992, 0.019120128631591796, 0.01923276710510254, 0.019084287643432618, 0.01907711982727051, 0.019142656326293944, 0.019140607833862306, 0.019058687210083008, 0.019076095581054688, 0.019161088943481445, 0.019091455459594727, 0.01904947280883789, 0.01905971145629883, 0.019100671768188478, 0.019079168319702147, 0.019120128631591796, 0.01906585693359375, 0.018800640106201173, 0.0191016960144043, 0.019168256759643554, 0.021114879608154297, 0.019520511627197267, 0.019134464263916014, 0.01889695930480957, 0.01913337516784668, 0.019062784194946288, 0.019168256759643554, 0.019506175994873046, 0.0189040641784668, 0.018456575393676757, 0.018568256378173827, 0.01847596740722656, 0.018541568756103514, 0.018523136138916017, 0.018531328201293946, 0.01897881507873535, 0.01904742431640625, 0.018505727767944336, 0.018555904388427736, 0.018508800506591795, 0.018526208877563476, 0.018465791702270508, 0.019215360641479492, 0.01844326400756836, 0.03929702377319336, 0.018728960037231446, 0.018301952362060548, 0.01826918411254883, 0.018327552795410155, 0.018184255599975586, 0.018796480178833008, 0.018395135879516602, 0.018572288513183592, 0.01848320007324219, 0.018505727767944336, 0.01849241638183594, 0.018431999206542968, 0.01860403251647949, 0.018431999206542968, 0.018523136138916017, 0.018487295150756835, 0.018504703521728515, 0.018507776260375978, 0.018421760559082033, 0.018283519744873047, 0.018692096710205077, 0.01823232078552246, 0.01826521682739258, 0.018296703338623047, 0.018242559432983398, 0.01857535934448242, 0.01842995262145996, 0.018525184631347655, 0.018471935272216796, 0.018486272811889647, 0.018513919830322266, 0.018495487213134765, 0.018522111892700196, 0.018548736572265623, 0.018480127334594726, 0.018588672637939452, 0.018471935272216796, 0.018526208877563476, 0.018544639587402344, 0.018559999465942383, 0.018527231216430663, 0.018585599899291993, 0.018502656936645507, 0.01850060844421387, 0.018550783157348632, 0.018537471771240235, 0.018516992568969725, 0.018516992568969725, 0.018473983764648438, 0.018694143295288086, 0.019397632598876953, 0.019135488510131835, 0.0184770565032959, 0.01862246322631836, 0.018573312759399413, 0.018544639587402344, 0.018519039154052733, 0.018516992568969725, 0.018408447265625, 0.018513919830322266, 0.018778112411499022, 0.018700288772583007, 0.03871846389770508, 0.018515968322753908, 0.01841971206665039, 0.01843404769897461, 0.018421760559082033, 0.018549760818481444, 0.018811904907226562, 0.018521087646484375, 0.018496511459350586, 0.018473983764648438, 0.018569215774536133, 0.018568191528320312, 0.0184268798828125, 0.01855897521972656, 0.019104768753051758, 0.01906073570251465, 0.01902387237548828, 0.019129344940185547, 0.018748416900634765, 0.01847500801086426, 0.018535423278808593, 0.018514944076538087, 0.018431999206542968, 0.0185743350982666, 0.01899622344970703, 0.01903104019165039, 0.019083263397216797, 0.01902079963684082, 0.018520063400268554, 0.0186060791015625, 0.01848422431945801, 0.018485248565673826, 0.018537471771240235, 0.018910207748413087, 0.018465791702270508, 0.018509855270385744, 0.018459615707397462, 0.01848320007324219, 0.018551807403564453, 0.018479103088378905, 0.018778112411499022, 0.018624576568603515, 0.018471872329711914, 0.018502656936645507, 0.018510847091674804, 0.018465791702270508, 0.018528255462646484, 0.018777120590209962, 0.01847702407836914, 0.018780223846435545, 0.019100608825683593, 0.019056640625, 0.018651136398315428, 0.01859993553161621, 0.018498559951782227, 0.018539520263671876, 0.018534400939941405, 0.018488319396972656, 0.018525184631347655, 0.018588672637939452, 0.018491392135620118, 0.018496511459350586, 0.018549760818481444, 0.038742015838623044, 0.018594816207885743, 0.018518016815185546, 0.018554880142211915, 0.018515968322753908, 0.018482175827026368, 0.018509824752807616, 0.018501632690429686, 0.018498559951782227, 0.018556928634643553, 0.018472959518432617, 0.018512895584106445, 0.01846784019470215, 0.018527231216430663, 0.018957311630249024, 0.018927616119384767, 0.01883443260192871, 0.018907136917114258, 0.01888876724243164, 0.018984895706176758, 0.019397632598876953, 0.018735103607177735, 0.019118112564086916, 0.019149856567382814, 0.018949056625366213, 0.01899622344970703, 0.018898944854736328, 0.018889728546142577, 0.01884876823425293, 0.018915327072143554, 0.018873344421386717, 0.01883852767944336, 0.01836031913757324, 0.01820364761352539, 0.018283519744873047, 0.018251775741577148, 0.018232416152954102, 0.018818080902099608, 0.018825088500976563, 0.01881395149230957, 0.01887129592895508, 0.01886617660522461, 0.0188570556640625, 0.01900124740600586, 0.019188735961914064, 0.019133440017700197, 0.018939903259277344, 0.018927616119384767, 0.018935808181762694, 0.01884160041809082, 0.01886310386657715, 0.018912479400634764, 0.01886297607421875, 0.018691999435424805, 0.018296831130981444, 0.01836031913757324, 0.018362464904785155, 0.018191263198852538, 0.018739200592041014, 0.018753536224365236, 0.018817024230957033, 0.019144704818725586, 0.018720767974853517, 0.0388853759765625, 0.018555904388427736, 0.018421823501586915, 0.018602943420410155, 0.018553855895996094, 0.018581504821777343, 0.01853753662109375, 0.01853331184387207, 0.018557952880859374, 0.018505727767944336, 0.018494464874267577, 0.018523136138916017, 0.018592767715454102, 0.018498559951782227, 0.018589696884155273, 0.018584575653076172, 0.018554912567138673, 0.01848214340209961, 0.018537471771240235, 0.018555904388427736, 0.018501632690429686, 0.01858371162414551, 0.01849839973449707, 0.018569215774536133, 0.018603008270263673, 0.0185948486328125, 0.018607072830200196, 0.01859891128540039, 0.018527231216430663, 0.018573312759399413, 0.018562047958374024, 0.018562047958374024, 0.018478080749511717, 0.018533376693725585, 0.018543615341186523, 0.018505727767944336, 0.018522111892700196, 0.018504703521728515, 0.018525184631347655, 0.018526208877563476, 0.018485279083251954, 0.018565088272094726, 0.01856716728210449, 0.01856716728210449, 0.018518016815185546, 0.018661376953125, 0.01846784019470215, 0.018555904388427736, 0.01861734390258789, 0.01846272087097168, 0.018493440628051756, 0.0185262393951416, 0.01848931121826172, 0.018480127334594726, 0.018537471771240235, 0.018504703521728515, 0.018534400939941405, 0.018581504821777343, 0.018494464874267577, 0.018588672637939452, 0.018421760559082033, 0.018534400939941405, 0.018485248565673826]",tokens/s,52.401436436423964,,,,, -bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11621,7 +11621,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 768.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11968,7 +11968,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11995,7 +11995,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12186,7 +12186,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GP ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1259.839488,3852.992512,0.0,3206.545408,3105.82784,s,10,0.1930521278381348,0.019305212783813475,0.0002542779813739284,0.01924176025390625,0.019335408782958984,0.019699080276489256,0.019990017471313477,"[0.02006275177001953, 0.019244255065917967, 0.019185632705688477, 0.01923993682861328, 0.01925459289550781, 0.01924358367919922, 0.019249919891357423, 0.019162208557128906, 0.019191839218139647, 0.0192174072265625]",tokens/s,13260.667098921807,kWh,2.2924400089528416e-07,1.255498719612745e-07,1.0659343152790816e-06,1.4207281881356402e-06,tokens/kWh,180189287.5342592,MB,1259.839488,3852.992512,0.0,3206.545408,3189.19168,s,10,11.02842041015625,1.102842041015625,0.019420409670015917,1.1083656005859375,1.1194994262695313,1.1242835876464843,1.128110916748047,"[1.1290677490234375, 1.0803919677734375, 1.106476806640625, 1.118436279296875, 1.115659912109375, 1.1168310546875, 1.11025439453125, 1.067032470703125, 1.078049560546875, 1.10622021484375]",tokens/s,57.12513456775939,kWh,1.29441376687585e-05,7.093055422513125e-06,3.657117912772184e-05,5.6608372218993465e-05,tokens/kWh,1112909.5844035237,,s,629,11.182288902282707,0.0177778837874129,0.0023677903324750755,0.017666048049926757,0.01791201286315918,0.018331238555908205,0.03661316253662109,"[0.018852863311767578, 0.018487295150756835, 0.018268159866333008, 0.018164735794067383, 0.01784115219116211, 0.018191360473632814, 0.01842995262145996, 0.01846784019470215, 0.017589248657226563, 0.017549312591552735, 0.01817804718017578, 0.018307071685791015, 0.01883750343322754, 0.018507776260375978, 0.018341888427734376, 0.018276351928710938, 0.018255872726440428, 0.018397184371948243, 0.018229248046875, 0.018317312240600587, 0.018367488861083983, 0.018478080749511717, 0.018318336486816408, 0.01850060844421387, 0.018241535186767577, 0.018151424407958985, 0.017863679885864257, 0.018349056243896485, 0.01829478454589844, 0.018984960556030273, 0.018490367889404297, 0.01776639938354492, 0.017897472381591797, 0.017772544860839845, 0.017763328552246094, 0.017752063751220702, 0.017862655639648437, 0.017739776611328126, 0.018132991790771484, 0.01807360076904297, 0.01779302406311035, 0.017617919921875, 0.017467391967773437, 0.017557504653930665, 0.017487871170043946, 0.017673215866088866, 0.017482751846313475, 0.018217983245849608, 0.018379776000976563, 0.018745344161987306, 0.01706188774108887, 0.016944128036499022, 0.01695542335510254, 0.016973791122436525, 0.01691756820678711, 0.016920511245727538, 0.016898048400878905, 0.016874496459960937, 0.016855039596557618, 0.016759807586669923, 0.016842752456665038, 0.016917503356933594, 0.036670463562011715, 0.01681407928466797, 0.01676089668273926, 0.01675052833557129, 0.01676595115661621, 0.016715776443481444, 0.01677516746520996, 0.01680486488342285, 0.01676905632019043, 0.016758752822875977, 0.016745471954345705, 0.01675775909423828, 0.016688127517700196, 0.016766048431396483, 0.016738208770751953, 0.01676595115661621, 0.01681817626953125, 0.016710655212402344, 0.016698368072509767, 0.016945152282714843, 0.016954368591308593, 0.01685606384277344, 0.016865280151367186, 0.016928768157958983, 0.016911359786987306, 0.01704140853881836, 0.0169881591796875, 0.016911359786987306, 0.016970752716064453, 0.016894975662231446, 0.016891904830932617, 0.01683251190185547, 0.016881664276123046, 0.016900096893310547, 0.016939008712768554, 0.016884735107421875, 0.016945152282714843, 0.016904191970825197, 0.016905216217041014, 0.016911359786987306, 0.01683967971801758, 0.01683967971801758, 0.016959487915039064, 0.01699635124206543, 0.016945152282714843, 0.017855552673339842, 0.01793529510498047, 0.018121728897094725, 0.019534847259521485, 0.018174976348876954, 0.017878015518188475, 0.017827840805053712, 0.017574911117553712, 0.01777663993835449, 0.017648639678955077, 0.01744588851928711, 0.017508352279663086, 0.017743871688842772, 0.017739776611328126, 0.01779199981689453, 0.01784320068359375, 0.01760870361328125, 0.01749504089355469, 0.03660800170898437, 0.016917503356933594, 0.01680588722229004, 0.01679974365234375, 0.01674553680419922, 0.0167259521484375, 0.016698432922363282, 0.01675872039794922, 0.01738956832885742, 0.017556480407714844, 0.017483776092529296, 0.017530879974365234, 0.017614879608154298, 0.01792099189758301, 0.017894399642944335, 0.017804288864135744, 0.017912832260131836, 0.017966079711914062, 0.017904640197753906, 0.017869823455810546, 0.01777561569213867, 0.017711103439331053, 0.01779302406311035, 0.017704959869384765, 0.01781862449645996, 0.01780019187927246, 0.01782067108154297, 0.017912832260131836, 0.0178155517578125, 0.017719295501708983, 0.01780838394165039, 0.017748992919921876, 0.017771520614624024, 0.017682432174682617, 0.01745715141296387, 0.017497087478637697, 0.017803264617919923, 0.017687551498413084, 0.0174704647064209, 0.017497087478637697, 0.01742233657836914, 0.01765888023376465, 0.017712127685546874, 0.017728511810302734, 0.017725439071655275, 0.01739468765258789, 0.017538047790527343, 0.017822719573974608, 0.01743974494934082, 0.017519615173339845, 0.01739571189880371, 0.017533952713012696, 0.01774284744262695, 0.017721343994140625, 0.017812480926513673, 0.01777561569213867, 0.017708032608032227, 0.017458175659179686, 0.01744486427307129, 0.01737932777404785, 0.01744179153442383, 0.017528863906860353, 0.017618911743164063, 0.037647361755371096, 0.017708032608032227, 0.01764249610900879, 0.017723392486572266, 0.01765478324890137, 0.017484800338745117, 0.017862655639648437, 0.017795072555541993, 0.017797119140625, 0.017827840805053712, 0.017979391098022462, 0.01779814338684082, 0.01779199981689453, 0.01777663993835449, 0.017755136489868165, 0.017686527252197267, 0.017796096801757814, 0.017769472122192383, 0.017747007369995117, 0.01774995231628418, 0.017722368240356445, 0.017703935623168944, 0.017748992919921876, 0.017733631134033204, 0.017507360458374022, 0.017522655487060546, 0.017795103073120117, 0.017766368865966796, 0.017722368240356445, 0.017752063751220702, 0.017715200424194336, 0.017731584548950196, 0.017733631134033204, 0.017713151931762695, 0.017748992919921876, 0.017754112243652344, 0.017743871688842772, 0.01765171241760254, 0.017835039138793946, 0.017753055572509766, 0.017719295501708983, 0.017780799865722657, 0.017803199768066408, 0.017796096801757814, 0.017803264617919923, 0.017753087997436523, 0.017773567199707033, 0.017846271514892577, 0.01773465538024902, 0.017744895935058593, 0.017744895935058593, 0.01779302406311035, 0.017810432434082032, 0.017755136489868165, 0.01779302406311035, 0.017524736404418945, 0.017514495849609374, 0.017754112243652344, 0.017769472122192383, 0.01780121612548828, 0.017702911376953127, 0.01778278350830078, 0.017833984375, 0.03754905700683594, 0.0177838077545166, 0.017723392486572266, 0.01768550491333008, 0.01768448066711426, 0.017692703247070313, 0.01792099189758301, 0.017960960388183594, 0.017671167373657228, 0.017859615325927735, 0.017755104064941407, 0.017901567459106444, 0.017895423889160156, 0.017762304306030274, 0.018351104736328124, 0.01788313674926758, 0.017729536056518554, 0.017673215866088866, 0.017712127685546874, 0.017727487564086913, 0.01766092872619629, 0.017682432174682617, 0.017710079193115236, 0.017709056854248048, 0.01781350326538086, 0.01775103950500488, 0.017681407928466796, 0.01776742362976074, 0.017525760650634766, 0.017476608276367187, 0.017549312591552735, 0.017491968154907226, 0.01744588851928711, 0.017564672470092774, 0.017537023544311522, 0.01749504089355469, 0.017494016647338868, 0.017838079452514647, 0.01777459144592285, 0.017760255813598632, 0.01780019187927246, 0.017656831741333007, 0.01779916763305664, 0.017733631134033204, 0.017695743560791014, 0.017674240112304687, 0.017529855728149413, 0.017499135971069335, 0.01740902328491211, 0.017492992401123047, 0.01742131233215332, 0.017546239852905272, 0.017521663665771483, 0.017451007843017577, 0.017605632781982423, 0.01802342414855957, 0.01777663993835449, 0.01780633544921875, 0.017872896194458008, 0.017796096801757814, 0.017735679626464843, 0.017752063751220702, 0.017717248916625978, 0.0375470085144043, 0.01798041534423828, 0.01781452751159668, 0.01777561569213867, 0.017777664184570312, 0.017763328552246094, 0.017730560302734375, 0.017747968673706056, 0.017732608795166017, 0.017708032608032227, 0.017707008361816406, 0.01781760025024414, 0.017702911376953127, 0.01776639938354492, 0.017696767807006835, 0.01769267272949219, 0.017847295761108398, 0.017771520614624024, 0.01760051155090332, 0.017572864532470703, 0.017490943908691405, 0.01780838394165039, 0.01778483200073242, 0.017707008361816406, 0.01746329689025879, 0.01763737678527832, 0.017810432434082032, 0.017733631134033204, 0.017777664184570312, 0.01759129524230957, 0.01744179153442383, 0.017794048309326172, 0.017801279067993163, 0.017785791397094727, 0.017781759262084963, 0.017750015258789064, 0.01772755241394043, 0.01774176025390625, 0.017785856246948242, 0.01778895950317383, 0.017408000946044923, 0.017538015365600585, 0.01767628860473633, 0.017797119140625, 0.017705984115600586, 0.01760665512084961, 0.01741107177734375, 0.017911808013916015, 0.017724416732788087, 0.017768447875976562, 0.01783500862121582, 0.01784217643737793, 0.01784934425354004, 0.017740800857543947, 0.017727487564086913, 0.01748684883117676, 0.017468416213989257, 0.017688575744628905, 0.017757183074951173, 0.017673215866088866, 0.017934335708618163, 0.01785753631591797, 0.017747968673706056, 0.03746406555175781, 0.017761280059814453, 0.01762918472290039, 0.01744076728820801, 0.01757798385620117, 0.017718271255493166, 0.017743871688842772, 0.01766912078857422, 0.017725439071655275, 0.017797119140625, 0.017682432174682617, 0.01775004768371582, 0.01771107292175293, 0.017797119140625, 0.017770496368408203, 0.017787904739379884, 0.01768550491333008, 0.017687551498413084, 0.017704959869384765, 0.01779814338684082, 0.017702911376953127, 0.01779302406311035, 0.017755136489868165, 0.017667072296142578, 0.01785241508483887, 0.017755136489868165, 0.017722368240356445, 0.017752063751220702, 0.017694719314575197, 0.017682432174682617, 0.017657855987548828, 0.017819648742675782, 0.0175861759185791, 0.017703935623168944, 0.017757183074951173, 0.01765171241760254, 0.017705984115600586, 0.017765375137329103, 0.01765990447998047, 0.01757900810241699, 0.018147327423095702, 0.02009600067138672, 0.01846886444091797, 0.018339839935302735, 0.017802240371704102, 0.01784115219116211, 0.01780121612548828, 0.017709056854248048, 0.017761280059814453, 0.0177807674407959, 0.01766294479370117, 0.017320959091186524, 0.016931840896606445, 0.01679769515991211, 0.016695295333862305, 0.0168089599609375, 0.01666662406921387, 0.016684032440185546, 0.016743423461914063, 0.016713727951049806, 0.01681100845336914, 0.016916479110717773, 0.016869375228881836, 0.036615169525146485, 0.016953344345092772, 0.016892927169799805, 0.016926719665527345, 0.016941055297851563, 0.016905216217041014, 0.016889856338500975, 0.016903200149536134, 0.016926687240600587, 0.016855039596557618, 0.016750591278076172, 0.016792575836181642, 0.016792575836181642, 0.01681715202331543, 0.016776191711425782, 0.016773120880126953, 0.01679052734375, 0.01677004814147949, 0.016762943267822267, 0.01679865646362305, 0.01759231948852539, 0.017087488174438475, 0.016904191970825197, 0.016900096893310547, 0.01684787178039551, 0.016928768157958983, 0.01703424072265625, 0.01700044822692871, 0.01696463966369629, 0.016958431243896486, 0.016948223114013672, 0.016913408279418944, 0.016921600341796874, 0.016924671173095703, 0.017105920791625977, 0.01699839973449707, 0.016933887481689454, 0.01695232009887695, 0.017077247619628907, 0.017062911987304686, 0.01700454330444336, 0.016949247360229493, 0.016945152282714843, 0.016932863235473633, 0.016895999908447267, 0.01685606384277344, 0.01699942398071289, 0.017191999435424803, 0.016944063186645507, 0.016890880584716796, 0.016948223114013672, 0.016889856338500975, 0.01700556755065918, 0.016946176528930663, 0.016942079544067384, 0.016912384033203123, 0.016894975662231446, 0.016902143478393555, 0.016888832092285155, 0.016898048400878905, 0.016878591537475587, 0.016914432525634765, 0.016924671173095703, 0.036580352783203124, 0.016909311294555664, 0.016905216217041014, 0.016869375228881836, 0.016904256820678712, 0.016869312286376954, 0.01676595115661621, 0.01683251190185547, 0.016913408279418944, 0.016890880584716796, 0.01684480094909668, 0.016977920532226562, 0.016920576095581053, 0.0168724479675293, 0.016862207412719727, 0.016934911727905275, 0.016939008712768554, 0.016947200775146484, 0.017140735626220704, 0.016925695419311524, 0.016917503356933594, 0.016869375228881836, 0.016957439422607423, 0.016869375228881836, 0.016745471954345705, 0.016750591278076172, 0.01684480094909668, 0.016875520706176757, 0.016962591171264647, 0.017020896911621095, 0.0169564151763916, 0.016917503356933594, 0.016889856338500975, 0.016925695419311524, 0.016928768157958983, 0.01686835289001465, 0.016888832092285155, 0.016927743911743166, 0.016861183166503906, 0.016863231658935548, 0.016946176528930663, 0.016723968505859374, 0.01701375961303711, 0.016940031051635742, 0.016894975662231446, 0.016903167724609376, 0.017778688430786133, 0.017624063491821287, 0.01766912078857422, 0.017708032608032227, 0.01763737678527832, 0.017683456420898438, 0.0176680965423584, 0.017656831741333007, 0.01767628860473633, 0.017722368240356445, 0.017747968673706056, 0.017617919921875, 0.017492992401123047, 0.017539072036743163, 0.017688575744628905, 0.017716224670410157, 0.017570816040039062, 0.0366561279296875, 0.01699942398071289, 0.016900096893310547, 0.017897472381591797, 0.01784524726867676, 0.01762611198425293, 0.017694719314575197, 0.01736396789550781, 0.01777663993835449, 0.017726463317871095, 0.017678335189819337, 0.01775103950500488, 0.01763020706176758, 0.01762303924560547, 0.01762099266052246, 0.01760256004333496, 0.01768448066711426, 0.017715200424194336, 0.01763430404663086, 0.017666048049926757, 0.01767628860473633, 0.01765068817138672, 0.01785958480834961, 0.017679359436035155, 0.01764352035522461, 0.017698816299438477, 0.0178657283782959, 0.01766912078857422, 0.01740083122253418, 0.017321983337402345, 0.017284095764160155, 0.017510400772094727, 0.017604608535766602, 0.01763839912414551, 0.01764556884765625, 0.017546239852905272, 0.017330175399780275, 0.017311744689941407, 0.01734758377075195, 0.017346559524536134, 0.017286144256591796, 0.017346559524536134, 0.017350656509399414, 0.017344512939453126, 0.017362943649291994, 0.018486272811889647, 0.018918399810791017, 0.017944576263427735, 0.017695743560791014, 0.01777561569213867, 0.017500160217285156, 0.017252351760864256, 0.017312768936157227, 0.017314815521240236, 0.01764659118652344, 0.017624063491821287, 0.01764352035522461, 0.01770086479187012, 0.017698816299438477, 0.01765376091003418, 0.017006591796875, 0.016745471954345705, 0.016749568939208984]",tokens/s,56.24966458088899,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1278.517248,6797.39392,0.0,6150.946816,5988.307968,s,10,0.30529612541198725,0.030529612541198726,0.00029578028387582366,0.0304357271194458,0.030562937927246094,0.030988860702514647,0.03132959892272949,"[0.031414783477783204, 0.030425983428955077, 0.030416704177856444, 0.030448543548583985, 0.030395135879516602, 0.030442911148071287, 0.03040652847290039, 0.030428543090820312, 0.03044870376586914, 0.03046828842163086]",tokens/s,8385.301308836995,kWh,3.621152406884342e-07,1.983538829961684e-07,1.7911295334801126e-06,2.351598657164715e-06,tokens/kWh,108862113.5328658,MB,1278.517248,6797.39392,0.0,6150.946816,6117.056512,s,10,13.401929077148438,1.3401929077148438,0.0071886328834873696,1.3391688232421877,1.348996240234375,1.3506322143554688,1.3519409936523439,"[1.3486326904296875, 1.3366287841796876, 1.33829296875, 1.3279244384765625, 1.3390897216796875, 1.331222900390625, 1.3392479248046876, 1.3414462890625, 1.3522681884765626, 1.3471751708984374]",tokens/s,47.0081580325783,kWh,1.6584539956587576e-05,9.086562614075947e-06,5.5923666638320975e-05,8.15947692089845e-05,tokens/kWh,772108.317858481,,s,629,13.652671457290648,0.021705360027489108,0.0036889444810717584,0.021188608169555666,0.021647769546508787,0.021866086959838868,0.052123201904296874,"[0.02315158462524414, 0.021507072448730468, 0.021169151306152344, 0.021189632415771483, 0.02108723258972168, 0.021118976593017577, 0.02105958366394043, 0.021150720596313476, 0.02145996856689453, 0.02214297676086426, 0.021568511962890623, 0.021142528533935546, 0.021331968307495116, 0.02123263931274414, 0.021134336471557616, 0.021182464599609374, 0.021183551788330077, 0.02118137550354004, 0.021149696350097655, 0.021133312225341795, 0.021165056228637694, 0.02110361671447754, 0.021190656661987304, 0.021172224044799806, 0.021176319122314453, 0.02168934440612793, 0.02230784034729004, 0.022359039306640623, 0.022817792892456053, 0.021707775115966797, 0.021389312744140625, 0.021339136123657225, 0.021132287979125978, 0.021615615844726564, 0.021440511703491212, 0.021191743850708007, 0.021265344619750978, 0.021350400924682617, 0.02145484733581543, 0.021112831115722656, 0.021153791427612305, 0.02129100799560547, 0.021210111618041993, 0.021176319122314453, 0.020975616455078124, 0.021134336471557616, 0.02128179168701172, 0.021424127578735352, 0.021164031982421876, 0.02147123146057129, 0.021137407302856445, 0.021136383056640624, 0.021198911666870116, 0.021197759628295898, 0.021564416885375977, 0.021275648117065428, 0.021738496780395508, 0.02145587158203125, 0.021147647857666017, 0.021404672622680664, 0.021176319122314453, 0.021209087371826172, 0.05206220626831055, 0.021217279434204102, 0.021114879608154297, 0.021175296783447265, 0.021149696350097655, 0.021136383056640624, 0.02111692810058594, 0.021132287979125978, 0.020974592208862306, 0.021183488845825195, 0.021378047943115236, 0.021217279434204102, 0.021171199798583985, 0.021161983489990235, 0.02151219177246094, 0.021953535079956055, 0.021575679779052736, 0.021283840179443358, 0.021097471237182617, 0.021204992294311522, 0.021110784530639647, 0.02124595260620117, 0.02128179168701172, 0.02112512016296387, 0.021160959243774414, 0.021156864166259767, 0.021206016540527343, 0.021168127059936523, 0.020938751220703124, 0.020883455276489257, 0.021525503158569336, 0.021367839813232422, 0.021205984115600585, 0.02125619125366211, 0.02105036735534668, 0.02100223922729492, 0.02109644889831543, 0.021209087371826172, 0.021560319900512694, 0.021194751739501954, 0.021209087371826172, 0.02127667236328125, 0.02122035217285156, 0.02151628875732422, 0.021166080474853514, 0.02123366355895996, 0.02181427192687988, 0.021175296783447265, 0.021175359725952147, 0.021204927444458007, 0.021198848724365234, 0.021195775985717775, 0.02119987106323242, 0.02116409683227539, 0.021214143753051758, 0.021202943801879884, 0.021249088287353515, 0.021211103439331056, 0.020974559783935545, 0.020967424392700194, 0.020986879348754883, 0.020982784271240236, 0.020980735778808594, 0.05229363250732422, 0.021223424911499023, 0.0212674560546875, 0.021357568740844726, 0.02119987106323242, 0.021359615325927735, 0.021277727127075194, 0.021219295501708986, 0.021334016799926758, 0.020937728881835937, 0.020953088760375976, 0.02096335983276367, 0.020942815780639647, 0.021489664077758788, 0.021152767181396484, 0.020988927841186524, 0.02105548858642578, 0.021020671844482423, 0.021046272277832033, 0.020988927841186524, 0.021013504028320314, 0.021296159744262695, 0.020981727600097658, 0.020942848205566408, 0.02104217529296875, 0.021198848724365234, 0.021210111618041993, 0.021167104721069335, 0.021308416366577147, 0.021634048461914062, 0.021225536346435547, 0.021304256439208986, 0.021321727752685548, 0.021328895568847657, 0.021288991928100586, 0.0211507511138916, 0.02122438430786133, 0.02168627166748047, 0.02142617607116699, 0.02145587158203125, 0.02143948745727539, 0.0212674560546875, 0.021184511184692383, 0.021235712051391603, 0.021432319641113282, 0.021646335601806642, 0.02123263931274414, 0.02124185562133789, 0.021122047424316406, 0.021097471237182617, 0.021448703765869142, 0.021118976593017577, 0.021316608428955077, 0.021188608169555666, 0.021110784530639647, 0.021191680908203125, 0.021106687545776368, 0.02122035217285156, 0.02162073516845703, 0.021327871322631836, 0.0216494083404541, 0.021144575119018554, 0.021132287979125978, 0.0521278076171875, 0.02114348793029785, 0.02109337615966797, 0.021195775985717775, 0.021113855361938477, 0.021104639053344726, 0.021105663299560547, 0.02108518409729004, 0.02085478401184082, 0.020959264755249025, 0.02088956832885742, 0.021190656661987304, 0.021115903854370118, 0.021073919296264648, 0.021061632156372072, 0.020925439834594727, 0.020864000320434572, 0.02088140869140625, 0.020875328063964842, 0.0211342716217041, 0.021313535690307618, 0.02089369583129883, 0.02084966468811035, 0.021094400405883788, 0.02087116813659668, 0.020932607650756836, 0.020891679763793945, 0.02094076728820801, 0.020950048446655274, 0.02113225555419922, 0.020961280822753905, 0.020908031463623047, 0.02084659194946289, 0.021204992294311522, 0.02112512016296387, 0.021218303680419923, 0.021189632415771483, 0.02124083137512207, 0.021145599365234375, 0.021166080474853514, 0.021150720596313476, 0.021180416107177736, 0.021139455795288087, 0.021213184356689452, 0.021121023178100586, 0.020936767578125, 0.020938688278198243, 0.020944896697998046, 0.020924415588378906, 0.021013504028320314, 0.021109760284423826, 0.021324800491333007, 0.02121625518798828, 0.021170175552368165, 0.021173248291015623, 0.021180448532104493, 0.021215200424194336, 0.021215232849121093, 0.021196800231933592, 0.021197824478149413, 0.021211135864257814, 0.020909055709838868, 0.02105958366394043, 0.05219225692749024, 0.021195775985717775, 0.021131263732910157, 0.021156864166259767, 0.021120000839233398, 0.02104934310913086, 0.02109644889831543, 0.021144575119018554, 0.021117952346801756, 0.021128192901611328, 0.021160959243774414, 0.021140512466430665, 0.021258207321166993, 0.02167500877380371, 0.02167091178894043, 0.021138431549072266, 0.021149696350097655, 0.02148863983154297, 0.02122956848144531, 0.021140544891357423, 0.021104576110839844, 0.021189632415771483, 0.02123161506652832, 0.02110054397583008, 0.021124095916748048, 0.021136383056640624, 0.021395456314086913, 0.021560319900512694, 0.021188608169555666, 0.021139455795288087, 0.021227519989013673, 0.0218787841796875, 0.02181427192687988, 0.021749759674072267, 0.022569984436035157, 0.021801984786987305, 0.021146623611450196, 0.021138431549072266, 0.021151744842529296, 0.021173248291015623, 0.02120806312561035, 0.021485567092895508, 0.021615615844726564, 0.02122547149658203, 0.021111808776855468, 0.021169151306152344, 0.021153791427612305, 0.021166080474853514, 0.02109443283081055, 0.02115171241760254, 0.021118976593017577, 0.021123071670532227, 0.021133312225341795, 0.021165056228637694, 0.021214208602905273, 0.020917280197143555, 0.020961248397827148, 0.021195775985717775, 0.020981760025024415, 0.021275648117065428, 0.020948991775512696, 0.020929536819458007, 0.020956159591674805, 0.05220454406738281, 0.021105663299560547, 0.021098495483398438, 0.021003263473510742, 0.020945920944213867, 0.020917247772216797, 0.02105753517150879, 0.02113747215270996, 0.021320640563964845, 0.02128998374938965, 0.021177343368530274, 0.02109644889831543, 0.021081087112426757, 0.021127168655395507, 0.02088652801513672, 0.020917247772216797, 0.021206016540527343, 0.021145599365234375, 0.02118662452697754, 0.02097862434387207, 0.021014528274536134, 0.020933631896972657, 0.02092748832702637, 0.021242879867553712, 0.021271551132202148, 0.020993024826049804, 0.021128192901611328, 0.021403648376464843, 0.02124799919128418, 0.02107187271118164, 0.020920320510864256, 0.020946943283081054, 0.02092748832702637, 0.020980735778808594, 0.020956159591674805, 0.021134336471557616, 0.021124095916748048, 0.02126233673095703, 0.021165056228637694, 0.021515264511108398, 0.0212490234375, 0.020935680389404295, 0.02106470489501953, 0.020941823959350587, 0.020913152694702147, 0.020934656143188478, 0.02102681541442871, 0.021110784530639647, 0.021444608688354492, 0.021179391860961915, 0.02127462387084961, 0.021159936904907226, 0.0212490234375, 0.021129215240478515, 0.021122047424316406, 0.021401599884033205, 0.021362688064575194, 0.021210111618041993, 0.021203968048095705, 0.021150720596313476, 0.021230592727661132, 0.021167104721069335, 0.021178367614746094, 0.052111358642578126, 0.021118976593017577, 0.020980735778808594, 0.0208855037689209, 0.02104425621032715, 0.021075935363769532, 0.021207040786743164, 0.021715967178344727, 0.021116992950439454, 0.021187519073486327, 0.021124095916748048, 0.021139488220214844, 0.021084192276000977, 0.020877248764038087, 0.020890623092651366, 0.020932607650756836, 0.02086911964416504, 0.021407743453979493, 0.02165862464904785, 0.02149171257019043, 0.02147327995300293, 0.02104012870788574, 0.02102374458312988, 0.02147942352294922, 0.021346303939819337, 0.02164735984802246, 0.021155839920043946, 0.021391359329223633, 0.02229350471496582, 0.021433343887329103, 0.021144575119018554, 0.021138431549072266, 0.021134336471557616, 0.021259263992309572, 0.021135360717773437, 0.0210831356048584, 0.021127168655395507, 0.021142528533935546, 0.021374975204467773, 0.021348352432250976, 0.021121023178100586, 0.021213247299194337, 0.021127103805541992, 0.021192703247070312, 0.021142528533935546, 0.021227519989013673, 0.02147942352294922, 0.021687295913696288, 0.021188608169555666, 0.021158912658691405, 0.021123071670532227, 0.021176319122314453, 0.021159936904907226, 0.021210111618041993, 0.021604352951049805, 0.02123161506652832, 0.021191680908203125, 0.02123980712890625, 0.02152038383483887, 0.02187980842590332, 0.02136479949951172, 0.02115782356262207, 0.021164031982421876, 0.05229363250732422, 0.02128691291809082, 0.021124160766601563, 0.02119980812072754, 0.021124095916748048, 0.021359615325927735, 0.02163199996948242, 0.02129100799560547, 0.021139455795288087, 0.021135360717773437, 0.021113855361938477, 0.0211015682220459, 0.021213184356689452, 0.02147331237792969, 0.021088224411010742, 0.021138431549072266, 0.02122550392150879, 0.02112406349182129, 0.021171199798583985, 0.021111808776855468, 0.021135360717773437, 0.0214835205078125, 0.021629951477050782, 0.021530624389648437, 0.021300224304199217, 0.02108518409729004, 0.02119987106323242, 0.021140480041503908, 0.021122047424316406, 0.02162892723083496, 0.021175296783447265, 0.02108415985107422, 0.021073919296264648, 0.021102592468261717, 0.021137407302856445, 0.02111692810058594, 0.021203968048095705, 0.021312511444091797, 0.021389312744140625, 0.021161983489990235, 0.021180416107177736, 0.021223424911499023, 0.021179391860961915, 0.021172224044799806, 0.021183488845825195, 0.021178367614746094, 0.021189632415771483, 0.02119990348815918, 0.021215200424194336, 0.022297599792480468, 0.021877824783325197, 0.021234624862670897, 0.021341184616088867, 0.021515264511108398, 0.02147430419921875, 0.021747711181640626, 0.021585920333862304, 0.021187583923339845, 0.021350400924682617, 0.021182464599609374, 0.021379072189331053, 0.02148249626159668, 0.021196800231933592, 0.05266636657714844, 0.021631040573120118, 0.021608383178710937, 0.02187468719482422, 0.021363712310791014, 0.021097471237182617, 0.02206822395324707, 0.021594112396240234, 0.02127667236328125, 0.021204992294311522, 0.021975040435791016, 0.021763071060180664, 0.021630975723266603, 0.02165247917175293, 0.021614591598510743, 0.02164531135559082, 0.021626880645751953, 0.02165043258666992, 0.021156864166259767, 0.021204992294311522, 0.021514240264892577, 0.021634048461914062, 0.021575679779052736, 0.02127462387084961, 0.02108415985107422, 0.021165056228637694, 0.021135360717773437, 0.021395456314086913, 0.021733375549316408, 0.02247065544128418, 0.02149478340148926, 0.021177343368530274, 0.021144575119018554, 0.021160959243774414, 0.021287935256958008, 0.021737472534179687, 0.022750207901000977, 0.02188800048828125, 0.021184576034545897, 0.021054399490356444, 0.021436416625976562, 0.021634048461914062, 0.02187264060974121, 0.021333055496215822, 0.021095359802246094, 0.02122035217285156, 0.021159999847412108, 0.02117420768737793, 0.021149696350097655, 0.02167705535888672, 0.021754880905151368, 0.02129817581176758, 0.0211015682220459, 0.021127168655395507, 0.021140480041503908, 0.02105548858642578, 0.02185625648498535, 0.021533695220947266, 0.021393407821655275, 0.02123776054382324, 0.021102592468261717, 0.021171199798583985, 0.021081087112426757, 0.052157440185546876, 0.021506048202514647, 0.021296127319335938, 0.02109644889831543, 0.02110361671447754, 0.021959680557250977, 0.021820415496826173, 0.021610496520996093, 0.021731327056884766, 0.022018047332763673, 0.022477823257446287, 0.021622783660888673, 0.021420032501220702, 0.02158799934387207, 0.021698528289794922, 0.021127168655395507, 0.02166169548034668, 0.021218303680419923, 0.021341184616088867, 0.02129206466674805, 0.021461984634399415, 0.021774335861206053, 0.02163302421569824, 0.021148672103881837, 0.02122035217285156, 0.021359615325927735, 0.02150297546386719, 0.02135865592956543, 0.021761983871459962, 0.02112512016296387, 0.02146816062927246, 0.021209087371826172, 0.021767168045043944, 0.021223424911499023, 0.021212160110473634, 0.021576736450195314, 0.02110358428955078, 0.021202943801879884, 0.021251071929931642, 0.021410816192626952, 0.021275648117065428, 0.021175296783447265, 0.021154815673828126, 0.021187583923339845, 0.021127168655395507, 0.021149696350097655, 0.02115385627746582, 0.021178304672241213, 0.021600255966186522, 0.021304319381713867, 0.021184511184692383, 0.021212160110473634, 0.021193727493286133, 0.020982847213745118, 0.020898752212524414, 0.020959232330322267, 0.020960256576538085, 0.020980735778808594, 0.021381120681762695, 0.02190540885925293, 0.021721088409423828, 0.02125823974609375, 0.021195775985717775]",tokens/s,46.07156936044984,,,main,False,False -bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12656,7 +12656,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpfvyql89x/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12949,7 +12949,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpgmr4nipx/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13465,7 +13465,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1278.38208,1713.897472,0.0,1067.450368,1018.226688,s,10,0.1711367034912109,0.017113670349121094,0.0006044720308494674,0.017062576293945312,0.017929014587402342,0.018202859115600586,0.01842193473815918,"[0.018476703643798827, 0.017089599609375, 0.016498239517211914, 0.017047168731689454, 0.017160512924194335, 0.016970815658569335, 0.017868160247802734, 0.01707798385620117, 0.016465215682983397, 0.016482303619384766]",tokens/s,14958.801635041858,kWh,1.9755586151162386e-07,1.0825147990690798e-07,5.015727721237629e-07,8.073801135422947e-07,tokens/kWh,317074938.6888254,MB,1278.38208,1713.897472,0.0,1067.450368,1059.39456,s,10,10.570229980468751,1.057022998046875,0.01434517760269983,1.0608662719726563,1.0653480590820312,1.071996600341797,1.0773154333496093,"[1.0786451416015626, 1.0614322509765626, 1.0485587158203125, 1.0598475341796876, 1.06387060546875, 1.0626763916015625, 1.061619140625, 1.06030029296875, 1.020062744140625, 1.0532171630859375]",tokens/s,59.601352209373765,kWh,1.2347086212728956e-05,6.765710445814322e-06,2.1629694932874916e-05,4.07424915914182e-05,tokens/kWh,1546297.183583883,,s,629,10.7047331676483,0.017018653684655508,0.0020937808719674184,0.01680588722229004,0.01699450836181641,0.017415782165527343,0.03315052490234375,"[0.017915903091430666, 0.0176363525390625, 0.017562623977661132, 0.01740902328491211, 0.017509376525878906, 0.01741619110107422, 0.01743974494934082, 0.01756159973144531, 0.01738444709777832, 0.017515520095825195, 0.01739776039123535, 0.017354751586914064, 0.017535999298095704, 0.01760665512084961, 0.0174202880859375, 0.017514495849609374, 0.01740390396118164, 0.017494016647338868, 0.017672191619873046, 0.017306623458862306, 0.017474559783935546, 0.01769267272949219, 0.01754521560668945, 0.01742233657836914, 0.016741376876831054, 0.016734207153320312, 0.016705535888671876, 0.016700416564941405, 0.017678335189819337, 0.017415168762207032, 0.0172042236328125, 0.016962560653686523, 0.016944128036499022, 0.016825344085693358, 0.016714752197265623, 0.016738304138183592, 0.01679769515991211, 0.01680281639099121, 0.016870399475097657, 0.016694271087646484, 0.0168222713470459, 0.01675468826293945, 0.016842752456665038, 0.016837631225585938, 0.016772096633911132, 0.01677414321899414, 0.016746496200561522, 0.016700416564941405, 0.01683865547180176, 0.016744447708129884, 0.016858112335205077, 0.01675468826293945, 0.016859136581420898, 0.01679667282104492, 0.016776191711425782, 0.01670867156982422, 0.016846847534179688, 0.01671878433227539, 0.016760831832885743, 0.01681715202331543, 0.016878591537475587, 0.01684787178039551, 0.034307071685791016, 0.01676288032531738, 0.016781312942504883, 0.016820287704467772, 0.01678335952758789, 0.01689695930480957, 0.016829439163208008, 0.01681920051574707, 0.016767999649047852, 0.016933887481689454, 0.01680793571472168, 0.017054719924926756, 0.016825344085693358, 0.016721920013427736, 0.01678339195251465, 0.016764896392822266, 0.016841728210449217, 0.016765983581542968, 0.016816095352172853, 0.01677107238769531, 0.017076223373413087, 0.016883712768554687, 0.01681612777709961, 0.01677414321899414, 0.016990207672119142, 0.016911359786987306, 0.01681305694580078, 0.016841728210449217, 0.016842752456665038, 0.016939008712768554, 0.01681100845336914, 0.01680793571472168, 0.01686425590515137, 0.016858112335205077, 0.016897024154663084, 0.01678643226623535, 0.01679871940612793, 0.0168222713470459, 0.016845823287963867, 0.01698099136352539, 0.01678745651245117, 0.016850944519042968, 0.01677414321899414, 0.01685606384277344, 0.01683660888671875, 0.016849920272827147, 0.016753664016723634, 0.016878591537475587, 0.016842752456665038, 0.016767999649047852, 0.016804927825927733, 0.0167126407623291, 0.0168222713470459, 0.017168384552001953, 0.016925695419311524, 0.01680179214477539, 0.01682636833190918, 0.016888832092285155, 0.016784383773803712, 0.016837631225585938, 0.016861183166503906, 0.01684480094909668, 0.016865280151367186, 0.03311513519287109, 0.016079872131347657, 0.016054271697998047, 0.01617407989501953, 0.016210943222045898, 0.01614028739929199, 0.01617100715637207, 0.016243711471557617, 0.016130048751831053, 0.016185375213623048, 0.016128992080688475, 0.01620479965209961, 0.016130048751831053, 0.01617203140258789, 0.016676864624023437, 0.01662156867980957, 0.016734272003173827, 0.016788415908813477, 0.0167956485748291, 0.01679871940612793, 0.016792575836181642, 0.01685196876525879, 0.01678335952758789, 0.01677824020385742, 0.016889856338500975, 0.016748544692993163, 0.016763904571533202, 0.016746496200561522, 0.01677107238769531, 0.016729087829589845, 0.016717824935913086, 0.016709632873535156, 0.016737279891967775, 0.01676697540283203, 0.0168089599609375, 0.016707584381103514, 0.016772096633911132, 0.01665433692932129, 0.016546815872192384, 0.016548864364624022, 0.01656831932067871, 0.01680076789855957, 0.01682636833190918, 0.01704550361633301, 0.01692163276672363, 0.01686832046508789, 0.016888832092285155, 0.016829439163208008, 0.016906240463256835, 0.016746496200561522, 0.01685196876525879, 0.016767999649047852, 0.01682022476196289, 0.016829439163208008, 0.01683148765563965, 0.016578559875488282, 0.016721920013427736, 0.01681920051574707, 0.0168089599609375, 0.01684889602661133, 0.01684889602661133, 0.016788480758666992, 0.01679052734375, 0.0344453125, 0.016760831832885743, 0.016747520446777343, 0.01675775909423828, 0.016744447708129884, 0.0168089599609375, 0.016782367706298828, 0.016795616149902343, 0.016745471954345705, 0.01677004814147949, 0.016849920272827147, 0.016886783599853517, 0.016883712768554687, 0.016792575836181642, 0.01679769515991211, 0.01678335952758789, 0.016873472213745116, 0.016793600082397463, 0.01678233528137207, 0.01682636833190918, 0.01681715202331543, 0.01680384063720703, 0.016781312942504883, 0.016855039596557618, 0.016827392578125, 0.01684889602661133, 0.01682022476196289, 0.016852991104125976, 0.01677414321899414, 0.016788543701171874, 0.016879552841186522, 0.01677004814147949, 0.01680281639099121, 0.01680793571472168, 0.01680076789855957, 0.016821247100830078, 0.01677926445007324, 0.016753664016723634, 0.01681920051574707, 0.0168724479675293, 0.01676288032531738, 0.016887807846069337, 0.016788480758666992, 0.016764928817749023, 0.016845823287963867, 0.01722163200378418, 0.01670662307739258, 0.016764863967895508, 0.016740352630615234, 0.016862207412719727, 0.016849920272827147, 0.016827392578125, 0.016772096633911132, 0.01676595115661621, 0.016769023895263673, 0.016880640029907225, 0.01683251190185547, 0.016883712768554687, 0.01680998420715332, 0.016850944519042968, 0.01681100845336914, 0.01679667282104492, 0.016827392578125, 0.034411518096923825, 0.016741376876831054, 0.01679052734375, 0.016880640029907225, 0.01681715202331543, 0.0167587833404541, 0.01679974365234375, 0.01681920051574707, 0.01677004814147949, 0.016827392578125, 0.016803871154785155, 0.017103839874267578, 0.016885759353637696, 0.01676192092895508, 0.01679046440124512, 0.016842752456665038, 0.01683558464050293, 0.0168222713470459, 0.016849920272827147, 0.017146879196166993, 0.01835212707519531, 0.01741926383972168, 0.016946176528930663, 0.016669696807861328, 0.016795679092407225, 0.016792543411254884, 0.016785408020019533, 0.01684480094909668, 0.01702400016784668, 0.01662566375732422, 0.016874496459960937, 0.016769023895263673, 0.017132543563842775, 0.017356800079345702, 0.016907264709472656, 0.01681612777709961, 0.01682329559326172, 0.01675775909423828, 0.016529407501220703, 0.016728063583374024, 0.016826431274414064, 0.01677510452270508, 0.01676700782775879, 0.016827360153198242, 0.016903167724609376, 0.01681612777709961, 0.016910335540771485, 0.01682329559326172, 0.016756736755371093, 0.01681305694580078, 0.01682329559326172, 0.016752639770507814, 0.016792575836181642, 0.016683008193969725, 0.016594944000244142, 0.016623615264892578, 0.016518144607543944, 0.01702809524536133, 0.01704140853881836, 0.017705984115600586, 0.017150976181030272, 0.01684787178039551, 0.016837631225585938, 0.03449651336669922, 0.016915456771850586, 0.017155071258544922, 0.016904191970825197, 0.016784383773803712, 0.01680281639099121, 0.01677516746520996, 0.017356800079345702, 0.016895008087158204, 0.016777183532714842, 0.016731136322021483, 0.016880735397338868, 0.016845727920532228, 0.016830495834350586, 0.01685091209411621, 0.01677414321899414, 0.01722470474243164, 0.016881696701049803, 0.01688470458984375, 0.0168222713470459, 0.01678643226623535, 0.01683967971801758, 0.01680281639099121, 0.016902143478393555, 0.016865280151367186, 0.016886783599853517, 0.016825344085693358, 0.016863231658935548, 0.016776191711425782, 0.01683046340942383, 0.01678233528137207, 0.016785408020019533, 0.016982015609741212, 0.016902143478393555, 0.01683558464050293, 0.016884735107421875, 0.016909311294555664, 0.016873472213745116, 0.016907264709472656, 0.016908287048339844, 0.01680899238586426, 0.01685807991027832, 0.01682022476196289, 0.016761856079101564, 0.016820255279541015, 0.016922592163085937, 0.016743423461914063, 0.01675468826293945, 0.01681920051574707, 0.016874496459960937, 0.016776191711425782, 0.01678233528137207, 0.016785408020019533, 0.016897024154663084, 0.016772096633911132, 0.01686944007873535, 0.016851903915405274, 0.016892927169799805, 0.016778303146362306, 0.016824256896972655, 0.016816192626953125, 0.01689900779724121, 0.01683660888671875, 0.036327423095703124, 0.01703219223022461, 0.01710393524169922, 0.016817087173461913, 0.016899072647094726, 0.016750591278076172, 0.016756767272949218, 0.016770015716552733, 0.0168222713470459, 0.016827392578125, 0.01684889602661133, 0.017040384292602538, 0.01682022476196289, 0.016733184814453125, 0.016862207412719727, 0.016788480758666992, 0.016781312942504883, 0.016882688522338866, 0.01682329559326172, 0.01659187126159668, 0.01661337661743164, 0.016764928817749023, 0.01683456039428711, 0.01681612777709961, 0.01678233528137207, 0.01680076789855957, 0.01684889602661133, 0.01679974365234375, 0.01685196876525879, 0.016784383773803712, 0.01676288032531738, 0.01678745651245117, 0.01684889602661133, 0.01682022476196289, 0.01683967971801758, 0.01679667282104492, 0.016899072647094726, 0.01677926445007324, 0.01672300720214844, 0.016747455596923828, 0.016737279891967775, 0.01683558464050293, 0.016785408020019533, 0.016846847534179688, 0.01685196876525879, 0.01682841682434082, 0.01677926445007324, 0.0168089599609375, 0.016728063583374024, 0.016881664276123046, 0.016862207412719727, 0.016842752456665038, 0.01679052734375, 0.0168724479675293, 0.016909311294555664, 0.016869375228881836, 0.01679974365234375, 0.01685203170776367, 0.01689900779724121, 0.016879615783691407, 0.016969728469848632, 0.01680281639099121, 0.016916479110717773, 0.03441049575805664, 0.01678233528137207, 0.016878591537475587, 0.016867328643798828, 0.016906272888183593, 0.016855072021484375, 0.016803775787353516, 0.01684889602661133, 0.016743423461914063, 0.01686425590515137, 0.016777215957641603, 0.01679155158996582, 0.01683148765563965, 0.01679871940612793, 0.016691200256347655, 0.01658060836791992, 0.016900096893310547, 0.016781312942504883, 0.01680588722229004, 0.01658880043029785, 0.016672767639160157, 0.016756736755371093, 0.016767999649047852, 0.01685196876525879, 0.01683456039428711, 0.01684480094909668, 0.01677414321899414, 0.016829439163208008, 0.0168089599609375, 0.0168407039642334, 0.016730112075805666, 0.01677824020385742, 0.01681510353088379, 0.016852991104125976, 0.01660825538635254, 0.016529407501220703, 0.016729087829589845, 0.016850944519042968, 0.016776191711425782, 0.016905216217041014, 0.016899072647094726, 0.016858112335205077, 0.016751615524291993, 0.01678335952758789, 0.01680486488342285, 0.01703321647644043, 0.01685606384277344, 0.0167956485748291, 0.01677824020385742, 0.016827455520629885, 0.016819135665893555, 0.01701171112060547, 0.016941055297851563, 0.01677516746520996, 0.016793600082397463, 0.01680384063720703, 0.01684377670288086, 0.01764659118652344, 0.01704652786254883, 0.01680691146850586, 0.016776191711425782, 0.016875520706176757, 0.016821247100830078, 0.033111038208007815, 0.016175104141235352, 0.01620582389831543, 0.0161474552154541, 0.016202751159667968, 0.016130048751831053, 0.01618841552734375, 0.01616896057128906, 0.01616998481750488, 0.016022527694702148, 0.016058368682861326, 0.016438272476196288, 0.016368640899658202, 0.016148544311523436, 0.016179136276245117, 0.016207872390747072, 0.016157695770263672, 0.016145408630371092, 0.016154624938964843, 0.016175104141235352, 0.016135168075561524, 0.016141311645507812, 0.01617715263366699, 0.016192512512207033, 0.016149503707885742, 0.01617203140258789, 0.01620582389831543, 0.01618841552734375, 0.01617817687988281, 0.01616486358642578, 0.016149503707885742, 0.016175167083740234, 0.016184255599975585, 0.016210943222045898, 0.016226303100585936, 0.016166912078857423, 0.01616281509399414, 0.01618022346496582, 0.01618022346496582, 0.016228351593017578, 0.016114688873291014, 0.016234495162963866, 0.01613926315307617, 0.01623040008544922, 0.01620582389831543, 0.01616896057128906, 0.016210943222045898, 0.01618841552734375, 0.016150527954101563, 0.01621708869934082, 0.01621708869934082, 0.016154624938964843, 0.016206911087036133, 0.016170944213867187, 0.01621504020690918, 0.016286720275878908, 0.01617100715637207, 0.016166912078857423, 0.01623859214782715, 0.016227327346801757, 0.016150527954101563, 0.01616998481750488, 0.01620992088317871, 0.03316428756713867, 0.01618022346496582, 0.01618841552734375, 0.0161976318359375, 0.01616383934020996, 0.0161474552154541, 0.01618022346496582, 0.016453632354736326, 0.016564224243164064, 0.016552959442138672, 0.016711679458618164, 0.016925695419311524, 0.01661235237121582, 0.016556032180786134, 0.016530431747436524, 0.016542720794677734, 0.016748544692993163, 0.01676288032531738, 0.01677516746520996, 0.01685196876525879, 0.016865280151367186, 0.01679052734375, 0.016910335540771485, 0.01702809524536133, 0.01679667282104492, 0.016855039596557618, 0.01680691146850586, 0.016842752456665038, 0.016894975662231446, 0.016906240463256835, 0.016724992752075195, 0.016850944519042968, 0.016756736755371093, 0.016846847534179688, 0.01680486488342285, 0.01681305694580078, 0.016785408020019533, 0.01678950309753418, 0.01676595115661621, 0.016663551330566406, 0.01663385581970215, 0.016707584381103514, 0.016869375228881836, 0.016927743911743166, 0.016885759353637696, 0.016894975662231446, 0.01682636833190918, 0.016753664016723634, 0.016761856079101564, 0.01681100845336914, 0.016777215957641603, 0.016864288330078126, 0.016827360153198242, 0.016713727951049806, 0.01682431983947754, 0.016841728210449217, 0.01683558464050293, 0.01680588722229004, 0.016679935455322266, 0.01680691146850586, 0.0167587833404541, 0.016886783599853517, 0.016726015090942382]",tokens/s,58.75905453682437,,,main,False,False -bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13727,7 +13727,7 @@ ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please req ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1155.096576,1455.947776,0.0,809.500672,723.533824,s,15,0.173844352722168,0.011589623514811197,0.0003799565019044181,0.01169200038909912,0.011827808189392091,0.012035174560546875,0.012418913116455078,"[0.012514847755432129, 0.011829600334167481, 0.011780799865722657, 0.011762816429138184, 0.011121055603027345, 0.011691712379455566, 0.011717568397521973, 0.011177023887634278, 0.011095871925354004, 0.011142815589904785, 0.011825119972229004, 0.01169200038909912, 0.011690336227416992, 0.011700032234191894, 0.011102751731872558]",tokens/s,22088.72442429554,kWh,1.3687280103232357e-07,7.499941463656173e-08,3.880742128667657e-07,5.99946428535651e-07,tokens/kWh,426704765.3318726,MB,1155.096576,1455.947776,0.0,809.500672,749.216256,s,15,10.518155212402345,0.7012103474934896,0.010939446017162301,0.7029287719726562,0.7118413452148438,0.7139651794433594,0.7139718444824219,"[0.7139735107421875, 0.7086609497070312, 0.7139616088867188, 0.7029287719726562, 0.7076468505859375, 0.7025215454101562, 0.70141259765625, 0.6947970581054688, 0.6885543212890625, 0.6679688720703125, 0.7027149047851563, 0.7057315063476562, 0.7060153198242187, 0.7033875732421875, 0.6978798217773438]",tokens/s,89.84465250006157,kWh,8.281585413480256e-06,4.537992137930547e-06,1.557408670500015e-05,2.839366425641095e-05,tokens/kWh,2218804.85135959,,s,944,10.667562987327564,0.011300384520474126,0.0014620249435332547,0.011165696144104004,0.011293696403503419,0.01140469717979431,0.022238607692718505,"[0.011946016311645508, 0.011907039642333984, 0.012149760246276856, 0.010962944030761718, 0.01061888027191162, 0.010612735748291016, 0.010579968452453613, 0.010592255592346191, 0.010800127983093261, 0.011176959991455078, 0.011200511932373047, 0.011199487686157226, 0.011238431930541993, 0.011479007720947265, 0.01143398380279541, 0.011431936264038087, 0.011265024185180664, 0.011222016334533692, 0.01135103988647461, 0.01123737621307373, 0.011149312019348144, 0.011206656455993653, 0.01123635196685791, 0.01124454402923584, 0.011256832122802735, 0.011200511932373047, 0.011184127807617187, 0.011299936294555665, 0.011191200256347657, 0.011272192001342773, 0.011169792175292969, 0.011192319869995117, 0.011144191741943359, 0.011204607963562012, 0.01117903995513916, 0.011198431968688964, 0.011172863960266113, 0.01124556827545166, 0.011199487686157226, 0.011174912452697755, 0.011217920303344727, 0.011229184150695801, 0.011172863960266113, 0.01120358371734619, 0.011223039627075194, 0.011226112365722657, 0.011174912452697755, 0.011188223838806152, 0.011195391654968261, 0.011210751533508301, 0.011783167839050293, 0.013306879997253418, 0.013592576026916504, 0.012147711753845216, 0.011752448081970214, 0.011341823577880859, 0.011339776039123535, 0.011367424011230469, 0.011053055763244628, 0.010977279663085938, 0.011018239974975585, 0.011346943855285644, 0.02332876777648926, 0.01091481590270996, 0.010882047653198243, 0.011111455917358399, 0.011453408241271973, 0.011422719955444336, 0.011257856369018555, 0.011213824272155762, 0.011287551879882812, 0.01124665641784668, 0.011330495834350586, 0.011282431602478027, 0.011202560424804688, 0.011185152053833008, 0.011208703994750976, 0.011286527633666991, 0.01124454402923584, 0.0112424955368042, 0.011131903648376466, 0.011344896316528321, 0.011210751533508301, 0.0112424955368042, 0.011263999938964844, 0.011302911758422851, 0.011213824272155762, 0.011325440406799316, 0.011231231689453124, 0.011256832122802735, 0.01124556827545166, 0.011246591567993165, 0.011188223838806152, 0.011262975692749023, 0.011303936004638672, 0.011294719696044921, 0.01122208023071289, 0.011206591606140136, 0.0112424955368042, 0.011308032035827637, 0.011212800025939941, 0.011331583976745606, 0.011241472244262696, 0.011345919609069824, 0.011208703994750976, 0.011266048431396485, 0.011198464393615723, 0.011206656455993653, 0.011239423751831054, 0.011283455848693847, 0.01124454402923584, 0.01136844825744629, 0.011285504341125489, 0.011257856369018555, 0.01122815990447998, 0.01116982364654541, 0.011263968467712403, 0.011196415901184082, 0.011222016334533692, 0.011268095970153809, 0.011254783630371093, 0.011423744201660157, 0.011216896057128906, 0.011208703994750976, 0.011232255935668945, 0.02350592041015625, 0.0112609281539917, 0.012005375862121583, 0.011383808135986329, 0.011304960250854493, 0.011283455848693847, 0.011226112365722657, 0.011202560424804688, 0.011263999938964844, 0.011246591567993165, 0.011610112190246581, 0.011381759643554687, 0.011267071723937988, 0.01233510398864746, 0.011413503646850585, 0.011245632171630859, 0.011271103858947754, 0.011197440147399903, 0.011265024185180664, 0.011243519783020019, 0.011241472244262696, 0.011213855743408203, 0.011274208068847656, 0.011226112365722657, 0.011177984237670899, 0.01124556827545166, 0.01125376033782959, 0.011276288032531738, 0.011593728065490723, 0.012101632118225097, 0.011258879661560058, 0.011426848411560058, 0.011340767860412597, 0.011250687599182128, 0.011226112365722657, 0.01131827163696289, 0.011140095710754394, 0.011276288032531738, 0.011244576454162597, 0.011317215919494628, 0.011682815551757812, 0.01140121555328369, 0.01131827163696289, 0.011275263786315918, 0.011298815727233886, 0.011229184150695801, 0.01125273609161377, 0.011293696403503419, 0.011226112365722657, 0.011249664306640626, 0.011280384063720703, 0.011209728240966797, 0.011223039627075194, 0.011274239540100098, 0.011243552207946777, 0.01130288028717041, 0.011293696403503419, 0.011228192329406738, 0.011246560096740722, 0.011202560424804688, 0.01121177577972412, 0.01123737621307373, 0.011276288032531738, 0.02346188735961914, 0.011205632209777832, 0.011248640060424805, 0.011196415901184082, 0.011185152053833008, 0.011194368362426758, 0.011103232383728028, 0.011118592262268067, 0.010899456024169921, 0.010898431777954102, 0.010889216423034668, 0.011025407791137695, 0.0111595516204834, 0.011188223838806152, 0.011126784324645997, 0.011218943595886231, 0.011179007530212403, 0.011150336265563965, 0.011134976387023926, 0.011125760078430176, 0.011155455589294434, 0.011191295623779298, 0.011165696144104004, 0.01113804817199707, 0.011150336265563965, 0.01121996784210205, 0.011215871810913085, 0.011169792175292969, 0.01112985610961914, 0.011187199592590333, 0.011195391654968261, 0.011234304428100587, 0.011191295623779298, 0.011135999679565429, 0.011169792175292969, 0.011190272331237794, 0.011175935745239257, 0.011165696144104004, 0.010827775955200195, 0.010895359992980956, 0.010907648086547851, 0.011126784324645997, 0.011133952140808106, 0.011169792175292969, 0.011165696144104004, 0.011185152053833008, 0.011100159645080567, 0.011106304168701172, 0.011142144203186035, 0.011210751533508301, 0.011107328414916993, 0.011194368362426758, 0.011146240234375, 0.011167743682861327, 0.011464703559875488, 0.011312128067016602, 0.011296768188476563, 0.011158528327941895, 0.01112166404724121, 0.011174912452697755, 0.01123532772064209, 0.011336704254150391, 0.011124735832214355, 0.022219776153564453, 0.011018239974975585, 0.011139072418212891, 0.01112985610961914, 0.011140095710754394, 0.011172863960266113, 0.0111278076171875, 0.011171872138977051, 0.011145183563232422, 0.011186176300048829, 0.011166720390319825, 0.011091967582702637, 0.011150336265563965, 0.011148287773132324, 0.011134976387023926, 0.011150336265563965, 0.011160575866699218, 0.011149312019348144, 0.0111278076171875, 0.01112063980102539, 0.01131007957458496, 0.01112985610961914, 0.011163647651672364, 0.01115340805053711, 0.01120358371734619, 0.011196415901184082, 0.011167743682861327, 0.011165696144104004, 0.011246591567993165, 0.011224063873291015, 0.011179007530212403, 0.011167743682861327, 0.011182080268859864, 0.011175935745239257, 0.011095040321350098, 0.011188223838806152, 0.01114521598815918, 0.011164671897888183, 0.011201536178588867, 0.01122815990447998, 0.01115340805053711, 0.011130880355834961, 0.011168767929077148, 0.011240447998046875, 0.01115340805053711, 0.01117081642150879, 0.011123711585998536, 0.011975680351257324, 0.014036992073059081, 0.011767807960510255, 0.011272192001342773, 0.011220992088317871, 0.011161600112915039, 0.011171839714050292, 0.011180031776428222, 0.011186176300048829, 0.011100159645080567, 0.01124454402923584, 0.011142144203186035, 0.011199487686157226, 0.011167776107788087, 0.011124704360961914, 0.011181056022644043, 0.023325727462768554, 0.01110524845123291, 0.011058176040649414, 0.010943488121032715, 0.010982399940490722, 0.010989567756652833, 0.011012096405029297, 0.011171839714050292, 0.011198464393615723, 0.011144224166870118, 0.011127776145935058, 0.011077695846557618, 0.011132863998413086, 0.011165696144104004, 0.011160575866699218, 0.011229184150695801, 0.01121177577972412, 0.011184127807617187, 0.011172863960266113, 0.01111244773864746, 0.011144191741943359, 0.011133952140808106, 0.011180031776428222, 0.011130880355834961, 0.011205632209777832, 0.011103232383728028, 0.011167743682861327, 0.01122713565826416, 0.011155455589294434, 0.01112063980102539, 0.011210751533508301, 0.011217920303344727, 0.011107328414916993, 0.011163647651672364, 0.01111244773864746, 0.011090944290161133, 0.011166720390319825, 0.011087871551513672, 0.011205632209777832, 0.01115340805053711, 0.011155455589294434, 0.01112063980102539, 0.011190272331237794, 0.01112985610961914, 0.011212800025939941, 0.01121177577972412, 0.011147263526916504, 0.011180031776428222, 0.011126784324645997, 0.011220992088317871, 0.011133983612060546, 0.01121174430847168, 0.011220992088317871, 0.011192319869995117, 0.011101183891296386, 0.011101183891296386, 0.011166751861572266, 0.011146240234375, 0.011119584083557128, 0.011091967582702637, 0.011132927894592285, 0.011150336265563965, 0.011187199592590333, 0.023353343963623048, 0.011106304168701172, 0.011140095710754394, 0.01115340805053711, 0.01093120002746582, 0.01091481590270996, 0.01095372772216797, 0.011004927635192872, 0.01112985610961914, 0.011101183891296386, 0.011104255676269532, 0.010960895538330078, 0.01094758415222168, 0.010938367843627929, 0.011023360252380371, 0.011216896057128906, 0.01113702392578125, 0.011067392349243164, 0.010962944030761718, 0.010912768363952637, 0.010910719871520995, 0.010944512367248535, 0.010978303909301757, 0.011214847564697266, 0.01115135955810547, 0.011186176300048829, 0.011133952140808106, 0.011190272331237794, 0.011031552314758301, 0.010978303909301757, 0.010934271812438966, 0.0111278076171875, 0.011119615554809571, 0.011147263526916504, 0.011101183891296386, 0.011167743682861327, 0.011180031776428222, 0.011135999679565429, 0.011117568016052246, 0.011113471984863281, 0.011185152053833008, 0.01125376033782959, 0.011240447998046875, 0.011155455589294434, 0.011187199592590333, 0.011194368362426758, 0.011183103561401368, 0.011171839714050292, 0.011149312019348144, 0.011119615554809571, 0.011183103561401368, 0.011135999679565429, 0.011386879920959473, 0.011439104080200196, 0.01154355239868164, 0.011313152313232423, 0.0112609281539917, 0.011216896057128906, 0.011262975692749023, 0.011165696144104004, 0.011180031776428222, 0.011125823974609376, 0.011188159942626954, 0.02225049591064453, 0.01052569580078125, 0.010568703651428223, 0.010543104171752929, 0.010622976303100586, 0.010621952056884766, 0.010568703651428223, 0.010541055679321289, 0.010622976303100586, 0.010554368019104005, 0.010582015991210938, 0.010565631866455078, 0.010530816078186036, 0.010882047653198243, 0.011097087860107421, 0.01112985610961914, 0.01112985610961914, 0.011177984237670899, 0.011173888206481934, 0.011111424446105958, 0.011111488342285156, 0.011157440185546876, 0.011141119956970215, 0.01115135955810547, 0.011143168449401856, 0.01112883186340332, 0.011140095710754394, 0.01136844825744629, 0.01112166404724121, 0.011107328414916993, 0.011119615554809571, 0.011123711585998536, 0.011150336265563965, 0.01120358371734619, 0.01112063980102539, 0.011146271705627441, 0.011225055694580079, 0.011163647651672364, 0.011206656455993653, 0.011156479835510253, 0.011208703994750976, 0.011188223838806152, 0.011194368362426758, 0.011114496231079102, 0.01114521598815918, 0.011217920303344727, 0.011176959991455078, 0.01121177577972412, 0.011148287773132324, 0.010975232124328613, 0.010976256370544434, 0.010999808311462403, 0.011003904342651367, 0.011169792175292969, 0.011200511932373047, 0.011224063873291015, 0.011240447998046875, 0.011239423751831054, 0.010964991569519043, 0.010970111846923827, 0.011025407791137695, 0.011190272331237794, 0.011210783958435059, 0.022187999725341797, 0.010556415557861328, 0.010572799682617188, 0.010587136268615722, 0.010652671813964844, 0.01060969638824463, 0.010614751815795898, 0.010569727897644043, 0.010598400115966796, 0.010579999923706054, 0.010570719718933105, 0.010619903564453125, 0.010581024169921875, 0.010570752143859862, 0.010605536460876464, 0.010647551536560058, 0.010609663963317872, 0.010669055938720704, 0.010605567932128907, 0.010513407707214355, 0.010582015991210938, 0.010595328330993652, 0.010590208053588868, 0.010577919960021973, 0.010600447654724121, 0.01055846405029297, 0.011377663612365722, 0.011400192260742188, 0.011234304428100587, 0.011131903648376466, 0.011103232383728028, 0.011122688293457032, 0.011140095710754394, 0.011132927894592285, 0.01112985610961914, 0.01115135955810547, 0.011166720390319825, 0.011167743682861327, 0.011173888206481934, 0.011142144203186035, 0.011146240234375, 0.011175935745239257, 0.01126195240020752, 0.011149312019348144, 0.011132927894592285, 0.011160575866699218, 0.01123840045928955, 0.011169792175292969, 0.011114496231079102, 0.011144191741943359, 0.011139072418212891, 0.010988544464111329, 0.010944512367248535, 0.010891263961791992, 0.010995712280273438, 0.011147263526916504, 0.011293696403503419, 0.011174912452697755, 0.011168767929077148, 0.011181056022644043, 0.01122815990447998, 0.011097087860107421, 0.011144191741943359, 0.022222848892211915, 0.010568703651428223, 0.010520640373229981, 0.010560447692871094, 0.010555392265319824, 0.010669055938720704, 0.010604543685913086, 0.010606592178344726, 0.010740736007690429, 0.010578944206237792, 0.010570752143859862, 0.010628095626831055, 0.010579968452453613, 0.010572799682617188, 0.010601471900939942, 0.01063424015045166, 0.010566656112670898, 0.010643551826477051, 0.010620832443237305, 0.010530816078186036, 0.010621952056884766, 0.010595328330993652, 0.010600447654724121, 0.010604543685913086, 0.010577919960021973, 0.0105830717086792, 0.010615776062011719, 0.010611712455749512, 0.010565631866455078, 0.010570752143859862, 0.010562560081481934, 0.01053593635559082, 0.010546175956726075, 0.01061888027191162, 0.010591232299804687, 0.010600447654724121, 0.010589183807373047, 0.010599424362182617, 0.010596351623535156, 0.010602527618408204, 0.010623968124389649, 0.010583040237426757, 0.010593279838562012, 0.010605567932128907, 0.010557439804077149, 0.01064140796661377, 0.010597375869750977, 0.010564607620239258, 0.010590208053588868, 0.010691583633422852, 0.010604543685913086, 0.010619903564453125, 0.01062604808807373, 0.010612735748291016, 0.010584063529968261, 0.010546175956726075, 0.010605567932128907, 0.010622976303100586, 0.010594304084777833, 0.010560511589050293, 0.010608639717102051, 0.010576895713806152, 0.010612768173217773, 0.023476192474365234, 0.011000831604003907, 0.010892288208007812, 0.011049983978271484, 0.011250687599182128, 0.011179007530212403, 0.01115443229675293, 0.011082847595214844, 0.01117686367034912, 0.011156479835510253, 0.011122688293457032, 0.011104255676269532, 0.011077631950378418, 0.011192319869995117, 0.011141119956970215, 0.011192319869995117, 0.01115340805053711, 0.011197471618652343, 0.011192288398742676, 0.0111278076171875, 0.011124735832214355, 0.011157504081726074, 0.011213824272155762, 0.01112883186340332, 0.0111595516204834, 0.011113471984863281, 0.011140128135681153, 0.011154399871826173, 0.01114521598815918, 0.011107328414916993, 0.011158528327941895, 0.01115340805053711, 0.011140095710754394, 0.011147263526916504, 0.011156479835510253, 0.011115519523620606, 0.011163647651672364, 0.011156479835510253, 0.01115135955810547, 0.01116262435913086, 0.011130880355834961, 0.011122688293457032, 0.011018239974975585, 0.010986495971679687, 0.011015232086181641, 0.011019200325012206, 0.0112609281539917, 0.011164671897888183, 0.011179007530212403, 0.011146240234375, 0.01121177577972412, 0.01115135955810547, 0.011171839714050292, 0.011141119956970215, 0.011281408309936524, 0.01120358371734619, 0.011191295623779298, 0.011168767929077148, 0.011190272331237794, 0.011324480056762695, 0.011292608261108398, 0.011140095710754394, 0.011161600112915039, 0.023377920150756838, 0.011315199851989746, 0.011220992088317871, 0.011183103561401368, 0.011199487686157226, 0.011177984237670899, 0.011420672416687011, 0.011294783592224121, 0.011265983581542969, 0.011155455589294434, 0.011230208396911622, 0.011168767929077148, 0.011123711585998536, 0.01124556827545166, 0.011197440147399903, 0.011199487686157226, 0.01115443229675293, 0.011144191741943359, 0.01113804817199707, 0.011110400199890137, 0.011140095710754394, 0.011111424446105958, 0.011141119956970215, 0.011123711585998536, 0.0111595516204834, 0.011122688293457032, 0.011305983543395997, 0.01131929588317871, 0.011192319869995117, 0.011145248413085937, 0.011159520149230957, 0.011119615554809571, 0.011196415901184082, 0.01122815990447998, 0.011187199592590333, 0.011146240234375, 0.0111595516204834, 0.011174976348876953, 0.011206591606140136, 0.011270144462585448, 0.01117081642150879, 0.011179007530212403, 0.011171839714050292, 0.01115443229675293, 0.011140095710754394, 0.011191295623779298, 0.011165696144104004, 0.011374591827392578, 0.01121996784210205, 0.011220992088317871, 0.011185152053833008, 0.011156479835510253, 0.011411456108093262, 0.011210751533508301, 0.01112166404724121, 0.011126784324645997, 0.011168767929077148, 0.011199487686157226, 0.011240447998046875, 0.011131903648376466, 0.011180031776428222, 0.01113702392578125, 0.011262975692749023, 0.023784448623657226, 0.011406335830688476, 0.011194368362426758, 0.010918911933898925, 0.010921983718872071, 0.011123711585998536, 0.01111244773864746, 0.011032575607299805, 0.011065343856811523, 0.011059200286865235, 0.011974656105041503, 0.01143295955657959, 0.011126784324645997, 0.011075648307800293, 0.01116870403289795, 0.011183103561401368, 0.011117568016052246, 0.011180031776428222, 0.011175935745239257, 0.011152383804321288, 0.011309056282043458, 0.011322367668151855, 0.011193344116210938, 0.011293696403503419, 0.011202560424804688, 0.011185152053833008, 0.011208703994750976, 0.011263999938964844, 0.01115135955810547, 0.011171839714050292, 0.011182080268859864, 0.011140095710754394, 0.011177984237670899, 0.011115519523620606, 0.011192352294921874, 0.011250656127929687, 0.011204607963562012, 0.011172863960266113, 0.011143168449401856, 0.011181056022644043, 0.011270144462585448, 0.011250687599182128, 0.01125376033782959, 0.011148287773132324, 0.011201536178588867, 0.011172863960266113, 0.01122815990447998, 0.011149312019348144, 0.011142144203186035, 0.011180031776428222, 0.011180031776428222, 0.011220992088317871, 0.011209728240966797, 0.011199487686157226, 0.011199487686157226, 0.011239423751831054, 0.011174912452697755, 0.01120358371734619, 0.011176992416381835, 0.011191264152526855, 0.011183103561401368, 0.01112063980102539, 0.011161600112915039, 0.02329599952697754, 0.011206656455993653, 0.011105279922485351, 0.011160575866699218, 0.011181056022644043, 0.011124735832214355, 0.010936320304870606, 0.01092300796508789, 0.011149312019348144, 0.01091481590270996, 0.010949631690979005, 0.01113702392578125, 0.011117568016052246, 0.011149312019348144, 0.01125376033782959, 0.011171839714050292, 0.011158528327941895, 0.011171839714050292, 0.01114521598815918, 0.011100159645080567, 0.011187264442443848, 0.011203519821166992, 0.011156479835510253, 0.011087871551513672, 0.011157504081726074, 0.011142144203186035, 0.011171839714050292, 0.011243519783020019, 0.011124735832214355, 0.011180031776428222, 0.011176959991455078, 0.011116543769836425, 0.011148287773132324, 0.011171839714050292, 0.01115443229675293, 0.01115340805053711, 0.011131903648376466, 0.011108415603637695, 0.011206591606140136, 0.011194368362426758, 0.011298815727233886, 0.01116262435913086, 0.011177984237670899, 0.011129920005798339, 0.01116051197052002, 0.011156479835510253, 0.011405311584472656, 0.011185152053833008, 0.011182080268859864, 0.011201536178588867, 0.011184127807617187, 0.011123711585998536, 0.011195391654968261, 0.01122713565826416, 0.011183199882507324, 0.011243424415588378, 0.011175935745239257, 0.01132748794555664, 0.011174912452697755, 0.011171839714050292, 0.011160575866699218, 0.011133952140808106, 0.011172863960266113, 0.022174720764160157, 0.010556415557861328, 0.01095372772216797, 0.011195391654968261, 0.011160575866699218, 0.011092032432556152, 0.01113491153717041, 0.01117084789276123, 0.011182047843933106, 0.011116543769836425, 0.011103232383728028, 0.011118592262268067, 0.011139072418212891, 0.011116543769836425, 0.011149312019348144, 0.011268095970153809, 0.011103232383728028, 0.011250687599182128, 0.011189248085021973, 0.011140159606933594, 0.011080639839172363, 0.011131903648376466, 0.011133952140808106, 0.011148287773132324, 0.01115443229675293, 0.011152383804321288, 0.011223039627075194, 0.011332608222961426, 0.011132927894592285, 0.010942463874816894, 0.010919936180114746, 0.010942463874816894, 0.010935296058654785, 0.01093734359741211, 0.011166720390319825, 0.010996800422668457, 0.010977215766906738, 0.010974207878112792, 0.010911744117736816, 0.010969087600708008, 0.01095577621459961, 0.011114496231079102, 0.011009023666381837, 0.01102233600616455, 0.010984448432922364, 0.0109619197845459, 0.0109486083984375, 0.010946559906005859, 0.010987520217895508, 0.011146240234375, 0.01111244773864746, 0.011174912452697755, 0.011181056022644043, 0.011263999938964844, 0.011266048431396485, 0.011032575607299805, 0.010979328155517578, 0.011013119697570802, 0.010957823753356934, 0.010988544464111329, 0.011087871551513672, 0.011174943923950196, 0.011246560096740722]",tokens/s,88.49256396436706,,,main,False,False bfloat16-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,,cuda,0,42,,,,,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,d35829e539df8480b726c647eeabf91e41eae047,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1192.706048,14382.792704,0.0,13736.3456,13487.531008,s,10,0.6736084442138672,0.06736084442138672,0.0002525542761618945,0.06724551773071288,0.06760586013793946,0.06782964820861817,0.06800867866516114,"[0.06805343627929687, 0.06725020599365235, 0.06722144317626953, 0.06722035217285156, 0.06722418975830079, 0.0672086410522461, 0.06755612945556641, 0.06724082946777343, 0.06725161743164063, 0.06738159942626953]",tokens/s,3800.4274174259212,kWh,7.965395170125742e-07,4.3641060099177397e-07,4.3883946367381145e-06,5.621344754742463e-06,tokens/kWh,45540704.43447271,MB,1192.706048,14382.792704,0.0,13736.3456,13661.254656,s,10,19.870003173828128,1.9870003173828124,0.001308161958882297,1.9866483154296875,1.9882964599609374,1.9892427124023437,1.9899997143554689,"[1.99018896484375, 1.988086181640625, 1.98498486328125, 1.98658154296875, 1.9861776123046875, 1.986427734375, 1.9865947265625, 1.98747412109375, 1.986701904296875, 1.9867855224609374]",tokens/s,31.706084517883102,kWh,2.3405527987313928e-05,1.282685471910084e-05,0.00010697394445306076,0.00014320632715947553,tokens/kWh,439924.6964126297,,s,629,20.443699218749988,0.03250190654809221,0.008036461391132918,0.03153817558288574,0.031657990264892584,0.03169198112487793,0.09906618194580079,"[0.03139174461364746, 0.03147878456115723, 0.03156172752380371, 0.03142348861694336, 0.031454208374023435, 0.031442943572998046, 0.03150131225585937, 0.031455232620239255, 0.03145011138916016, 0.03141228866577148, 0.03145619201660156, 0.03143168067932129, 0.03148598480224609, 0.031468511581420895, 0.0316549129486084, 0.03184639930725098, 0.03200921630859375, 0.0318474235534668, 0.031547391891479495, 0.031475711822509765, 0.031494144439697266, 0.03146956825256347, 0.03148492813110351, 0.03150233650207519, 0.03157401657104492, 0.03155046463012695, 0.03151263999938965, 0.03151865577697754, 0.03153817558288574, 0.03150233650207519, 0.03153510475158691, 0.03154742431640625, 0.03157705688476563, 0.031559680938720705, 0.03156070327758789, 0.03156480026245117, 0.031591424942016604, 0.03154332733154297, 0.03157910346984863, 0.031579135894775394, 0.03159552001953125, 0.03159040069580078, 0.03161087989807129, 0.0316753921508789, 0.031692800521850584, 0.03162732887268067, 0.031632320404052734, 0.03161497688293457, 0.031628288269042966, 0.03162112045288086, 0.03166924858093262, 0.03167033576965332, 0.03166201591491699, 0.03164262390136719, 0.031747072219848634, 0.031752191543579104, 0.03167334365844727, 0.03168767929077149, 0.03163852882385254, 0.03165798377990723, 0.03169075202941894, 0.03166521644592285, 0.09908729553222656, 0.03136614418029785, 0.03138252830505371, 0.03139174461364746, 0.03137126350402832, 0.03139788818359375, 0.03159756851196289, 0.03200307083129883, 0.031575040817260744, 0.031497215270996096, 0.03148492813110351, 0.031461408615112305, 0.03139785575866699, 0.031455232620239255, 0.03140201568603516, 0.031416288375854494, 0.031455232620239255, 0.031460351943969726, 0.03146137619018555, 0.031437824249267575, 0.03146342468261719, 0.031453184127807614, 0.031441919326782225, 0.03158732795715332, 0.03161907196044922, 0.03153919982910156, 0.03149516868591309, 0.031509504318237305, 0.03149516868591309, 0.031475711822509765, 0.03150028800964356, 0.03147776031494141, 0.03151667213439941, 0.031541248321533204, 0.03151974487304687, 0.03149926376342774, 0.031531007766723636, 0.031543296813964845, 0.03154841613769531, 0.03161907196044922, 0.03155251121520996, 0.03156684875488281, 0.031601696014404296, 0.03161596870422363, 0.03155251121520996, 0.03158016014099121, 0.03159654426574707, 0.03159040069580078, 0.03159552001953125, 0.03159555244445801, 0.031607776641845706, 0.031604736328125, 0.031665151596069335, 0.03163033676147461, 0.03159859275817871, 0.03167027282714844, 0.03164262390136719, 0.031623231887817384, 0.03163852882385254, 0.031638463973999026, 0.03161702346801758, 0.03202969741821289, 0.03202252960205078, 0.09912422180175781, 0.03140096092224121, 0.031352832794189454, 0.03136614418029785, 0.03136511993408203, 0.0313753604888916, 0.0313856315612793, 0.03141423988342285, 0.031388671875, 0.03137945556640625, 0.03140300750732422, 0.03141529655456543, 0.03138252830505371, 0.03142348861694336, 0.031421440124511715, 0.03143270492553711, 0.031415327072143555, 0.03142243194580078, 0.03143577575683594, 0.031425535202026365, 0.03141529655456543, 0.03144601631164551, 0.03142758369445801, 0.03144806480407715, 0.031441919326782225, 0.031455232620239255, 0.031476736068725586, 0.031492095947265625, 0.03147980880737305, 0.031508480072021484, 0.03156377601623535, 0.031664127349853514, 0.03156684875488281, 0.031508480072021484, 0.031514623641967776, 0.031575040817260744, 0.031529983520507815, 0.03155353546142578, 0.031541248321533204, 0.0315228157043457, 0.031525888442993165, 0.03153203201293946, 0.03153408050537109, 0.031529983520507815, 0.031591424942016604, 0.03157401657104492, 0.03159654426574707, 0.03160166358947754, 0.0315863037109375, 0.03157811164855957, 0.0315729923248291, 0.03157606315612793, 0.0315729923248291, 0.03157708740234375, 0.03160681533813477, 0.03164975929260254, 0.03162521553039551, 0.03162009620666504, 0.03159859275817871, 0.03159449577331543, 0.03158425521850586, 0.03159654426574707, 0.03161087989807129, 0.09906175994873047, 0.0313702392578125, 0.03139993667602539, 0.031325183868408206, 0.03136000061035156, 0.03138662338256836, 0.031409151077270506, 0.03139891242980957, 0.03141119956970215, 0.03138457679748535, 0.03141836738586426, 0.03139174461364746, 0.03139481544494629, 0.031442943572998046, 0.031409151077270506, 0.03146342468261719, 0.03142963218688965, 0.031442943572998046, 0.031457279205322264, 0.03146444892883301, 0.03148595237731933, 0.031497215270996096, 0.03148185539245606, 0.03143168067932129, 0.0314521598815918, 0.031494144439697266, 0.031453184127807614, 0.031494144439697266, 0.031498239517211916, 0.031509504318237305, 0.03150233650207519, 0.031512575149536134, 0.03153919982910156, 0.03155251121520996, 0.03150441551208496, 0.03152892875671387, 0.03157708740234375, 0.03157811164855957, 0.03159756851196289, 0.03159449577331543, 0.03163443183898926, 0.03154944038391113, 0.03155046463012695, 0.03158937644958496, 0.03158527946472168, 0.03158937644958496, 0.03158732795715332, 0.03159244728088379, 0.03159859275817871, 0.03162521553039551, 0.03164262390136719, 0.03161395263671875, 0.031666175842285156, 0.031734783172607424, 0.031627264022827145, 0.03164262390136719, 0.03179827117919922, 0.03181875228881836, 0.031697919845581055, 0.03163852882385254, 0.03163750457763672, 0.03167231941223145, 0.031627264022827145, 0.09906790161132813, 0.031389696121215824, 0.03135078430175781, 0.03138662338256836, 0.03139379119873047, 0.03141324806213379, 0.031547391891479495, 0.03158937644958496, 0.03138150405883789, 0.031410175323486327, 0.03142963218688965, 0.03141119956970215, 0.03142560005187988, 0.031449024200439456, 0.03146956825256347, 0.0314450569152832, 0.03142854309082031, 0.03147776031494141, 0.03141119956970215, 0.031480831146240236, 0.03143475151062012, 0.03144806480407715, 0.031474687576293944, 0.031456256866455076, 0.03144499206542969, 0.03148902320861816, 0.0314726390838623, 0.031488000869750975, 0.03144908714294434, 0.031503360748291014, 0.031508480072021484, 0.03150028800964356, 0.031525888442993165, 0.03153715133666992, 0.03158016014099121, 0.03167334365844727, 0.03153919982910156, 0.031575040817260744, 0.031575040817260744, 0.03156076812744141, 0.03154323196411133, 0.03155353546142578, 0.03156377601623535, 0.03156582450866699, 0.03157196807861328, 0.03156588745117188, 0.031603647232055665, 0.03156582450866699, 0.03158732795715332, 0.03161907196044922, 0.03158323287963867, 0.03164262390136719, 0.03158432006835937, 0.031652799606323244, 0.03164159965515137, 0.0316231689453125, 0.03161087989807129, 0.031665151596069335, 0.031615999221801756, 0.03159756851196289, 0.03160576057434082, 0.03161292839050293, 0.031666175842285156, 0.09904742431640624, 0.03136409568786621, 0.03138252830505371, 0.031339519500732424, 0.03135385513305664, 0.03144806480407715, 0.03139583969116211, 0.03142457580566406, 0.03139065551757812, 0.03141324806213379, 0.03146240043640137, 0.031441919326782225, 0.03142860794067383, 0.0314204158782959, 0.03144806480407715, 0.031437824249267575, 0.03143065643310547, 0.03145011138916016, 0.03143475151062012, 0.03148287963867188, 0.031458303451538085, 0.0315043830871582, 0.031456256866455076, 0.03144499206542969, 0.03142963218688965, 0.031509504318237305, 0.03147372817993164, 0.03150943946838379, 0.031508480072021484, 0.031514623641967776, 0.0314839038848877, 0.03146854400634766, 0.03151872062683105, 0.031535135269165036, 0.031506399154663085, 0.03158016014099121, 0.03155046463012695, 0.03163443183898926, 0.03157094383239746, 0.03156172752380371, 0.03168358421325684, 0.03168870353698731, 0.03152383995056152, 0.03157094383239746, 0.03155558395385742, 0.0315863037109375, 0.03160166358947754, 0.03159756851196289, 0.031591424942016604, 0.03159756851196289, 0.03162214469909668, 0.0316231689453125, 0.03158323287963867, 0.03162931251525879, 0.0315996150970459, 0.03165286445617676, 0.03163340759277344, 0.03167027282714844, 0.031663103103637694, 0.031632383346557616, 0.03163552093505859, 0.03165996742248535, 0.03170816040039062, 0.09976012420654297, 0.031454208374023435, 0.03139379119873047, 0.03139276885986328, 0.03138355255126953, 0.03136511993408203, 0.03139891242980957, 0.03139174461364746, 0.031399999618530273, 0.03145209693908691, 0.031406080245971676, 0.03146240043640137, 0.03140096092224121, 0.03144499206542969, 0.03143577575683594, 0.03143065643310547, 0.031470592498779294, 0.03145011138916016, 0.03150131225585937, 0.031442943572998046, 0.03143475151062012, 0.0314521598815918, 0.03148595237731933, 0.031471616744995115, 0.031455232620239255, 0.03149516868591309, 0.031457279205322264, 0.031475711822509765, 0.031470592498779294, 0.03151769638061523, 0.031503360748291014, 0.031527936935424807, 0.031513599395751955, 0.03155459213256836, 0.031517663955688474, 0.03159449577331543, 0.03156172752380371, 0.031547391891479495, 0.031562751770019534, 0.03156070327758789, 0.03158220863342285, 0.031558656692504884, 0.0315361270904541, 0.0315228157043457, 0.03155046463012695, 0.03160780715942383, 0.03159040069580078, 0.03178291130065918, 0.03182080078125, 0.03165801620483399, 0.031572959899902345, 0.03167436790466309, 0.031611904144287106, 0.03163852882385254, 0.031647743225097655, 0.0315996150970459, 0.03160063934326172, 0.03161907196044922, 0.03160780715942383, 0.031643648147583005, 0.03161087989807129, 0.031649791717529296, 0.031632383346557616, 0.09910886383056641, 0.03138150405883789, 0.0314071044921875, 0.03135897636413574, 0.03136716842651367, 0.03142963218688965, 0.03143475151062012, 0.031421440124511715, 0.03140403175354004, 0.03141222381591797, 0.03143577575683594, 0.031389696121215824, 0.03139686393737793, 0.0314521598815918, 0.031426559448242186, 0.031510528564453126, 0.03149004745483398, 0.031457279205322264, 0.03144601631164551, 0.03152076721191406, 0.031457279205322264, 0.03157708740234375, 0.03172761535644531, 0.03187711906433106, 0.031529983520507815, 0.03153510475158691, 0.03146342468261719, 0.03148287963867188, 0.03147980880737305, 0.031541248321533204, 0.03152076721191406, 0.03151667213439941, 0.03154944038391113, 0.03156684875488281, 0.03152383995056152, 0.03153715133666992, 0.03155251121520996, 0.03159756851196289, 0.0315729923248291, 0.03158527946472168, 0.03160783958435059, 0.031556575775146485, 0.031541248321533204, 0.03158835220336914, 0.031558656692504884, 0.03158323287963867, 0.03159654426574707, 0.031661056518554685, 0.03171327972412109, 0.03164262390136719, 0.03163340759277344, 0.03160985565185547, 0.03162419128417969, 0.031628288269042966, 0.031604736328125, 0.03163750457763672, 0.03163955116271973, 0.03164159965515137, 0.03167647933959961, 0.031638463973999026, 0.031616031646728514, 0.03169487953186035, 0.031621055603027345, 0.09908428955078125, 0.03140918350219726, 0.031331296920776366, 0.03136819267272949, 0.03139481544494629, 0.03138764762878418, 0.03138457679748535, 0.03139174461364746, 0.0314019832611084, 0.031445024490356445, 0.0314050235748291, 0.03144908714294434, 0.03142451286315918, 0.03150028800964356, 0.03148595237731933, 0.03143270492553711, 0.03142860794067383, 0.03146751976013184, 0.03142451286315918, 0.03145011138916016, 0.0314337272644043, 0.031441919326782225, 0.03146137619018555, 0.031475711822509765, 0.03148595237731933, 0.03146444892883301, 0.031474687576293944, 0.0314726390838623, 0.031480831146240236, 0.03151872062683105, 0.0315729923248291, 0.03156787109375, 0.03154431915283203, 0.03152179145812988, 0.03156172752380371, 0.03151155281066895, 0.03150643157958984, 0.031562751770019534, 0.03154841613769531, 0.03154841613769531, 0.0315228157043457, 0.03158323287963867, 0.031559680938720705, 0.03156070327758789, 0.03158016014099121, 0.031645696640014646, 0.03160166358947754, 0.03160166358947754, 0.0317890567779541, 0.03191500854492187, 0.03163033676147461, 0.03160678482055664, 0.03164672088623047, 0.031628288269042966, 0.03163955116271973, 0.031662080764770506, 0.03161497688293457, 0.03163852882385254, 0.03161087989807129, 0.03161907196044922, 0.031628288269042966, 0.031664127349853514, 0.031714303970336914, 0.09944064331054688, 0.03139481544494629, 0.03144396781921387, 0.031352832794189454, 0.0313702392578125, 0.03138047981262207, 0.03140812873840332, 0.03138665580749512, 0.03141321563720703, 0.03141324806213379, 0.031466527938842775, 0.03147772789001465, 0.03141939163208008, 0.03143168067932129, 0.03142758369445801, 0.0314768009185791, 0.03142035293579101, 0.03143680000305176, 0.03144806480407715, 0.031477792739868164, 0.03144803237915039, 0.03148902320861816, 0.03144499206542969, 0.031525888442993165, 0.03147878456115723, 0.03148287963867188, 0.0314654712677002, 0.031528959274291994, 0.031498239517211916, 0.0315545597076416, 0.03154431915283203, 0.031546367645263675, 0.03153715133666992, 0.03155660820007324, 0.03152179145812988, 0.03159654426574707, 0.031591424942016604, 0.031575040817260744, 0.03157196807861328, 0.03157606315612793, 0.031581216812133786, 0.031552480697631835, 0.0315996150970459, 0.03160166358947754, 0.03155353546142578, 0.03160883140563965, 0.03160268783569336, 0.03161395263671875, 0.03159859275817871, 0.03161087989807129, 0.0316753921508789, 0.031649791717529296, 0.03160268783569336, 0.031627264022827145, 0.031665151596069335, 0.031649791717529296, 0.03161087989807129, 0.031648767471313476, 0.03165798377990723, 0.031665151596069335, 0.03160678482055664, 0.03167231941223145, 0.03162419128417969]",tokens/s,30.767425859166956,,,main,False,False -bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,flash_attention_2,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14433,7 +14433,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1146.896384,7590.117376,0.0,6949.961728,6804.719616,s,10,0.5250520591735839,0.0525052059173584,0.00029772145613171885,0.05241555213928223,0.0527534252166748,0.05302911167144775,0.05324966083526611,"[0.0533047981262207, 0.05244457626342774, 0.05252767944335938, 0.05232595062255859, 0.05269216156005859, 0.05252345657348633, 0.05238652801513672, 0.05235254287719727, 0.05219772720336914, 0.05229663848876953]",tokens/s,4875.706999472323,kWh,6.486730596913521e-07,3.5538707580584384e-07,4.55767820594539e-06,5.561738341442585e-06,tokens/kWh,46028774.50966159,MB,1146.896384,7590.117376,0.0,6949.961728,6904.4864,s,10,12.46740625,1.2467406249999997,0.012987626456919968,1.2483598022460938,1.2645401611328124,1.2647395629882812,1.2648990844726562,"[1.2483555908203126, 1.230513671875, 1.2362603759765625, 1.259568359375, 1.2252486572265624, 1.24107470703125, 1.26493896484375, 1.264495849609375, 1.248364013671875, 1.2485860595703124]",tokens/s,50.531761568289326,kWh,1.4604661830117939e-05,8.003160860817084e-06,5.1301283147852863e-05,7.390910583878787e-05,tokens/kWh,852398.3518000739,,s,629,12.919056360244769,0.020539040318354136,0.006232593580653284,0.019941375732421874,0.02016624641418457,0.020408320617675783,0.07202649169921875,"[0.02001203155517578, 0.019917823791503905, 0.020178943634033202, 0.019944448471069336, 0.020190208435058594, 0.020106239318847655, 0.01999667167663574, 0.019758079528808595, 0.019920896530151368, 0.020060159683227538, 0.020146175384521483, 0.019851264953613282, 0.019941375732421874, 0.019960832595825196, 0.02008678436279297, 0.01985740852355957, 0.019977216720581056, 0.02083839988708496, 0.02002841567993164, 0.019766271591186522, 0.019830848693847655, 0.019814336776733398, 0.01962700843811035, 0.02046771240234375, 0.020779008865356444, 0.02088140869140625, 0.020343807220458983, 0.020195327758789062, 0.02003660774230957, 0.01986457633972168, 0.020007936477661133, 0.02002227210998535, 0.019947519302368166, 0.019885055541992186, 0.019818496704101563, 0.019926015853881835, 0.019943424224853516, 0.01983590316772461, 0.019307519912719725, 0.01928294372558594, 0.01922969627380371, 0.01926553535461426, 0.01923891258239746, 0.019289087295532227, 0.01924198341369629, 0.019115007400512696, 0.01926348876953125, 0.01917647933959961, 0.019151840209960937, 0.01966592025756836, 0.019796991348266603, 0.020001792907714845, 0.020025344848632814, 0.019751935958862304, 0.019893247604370116, 0.01924812889099121, 0.019171327590942384, 0.01900851249694824, 0.01923686408996582, 0.01926553535461426, 0.019531776428222656, 0.019997695922851562, 0.07219097900390625, 0.01925324821472168, 0.019333120346069335, 0.019190784454345702, 0.01925632095336914, 0.01922355270385742, 0.019514368057250975, 0.019652639389038086, 0.01967100715637207, 0.019956735610961913, 0.02003455924987793, 0.01984716796875, 0.01964134407043457, 0.019927040100097656, 0.020113407135009767, 0.02002841567993164, 0.020016128540039063, 0.019959808349609375, 0.020007936477661133, 0.019955711364746095, 0.019994623184204103, 0.01983180809020996, 0.019955711364746095, 0.01966192054748535, 0.019778463363647462, 0.019951648712158203, 0.019997663497924804, 0.020172800064086914, 0.020002815246582033, 0.019902463912963866, 0.019948543548583983, 0.019762176513671875, 0.01901568031311035, 0.018940927505493164, 0.01925017547607422, 0.01926246452331543, 0.019194879531860352, 0.019817472457885742, 0.01988915252685547, 0.01988403129577637, 0.01990553665161133, 0.019770368576049805, 0.019280895233154297, 0.01922047996520996, 0.01920204734802246, 0.019140607833862306, 0.019219455718994142, 0.01922355270385742, 0.01923891258239746, 0.019162111282348633, 0.018958335876464845, 0.01900339126586914, 0.01901158332824707, 0.018970624923706055, 0.018885631561279297, 0.019139583587646485, 0.019144704818725586, 0.019174400329589843, 0.019397760391235353, 0.01949171257019043, 0.01920921516418457, 0.019118080139160155, 0.019160064697265625, 0.07202508544921875, 0.018894847869873048, 0.019054592132568358, 0.019107839584350587, 0.018957311630249024, 0.018782207489013672, 0.01904128074645996, 0.019203071594238282, 0.01928294372558594, 0.01923993682861328, 0.019151872634887695, 0.019175424575805664, 0.019324928283691405, 0.01924505615234375, 0.019117055892944337, 0.01942425537109375, 0.01925119972229004, 0.019250303268432616, 0.0190882568359375, 0.01927168083190918, 0.019301408767700194, 0.019954656600952147, 0.02008064079284668, 0.019869695663452147, 0.01969049644470215, 0.018979839324951172, 0.02008780860900879, 0.020140031814575195, 0.019156991958618166, 0.018994176864624023, 0.019647487640380858, 0.019810304641723633, 0.02004787254333496, 0.01964134407043457, 0.019693632125854493, 0.019654592514038085, 0.01924505615234375, 0.020039680480957032, 0.019739648818969727, 0.01920921516418457, 0.01969049644470215, 0.01986764717102051, 0.020494335174560546, 0.021016576766967773, 0.020304895401000975, 0.019336191177368164, 0.019478527069091797, 0.02000383949279785, 0.01992192077636719, 0.01986457633972168, 0.02100432014465332, 0.01991267204284668, 0.019941375732421874, 0.019893247604370116, 0.019910655975341796, 0.0198287353515625, 0.0198973445892334, 0.019909631729125975, 0.01985228729248047, 0.019825664520263672, 0.019957759857177734, 0.01999564743041992, 0.02001817512512207, 0.07238246154785156, 0.019957759857177734, 0.019902463912963866, 0.020093952178955078, 0.020057088851928712, 0.019907583236694337, 0.020074495315551756, 0.020059135437011717, 0.021238784790039062, 0.02005504035949707, 0.019870719909667968, 0.020012063980102537, 0.02010927963256836, 0.0198984317779541, 0.019869632720947265, 0.019898399353027344, 0.019865568161010743, 0.020091903686523437, 0.019911680221557617, 0.019937280654907227, 0.02002022361755371, 0.01991372871398926, 0.020063232421875, 0.019964927673339843, 0.019887104034423828, 0.019944448471069336, 0.019991552352905274, 0.020008960723876954, 0.019955711364746095, 0.019991552352905274, 0.019942399978637695, 0.019899391174316407, 0.02005401611328125, 0.019877887725830077, 0.019944448471069336, 0.019935232162475586, 0.02000592041015625, 0.020004831314086913, 0.019983360290527344, 0.019818496704101563, 0.019959808349609375, 0.019952640533447266, 0.020152320861816408, 0.019940351486206053, 0.019950592041015625, 0.02000486373901367, 0.02004991912841797, 0.020136959075927736, 0.0202608642578125, 0.019934207916259765, 0.020315135955810547, 0.0200898551940918, 0.01987379264831543, 0.020063232421875, 0.019895296096801757, 0.019212287902832033, 0.019652671813964843, 0.019998655319213868, 0.019939327239990236, 0.02009600067138672, 0.020068351745605468, 0.0200130558013916, 0.019969024658203126, 0.07241318511962891, 0.01904332733154297, 0.01923174476623535, 0.019211263656616212, 0.01924403190612793, 0.019094528198242186, 0.019289087295532227, 0.01921027183532715, 0.019331039428710936, 0.019260416030883788, 0.01920204734802246, 0.019292160034179686, 0.019215360641479492, 0.019150848388671874, 0.01923379135131836, 0.01922559928894043, 0.019212287902832033, 0.019054592132568358, 0.019098655700683594, 0.019611616134643555, 0.019408031463623045, 0.01974870491027832, 0.01988403129577637, 0.019969024658203126, 0.01985740852355957, 0.019738624572753907, 0.019742719650268553, 0.019945472717285157, 0.01987481689453125, 0.019870719909667968, 0.01982156753540039, 0.020006912231445313, 0.019929088592529298, 0.019954687118530275, 0.019720191955566405, 0.020059135437011717, 0.02003763198852539, 0.019894271850585937, 0.019833856582641602, 0.019713024139404296, 0.01978982353210449, 0.019909664154052733, 0.01973142433166504, 0.01904640007019043, 0.019156991958618166, 0.020281343460083007, 0.01999977684020996, 0.019775455474853515, 0.019183616638183593, 0.018946048736572265, 0.01914591979980469, 0.01901958465576172, 0.019117055892944337, 0.018998271942138673, 0.019002431869506835, 0.01896339225769043, 0.01901055908203125, 0.018921472549438476, 0.018982048034667968, 0.019191648483276365, 0.01922870445251465, 0.01903100776672363, 0.018943071365356445, 0.07202703857421874, 0.01900441551208496, 0.01923686408996582, 0.01922665596008301, 0.019155935287475587, 0.018956287384033203, 0.018982912063598634, 0.0192491512298584, 0.019178495407104493, 0.019080192565917968, 0.018968576431274413, 0.018927616119384767, 0.0192542724609375, 0.019091487884521485, 0.019215328216552734, 0.019087360382080077, 0.01923788833618164, 0.01926246452331543, 0.01920204734802246, 0.01924300765991211, 0.019283967971801756, 0.019296255111694336, 0.020113407135009767, 0.019927040100097656, 0.020048896789550782, 0.019900415420532228, 0.019983360290527344, 0.019952640533447266, 0.019955711364746095, 0.01999564743041992, 0.0198604793548584, 0.020100095748901366, 0.01997113609313965, 0.01997920036315918, 0.020008960723876954, 0.019916799545288084, 0.019941375732421874, 0.02003455924987793, 0.02003660774230957, 0.019875904083251954, 0.020047807693481447, 0.01999564743041992, 0.019990528106689453, 0.019969024658203126, 0.0198656005859375, 0.019968000411987305, 0.02000588798522949, 0.019917823791503905, 0.020028511047363282, 0.0198255672454834, 0.019981311798095702, 0.019994623184204103, 0.019960832595825196, 0.019958784103393554, 0.01986457633972168, 0.019921951293945313, 0.020371423721313477, 0.019973119735717772, 0.019904512405395508, 0.019994623184204103, 0.019764223098754884, 0.02044108772277832, 0.020068351745605468, 0.07218994903564453, 0.019598335266113282, 0.02006220817565918, 0.020033536911010744, 0.02027212715148926, 0.019972192764282228, 0.020255647659301757, 0.0200130558013916, 0.020008960723876954, 0.020091903686523437, 0.02043084716796875, 0.02087833595275879, 0.020446207046508787, 0.01923276710510254, 0.0192225284576416, 0.019194879531860352, 0.02007244873046875, 0.02005504035949707, 0.019969024658203126, 0.020001792907714845, 0.019997695922851562, 0.019978239059448243, 0.02004991912841797, 0.020009983062744142, 0.02002943992614746, 0.02000486373901367, 0.020122623443603514, 0.019983360290527344, 0.02002841567993164, 0.02009600067138672, 0.019938304901123048, 0.019977216720581056, 0.020074495315551756, 0.019977279663085937, 0.020011968612670898, 0.020025344848632814, 0.020017215728759766, 0.019993535995483397, 0.020084735870361328, 0.02003455924987793, 0.02006937599182129, 0.019999744415283204, 0.02001203155517578, 0.019973119735717772, 0.019904512405395508, 0.020011007308959963, 0.019974143981933593, 0.019959808349609375, 0.020008960723876954, 0.020017152786254884, 0.02002739143371582, 0.020106239318847655, 0.020187135696411132, 0.019989503860473632, 0.020198400497436524, 0.019927040100097656, 0.02063155174255371, 0.02286079978942871, 0.020477951049804686, 0.020207616806030275, 0.020115455627441405, 0.020180992126464844, 0.020106239318847655, 0.07206092834472656, 0.01927065658569336, 0.019795967102050782, 0.019978239059448243, 0.019975168228149414, 0.019997695922851562, 0.019992576599121094, 0.019945472717285157, 0.020053056716918944, 0.01997715187072754, 0.019981311798095702, 0.02001919937133789, 0.020067327499389647, 0.020059167861938478, 0.02005193519592285, 0.020033536911010744, 0.02000486373901367, 0.020031488418579102, 0.02000588798522949, 0.019801088333129883, 0.0200447998046875, 0.019973119735717772, 0.02004377555847168, 0.02001203155517578, 0.020011007308959963, 0.020727807998657227, 0.021276735305786134, 0.020643775939941406, 0.020557823181152343, 0.020023296356201172, 0.020039680480957032, 0.02001817512512207, 0.020174848556518556, 0.020178943634033202, 0.02007756805419922, 0.02008166313171387, 0.01992192077636719, 0.019985408782958985, 0.02028339195251465, 0.019945472717285157, 0.020362239837646484, 0.019999744415283204, 0.019985408782958985, 0.020082687377929686, 0.02003865623474121, 0.019924991607666014, 0.020124671936035156, 0.020141056060791016, 0.020178943634033202, 0.020237312316894532, 0.020389888763427736, 0.019959808349609375, 0.020335615158081053, 0.02001919937133789, 0.02008166313171387, 0.019998720169067383, 0.020025344848632814, 0.020146175384521483, 0.020121599197387697, 0.02003455924987793, 0.01986764717102051, 0.01964851188659668, 0.020051967620849608, 0.0719482879638672, 0.019336191177368164, 0.020343807220458983, 0.019769344329833984, 0.019886079788208007, 0.01992192077636719, 0.019968000411987305, 0.019962879180908204, 0.020226079940795897, 0.019179487228393556, 0.019164159774780275, 0.019190784454345702, 0.020024320602416993, 0.020131839752197265, 0.019924991607666014, 0.01988198471069336, 0.01987900733947754, 0.020139936447143555, 0.019946495056152345, 0.019981311798095702, 0.019935232162475586, 0.01984819221496582, 0.020164608001708984, 0.019993600845336915, 0.019963903427124022, 0.01998028755187988, 0.019993600845336915, 0.019937280654907227, 0.019951616287231445, 0.019984384536743165, 0.019970048904418947, 0.020001792907714845, 0.019940351486206053, 0.019949567794799804, 0.019918880462646484, 0.019850208282470704, 0.019313663482666017, 0.019999744415283204, 0.019938304901123048, 0.019940351486206053, 0.01970278358459473, 0.019954687118530275, 0.02001817512512207, 0.019991552352905274, 0.019204095840454103, 0.019280895233154297, 0.019191808700561523, 0.019561471939086913, 0.019730432510375977, 0.019378175735473634, 0.019208192825317383, 0.019191808700561523, 0.01923072052001953, 0.019883007049560548, 0.019954687118530275, 0.01997007942199707, 0.01986249542236328, 0.019948543548583983, 0.019914751052856446, 0.020074495315551756, 0.020033536911010744, 0.019947551727294923, 0.019976160049438477, 0.07208448028564453, 0.01923072052001953, 0.01940787124633789, 0.02044825553894043, 0.020187135696411132, 0.019972127914428713, 0.020087776184082032, 0.020221952438354493, 0.020182016372680665, 0.019933183670043944, 0.020011007308959963, 0.019963903427124022, 0.02004991912841797, 0.02009702491760254, 0.019928064346313477, 0.01998847961425781, 0.019945472717285157, 0.020016128540039063, 0.02001203155517578, 0.01991372871398926, 0.01984614372253418, 0.01943244743347168, 0.019556352615356445, 0.020009983062744142, 0.020107263565063475, 0.019978239059448243, 0.02003763198852539, 0.019968000411987305, 0.019811328887939454, 0.01922150421142578, 0.019307519912719725, 0.0192225284576416, 0.019108863830566408, 0.019908607482910155, 0.019689472198486328, 0.01993011283874512, 0.019190784454345702, 0.019216384887695313, 0.02003455924987793, 0.019999744415283204, 0.019957759857177734, 0.019915775299072267, 0.019923967361450197, 0.020130815505981444, 0.020239360809326173, 0.0200263671875, 0.019920896530151368, 0.019970048904418947, 0.020420608520507814, 0.020454399108886717, 0.020231168746948244, 0.02004377555847168, 0.020009983062744142, 0.02082099151611328, 0.019579904556274414, 0.01919491195678711, 0.01927471923828125, 0.019171327590942384, 0.019357696533203125, 0.019185663223266602, 0.019305503845214844, 0.01927676773071289, 0.01925939178466797]",tokens/s,48.6877665411844,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14470,7 +14470,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 1.50 GiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14889,7 +14889,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 100.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14916,7 +14916,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15070,7 +15070,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 432.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15111,7 +15111,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 44.00 MiB. GP ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1135.607808,6809.976832,0.0,6169.821184,6038.342656,s,10,0.4774934082031251,0.0477493408203125,0.0009943449519893373,0.0473834228515625,0.04809864120483399,0.049400473022460936,0.0504419384765625,"[0.05070230484008789, 0.04743491363525391, 0.04748246383666992, 0.04735868835449219, 0.04733385467529297, 0.04736966323852539, 0.04780934524536133, 0.047322303771972656, 0.04728268814086914, 0.04739718246459961]",tokens/s,5361.330556653423,kWh,5.837328303819417e-07,3.198557834281386e-07,4.061099307990119e-06,4.9646879218001994e-06,tokens/kWh,51564167.58360397,MB,1135.607808,6809.976832,0.0,6169.821184,6138.651648,s,10,11.794597167968748,1.179459716796875,0.006327697856050478,1.1794486694335937,1.1890898925781248,1.1891889892578125,1.1892682666015624,"[1.18018701171875, 1.173260986328125, 1.1726756591796874, 1.1892880859375, 1.18906787109375, 1.18304736328125, 1.18335009765625, 1.171730224609375, 1.173279541015625, 1.1787103271484376]",tokens/s,53.41428715436983,kWh,1.3728726102017813e-05,7.521564732477926e-06,4.8728887368410494e-05,6.997917820290624e-05,tokens/kWh,900267.7884745953,,s,629,12.200843275070193,0.019397207114579002,0.005667996487495601,0.018815999984741212,0.019077529525756837,0.019170470046997068,0.06604796112060547,"[0.018860031127929687, 0.01820159912109375, 0.018263103485107422, 0.018254783630371092, 0.018280448913574218, 0.01824665641784668, 0.018233343124389647, 0.018058240890502928, 0.01822412872314453, 0.018313215255737304, 0.018745344161987306, 0.01906585693359375, 0.018976768493652343, 0.01902796745300293, 0.018961408615112304, 0.01888768005371094, 0.018465791702270508, 0.01899519920349121, 0.018923519134521484, 0.01904025650024414, 0.018963455200195312, 0.018961408615112304, 0.01903104019165039, 0.018988031387329102, 0.018931711196899414, 0.018941951751708985, 0.01897881507873535, 0.018932735443115235, 0.0192491512298584, 0.018677759170532226, 0.01861631965637207, 0.018945024490356444, 0.01905971145629883, 0.018983936309814452, 0.019283967971801756, 0.01957683181762695, 0.018371583938598633, 0.018293760299682618, 0.018929664611816405, 0.019009536743164062, 0.018663423538208008, 0.018997247695922852, 0.018293792724609376, 0.01886409568786621, 0.019144704818725586, 0.018929664611816405, 0.018981887817382814, 0.018988031387329102, 0.019079168319702147, 0.01901055908203125, 0.018824192047119142, 0.018156543731689453, 0.018318336486816408, 0.01808793640136719, 0.018749439239501953, 0.018118656158447266, 0.01802444839477539, 0.01808896064758301, 0.018191360473632814, 0.01827020835876465, 0.018667520523071288, 0.018153472900390624, 0.06681088256835938, 0.0190515193939209, 0.01866547203063965, 0.018137088775634767, 0.01803775978088379, 0.01803980827331543, 0.01862348747253418, 0.01824870491027832, 0.018101247787475586, 0.018316287994384766, 0.018667520523071288, 0.018697216033935548, 0.018601984024047852, 0.01819443130493164, 0.0182609920501709, 0.018258943557739257, 0.01820569610595703, 0.018892831802368164, 0.018950111389160158, 0.018149375915527344, 0.01804083251953125, 0.018135040283203126, 0.018309120178222657, 0.01881088066101074, 0.01863065528869629, 0.018777088165283205, 0.01865830421447754, 0.018637823104858398, 0.018715648651123046, 0.018753536224365236, 0.01864089584350586, 0.018405376434326173, 0.01865216064453125, 0.01880681610107422, 0.019593183517456054, 0.01900851249694824, 0.018219039916992186, 0.018270175933837892, 0.01845145606994629, 0.01905356788635254, 0.018687999725341797, 0.018817024230957033, 0.018799615859985352, 0.018713600158691408, 0.018571264266967775, 0.018708480834960937, 0.018928640365600585, 0.019119104385375976, 0.018889728546142577, 0.01868083190917969, 0.018626560211181642, 0.018927616119384767, 0.019162111282348633, 0.01904435157775879, 0.0184770565032959, 0.018840576171875, 0.01865011215209961, 0.01862246322631836, 0.018678783416748047, 0.018657312393188477, 0.01845756721496582, 0.01865216064453125, 0.018538496017456055, 0.06657126617431641, 0.018985984802246093, 0.018959360122680666, 0.018860031127929687, 0.01867263984680176, 0.018735103607177735, 0.018662399291992187, 0.018573312759399413, 0.018103296279907227, 0.018061311721801757, 0.01818623924255371, 0.018722816467285155, 0.018343936920166014, 0.01844633674621582, 0.01801318359375, 0.01801523208618164, 0.018548736572265623, 0.01869824028015137, 0.018249727249145507, 0.01810534477233887, 0.018167808532714845, 0.01859071922302246, 0.018520063400268554, 0.01808076858520508, 0.018033664703369142, 0.01803980827331543, 0.018043903350830077, 0.01865727996826172, 0.01861427116394043, 0.018669567108154296, 0.018922496795654296, 0.018719743728637696, 0.01884979248046875, 0.019076095581054688, 0.01896348762512207, 0.018963424682617188, 0.018662399291992187, 0.01864396858215332, 0.018752511978149415, 0.01906483268737793, 0.0190699520111084, 0.01903001594543457, 0.01820364761352539, 0.01881395149230957, 0.018949119567871094, 0.01860915184020996, 0.01899622344970703, 0.01905971145629883, 0.018913280487060546, 0.019083263397216797, 0.018952192306518553, 0.01900441551208496, 0.018869247436523438, 0.018991104125976564, 0.018751487731933594, 0.01859993553161621, 0.01860710334777832, 0.01862758445739746, 0.018217983245849608, 0.018728960037231446, 0.018471935272216796, 0.018025472640991212, 0.018077695846557617, 0.06666547393798829, 0.018998271942138673, 0.018953216552734374, 0.01898700714111328, 0.018997312545776367, 0.0189869441986084, 0.01901055908203125, 0.018976768493652343, 0.01903001594543457, 0.01905766487121582, 0.018745344161987306, 0.018581504821777343, 0.01905971145629883, 0.018914304733276367, 0.018335744857788085, 0.018735103607177735, 0.018723840713500976, 0.01833580780029297, 0.018297792434692383, 0.019166208267211913, 0.019772415161132813, 0.019360767364501954, 0.019019775390625, 0.01900748825073242, 0.019087360382080077, 0.019136512756347656, 0.019054655075073243, 0.018967487335205077, 0.019082239151000976, 0.018661376953125, 0.018249727249145507, 0.018970624923706055, 0.018962432861328125, 0.018990079879760743, 0.018723840713500976, 0.018736127853393555, 0.018704383850097657, 0.01864806365966797, 0.018122751235961913, 0.0182609920501709, 0.01825484848022461, 0.019062784194946288, 0.01905971145629883, 0.018982912063598634, 0.019099647521972657, 0.01906892776489258, 0.01903206443786621, 0.0190382080078125, 0.0190382080078125, 0.01888768005371094, 0.018906112670898437, 0.018915327072143554, 0.019072000503540038, 0.019013631820678712, 0.01903206443786621, 0.01883647918701172, 0.018559999465942383, 0.01859584045410156, 0.019309568405151366, 0.0184453125, 0.018461696624755858, 0.018941951751708985, 0.018948095321655273, 0.06608179473876953, 0.018217983245849608, 0.018239488601684572, 0.018945024490356444, 0.01887539291381836, 0.018920448303222655, 0.018355199813842774, 0.018497535705566406, 0.018817024230957033, 0.019091455459594727, 0.018965503692626954, 0.02003455924987793, 0.019146751403808594, 0.0192491512298584, 0.01925632095336914, 0.01902387237548828, 0.019338239669799806, 0.018962432861328125, 0.01880268859863281, 0.018778112411499022, 0.018699264526367186, 0.018287616729736327, 0.018184192657470705, 0.01823641586303711, 0.01885593605041504, 0.018897920608520507, 0.018953216552734374, 0.018189311981201172, 0.018888704299926756, 0.018701311111450195, 0.01868191909790039, 0.01860908889770508, 0.01937411117553711, 0.01910780715942383, 0.01886720085144043, 0.019124223709106446, 0.01897881507873535, 0.018874368667602538, 0.018968576431274413, 0.018908159255981445, 0.019173376083374022, 0.01899929618835449, 0.018988031387329102, 0.018882560729980468, 0.01900032043457031, 0.019013696670532227, 0.019054527282714843, 0.019087360382080077, 0.018229248046875, 0.0186429443359375, 0.019054592132568358, 0.019001344680786132, 0.018860031127929687, 0.019013631820678712, 0.018942975997924806, 0.018856960296630858, 0.0189901123046875, 0.01891017532348633, 0.01904947280883789, 0.01904025650024414, 0.018720767974853517, 0.01887539291381836, 0.019080192565917968, 0.06615142059326172, 0.018611200332641603, 0.018726911544799805, 0.018984960556030273, 0.018958335876464845, 0.018914304733276367, 0.018965503692626954, 0.018952192306518553, 0.018900991439819336, 0.018972671508789063, 0.018924543380737305, 0.018249727249145507, 0.018700288772583007, 0.018601984024047852, 0.01881497573852539, 0.018936832427978514, 0.018952192306518553, 0.018966527938842775, 0.018907136917114258, 0.018965503692626954, 0.018965503692626954, 0.018143232345581056, 0.018321407318115233, 0.019087360382080077, 0.018951168060302736, 0.018892799377441406, 0.018966527938842775, 0.018964479446411133, 0.018994176864624023, 0.01901468849182129, 0.0189071044921875, 0.018916351318359375, 0.018897983551025392, 0.019040191650390625, 0.018933759689331055, 0.018259967803955078, 0.018447359085083007, 0.018910207748413087, 0.01903923225402832, 0.01903513526916504, 0.01906175994873047, 0.018586624145507814, 0.01802444839477539, 0.01800704002380371, 0.0185784969329834, 0.018645952224731446, 0.018989055633544923, 0.018934783935546876, 0.018888704299926756, 0.018950143814086915, 0.018685951232910156, 0.018691072463989256, 0.01862041664123535, 0.018086912155151368, 0.01806540870666504, 0.018337791442871093, 0.018601984024047852, 0.019013631820678712, 0.018922496795654296, 0.0190699520111084, 0.018918399810791017, 0.018991104125976564, 0.018928640365600585, 0.06671462249755859, 0.018685951232910156, 0.018954240798950195, 0.018382848739624022, 0.019070976257324217, 0.01903206443786621, 0.01901670455932617, 0.019183616638183593, 0.018979839324951172, 0.018979839324951172, 0.018966527938842775, 0.01902694320678711, 0.018280448913574218, 0.01823744010925293, 0.018257919311523436, 0.01826201629638672, 0.018911231994628908, 0.019140607833862306, 0.018893823623657227, 0.019054592132568358, 0.018990079879760743, 0.01943449592590332, 0.01904742431640625, 0.018984960556030273, 0.018976768493652343, 0.018956287384033203, 0.019194879531860352, 0.019194879531860352, 0.018782207489013672, 0.018817024230957033, 0.01907711982727051, 0.01903001594543457, 0.01901875114440918, 0.01923891258239746, 0.01904537582397461, 0.01900748825073242, 0.01906790351867676, 0.019137535095214844, 0.0190382080078125, 0.019074047088623047, 0.018944000244140623, 0.019075071334838867, 0.01898700714111328, 0.019161088943481445, 0.019050495147705078, 0.018089984893798827, 0.018110464096069336, 0.018667520523071288, 0.018697216033935548, 0.018670623779296874, 0.018024415969848634, 0.018111488342285157, 0.018076671600341796, 0.01868185615539551, 0.01863987159729004, 0.01863270378112793, 0.018149375915527344, 0.0181790714263916, 0.01808076858520508, 0.018505727767944336, 0.01869004821777344, 0.01815657615661621, 0.01865519905090332, 0.0659609603881836, 0.01808896064758301, 0.01805619239807129, 0.01802239990234375, 0.018561023712158203, 0.01865011215209961, 0.018059263229370116, 0.018069503784179687, 0.018182144165039063, 0.018750463485717773, 0.018792448043823243, 0.018719743728637696, 0.018736127853393555, 0.018737152099609376, 0.018964479446411133, 0.018994176864624023, 0.019086336135864256, 0.019072063446044923, 0.019002304077148438, 0.01903615951538086, 0.019057695388793944, 0.018808799743652342, 0.018982912063598634, 0.0186429443359375, 0.01821900749206543, 0.018387968063354493, 0.018792448043823243, 0.018677759170532226, 0.01863372802734375, 0.01869004821777344, 0.018272256851196288, 0.018753536224365236, 0.01869004821777344, 0.018692096710205077, 0.018588672637939452, 0.01878118324279785, 0.01863065528869629, 0.018646015167236327, 0.01863065528869629, 0.01863577651977539, 0.01839308738708496, 0.01812384033203125, 0.018821056365966798, 0.018264064788818358, 0.018118656158447266, 0.018898944854736328, 0.018976768493652343, 0.0192675838470459, 0.019065887451171874, 0.018428895950317385, 0.018131967544555663, 0.018748416900634765, 0.018918399810791017, 0.01904844856262207, 0.018873344421386717, 0.01802342414855957, 0.01803980827331543, 0.018035711288452147, 0.018663423538208008, 0.018504703521728515, 0.018051071166992186, 0.018264064788818358, 0.018684928894042968, 0.06591385650634765, 0.018900991439819336, 0.018989055633544923, 0.020364288330078126, 0.019544063568115236, 0.019111936569213867, 0.018975744247436522, 0.018963455200195312, 0.019095552444458007, 0.018962432861328125, 0.018954240798950195, 0.018692096710205077, 0.018127872467041017, 0.01810534477233887, 0.018061311721801757, 0.01816268730163574, 0.01823744010925293, 0.018326528549194337, 0.01825484848022461, 0.018310144424438478, 0.018256895065307616, 0.018966527938842775, 0.01843404769897461, 0.018304000854492186, 0.018268159866333008, 0.018316287994384766, 0.018229280471801758, 0.018230239868164064, 0.01821286392211914, 0.01821286392211914, 0.018307071685791015, 0.01821388816833496, 0.01826201629638672, 0.018505727767944336, 0.01804287910461426, 0.018156543731689453, 0.01820159912109375, 0.018751487731933594, 0.018697216033935548, 0.018619392395019533, 0.01816268730163574, 0.01845043182373047, 0.019002368927001953, 0.018946048736572265, 0.019096576690673828, 0.019124223709106446, 0.019729408264160156, 0.01916214370727539, 0.019052511215209962, 0.019096576690673828, 0.0190515193939209, 0.018974720001220705, 0.01887129592895508, 0.01865216064453125, 0.01819647979736328, 0.018336767196655272, 0.01823641586303711, 0.01865318489074707, 0.018937856674194335, 0.018952192306518553, 0.01818726348876953, 0.018258943557739257, 0.01822412872314453, 0.06621593475341797, 0.01827020835876465, 0.01817087936401367, 0.018356224060058594, 0.018256895065307616, 0.01828556823730469, 0.018249727249145507, 0.018239488601684572, 0.018508800506591795, 0.01907302474975586, 0.019140607833862306, 0.018952192306518553, 0.018944000244140623, 0.01900339126586914, 0.019013631820678712, 0.019014656066894533, 0.01903411293029785, 0.018811904907226562, 0.018576383590698242, 0.01826304054260254, 0.018288639068603514, 0.018239488601684572, 0.018215936660766603, 0.01822822380065918, 0.018255872726440428, 0.018815999984741212, 0.018967552185058592, 0.018936832427978514, 0.01903513526916504, 0.01903411293029785, 0.01899212837219238, 0.018981887817382814, 0.01904025650024414, 0.019005504608154297, 0.019173311233520507, 0.01900851249694824, 0.018981887817382814, 0.019019775390625, 0.01888768005371094, 0.01864908790588379, 0.018773056030273436, 0.018710464477539063, 0.018217983245849608, 0.01883443260192871, 0.018584575653076172, 0.018268159866333008, 0.018309120178222657, 0.01819647979736328, 0.01826918411254883, 0.01820364761352539, 0.01822003173828125, 0.018199552536010744, 0.01909350395202637, 0.01904947280883789, 0.01898089599609375, 0.01892246437072754, 0.018953216552734374, 0.01902694320678711, 0.018982912063598634, 0.018974720001220705, 0.019098623275756836, 0.019100671768188478, 0.019013631820678712]",tokens/s,51.55381360280454,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1146.540032,12579.241984,0.0,11939.086336,11714.427904,s,10,1.006503044128418,0.1006503044128418,0.0014761109927726888,0.10015551757812499,0.10116352462768555,0.10308752250671387,0.10462672080993653,"[0.10501152038574219, 0.10034623718261719, 0.100162109375, 0.10001036834716796, 0.09981382751464844, 0.10006797027587891, 0.09984611511230469, 0.10014892578125, 0.10036000061035157, 0.10073596954345704]",tokens/s,2543.459768884091,kWh,1.206257881896566e-06,6.609661215741058e-07,8.43877375781639e-06,1.0305997761287061e-05,tokens/kWh,24839904.483739138,MB,1146.540032,12579.241984,0.0,11939.086336,11881.618432,s,10,18.944419067382814,1.8944419067382814,0.005874336927729204,1.8952128906249999,1.9003347290039063,1.9019706481933594,1.9032793835449218,"[1.894547119140625, 1.8846368408203125, 1.886148681640625, 1.895878662109375, 1.9036065673828124, 1.8965111083984374, 1.894208984375, 1.8890150146484375, 1.89997119140625, 1.8998948974609375]",tokens/s,33.25517651183563,kWh,2.219931550273279e-05,1.2164041081180825e-05,8.769219203578269e-05,0.00012205554861969631,tokens/kWh,516158.42714612634,,s,629,19.815456756591768,0.031503110900781874,0.011961892980804514,0.030057472229003908,0.030304045104980467,0.03044679718017578,0.13060185424804688,"[0.029978624343872072, 0.02998476791381836, 0.0300001277923584, 0.029764608383178712, 0.029841407775878907, 0.029799423217773437, 0.02977484893798828, 0.02978816032409668, 0.02979635238647461, 0.029853696823120116, 0.029908992767333983, 0.029791231155395507, 0.029829120635986327, 0.02979020881652832, 0.02987014389038086, 0.029915071487426757, 0.030509056091308592, 0.03040870475769043, 0.030235647201538086, 0.03013324737548828, 0.030127103805541993, 0.030176256179809572, 0.030445568084716795, 0.03039129638671875, 0.030096384048461915, 0.030216192245483397, 0.03016089630126953, 0.03013734436035156, 0.030187519073486328, 0.030100479125976562, 0.030242816925048828, 0.030076927185058593, 0.03015475273132324, 0.030273536682128906, 0.030233600616455077, 0.030192640304565428, 0.02995199966430664, 0.029953023910522462, 0.030202880859375, 0.029867008209228517, 0.030085119247436523, 0.02996735954284668, 0.029906944274902345, 0.029858816146850587, 0.029901824951171874, 0.03000531196594238, 0.03026425552368164, 0.030286848068237306, 0.03002060890197754, 0.02995097541809082, 0.029936639785766602, 0.029880319595336914, 0.029921279907226563, 0.02994486427307129, 0.029961183547973634, 0.030107648849487304, 0.03036057662963867, 0.030501888275146483, 0.03057868766784668, 0.030091264724731445, 0.03002572822570801, 0.030067712783813476, 0.13073715209960937, 0.029765695571899415, 0.029743040084838867, 0.02997452735900879, 0.030110719680786133, 0.030036991119384765, 0.02981177520751953, 0.029830080032348633, 0.02976153564453125, 0.029823999404907226, 0.02976870346069336, 0.029879295349121093, 0.02977791976928711, 0.0297574405670166, 0.029775871276855468, 0.029839359283447265, 0.02979327964782715, 0.029967391967773437, 0.029869024276733398, 0.02975129508972168, 0.02994790458679199, 0.029709312438964845, 0.02979327964782715, 0.029896703720092774, 0.029882368087768556, 0.029807615280151366, 0.029880319595336914, 0.02993356704711914, 0.029826047897338868, 0.029840383529663086, 0.029821952819824218, 0.029833215713500977, 0.029850688934326172, 0.029847488403320313, 0.029826047897338868, 0.029929471969604493, 0.029916160583496092, 0.029930496215820314, 0.029920255661010742, 0.029881376266479492, 0.029929439544677736, 0.029944831848144532, 0.030050304412841795, 0.030027776718139648, 0.030212095260620117, 0.0301844482421875, 0.029954048156738283, 0.02993152046203613, 0.02996940803527832, 0.030042112350463866, 0.030015487670898438, 0.029965311050415038, 0.030015487670898438, 0.02997452735900879, 0.02995916748046875, 0.029973503112792968, 0.02999603271484375, 0.029978624343872072, 0.029970432281494142, 0.02997452735900879, 0.0299233283996582, 0.02995916748046875, 0.03000115203857422, 0.130845703125, 0.029944831848144532, 0.029833215713500977, 0.02979327964782715, 0.029833215713500977, 0.029821952819824218, 0.02981068801879883, 0.02979532814025879, 0.02978508758544922, 0.029842432022094727, 0.029856767654418945, 0.029877248764038085, 0.02983526420593262, 0.029823999404907226, 0.02979020881652832, 0.029911039352416992, 0.02992742347717285, 0.029937664031982423, 0.03018342399597168, 0.03017420768737793, 0.02977996826171875, 0.029846527099609374, 0.029900800704956054, 0.02995814323425293, 0.029863935470581054, 0.029896703720092774, 0.029901824951171874, 0.029891584396362306, 0.029853696823120116, 0.029862911224365234, 0.029878271102905272, 0.029894655227661132, 0.029921279907226563, 0.029889535903930665, 0.030102527618408204, 0.02992742347717285, 0.02996633529663086, 0.029872127532958984, 0.029933631896972655, 0.02987718391418457, 0.029938688278198244, 0.029925376892089843, 0.02994892883300781, 0.029962240219116212, 0.029937664031982423, 0.029945856094360353, 0.02992639923095703, 0.02996735954284668, 0.02995916748046875, 0.030113792419433592, 0.030280704498291015, 0.03016499137878418, 0.029962240219116212, 0.029899776458740233, 0.030013439178466796, 0.029929471969604493, 0.03001651191711426, 0.02995916748046875, 0.03000217628479004, 0.02995916748046875, 0.02999091148376465, 0.029917184829711913, 0.02998784065246582, 0.13061016845703124, 0.03016499137878418, 0.02983526420593262, 0.02998067283630371, 0.030138368606567382, 0.03013324737548828, 0.030097408294677733, 0.03011686325073242, 0.030065664291381834, 0.02998579216003418, 0.029834239959716798, 0.029853696823120116, 0.029867040634155274, 0.029844480514526366, 0.02983216094970703, 0.029858816146850587, 0.02994486427307129, 0.030125024795532227, 0.03013734436035156, 0.030144512176513674, 0.030113792419433592, 0.029829120635986327, 0.02980659294128418, 0.029863935470581054, 0.029894655227661132, 0.029907968521118163, 0.029873151779174805, 0.029925376892089843, 0.029872127532958984, 0.030090240478515624, 0.030204927444458008, 0.02993561553955078, 0.02994380760192871, 0.03017932891845703, 0.029865983963012696, 0.029856767654418945, 0.030028799057006835, 0.03021824073791504, 0.029944831848144532, 0.029907968521118163, 0.029954048156738283, 0.029887487411499023, 0.029911039352416992, 0.0299683837890625, 0.030538751602172853, 0.03097599983215332, 0.030530559539794923, 0.030280704498291015, 0.030288896560668944, 0.030272512435913085, 0.030460927963256838, 0.030239744186401366, 0.030265344619750976, 0.030255104064941408, 0.030307327270507813, 0.030318592071533205, 0.030271488189697264, 0.03032268714904785, 0.03033087921142578, 0.03037696075439453, 0.030303232192993163, 0.030002208709716798, 0.03001955223083496, 0.13050469970703124, 0.030180383682250976, 0.03005955123901367, 0.03061958312988281, 0.030086143493652344, 0.030023679733276368, 0.030071807861328126, 0.03011686325073242, 0.030117887496948242, 0.030096384048461915, 0.03012505531311035, 0.030079999923706056, 0.030111743927001954, 0.03011177635192871, 0.030183391571044924, 0.030135295867919923, 0.03018854331970215, 0.030288896560668944, 0.030303232192993163, 0.030144512176513674, 0.030093311309814453, 0.030151679992675783, 0.030102527618408204, 0.03014041519165039, 0.030076927185058593, 0.030166015625, 0.030103551864624024, 0.030126079559326172, 0.03015884780883789, 0.03021516799926758, 0.03018649673461914, 0.030176256179809572, 0.03015884780883789, 0.030277664184570313, 0.030210079193115233, 0.03016294479370117, 0.030209983825683594, 0.029880319595336914, 0.03016908836364746, 0.030240768432617186, 0.030219263076782226, 0.030219263076782226, 0.03021516799926758, 0.030252031326293945, 0.030213119506835938, 0.03021824073791504, 0.030202880859375, 0.03021824073791504, 0.030235647201538086, 0.030255104064941408, 0.030858240127563476, 0.030486528396606444, 0.030264320373535155, 0.030298112869262695, 0.03033497619628906, 0.03022233581542969, 0.030257152557373046, 0.03041996765136719, 0.030442527770996094, 0.030309343338012697, 0.03023155212402344, 0.030263296127319338, 0.030234624862670898, 0.13058047485351562, 0.029891584396362306, 0.03017932891845703, 0.030146623611450197, 0.029785024642944337, 0.02981068801879883, 0.029844480514526366, 0.030144512176513674, 0.030112768173217775, 0.030102527618408204, 0.029812736511230467, 0.03016499137878418, 0.030136320114135744, 0.030095359802246095, 0.030069759368896484, 0.030090240478515624, 0.030104576110839845, 0.03021516799926758, 0.030311424255371092, 0.030297088623046874, 0.030101503372192383, 0.030103551864624024, 0.030187519073486328, 0.03012505531311035, 0.03054899215698242, 0.03037593650817871, 0.0301977596282959, 0.029902847290039062, 0.029914112091064454, 0.029858816146850587, 0.029922304153442384, 0.029885440826416015, 0.029865983963012696, 0.029884416580200194, 0.02996019172668457, 0.029877248764038085, 0.0299683837890625, 0.0299233283996582, 0.029904895782470704, 0.029901824951171874, 0.030048320770263672, 0.03023251152038574, 0.030196735382080078, 0.03019161605834961, 0.030136320114135744, 0.030200832366943358, 0.030258176803588867, 0.030248960494995116, 0.030303232192993163, 0.030196735382080078, 0.030254079818725587, 0.030005247116088866, 0.02996633529663086, 0.02999603271484375, 0.02999295997619629, 0.02998067283630371, 0.03002572822570801, 0.02999193572998047, 0.030057472229003908, 0.030993408203125, 0.030447616577148437, 0.030279680252075194, 0.030340095520019532, 0.13070851135253905, 0.030311391830444335, 0.03015475273132324, 0.030101503372192383, 0.03001759910583496, 0.03007174491882324, 0.03014143943786621, 0.030125120162963866, 0.030121919631958007, 0.030091264724731445, 0.030076927185058593, 0.0301977596282959, 0.030107648849487304, 0.029829120635986327, 0.02982809638977051, 0.029864959716796875, 0.030187519073486328, 0.030264352798461912, 0.030161888122558593, 0.029851648330688478, 0.02994892883300781, 0.029853696823120116, 0.029860864639282225, 0.029894655227661132, 0.029905920028686524, 0.029837312698364257, 0.029860864639282225, 0.029865983963012696, 0.02981888008117676, 0.029815807342529296, 0.029879295349121093, 0.029902847290039062, 0.029905920028686524, 0.030094335556030274, 0.030224384307861327, 0.030198848724365235, 0.030155712127685547, 0.03016908836364746, 0.030211103439331054, 0.03024073600769043, 0.03023155212402344, 0.029940799713134767, 0.029922239303588866, 0.02992848014831543, 0.029909984588623047, 0.029886463165283202, 0.02993152046203613, 0.02991417694091797, 0.030027711868286133, 0.030088191986083986, 0.029936639785766602, 0.030110719680786133, 0.030238719940185548, 0.030225408554077147, 0.030309375762939454, 0.029945856094360353, 0.03038412857055664, 0.030298112869262695, 0.030292991638183595, 0.030281728744506835, 0.030239744186401366, 0.030275583267211914, 0.029917184829711913, 0.13073408508300782, 0.029876224517822264, 0.029859840393066408, 0.029792255401611328, 0.029832191467285156, 0.02977894401550293, 0.02978713607788086, 0.02978201675415039, 0.029799423217773437, 0.029830144882202148, 0.029783039093017577, 0.029826047897338868, 0.029838336944580077, 0.029825023651123047, 0.029800447463989257, 0.029877248764038085, 0.029861888885498046, 0.029892608642578124, 0.029954048156738283, 0.03019878387451172, 0.03019878387451172, 0.03018547248840332, 0.03014860725402832, 0.0301711368560791, 0.029879295349121093, 0.029848575592041016, 0.029851680755615236, 0.029930463790893556, 0.029877248764038085, 0.029898752212524415, 0.029889535903930665, 0.029932544708251952, 0.029911039352416992, 0.0299182071685791, 0.029944831848144532, 0.029901824951171874, 0.02992742347717285, 0.029914112091064454, 0.03021107292175293, 0.02994175910949707, 0.02993971252441406, 0.029925376892089843, 0.029911039352416992, 0.02999091148376465, 0.029897727966308595, 0.029925376892089843, 0.02999603271484375, 0.029981695175170898, 0.02995916748046875, 0.02994790458679199, 0.030107648849487304, 0.03035852813720703, 0.030232576370239257, 0.02998886489868164, 0.02994380760192871, 0.030064640045166017, 0.030262271881103517, 0.030261247634887696, 0.030287872314453124, 0.030321664810180664, 0.03013324737548828, 0.030310400009155275, 0.03002572822570801, 0.1309102020263672, 0.02998067283630371, 0.029817855834960938, 0.03002060890197754, 0.03015372848510742, 0.030138368606567382, 0.030254079818725587, 0.03015372848510742, 0.030114816665649413, 0.03018649673461914, 0.030086143493652344, 0.02996019172668457, 0.029797376632690428, 0.029851648330688478, 0.030028799057006835, 0.030102527618408204, 0.02980454444885254, 0.030031871795654298, 0.029996063232421873, 0.03030729675292969, 0.030134271621704102, 0.030081024169921877, 0.029820928573608397, 0.029830144882202148, 0.030049280166625978, 0.03015065574645996, 0.029854719161987304, 0.02981888008117676, 0.029825088500976562, 0.030040000915527342, 0.03020185661315918, 0.03015782356262207, 0.030061567306518554, 0.029915136337280275, 0.030023679733276368, 0.03017011260986328, 0.029944831848144532, 0.03016499137878418, 0.03017523193359375, 0.03022336006164551, 0.03019878387451172, 0.03017318344116211, 0.03021830368041992, 0.030189504623413087, 0.03030940818786621, 0.030727136611938478, 0.03075584030151367, 0.030487552642822265, 0.030320640563964843, 0.030309375762939454, 0.030317567825317384, 0.030608383178710938, 0.03037286376953125, 0.030500864028930662, 0.030321664810180664, 0.030278656005859376, 0.03022336006164551, 0.030245887756347657, 0.030270463943481447, 0.030130176544189452, 0.030299135208129883, 0.030283775329589844, 0.030453760147094725, 0.13132389831542968, 0.029904895782470704, 0.029965311050415038, 0.030178367614746095, 0.03011577606201172, 0.02997760009765625, 0.030064640045166017, 0.030078975677490235, 0.03015782356262207, 0.030139392852783203, 0.030113792419433592, 0.030087167739868165, 0.03013222312927246, 0.030165056228637695, 0.030129087448120116, 0.030071807861328126, 0.03018342399597168, 0.030198816299438477, 0.030581727981567382, 0.030224447250366212, 0.03013420867919922, 0.03013324737548828, 0.030120960235595705, 0.03015475273132324, 0.030104576110839845, 0.03015577507019043, 0.030143487930297853, 0.030134271621704102, 0.030100479125976562, 0.030192640304565428, 0.03020595169067383, 0.03019059181213379, 0.0301977596282959, 0.030177312850952147, 0.030157791137695313, 0.03016908836364746, 0.03017011260986328, 0.030154783248901366, 0.03024995231628418, 0.030154783248901366, 0.030143455505371095, 0.02993971252441406, 0.02994380760192871, 0.029937664031982423, 0.0299683837890625, 0.02994175910949707, 0.03023155212402344, 0.02997760009765625, 0.02997248077392578, 0.030139392852783203, 0.030229503631591798, 0.030508031845092775, 0.030713855743408205, 0.030452735900878908, 0.030265344619750976, 0.030195711135864257, 0.030221343994140625, 0.03004105567932129, 0.03031449508666992, 0.030327808380126952, 0.030243839263916016, 0.030029823303222656, 0.02996428871154785]",tokens/s,31.742896857058682,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15665,7 +15665,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 896.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1176.375296,9272.03328,0.0,8631.877632,8498.684928,s,10,0.5381008338928224,0.053810083389282226,0.00035037334611295,0.05368016052246094,0.05422293357849121,0.05444493808746338,0.05462254169464112,"[0.05466694259643555, 0.05365244674682617, 0.05390115356445312, 0.05355408096313476, 0.05359164810180664, 0.05367539215087891, 0.053816032409667966, 0.05338460922241211, 0.05417359924316406, 0.05368492889404297]",tokens/s,4757.472649651933,kWh,6.675134018316108e-07,3.6573236569800707e-07,4.647660654723297e-06,5.680906422252914e-06,tokens/kWh,45063231.28246081,MB,1176.702976,9272.03328,0.0,8631.877632,8597.292544,s,10,13.016972412109373,1.3016972412109373,0.008260526575922323,1.2994389038085936,1.31171298828125,1.3148390380859374,1.3173398779296874,"[1.3035977783203125, 1.296099609375, 1.2949403076171875, 1.3008309326171874, 1.311018310546875, 1.317965087890625, 1.3095673828125, 1.298046875, 1.291166259765625, 1.2937398681640624]",tokens/s,48.39835101854609,kWh,1.532481990768636e-05,8.397759716100169e-06,5.2373523851075e-05,7.609610347486153e-05,tokens/kWh,827900.4722076493,,s,629,13.478948848724393,0.021429171460611074,0.006443288456210155,0.020567039489746093,0.021075148391723635,0.021187609863281248,0.0746680322265625,"[0.020566015243530272, 0.020549631118774413, 0.020551679611206054, 0.020537343978881836, 0.020470783233642577, 0.02045644760131836, 0.02043084716796875, 0.020504575729370117, 0.020477951049804686, 0.020528127670288086, 0.02048409652709961, 0.020271104812622072, 0.020376575469970702, 0.020439071655273436, 0.02044108772277832, 0.02051171112060547, 0.020512767791748047, 0.020567039489746093, 0.020534271240234374, 0.02043187141418457, 0.02045747184753418, 0.020431936264038084, 0.020485055923461913, 0.021149696350097655, 0.021112831115722656, 0.02450432014465332, 0.022615039825439453, 0.021727231979370116, 0.02063667106628418, 0.02065407943725586, 0.020559871673583984, 0.02058956718444824, 0.020658176422119142, 0.02061516761779785, 0.020625408172607423, 0.020585472106933594, 0.020552703857421875, 0.020546560287475587, 0.020584447860717774, 0.020582399368286132, 0.020556800842285155, 0.020527103424072265, 0.020550655364990233, 0.020534271240234374, 0.02058956718444824, 0.020599807739257812, 0.020612096786499022, 0.020561920166015626, 0.020567039489746093, 0.020618240356445314, 0.020618240356445314, 0.020588544845581053, 0.020548608779907225, 0.020584447860717774, 0.020509695053100584, 0.020578336715698243, 0.02053424072265625, 0.02061516761779785, 0.020540416717529295, 0.020595712661743162, 0.020567039489746093, 0.020543487548828124, 0.07469465637207032, 0.02064691162109375, 0.020553728103637696, 0.020607999801635742, 0.020572160720825194, 0.02058956718444824, 0.020641792297363282, 0.020555776596069338, 0.02062335968017578, 0.02040934371948242, 0.020392959594726562, 0.020546560287475587, 0.020552703857421875, 0.020520959854125977, 0.020706304550170897, 0.020552703857421875, 0.02057318305969238, 0.020582399368286132, 0.020609024047851563, 0.020525056838989256, 0.020570144653320313, 0.02054550361633301, 0.020600831985473633, 0.020516864776611327, 0.020562944412231447, 0.020488191604614257, 0.020559871673583984, 0.02057222366333008, 0.020487104415893555, 0.020580352783203124, 0.02108006477355957, 0.020548608779907225, 0.02049843215942383, 0.02057219123840332, 0.020572128295898436, 0.020519935607910156, 0.020564992904663085, 0.020548608779907225, 0.02052403259277344, 0.02050662422180176, 0.020568063735961914, 0.020541439056396483, 0.020559871673583984, 0.02062335968017578, 0.020572160720825194, 0.020575231552124023, 0.020634624481201173, 0.020550655364990233, 0.020582496643066408, 0.020598751068115234, 0.02036934471130371, 0.020374528884887694, 0.020591615676879883, 0.02053222465515137, 0.020577280044555665, 0.020587520599365236, 0.020571168899536134, 0.020640735626220704, 0.020626432418823244, 0.020609024047851563, 0.020547615051269532, 0.020557792663574218, 0.020529151916503906, 0.07494348907470703, 0.020534271240234374, 0.020545536041259766, 0.020559871673583984, 0.02051481628417969, 0.020557823181152343, 0.020520959854125977, 0.020517887115478514, 0.020529151916503906, 0.020557823181152343, 0.020545536041259766, 0.020582399368286132, 0.02062848091125488, 0.020505599975585938, 0.020634624481201173, 0.020554752349853517, 0.020562944412231447, 0.020658176422119142, 0.020600831985473633, 0.020579328536987306, 0.020617216110229493, 0.0206561279296875, 0.020562944412231447, 0.020582399368286132, 0.020591615676879883, 0.02060697555541992, 0.020595712661743162, 0.020576255798339844, 0.020572160720825194, 0.020534271240234374, 0.020519935607910156, 0.020592639923095703, 0.020397056579589845, 0.020350976943969725, 0.020366336822509764, 0.0204400634765625, 0.020536319732666015, 0.02059775924682617, 0.020561920166015626, 0.02057318305969238, 0.020591615676879883, 0.020641792297363282, 0.020561920166015626, 0.020542463302612304, 0.02052511978149414, 0.02057209587097168, 0.020528127670288086, 0.020555776596069338, 0.020601856231689454, 0.020525056838989256, 0.02054969596862793, 0.020471744537353516, 0.020505599975585938, 0.02050048065185547, 0.0204902400970459, 0.020550655364990233, 0.020518911361694335, 0.020539392471313478, 0.020551679611206054, 0.020489248275756836, 0.020499423980712892, 0.02048204803466797, 0.020580352783203124, 0.07464959716796875, 0.020530176162719727, 0.020580352783203124, 0.020523008346557618, 0.020505599975585938, 0.020578304290771485, 0.02047488021850586, 0.020584447860717774, 0.020505599975585938, 0.020541439056396483, 0.020504575729370117, 0.020620288848876952, 0.020545536041259766, 0.020521984100341797, 0.02052403259277344, 0.020555776596069338, 0.020548608779907225, 0.020973567962646485, 0.02058137512207031, 0.020742143630981445, 0.02045952033996582, 0.020552736282348633, 0.020530208587646485, 0.02051068878173828, 0.02051375961303711, 0.020632575988769532, 0.020578304290771485, 0.020504575729370117, 0.020550655364990233, 0.020550655364990233, 0.020540416717529295, 0.02058038330078125, 0.020603872299194335, 0.020570175170898437, 0.02053215980529785, 0.020548608779907225, 0.020898815155029296, 0.020549631118774413, 0.020494335174560546, 0.020764671325683593, 0.020560895919799805, 0.02067865562438965, 0.020518911361694335, 0.020534271240234374, 0.021013504028320314, 0.020619264602661135, 0.02058963203430176, 0.02100524711608887, 0.020634624481201173, 0.020925439834594727, 0.02063155174255371, 0.020593664169311524, 0.020974592208862306, 0.02051481628417969, 0.020537343978881836, 0.02088960075378418, 0.020540416717529295, 0.02067865562438965, 0.021784576416015625, 0.021168127059936523, 0.02091929626464844, 0.020542463302612304, 0.020528127670288086, 0.07467520141601562, 0.020539392471313478, 0.02046771240234375, 0.020999168395996092, 0.02051481628417969, 0.020676607131958007, 0.020575231552124023, 0.02050048065185547, 0.020880384445190428, 0.020543487548828124, 0.020577280044555665, 0.021061632156372072, 0.02063871955871582, 0.020769792556762694, 0.020593664169311524, 0.020521984100341797, 0.020977664947509765, 0.021168127059936523, 0.02085785675048828, 0.02107596778869629, 0.020574207305908202, 0.02063155174255371, 0.021082111358642578, 0.021113855361938477, 0.020913152694702147, 0.020544511795043945, 0.020685823440551757, 0.021234687805175782, 0.021046272277832033, 0.020544511795043945, 0.02065715217590332, 0.02063667106628418, 0.020624383926391602, 0.020548608779907225, 0.020567039489746093, 0.020832256317138673, 0.021118976593017577, 0.021170175552368165, 0.02109644889831543, 0.020596736907958983, 0.020917247772216797, 0.020561920166015626, 0.02053222465515137, 0.02099404716491699, 0.020546560287475587, 0.02049849510192871, 0.021067712783813478, 0.020572160720825194, 0.020786176681518553, 0.02050662422180176, 0.02066739273071289, 0.021170175552368165, 0.020701183319091796, 0.020668415069580077, 0.021178367614746094, 0.021135360717773437, 0.021204992294311522, 0.020900863647460938, 0.021169151306152344, 0.02084966468811035, 0.020819007873535158, 0.02115987205505371, 0.021189632415771483, 0.07523020935058594, 0.021137407302856445, 0.02123673629760742, 0.021105663299560547, 0.02109951972961426, 0.02107494354248047, 0.021194751739501954, 0.02106470489501953, 0.021117952346801756, 0.02072985649108887, 0.021176319122314453, 0.021184576034545897, 0.021138368606567384, 0.02123776054382324, 0.021234687805175782, 0.02108006477355957, 0.020992000579833983, 0.021158912658691405, 0.02087120056152344, 0.020984800338745117, 0.02116204833984375, 0.020683712005615234, 0.02109951972961426, 0.021008384704589843, 0.02205081558227539, 0.02228428840637207, 0.020999168395996092, 0.02060492706298828, 0.020634624481201173, 0.020538368225097657, 0.020548608779907225, 0.02057318305969238, 0.020527103424072265, 0.02051481628417969, 0.020764671325683593, 0.020509695053100584, 0.02066329574584961, 0.02053222465515137, 0.020508672714233397, 0.020891647338867187, 0.020736000061035157, 0.020557823181152343, 0.021029888153076173, 0.020518911361694335, 0.020738048553466795, 0.020540416717529295, 0.020601856231689454, 0.02089574432373047, 0.020503551483154296, 0.020599807739257812, 0.020755456924438476, 0.020509695053100584, 0.020668415069580077, 0.020478975296020507, 0.020530176162719727, 0.02085171127319336, 0.022261760711669923, 0.021413888931274414, 0.02106982421875, 0.021121023178100586, 0.020837440490722656, 0.02063046455383301, 0.02067251205444336, 0.07491276550292969, 0.020529151916503906, 0.020552703857421875, 0.02084966468811035, 0.020824064254760744, 0.02062131118774414, 0.02108415985107422, 0.021398527145385742, 0.02079641532897949, 0.02062745666503906, 0.02059878349304199, 0.021196800231933592, 0.020633600234985353, 0.020704288482666016, 0.021265375137329103, 0.02125209617614746, 0.02085478401184082, 0.020590591430664062, 0.02062131118774414, 0.02105036735534668, 0.020569087982177735, 0.020583423614501953, 0.020937728881835937, 0.020891647338867187, 0.020775936126708985, 0.02057318305969238, 0.020725759506225586, 0.020574272155761717, 0.020619199752807616, 0.020628543853759767, 0.02058028793334961, 0.020948991775512696, 0.020566015243530272, 0.020997119903564454, 0.020587520599365236, 0.02064896011352539, 0.0206080322265625, 0.020601823806762697, 0.020732927322387695, 0.02103910446166992, 0.020678688049316406, 0.02053424072265625, 0.020520959854125977, 0.020592639923095703, 0.02064793586730957, 0.022018047332763673, 0.02227302360534668, 0.021588991165161133, 0.02166374397277832, 0.020633600234985353, 0.02060492706298828, 0.020571136474609376, 0.020560895919799805, 0.020553728103637696, 0.020620288848876952, 0.020539392471313478, 0.02048204803466797, 0.020478975296020507, 0.020536319732666015, 0.020562944412231447, 0.020546560287475587, 0.020534271240234374, 0.02050764846801758, 0.07447142028808594, 0.020553728103637696, 0.020554752349853517, 0.020596736907958983, 0.020541439056396483, 0.02058137512207031, 0.020570112228393556, 0.020577280044555665, 0.020556800842285155, 0.020595712661743162, 0.020559871673583984, 0.020493343353271486, 0.020548576354980468, 0.02068377685546875, 0.020345855712890625, 0.020436992645263673, 0.02061516761779785, 0.020519935607910156, 0.020559871673583984, 0.0204769287109375, 0.020556800842285155, 0.020539392471313478, 0.0208721923828125, 0.02109132766723633, 0.020596736907958983, 0.020518911361694335, 0.020555776596069338, 0.020591615676879883, 0.020634624481201173, 0.021158912658691405, 0.020550655364990233, 0.02083328056335449, 0.020619264602661135, 0.020574207305908202, 0.02051584053039551, 0.02067865562438965, 0.02049228858947754, 0.020544544219970703, 0.020487136840820312, 0.020590591430664062, 0.0206059513092041, 0.02053222465515137, 0.020585472106933594, 0.020562944412231447, 0.020495359420776366, 0.020488191604614257, 0.020595712661743162, 0.020534271240234374, 0.020638751983642577, 0.02068681526184082, 0.020571136474609376, 0.020542463302612304, 0.020577280044555665, 0.020545536041259766, 0.020658176422119142, 0.020560895919799805, 0.02059878349304199, 0.02082918357849121, 0.020584447860717774, 0.02065203285217285, 0.020741119384765624, 0.020510719299316405, 0.020554752349853517, 0.07528959655761719, 0.020634624481201173, 0.02071347236633301, 0.020583423614501953, 0.02048409652709961, 0.020546560287475587, 0.020542463302612304, 0.020498464584350586, 0.020568031311035156, 0.020496383666992187, 0.020564992904663085, 0.020497440338134765, 0.020544479370117188, 0.020550655364990233, 0.020541439056396483, 0.020519935607910156, 0.020454399108886717, 0.020319232940673827, 0.020497407913208008, 0.020312063217163084, 0.020316160202026368, 0.02050048065185547, 0.020551679611206054, 0.020568063735961914, 0.020388864517211915, 0.020374528884887694, 0.02064691162109375, 0.020544511795043945, 0.020530176162719727, 0.02048204803466797, 0.02053222465515137, 0.020519935607910156, 0.020365312576293947, 0.020348928451538087, 0.020312063217163084, 0.020257791519165038, 0.020295679092407228, 0.02046976089477539, 0.020517887115478514, 0.020527103424072265, 0.020571136474609376, 0.020427776336669923, 0.02031820869445801, 0.020382720947265624, 0.02045849609375, 0.020477951049804686, 0.020531200408935548, 0.020575231552124023, 0.02048307228088379, 0.020536319732666015, 0.020486143112182616, 0.020479999542236327, 0.020476959228515626, 0.02046771240234375, 0.020469728469848632, 0.02046156883239746, 0.020518943786621092, 0.0205383358001709, 0.020504575729370117, 0.02061414337158203, 0.020493312835693358, 0.02042367935180664, 0.020468736648559572, 0.07475609588623047, 0.020543487548828124, 0.020519935607910156, 0.020510719299316405, 0.020563968658447264, 0.02059878349304199, 0.02046463966369629, 0.02051584053039551, 0.02050764846801758, 0.02055990409851074, 0.02055369567871094, 0.020576255798339844, 0.020520959854125977, 0.020556800842285155, 0.020510719299316405, 0.020493312835693358, 0.020632575988769532, 0.02030182456970215, 0.02045849609375, 0.020544511795043945, 0.02050048065185547, 0.020558847427368163, 0.020546560287475587, 0.020553728103637696, 0.020557823181152343, 0.02041753578186035, 0.02035916709899902, 0.020444160461425782, 0.020580352783203124, 0.020535295486450195, 0.020553728103637696, 0.020706304550170897, 0.020562944412231447, 0.020577280044555665, 0.020402175903320312, 0.02033459281921387, 0.020555776596069338, 0.020366336822509764, 0.02033459281921387, 0.020519935607910156, 0.020619264602661135, 0.020478975296020507, 0.020486143112182616, 0.020596736907958983, 0.020510719299316405, 0.020576255798339844, 0.020539392471313478, 0.020502527236938475, 0.02052716827392578, 0.02056595230102539, 0.020568063735961914, 0.020594688415527345, 0.020583423614501953, 0.020546560287475587, 0.020617216110229493, 0.020567039489746093, 0.02046976089477539, 0.020561983108520508, 0.020559808731079102, 0.020592639923095703, 0.020546560287475587, 0.0206059513092041, 0.020554752349853517]",tokens/s,46.66535996681432,,,main,False,False float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1069.146112,6302.466048,0.0,5662.3104,5460.681728,s,11,0.5416408309936523,0.049240075544877485,0.0002039274825441834,0.049216064453125,0.04947452926635742,0.049598928451538085,0.04969844779968262,"[0.04972332763671875, 0.04910361480712891, 0.049191200256347656, 0.049352672576904295, 0.04921939086914062, 0.04907404708862305, 0.04901577758789062, 0.04947452926635742, 0.049216064453125, 0.04900406265258789, 0.04926614379882813]",tokens/s,5199.017206354227,kWh,6.003607692721196e-07,3.289697738260924e-07,4.197974311563383e-06,5.127304854661595e-06,tokens/kWh,49928765.16153557,MB,1069.473792,6304.5632,0.0,5664.407552,5560.067072,s,11,10.435399841308593,0.9486727128462357,0.009900882562527741,0.94589892578125,0.9587894287109375,0.9673311767578125,0.9741645751953125,"[0.9758729248046875, 0.9587894287109375, 0.94589892578125, 0.9464432373046875, 0.9484384155273438, 0.9483360595703125, 0.9409358520507812, 0.9458143310546875, 0.9432595825195312, 0.9397677612304688, 0.9418433227539063]",tokens/s,66.40857183610305,kWh,1.1131132538212399e-05,6.099426448319144e-06,4.237011849843808e-05,5.9600677484969626e-05,tokens/kWh,1057034.964340592,,s,692,10.911567865371714,0.015768161655161424,0.005921252649960189,0.014926848411560058,0.01554616289138794,0.015649655580520627,0.06447429748535156,"[0.015466496467590332, 0.015452159881591796, 0.015409152030944824, 0.01564367961883545, 0.01561302375793457, 0.015385472297668457, 0.015537152290344238, 0.015535103797912597, 0.01549721622467041, 0.015544320106506348, 0.015419391632080079, 0.015547391891479492, 0.015483903884887695, 0.01570918369293213, 0.015637503623962403, 0.015462400436401368, 0.016269311904907227, 0.016055295944213867, 0.015590399742126464, 0.015546367645263673, 0.015410176277160645, 0.01600921630859375, 0.016192512512207033, 0.016316415786743164, 0.015418368339538574, 0.015455231666564942, 0.015405055999755859, 0.015399935722351075, 0.015276032447814942, 0.01509887981414795, 0.015958016395568847, 0.016149503707885742, 0.016210943222045898, 0.01581772804260254, 0.015572992324829102, 0.015656959533691405, 0.015594495773315429, 0.015569919586181641, 0.015515647888183593, 0.015493120193481445, 0.015482912063598633, 0.015476703643798828, 0.015505439758300781, 0.015480799674987793, 0.015508480072021484, 0.015531007766723632, 0.015484928131103515, 0.015503359794616698, 0.015458304405212403, 0.015500288009643554, 0.015516672134399414, 0.014855168342590331, 0.014912511825561523, 0.014935040473937988, 0.014874624252319337, 0.014935040473937988, 0.014898176193237305, 0.014906368255615234, 0.014920703887939453, 0.01488486385345459, 0.014918656349182128, 0.014973952293395995, 0.06447615814208985, 0.01490944004058838, 0.014971903800964356, 0.014898176193237305, 0.014882816314697265, 0.014928895950317383, 0.01484390354156494, 0.01491763210296631, 0.014879743576049804, 0.01491152000427246, 0.014884832382202149, 0.014937088012695313, 0.014904319763183594, 0.01539583969116211, 0.015571968078613281, 0.014867456436157226, 0.014892031669616699, 0.01569587230682373, 0.01560371208190918, 0.01556275177001953, 0.015575039863586425, 0.01557094383239746, 0.015542271614074708, 0.015524864196777344, 0.015542271614074708, 0.015494144439697266, 0.015512576103210449, 0.015536128044128418, 0.015515647888183593, 0.015550463676452637, 0.015569919586181641, 0.015571968078613281, 0.015553536415100098, 0.015637503623962403, 0.015519743919372558, 0.015511551856994628, 0.01496780776977539, 0.01489408016204834, 0.01490227222442627, 0.014896160125732421, 0.01495036792755127, 0.014946304321289062, 0.01488486385345459, 0.01489305591583252, 0.01498624038696289, 0.014883839607238769, 0.014896127700805664, 0.014916607856750488, 0.01491763210296631, 0.01491763210296631, 0.014923775672912597, 0.014921728134155274, 0.014943231582641601, 0.014945280075073243, 0.015452159881591796, 0.01568569564819336, 0.015549375534057617, 0.01548902416229248, 0.015498240470886231, 0.015519743919372558, 0.015463423728942872, 0.015507488250732422, 0.015511520385742187, 0.06459801483154297, 0.014866432189941407, 0.014921728134155274, 0.01486847972869873, 0.014851072311401366, 0.0148787202835083, 0.01490944004058838, 0.014738431930541992, 0.015633440017700194, 0.015939552307128905, 0.01549516773223877, 0.015382528305053711, 0.015470656394958496, 0.015563712120056153, 0.015571968078613281, 0.015552512168884277, 0.01552998447418213, 0.015518719673156739, 0.01491763210296631, 0.014932991981506348, 0.014873600006103516, 0.014953472137451173, 0.014953472137451173, 0.014960639953613282, 0.014924799919128418, 0.014916607856750488, 0.014910495758056641, 0.014919648170471192, 0.01491763210296631, 0.014857215881347656, 0.014937088012695313, 0.01475276756286621, 0.01467193603515625, 0.01479468822479248, 0.014816255569458007, 0.014759936332702638, 0.014935040473937988, 0.01499238395690918, 0.014921728134155274, 0.01488691234588623, 0.014950495719909668, 0.014933919906616211, 0.014950400352478027, 0.014950400352478027, 0.014941184043884278, 0.01496780776977539, 0.014929951667785645, 0.014911487579345703, 0.01494217586517334, 0.014914560317993163, 0.014871552467346191, 0.014899200439453125, 0.014942208290100097, 0.014912511825561523, 0.014934016227722167, 0.014888992309570312, 0.014936032295227051, 0.014898176193237305, 0.014971903800964356, 0.015010815620422363, 0.014898176193237305, 0.014870528221130372, 0.014892031669616699, 0.06479666900634766, 0.014880767822265625, 0.01489305591583252, 0.014887935638427734, 0.014871552467346191, 0.014877696037292481, 0.014881792068481446, 0.014942208290100097, 0.014839808464050292, 0.014928895950317383, 0.01490227222442627, 0.014948351860046387, 0.014832639694213867, 0.014957568168640138, 0.014861311912536621, 0.014897151947021485, 0.014883839607238769, 0.014913536071777344, 0.014921728134155274, 0.01488486385345459, 0.014884896278381347, 0.014939104080200195, 0.014862336158752442, 0.014904319763183594, 0.014907391548156738, 0.014926848411560058, 0.014865407943725586, 0.014946335792541504, 0.014877663612365723, 0.014916607856750488, 0.014857215881347656, 0.014990336418151855, 0.014881792068481446, 0.014904319763183594, 0.014939135551452636, 0.014912511825561523, 0.014926848411560058, 0.014905344009399414, 0.015431679725646973, 0.015286272048950195, 0.015348735809326173, 0.015511551856994628, 0.015484928131103515, 0.014880767822265625, 0.014945280075073243, 0.014945280075073243, 0.014900223731994629, 0.01487564754486084, 0.014929920196533204, 0.015058943748474121, 0.016536575317382812, 0.01590169620513916, 0.015571968078613281, 0.015552512168884277, 0.014918656349182128, 0.015006719589233398, 0.014899200439453125, 0.014918656349182128, 0.014906368255615234, 0.014958592414855957, 0.014963711738586426, 0.015045632362365722, 0.01489305591583252, 0.06461644744873046, 0.01490944004058838, 0.014944255828857422, 0.01489305591583252, 0.01485209560394287, 0.01490944004058838, 0.01485209560394287, 0.014946304321289062, 0.014940159797668457, 0.015078399658203125, 0.014919679641723632, 0.01488588809967041, 0.014927871704101562, 0.015002623558044433, 0.014949376106262208, 0.014926848411560058, 0.015009792327880859, 0.014929920196533204, 0.014906368255615234, 0.014954527854919434, 0.014886879920959472, 0.014934016227722167, 0.014963711738586426, 0.01490944004058838, 0.014904319763183594, 0.014926912307739258, 0.0148919677734375, 0.014925824165344239, 0.015003647804260254, 0.014911487579345703, 0.014914560317993163, 0.015885343551635744, 0.015710176467895506, 0.015551487922668457, 0.015532032012939453, 0.015730751991271972, 0.015538175582885743, 0.01551968002319336, 0.015539199829101562, 0.015569919586181641, 0.015536128044128418, 0.01553929615020752, 0.01546127986907959, 0.015491071701049805, 0.015462431907653808, 0.01535587215423584, 0.015316991806030274, 0.014738431930541992, 0.014819328308105468, 0.014774271965026856, 0.01472000026702881, 0.01488588809967041, 0.014773311614990235, 0.014855104446411133, 0.014779392242431641, 0.014765055656433105, 0.014751744270324708, 0.014681183815002441, 0.014832544326782226, 0.014737407684326171, 0.014706687927246094, 0.01490124797821045, 0.01472000026702881, 0.064395263671875, 0.014733311653137206, 0.014753791809082031, 0.01484390354156494, 0.014733311653137206, 0.014731264114379883, 0.014766079902648926, 0.014715904235839844, 0.014674943923950195, 0.015159296035766602, 0.015340543746948243, 0.015343647956848144, 0.01535689640045166, 0.014739456176757813, 0.01470361614227295, 0.014748671531677245, 0.014736384391784667, 0.014743552207946778, 0.014749695777893066, 0.014721023559570312, 0.014706687927246094, 0.014750720024108887, 0.014914560317993163, 0.01475487995147705, 0.014751680374145508, 0.014948351860046387, 0.014908415794372559, 0.014971903800964356, 0.014979071617126465, 0.01491763210296631, 0.015027199745178223, 0.014943231582641601, 0.014940159797668457, 0.014909472465515137, 0.014960607528686524, 0.01491763210296631, 0.014880767822265625, 0.014926848411560058, 0.014947327613830566, 0.014925824165344239, 0.014922752380371093, 0.01489408016204834, 0.014962688446044922, 0.014944255828857422, 0.01493712043762207, 0.014937055587768555, 0.014938112258911132, 0.014920703887939453, 0.015631360054016115, 0.01577881622314453, 0.015610879898071289, 0.015682559967041015, 0.01562009620666504, 0.015463423728942872, 0.015535103797912597, 0.015501312255859375, 0.015506431579589844, 0.015511551856994628, 0.015514623641967774, 0.015455231666564942, 0.01547264003753662, 0.015518719673156739, 0.015492095947265624, 0.06447411346435547, 0.01490944004058838, 0.014888959884643555, 0.014881792068481446, 0.01489510440826416, 0.014937088012695313, 0.014931967735290527, 0.014869503974914551, 0.014914560317993163, 0.014933024406433106, 0.014920672416687012, 0.014958592414855957, 0.014944255828857422, 0.014915583610534668, 0.014934016227722167, 0.014908415794372559, 0.014919679641723632, 0.01578495979309082, 0.014949407577514649, 0.014887904167175293, 0.014928895950317383, 0.014895135879516602, 0.014882783889770507, 0.014935040473937988, 0.014922752380371093, 0.0148787202835083, 0.014945280075073243, 0.014916607856750488, 0.014943231582641601, 0.014979071617126465, 0.014854144096374512, 0.015034367561340332, 0.014910464286804198, 0.014867456436157226, 0.014892031669616699, 0.014922752380371093, 0.014870528221130372, 0.014866496086120605, 0.014942144393920899, 0.0149749755859375, 0.014896127700805664, 0.014906368255615234, 0.014871552467346191, 0.01488588809967041, 0.014901311874389648, 0.014875583648681641, 0.014907391548156738, 0.014914560317993163, 0.014921728134155274, 0.0149616641998291, 0.014934016227722167, 0.014889984130859376, 0.014844927787780762, 0.014989312171936036, 0.01489510440826416, 0.014850048065185547, 0.014922752380371093, 0.014891008377075195, 0.014979071617126465, 0.014915583610534668, 0.01488588809967041, 0.014823488235473633, 0.015007679939270019, 0.06488166046142578, 0.014866432189941407, 0.014928895950317383, 0.014915583610534668, 0.014842880249023438, 0.014896127700805664, 0.01491763210296631, 0.014919679641723632, 0.014911487579345703, 0.01486847972869873, 0.014931967735290527, 0.014904319763183594, 0.014881855964660644, 0.014996416091918945, 0.01489408016204834, 0.014889984130859376, 0.014892031669616699, 0.015239168167114257, 0.015006719589233398, 0.01488486385345459, 0.014914560317993163, 0.01487564754486084, 0.014862336158752442, 0.014881792068481446, 0.01490329647064209, 0.014953472137451173, 0.0149616641998291, 0.01499443244934082, 0.014874624252319337, 0.014939135551452636, 0.014826496124267579, 0.014888959884643555, 0.014913536071777344, 0.01489510440826416, 0.014921759605407715, 0.014885855674743651, 0.014938112258911132, 0.014913599967956543, 0.015020000457763671, 0.015045599937438965, 0.015779840469360353, 0.015703040122985838, 0.015554559707641602, 0.015608832359313965, 0.015474687576293946, 0.015490048408508301, 0.015453184127807617, 0.014861311912536621, 0.014911487579345703, 0.014962688446044922, 0.01490227222442627, 0.014925824165344239, 0.014941184043884278, 0.014951423645019531, 0.014910464286804198, 0.01497599983215332, 0.015137791633605957, 0.015069184303283691, 0.014981120109558106, 0.014958592414855957, 0.015031295776367188, 0.014938112258911132, 0.014900223731994629, 0.06468300628662109, 0.01485209560394287, 0.01490227222442627, 0.01488588809967041, 0.014870528221130372, 0.014918656349182128, 0.01489510440826416, 0.014941247940063476, 0.014903231620788574, 0.014904319763183594, 0.014950400352478027, 0.014951423645019531, 0.014906368255615234, 0.015022080421447754, 0.014942272186279297, 0.01490937614440918, 0.01496678352355957, 0.015029248237609863, 0.014957568168640138, 0.014919679641723632, 0.014940159797668457, 0.014946304321289062, 0.014925824165344239, 0.014952447891235352, 0.014958592414855957, 0.014940159797668457, 0.014943231582641601, 0.014871552467346191, 0.014995455741882324, 0.015989760398864745, 0.015640576362609862, 0.0155863676071167, 0.015521727561950683, 0.014954496383666992, 0.014952447891235352, 0.014937088012695313, 0.014951423645019531, 0.014905344009399414, 0.0149934720993042, 0.014972928047180176, 0.01499232006072998, 0.014906368255615234, 0.01506611156463623, 0.014930944442749023, 0.014904319763183594, 0.014918656349182128, 0.014872575759887695, 0.014934016227722167, 0.014810111999511719, 0.014921728134155274, 0.01498624038696289, 0.0147640323638916, 0.01476095962524414, 0.014840928077697754, 0.014763936042785645, 0.01469542407989502, 0.014821375846862793, 0.014828543663024902, 0.01490329647064209, 0.014899231910705567, 0.014949343681335448, 0.014919743537902833, 0.014947263717651368, 0.06439936065673828, 0.014844927787780762, 0.014942208290100097, 0.01490944004058838, 0.014800895690917968, 0.015002623558044433, 0.01486847972869873, 0.01477017593383789, 0.014844063758850098, 0.014779232025146484, 0.014717951774597168, 0.014718976020812988, 0.01470361614227295, 0.014924799919128418, 0.015164416313171386, 0.014927871704101562, 0.014970879554748535, 0.01498419189453125, 0.01490944004058838, 0.014908415794372559, 0.014924799919128418, 0.014919679641723632, 0.015027199745178223, 0.014910464286804198, 0.014964735984802247, 0.01495961570739746, 0.014914655685424804, 0.014933919906616211, 0.01497599983215332, 0.014932991981506348, 0.014882816314697265, 0.01489510440826416, 0.01489408016204834, 0.01489408016204834, 0.014948351860046387, 0.014960639953613282, 0.014888992309570312, 0.014881759643554688, 0.01488691234588623, 0.015017984390258789, 0.014880767822265625, 0.014923775672912597, 0.014989312171936036, 0.014939135551452636, 0.014932991981506348, 0.0148602876663208, 0.01491155242919922, 0.0148919677734375, 0.01500057601928711, 0.014928895950317383, 0.014914560317993163, 0.014861311912536621, 0.014876704216003418, 0.014899168014526367, 0.01489510440826416, 0.014907391548156738, 0.014872575759887695, 0.014927871704101562, 0.01496678352355957, 0.014921728134155274, 0.014941184043884278, 0.014993408203125, 0.014877696037292481, 0.06474547576904296, 0.01487667179107666, 0.014892031669616699, 0.014938112258911132, 0.01490124797821045, 0.014965760231018066, 0.014887935638427734, 0.0149616641998291, 0.014910464286804198, 0.014864383697509765, 0.014977024078369141, 0.014883839607238769, 0.014953472137451173, 0.0148787202835083, 0.014889984130859376, 0.014887935638427734, 0.014926912307739258, 0.014844863891601563, 0.015262720108032226, 0.015459327697753907, 0.01539583969116211, 0.015394816398620606, 0.015446016311645508, 0.014866432189941407, 0.014922752380371093, 0.01488588809967041, 0.014871552467346191, 0.014867456436157226, 0.014971903800964356, 0.014883839607238769, 0.014888959884643555, 0.014944255828857422, 0.014881792068481446, 0.0148787202835083, 0.014930944442749023, 0.014898176193237305, 0.014898176193237305, 0.014884896278381347, 0.014864352226257324, 0.014872575759887695, 0.01490944004058838, 0.01486847972869873, 0.014908415794372559, 0.014893088340759277, 0.014870495796203613, 0.01489510440826416, 0.014923775672912597, 0.01487564754486084, 0.014930944442749023, 0.01488588809967041, 0.014926848411560058, 0.014889984130859376, 0.014872575759887695, 0.01490124797821045, 0.014919679641723632, 0.014920703887939453, 0.014881792068481446, 0.014985216140747071, 0.014936063766479492, 0.014874624252319337, 0.01490124797821045, 0.014910495758056641, 0.014888928413391114]",tokens/s,63.4189337900825,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15938,7 +15938,7 @@ OSError: r is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -16012,7 +16012,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 316.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -16528,7 +16528,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1149.001728,2605.187072,0.0,1965.031424,1890.996224,s,10,0.20988812828063966,0.020988812828063967,0.0005355747849219581,0.020890799522399903,0.021313901138305665,0.021906086540222167,0.022379834861755373,"[0.022498271942138673, 0.02067964744567871, 0.020683040618896486, 0.02118230438232422, 0.020906368255615235, 0.020609184265136717, 0.020998016357421875, 0.020876991271972657, 0.020904607772827148, 0.02054969596862793]",tokens/s,12196.973792519819,kWh,2.475870372725344e-07,1.356627377859443e-07,1.5029781415848897e-06,1.8862279166433682e-06,tokens/kWh,135720608.17314383,MB,1149.001728,2605.187072,0.0,1965.031424,1948.84096,s,10,11.4994541015625,1.14994541015625,0.02259531597582969,1.1540154418945314,1.1748104248046876,1.1748933715820313,1.1749597290039064,"[1.132986083984375, 1.1180478515625, 1.1256864013671875, 1.174976318359375, 1.1742025146484374, 1.1569588623046876, 1.1747919921875, 1.1707353515625, 1.151072021484375, 1.1199967041015626]",tokens/s,54.78520931827521,kWh,1.357429260460475e-05,7.438304037412697e-06,2.742707278021723e-05,4.843966942223466e-05,tokens/kWh,1300586.9105102087,,s,629,11.666427900314329,0.01854758012768574,0.0025782765884541027,0.018424800872802734,0.01868211135864258,0.018801049423217774,0.03874668548583985,"[0.019333120346069335, 0.01822719955444336, 0.018052095413208007, 0.01785036849975586, 0.017912832260131836, 0.01824563217163086, 0.01822003173828125, 0.017979391098022462, 0.018020351409912108, 0.01784115219116211, 0.017582080841064454, 0.018381824493408205, 0.017802240371704102, 0.01782374382019043, 0.017821695327758787, 0.017728511810302734, 0.017769472122192383, 0.01779302406311035, 0.017754112243652344, 0.01775103950500488, 0.0178288631439209, 0.017802240371704102, 0.018108415603637695, 0.01839308738708496, 0.018181119918823242, 0.0177838077545166, 0.017814559936523436, 0.017944543838500977, 0.01792617607116699, 0.017824735641479492, 0.01819545555114746, 0.018593791961669923, 0.01862860870361328, 0.01864396858215332, 0.017701887130737306, 0.01794867134094238, 0.017785856246948242, 0.017640447616577147, 0.017756160736083985, 0.017952768325805665, 0.018299903869628906, 0.01778892707824707, 0.0178155517578125, 0.01779814338684082, 0.01785651206970215, 0.01778892707824707, 0.0176312313079834, 0.01764352035522461, 0.018555967330932618, 0.018244543075561524, 0.017858591079711914, 0.017780704498291014, 0.01780940818786621, 0.01779916763305664, 0.017780736923217775, 0.017581056594848633, 0.01767731285095215, 0.017687551498413084, 0.017617919921875, 0.01763839912414551, 0.01765376091003418, 0.017624063491821287, 0.038637569427490234, 0.01761894416809082, 0.017738752365112305, 0.017760320663452147, 0.018018239974975585, 0.017758207321166994, 0.01780531120300293, 0.017740800857543947, 0.01776742362976074, 0.017753087997436523, 0.01776131248474121, 0.0177161922454834, 0.01757798385620117, 0.017582080841064454, 0.01756979179382324, 0.01768448066711426, 0.017797119140625, 0.017819648742675782, 0.017902591705322265, 0.017725439071655275, 0.017622016906738282, 0.017764352798461915, 0.01778892707824707, 0.01756159973144531, 0.01760153579711914, 0.017518592834472657, 0.01761177635192871, 0.01767430305480957, 0.017612735748291017, 0.017827840805053712, 0.01821388816833496, 0.01779199981689453, 0.017655807495117186, 0.017730560302734375, 0.01763430404663086, 0.017555456161499023, 0.01758515167236328, 0.017778688430786133, 0.017779712677001954, 0.017760255813598632, 0.017878015518188475, 0.018019327163696287, 0.01798041534423828, 0.01784832000732422, 0.017753087997436523, 0.017728511810302734, 0.017722368240356445, 0.017688575744628905, 0.017781759262084963, 0.017738752365112305, 0.01775103950500488, 0.017705984115600586, 0.017722368240356445, 0.01775103950500488, 0.017740800857543947, 0.017727487564086913, 0.017756160736083985, 0.01780636787414551, 0.01770800018310547, 0.017709056854248048, 0.017884159088134767, 0.017832000732421874, 0.01772640037536621, 0.03880038452148438, 0.01774284744262695, 0.01780940818786621, 0.01776639938354492, 0.01764454460144043, 0.017582080841064454, 0.01765990447998047, 0.017566719055175782, 0.017531904220581054, 0.01756979179382324, 0.017554431915283202, 0.017549312591552735, 0.01776639938354492, 0.017694719314575197, 0.017758207321166994, 0.017738752365112305, 0.017740800857543947, 0.017714208602905273, 0.0177161922454834, 0.017741823196411134, 0.017597471237182617, 0.017669088363647462, 0.0177807674407959, 0.017611743927001954, 0.01764249610900879, 0.017719295501708983, 0.017574911117553712, 0.01759846305847168, 0.017722368240356445, 0.017743871688842772, 0.017709056854248048, 0.017756160736083985, 0.017721343994140625, 0.017657855987548828, 0.01784832000732422, 0.017743871688842772, 0.017755136489868165, 0.01762918472290039, 0.017661951065063478, 0.017686527252197267, 0.017732608795166017, 0.017581056594848633, 0.017687551498413084, 0.017762304306030274, 0.01760153579711914, 0.01761587142944336, 0.017802240371704102, 0.01785958480834961, 0.018259967803955078, 0.018498559951782227, 0.018420736312866212, 0.01839411163330078, 0.018342912673950194, 0.018408447265625, 0.018513919830322266, 0.018503679275512695, 0.01840127944946289, 0.018415615081787108, 0.018391040802001952, 0.01841459274291992, 0.018415615081787108, 0.01843404769897461, 0.01845248031616211, 0.04163174438476563, 0.019153919219970703, 0.01859071922302246, 0.018542591094970702, 0.01859686470031738, 0.018619392395019533, 0.018637823104858398, 0.018589696884155273, 0.018534400939941405, 0.018463743209838866, 0.018759679794311524, 0.01862348747253418, 0.018619392395019533, 0.01861427116394043, 0.01860710334777832, 0.01860505676269531, 0.01863167953491211, 0.01862348747253418, 0.018504703521728515, 0.018364416122436524, 0.018339839935302735, 0.01841868782043457, 0.018386943817138672, 0.018563072204589845, 0.01875660705566406, 0.01859584045410156, 0.018408447265625, 0.018340864181518556, 0.01860915184020996, 0.018661376953125, 0.01863987159729004, 0.0185743350982666, 0.01846784019470215, 0.018307071685791015, 0.019081247329711913, 0.019641311645507812, 0.01926246452331543, 0.01869004821777344, 0.01863065528869629, 0.018507776260375978, 0.01863680076599121, 0.01880473518371582, 0.01883033561706543, 0.018593791961669923, 0.01878118324279785, 0.018426944732666015, 0.018834367752075195, 0.01860505676269531, 0.01864806365966797, 0.018687007904052734, 0.01850262451171875, 0.018469888687133788, 0.018565120697021483, 0.018506752014160157, 0.01887539291381836, 0.018353151321411132, 0.018382848739624022, 0.01844428825378418, 0.018586624145507814, 0.018646015167236327, 0.01864499282836914, 0.01856716728210449, 0.01858252716064453, 0.03990118408203125, 0.018689023971557618, 0.018391040802001952, 0.018456575393676757, 0.018364416122436524, 0.01858252716064453, 0.01844223976135254, 0.018463743209838866, 0.01861734390258789, 0.018555904388427736, 0.01858252716064453, 0.01862041664123535, 0.018593791961669923, 0.018715648651123046, 0.018874368667602538, 0.018565120697021483, 0.01862553596496582, 0.018663423538208008, 0.018584640502929687, 0.018612159729003906, 0.0186562557220459, 0.01880985641479492, 0.0186429443359375, 0.018539520263671876, 0.018581504821777343, 0.018557952880859374, 0.01856518363952637, 0.01857836723327637, 0.01840332794189453, 0.0184597110748291, 0.018613183975219726, 0.018578432083129884, 0.01918569564819336, 0.019470304489135743, 0.01905971145629883, 0.018675712585449217, 0.01864396858215332, 0.018569215774536133, 0.01860812759399414, 0.018555904388427736, 0.01862758445739746, 0.01862246322631836, 0.018585599899291993, 0.018675712585449217, 0.01862246322631836, 0.018536447525024414, 0.018603008270263673, 0.0186746883392334, 0.018584575653076172, 0.01866035270690918, 0.01864396858215332, 0.018626560211181642, 0.01863065528869629, 0.018704383850097657, 0.018695167541503906, 0.018736127853393555, 0.01862041664123535, 0.01860095977783203, 0.018619392395019533, 0.018626560211181642, 0.018601984024047852, 0.01860710334777832, 0.018592767715454102, 0.03876761627197266, 0.01776742362976074, 0.017737728118896484, 0.017914880752563478, 0.01778892707824707, 0.0177541446685791, 0.017742816925048827, 0.017729536056518554, 0.017701887130737306, 0.017730592727661133, 0.017787872314453126, 0.017812480926513673, 0.01783193588256836, 0.017712127685546874, 0.01775103950500488, 0.017693695068359376, 0.017688575744628905, 0.017854463577270507, 0.017709056854248048, 0.017733631134033204, 0.017722368240356445, 0.01789952087402344, 0.018052095413208007, 0.018482175827026368, 0.018684928894042968, 0.018456575393676757, 0.018522111892700196, 0.018542591094970702, 0.018708480834960937, 0.018362367630004883, 0.018495487213134765, 0.018905088424682616, 0.01884364891052246, 0.01866444778442383, 0.01839107131958008, 0.018424800872802734, 0.0186060791015625, 0.018566144943237304, 0.01942835235595703, 0.02079539108276367, 0.019021823883056642, 0.018715648651123046, 0.0186296329498291, 0.018586624145507814, 0.018601984024047852, 0.018651136398315428, 0.018679807662963867, 0.018680864334106446, 0.01868796730041504, 0.018542591094970702, 0.01862246322631836, 0.01862451171875, 0.018569215774536133, 0.01861840057373047, 0.018598880767822266, 0.018554880142211915, 0.0186296329498291, 0.01868191909790039, 0.018625471115112306, 0.018704383850097657, 0.018687999725341797, 0.018564096450805666, 0.018556928634643553, 0.0400711669921875, 0.018571264266967775, 0.018587648391723634, 0.018572288513183592, 0.018428928375244142, 0.018332672119140626, 0.01866649627685547, 0.01860710334777832, 0.01863680076599121, 0.018708480834960937, 0.01861734390258789, 0.01908531188964844, 0.01901260757446289, 0.018676736831665038, 0.01863372802734375, 0.018671615600585938, 0.01861529541015625, 0.018715648651123046, 0.01864089584350586, 0.018570240020751954, 0.018610176086425782, 0.01864396858215332, 0.018682880401611326, 0.01864192008972168, 0.018651136398315428, 0.0186746883392334, 0.01865011215209961, 0.018670591354370117, 0.01861734390258789, 0.018594816207885743, 0.018662399291992187, 0.018581504821777343, 0.018634752273559572, 0.018699264526367186, 0.018626560211181642, 0.01860406494140625, 0.018647008895874024, 0.01859686470031738, 0.018637823104858398, 0.01860812759399414, 0.018795520782470702, 0.018687999725341797, 0.018673728942871094, 0.018618303298950194, 0.018572288513183592, 0.018663423538208008, 0.01857535934448242, 0.018826271057128908, 0.018756576538085937, 0.01859891128540039, 0.018700288772583007, 0.01866035270690918, 0.01861734390258789, 0.01856924819946289, 0.01868182373046875, 0.018526208877563476, 0.018509824752807616, 0.018593791961669923, 0.018587648391723634, 0.018561023712158203, 0.018670591354370117, 0.0186060791015625, 0.01861324882507324, 0.039984127044677735, 0.018713600158691408, 0.01862041664123535, 0.0183767032623291, 0.018524160385131837, 0.018594816207885743, 0.01860095977783203, 0.018676736831665038, 0.01866963195800781, 0.018686912536621095, 0.018390016555786134, 0.01855897521972656, 0.01864089584350586, 0.01864192008972168, 0.01879347229003906, 0.01859071922302246, 0.018584575653076172, 0.01865318489074707, 0.01863167953491211, 0.018603008270263673, 0.01867366409301758, 0.01859071922302246, 0.018578432083129884, 0.018663423538208008, 0.01860095977783203, 0.01854060745239258, 0.018655168533325196, 0.018570240020751954, 0.01843404769897461, 0.018512895584106445, 0.018512895584106445, 0.018358272552490236, 0.01847091293334961, 0.018565120697021483, 0.01859891128540039, 0.018535423278808593, 0.018537471771240235, 0.01882521629333496, 0.018682880401611326, 0.018626560211181642, 0.018692096710205077, 0.01851910400390625, 0.018366399765014647, 0.018589759826660158, 0.018634687423706053, 0.01860505676269531, 0.018725887298583984, 0.018481184005737304, 0.018608095169067383, 0.01856716728210449, 0.018620479583740236, 0.018594751358032225, 0.01855897521972656, 0.018555904388427736, 0.018577407836914063, 0.018571264266967775, 0.01863167953491211, 0.01864089584350586, 0.018587648391723634, 0.018375680923461913, 0.01840230369567871, 0.018354175567626953, 0.018328575134277342, 0.03990220642089844, 0.018370559692382812, 0.018415615081787108, 0.018601984024047852, 0.018585599899291993, 0.018576383590698242, 0.018538496017456055, 0.018528255462646484, 0.018397184371948243, 0.018325504302978517, 0.01838387107849121, 0.01856211280822754, 0.01857529640197754, 0.018578432083129884, 0.018413568496704103, 0.018363391876220703, 0.018530399322509765, 0.018537376403808595, 0.01856716728210449, 0.018585599899291993, 0.018489343643188477, 0.018374656677246092, 0.018332672119140626, 0.018350080490112306, 0.018700288772583007, 0.018562047958374024, 0.018525184631347655, 0.018594816207885743, 0.018594816207885743, 0.01861529541015625, 0.018557952880859374, 0.018594816207885743, 0.018573312759399413, 0.0186060791015625, 0.018559999465942383, 0.018603008270263673, 0.01859993553161621, 0.0186060791015625, 0.01860095977783203, 0.01859174346923828, 0.018563072204589845, 0.018676736831665038, 0.017737728118896484, 0.017745920181274414, 0.0178657283782959, 0.01777663993835449, 0.01773465538024902, 0.017696767807006835, 0.01761894416809082, 0.017663999557495116, 0.017772544860839845, 0.017740800857543947, 0.017760255813598632, 0.017765439987182618, 0.01773356819152832, 0.01776639938354492, 0.017708032608032227, 0.017718271255493166, 0.017731584548950196, 0.017696767807006835, 0.017729536056518554, 0.017686527252197267, 0.017754112243652344, 0.03869286346435547, 0.017753087997436523, 0.017721343994140625, 0.017739776611328126, 0.017718271255493166, 0.017733663558959962, 0.017747936248779298, 0.017736703872680663, 0.01775923156738281, 0.017720319747924804, 0.017721343994140625, 0.017748992919921876, 0.0178606071472168, 0.017722368240356445, 0.017768447875976562, 0.017736703872680663, 0.017787904739379884, 0.017738752365112305, 0.017804288864135744, 0.018058240890502928, 0.0177838077545166, 0.017785856246948242, 0.01783302307128906, 0.017777599334716798, 0.01784012794494629, 0.017761280059814453, 0.0178155517578125, 0.017747968673706056, 0.017820735931396485, 0.01770284843444824, 0.017772544860839845, 0.01770086479187012, 0.017787904739379884, 0.017833984375, 0.01824870491027832, 0.017516544342041016, 0.017665023803710937, 0.017746944427490235, 0.017778688430786133, 0.017886207580566405, 0.01787494468688965, 0.017752063751220702, 0.017787904739379884, 0.017755136489868165, 0.017726463317871095, 0.017755136489868165, 0.01779199981689453, 0.017703935623168944, 0.017765375137329103, 0.017661951065063478, 0.01773472023010254, 0.018082752227783203, 0.017748992919921876, 0.017746944427490235, 0.017741823196411134, 0.017689599990844726, 0.017728511810302734, 0.017745920181274414, 0.017687551498413084, 0.017747968673706056, 0.017731584548950196, 0.017731584548950196, 0.01779814338684082]",tokens/s,53.915389129782625,,,main,False,False -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -16824,7 +16824,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 64.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -16926,7 +16926,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 172.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1099.210752,1418.19904,0.0,778.043392,601.166848,s,19,0.18128963279724122,0.009541559620907433,0.00031779269862037553,0.009458847999572754,0.009708159828186035,0.010222595596313475,0.010578356246948242,"[0.010667296409606933, 0.010173184394836426, 0.009462528228759766, 0.009500479698181153, 0.009489824295043945, 0.009458847999572754, 0.009479328155517578, 0.00946134376525879, 0.009448384284973145, 0.009422528266906737, 0.009321439743041992, 0.009472767829895019, 0.00940332794189453, 0.009591903686523438, 0.009431520462036132, 0.009341440200805663, 0.009337087631225587, 0.00940169620513916, 0.009424703598022461]",tokens/s,26829.99532267803,kWh,1.1473024290099243e-07,6.284304177807445e-08,5.747377844931974e-07,7.523110691722643e-07,tokens/kWh,340284771.1408339,MB,1099.538432,1418.19904,0.0,778.043392,620.356608,s,19,9.940461730957031,0.5231821963661595,0.006936860479069573,0.5211837768554688,0.5285371215820313,0.5336169982910156,0.5456140246582031,"[0.54861328125, 0.5200792236328124, 0.5259476318359375, 0.5217031860351562, 0.5217098388671875, 0.5319507446289062, 0.5205490112304687, 0.5212012939453124, 0.52119287109375, 0.521158447265625, 0.5182402954101563, 0.5211837768554688, 0.5231117553710938, 0.5208383178710938, 0.5200266723632813, 0.515190673828125, 0.5189518432617187, 0.521129150390625, 0.5276837158203125]",tokens/s,120.41694162679075,kWh,6.329118204940518e-06,3.4681341634669864e-06,1.1839526532875298e-05,2.1636778901282805e-05,tokens/kWh,2911708.8216982633,,s,1196,10.10165556907654,0.008446200308592423,0.0012225374068946526,0.008260607719421387,0.008543744087219238,0.008691968202590942,0.018165042972564698,"[0.009374719619750976, 0.009341952323913574, 0.00898252773284912, 0.008909824371337891, 0.008951807975769043, 0.008835071563720704, 0.008820735931396484, 0.008653887748718262, 0.00871724796295166, 0.008668160438537598, 0.008626175880432128, 0.008643584251403809, 0.008753151893615722, 0.008596511840820312, 0.00855446434020996, 0.008706175804138183, 0.008697728157043456, 0.008782848358154297, 0.008606719970703124, 0.008599552154541015, 0.008691712379455567, 0.008648703575134278, 0.008502304077148437, 0.008571871757507324, 0.008642560005187988, 0.00858726406097412, 0.008644607543945313, 0.008657919883728027, 0.008777728080749512, 0.008855551719665527, 0.008734720230102539, 0.008647680282592773, 0.008728575706481934, 0.008791040420532227, 0.008677375793457032, 0.00858521556854248, 0.008523776054382324, 0.008739935874938964, 0.008729503631591798, 0.008763392448425293, 0.008556544303894043, 0.0087326717376709, 0.008560640335083008, 0.008565759658813477, 0.00858521556854248, 0.008911871910095214, 0.00882380771636963, 0.008761343955993652, 0.008822848320007324, 0.008705056190490723, 0.008563615798950196, 0.008794112205505371, 0.00852684783935547, 0.008408063888549805, 0.008756223678588868, 0.008516608238220215, 0.00838144016265869, 0.008282112121582032, 0.008230912208557128, 0.008299584388732911, 0.008713184356689453, 0.008756192207336425, 0.019679231643676756, 0.008623104095458984, 0.008548352241516113, 0.008539135932922364, 0.008547327995300292, 0.008492032051086425, 0.008426624298095703, 0.008375167846679688, 0.00839577579498291, 0.008140800476074218, 0.00809062385559082, 0.00813980770111084, 0.008227807998657227, 0.008119296073913575, 0.00810700798034668, 0.008113151550292968, 0.0081397762298584, 0.008121343612670898, 0.008101920127868652, 0.008088543891906738, 0.008086527824401855, 0.00819814395904541, 0.00820531177520752, 0.008143872261047362, 0.008115391731262207, 0.008131391525268554, 0.008151040077209473, 0.008119296073913575, 0.00808448028564453, 0.008117376327514648, 0.008102848052978515, 0.008093631744384766, 0.0081080322265625, 0.008170495986938477, 0.008227840423583984, 0.008317952156066894, 0.008291328430175781, 0.008214624404907226, 0.008226719856262207, 0.008170495986938477, 0.008251551628112794, 0.008176480293273926, 0.008205344200134277, 0.008396767616271972, 0.008272895812988281, 0.008242207527160645, 0.008279007911682129, 0.00828825569152832, 0.008260607719421387, 0.008252479553222657, 0.008307647705078126, 0.008246272087097169, 0.00828825569152832, 0.008223744392395019, 0.008246272087097169, 0.008268799781799317, 0.008292351722717285, 0.00822169589996338, 0.008245247840881348, 0.008274944305419921, 0.008256511688232422, 0.008257535934448243, 0.00832102394104004, 0.01823027229309082, 0.008279040336608886, 0.008247296333312988, 0.008256511688232422, 0.00828006362915039, 0.008269824028015137, 0.008269824028015137, 0.00892416000366211, 0.008607744216918945, 0.009148447990417481, 0.008775648117065429, 0.008606719970703124, 0.008601632118225098, 0.008605664253234863, 0.008607744216918945, 0.00860262393951416, 0.008560640335083008, 0.008464384078979491, 0.008464384078979491, 0.00827187156677246, 0.008309760093688966, 0.008245311737060546, 0.008244159698486329, 0.008264703750610352, 0.008203264236450195, 0.008270848274230956, 0.008250368118286134, 0.008244223594665527, 0.008243200302124023, 0.008232959747314453, 0.008248319625854492, 0.008265727996826172, 0.008259584426879883, 0.008257535934448243, 0.008257535934448243, 0.008247296333312988, 0.008263680458068847, 0.008260607719421387, 0.00829030418395996, 0.008268799781799317, 0.008265727996826172, 0.008272895812988281, 0.008247296333312988, 0.008276991844177246, 0.008282112121582032, 0.008230912208557128, 0.008289279937744141, 0.008313856124877929, 0.008282112121582032, 0.008260607719421387, 0.00830463981628418, 0.008284159660339355, 0.008274944305419921, 0.008253439903259278, 0.008228863716125488, 0.008272895812988281, 0.008279040336608886, 0.008253439903259278, 0.008260607719421387, 0.008266752243041992, 0.008334336280822753, 0.008247296333312988, 0.008259584426879883, 0.0182794246673584, 0.00820633602142334, 0.008225791931152344, 0.008299519538879394, 0.008212639808654785, 0.008256352424621582, 0.008260607719421387, 0.008264703750610352, 0.008232959747314453, 0.008239168167114258, 0.008250304222106934, 0.008249343872070313, 0.008294400215148925, 0.008276991844177246, 0.008249343872070313, 0.008491007804870606, 0.008254464149475099, 0.00830463981628418, 0.00821350383758545, 0.008239104270935058, 0.008261631965637208, 0.008259584426879883, 0.008202239990234375, 0.008267775535583496, 0.008233983993530274, 0.008242176055908204, 0.0082227201461792, 0.008220671653747558, 0.008255488395690918, 0.008260607719421387, 0.008225791931152344, 0.00824118423461914, 0.008240096092224121, 0.008202239990234375, 0.008259584426879883, 0.008227840423583984, 0.008187040328979493, 0.008275808334350586, 0.008254464149475099, 0.008259584426879883, 0.008279040336608886, 0.008245247840881348, 0.008272895812988281, 0.008253439903259278, 0.008224767684936523, 0.008265727996826172, 0.008223744392395019, 0.008734784126281737, 0.008686528205871583, 0.008285280227661132, 0.00826255989074707, 0.008541184425354004, 0.008240127563476562, 0.008256511688232422, 0.008248384475708009, 0.008268735885620116, 0.008258560180664062, 0.008247296333312988, 0.008283136367797851, 0.008293375968933106, 0.008229887962341309, 0.008283136367797851, 0.008249343872070313, 0.01817087936401367, 0.008275967597961426, 0.008310784339904785, 0.008278016090393067, 0.008229887962341309, 0.008243200302124023, 0.008264703750610352, 0.008226816177368163, 0.00819814395904541, 0.008283136367797851, 0.008235008239746093, 0.008256511688232422, 0.008253472328186035, 0.008289248466491698, 0.008261631965637208, 0.008209407806396484, 0.008203264236450195, 0.00830463981628418, 0.008283136367797851, 0.00830463981628418, 0.008259584426879883, 0.008258560180664062, 0.008285183906555176, 0.008246272087097169, 0.008251392364501953, 0.008258560180664062, 0.008282112121582032, 0.008279040336608886, 0.008324095726013184, 0.008264703750610352, 0.008340479850769043, 0.008230912208557128, 0.008253439903259278, 0.008284159660339355, 0.008294400215148925, 0.008235008239746093, 0.008648703575134278, 0.008472576141357421, 0.008267775535583496, 0.008245247840881348, 0.00824732780456543, 0.008243167877197265, 0.008246272087097169, 0.008273920059204102, 0.008250368118286134, 0.008254464149475099, 0.008242176055908204, 0.008235008239746093, 0.008251392364501953, 0.008281087875366211, 0.008275967597961426, 0.008253439903259278, 0.008253439903259278, 0.008263680458068847, 0.008281087875366211, 0.008279040336608886, 0.008341504096984862, 0.008265727996826172, 0.008293375968933106, 0.008285183906555176, 0.00828006362915039, 0.008281120300292969, 0.00827184009552002, 0.01822822380065918, 0.008253439903259278, 0.008268799781799317, 0.008255488395690918, 0.008249343872070313, 0.008263680458068847, 0.008273920059204102, 0.008833024024963379, 0.008607744216918945, 0.008302623748779298, 0.009087967872619628, 0.008692735671997071, 0.008630271911621093, 0.008647680282592773, 0.008583168029785156, 0.00857907199859619, 0.008657919883728027, 0.00859545612335205, 0.008541184425354004, 0.008582143783569337, 0.008301568031311036, 0.008240127563476562, 0.008259584426879883, 0.008261631965637208, 0.008287232398986816, 0.008243200302124023, 0.008281120300292969, 0.00828003215789795, 0.008276991844177246, 0.008253439903259278, 0.008279040336608886, 0.008301568031311036, 0.008278016090393067, 0.00841113567352295, 0.008590335845947266, 0.008523776054382324, 0.008557567596435547, 0.008532992362976074, 0.00856169605255127, 0.008530912399291992, 0.008527872085571288, 0.008546303749084473, 0.008477696418762207, 0.008524800300598144, 0.008511487960815429, 0.008522751808166504, 0.00858726406097412, 0.008498175621032715, 0.008551424026489257, 0.008538111686706543, 0.008540160179138183, 0.008570879936218261, 0.008535103797912598, 0.008517567634582519, 0.00852070426940918, 0.008295424461364746, 0.008268799781799317, 0.008265727996826172, 0.008264703750610352, 0.008258560180664062, 0.008170495986938477, 0.008286208152770995, 0.008284159660339355, 0.01825382423400879, 0.00827187156677246, 0.008236031532287597, 0.008315903663635254, 0.008229887962341309, 0.008308735847473145, 0.008258560180664062, 0.00818380832672119, 0.008256511688232422, 0.008237055778503418, 0.008228863716125488, 0.008302592277526855, 0.008220671653747558, 0.008227840423583984, 0.008262656211853027, 0.00822169589996338, 0.008216575622558593, 0.008283136367797851, 0.008227840423583984, 0.008270848274230956, 0.008218624114990235, 0.0082227201461792, 0.00830668830871582, 0.00829030418395996, 0.008225791931152344, 0.008224767684936523, 0.008252415657043457, 0.008263680458068847, 0.008240127563476562, 0.008244223594665527, 0.008218624114990235, 0.008226816177368163, 0.008229887962341309, 0.008391679763793946, 0.008232959747314453, 0.008275967597961426, 0.008237055778503418, 0.008248319625854492, 0.008253439903259278, 0.008256511688232422, 0.008274944305419921, 0.008299551963806152, 0.008275936126708985, 0.008269824028015137, 0.00827187156677246, 0.008254464149475099, 0.008332287788391114, 0.008243200302124023, 0.008244223594665527, 0.008263680458068847, 0.00829849624633789, 0.008225791931152344, 0.008239104270935058, 0.008235008239746093, 0.008260640144348144, 0.00825443172454834, 0.008270848274230956, 0.008255488395690918, 0.008254464149475099, 0.008239104270935058, 0.008226816177368163, 0.008227840423583984, 0.00826576042175293, 0.018193376541137694, 0.00828825569152832, 0.008286208152770995, 0.008251392364501953, 0.008233983993530274, 0.008267775535583496, 0.008228863716125488, 0.008241151809692383, 0.008246272087097169, 0.008246272087097169, 0.00822169589996338, 0.008268832206726074, 0.008218591690063476, 0.00830668830871582, 0.008258560180664062, 0.008254464149475099, 0.00830668830871582, 0.008252415657043457, 0.008232959747314453, 0.008278016090393067, 0.008299519538879394, 0.00831283187866211, 0.008294400215148925, 0.008301568031311036, 0.008254464149475099, 0.008244223594665527, 0.008294400215148925, 0.008240127563476562, 0.008287232398986816, 0.008254464149475099, 0.008260607719421387, 0.008243200302124023, 0.008245247840881348, 0.008272895812988281, 0.008315903663635254, 0.008256511688232422, 0.008284159660339355, 0.008240127563476562, 0.008261631965637208, 0.008267775535583496, 0.008342528343200683, 0.008256511688232422, 0.008248319625854492, 0.008238080024719239, 0.008262656211853027, 0.008236031532287597, 0.008259584426879883, 0.008233983993530274, 0.008241151809692383, 0.008266752243041992, 0.008250368118286134, 0.008266752243041992, 0.00831283187866211, 0.008244223594665527, 0.00828825569152832, 0.008256511688232422, 0.00831283187866211, 0.008262656211853027, 0.008262656211853027, 0.008285183906555176, 0.008299519538879394, 0.008274944305419921, 0.008281087875366211, 0.01820569610595703, 0.008229887962341309, 0.00823094367980957, 0.008260576248168946, 0.008353792190551757, 0.008250368118286134, 0.008275967597961426, 0.008233983993530274, 0.008249343872070313, 0.008237055778503418, 0.008249343872070313, 0.008247296333312988, 0.008253439903259278, 0.008258560180664062, 0.008302592277526855, 0.008341504096984862, 0.008279040336608886, 0.008278016090393067, 0.008294400215148925, 0.008275967597961426, 0.008252415657043457, 0.008259584426879883, 0.008219648361206054, 0.008256511688232422, 0.008255488395690918, 0.008236031532287597, 0.008247296333312988, 0.008257535934448243, 0.008263680458068847, 0.008250368118286134, 0.008255488395690918, 0.008258560180664062, 0.008245280265808105, 0.008268832206726074, 0.00829638385772705, 0.008232959747314453, 0.008276991844177246, 0.008228863716125488, 0.008301568031311036, 0.008267775535583496, 0.008273920059204102, 0.008217599868774414, 0.008269824028015137, 0.008226816177368163, 0.008269824028015137, 0.008263680458068847, 0.008262751579284668, 0.008260512351989746, 0.008248319625854492, 0.008247296333312988, 0.008258560180664062, 0.008254464149475099, 0.008276000022888183, 0.008285152435302734, 0.008275967597961426, 0.008274944305419921, 0.008281087875366211, 0.008282112121582032, 0.008400896072387695, 0.008282112121582032, 0.008293375968933106, 0.00827187156677246, 0.008256511688232422, 0.018164735794067383, 0.008260607719421387, 0.008272895812988281, 0.008287232398986816, 0.008270848274230956, 0.008302592277526855, 0.008247360229492188, 0.008172479629516602, 0.008267775535583496, 0.008261631965637208, 0.008254464149475099, 0.008275967597961426, 0.008264703750610352, 0.008273920059204102, 0.008242176055908204, 0.008243200302124023, 0.008266752243041992, 0.008264703750610352, 0.00829036808013916, 0.008282048225402832, 0.00830668830871582, 0.008285183906555176, 0.008259584426879883, 0.008289279937744141, 0.008336383819580078, 0.008287232398986816, 0.00829644775390625, 0.008286208152770995, 0.008261631965637208, 0.008265727996826172, 0.008287232398986816, 0.008233983993530274, 0.00827187156677246, 0.008266752243041992, 0.008247296333312988, 0.008273920059204102, 0.008267775535583496, 0.008241151809692383, 0.008266752243041992, 0.008245247840881348, 0.008253439903259278, 0.008235008239746093, 0.008260607719421387, 0.008245247840881348, 0.008230912208557128, 0.008262656211853027, 0.008231936454772949, 0.008236031532287597, 0.008246272087097169, 0.008274944305419921, 0.008273920059204102, 0.008291328430175781, 0.008248319625854492, 0.008237055778503418, 0.008273920059204102, 0.008253439903259278, 0.008262656211853027, 0.008274944305419921, 0.008230912208557128, 0.00818380832672119, 0.008248319625854492, 0.008333312034606934, 0.008364031791687012, 0.018010112762451173, 0.008252415657043457, 0.008248319625854492, 0.008126463890075684, 0.008143872261047362, 0.008130720138549805, 0.008145759582519532, 0.008138751983642578, 0.008130559921264649, 0.008114175796508789, 0.008159232139587403, 0.008241151809692383, 0.008241151809692383, 0.008268799781799317, 0.008259584426879883, 0.008299519538879394, 0.008261631965637208, 0.008254464149475099, 0.008275967597961426, 0.008261631965637208, 0.008248319625854492, 0.008251392364501953, 0.00828006362915039, 0.008257535934448243, 0.008263680458068847, 0.008229887962341309, 0.008257535934448243, 0.008238080024719239, 0.008243200302124023, 0.008230912208557128, 0.008254464149475099, 0.008236031532287597, 0.008243200302124023, 0.008351743698120117, 0.008211520195007324, 0.00822265625, 0.008246272087097169, 0.008143872261047362, 0.008269824028015137, 0.008196224212646484, 0.008152095794677735, 0.008130399703979493, 0.008174592018127442, 0.008231936454772949, 0.00811520004272461, 0.008135680198669434, 0.008121343612670898, 0.008111104011535645, 0.008113151550292968, 0.008143903732299805, 0.008210399627685546, 0.008335359573364258, 0.008289279937744141, 0.008229887962341309, 0.008262656211853027, 0.008245247840881348, 0.008256511688232422, 0.008264703750610352, 0.008208383560180664, 0.008260607719421387, 0.008257535934448243, 0.008262656211853027, 0.008251392364501953, 0.01820876884460449, 0.008242176055908204, 0.008248319625854492, 0.008261664390563965, 0.00826364803314209, 0.00828006362915039, 0.008257535934448243, 0.008263680458068847, 0.008263680458068847, 0.008246272087097169, 0.008267775535583496, 0.008261631965637208, 0.008264703750610352, 0.008250368118286134, 0.008220671653747558, 0.008250368118286134, 0.008259584426879883, 0.008293375968933106, 0.00827187156677246, 0.00824732780456543, 0.008262623786926269, 0.008279040336608886, 0.008256511688232422, 0.008223744392395019, 0.008279040336608886, 0.008262656211853027, 0.008246272087097169, 0.008211456298828124, 0.008282112121582032, 0.008328191757202149, 0.008286208152770995, 0.008257535934448243, 0.008263680458068847, 0.008257535934448243, 0.008260607719421387, 0.008247296333312988, 0.008276991844177246, 0.008243200302124023, 0.008254464149475099, 0.008244223594665527, 0.008267775535583496, 0.008250368118286134, 0.00827187156677246, 0.00827187156677246, 0.008227871894836427, 0.008282079696655273, 0.008252415657043457, 0.008317952156066894, 0.00827187156677246, 0.008285183906555176, 0.008283167839050294, 0.008286175727844238, 0.008297472000122071, 0.008247296333312988, 0.008263680458068847, 0.008294400215148925, 0.008332287788391114, 0.008286208152770995, 0.008245247840881348, 0.008226816177368163, 0.008258560180664062, 0.00830361557006836, 0.008311807632446289, 0.018150400161743165, 0.008276991844177246, 0.008250368118286134, 0.008244223594665527, 0.008231936454772949, 0.008278016090393067, 0.00830463981628418, 0.008255488395690918, 0.008275967597961426, 0.008246272087097169, 0.008251392364501953, 0.008256511688232422, 0.008247296333312988, 0.008232959747314453, 0.008241151809692383, 0.008236031532287597, 0.008262656211853027, 0.008223744392395019, 0.008266752243041992, 0.008220671653747558, 0.008392704010009766, 0.008273920059204102, 0.009038847923278808, 0.008653823852539062, 0.008942591667175292, 0.008707072257995606, 0.008226816177368163, 0.008275967597961426, 0.008268799781799317, 0.008255488395690918, 0.008282112121582032, 0.008281087875366211, 0.008262656211853027, 0.00825654411315918, 0.008259552001953125, 0.008263680458068847, 0.008278016090393067, 0.008258560180664062, 0.008203264236450195, 0.008269856452941895, 0.008262623786926269, 0.008243200302124023, 0.008261631965637208, 0.008275967597961426, 0.008255488395690918, 0.00828825569152832, 0.008344575881958008, 0.008249343872070313, 0.008250368118286134, 0.008258560180664062, 0.008242176055908204, 0.008231936454772949, 0.008238080024719239, 0.008239104270935058, 0.008244223594665527, 0.008287232398986816, 0.008236031532287597, 0.00828006362915039, 0.008251392364501953, 0.008250368118286134, 0.008258560180664062, 0.008235008239746093, 0.008243200302124023, 0.018347007751464844, 0.008209407806396484, 0.008228863716125488, 0.008250399589538574, 0.008270815849304199, 0.008247360229492188, 0.008241087913513183, 0.008245247840881348, 0.008245247840881348, 0.008250368118286134, 0.008251392364501953, 0.008256511688232422, 0.008257535934448243, 0.008261631965637208, 0.008254464149475099, 0.00831283187866211, 0.008270848274230956, 0.008245247840881348, 0.008252415657043457, 0.008345600128173827, 0.008294400215148925, 0.008231936454772949, 0.008241151809692383, 0.008269824028015137, 0.008252415657043457, 0.008218624114990235, 0.008262656211853027, 0.008224767684936523, 0.008268799781799317, 0.008240127563476562, 0.008239104270935058, 0.008248319625854492, 0.008243200302124023, 0.008249343872070313, 0.008241151809692383, 0.008238080024719239, 0.008258560180664062, 0.008310784339904785, 0.00827187156677246, 0.008261631965637208, 0.008246272087097169, 0.008256511688232422, 0.008317952156066894, 0.008241151809692383, 0.008267775535583496, 0.008247296333312988, 0.008281087875366211, 0.008220671653747558, 0.008229887962341309, 0.008263680458068847, 0.008291328430175781, 0.008250368118286134, 0.008269824028015137, 0.008261631965637208, 0.008251392364501953, 0.008286208152770995, 0.00830463981628418, 0.008269856452941895, 0.008304608345031739, 0.008263680458068847, 0.008243200302124023, 0.0083374080657959, 0.008251392364501953, 0.018182144165039063, 0.008268799781799317, 0.008261631965637208, 0.008259584426879883, 0.008261631965637208, 0.008248319625854492, 0.008278016090393067, 0.008243200302124023, 0.008264703750610352, 0.008235008239746093, 0.008235008239746093, 0.008265727996826172, 0.008253439903259278, 0.008250368118286134, 0.008236031532287597, 0.008249343872070313, 0.008252415657043457, 0.008251392364501953, 0.008241151809692383, 0.008301568031311036, 0.00827187156677246, 0.00822169589996338, 0.0083056640625, 0.008263680458068847, 0.008261631965637208, 0.008260640144348144, 0.008270815849304199, 0.008263680458068847, 0.00827187156677246, 0.008251456260681153, 0.008375231742858886, 0.008242176055908204, 0.008281087875366211, 0.00827187156677246, 0.008270848274230956, 0.008238080024719239, 0.008334336280822753, 0.008263680458068847, 0.008259584426879883, 0.008246272087097169, 0.008259584426879883, 0.008241151809692383, 0.008274944305419921, 0.008270848274230956, 0.008179712295532226, 0.0082227201461792, 0.008263680458068847, 0.008264703750610352, 0.008261631965637208, 0.008284223556518555, 0.008256447792053223, 0.008124608039855957, 0.008112095832824706, 0.008235872268676758, 0.008159232139587403, 0.008217599868774414, 0.008226816177368163, 0.008240127563476562, 0.008168448448181152, 0.008169471740722656, 0.008161279678344726, 0.008228863716125488, 0.008196096420288086, 0.017939456939697264, 0.008122367858886719, 0.008211456298828124, 0.008127488136291505, 0.008119359970092773, 0.008130496025085449, 0.008157183647155761, 0.008208383560180664, 0.008144895553588867, 0.008172575950622558, 0.00817353630065918, 0.008156160354614257, 0.00820736026763916, 0.008217599868774414, 0.008183967590332032, 0.008142687797546387, 0.008202239990234375, 0.008152064323425292, 0.00819711971282959, 0.008274944305419921, 0.008174592018127442, 0.008112128257751466, 0.00818073558807373, 0.00819916820526123, 0.00823516845703125, 0.008149951934814452, 0.008150943756103516, 0.008125439643859863, 0.008324095726013184, 0.008185855865478516, 0.008195072174072266, 0.008164352416992187, 0.008177663803100586, 0.008135680198669434, 0.008136799812316894, 0.008122400283813476, 0.008171520233154296, 0.0081396484375, 0.0081397762298584, 0.008121439933776856, 0.008140704154968263, 0.008135680198669434, 0.008162303924560547, 0.008113216400146485, 0.008119232177734375, 0.008119359970092773, 0.008203200340270996, 0.008232959747314453, 0.008279040336608886, 0.00831488037109375, 0.008162303924560547, 0.008119296073913575, 0.008132608413696289, 0.008130559921264649, 0.008120320320129394, 0.008141823768615723, 0.008146944046020508, 0.008129535675048828, 0.008233983993530274, 0.008136896133422852, 0.00821331214904785, 0.008194047927856446, 0.008301631927490234, 0.01793324851989746, 0.008103936195373536, 0.008190976142883302, 0.008193023681640625, 0.008126463890075684, 0.008236031532287597, 0.008135807991027833, 0.008161151885986328, 0.008132608413696289, 0.008160256385803222, 0.008151040077209473, 0.008136704444885253, 0.0081080322265625, 0.008133631706237793, 0.00812339210510254, 0.00819711971282959, 0.008269824028015137, 0.00824630355834961, 0.008289248466491698, 0.00830463981628418, 0.008291328430175781, 0.008240127563476562, 0.008193023681640625, 0.008219648361206054, 0.008166399955749512, 0.008212479591369629, 0.008227840423583984, 0.008219648361206054, 0.008252415657043457, 0.008233983993530274, 0.008275967597961426, 0.008282112121582032, 0.00840499210357666, 0.008261631965637208, 0.008258560180664062, 0.008220671653747558, 0.00829849624633789, 0.008245247840881348, 0.008240256309509277, 0.008218496322631837, 0.008245247840881348, 0.008250368118286134, 0.008265727996826172, 0.008242176055908204, 0.008274944305419921, 0.008273920059204102, 0.008227840423583984, 0.008287263870239257, 0.008283103942871094, 0.008279040336608886, 0.008276991844177246, 0.008240192413330079, 0.00831174373626709, 0.008235008239746093, 0.008254464149475099, 0.008281087875366211, 0.008260671615600585, 0.008253376007080078, 0.008255488395690918, 0.008173567771911621, 0.008254528045654297, 0.008232895851135253, 0.008352767944335938, 0.018137088775634767, 0.008227840423583984, 0.008237055778503418, 0.008284159660339355, 0.008263680458068847, 0.008300607681274413, 0.008325056076049804, 0.008283167839050294, 0.008232928276062012, 0.008283136367797851, 0.008275967597961426, 0.008245247840881348, 0.008259584426879883, 0.008240127563476562, 0.008249343872070313, 0.008235008239746093, 0.008270879745483399, 0.008240096092224121, 0.008264703750610352, 0.008230912208557128, 0.008284159660339355, 0.008253439903259278, 0.008239104270935058, 0.008243200302124023, 0.008256511688232422, 0.008263680458068847, 0.008289279937744141, 0.008407039642333984, 0.008258560180664062, 0.008266752243041992, 0.008258560180664062, 0.00821350383758545, 0.008254464149475099, 0.008285183906555176, 0.008257535934448243, 0.008248319625854492, 0.008238080024719239, 0.008255488395690918, 0.008238080024719239, 0.008251392364501953, 0.008287232398986816, 0.008246272087097169, 0.008261631965637208, 0.008270879745483399, 0.008243167877197265, 0.008255488395690918, 0.008223744392395019, 0.00821555233001709, 0.008285183906555176, 0.008273920059204102, 0.008279040336608886, 0.008268799781799317, 0.00827187156677246, 0.008263680458068847, 0.008249343872070313, 0.008263680458068847, 0.008266752243041992, 0.008262656211853027, 0.008340479850769043, 0.008260607719421387, 0.008266783714294434, 0.008297439575195312, 0.008256511688232422, 0.018292736053466797, 0.008232959747314453, 0.008257535934448243, 0.008282112121582032, 0.00824015998840332, 0.008266719818115234, 0.008231936454772949, 0.008250368118286134, 0.008265727996826172, 0.008268799781799317, 0.008261631965637208, 0.008258560180664062, 0.008260640144348144, 0.008283103942871094, 0.008264703750610352, 0.008270848274230956, 0.008270848274230956, 0.008254464149475099, 0.008262656211853027, 0.008273920059204102, 0.008258560180664062, 0.008285216331481934, 0.008334303855895996, 0.008269824028015137, 0.008264703750610352, 0.008238080024719239, 0.008263680458068847, 0.008260671615600585, 0.008264639854431152, 0.008259584426879883, 0.008281087875366211, 0.008258560180664062, 0.008289312362670899, 0.008259552001953125, 0.008261631965637208, 0.008243200302124023, 0.008263680458068847, 0.008391679763793946, 0.00827187156677246, 0.00821452808380127, 0.008239104270935058, 0.008252415657043457, 0.008282112121582032, 0.008272895812988281, 0.00822169589996338, 0.008888319969177246, 0.009343999862670899, 0.008645631790161134, 0.00860371208190918, 0.008592320442199706, 0.00861184024810791, 0.008594431877136231, 0.008583168029785156, 0.00859340763092041, 0.008556544303894043, 0.008529919624328614, 0.008550399780273438, 0.008525856018066406, 0.008541152000427247, 0.00851865577697754, 0.00852070426940918, 0.008537088394165039, 0.008536064147949218]",tokens/s,118.39643430936485,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -17353,7 +17353,7 @@ OSError: 0 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,eager,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -17586,7 +17586,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -17643,7 +17643,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, float32-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1140.084736,7590.117376,0.0,6949.961728,6804.719616,s,10,0.5006790390014648,0.050067903900146496,0.00021801737872100758,0.050056127548217776,0.050273246383666995,0.05040283184051514,0.050506500205993655,"[0.05053241729736328, 0.04974985504150391, 0.05010054397583008, 0.050112926483154296, 0.04992403030395508, 0.05001171112060547, 0.05024444961547852, 0.04996432113647461, 0.05022534561157226, 0.04981343841552734]",tokens/s,5113.056071022197,kWh,6.257487843136168e-07,3.4287514085701243e-07,4.363943820370283e-06,5.332567745540912e-06,tokens/kWh,48006891.27935918,MB,1140.084736,7590.117376,0.0,6949.961728,6904.4864,s,10,11.425099365234374,1.1425099365234375,0.010515235760784074,1.139412780761719,1.1487585693359375,1.1606250122070312,1.1701181665039062,"[1.172491455078125, 1.1392100830078125, 1.14612158203125, 1.1408779296875, 1.1381038818359375, 1.13300341796875, 1.1394222412109376, 1.140939453125, 1.1394033203125, 1.1355260009765624]",tokens/s,55.14175236996516,kWh,1.3590661998288481e-05,7.447281955385169e-06,5.087762815062976e-05,7.191557210430343e-05,tokens/kWh,876027.2380038548,,s,629,11.85613618469239,0.0188491831235173,0.0059958730071987595,0.01808076858520508,0.01849671630859375,0.018830361557006837,0.0684296826171875,"[0.019128320693969726, 0.018716672897338867, 0.018827327728271483, 0.019075008392333986, 0.018792448043823243, 0.019017791748046874, 0.01915078353881836, 0.018762752532958983, 0.019168256759643554, 0.018784255981445314, 0.018940927505493164, 0.01922870445251465, 0.01889686393737793, 0.019079168319702147, 0.01863372802734375, 0.01882624053955078, 0.018778240203857422, 0.018760576248168945, 0.01909760093688965, 0.019019775390625, 0.018817024230957033, 0.018926591873168946, 0.018637823104858398, 0.018710527420043945, 0.018832384109497072, 0.019763200759887696, 0.019982336044311523, 0.020032512664794923, 0.019074047088623047, 0.018647039413452148, 0.018619392395019533, 0.018579456329345705, 0.018534400939941405, 0.018496511459350586, 0.01822412872314453, 0.01807360076904297, 0.01818726348876953, 0.018151487350463867, 0.018004928588867188, 0.018173952102661133, 0.018177024841308592, 0.018158592224121094, 0.01805414390563965, 0.018174976348876954, 0.018119680404663087, 0.01805721664428711, 0.018755584716796874, 0.01863577651977539, 0.018514944076538087, 0.01878835105895996, 0.018344959259033202, 0.017921024322509766, 0.017903615951538086, 0.017950719833374023, 0.017906688690185548, 0.017898496627807618, 0.018148351669311523, 0.018111488342285157, 0.01798259162902832, 0.018066303253173828, 0.018177024841308592, 0.018106367111206053, 0.06830079650878906, 0.018106367111206053, 0.018067455291748045, 0.017963008880615236, 0.018077695846557617, 0.018069503784179687, 0.01807360076904297, 0.01803264045715332, 0.01860095977783203, 0.018140159606933593, 0.017985536575317384, 0.01814630317687988, 0.01815449523925781, 0.018115583419799804, 0.01819340705871582, 0.018121728897094725, 0.018110464096069336, 0.017941503524780272, 0.01788211250305176, 0.017958911895751953, 0.017770496368408203, 0.017876991271972655, 0.018058240890502928, 0.0180992317199707, 0.01808483123779297, 0.018107391357421874, 0.01794867134094238, 0.018137088775634767, 0.01806540870666504, 0.018068479537963866, 0.01798041534423828, 0.018128896713256838, 0.018043903350830077, 0.01803264045715332, 0.018027519226074217, 0.01807257652282715, 0.018084863662719726, 0.018000896453857423, 0.018060287475585936, 0.018131967544555663, 0.0180316162109375, 0.018078720092773438, 0.018043935775756834, 0.018157535552978516, 0.018124799728393554, 0.01805516815185547, 0.018113536834716795, 0.018082815170288084, 0.0179814395904541, 0.0180633602142334, 0.018091007232666014, 0.01824460792541504, 0.018130943298339842, 0.018135040283203126, 0.01816582489013672, 0.018078655242919923, 0.018099199295043944, 0.018081792831420897, 0.018086912155151368, 0.01812588882446289, 0.018087871551513673, 0.018076671600341796, 0.01803264045715332, 0.06877184295654297, 0.01818009567260742, 0.01815449523925781, 0.018140159606933593, 0.01817087936401367, 0.018125823974609375, 0.018109439849853515, 0.018101247787475586, 0.01813811111450195, 0.01815449523925781, 0.018084863662719726, 0.018122751235961913, 0.018136064529418947, 0.018203680038452148, 0.018161632537841796, 0.018106367111206053, 0.01817087936401367, 0.018110464096069336, 0.018158592224121094, 0.019107839584350587, 0.01982975959777832, 0.018743295669555664, 0.018694143295288086, 0.01846067237854004, 0.018070528030395508, 0.01810950469970703, 0.018100160598754883, 0.01823846435546875, 0.018116607666015624, 0.017953792572021485, 0.017936384201049805, 0.01789030456542969, 0.018586624145507814, 0.018497535705566406, 0.018386943817138672, 0.018117631912231445, 0.018117631912231445, 0.018074623107910158, 0.018149375915527344, 0.018030624389648437, 0.018158559799194337, 0.018096128463745118, 0.018109439849853515, 0.018089984893798827, 0.018109439849853515, 0.018053119659423827, 0.018092031478881835, 0.018106367111206053, 0.01804083251953125, 0.01808896064758301, 0.018084863662719726, 0.018103296279907227, 0.018119680404663087, 0.018060287475585936, 0.0179814395904541, 0.018111488342285157, 0.018111488342285157, 0.018145280838012694, 0.018089984893798827, 0.01802444839477539, 0.017958911895751953, 0.01807974433898926, 0.017923072814941408, 0.06848921966552735, 0.017969152450561524, 0.018156543731689453, 0.018137088775634767, 0.018114559173583983, 0.018157567977905274, 0.018149375915527344, 0.018142240524291992, 0.01813603210449219, 0.01801523208618164, 0.018109439849853515, 0.018190336227416993, 0.018131967544555663, 0.018111488342285157, 0.018132991790771484, 0.018110464096069336, 0.018100223541259765, 0.018130943298339842, 0.018118656158447266, 0.018069503784179687, 0.018155519485473632, 0.01804902458190918, 0.018118719100952148, 0.018103231430053712, 0.018107391357421874, 0.018086912155151368, 0.018100223541259765, 0.01800499153137207, 0.018089984893798827, 0.018149375915527344, 0.018139135360717772, 0.018145280838012694, 0.01820876884460449, 0.018140159606933593, 0.018197504043579102, 0.018129919052124025, 0.018044927597045898, 0.018062335968017578, 0.018112512588500978, 0.01804083251953125, 0.018036735534667968, 0.018121728897094725, 0.01802444839477539, 0.018093055725097656, 0.018089984893798827, 0.01804800033569336, 0.01816985511779785, 0.018128896713256838, 0.018111488342285157, 0.018183168411254884, 0.018137088775634767, 0.01802444839477539, 0.01807257652282715, 0.018106367111206053, 0.018100223541259765, 0.01807974433898926, 0.018100223541259765, 0.018089984893798827, 0.018069503784179687, 0.018041856765747072, 0.018118656158447266, 0.018075647354125975, 0.018131967544555663, 0.06849740600585938, 0.018106367111206053, 0.018107391357421874, 0.01798963165283203, 0.018081792831420897, 0.01807974433898926, 0.018174976348876954, 0.0180316162109375, 0.01810534477233887, 0.017978368759155275, 0.017912832260131836, 0.017935359954833984, 0.01790771293640137, 0.01802649688720703, 0.018109439849853515, 0.018089984893798827, 0.018075647354125975, 0.01807974433898926, 0.018053119659423827, 0.018087968826293946, 0.01807459259033203, 0.018111488342285157, 0.018136064529418947, 0.018098175048828127, 0.018067455291748045, 0.01820979118347168, 0.018068479537963866, 0.018091007232666014, 0.018094079971313477, 0.018070528030395508, 0.01805516815185547, 0.018127872467041017, 0.018123775482177733, 0.018084863662719726, 0.018061311721801757, 0.018157567977905274, 0.018159616470336915, 0.01818009567260742, 0.01823232078552246, 0.018027519226074217, 0.01813811111450195, 0.018119680404663087, 0.01807974433898926, 0.01807257652282715, 0.01803980827331543, 0.018052095413208007, 0.018075647354125975, 0.01807360076904297, 0.018108415603637695, 0.018044927597045898, 0.018100223541259765, 0.018062335968017578, 0.018059263229370116, 0.018107391357421874, 0.017929216384887696, 0.018058240890502928, 0.018093055725097656, 0.018063455581665038, 0.01787910461425781, 0.017816415786743166, 0.01783296012878418, 0.017787967681884766, 0.01780851173400879, 0.06847980499267578, 0.017846336364746095, 0.017819583892822264, 0.017937408447265626, 0.018001920700073244, 0.0178606071472168, 0.017881088256835938, 0.01784217643737793, 0.017906688690185548, 0.017862655639648437, 0.017905759811401366, 0.017723295211791994, 0.017756160736083985, 0.017683456420898438, 0.01784934425354004, 0.017827840805053712, 0.01780838394165039, 0.017789951324462892, 0.017821695327758787, 0.017896448135375977, 0.017829887390136717, 0.017893375396728514, 0.017925119400024413, 0.017863679885864257, 0.017886207580566405, 0.017743871688842772, 0.017869823455810546, 0.01786777687072754, 0.018125823974609375, 0.018114559173583983, 0.018094079971313477, 0.017966079711914062, 0.01798860740661621, 0.018176000595092775, 0.01807360076904297, 0.018068479537963866, 0.018081792831420897, 0.018074623107910158, 0.018139135360717772, 0.01803264045715332, 0.01813811111450195, 0.01817190361022949, 0.018089984893798827, 0.018070528030395508, 0.0180316162109375, 0.01803468894958496, 0.01805721664428711, 0.018130943298339842, 0.01807974433898926, 0.018098175048828127, 0.018096128463745118, 0.01805721664428711, 0.018028543472290038, 0.01799884796142578, 0.01802448081970215, 0.01806230354309082, 0.01804800033569336, 0.018084863662719726, 0.018067455291748045, 0.018095104217529297, 0.018132991790771484, 0.01805619239807129, 0.01804595184326172, 0.06893055725097656, 0.018086912155151368, 0.018109439849853515, 0.018062400817871093, 0.01811244773864746, 0.018067455291748045, 0.018107391357421874, 0.0180633602142334, 0.018149375915527344, 0.01965363121032715, 0.01887846374511719, 0.018144256591796876, 0.018123775482177733, 0.01803980827331543, 0.018140159606933593, 0.018091007232666014, 0.018139135360717772, 0.018156543731689453, 0.01805414390563965, 0.018132991790771484, 0.018099199295043944, 0.01807360076904297, 0.018393119812011718, 0.01804182434082031, 0.018082815170288084, 0.018066432952880858, 0.018116607666015624, 0.01801318359375, 0.018077695846557617, 0.018155519485473632, 0.018036735534667968, 0.018053184509277342, 0.018037696838378907, 0.017902591705322265, 0.01803980827331543, 0.017898496627807618, 0.01797427177429199, 0.017843231201171875, 0.017934303283691406, 0.01778892707824707, 0.017796096801757814, 0.017864704132080078, 0.018084863662719726, 0.018060287475585936, 0.018150400161743165, 0.018035711288452147, 0.01799679946899414, 0.018076671600341796, 0.018068479537963866, 0.01801625633239746, 0.018051071166992186, 0.01803059196472168, 0.017979391098022462, 0.01799884796142578, 0.018145280838012694, 0.01788313674926758, 0.017934335708618163, 0.01786777687072754, 0.018001920700073244, 0.017827840805053712, 0.017932287216186525, 0.017960960388183594, 0.018084863662719726, 0.0684933090209961, 0.0180633602142334, 0.018077695846557617, 0.018027519226074217, 0.018093055725097656, 0.01817190361022949, 0.018116607666015624, 0.018085887908935547, 0.018104320526123048, 0.01807257652282715, 0.01805516815185547, 0.018206720352172853, 0.018151424407958985, 0.017871904373168945, 0.017973215103149413, 0.017872896194458008, 0.017968128204345703, 0.01804697608947754, 0.017892351150512697, 0.017953792572021485, 0.01781452751159668, 0.01788211250305176, 0.018206720352172853, 0.017905664443969727, 0.018542591094970702, 0.018733055114746093, 0.018140159606933593, 0.018044927597045898, 0.01822105598449707, 0.018095104217529297, 0.01805516815185547, 0.018124799728393554, 0.01921331214904785, 0.018689023971557618, 0.01821388816833496, 0.01807974433898926, 0.01796505546569824, 0.017906688690185548, 0.017932287216186525, 0.018082815170288084, 0.018044927597045898, 0.01804083251953125, 0.01802137565612793, 0.018103296279907227, 0.018066432952880858, 0.018043903350830077, 0.018009088516235353, 0.018207744598388673, 0.018070528030395508, 0.01823027229309082, 0.018050048828125, 0.018051071166992186, 0.018050048828125, 0.01805414390563965, 0.018117631912231445, 0.01810534477233887, 0.018100223541259765, 0.018028543472290038, 0.018060287475585936, 0.01803775978088379, 0.018159616470336915, 0.018053119659423827, 0.018083839416503905, 0.06882508850097656, 0.01818726348876953, 0.01808896064758301, 0.01810534477233887, 0.018069503784179687, 0.018166784286499024, 0.018832384109497072, 0.01859993553161621, 0.018597888946533202, 0.018544639587402344, 0.018471935272216796, 0.018231296539306642, 0.01838591957092285, 0.01805414390563965, 0.01806540870666504, 0.018061311721801757, 0.018069503784179687, 0.018025472640991212, 0.01804607963562012, 0.01806732749938965, 0.018081792831420897, 0.018008064270019532, 0.01819647979736328, 0.018127872467041017, 0.01803980827331543, 0.018136064529418947, 0.01803878402709961, 0.018035711288452147, 0.018133056640625, 0.018062271118164063, 0.017993728637695314, 0.01803059196472168, 0.018121728897094725, 0.01807257652282715, 0.01804697608947754, 0.018069503784179687, 0.018126848220825196, 0.01802956771850586, 0.01801420783996582, 0.018051071166992186, 0.018060287475585936, 0.018081792831420897, 0.018119680404663087, 0.01801625633239746, 0.018084863662719726, 0.017904640197753906, 0.017957887649536132, 0.017886207580566405, 0.01803059196472168, 0.018050048828125, 0.01805731201171875, 0.017820575714111327, 0.017976320266723633, 0.017796096801757814, 0.0178288631439209, 0.01787494468688965, 0.017947647094726564, 0.017859647750854492, 0.017836992263793944, 0.01785139274597168, 0.017911808013916015, 0.017961984634399415, 0.01783193588256836, 0.0682425308227539, 0.017862560272216797, 0.017928192138671875, 0.01787392044067383, 0.017937408447265626, 0.01786684799194336, 0.017915807723999023, 0.01788313674926758, 0.01785139274597168, 0.017839103698730468, 0.017932287216186525, 0.017868864059448243, 0.017847328186035155, 0.017670047760009765, 0.01782067108154297, 0.017663999557495116, 0.017683456420898438, 0.01779097557067871, 0.017905664443969727, 0.01786777687072754, 0.018234560012817383, 0.017982271194458006, 0.0180316162109375, 0.01802956771850586, 0.01803878402709961, 0.018051071166992186, 0.018207744598388673, 0.018096128463745118, 0.018066432952880858, 0.018156543731689453, 0.018084863662719726, 0.018106399536132814, 0.01809814453125, 0.018070528030395508, 0.01820364761352539, 0.01842585563659668, 0.018111488342285157, 0.018124799728393554, 0.018110464096069336, 0.018114559173583983, 0.018081792831420897, 0.018051071166992186, 0.01806550407409668, 0.01803664016723633, 0.018094079971313477, 0.018122751235961913, 0.01808076858520508, 0.018070528030395508, 0.01802649688720703, 0.01805721664428711, 0.018126848220825196, 0.018035711288452147, 0.01803468894958496, 0.018144256591796876, 0.018362367630004883, 0.01802239990234375, 0.018076671600341796, 0.01804083251953125, 0.01805619239807129, 0.018027519226074217, 0.01806438446044922, 0.018020383834838866, 0.01806844711303711]",tokens/s,53.052696949627695,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -17680,7 +17680,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 1.50 GiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -18125,7 +18125,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 100.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -18152,7 +18152,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -18295,7 +18295,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -18334,7 +18334,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 44.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -18903,7 +18903,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -19168,7 +19168,7 @@ OSError: r is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,Qwen/Qwen1.5-7B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -19232,7 +19232,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -19788,7 +19788,7 @@ OSError: google/recurrentgemma-7b is not a local folder and is not a valid model If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -19815,7 +19815,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -20193,7 +20193,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -20321,7 +20321,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPT2LMHeadModel does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,01-ai/Yi-6B,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -20729,7 +20729,7 @@ OSError: 0 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +float32-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float32,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -20952,7 +20952,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -21009,7 +21009,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, bfloat16-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,MB,1263.665152,4276.617216,0.0,3632.267264,3408.132096,s,10,0.19949964714050294,0.019949964714050295,0.0002634110509256969,0.019849407196044923,0.020051040267944336,0.020388576316833495,0.020658605155944823,"[0.020726112365722655, 0.01983452796936035, 0.019976032257080077, 0.019860031127929688, 0.019847679138183593, 0.019810176849365233, 0.019833568572998048, 0.01993631935119629, 0.01985113525390625, 0.019824064254760743]",tokens/s,12832.102896889093,kWh,2.367375350109959e-07,1.297196071719301e-07,1.1859149990541087e-06,1.552372141237035e-06,tokens/kWh,164908911.465006,MB,1263.665152,4278.714368,0.0,3632.267264,3474.14784,s,10,11.340074707031249,1.134007470703125,0.017892579925606965,1.1269158325195312,1.1675471557617187,1.1677104675292969,1.1678411169433593,"[1.1675108642578125, 1.1208251953125, 1.12455859375, 1.130638671875, 1.1247318115234375, 1.1166605224609376, 1.167873779296875, 1.1390423583984375, 1.119133056640625, 1.129099853515625]",tokens/s,55.55518956232075,kWh,1.3274553809949536e-05,7.274062950731538e-06,3.679097649354576e-05,5.733959325422684e-05,tokens/kWh,1098717.246225947,,s,629,11.499658243179313,0.01828244553764598,0.0024469567160428553,0.017869823455810546,0.018637005233764648,0.01880657958984375,0.038081085357666015,"[0.01884774398803711, 0.01860812759399414, 0.0185743350982666, 0.0184453125, 0.01886617660522461, 0.01882931137084961, 0.01926348876953125, 0.01861427116394043, 0.01864192008972168, 0.018528255462646484, 0.018469888687133788, 0.018545663833618165, 0.01879859161376953, 0.018585599899291993, 0.018709503173828124, 0.018785280227661134, 0.01863580894470215, 0.018593759536743165, 0.018550783157348632, 0.018528255462646484, 0.01887846374511719, 0.018746368408203123, 0.018538496017456055, 0.018334720611572267, 0.01846067237854004, 0.018226175308227538, 0.01903411293029785, 0.022605823516845702, 0.01947648048400879, 0.01963827133178711, 0.018687999725341797, 0.018511871337890624, 0.018373632431030275, 0.01879449653625488, 0.018298879623413086, 0.018463743209838866, 0.018551807403564453, 0.018501632690429686, 0.018333696365356447, 0.018592767715454102, 0.018457599639892578, 0.018525184631347655, 0.018374656677246092, 0.018017280578613282, 0.017897472381591797, 0.01785651206970215, 0.018173952102661133, 0.018081792831420897, 0.017719295501708983, 0.017928192138671875, 0.01800294494628906, 0.018084863662719726, 0.018745344161987306, 0.018118656158447266, 0.01796505546569824, 0.017923072814941408, 0.01820262336730957, 0.0178606071472168, 0.017819648742675782, 0.017984512329101563, 0.017900543212890627, 0.01789030456542969, 0.0381952018737793, 0.017870847702026366, 0.01784524726867676, 0.017769472122192383, 0.018292736053466797, 0.018486335754394533, 0.017918912887573243, 0.01781760025024414, 0.017871904373168945, 0.017918943405151367, 0.017780736923217775, 0.017930240631103517, 0.01782374382019043, 0.01784524726867676, 0.017900543212890627, 0.01780940818786621, 0.017913856506347657, 0.01780735969543457, 0.01763020706176758, 0.017731584548950196, 0.017713151931762695, 0.01789132881164551, 0.017888256072998047, 0.017942527770996093, 0.01791391944885254, 0.01770182418823242, 0.017869823455810546, 0.017869823455810546, 0.017781759262084963, 0.017686527252197267, 0.01782476806640625, 0.017871871948242187, 0.017888256072998047, 0.017958911895751953, 0.01779199981689453, 0.01788729667663574, 0.017980352401733398, 0.017812480926513673, 0.017787904739379884, 0.017778688430786133, 0.017715200424194336, 0.017729536056518554, 0.017673248291015624, 0.017790016174316407, 0.017834911346435545, 0.017758304595947266, 0.01781443214416504, 0.017750015258789064, 0.01765888023376465, 0.017469440460205078, 0.017543167114257813, 0.017501184463500977, 0.017442815780639647, 0.017552383422851564, 0.017527807235717775, 0.017427455902099608, 0.017501184463500977, 0.01761075210571289, 0.01739571189880371, 0.01741619110107422, 0.017682464599609374, 0.017755104064941407, 0.018130943298339842, 0.038694911956787106, 0.017531904220581054, 0.01760166358947754, 0.018078592300415038, 0.018190336227416993, 0.0181146240234375, 0.018285503387451173, 0.017655807495117186, 0.017790048599243165, 0.017686431884765624, 0.017709056854248048, 0.017706144332885743, 0.017629024505615234, 0.01817308807373047, 0.01808880043029785, 0.01805721664428711, 0.017702911376953127, 0.01760870361328125, 0.017690624237060547, 0.018164735794067383, 0.018199615478515625, 0.01773049545288086, 0.017687551498413084, 0.017640512466430665, 0.017819583892822264, 0.017896448135375977, 0.017967103958129883, 0.017886207580566405, 0.01800601577758789, 0.01783705520629883, 0.01794867134094238, 0.017704959869384765, 0.0178657283782959, 0.018155519485473632, 0.01804697608947754, 0.017895423889160156, 0.017922048568725587, 0.01785958480834961, 0.017952768325805665, 0.01794047927856445, 0.017904640197753906, 0.017993728637695314, 0.017728511810302734, 0.017718271255493166, 0.017839103698730468, 0.017970176696777345, 0.017928192138671875, 0.017917951583862304, 0.01785753631591797, 0.017836032867431642, 0.01789132881164551, 0.017528831481933595, 0.017690624237060547, 0.017417375564575194, 0.017654624938964844, 0.017476608276367187, 0.017512447357177736, 0.017732608795166017, 0.017730560302734375, 0.017711103439331053, 0.017747968673706056, 0.01785036849975586, 0.017827840805053712, 0.038109184265136715, 0.017896448135375977, 0.017785856246948242, 0.017747968673706056, 0.017863679885864257, 0.01794047927856445, 0.01781657600402832, 0.0178288631439209, 0.018043903350830077, 0.01860403251647949, 0.01850992012023926, 0.018368480682373046, 0.017993663787841795, 0.017885183334350584, 0.01863987159729004, 0.01838083267211914, 0.018540512084960936, 0.018571264266967775, 0.01838595199584961, 0.01786467170715332, 0.017897472381591797, 0.017870847702026366, 0.017821855545043945, 0.01789423942565918, 0.01785753631591797, 0.01785241508483887, 0.017904640197753906, 0.017864704132080078, 0.017886207580566405, 0.017912832260131836, 0.017862655639648437, 0.01785241508483887, 0.01790771293640137, 0.0176629753112793, 0.017690624237060547, 0.01764352035522461, 0.017605632781982423, 0.017690624237060547, 0.017778688430786133, 0.017887231826782226, 0.01764352035522461, 0.017689599990844726, 0.017674240112304687, 0.017656831741333007, 0.017714176177978515, 0.01765990447998047, 0.01776639938354492, 0.017946624755859376, 0.018001920700073244, 0.01787392044067383, 0.017918975830078124, 0.0178606071472168, 0.017961984634399415, 0.017880064010620117, 0.017879039764404296, 0.01788313674926758, 0.017908735275268553, 0.017911808013916015, 0.017887231826782226, 0.017933311462402343, 0.017961984634399415, 0.01840947151184082, 0.018391040802001952, 0.03800883102416992, 0.017836032867431642, 0.017935359954833984, 0.0178657283782959, 0.0178606071472168, 0.01785958480834961, 0.017929216384887696, 0.01760972785949707, 0.018041856765747072, 0.017744895935058593, 0.017680383682250975, 0.017710079193115236, 0.017723392486572266, 0.017844224929809572, 0.017727487564086913, 0.017938432693481447, 0.017853439331054686, 0.017928192138671875, 0.017920000076293945, 0.017926143646240233, 0.01784115219116211, 0.01781350326538086, 0.017854463577270507, 0.0178288631439209, 0.01781350326538086, 0.017777664184570312, 0.017692703247070313, 0.01784419250488281, 0.017942623138427736, 0.017874847412109374, 0.017878015518188475, 0.01783705520629883, 0.017672191619873046, 0.01756979179382324, 0.017896448135375977, 0.01780940818786621, 0.017870847702026366, 0.017880064010620117, 0.017829887390136717, 0.017933311462402343, 0.017691648483276368, 0.017722400665283203, 0.01789743995666504, 0.017904640197753906, 0.017884159088134767, 0.01786675262451172, 0.017715200424194336, 0.01763942337036133, 0.017699840545654297, 0.017613824844360353, 0.0176363525390625, 0.01824051284790039, 0.019198976516723632, 0.018696191787719727, 0.018225151062011717, 0.01825798416137695, 0.017827903747558594, 0.017515392303466798, 0.01743155288696289, 0.01765376091003418, 0.017583168029785157, 0.01768953514099121, 0.017624063491821287, 0.037850112915039064, 0.017649663925170898, 0.017490943908691405, 0.01738444709777832, 0.017460224151611328, 0.017447935104370118, 0.017443840026855468, 0.01741414451599121, 0.01745305633544922, 0.01765990447998047, 0.017498111724853514, 0.017584127426147463, 0.01738764762878418, 0.017346431732177734, 0.01743052864074707, 0.017526784896850587, 0.01743769645690918, 0.017398847579956054, 0.017348575592041015, 0.017396703720092773, 0.017445087432861328, 0.017409824371337892, 0.01741414451599121, 0.01765171241760254, 0.017551359176635743, 0.01760358428955078, 0.017538047790527343, 0.017663999557495116, 0.01763532829284668, 0.01761587142944336, 0.017689599990844726, 0.01764352035522461, 0.017552383422851564, 0.017697792053222656, 0.017903615951538086, 0.018372671127319336, 0.018416576385498047, 0.018362367630004883, 0.018371583938598633, 0.018062335968017578, 0.01831532859802246, 0.01775814437866211, 0.01764249610900879, 0.017722368240356445, 0.017723392486572266, 0.017605632781982423, 0.01780121612548828, 0.017976383209228515, 0.017961919784545897, 0.017884159088134767, 0.01782476806640625, 0.017724416732788087, 0.017570816040039062, 0.017950719833374023, 0.018543615341186523, 0.01839308738708496, 0.018249727249145507, 0.01800294494628906, 0.01761177635192871, 0.017582080841064454, 0.01784012794494629, 0.01781760025024414, 0.01779622459411621, 0.0381162223815918, 0.017648767471313477, 0.017774463653564453, 0.01767945671081543, 0.017769376754760743, 0.01789132881164551, 0.01785139274597168, 0.017900543212890627, 0.017864704132080078, 0.017830911636352538, 0.017892351150512697, 0.017859615325927735, 0.01850569534301758, 0.019381248474121093, 0.01983795166015625, 0.018976768493652343, 0.01903001594543457, 0.019076095581054688, 0.01887948799133301, 0.0186296329498291, 0.018323455810546875, 0.01839206314086914, 0.018590879440307618, 0.018406240463256836, 0.019336191177368164, 0.02059878349304199, 0.018817184448242187, 0.018888544082641602, 0.018720767974853517, 0.0186296329498291, 0.01865216064453125, 0.01863372802734375, 0.01859280014038086, 0.01830089569091797, 0.01840127944946289, 0.018692096710205077, 0.018547712326049806, 0.018573312759399413, 0.018811904907226562, 0.018626560211181642, 0.0186296329498291, 0.018333696365356447, 0.018538496017456055, 0.018677759170532226, 0.018694143295288086, 0.018552831649780274, 0.018734079360961914, 0.01859174346923828, 0.01863065528869629, 0.018649120330810547, 0.01866851234436035, 0.01824665641784668, 0.018742271423339844, 0.018626560211181642, 0.018760704040527345, 0.01863680076599121, 0.018888704299926756, 0.018742271423339844, 0.01761996841430664, 0.017937408447265626, 0.018754560470581053, 0.018646015167236327, 0.018474048614501953, 0.038491073608398436, 0.01862451171875, 0.018637823104858398, 0.018675712585449217, 0.0182476806640625, 0.018325504302978517, 0.018504703521728515, 0.018465791702270508, 0.018594816207885743, 0.018499584197998048, 0.0184770565032959, 0.01846272087097168, 0.018290687561035156, 0.018323455810546875, 0.01768448066711426, 0.017693695068359376, 0.017713151931762695, 0.01822412872314453, 0.017855487823486327, 0.0178288631439209, 0.01780735969543457, 0.017899551391601563, 0.01788822364807129, 0.01756985664367676, 0.017651647567749024, 0.017674240112304687, 0.017722368240356445, 0.017649663925170898, 0.01773465538024902, 0.017724416732788087, 0.01764352035522461, 0.017772544860839845, 0.01773465538024902, 0.017633279800415038, 0.017681407928466796, 0.017698816299438477, 0.017729536056518554, 0.01789132881164551, 0.01787392044067383, 0.01785753631591797, 0.017871871948242187, 0.017898496627807618, 0.01784217643737793, 0.0178606071472168, 0.01786777687072754, 0.017723392486572266, 0.01780531120300293, 0.018555904388427736, 0.01862144088745117, 0.018695167541503906, 0.018594816207885743, 0.018564096450805666, 0.018564096450805666, 0.018296831130981444, 0.017720319747924804, 0.017872896194458008, 0.01793846321105957, 0.018650079727172853, 0.01860710334777832, 0.01869004821777344, 0.018527231216430663, 0.017903615951538086, 0.017898496627807618, 0.038161407470703124, 0.017778688430786133, 0.017705984115600586, 0.017822719573974608, 0.017829887390136717, 0.017661951065063478, 0.017723392486572266, 0.01765068817138672, 0.01819340705871582, 0.01789030456542969, 0.01787596893310547, 0.017937408447265626, 0.017811456680297853, 0.017887231826782226, 0.0178155517578125, 0.01800704002380371, 0.01803059196472168, 0.01794047927856445, 0.01775923156738281, 0.01770086479187012, 0.017928192138671875, 0.01822208023071289, 0.01841663932800293, 0.018326528549194337, 0.018387968063354493, 0.01815449523925781, 0.01760870361328125, 0.017763328552246094, 0.017912832260131836, 0.017625087738037108, 0.017707008361816406, 0.017468416213989257, 0.01760665512084961, 0.01775923156738281, 0.017729536056518554, 0.017589248657226563, 0.01760051155090332, 0.017683456420898438, 0.017667072296142578, 0.017540096282958984, 0.017803264617919923, 0.017765375137329103, 0.017762304306030274, 0.017482751846313475, 0.01758720016479492, 0.01759129524230957, 0.017686527252197267, 0.017584127426147463, 0.01745305633544922, 0.01744076728820801, 0.017519775390625, 0.01746211242675781, 0.01759040069580078, 0.017678207397460937, 0.01759436798095703, 0.01762713623046875, 0.017665023803710937, 0.017604608535766602, 0.017708032608032227, 0.01757900810241699, 0.0176629753112793, 0.01758515167236328, 0.017698816299438477, 0.03851980972290039, 0.017605632781982423, 0.017757247924804688, 0.017596351623535157, 0.017496063232421876, 0.017955839157104494, 0.018095104217529297, 0.017902591705322265, 0.017825920104980467, 0.01748467254638672, 0.0174716796875, 0.017650495529174803, 0.017705087661743165, 0.017675167083740236, 0.017752031326293945, 0.01775721549987793, 0.0176977596282959, 0.017581056594848633, 0.01759231948852539, 0.01760883140563965, 0.01782156753540039, 0.01788313674926758, 0.017892351150512697, 0.017768447875976562, 0.01786675262451172, 0.017938528060913086, 0.01778268814086914, 0.017868799209594728, 0.017764352798461915, 0.017687551498413084, 0.017722368240356445, 0.017803264617919923, 0.018184192657470705, 0.018524160385131837, 0.018824192047119142, 0.01858252716064453, 0.01865318489074707, 0.018654207229614257, 0.018593791961669923, 0.018655263900756835, 0.01783600044250488, 0.01798246383666992, 0.018033664703369142, 0.018075647354125975, 0.01785241508483887, 0.017914880752563478, 0.017885183334350584, 0.017909759521484374, 0.017753087997436523, 0.017929216384887696, 0.017915903091430666, 0.017873952865600586, 0.017821664810180663, 0.017863679885864257, 0.01786675262451172, 0.017855487823486327, 0.01789132881164551, 0.0182609920501709, 0.01783705520629883, 0.017803264617919923, 0.017864704132080078, 0.017904640197753906, 0.017830911636352538]",tokens/s,54.697277666757834,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -21046,7 +21046,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 768.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -21491,7 +21491,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 136.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -21518,7 +21518,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -21702,7 +21702,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -21729,7 +21729,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-14B,Qwen/Qwen1.5-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -22188,7 +22188,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -22481,7 +22481,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -23037,7 +23037,7 @@ OSError: google/recurrentgemma-7b is not a local folder and is not a valid model If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -23064,7 +23064,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -23404,7 +23404,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -23876,7 +23876,7 @@ OSError: 0 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +bfloat16-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,bfloat16,True,False,,sdpa,,False,,False,forward,,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch