diff --git "a/perf-df-gptq-1xA10.csv" "b/perf-df-gptq-1xA10.csv" --- "a/perf-df-gptq-1xA10.csv" +++ "b/perf-df-gptq-1xA10.csv" @@ -229,7 +229,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1452.27776,2426.929152,0.0,1780.482048,1621.410304,s,10,0.7058684158325197,0.07058684158325197,0.0002928328251085697,0.07070625686645508,0.07090748825073243,0.07092075080871582,0.07093136085510254,"[0.07084793853759766, 0.07022029113769532, 0.07070374298095704, 0.07075682830810547, 0.07070877075195313, 0.07024800109863281, 0.07009436798095703, 0.07044992065429688, 0.07093401336669922, 0.070904541015625]",tokens/s,3626.738273847639,kWh,8.283197671383412e-07,4.537640673683759e-07,3.7133809815803766e-06,4.995464816087093e-06,tokens/kWh,51246482.444555126,MB,1452.27776,2426.929152,0.0,1780.482048,1723.272192,s,10,38.030801269531246,3.8030801269531245,0.0020719084747050052,3.802459106445313,3.8038822265625,3.806487744140625,3.808572158203125,"[3.80909326171875, 3.80165771484375, 3.8017744140625, 3.80289404296875, 3.802064697265625, 3.80330322265625, 3.80299072265625, 3.802145751953125, 3.8027724609375, 3.80210498046875]",tokens/s,16.565520025073223,kWh,4.4899062949067045e-05,2.4607307695079302e-05,0.00019991478192082245,0.0002694211525649688,tokens/kWh,233834.6466126413,,s,629,38.604979228973406,0.06137516570584005,0.008407749083102415,0.06034739303588867,0.06043340682983398,0.060483585357666014,0.13084835754394533,"[0.06085631942749024, 0.06110108947753906, 0.06107440185546875, 0.06086963272094727, 0.06092902374267578, 0.060903423309326174, 0.06097305679321289, 0.06068940734863281, 0.060516353607177734, 0.06042521667480469, 0.06034124755859375, 0.06029312133789062, 0.06031155014038086, 0.06031769561767578, 0.060355583190917966, 0.060298240661621094, 0.06034534454345703, 0.06035148620605469, 0.060369918823242184, 0.060316673278808595, 0.06033817672729492, 0.060298240661621094, 0.06036377716064453, 0.06029312133789062, 0.06034534454345703, 0.06041497421264649, 0.060398593902587894, 0.06032076644897461, 0.060527614593505856, 0.06036479949951172, 0.06035148620605469, 0.06034329605102539, 0.06039961624145508, 0.06036172866821289, 0.06039654541015625, 0.06041088104248047, 0.06041190338134766, 0.06058393478393555, 0.060483585357666014, 0.06034431838989258, 0.06034841537475586, 0.06036070251464844, 0.060386302947998044, 0.0604139518737793, 0.06040681457519531, 0.06037091064453125, 0.06038323211669922, 0.06031872177124024, 0.06034739303588867, 0.0603135986328125, 0.0604068489074707, 0.060338111877441404, 0.060382209777832034, 0.06051123046875, 0.06040883255004883, 0.060437503814697265, 0.06033407974243164, 0.06041190338134766, 0.060483585357666014, 0.06047129440307617, 0.060355583190917966, 0.060355583190917966, 0.13078938293457032, 0.060290046691894535, 0.06030131149291992, 0.060345375061035156, 0.06027977752685547, 0.06031769561767578, 0.060283905029296876, 0.06029721450805664, 0.06027775955200195, 0.06029414367675781, 0.06029926300048828, 0.06034022521972656, 0.06031872177124024, 0.06030131149291992, 0.06027468872070312, 0.060336128234863284, 0.06028902435302735, 0.06032281494140625, 0.06031257629394531, 0.06031564712524414, 0.06030438232421875, 0.06032998275756836, 0.06030438232421875, 0.06031155014038086, 0.06029107284545898, 0.06038016128540039, 0.06032179260253906, 0.06031872177124024, 0.060393470764160156, 0.06030847930908203, 0.06031564712524414, 0.06033715057373047, 0.06030131149291992, 0.06035763168334961, 0.06037811279296875, 0.06036787033081055, 0.06038019180297852, 0.06038729476928711, 0.06032998275756836, 0.060353534698486325, 0.06034841537475586, 0.0603351058959961, 0.060355583190917966, 0.06032998275756836, 0.06028083038330078, 0.06039961624145508, 0.06038016128540039, 0.06034841537475586, 0.06039449691772461, 0.06032793426513672, 0.06034841537475586, 0.060371967315673826, 0.06034431838989258, 0.06035865783691406, 0.06034636688232422, 0.06037401580810547, 0.060424190521240234, 0.06042726516723633, 0.060418048858642576, 0.06036377716064453, 0.060437503814697265, 0.060439552307128906, 0.060385280609130856, 0.13130752563476564, 0.060295169830322265, 0.060316673278808595, 0.0603238410949707, 0.060283905029296876, 0.0603135986328125, 0.06028902435302735, 0.060281856536865235, 0.06031872177124024, 0.06030438232421875, 0.06033407974243164, 0.060260353088378904, 0.06028799819946289, 0.060267520904541017, 0.06033919906616211, 0.06032998275756836, 0.06028287887573242, 0.060295169830322265, 0.060303359985351565, 0.060314624786376954, 0.0603054084777832, 0.06033919906616211, 0.06035763168334961, 0.06036070251464844, 0.06036684799194336, 0.06033715057373047, 0.06031052780151367, 0.06033817672729492, 0.060349441528320315, 0.06032076644897461, 0.06037299346923828, 0.06037299346923828, 0.060331008911132813, 0.0604139518737793, 0.060386302947998044, 0.06040576171875, 0.060388351440429686, 0.060371967315673826, 0.06029926300048828, 0.060388351440429686, 0.06034739303588867, 0.060368896484375, 0.060353534698486325, 0.0603351058959961, 0.06035148620605469, 0.060365825653076174, 0.060401664733886716, 0.060434432983398435, 0.06033407974243164, 0.06032998275756836, 0.06036275100708008, 0.06036684799194336, 0.06034124755859375, 0.060365825653076174, 0.06031257629394531, 0.06035968017578125, 0.060437503814697265, 0.06032588958740234, 0.06041088104248047, 0.060349441528320315, 0.060385280609130856, 0.060316673278808595, 0.06039244842529297, 0.13137100219726563, 0.06030847930908203, 0.0605747184753418, 0.060353534698486325, 0.06028902435302735, 0.0603351058959961, 0.060286975860595705, 0.06034739303588867, 0.06035865783691406, 0.06034534454345703, 0.060290046691894535, 0.060437503814697265, 0.060319744110107425, 0.06032076644897461, 0.06028595352172852, 0.06031872177124024, 0.06034841537475586, 0.060665855407714846, 0.06037811279296875, 0.06034329605102539, 0.06041292953491211, 0.0603351058959961, 0.0603422737121582, 0.060314624786376954, 0.06029619216918945, 0.06047129440307617, 0.06029107284545898, 0.06037811279296875, 0.06033203125, 0.06035968017578125, 0.0603422737121582, 0.0603504638671875, 0.0603770866394043, 0.060416000366210934, 0.06037094497680664, 0.0603422737121582, 0.06034431838989258, 0.06035865783691406, 0.0603351058959961, 0.06036380767822266, 0.06032892990112305, 0.060375072479248046, 0.060357601165771484, 0.06030131149291992, 0.060316673278808595, 0.060349441528320315, 0.06036070251464844, 0.06037811279296875, 0.0603054084777832, 0.06037913513183594, 0.06035968017578125, 0.060391422271728515, 0.06034739303588867, 0.06038016128540039, 0.06030847930908203, 0.06037299346923828, 0.06040371322631836, 0.060298240661621094, 0.060401664733886716, 0.06032691192626953, 0.06044160079956055, 0.06035456085205078, 0.06043340682983398, 0.13126860046386718, 0.060421119689941405, 0.060418048858642576, 0.060375038146972655, 0.06035148620605469, 0.06029926300048828, 0.06027264022827149, 0.06029107284545898, 0.06031564712524414, 0.06032691192626953, 0.0603238410949707, 0.060284927368164064, 0.060268543243408204, 0.0603504638671875, 0.06031872177124024, 0.060319744110107425, 0.060298240661621094, 0.0603135986328125, 0.060462078094482424, 0.06032998275756836, 0.06028595352172852, 0.060331008911132813, 0.060298240661621094, 0.06032896041870117, 0.0602716178894043, 0.06033203125, 0.060316673278808595, 0.06044160079956055, 0.06032179260253906, 0.06032896041870117, 0.06029619216918945, 0.06037606430053711, 0.06034739303588867, 0.06036377716064453, 0.06037606430053711, 0.060398593902587894, 0.060385280609130856, 0.060369918823242184, 0.06032793426513672, 0.060407806396484375, 0.06034534454345703, 0.06032076644897461, 0.060368896484375, 0.06032588958740234, 0.060336128234863284, 0.06036070251464844, 0.06034329605102539, 0.060352512359619144, 0.0603238410949707, 0.060331008911132813, 0.06036172866821289, 0.06038323211669922, 0.06035456085205078, 0.06036684799194336, 0.06034841537475586, 0.0603422737121582, 0.060440574645996094, 0.06031052780151367, 0.060365825653076174, 0.06039039993286133, 0.06045798492431641, 0.060355583190917966, 0.060402687072753904, 0.13087129211425783, 0.06030643081665039, 0.0603504638671875, 0.06032486343383789, 0.06032486343383789, 0.06033407974243164, 0.06032691192626953, 0.06029721450805664, 0.060313663482666015, 0.060308414459228514, 0.06031155014038086, 0.06032588958740234, 0.060319744110107425, 0.060314624786376954, 0.06033203125, 0.060319744110107425, 0.06032896041870117, 0.06031769561767578, 0.06030745697021484, 0.06030131149291992, 0.06047436904907227, 0.060437503814697265, 0.06032076644897461, 0.060355583190917966, 0.06031872177124024, 0.06036275100708008, 0.06032179260253906, 0.06032793426513672, 0.0603054084777832, 0.060298240661621094, 0.060365825653076174, 0.06033407974243164, 0.060319744110107425, 0.060388351440429686, 0.06036172866821289, 0.060455936431884766, 0.060388351440429686, 0.060401664733886716, 0.060369918823242184, 0.06038323211669922, 0.06042828750610352, 0.06038937759399414, 0.06034124755859375, 0.06034739303588867, 0.060298240661621094, 0.060333057403564455, 0.06038118362426758, 0.060369918823242184, 0.060900352478027345, 0.06035865783691406, 0.060401664733886716, 0.06035968017578125, 0.0603351058959961, 0.06038118362426758, 0.06034431838989258, 0.06064025497436523, 0.06043545532226562, 0.060369918823242184, 0.06040883255004883, 0.060340255737304685, 0.06042825698852539, 0.06035660934448242, 0.06040063858032227, 0.1306808319091797, 0.06031257629394531, 0.06034636688232422, 0.06033407974243164, 0.0605665283203125, 0.06033715057373047, 0.060300289154052736, 0.06031052780151367, 0.060319744110107425, 0.06033203125, 0.06029312133789062, 0.06029312133789062, 0.06032691192626953, 0.06031155014038086, 0.06032691192626953, 0.0603238410949707, 0.06026137542724609, 0.06029312133789062, 0.060290046691894535, 0.06035660934448242, 0.06030950546264648, 0.06033715057373047, 0.06032896041870117, 0.06042214584350586, 0.06032179260253906, 0.06034431838989258, 0.06031465530395508, 0.06036374282836914, 0.060314624786376954, 0.06031052780151367, 0.06032281494140625, 0.060391422271728515, 0.06036172866821289, 0.06047334289550781, 0.06042726516723633, 0.06035865783691406, 0.060375038146972655, 0.06036479949951172, 0.06031769561767578, 0.06042931365966797, 0.06032799911499023, 0.060305343627929685, 0.06035968017578125, 0.060355583190917966, 0.06034022521972656, 0.06037606430053711, 0.060352512359619144, 0.06036684799194336, 0.06035456085205078, 0.06041088104248047, 0.06037811279296875, 0.0605849609375, 0.060386302947998044, 0.06039654541015625, 0.06033817672729492, 0.06036275100708008, 0.060475391387939455, 0.06033817672729492, 0.06041907119750976, 0.06040371322631836, 0.06048255920410156, 0.060510208129882816, 0.06039449691772461, 0.1310750732421875, 0.06031872177124024, 0.06032793426513672, 0.0603770866394043, 0.06030643081665039, 0.060368896484375, 0.06032179260253906, 0.06029926300048828, 0.06029209518432617, 0.06032998275756836, 0.06029209518432617, 0.06027980804443359, 0.06030131149291992, 0.06029721450805664, 0.060278785705566405, 0.06032793426513672, 0.060319744110107425, 0.06037094497680664, 0.0603422737121582, 0.06032896041870117, 0.060290046691894535, 0.06036070251464844, 0.06032588958740234, 0.0603351058959961, 0.060300289154052736, 0.06034739303588867, 0.0603054084777832, 0.060368896484375, 0.06031564712524414, 0.06034431838989258, 0.06032076644897461, 0.06033919906616211, 0.06032691192626953, 0.06046105575561524, 0.060409854888916016, 0.060349441528320315, 0.060369918823242184, 0.06035865783691406, 0.06032588958740234, 0.06040576171875, 0.060365825653076174, 0.06039244842529297, 0.06039039993286133, 0.060307487487792966, 0.060313568115234376, 0.06037606430053711, 0.06039961624145508, 0.06036377716064453, 0.06030847930908203, 0.0603054084777832, 0.06034534454345703, 0.06039039993286133, 0.06034534454345703, 0.06036684799194336, 0.0603135986328125, 0.06032486343383789, 0.06041292953491211, 0.06047948837280273, 0.06050099182128906, 0.06032076644897461, 0.06044467163085938, 0.06034329605102539, 0.060402687072753904, 0.13153587341308592, 0.060286975860595705, 0.06035148620605469, 0.06040576171875, 0.060283905029296876, 0.060349441528320315, 0.06030847930908203, 0.060300289154052736, 0.06027980804443359, 0.06028287887573242, 0.060319744110107425, 0.06028595352172852, 0.060298240661621094, 0.06046003341674805, 0.06037404632568359, 0.06031766510009766, 0.06031872177124024, 0.060281856536865235, 0.060303359985351565, 0.060319744110107425, 0.06031872177124024, 0.06056959915161133, 0.06044979095458984, 0.06042009735107422, 0.06029619216918945, 0.06031872177124024, 0.06032896041870117, 0.06034841537475586, 0.06034534454345703, 0.060450817108154295, 0.06033615875244141, 0.060324832916259764, 0.060365825653076174, 0.06036684799194336, 0.06040063858032227, 0.06036070251464844, 0.06043340682983398, 0.06036684799194336, 0.06036073684692383, 0.06044566345214844, 0.06033919906616211, 0.060365825653076174, 0.06036377716064453, 0.060349441528320315, 0.06034841537475586, 0.060369918823242184, 0.060381214141845704, 0.06036067199707031, 0.06034431838989258, 0.06032793426513672, 0.060352512359619144, 0.06038118362426758, 0.06040063858032227, 0.06036377716064453, 0.06029721450805664, 0.060442623138427735, 0.06050201416015625, 0.06030643081665039, 0.060416000366210934, 0.06033817672729492, 0.060416000366210934, 0.06034841537475586, 0.06035763168334961, 0.13149183654785157, 0.06031257629394531, 0.06034431838989258, 0.06033817672729492, 0.060267520904541017, 0.06034124755859375, 0.06029209518432617, 0.06033203125, 0.06028799819946289, 0.060331008911132813, 0.06029926300048828, 0.06027980804443359, 0.06034739303588867, 0.06031564712524414, 0.06030745697021484, 0.06032486343383789, 0.0602716178894043, 0.060336128234863284, 0.0603504638671875, 0.06035456085205078, 0.06035763168334961, 0.06036479949951172, 0.06032691192626953, 0.06033715057373047, 0.06032486343383789, 0.06039244842529297, 0.060295169830322265, 0.060368896484375, 0.06030847930908203, 0.06031872177124024, 0.0603351058959961, 0.06036479949951172, 0.06035865783691406, 0.06043033599853516, 0.06038016128540039, 0.0603422737121582, 0.060401664733886716, 0.060371967315673826, 0.0603135986328125, 0.06046003341674805, 0.060333057403564455, 0.060369918823242184, 0.06035456085205078, 0.060352512359619144, 0.0603422737121582, 0.06037094497680664, 0.060349441528320315, 0.060386302947998044, 0.06028902435302735, 0.06037913513183594, 0.06036172866821289, 0.06035865783691406, 0.06035763168334961, 0.060386302947998044, 0.06034124755859375, 0.06039039993286133, 0.06043135833740235, 0.06030131149291992, 0.06038425445556641, 0.06040678405761719, 0.0603873291015625, 0.06038118362426758, 0.06035660934448242]",tokens/s,16.29323503243669,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -271,7 +271,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -546,7 +546,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2472.198144,9028.763648,0.0,8382.316544,7939.283456,s,10,6.1620845947265614,0.6162084594726561,0.0011803163390921698,0.6157510986328125,0.6171693908691406,0.6182771697998047,0.619163392944336,"[0.6155607299804687, 0.6193849487304688, 0.6151373901367188, 0.61642626953125, 0.6163339233398437, 0.6153935546875, 0.6169232177734375, 0.61542236328125, 0.6157410278320312, 0.6157611694335937]",tokens/s,415.4438259725966,kWh,7.26740989420149e-06,3.982230299933529e-06,3.644055202823391e-05,4.769019222236893e-05,tokens/kWh,5367980.040976309,MB,2472.779776,9028.763648,0.0,8382.316544,8233.862144,s,10,338.1937578125,33.81937578125,0.0011103055076484762,33.81973828125,33.820403125000006,33.820582421875,33.820725859375,"[33.81766796875, 33.82027734375, 33.8172265625, 33.82005078125, 33.81898828125, 33.8189453125, 33.82076171875, 33.8196171875, 33.819859375, 33.82036328125]",tokens/s,1.8628374576602387,kWh,0.0003992788065804376,0.0002188394560563585,0.0020008438000201846,0.0026189620626569806,tokens/kWh,24055.331269703638,,s,629,343.20291455078103,0.5456326145481419,0.07326185412723402,0.5368145751953125,0.53704130859375,0.5371090820312501,1.15295984375,"[0.5366138916015625, 0.5368842163085937, 0.5365032958984375, 0.5366651000976562, 0.5367828369140625, 0.5368319702148437, 0.536500244140625, 0.5367971801757813, 0.5363753051757812, 0.5366179809570313, 0.5364674682617188, 0.5371392211914062, 0.5367183227539063, 0.5364090576171875, 0.5364602661132812, 0.5366578979492187, 0.5366630249023437, 0.536791015625, 0.5366445922851563, 0.5370265502929688, 0.5366231079101562, 0.5366497192382812, 0.5368258666992187, 0.5368289184570313, 0.53686376953125, 0.5370091552734375, 0.5367439575195313, 0.5367941284179687, 0.5368914184570313, 0.5367971801757813, 0.5367203979492188, 0.5368668212890625, 0.5365339965820313, 0.5368442993164062, 0.5367408447265625, 0.5368688354492187, 0.5367408447265625, 0.536742919921875, 0.5369190673828125, 0.5368074340820312, 0.5366405029296875, 0.53675830078125, 0.5370582885742188, 0.536890380859375, 0.5368985595703125, 0.5370951538085937, 0.5367521362304688, 0.5369446411132812, 0.5367613525390625, 0.5367725830078125, 0.5367357177734375, 0.5368411865234375, 0.5369948120117187, 0.53707470703125, 0.5366937866210938, 0.5367838745117187, 0.536826904296875, 0.536911865234375, 0.5365247802734375, 0.5369384765625, 0.5366364135742188, 0.5370501098632813, 1.15660498046875, 0.5366784057617188, 0.5368135375976563, 0.5366763305664063, 0.5366251220703125, 0.5368370971679688, 0.5370634155273437, 0.5365442504882812, 0.5366190185546875, 0.5368156127929687, 0.5367030029296875, 0.5364469604492188, 0.5368719482421875, 0.5366988525390625, 0.5366445922851563, 0.5365545043945312, 0.5366353759765625, 0.5364398193359375, 0.5369077758789063, 0.5365125122070312, 0.5368678588867187, 0.5366210327148437, 0.5368207397460938, 0.536680419921875, 0.5368309936523438, 0.536911865234375, 0.5369497680664063, 0.5366302490234375, 0.53686376953125, 0.5368207397460938, 0.536680419921875, 0.5367817993164062, 0.5367613525390625, 0.5367193603515625, 0.5368514404296875, 0.5367982177734375, 0.5366886596679687, 0.536848388671875, 0.5369354248046875, 0.5368606567382812, 0.5369999389648438, 0.5371566162109375, 0.53701123046875, 0.5369251708984375, 0.5368104858398437, 0.53696923828125, 0.5372661743164062, 0.536816650390625, 0.537080810546875, 0.5368729858398438, 0.5366435546875, 0.536722412109375, 0.5367203979492188, 0.5367930908203125, 0.5370296020507812, 0.5369251708984375, 0.5368258666992187, 0.5370675048828125, 0.53682177734375, 0.53661083984375, 0.5368616943359374, 0.5369835815429688, 0.5369682006835937, 1.153555419921875, 0.5368104858398437, 0.536627197265625, 0.536426513671875, 0.5363701782226562, 0.5369129028320313, 0.5365360717773437, 0.5366179809570313, 0.5368780517578124, 0.5365504150390625, 0.5364623413085937, 0.5367234497070312, 0.5365247802734375, 0.5367357177734375, 0.5363619995117187, 0.536642578125, 0.5365043334960937, 0.5364766845703125, 0.5367828369140625, 0.5367777099609375, 0.5369682006835937, 0.53671728515625, 0.5365309448242187, 0.5366927490234376, 0.5366732788085937, 0.536748046875, 0.5368842163085937, 0.536711181640625, 0.5367275390625, 0.53675927734375, 0.5366507568359375, 0.5366527709960938, 0.537112548828125, 0.5368504028320312, 0.5369364624023437, 0.5365678100585938, 0.53705419921875, 0.5368473510742188, 0.536774658203125, 0.5369088134765625, 0.5367490844726562, 0.5367777099609375, 0.5368207397460938, 0.5368125610351563, 0.5366998901367187, 0.536975341796875, 0.5371187133789063, 0.5369088134765625, 0.5368340454101562, 0.5368309936523438, 0.53690673828125, 0.536796142578125, 0.53704296875, 0.5369108276367187, 0.5367992553710937, 0.5370162963867188, 0.536953857421875, 0.5370245361328125, 0.5368985595703125, 0.5370009765625, 0.5369139404296875, 0.5368319702148437, 0.5369620361328125, 1.1536087646484374, 0.53665380859375, 0.5366210327148437, 0.5366343383789063, 0.5365514526367188, 0.5365883178710937, 0.5368463134765625, 0.5365125122070312, 0.536474609375, 0.5365883178710937, 0.5365862426757813, 0.5367296142578125, 0.536585205078125, 0.5366619873046875, 0.5366722412109375, 0.5366896362304687, 0.536859619140625, 0.5366026000976563, 0.5365985107421875, 0.536827880859375, 0.5368822021484375, 0.5369682006835937, 0.5367449340820313, 0.5368606567382812, 0.5368719482421875, 0.5370050659179687, 0.5367684936523438, 0.5367459716796875, 0.5366732788085937, 0.5369620361328125, 0.5367920532226562, 0.536858642578125, 0.5368411865234375, 0.5368186645507812, 0.5369609985351562, 0.5366333618164062, 0.5371084594726563, 0.5371924438476563, 0.5368709106445313, 0.5368985595703125, 0.5369241333007813, 0.5369467163085937, 0.5369282836914062, 0.5369384765625, 0.5369241333007813, 0.5369927978515625, 0.5370040283203125, 0.5367838745117187, 0.5370081176757813, 0.5368955078125, 0.5368207397460938, 0.5365821533203124, 0.537007080078125, 0.5369712524414062, 0.5370511474609375, 0.5367930908203125, 0.5369210815429688, 0.5370777587890625, 0.5367992553710937, 0.5366517944335938, 0.5372211303710938, 0.5368811645507813, 0.5370286254882812, 1.1536322021484375, 0.5366702270507813, 0.53663232421875, 0.5365944213867188, 0.5365933837890625, 0.536669189453125, 0.536585205078125, 0.5366456298828125, 0.5366261596679688, 0.5365391235351562, 0.53661181640625, 0.5368197021484375, 0.5366292724609375, 0.5367183227539063, 0.5363814697265625, 0.5366312866210937, 0.5366026000976563, 0.5365043334960937, 0.5368678588867187, 0.5367070922851562, 0.5367982177734375, 0.5367142333984375, 0.5365944213867188, 0.5368822021484375, 0.536764404296875, 0.5367572631835937, 0.5367725830078125, 0.5366507568359375, 0.5366978759765625, 0.5369057006835938, 0.5367367553710938, 0.5368442993164062, 0.5371207885742187, 0.5368688354492187, 0.536975341796875, 0.5372374877929688, 0.5367255249023437, 0.5367490844726562, 0.536848388671875, 0.5366968383789062, 0.5368555297851563, 0.5369047241210938, 0.53682275390625, 0.5368104858398437, 0.536953857421875, 0.537143310546875, 0.5368975219726563, 0.536890380859375, 0.5368473510742188, 0.5369108276367187, 0.5374085083007812, 0.5368156127929687, 0.5369896850585938, 0.5370736694335938, 0.53701123046875, 0.536953857421875, 0.5369763793945312, 0.5369876708984375, 0.536796142578125, 0.5366702270507813, 0.5370020141601562, 0.5367982177734375, 0.5368473510742188, 1.152858154296875, 0.5366998901367187, 0.536784912109375, 0.5366947631835938, 0.5364510498046875, 0.5367101440429688, 0.536543212890625, 0.5367490844726562, 0.5370357666015625, 0.5365555419921875, 0.5364735717773438, 0.5366886596679687, 0.5366702270507813, 0.5367971801757813, 0.53650634765625, 0.5364786987304687, 0.5364551391601563, 0.53650537109375, 0.5366241455078125, 0.5367337036132812, 0.536774658203125, 0.536574951171875, 0.5367716064453125, 0.537091064453125, 0.5366610107421875, 0.5366507568359375, 0.5369815063476563, 0.53661083984375, 0.5366763305664063, 0.5367521362304688, 0.5368361206054687, 0.5368678588867187, 0.5367817993164062, 0.5368463134765625, 0.53707470703125, 0.5366835327148437, 0.536911865234375, 0.536911865234375, 0.5369528198242187, 0.5372190551757813, 0.5367623901367188, 0.5369282836914062, 0.5367654418945312, 0.5368145751953125, 0.5367992553710937, 0.5374668579101562, 0.5371975708007812, 0.5368463134765625, 0.5368524780273437, 0.5368186645507812, 0.53682177734375, 0.5366507568359375, 0.5368616943359374, 0.5369609985351562, 0.53703271484375, 0.53690673828125, 0.5366845703125, 0.5369559326171875, 0.5369241333007813, 0.5368340454101562, 0.5370408935546875, 0.5368411865234375, 0.5369733276367188, 1.1542640380859375, 0.5367930297851562, 0.5365872802734375, 0.5365022583007812, 0.5364807739257812, 0.5368463134765625, 0.5367828369140625, 0.53667431640625, 0.5366415405273437, 0.536479736328125, 0.5365933837890625, 0.5367357788085938, 0.5367592163085938, 0.5366712036132812, 0.5365125122070312, 0.53663232421875, 0.5365278930664062, 0.5366466064453125, 0.5368186645507812, 0.5369241333007813, 0.5368115234375, 0.5370562744140625, 0.5365463256835937, 0.53686376953125, 0.5368514404296875, 0.5370029907226562, 0.5367296142578125, 0.5366886596679687, 0.5367869262695313, 0.53690673828125, 0.5365933837890625, 0.5370020141601562, 0.5367613525390625, 0.53675830078125, 0.5370337524414063, 0.5367316284179687, 0.5368453369140626, 0.5372139282226562, 0.5369917602539063, 0.5368442993164062, 0.5370337524414063, 0.5368677978515625, 0.5368289184570313, 0.5368361206054687, 0.5369886474609376, 0.5370162963867188, 0.5369282836914062, 0.5369343872070312, 0.53701220703125, 0.53686474609375, 0.536791015625, 0.5366722412109375, 0.5372876586914063, 0.5370562744140625, 0.536880126953125, 0.5368606567382812, 0.5368473510742188, 0.537079833984375, 0.53690673828125, 0.5369467163085937, 0.537111572265625, 0.5369978637695313, 0.5370552368164062, 1.152716796875, 0.5368033447265625, 0.5367306518554688, 0.5365186767578125, 0.5366087646484375, 0.5367869262695313, 0.5367767333984375, 0.5365708618164062, 0.536700927734375, 0.536627197265625, 0.5366988525390625, 0.5369722900390625, 0.5366445922851563, 0.5371094970703125, 0.5364469604492188, 0.5367030029296875, 0.5365135498046875, 0.5366927490234376, 0.5367418823242187, 0.53669580078125, 0.536995849609375, 0.53682275390625, 0.536711181640625, 0.5367275390625, 0.5368361206054687, 0.5368811645507813, 0.5368923950195312, 0.5364756469726563, 0.5366835327148437, 0.5368811645507813, 0.5367879638671875, 0.5370316772460938, 0.5366599731445313, 0.5367091064453124, 0.5370531616210937, 0.5366466674804687, 0.5371033325195312, 0.5367869262695313, 0.5368330078125, 0.536774658203125, 0.5366927490234376, 0.5368402099609375, 0.53682275390625, 0.5367613525390625, 0.5369722900390625, 0.5369609985351562, 0.5371791381835938, 0.5367408447265625, 0.536826904296875, 0.5367736206054687, 0.5367817993164062, 0.5369354248046875, 0.53711669921875, 0.5369241333007813, 0.5370623779296875, 0.5369609985351562, 0.5368463134765625, 0.5369671630859375, 0.536806396484375, 0.5367019653320313, 0.5371135864257812, 0.5369528198242187, 0.53698046875, 1.1529993896484374, 0.5368299560546875, 0.5370091552734375, 0.5366353759765625, 0.5365616455078125, 0.5366558837890625, 0.536468505859375, 0.5366865844726563, 0.5366210327148437, 0.5364827880859375, 0.5366784057617188, 0.5368084716796875, 0.536574951171875, 0.5367470092773438, 0.5364807739257812, 0.5367275390625, 0.5365463256835937, 0.5368893432617188, 0.5367490844726562, 0.5366538696289063, 0.5368145751953125, 0.5367122192382813, 0.536791015625, 0.5369139404296875, 0.5368699340820312, 0.5368401489257812, 0.5368053588867188, 0.5367551879882813, 0.536791015625, 0.5368115234375, 0.5367808227539063, 0.5368504028320312, 0.5366098022460938, 0.536838134765625, 0.5369794311523437, 0.5365626831054687, 0.5368678588867187, 0.5369231567382813, 0.5369620361328125, 0.5371443481445313, 0.5370357666015625, 0.5368053588867188, 0.5367316284179687, 0.5368494262695312, 0.5369876708984375, 0.5369149169921875, 0.5369937744140625, 0.5367244873046875, 0.53688525390625, 0.5367900390625, 0.5368668212890625, 0.5369569091796875, 0.536943603515625, 0.5370767211914063, 0.53701123046875, 0.5369712524414062, 0.5369476928710938, 0.53696923828125, 0.53707568359375, 0.53675830078125, 0.5370572509765625, 0.5368934326171875, 0.5369927978515625, 1.1530455322265625, 0.5370398559570313, 0.536827880859375, 0.53667431640625, 0.5366241455078125, 0.5367562255859375, 0.5369866333007812, 0.5366569213867187, 0.536543212890625, 0.5366067504882812, 0.5366558837890625, 0.536690673828125, 0.5370921020507813, 0.5367521362304688, 0.5363927001953125, 0.5365883178710937, 0.5365883178710937, 0.5364572143554688, 0.5367060546875, 0.5369026489257812, 0.5367357177734375, 0.5368033447265625, 0.5366507568359375, 0.5368565673828125, 0.5367767333984375, 0.536933349609375, 0.5369292602539063, 0.5367490844726562, 0.5366527709960938, 0.53698046875, 0.536901611328125, 0.5368790893554688, 0.5367060546875, 0.5367142333984375, 0.5368207397460938, 0.5366681518554688, 0.536848388671875, 0.5368248291015625, 0.536901611328125, 0.5367930908203125, 0.5368125610351563, 0.5368084716796875, 0.5369190673828125, 0.5370562744140625, 0.5367808227539063, 0.53699072265625, 0.5371361083984375, 0.5368914184570313, 0.5368616943359374, 0.5369302978515625, 0.5368442993164062, 0.536732666015625, 0.5369579467773438, 0.5370081176757813, 0.5371074829101562, 0.5370603637695313, 0.5368606567382812, 0.5370623779296875, 0.5368893432617188, 0.5367756958007812, 0.5370880126953125, 0.5369036865234375, 0.5369425659179687]",tokens/s,1.8327350186501152,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -575,7 +575,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -734,7 +734,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp2jevaacg/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1132,7 +1132,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpgtu01ugm/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2073,7 +2073,7 @@ ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please req ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1279.791104,1019.74016,0.0,373.293056,277.410816,s,10,0.3806483840942383,0.03806483840942383,0.0010374051807092283,0.03839825630187988,0.038631001281738284,0.03909522857666015,0.03946661041259766,"[0.03955945587158203, 0.03852783966064453, 0.03837411117553711, 0.038422401428222654, 0.03847126388549805, 0.038292991638183595, 0.03834467315673828, 0.03843852615356445, 0.036084991455078125, 0.03613212966918945]",tokens/s,6725.366787229584,kWh,4.3617786259962035e-07,2.3900503699099743e-07,1.1732389459702955e-06,1.8484218455609134e-06,tokens/kWh,138496523.73174343,MB,1280.086016,1019.74016,0.0,373.293056,323.047424,s,10,23.712255126953128,2.371225512695313,0.003882883813537052,2.3718930664062503,2.376249291992188,2.3762570922851562,2.3762633325195313,"[2.362728515625, 2.37624755859375, 2.373291748046875, 2.367230224609375, 2.371970947265625, 2.371815185546875, 2.376264892578125, 2.37161865234375, 2.37223681640625, 2.3688505859375]",tokens/s,26.56854004931377,kWh,2.7777078536237154e-05,1.5222735153430559e-05,6.781520513363374e-05,0.00011081501882330147,tokens/kWh,568514.9961527847,,s,629,24.01549103546141,0.03818043089898478,0.004509392475924156,0.03761459350585938,0.03786444854736328,0.03806208114624023,0.07505829895019543,"[0.03638784027099609, 0.03659775924682617, 0.036880382537841795, 0.03663974380493164, 0.03670937728881836, 0.03717529678344727, 0.036736000061035154, 0.03655987167358398, 0.03629260635375976, 0.03701657485961914, 0.03663974380493164, 0.0363612174987793, 0.03665100860595703, 0.038844417572021485, 0.03839078521728516, 0.038013950347900394, 0.03768627166748047, 0.03772825622558594, 0.037766143798828124, 0.03749990463256836, 0.037585918426513674, 0.03749683380126953, 0.03755929565429687, 0.03760537719726562, 0.037743614196777346, 0.03752140808105469, 0.0375551986694336, 0.03762995147705078, 0.037495807647705076, 0.03751935958862305, 0.03787571334838867, 0.03770880126953125, 0.0376360969543457, 0.03753472137451172, 0.037599231719970705, 0.03759718322753906, 0.03856595230102539, 0.03801900863647461, 0.037601280212402347, 0.03762073516845703, 0.0376545295715332, 0.03752243041992188, 0.03753881454467774, 0.03760435104370117, 0.03775795364379883, 0.037550079345703126, 0.03786649703979492, 0.03831398391723633, 0.03769036865234375, 0.037601280212402347, 0.03755929565429687, 0.037498878479003905, 0.03746303939819336, 0.03752345657348633, 0.03794432067871094, 0.037713920593261716, 0.03760435104370117, 0.037495807647705076, 0.037591041564941405, 0.03748454284667969, 0.03763916778564453, 0.03788083267211914, 0.07681126403808594, 0.03763711929321289, 0.03831808090209961, 0.03748966217041016, 0.03763507080078125, 0.03758489608764649, 0.03759308624267578, 0.037579776763916016, 0.03772419357299805, 0.03755926513671875, 0.03763097763061524, 0.03747020721435547, 0.03756748962402344, 0.03761663818359375, 0.037901313781738284, 0.03767398452758789, 0.03793817520141601, 0.03772415924072266, 0.037601280212402347, 0.03751321411132812, 0.03760332870483398, 0.03770470428466797, 0.03760435104370117, 0.03767705535888672, 0.03810611343383789, 0.0376995849609375, 0.03758694458007812, 0.03785728073120117, 0.03757878494262695, 0.03765244674682617, 0.03786956787109375, 0.03756032180786133, 0.037838848114013675, 0.037773311614990236, 0.037525505065917966, 0.037884929656982425, 0.03775692749023438, 0.03761151885986328, 0.0378869743347168, 0.03772518539428711, 0.037797889709472655, 0.03774566268920899, 0.037759998321533206, 0.03785113525390625, 0.038040576934814455, 0.03770470428466797, 0.0377077751159668, 0.03755929565429687, 0.037599231719970705, 0.037661697387695314, 0.037553150177001955, 0.037823486328125, 0.03759820938110352, 0.03769651031494141, 0.037648384094238284, 0.03772313690185547, 0.03791667175292969, 0.03774259185791016, 0.03760435104370117, 0.03763711929321289, 0.03762483215332031, 0.03786444854736328, 0.03789516830444336, 0.07634636688232421, 0.03765350341796875, 0.03746713638305664, 0.03770470428466797, 0.03758796691894531, 0.03751116943359375, 0.03785318374633789, 0.037889022827148434, 0.03752345657348633, 0.0374774398803711, 0.037649345397949216, 0.03765760040283203, 0.037743614196777346, 0.0377149429321289, 0.03764329528808594, 0.03746198272705078, 0.037561344146728515, 0.037763072967529294, 0.03768832015991211, 0.03759823989868164, 0.037550048828125, 0.03745382308959961, 0.03772825622558594, 0.03768012619018555, 0.03773952102661133, 0.03763097763061524, 0.03759820938110352, 0.03765657424926758, 0.0381921272277832, 0.037905406951904294, 0.03770675277709961, 0.03752345657348633, 0.03779072189331055, 0.0376627197265625, 0.03760844802856445, 0.03748761749267578, 0.03757670211791992, 0.0375654411315918, 0.03750713729858399, 0.03755820846557617, 0.0375551986694336, 0.03763302230834961, 0.037645313262939455, 0.03746099090576172, 0.03763507080078125, 0.0376258544921875, 0.03769651031494141, 0.0375203857421875, 0.037610496520996094, 0.037612545013427735, 0.03768627166748047, 0.03968307113647461, 0.03810201644897461, 0.03744870376586914, 0.03748761749267578, 0.03760844802856445, 0.037705726623535156, 0.0375654411315918, 0.037533695220947266, 0.037482494354248046, 0.037713920593261716, 0.037579776763916016, 0.037495807647705076, 0.07634534454345702, 0.03747840118408203, 0.037424129486083986, 0.03769753646850586, 0.03747020721435547, 0.03760639953613281, 0.03746815872192383, 0.03759308624267578, 0.03759001541137695, 0.037582847595214845, 0.037572608947753904, 0.037525505065917966, 0.03752959823608398, 0.03775590515136719, 0.03758899307250976, 0.037459968566894535, 0.0377446403503418, 0.037495807647705076, 0.03755929565429687, 0.037438465118408204, 0.03756851196289063, 0.037602302551269534, 0.037515262603759765, 0.03751321411132812, 0.03760639953613281, 0.037541889190673826, 0.03759206390380859, 0.03758694458007812, 0.03766681671142578, 0.037548030853271484, 0.037512191772460936, 0.037582847595214845, 0.037493759155273435, 0.037449726104736326, 0.03746815872192383, 0.03757056045532227, 0.037823486328125, 0.037482494354248046, 0.037579776763916016, 0.03767091369628906, 0.03766886520385742, 0.03751424026489258, 0.03748147201538086, 0.037495807647705076, 0.03765555191040039, 0.03773132705688476, 0.03749273681640625, 0.03760639953613281, 0.03768217468261719, 0.037533695220947266, 0.037394432067871096, 0.03765350341796875, 0.03762278366088867, 0.03748966217041016, 0.03768729782104492, 0.0376514892578125, 0.03766985702514648, 0.03745280075073242, 0.037541889190673826, 0.03762073516845703, 0.03751731109619141, 0.037572608947753904, 0.03764223861694336, 0.07663206481933593, 0.037482494354248046, 0.03758694458007812, 0.03754598236083984, 0.03749785614013672, 0.03743641662597656, 0.03762790298461914, 0.03742105484008789, 0.03755929565429687, 0.037628929138183595, 0.03765350341796875, 0.03744768142700195, 0.03746918487548828, 0.03767500686645508, 0.03757670211791992, 0.03751222229003906, 0.03749884796142578, 0.03762995147705078, 0.03771596908569336, 0.03759718322753906, 0.03778252792358398, 0.03771596908569336, 0.03758182525634766, 0.03766886520385742, 0.0377262077331543, 0.03749683380126953, 0.0374917106628418, 0.037571582794189456, 0.03759513473510742, 0.03750912094116211, 0.037544960021972655, 0.0376627197265625, 0.03753574371337891, 0.037466110229492186, 0.03755212783813477, 0.03748044967651367, 0.03741798400878906, 0.03748761749267578, 0.03759414291381836, 0.03739030456542969, 0.03777740859985351, 0.03743129730224609, 0.037466110229492186, 0.037466110229492186, 0.03757056045532227, 0.03745177459716797, 0.03782451248168945, 0.03930624008178711, 0.03780198287963867, 0.037553150177001955, 0.037572608947753904, 0.037628929138183595, 0.03765350341796875, 0.03766579055786133, 0.03758796691894531, 0.03759718322753906, 0.03784396743774414, 0.038711296081542966, 0.03820236968994141, 0.037833728790283204, 0.03772931289672852, 0.03775075149536133, 0.03758796691894531, 0.07656038665771485, 0.03744153594970703, 0.037580799102783204, 0.03758489608764649, 0.03763507080078125, 0.03759308624267578, 0.03766681671142578, 0.0375470085144043, 0.0374835205078125, 0.03772313690185547, 0.037703678131103514, 0.037541889190673826, 0.0375551986694336, 0.037652511596679684, 0.037497825622558593, 0.03801190567016602, 0.037644287109375, 0.03753881454467774, 0.03759718322753906, 0.03752345657348633, 0.037541889190673826, 0.03757363128662109, 0.0374835205078125, 0.03762176132202148, 0.03808870315551758, 0.03783782577514649, 0.03767705535888672, 0.03754086303710937, 0.03700121688842774, 0.03764223861694336, 0.03752959823608398, 0.037765121459960936, 0.03765862274169922, 0.03768012619018555, 0.03785116958618164, 0.03752854537963867, 0.03746201705932617, 0.037591041564941405, 0.03764022445678711, 0.03770057678222656, 0.03760435104370117, 0.037569534301757815, 0.03757875061035156, 0.037550079345703126, 0.03768320083618164, 0.03784806442260742, 0.037806079864501956, 0.037579776763916016, 0.03758796691894531, 0.03765248107910156, 0.03765862274169922, 0.03774566268920899, 0.03764121627807617, 0.03772313690185547, 0.03784294509887695, 0.037817344665527344, 0.03783987045288086, 0.03761971282958984, 0.03763814544677734, 0.03765862274169922, 0.0377077751159668, 0.03765862274169922, 0.03762995147705078, 0.07627571105957032, 0.03767603302001953, 0.03746713638305664, 0.03762278366088867, 0.03759001541137695, 0.037689342498779296, 0.03778559875488281, 0.03764019012451172, 0.03788390350341797, 0.03751424026489258, 0.03749273681640625, 0.03770880126953125, 0.03754291152954101, 0.037703678131103514, 0.03769036865234375, 0.037574657440185545, 0.03785420989990235, 0.03765657424926758, 0.03755929565429687, 0.037599231719970705, 0.03741593551635742, 0.03765555191040039, 0.0375838737487793, 0.038059009552001956, 0.03752243041992188, 0.03781324768066406, 0.03762790298461914, 0.03782761764526367, 0.03764118576049805, 0.03759513473510742, 0.03786444854736328, 0.037617664337158206, 0.03751731109619141, 0.037566463470458986, 0.03778252792358398, 0.037596160888671876, 0.03751321411132812, 0.03806412887573242, 0.03936972808837891, 0.03786342239379883, 0.03775897598266602, 0.037601280212402347, 0.03762483215332031, 0.03776921463012695, 0.03782860946655273, 0.037449726104736326, 0.037705726623535156, 0.0375654411315918, 0.03755929565429687, 0.03753267288208008, 0.03763302230834961, 0.03753881454467774, 0.037580799102783204, 0.03768832015991211, 0.037574657440185545, 0.037531646728515625, 0.03764940643310547, 0.03754393768310547, 0.03761151885986328, 0.037515262603759765, 0.03761459350585938, 0.03788288116455078, 0.03977830505371094, 0.07646412658691407, 0.03738828659057617, 0.03792588806152344, 0.037748737335205076, 0.03746713638305664, 0.037897216796875, 0.037553150177001955, 0.03762176132202148, 0.037343231201171875, 0.037651454925537106, 0.037405696868896485, 0.0374835205078125, 0.03782758331298828, 0.03766681671142578, 0.03759001541137695, 0.0375838737487793, 0.0378081283569336, 0.03773235321044922, 0.03765760040283203, 0.03766886520385742, 0.03765964889526367, 0.03771187210083008, 0.03750707244873047, 0.03753472137451172, 0.037531646728515625, 0.03755110549926758, 0.03769343948364258, 0.03748659133911133, 0.037596160888671876, 0.03750707244873047, 0.03753881454467774, 0.037820415496826174, 0.03772415924072266, 0.03755929565429687, 0.037553150177001955, 0.03769036865234375, 0.03775897598266602, 0.037594112396240234, 0.03757875061035156, 0.03759513473510742, 0.03752755355834961, 0.03780505752563477, 0.03753267288208008, 0.03759820938110352, 0.03762790298461914, 0.03850447845458985, 0.03777225494384766, 0.03763916778564453, 0.03750297546386719, 0.037561344146728515, 0.037645313262939455, 0.03790848159790039, 0.037582847595214845, 0.037553150177001955, 0.037664768218994144, 0.03753574371337891, 0.03770675277709961, 0.03749075317382813, 0.03756332778930664, 0.03781324768066406, 0.03759718322753906, 0.037773311614990236, 0.037571582794189456, 0.07186124420166015, 0.037190654754638675, 0.03747225570678711, 0.03754908752441406, 0.037579742431640625, 0.03760639953613281, 0.03770982360839844, 0.03739136123657227, 0.03773440170288086, 0.0376360969543457, 0.03769241714477539, 0.03769241714477539, 0.03752345657348633, 0.03768627166748047, 0.03762688064575195, 0.03752140808105469, 0.037733375549316404, 0.03770675277709961, 0.037754878997802735, 0.03753472137451172, 0.03745792007446289, 0.03804876708984375, 0.03773747253417969, 0.037572608947753904, 0.03754291152954101, 0.03757056045532227, 0.03814297485351562, 0.03765350341796875, 0.03766988754272461, 0.03759001541137695, 0.03777536010742188, 0.03753472137451172, 0.037574657440185545, 0.03759718322753906, 0.03750707244873047, 0.03766886520385742, 0.03754393768310547, 0.03745280075073242, 0.03773952102661133, 0.03783168029785156, 0.03754086303710937, 0.03762688064575195, 0.037483551025390624, 0.03754902267456055, 0.03769036865234375, 0.03759718322753906, 0.037456897735595705, 0.03748761749267578, 0.038013950347900394, 0.037768192291259765, 0.03762073516845703, 0.03767705535888672, 0.037928958892822266, 0.03753779220581055, 0.03802828979492188, 0.037585918426513674, 0.039975936889648435, 0.03809791946411133, 0.03769343948364258, 0.037564414978027344, 0.03762278366088867, 0.03766988754272461, 0.03770470428466797, 0.07192781066894531, 0.03523993682861328, 0.03559731292724609, 0.038373374938964845, 0.03784908676147461, 0.03781017684936523, 0.03761663818359375, 0.037765121459960936, 0.037564414978027344, 0.037615615844726565, 0.03755417633056641, 0.038040576934814455, 0.03775692749023438, 0.03775692749023438, 0.037582847595214845, 0.037716991424560545, 0.03759308624267578, 0.03775283050537109, 0.03753062438964844, 0.03760435104370117, 0.03752447891235351, 0.037569534301757815, 0.03765964889526367, 0.03779891204833984, 0.03758182525634766, 0.037698558807373043, 0.037585918426513674, 0.03753062438964844, 0.03767398452758789, 0.037664768218994144, 0.0376627197265625, 0.037787647247314454, 0.03765657424926758, 0.037833728790283204, 0.037781505584716796, 0.037684223175048825, 0.037719039916992186, 0.03794944000244141, 0.037571582794189456, 0.03768627166748047, 0.037579776763916016, 0.037574657440185545, 0.03738521575927734, 0.03818188858032227, 0.0376627197265625, 0.03749990463256836, 0.037795841217041014, 0.03761151885986328, 0.037803009033203126, 0.0377077751159668, 0.03760435104370117, 0.03778867340087891, 0.037615615844726565, 0.03751731109619141, 0.03779379272460937, 0.03823820877075195, 0.03774054336547852, 0.03762688064575195, 0.03750400161743164, 0.037553150177001955, 0.037956607818603515, 0.03771289443969727, 0.037722110748291016]",tokens/s,26.191427819286087,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2018.93888,5274.861568,0.0,4628.414464,4279.515648,s,10,3.0976935119628908,0.30976935119628907,0.0007532381367795811,0.30955081176757815,0.3109149108886719,0.3109400146484375,0.31096009765625,"[0.3092184143066406, 0.3087049560546875, 0.3094501647949219, 0.3094185485839844, 0.30965145874023436, 0.3089674072265625, 0.30985198974609374, 0.31055612182617187, 0.31090933227539064, 0.3109651184082031]",tokens/s,826.4213325539185,kWh,3.650658585206427e-06,1.999819851334905e-06,1.778628527278799e-05,2.3436763709329325e-05,tokens/kWh,10923009.813769454,MB,2023.018496,5274.861568,0.0,4628.414464,4463.184384,s,10,173.1631875,17.31631875,0.0012339063464567417,17.316303710937497,17.317707031250002,17.318111328125,17.318434765625,"[17.316962890625, 17.318515625, 17.315822265625, 17.315650390625, 17.3176171875, 17.314611328125, 17.314958984375, 17.315001953125, 17.31678515625, 17.31726171875]",tokens/s,3.6381866671286587,kWh,0.00020445280675334164,0.00011205687635914728,0.0009877879636304137,0.0013042976467429026,tokens/kWh,48301.8582125973,,s,629,175.6766339416501,0.27929512550341873,0.0368162753573841,0.2748590087890625,0.2750474243164063,0.27512094726562497,0.584229677734375,"[0.27480780029296875, 0.27519488525390623, 0.2747535400390625, 0.27477914428710937, 0.2746726379394531, 0.2748118896484375, 0.2747586669921875, 0.2748160095214844, 0.2747463684082031, 0.2747105407714844, 0.27477197265625, 0.27491226196289065, 0.274830322265625, 0.2747412414550781, 0.27507815551757814, 0.2747678833007812, 0.2746982421875, 0.2746357727050781, 0.2747279357910156, 0.27481292724609374, 0.2747586669921875, 0.27477093505859373, 0.27480267333984376, 0.2746726379394531, 0.2747781066894531, 0.2748037109375, 0.27487026977539064, 0.27482623291015623, 0.2746746826171875, 0.27491021728515624, 0.2748866577148438, 0.27482110595703124, 0.27484774780273435, 0.2748170166015625, 0.2750279541015625, 0.2748395385742187, 0.2748538818359375, 0.2749214782714844, 0.27490509033203125, 0.27485696411132815, 0.2751037292480469, 0.27481396484375, 0.2748989562988281, 0.2748948364257813, 0.2748395385742187, 0.2748098449707031, 0.2748395385742187, 0.2748293151855469, 0.2749644775390625, 0.27495526123046876, 0.27507815551757814, 0.274988037109375, 0.27512933349609375, 0.2749081726074219, 0.2749347839355469, 0.2750074768066406, 0.2749327392578125, 0.27496551513671874, 0.2750074768066406, 0.27491226196289065, 0.27496139526367186, 0.2749798278808594, 0.5839052734375, 0.2747381896972656, 0.2746685485839844, 0.27480780029296875, 0.27462655639648437, 0.2747535400390625, 0.274850830078125, 0.27504537963867187, 0.274882568359375, 0.2748190612792969, 0.27480575561523435, 0.2750689392089844, 0.2746798095703125, 0.27480575561523435, 0.2747381896972656, 0.2746982421875, 0.27474533081054686, 0.27501773071289065, 0.27489996337890626, 0.27472384643554687, 0.2748252258300781, 0.2747166748046875, 0.27489996337890626, 0.2747637634277344, 0.274777099609375, 0.27480679321289064, 0.27467572021484377, 0.2747699279785156, 0.2752634887695313, 0.274724853515625, 0.27481497192382814, 0.27493170166015624, 0.27494503784179686, 0.2750330810546875, 0.27488153076171873, 0.2748272705078125, 0.27500338745117187, 0.27493991088867187, 0.27490609741210936, 0.2749163513183594, 0.274882568359375, 0.2751744079589844, 0.2748958740234375, 0.27488461303710937, 0.27520510864257813, 0.2751129455566406, 0.2748856201171875, 0.27490609741210936, 0.275019775390625, 0.27484979248046876, 0.2750382080078125, 0.27491021728515624, 0.2749378662109375, 0.2749491271972656, 0.27492352294921873, 0.2749931640625, 0.27498086547851563, 0.2750474243164063, 0.27502899169921874, 0.27504434204101563, 0.2750597229003906, 0.27496856689453125, 0.27508428955078124, 0.584416259765625, 0.274740234375, 0.27477093505859373, 0.27474432373046875, 0.2746582946777344, 0.274671630859375, 0.27458456420898436, 0.2746163330078125, 0.27472384643554687, 0.2748641357421875, 0.27479244995117186, 0.27462451171875, 0.2746695556640625, 0.27464703369140625, 0.27479449462890626, 0.2748590087890625, 0.27491738891601564, 0.27486004638671874, 0.2748610534667969, 0.27494503784179686, 0.27507815551757814, 0.27484979248046876, 0.27474533081054686, 0.2746705932617188, 0.2747832336425781, 0.27472897338867186, 0.27475762939453124, 0.27486209106445314, 0.27463372802734376, 0.27475762939453124, 0.27485491943359375, 0.27470437622070315, 0.2752000122070313, 0.27496139526367186, 0.27483444213867186, 0.27475149536132815, 0.274956298828125, 0.2748334045410156, 0.2748170166015625, 0.27502590942382815, 0.27488973999023436, 0.2748323974609375, 0.2750013427734375, 0.27480575561523435, 0.27495834350585935, 0.27506585693359376, 0.27492556762695314, 0.27496139526367186, 0.2750126037597656, 0.274946044921875, 0.27499725341796877, 0.2750269470214844, 0.27491943359375, 0.274872314453125, 0.2750965881347656, 0.27492556762695314, 0.2749634704589844, 0.27498699951171873, 0.27491943359375, 0.2748856201171875, 0.27493887329101563, 0.2748241882324219, 0.2750003051757812, 0.5843558349609375, 0.2747125854492187, 0.27458660888671876, 0.2746224670410156, 0.27471463012695313, 0.2747678833007812, 0.27459890747070315, 0.274724853515625, 0.2747064208984375, 0.2748293151855469, 0.2745948181152344, 0.27471768188476564, 0.2746695556640625, 0.274840576171875, 0.2746921081542969, 0.27465625, 0.27479345703125, 0.27493170166015624, 0.27479962158203125, 0.27482110595703124, 0.27469515991210935, 0.2747832336425781, 0.27483544921875, 0.2746767272949219, 0.2749491271972656, 0.2748887023925781, 0.2751119384765625, 0.274798583984375, 0.274882568359375, 0.27497369384765624, 0.2748590087890625, 0.2748887023925781, 0.27485491943359375, 0.27481805419921873, 0.2747545471191406, 0.27473919677734376, 0.2748108825683594, 0.2749921264648437, 0.27480575561523435, 0.27500543212890627, 0.2749286499023437, 0.27485592651367186, 0.27485592651367186, 0.2749378662109375, 0.2750003051757812, 0.27502490234375, 0.274951171875, 0.2748856201171875, 0.27495013427734377, 0.27501568603515625, 0.2748528747558594, 0.2749849548339844, 0.2749573059082031, 0.27474945068359374, 0.275061767578125, 0.2749859924316406, 0.2748241882324219, 0.2749593505859375, 0.27505050659179686, 0.27484262084960936, 0.27493374633789064, 0.27519793701171874, 0.2749962158203125, 0.5846722412109375, 0.2751098937988281, 0.2747699279785156, 0.27475045776367185, 0.2746501159667969, 0.2746787719726563, 0.27471463012695313, 0.27465728759765623, 0.27470132446289064, 0.27461016845703123, 0.2745999450683594, 0.27479449462890626, 0.27462860107421877, 0.27524710083007814, 0.27474227905273435, 0.27471563720703124, 0.2747084655761719, 0.2747586669921875, 0.27493170166015624, 0.2747279357910156, 0.27475762939453124, 0.27473715209960936, 0.2749266052246094, 0.27491021728515624, 0.27474945068359374, 0.2747340698242188, 0.2748794860839844, 0.27477093505859373, 0.2746552429199219, 0.27478732299804687, 0.2748487548828125, 0.2747637634277344, 0.27487026977539064, 0.27508837890625, 0.2749296569824219, 0.2750433349609375, 0.2749542541503906, 0.2748968811035156, 0.2750556030273438, 0.2749910888671875, 0.2748293151855469, 0.27487744140625, 0.274988037109375, 0.27488973999023436, 0.2748190612792969, 0.274951171875, 0.27495321655273436, 0.2749941711425781, 0.2749327392578125, 0.27491122436523435, 0.27508428955078124, 0.275135498046875, 0.2751324157714844, 0.27496551513671874, 0.2749819030761719, 0.27534335327148435, 0.27508633422851564, 0.27493887329101563, 0.27502899169921874, 0.27484466552734377, 0.2749542541503906, 0.2749644775390625, 0.27518875122070313, 0.5838991088867187, 0.2747965393066406, 0.2747125854492187, 0.2746480712890625, 0.2746009521484375, 0.2746828918457031, 0.274661376953125, 0.2748989562988281, 0.27465625, 0.2748037109375, 0.2750279541015625, 0.2747432861328125, 0.27462042236328127, 0.27462860107421877, 0.2747166748046875, 0.2746152954101562, 0.2747555847167969, 0.27466033935546874, 0.27464602661132814, 0.27475967407226565, 0.2746582946777344, 0.27469720458984376, 0.27470233154296875, 0.27486514282226565, 0.2747135925292969, 0.2746306457519531, 0.2747412414550781, 0.2748241882324219, 0.2748016662597656, 0.27485696411132815, 0.27470745849609374, 0.2747832336425781, 0.2747484130859375, 0.2749245300292969, 0.2750597229003906, 0.27482009887695313, 0.27485696411132815, 0.27487026977539064, 0.27477914428710937, 0.2749245300292969, 0.27483544921875, 0.27486309814453125, 0.2748436584472656, 0.2752860107421875, 0.2748108825683594, 0.274850830078125, 0.27503103637695314, 0.2750986328125, 0.27488153076171873, 0.27500543212890627, 0.27479962158203125, 0.27487335205078123, 0.27490609741210936, 0.2748805236816406, 0.27498904418945314, 0.2749696044921875, 0.275061767578125, 0.27496652221679685, 0.2752071533203125, 0.27506585693359376, 0.2748989562988281, 0.2749747314453125, 0.275009521484375, 0.5848350830078125, 0.274661376953125, 0.27465728759765623, 0.274671630859375, 0.27472076416015623, 0.27464190673828126, 0.2747658386230469, 0.2749992980957031, 0.27464089965820315, 0.2746787719726563, 0.2746992492675781, 0.27476275634765623, 0.27472280883789063, 0.274703369140625, 0.2751016845703125, 0.2746326904296875, 0.2748692626953125, 0.27478427124023436, 0.27471563720703124, 0.2749327392578125, 0.274724853515625, 0.2748958740234375, 0.27469619750976565, 0.2748713073730469, 0.2746582946777344, 0.2747832336425781, 0.274766845703125, 0.2748231811523438, 0.2745599975585937, 0.274904052734375, 0.2747361145019531, 0.27491738891601564, 0.2747361145019531, 0.2747484130859375, 0.27478118896484377, 0.2747330627441406, 0.2748784790039063, 0.27495321655273436, 0.2748743591308594, 0.27503717041015624, 0.27474945068359374, 0.2748610534667969, 0.27481396484375, 0.27484161376953126, 0.274904052734375, 0.2748221435546875, 0.27497369384765624, 0.2749921264648437, 0.2749798278808594, 0.2750535583496094, 0.2748856201171875, 0.27496755981445314, 0.27506381225585935, 0.2750412902832031, 0.27498699951171873, 0.27489178466796876, 0.27523379516601565, 0.2749358215332031, 0.2749378662109375, 0.27495013427734377, 0.27487640380859374, 0.27494708251953126, 0.2748856201171875, 0.5855928344726562, 0.27480679321289064, 0.27471768188476564, 0.27463986206054686, 0.2746644592285156, 0.2746173400878906, 0.2747852783203125, 0.27456820678710936, 0.2747197570800781, 0.2749071350097656, 0.2747197570800781, 0.27468185424804686, 0.2749153137207031, 0.274634765625, 0.274661376953125, 0.2746644592285156, 0.27470950317382814, 0.2747084655761719, 0.2747883605957031, 0.2747637634277344, 0.274766845703125, 0.2748252258300781, 0.2747115478515625, 0.2748784790039063, 0.274845703125, 0.2747166748046875, 0.27474945068359374, 0.2747279357910156, 0.27474533081054686, 0.2749153137207031, 0.27497268676757813, 0.27476171875, 0.27479757690429685, 0.2747678833007812, 0.2748170166015625, 0.27481805419921873, 0.27477914428710937, 0.27481292724609374, 0.27478118896484377, 0.2751651916503906, 0.274798583984375, 0.2750699462890625, 0.27489178466796876, 0.27470437622070315, 0.2748518371582031, 0.27496551513671874, 0.27504537963867187, 0.27489279174804687, 0.2750320739746094, 0.27494503784179686, 0.2748374938964844, 0.27502386474609375, 0.27486514282226565, 0.2750382080078125, 0.2750474243164063, 0.27487026977539064, 0.27480679321289064, 0.2749706115722656, 0.2748958740234375, 0.27530035400390623, 0.2749931640625, 0.27500851440429686, 0.2750474243164063, 0.5861458129882813, 0.2749306945800781, 0.2746849365234375, 0.27467572021484377, 0.2746921081542969, 0.27488461303710937, 0.2747361145019531, 0.27468084716796876, 0.2749081726074219, 0.2747637634277344, 0.2749716491699219, 0.27516619873046877, 0.2746982421875, 0.2747187194824219, 0.27472897338867186, 0.2747330627441406, 0.27489279174804687, 0.27473202514648437, 0.2747821960449219, 0.2746531982421875, 0.27479757690429685, 0.2748590087890625, 0.2747268981933594, 0.2749153137207031, 0.2749378662109375, 0.2747852783203125, 0.2748743591308594, 0.2749091796875, 0.2748692626953125, 0.2747084655761719, 0.274798583984375, 0.2747832336425781, 0.2749624328613281, 0.27485797119140626, 0.2748938293457031, 0.2748518371582031, 0.27482827758789063, 0.2749358215332031, 0.27496038818359375, 0.27486309814453125, 0.2749624328613281, 0.27491738891601564, 0.2747637634277344, 0.2748323974609375, 0.2748641357421875, 0.27496038818359375, 0.2748948364257813, 0.27498086547851563, 0.27496038818359375, 0.274882568359375, 0.2749347839355469, 0.2749542541503906, 0.27495321655273436, 0.27502285766601564, 0.27491326904296876, 0.274830322265625, 0.27490609741210936, 0.2749593505859375, 0.27500851440429686, 0.2748876647949219, 0.2749491271972656, 0.27524505615234374, 0.2749306945800781, 0.5859583740234375, 0.2748016662597656, 0.27472384643554687, 0.2748016662597656, 0.27469619750976565, 0.27500543212890627, 0.2750474243164063, 0.27471563720703124, 0.27477297973632814, 0.27522457885742185, 0.27464190673828126, 0.27464703369140625, 0.27475250244140625, 0.2746992492675781, 0.274629638671875, 0.274724853515625, 0.27475045776367185, 0.27475250244140625, 0.2747535400390625, 0.2747340698242188, 0.2750074768066406, 0.27490509033203125, 0.27470437622070315, 0.2745856018066406, 0.27472897338867186, 0.2749440002441406, 0.27473919677734376, 0.27478631591796876, 0.27475045776367185, 0.2747699279785156, 0.274904052734375, 0.2747361145019531, 0.2749224853515625, 0.27496856689453125, 0.27500543212890627, 0.274766845703125, 0.274840576171875, 0.27487640380859374, 0.2749358215332031, 0.2749900817871094, 0.27512628173828124, 0.274840576171875, 0.2750873718261719, 0.27478936767578127, 0.2749931640625, 0.27499725341796877, 0.27488973999023436, 0.27509963989257813, 0.27493374633789064, 0.2749347839355469, 0.27490609741210936, 0.27505255126953126, 0.2748907470703125, 0.2751744079589844, 0.2749296569824219, 0.27491738891601564, 0.2752255859375, 0.2749153137207031, 0.27497268676757813, 0.2749819030761719, 0.27503616333007813, 0.2748958740234375, 0.27490509033203125]",tokens/s,3.580442008064188,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2503,7 +2503,7 @@ OSError: 8 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2631,7 +2631,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2690,7 +2690,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1443.033088,2426.929152,0.0,1780.482048,1621.410304,s,10,0.7064453201293945,0.07064453201293945,0.0003248218262012489,0.07069225692749023,0.07102691879272462,0.07108656425476075,0.07113428062438965,"[0.07067814636230468, 0.07048297882080078, 0.07070636749267578, 0.07082745361328124, 0.07072316741943359, 0.0701509780883789, 0.07005088043212891, 0.07066547393798828, 0.07101366424560547, 0.07114620971679687]",tokens/s,3623.7765713149647,kWh,8.268914728916795e-07,4.5309206087460127e-07,3.6233265194823877e-06,4.903310053248669e-06,tokens/kWh,52209629.25450497,MB,1443.033088,2426.929152,0.0,1780.482048,1723.272192,s,10,38.051747802734376,3.8051747802734375,0.001633156200721129,3.8047684326171876,3.806425,3.8079255126953124,3.8091259228515626,"[3.809426025390625, 3.806091552734375, 3.80473193359375, 3.804946533203125, 3.80474365234375, 3.8032763671875, 3.805872802734375, 3.804793212890625, 3.803895263671875, 3.803970458984375]",tokens/s,16.556401121599166,kWh,4.491550297112361e-05,2.461545215875538e-05,0.0001956961107137206,0.00026522706584359966,tokens/kWh,237532.31895703336,,s,629,38.62667061996455,0.06140965122410906,0.00841844016264897,0.0603873291015625,0.06047764434814453,0.060527001953125,0.13100806091308592,"[0.060862464904785155, 0.06109286499023438, 0.0608983039855957, 0.06086656188964844, 0.060881919860839843, 0.06083071899414062, 0.06078668975830078, 0.06044364929199219, 0.0603422737121582, 0.06038323211669922, 0.06031872177124024, 0.0603955192565918, 0.0603135986328125, 0.060398593902587894, 0.06039449691772461, 0.060418048858642576, 0.06032486343383789, 0.060388351440429686, 0.06033203125, 0.060393470764160156, 0.06033919906616211, 0.0603955192565918, 0.06034022521972656, 0.06041907119750976, 0.06042009735107422, 0.06042009735107422, 0.06048051071166992, 0.06038016128540039, 0.06032998275756836, 0.06038323211669922, 0.060336128234863284, 0.0603955192565918, 0.06038937759399414, 0.060442623138427735, 0.06038323211669922, 0.06042214584350586, 0.06044569778442383, 0.0603770866394043, 0.06039654541015625, 0.0604139518737793, 0.06038425445556641, 0.060472320556640625, 0.0603422737121582, 0.06043648147583008, 0.06041702270507812, 0.06047334289550781, 0.06039039993286133, 0.06040883255004883, 0.0603504638671875, 0.06043545532226562, 0.06041190338134766, 0.060439552307128906, 0.060388351440429686, 0.060398593902587894, 0.06063206481933594, 0.06055936050415039, 0.06030438232421875, 0.06042726516723633, 0.06035660934448242, 0.060483585357666014, 0.06033919906616211, 0.06044364929199219, 0.131093505859375, 0.060385280609130856, 0.06040576171875, 0.06040063858032227, 0.0603873291015625, 0.060388351440429686, 0.06041292953491211, 0.06032793426513672, 0.06038425445556641, 0.06032486343383789, 0.060647422790527344, 0.0603873291015625, 0.06036070251464844, 0.060319744110107425, 0.06039244842529297, 0.06034841537475586, 0.06040371322631836, 0.06038323211669922, 0.06036275100708008, 0.06034636688232422, 0.060508159637451174, 0.060393470764160156, 0.06036479949951172, 0.060565502166748046, 0.06042521667480469, 0.06041190338134766, 0.060416000366210934, 0.06036377716064453, 0.060386302947998044, 0.06051123046875, 0.06042521667480469, 0.06036172866821289, 0.060401664733886716, 0.06039961624145508, 0.060450817108154295, 0.06038425445556641, 0.06045491027832031, 0.06037401580810547, 0.06037401580810547, 0.06039449691772461, 0.060416000366210934, 0.06036275100708008, 0.060416000366210934, 0.06038016128540039, 0.060416000366210934, 0.06041497421264649, 0.060485633850097656, 0.06034431838989258, 0.060418048858642576, 0.06036479949951172, 0.060453887939453124, 0.06043648147583008, 0.060470272064208984, 0.060470272064208984, 0.06043852615356445, 0.06045183944702148, 0.06048153686523437, 0.06035660934448242, 0.06048255920410156, 0.06037606430053711, 0.06047641754150391, 0.060391422271728515, 0.06046515274047851, 0.13133619689941406, 0.06032896041870117, 0.060440574645996094, 0.060365825653076174, 0.060365825653076174, 0.060278785705566405, 0.06043135833740235, 0.060314624786376954, 0.06039654541015625, 0.060281856536865235, 0.060382209777832034, 0.06027571105957031, 0.06045491027832031, 0.06034329605102539, 0.0603955192565918, 0.06031872177124024, 0.060424190521240234, 0.06031769561767578, 0.060407806396484375, 0.060319744110107425, 0.060404735565185545, 0.06034841537475586, 0.0604139518737793, 0.06036787033081055, 0.06044160079956055, 0.06036275100708008, 0.06041907119750976, 0.0602716178894043, 0.06040576171875, 0.06030131149291992, 0.060447742462158206, 0.06032793426513672, 0.060426239013671876, 0.06038425445556641, 0.060459007263183595, 0.06040371322631836, 0.060437503814697265, 0.06034431838989258, 0.060426239013671876, 0.0604139518737793, 0.06040576171875, 0.06034022521972656, 0.060424190521240234, 0.06030950546264648, 0.06039244842529297, 0.06034739303588867, 0.060590080261230465, 0.0603504638671875, 0.060371967315673826, 0.06032281494140625, 0.060440574645996094, 0.06040371322631836, 0.06040576171875, 0.0603770866394043, 0.06047129440307617, 0.06047129440307617, 0.06056447982788086, 0.060295169830322265, 0.06043033599853516, 0.06050099182128906, 0.060493824005126956, 0.06031769561767578, 0.06044467163085938, 0.1314723815917969, 0.06027468872070312, 0.06040371322631836, 0.06033919906616211, 0.06039961624145508, 0.060336128234863284, 0.06038937759399414, 0.0603054084777832, 0.06041702270507812, 0.06029619216918945, 0.060365825653076174, 0.06028902435302735, 0.06035456085205078, 0.06032691192626953, 0.060402687072753904, 0.06038016128540039, 0.060593151092529295, 0.06032179260253906, 0.06037913513183594, 0.06032076644897461, 0.060369918823242184, 0.06031872177124024, 0.060401664733886716, 0.06030643081665039, 0.06041292953491211, 0.060353534698486325, 0.0603955192565918, 0.06035660934448242, 0.060477439880371096, 0.060314624786376954, 0.06036377716064453, 0.06034739303588867, 0.06039961624145508, 0.0603770866394043, 0.06041190338134766, 0.06040678405761719, 0.060439552307128906, 0.06037299346923828, 0.06051737594604492, 0.06039756774902344, 0.060401664733886716, 0.060365825653076174, 0.06041702270507812, 0.06029107284545898, 0.06039756774902344, 0.06039961624145508, 0.060440574645996094, 0.06035968017578125, 0.06040371322631836, 0.06081024169921875, 0.06043545532226562, 0.06036377716064453, 0.06042214584350586, 0.060398593902587894, 0.06039756774902344, 0.06041088104248047, 0.06047641754150391, 0.06032998275756836, 0.060508159637451174, 0.06032793426513672, 0.060478462219238284, 0.06045286560058594, 0.060424190521240234, 0.13134439086914063, 0.060319744110107425, 0.06065868759155273, 0.06037299346923828, 0.06037913513183594, 0.06034534454345703, 0.06036070251464844, 0.060300289154052736, 0.060336128234863284, 0.06033203125, 0.06041907119750976, 0.06032998275756836, 0.06041292953491211, 0.06030438232421875, 0.060385280609130856, 0.06034124755859375, 0.06038118362426758, 0.06038323211669922, 0.06038016128540039, 0.06030643081665039, 0.06037299346923828, 0.06034841537475586, 0.060352512359619144, 0.06032793426513672, 0.06032896041870117, 0.0603770866394043, 0.06038016128540039, 0.060447742462158206, 0.060442623138427735, 0.06034124755859375, 0.06038323211669922, 0.06034534454345703, 0.060368896484375, 0.06038323211669922, 0.06043033599853516, 0.060409854888916016, 0.060450817108154295, 0.06035660934448242, 0.06036275100708008, 0.06036787033081055, 0.060416000366210934, 0.060407806396484375, 0.06048153686523437, 0.06033817672729492, 0.06041088104248047, 0.06042521667480469, 0.06047334289550781, 0.060470272064208984, 0.060386302947998044, 0.0603238410949707, 0.0603873291015625, 0.060407806396484375, 0.060401664733886716, 0.06037913513183594, 0.06041292953491211, 0.06043033599853516, 0.060494846343994144, 0.060382209777832034, 0.06050611114501953, 0.06032691192626953, 0.06048767852783203, 0.06034124755859375, 0.06044672012329102, 0.13078834533691405, 0.06037401580810547, 0.06045798492431641, 0.06032896041870117, 0.060333057403564455, 0.0602716178894043, 0.06035660934448242, 0.06029414367675781, 0.06038937759399414, 0.06027571105957031, 0.060371967315673826, 0.06027775955200195, 0.060355583190917966, 0.06030745697021484, 0.06040576171875, 0.060273662567138675, 0.06036275100708008, 0.06027264022827149, 0.06040371322631836, 0.060298240661621094, 0.06044364929199219, 0.060303359985351565, 0.06036479949951172, 0.0603422737121582, 0.06037094497680664, 0.06032691192626953, 0.06041292953491211, 0.06036377716064453, 0.06040063858032227, 0.060319744110107425, 0.06039654541015625, 0.06034636688232422, 0.06046515274047851, 0.06034124755859375, 0.06042521667480469, 0.060331008911132813, 0.060421119689941405, 0.060319744110107425, 0.060369918823242184, 0.06033919906616211, 0.06044876861572265, 0.06032998275756836, 0.06040883255004883, 0.06030745697021484, 0.060352512359619144, 0.06034534454345703, 0.06046105575561524, 0.06034022521972656, 0.06038425445556641, 0.06030233764648438, 0.0603955192565918, 0.06034022521972656, 0.060432384490966794, 0.060401664733886716, 0.06036684799194336, 0.060398593902587894, 0.06045695877075195, 0.06030643081665039, 0.06044569778442383, 0.060319744110107425, 0.060483585357666014, 0.06034329605102539, 0.060464126586914066, 0.13067263793945313, 0.06030643081665039, 0.060472320556640625, 0.06033817672729492, 0.06035763168334961, 0.06032998275756836, 0.06036377716064453, 0.06032076644897461, 0.060353534698486325, 0.060290046691894535, 0.06033817672729492, 0.06061875152587891, 0.06047334289550781, 0.060336128234863284, 0.06037299346923828, 0.060353534698486325, 0.06040678405761719, 0.06029414367675781, 0.06038016128540039, 0.060298240661621094, 0.060369918823242184, 0.06039449691772461, 0.060434432983398435, 0.0603351058959961, 0.060401664733886716, 0.060386302947998044, 0.060424190521240234, 0.0603351058959961, 0.060398593902587894, 0.06045798492431641, 0.060388351440429686, 0.060486656188964844, 0.06044467163085938, 0.06045695877075195, 0.06081228637695312, 0.06042316818237305, 0.06042009735107422, 0.06036684799194336, 0.06051430511474609, 0.06054912185668945, 0.06043340682983398, 0.06039039993286133, 0.06043135833740235, 0.0603504638671875, 0.060386302947998044, 0.06043545532226562, 0.06046105575561524, 0.06042521667480469, 0.06042828750610352, 0.0603422737121582, 0.060382209777832034, 0.060402687072753904, 0.060439552307128906, 0.060388351440429686, 0.0604846076965332, 0.06035148620605469, 0.06046515274047851, 0.06030950546264648, 0.06046105575561524, 0.060349441528320315, 0.06053273773193359, 0.060386302947998044, 0.06046105575561524, 0.13130650329589844, 0.06029107284545898, 0.060402687072753904, 0.060393470764160156, 0.06034841537475586, 0.06032896041870117, 0.060375038146972655, 0.06038118362426758, 0.06033919906616211, 0.06034534454345703, 0.06034124755859375, 0.060286975860595705, 0.060375038146972655, 0.060331008911132813, 0.060398593902587894, 0.06032179260253906, 0.06035456085205078, 0.060369918823242184, 0.06035865783691406, 0.06031257629394531, 0.060353534698486325, 0.0603238410949707, 0.06037913513183594, 0.06042009735107422, 0.060404735565185545, 0.06040371322631836, 0.060402687072753904, 0.060303359985351565, 0.060388351440429686, 0.06038425445556641, 0.06051737594604492, 0.06032998275756836, 0.06035968017578125, 0.0603770866394043, 0.06051123046875, 0.06038118362426758, 0.06041497421264649, 0.060391422271728515, 0.060371967315673826, 0.06039449691772461, 0.06054297637939453, 0.06048972702026367, 0.060459007263183595, 0.06039449691772461, 0.06039449691772461, 0.06041497421264649, 0.06044979095458984, 0.060393470764160156, 0.06040576171875, 0.060398593902587894, 0.060398593902587894, 0.06036070251464844, 0.06041702270507812, 0.060416000366210934, 0.06040063858032227, 0.060398593902587894, 0.06043852615356445, 0.06038937759399414, 0.060440574645996094, 0.060385280609130856, 0.0604846076965332, 0.06034022521972656, 0.060485633850097656, 0.13168946838378906, 0.06034329605102539, 0.060402687072753904, 0.06033817672729492, 0.06036070251464844, 0.0602716178894043, 0.060365825653076174, 0.060290046691894535, 0.06037401580810547, 0.060283905029296876, 0.06038323211669922, 0.060278785705566405, 0.060388351440429686, 0.060284927368164064, 0.0603873291015625, 0.06031872177124024, 0.06036479949951172, 0.06031769561767578, 0.060355583190917966, 0.06030131149291992, 0.06037094497680664, 0.060290046691894535, 0.060386302947998044, 0.0603054084777832, 0.06042726516723633, 0.06039961624145508, 0.06043852615356445, 0.06031155014038086, 0.06038016128540039, 0.06030233764648438, 0.060385280609130856, 0.060349441528320315, 0.06043545532226562, 0.06037094497680664, 0.060391422271728515, 0.060353534698486325, 0.060434432983398435, 0.060365825653076174, 0.06041292953491211, 0.06035148620605469, 0.06042931365966797, 0.06034534454345703, 0.06041702270507812, 0.060352512359619144, 0.06036479949951172, 0.06036787033081055, 0.06043545532226562, 0.06035968017578125, 0.06058803176879883, 0.06035763168334961, 0.06043852615356445, 0.06034739303588867, 0.06043340682983398, 0.06037299346923828, 0.06035148620605469, 0.06041292953491211, 0.06050406265258789, 0.06034636688232422, 0.06049280166625977, 0.06036787033081055, 0.06051839828491211, 0.06036479949951172, 0.06042521667480469, 0.1317969970703125, 0.06032691192626953, 0.060418048858642576, 0.06032998275756836, 0.060472320556640625, 0.06033407974243164, 0.060442623138427735, 0.06029414367675781, 0.06045183944702148, 0.060319744110107425, 0.060365825653076174, 0.060295169830322265, 0.06056140899658203, 0.06048153686523437, 0.060404735565185545, 0.06032486343383789, 0.0603873291015625, 0.06028595352172852, 0.060391422271728515, 0.060333057403564455, 0.06036070251464844, 0.06033203125, 0.06034431838989258, 0.06033407974243164, 0.060333057403564455, 0.06036787033081055, 0.06043340682983398, 0.0603955192565918, 0.060391422271728515, 0.060278785705566405, 0.06038937759399414, 0.06032179260253906, 0.06048767852783203, 0.06038323211669922, 0.06038937759399414, 0.060386302947998044, 0.060402687072753904, 0.060371967315673826, 0.06035763168334961, 0.060353534698486325, 0.0603955192565918, 0.06033203125, 0.060382209777832034, 0.060314624786376954, 0.06036377716064453, 0.06034534454345703, 0.060485633850097656, 0.06040063858032227, 0.060355583190917966, 0.06032691192626953, 0.06041088104248047, 0.0603422737121582, 0.060416000366210934, 0.060375038146972655, 0.06036172866821289, 0.060353534698486325, 0.06043852615356445, 0.060333057403564455, 0.060437503814697265, 0.060333057403564455, 0.060472320556640625, 0.06032076644897461, 0.06042828750610352]",tokens/s,16.284085319921278,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2732,7 +2732,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3151,7 +3151,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2508.439552,9028.763648,0.0,8382.316544,7939.283456,s,10,6.160718749999999,0.616071875,0.0005587926705172899,0.6159664306640624,0.6168901062011719,0.6170025787353516,0.6170925567626954,"[0.6156204223632813, 0.6152601318359375, 0.616293212890625, 0.616349365234375, 0.6168651123046875, 0.6156136474609375, 0.6171150512695313, 0.6156689453125, 0.6160707397460937, 0.6158621215820312]",tokens/s,415.5359307710646,kWh,7.268102301491631e-06,3.982613158030504e-06,3.636417941812856e-05,4.7614894877650695e-05,tokens/kWh,5376468.868781654,MB,2508.439552,9028.763648,0.0,8382.316544,8233.862144,s,10,338.1393203125,33.813932031250005,0.0010795636913464568,33.813869140625,33.815248046875,33.8156611328125,33.8159916015625,"[33.814296875, 33.8134296875, 33.81288671875, 33.81338671875, 33.81515625, 33.81607421875, 33.81354296875, 33.8141953125, 33.81207421875, 33.81427734375]",tokens/s,1.8631373583461677,kWh,0.0003992101543976201,0.00021880183463568006,0.001996706808148448,0.002614718797181748,tokens/kWh,24094.36917954772,,s,629,343.1469031372071,0.5455435661958776,0.07325723275974945,0.5367101440429688,0.5369858032226562,0.5370556152343751,1.1530929980468752,"[0.5367275390625, 0.5368402099609375, 0.5365862426757813, 0.5366046752929687, 0.536500244140625, 0.5371934814453125, 0.536384521484375, 0.5366179809570313, 0.536248291015625, 0.5367777099609375, 0.5363753051757812, 0.5365616455078125, 0.5364766845703125, 0.536511474609375, 0.536501220703125, 0.5365964965820312, 0.5364152221679688, 0.5368176879882812, 0.5364592895507813, 0.53726513671875, 0.5365821533203124, 0.5365892944335937, 0.5366384887695312, 0.5367306518554688, 0.5367725830078125, 0.5367736206054687, 0.536537109375, 0.5368207397460938, 0.5368606567382812, 0.536796142578125, 0.5370715942382812, 0.536573974609375, 0.5366773681640625, 0.5369937744140625, 0.5366261596679688, 0.5366814575195312, 0.536542236328125, 0.536753173828125, 0.5365514526367188, 0.5366978759765625, 0.5366343383789063, 0.5366548461914062, 0.536574951171875, 0.536732666015625, 0.5368053588867188, 0.5372456665039063, 0.5365043334960937, 0.536658935546875, 0.5365442504882812, 0.5368237915039062, 0.5367091064453124, 0.5369200439453125, 0.5367900390625, 0.536974365234375, 0.536680419921875, 0.5367030029296875, 0.536838134765625, 0.5368370971679688, 0.5364818115234375, 0.5369077758789063, 0.5365545043945312, 0.537333740234375, 1.1534786376953126, 0.5365780639648438, 0.53663232421875, 0.5365493774414063, 0.53699072265625, 0.5363886108398438, 0.53667431640625, 0.5364254760742188, 0.5368883056640625, 0.536364013671875, 0.5366353759765625, 0.5364930419921875, 0.536753173828125, 0.5364725952148437, 0.5367275390625, 0.5364439086914062, 0.5366087646484375, 0.5365186767578125, 0.5367808227539063, 0.5370828857421875, 0.5370153198242188, 0.5366220703125, 0.536890380859375, 0.536553466796875, 0.5365698852539063, 0.5367623901367188, 0.536953857421875, 0.5364602661132812, 0.5366998901367187, 0.5367900390625, 0.5366876220703125, 0.536721435546875, 0.5366661376953125, 0.5365401611328126, 0.5368104858398437, 0.5369794311523437, 0.5364541625976562, 0.5367470092773438, 0.5366578979492187, 0.5365872802734375, 0.5365186767578125, 0.5366548461914062, 0.5367060546875, 0.536869873046875, 0.5368473510742188, 0.5368237915039062, 0.5370029907226562, 0.5366814575195312, 0.5366640625, 0.536711181640625, 0.5369948120117187, 0.5366640625, 0.53665380859375, 0.5369343872070312, 0.5368289184570313, 0.5369354248046875, 0.5367367553710938, 0.5368514404296875, 0.5366773681640625, 0.5364838256835938, 0.5370531616210937, 0.536764404296875, 0.5369661254882813, 1.153585205078125, 0.5365933837890625, 0.536732666015625, 0.5365339965820313, 0.5364029541015625, 0.5366087646484375, 0.536580078125, 0.536395751953125, 0.5366435546875, 0.536427490234375, 0.5366814575195312, 0.5365401611328126, 0.536468505859375, 0.5364285278320312, 0.5363650512695313, 0.5365278930664062, 0.5366292724609375, 0.536468505859375, 0.53648486328125, 0.5365299072265625, 0.536774658203125, 0.5366128540039062, 0.5364951171875, 0.5365032958984375, 0.5370203857421875, 0.5368985595703125, 0.5366384887695312, 0.53654833984375, 0.5365933837890625, 0.5366578979492187, 0.5368760375976562, 0.5367705688476563, 0.5369251708984375, 0.5367439575195313, 0.5370286254882812, 0.536648681640625, 0.5368391723632813, 0.5365596313476563, 0.5367285766601563, 0.5368002319335937, 0.5367418823242187, 0.536711181640625, 0.5370572509765625, 0.5369292602539063, 0.5370091552734375, 0.5367725830078125, 0.5368463134765625, 0.5368094482421875, 0.5370818481445313, 0.5366753540039062, 0.536648681640625, 0.5365473022460937, 0.536875, 0.5368524780273437, 0.5368463134765625, 0.5367808227539063, 0.5367828369140625, 0.5369835815429688, 0.536774658203125, 0.5366026000976563, 0.5368453369140626, 0.5368463134765625, 0.5369210815429688, 1.1540050048828125, 0.5364725952148437, 0.53673779296875, 0.5366005859375, 0.5364121704101562, 0.536585205078125, 0.5365657348632813, 0.536479736328125, 0.5365575561523438, 0.5364387817382813, 0.53650537109375, 0.5364961547851562, 0.5364869384765625, 0.5364664306640625, 0.53623193359375, 0.5365073852539063, 0.536711181640625, 0.5362216796875, 0.5367019653320313, 0.5367122192382813, 0.5370501098632813, 0.5365514526367188, 0.5365924072265625, 0.536627197265625, 0.5366241455078125, 0.5367828369140625, 0.536974365234375, 0.5365841674804688, 0.5367449340820313, 0.5366261596679688, 0.5368453369140626, 0.5365964965820312, 0.5367470092773438, 0.5365155639648438, 0.5371627807617188, 0.5366190185546875, 0.536848388671875, 0.5366978759765625, 0.5367684936523438, 0.5366149291992187, 0.53686376953125, 0.5367787475585938, 0.536605712890625, 0.536658935546875, 0.537101318359375, 0.5369169921875, 0.5370439453125, 0.5366231079101562, 0.5367838745117187, 0.5364930419921875, 0.5368565673828125, 0.5365831909179688, 0.53692724609375, 0.5368319702148437, 0.536985595703125, 0.5367777099609375, 0.5369139404296875, 0.5370009765625, 0.53667431640625, 0.5366384887695312, 0.537037841796875, 0.53671630859375, 0.5374013671875, 1.1540633544921874, 0.5367183227539063, 0.536574951171875, 0.536537109375, 0.5365288696289062, 0.5366630249023437, 0.5365299072265625, 0.5365872802734375, 0.536580078125, 0.5363568725585938, 0.5365626831054687, 0.5364346923828125, 0.5366302490234375, 0.5368023071289062, 0.5364418334960938, 0.5364080810546875, 0.5367490844726562, 0.5365657348632813, 0.5367930908203125, 0.5366405029296875, 0.5367767333984375, 0.5366978759765625, 0.5365657348632813, 0.536595458984375, 0.5366210327148437, 0.5367633666992188, 0.536784912109375, 0.5366651000976562, 0.5367859497070312, 0.5371719970703125, 0.5367050170898438, 0.5367664794921875, 0.5368033447265625, 0.536553466796875, 0.5369395141601563, 0.536732666015625, 0.53688525390625, 0.5366098022460938, 0.5369927978515625, 0.53686474609375, 0.536637451171875, 0.5367244873046875, 0.5367367553710938, 0.536932373046875, 0.5369876708984375, 0.5368678588867187, 0.5369866333007812, 0.5367203979492188, 0.536911865234375, 0.5367930908203125, 0.536795166015625, 0.53676953125, 0.5369784545898437, 0.5369251708984375, 0.5369395141601563, 0.5367275390625, 0.5368350830078125, 0.5369600219726562, 0.5368616943359374, 0.5367131958007813, 0.5370685424804688, 0.5368555297851563, 0.5369569091796875, 1.1528519287109376, 0.5366435546875, 0.5369937744140625, 0.5363619995117187, 0.5365892944335937, 0.5367408447265625, 0.5365821533203124, 0.5365094604492188, 0.5366405029296875, 0.5366814575195312, 0.5365616455078125, 0.5363660888671875, 0.5365811157226562, 0.536352783203125, 0.5364869384765625, 0.5365616455078125, 0.536542236328125, 0.5364357299804687, 0.5367992553710937, 0.5366364135742188, 0.5370408935546875, 0.5365309448242187, 0.536680419921875, 0.53680126953125, 0.5368299560546875, 0.53682177734375, 0.5368463134765625, 0.536953857421875, 0.536647705078125, 0.5366937866210938, 0.536774658203125, 0.5366599731445313, 0.5367675170898437, 0.5365504150390625, 0.5369415893554688, 0.5367234497070312, 0.5368084716796875, 0.5367572631835937, 0.5370081176757813, 0.5367817993164062, 0.5370224609375, 0.5365892944335937, 0.5371392211914062, 0.5368504028320312, 0.5367367553710938, 0.536853515625, 0.5368709106445313, 0.5366343383789063, 0.5370214233398437, 0.53682177734375, 0.5368330078125, 0.5367500610351562, 0.5370347290039063, 0.5369180297851562, 0.5368719482421875, 0.5372518310546875, 0.5368914184570313, 0.53684326171875, 0.5368125610351563, 0.5365391235351562, 0.5370531616210937, 0.5368402099609375, 0.5369620361328125, 1.1546163330078125, 0.536658935546875, 0.5366415405273437, 0.536447998046875, 0.5365227661132812, 0.536553466796875, 0.536580078125, 0.5364224243164063, 0.5365206909179687, 0.5364510498046875, 0.5366732788085937, 0.5365524291992188, 0.5365759887695313, 0.5364756469726563, 0.5366435546875, 0.5366773681640625, 0.5367859497070312, 0.5364930419921875, 0.5368688354492187, 0.536500244140625, 0.5367050170898438, 0.5366671142578125, 0.53652685546875, 0.5368053588867188, 0.5367131958007813, 0.5365862426757813, 0.536685546875, 0.5364827880859375, 0.5367152709960937, 0.536784912109375, 0.536742919921875, 0.5370839233398438, 0.5367019653320313, 0.5365166015625, 0.5370787963867187, 0.536642578125, 0.5367982177734375, 0.5366282348632813, 0.5368688354492187, 0.5366445922851563, 0.5366927490234376, 0.5366364135742188, 0.536685546875, 0.5366190185546875, 0.5369262084960937, 0.536896484375, 0.5369661254882813, 0.5370511474609375, 0.5368923950195312, 0.5365504150390625, 0.5368422241210937, 0.5365524291992188, 0.5370234985351563, 0.5368914184570313, 0.536858642578125, 0.5367828369140625, 0.537111572265625, 0.5368043823242188, 0.5368790893554688, 0.5366845703125, 0.5368555297851563, 0.5367388305664063, 0.5367633666992188, 1.15302294921875, 0.5366507568359375, 0.5366722412109375, 0.5365514526367188, 0.5365708618164062, 0.5365339965820313, 0.5365176391601563, 0.536474609375, 0.53650634765625, 0.536406005859375, 0.5365780639648438, 0.5364827880859375, 0.5365155639648438, 0.5366302490234375, 0.5365411987304688, 0.53661181640625, 0.53669580078125, 0.5366046752929687, 0.5367203979492188, 0.5367101440429688, 0.5368402099609375, 0.5367101440429688, 0.5366998901367187, 0.53671630859375, 0.536595458984375, 0.5367562255859375, 0.5368985595703125, 0.5366599731445313, 0.5367234497070312, 0.5370531616210937, 0.5366917114257812, 0.536722412109375, 0.5367152709960937, 0.5366405029296875, 0.5371094970703125, 0.5365759887695313, 0.5367122192382813, 0.536563720703125, 0.536896484375, 0.5368258666992187, 0.5366241455078125, 0.5366507568359375, 0.5368135375976563, 0.5368985595703125, 0.5373378295898438, 0.5368923950195312, 0.5369620361328125, 0.53663232421875, 0.5367705688476563, 0.5364285278320312, 0.5368125610351563, 0.5366394653320312, 0.5368565673828125, 0.5369262084960937, 0.5369302978515625, 0.53675927734375, 0.536806396484375, 0.53671630859375, 0.5367633666992188, 0.5365729370117187, 0.5371934814453125, 0.5368186645507812, 0.5369548950195312, 1.1533997802734375, 0.53671630859375, 0.53673779296875, 0.5364172973632813, 0.536574951171875, 0.5363138427734375, 0.5363886108398438, 0.5361704711914063, 0.5366302490234375, 0.5365309448242187, 0.5365596313476563, 0.5366067504882812, 0.5365155639648438, 0.5364132080078124, 0.536447998046875, 0.5364818115234375, 0.5365022583007812, 0.5364735717773438, 0.5368237915039062, 0.536616943359375, 0.5367705688476563, 0.536542236328125, 0.5366087646484375, 0.536531982421875, 0.5365647583007812, 0.5367039794921875, 0.5368289184570313, 0.5365718994140625, 0.5365913696289063, 0.5367357177734375, 0.5367091064453124, 0.53696923828125, 0.5366077270507813, 0.53646337890625, 0.5369139404296875, 0.5366016235351563, 0.5370368041992187, 0.5366947631835938, 0.5368391723632813, 0.5366251220703125, 0.5368023071289062, 0.5366968383789062, 0.5369682006835937, 0.5367828369140625, 0.536826904296875, 0.5367490844726562, 0.5370368041992187, 0.5367285766601563, 0.5367756958007812, 0.5366456298828125, 0.5367992553710937, 0.5366619873046875, 0.536922119140625, 0.5367510986328125, 0.5369384765625, 0.5367971801757813, 0.5368422241210937, 0.5369241333007813, 0.5370265502929688, 0.5366046752929687, 0.5369190673828125, 0.5368411865234375, 0.5369835815429688, 1.1531202392578126, 0.53661083984375, 0.5367684936523438, 0.5365657348632813, 0.5369896850585938, 0.5365094604492188, 0.53654833984375, 0.5365493774414063, 0.5365933837890625, 0.5366210327148437, 0.5365616455078125, 0.5365084228515625, 0.5366517944335938, 0.5366220703125, 0.5364132080078124, 0.5364838256835938, 0.5366138916015625, 0.5365933837890625, 0.53676953125, 0.536748046875, 0.5367439575195313, 0.5366619873046875, 0.5366067504882812, 0.5366651000976562, 0.5365985107421875, 0.5369415893554688, 0.5369354248046875, 0.5366578979492187, 0.5367019653320313, 0.53654833984375, 0.5365944213867188, 0.5366261596679688, 0.5367490844726562, 0.5364592895507813, 0.5370839233398438, 0.5365892944335937, 0.536669189453125, 0.536764404296875, 0.5369815063476563, 0.5365975341796875, 0.5367654418945312, 0.536732666015625, 0.5365339965820313, 0.5369415893554688, 0.5369088134765625, 0.53686376953125, 0.5369405517578125, 0.536875, 0.536875, 0.53671728515625, 0.5368309936523438, 0.53659033203125, 0.5371371459960937, 0.5367890014648438, 0.5368709106445313, 0.5366651000976562, 0.536742919921875, 0.5368914184570313, 0.5369149169921875, 0.5366353759765625, 0.537017333984375, 0.536827880859375, 0.5368944702148437]",tokens/s,1.8330341735548021,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3180,7 +3180,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3222,7 +3222,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 96.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3293,7 +3293,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3365,7 +3365,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3423,7 +3423,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3480,7 +3480,7 @@ ChildProcessError: Traceback (most recent call last): RuntimeError: The size of tensor a (60) must match the size of tensor b (32) at non-singleton dimension 2 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3509,7 +3509,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3571,7 +3571,7 @@ ValueError: OPTForCausalLM does not support an attention implementation through ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1712.406528,5964.824576,0.0,5318.377472,5048.328704,s,10,3.4079426879882817,0.34079426879882807,0.0009322383785557938,0.3405218048095703,0.3422028137207031,0.34224659423828124,0.34228161865234374,"[0.33942413330078125, 0.33999859619140627, 0.3406260070800781, 0.3401291809082031, 0.3404176025390625, 0.3401330871582031, 0.34098886108398435, 0.34174176025390623, 0.3421930847167969, 0.3422903747558594]",tokens/s,751.1863415494162,kWh,4.007971358520013e-06,2.195722103369917e-06,2.097056307273452e-05,2.717425653462445e-05,tokens/kWh,9420680.918126099,MB,1712.406528,5964.824576,0.0,5318.377472,5098.333184,s,10,188.29865429687501,18.8298654296875,0.0006673439432085487,18.82976953125,18.8306990234375,18.83094423828125,18.831140410156248,"[18.829521484375, 18.829875, 18.82865625, 18.829970703125, 18.8296640625, 18.83021484375, 18.83064453125, 18.829525390625, 18.831189453125, 18.829392578125]",tokens/s,3.3457488177622916,kWh,0.0002223270491041519,0.00012185299643231701,0.0011337806940608672,0.0014779607395973358,tokens/kWh,42626.30143826695,,s,629,191.06832772827136,0.3037652269129912,0.04052678174686376,0.29887591552734377,0.29901171264648435,0.2990915771484375,0.6395446215820312,"[0.29933056640625, 0.2987888488769531, 0.2987939758300781, 0.2988922729492188, 0.29877658081054687, 0.2988226623535156, 0.2988328857421875, 0.29880218505859374, 0.29885440063476565, 0.2987530212402344, 0.2989803466796875, 0.298977294921875, 0.2988349304199219, 0.2987745361328125, 0.29902029418945314, 0.29885336303710935, 0.2988748779296875, 0.2988410949707031, 0.29882571411132813, 0.2988380126953125, 0.2986956787109375, 0.2987386779785156, 0.29879092407226565, 0.29889739990234376, 0.298745849609375, 0.29895782470703125, 0.2990745544433594, 0.2989486083984375, 0.2988134460449219, 0.2987888488769531, 0.29889535522460936, 0.2988922729492188, 0.29888717651367186, 0.298935302734375, 0.29891070556640625, 0.29895269775390626, 0.2988277893066406, 0.2988584899902344, 0.29889535522460936, 0.29889739990234376, 0.2988717956542969, 0.2989383544921875, 0.2988564453125, 0.29891787719726565, 0.29884210205078127, 0.2988062744140625, 0.2989690856933594, 0.29877862548828127, 0.29881036376953124, 0.29896807861328123, 0.29892098999023436, 0.29895574951171877, 0.29893939208984377, 0.2987796630859375, 0.29892095947265623, 0.2987734985351562, 0.29884722900390626, 0.2990111083984375, 0.2989178466796875, 0.2988789672851562, 0.2989742431640625, 0.2987898559570312, 0.6390538330078125, 0.29874993896484375, 0.2987274169921875, 0.2988769226074219, 0.2988164978027344, 0.29890048217773435, 0.29877658081054687, 0.29891278076171873, 0.2988912658691406, 0.29923736572265625, 0.29886669921875, 0.29883392333984377, 0.29876019287109373, 0.29876223754882814, 0.2988769226074219, 0.29878170776367186, 0.2988779602050781, 0.2988216247558594, 0.2988441467285156, 0.29885952758789064, 0.2987755432128906, 0.29880935668945313, 0.29883392333984377, 0.2988554382324219, 0.29924658203125, 0.2988380126953125, 0.29887387084960937, 0.2988840942382813, 0.2993029174804688, 0.2988584899902344, 0.29871514892578127, 0.2991124572753906, 0.298829833984375, 0.29886053466796875, 0.29897317504882814, 0.2988963928222656, 0.29880523681640625, 0.29894247436523436, 0.29871514892578127, 0.29875201416015623, 0.29894451904296876, 0.29874789428710935, 0.29877658081054687, 0.29893017578125, 0.29889434814453125, 0.2990643310546875, 0.29894451904296876, 0.29872845458984376, 0.2989383544921875, 0.29896295166015624, 0.29883187866210936, 0.2988687438964844, 0.29901412963867186, 0.29885748291015624, 0.2988984375, 0.2988963928222656, 0.29887387084960937, 0.29880935668945313, 0.29898651123046877, 0.2988912658691406, 0.2989486083984375, 0.29897830200195313, 0.2988584899902344, 0.6398330688476562, 0.298903564453125, 0.29880218505859374, 0.2988216247558594, 0.2987315063476563, 0.298777587890625, 0.29872332763671877, 0.2986977233886719, 0.2987960205078125, 0.29879092407226565, 0.298724365234375, 0.29893121337890627, 0.2992732238769531, 0.298814453125, 0.29890866088867185, 0.2988011474609375, 0.2988769226074219, 0.2987274169921875, 0.2989219970703125, 0.29873458862304686, 0.29885235595703125, 0.2989690856933594, 0.2988492736816406, 0.2988031921386719, 0.298861572265625, 0.2988554382324219, 0.2988062744140625, 0.2989547424316406, 0.29884622192382815, 0.2987438049316406, 0.2989117431640625, 0.2988840942382813, 0.29899468994140627, 0.2988564453125, 0.298814453125, 0.298819580078125, 0.29902542114257813, 0.2989270935058594, 0.2990540771484375, 0.2987673645019531, 0.298851318359375, 0.29884518432617185, 0.29885235595703125, 0.2987008056640625, 0.29885336303710935, 0.29878170776367186, 0.2989045715332031, 0.2989260864257812, 0.2990172119140625, 0.298829833984375, 0.29888101196289063, 0.2989363098144531, 0.2988584899902344, 0.2988636169433594, 0.29881753540039063, 0.2989066162109375, 0.29882571411132813, 0.29906329345703125, 0.29887387084960937, 0.2989240417480469, 0.2989045715332031, 0.29904177856445313, 0.29881857299804687, 0.6396661987304687, 0.29890866088867185, 0.29881549072265623, 0.29886465454101563, 0.29882366943359373, 0.2986588134765625, 0.2989281311035156, 0.29879806518554686, 0.2988359680175781, 0.29883905029296876, 0.2988031921386719, 0.2987530212402344, 0.2989854736328125, 0.299030517578125, 0.2989547424316406, 0.29914111328125, 0.2987806701660156, 0.29885952758789064, 0.298814453125, 0.2987591552734375, 0.2988308410644531, 0.29881857299804687, 0.2988554382324219, 0.2988011474609375, 0.2988554382324219, 0.29885440063476565, 0.29888204956054687, 0.29879910278320315, 0.2987745361328125, 0.29876837158203123, 0.29885235595703125, 0.29886770629882814, 0.2988062744140625, 0.29895065307617186, 0.29894552612304687, 0.29887591552734377, 0.29902847290039064, 0.298745849609375, 0.29884722900390626, 0.29874176025390625, 0.29884005737304686, 0.29894247436523436, 0.29886257934570315, 0.2987888488769531, 0.2990315551757812, 0.29888717651367186, 0.29897216796875, 0.29883905029296876, 0.2989117431640625, 0.2990663757324219, 0.2989895629882813, 0.29887591552734377, 0.298925048828125, 0.2989056091308594, 0.29893939208984377, 0.29892095947265623, 0.29896395874023435, 0.2988687438964844, 0.2989609069824219, 0.29890765380859374, 0.2990243835449219, 0.2988287963867188, 0.29879092407226565, 0.6396928100585938, 0.29893939208984377, 0.29887283325195313, 0.29885440063476565, 0.2989383544921875, 0.29895269775390626, 0.2988441467285156, 0.2988482666015625, 0.2989619140625, 0.29886257934570315, 0.29881549072265623, 0.2989117431640625, 0.2988072814941406, 0.29882571411132813, 0.2989240417480469, 0.29879296875, 0.2987960205078125, 0.29870181274414065, 0.29926913452148435, 0.29865475463867186, 0.2987703857421875, 0.29870181274414065, 0.2988308410644531, 0.29877658081054687, 0.2987857971191406, 0.29879193115234376, 0.298919921875, 0.2987591552734375, 0.2987806701660156, 0.298919921875, 0.2987694091796875, 0.29903359985351563, 0.29880523681640625, 0.2990408020019531, 0.2988523254394531, 0.29895782470703125, 0.2988554382324219, 0.2988769226074219, 0.2992046203613281, 0.298925048828125, 0.29894247436523436, 0.29885336303710935, 0.2989486083984375, 0.29879296875, 0.29882571411132813, 0.2989240417480469, 0.29892913818359373, 0.29900799560546876, 0.29885952758789064, 0.29900390625, 0.29886257934570315, 0.2988062744140625, 0.29906842041015624, 0.29885336303710935, 0.29888101196289063, 0.2989056091308594, 0.2989219970703125, 0.2988503112792969, 0.2989137878417969, 0.29884005737304686, 0.2988789672851562, 0.29902029418945314, 0.29878988647460936, 0.6392319946289062, 0.29863629150390625, 0.2988984375, 0.29874688720703124, 0.29883392333984377, 0.2988584899902344, 0.29881857299804687, 0.29880218505859374, 0.29896600341796875, 0.29875506591796874, 0.2988287963867188, 0.2988226623535156, 0.2989434814453125, 0.29880218505859374, 0.2989854736328125, 0.2987653198242187, 0.29884005737304686, 0.29883187866210936, 0.2989383544921875, 0.29896295166015624, 0.298893310546875, 0.2992261047363281, 0.29879296875, 0.29881036376953124, 0.2989045715332031, 0.29901412963867186, 0.2989537353515625, 0.29883905029296876, 0.2987694091796875, 0.2988349304199219, 0.29899468994140627, 0.2988861389160156, 0.29891787719726565, 0.29884518432617185, 0.2990172119140625, 0.2990745544433594, 0.29899264526367186, 0.29902029418945314, 0.29891278076171873, 0.2989342651367187, 0.2989270935058594, 0.2989557800292969, 0.2988441467285156, 0.29887081909179686, 0.2988656311035156, 0.29873358154296875, 0.29878680419921877, 0.2989465637207031, 0.2988912658691406, 0.298819580078125, 0.29887283325195313, 0.29891583251953124, 0.29893121337890627, 0.29889022827148437, 0.2988380126953125, 0.299009033203125, 0.2988410949707031, 0.2991124572753906, 0.29884622192382815, 0.2988584899902344, 0.29896395874023435, 0.29894451904296876, 0.29879705810546875, 0.6401351928710938, 0.2988328857421875, 0.2987581481933594, 0.2988840942382813, 0.29872024536132813, 0.29893939208984377, 0.2988707885742187, 0.29880523681640625, 0.2989363098144531, 0.2989342651367187, 0.2987714538574219, 0.29879092407226565, 0.2989823913574219, 0.29887387084960937, 0.2987889099121094, 0.2989301452636719, 0.2989189147949219, 0.2988789672851562, 0.2987530212402344, 0.29921383666992185, 0.2988380126953125, 0.2988083190917969, 0.29881753540039063, 0.2987274169921875, 0.2992404479980469, 0.298977294921875, 0.2987080078125, 0.2987703857421875, 0.2988779602050781, 0.2988001403808594, 0.29884722900390626, 0.2989322509765625, 0.2988431091308594, 0.2989875183105469, 0.29900698852539065, 0.298787841796875, 0.2988380126953125, 0.2987663269042969, 0.29904177856445313, 0.2989045715332031, 0.2989270935058594, 0.2989854736328125, 0.2989793395996094, 0.2989168701171875, 0.29900286865234377, 0.2988134460449219, 0.2988492736816406, 0.2989066162109375, 0.2990909729003906, 0.29900079345703123, 0.2989189147949219, 0.2989056091308594, 0.29892913818359373, 0.29894964599609375, 0.2989598693847656, 0.29893017578125, 0.29894659423828124, 0.2988861083984375, 0.29896295166015624, 0.29888204956054687, 0.2989690856933594, 0.2989066162109375, 0.2988011474609375, 0.64087451171875, 0.2988800048828125, 0.2988359680175781, 0.29870693969726564, 0.2989414367675781, 0.2989752197265625, 0.29876223754882814, 0.2988226623535156, 0.29887591552734377, 0.2989363098144531, 0.298777587890625, 0.29906228637695315, 0.2988083190917969, 0.29909197998046877, 0.2988851318359375, 0.29884210205078127, 0.29921075439453126, 0.2988840942382813, 0.2988717956542969, 0.2987806701660156, 0.29890252685546875, 0.2987325439453125, 0.2989486083984375, 0.29873358154296875, 0.2988717956542969, 0.29869158935546875, 0.29887591552734377, 0.2990520324707031, 0.2989014892578125, 0.2989066162109375, 0.2988083190917969, 0.2989014892578125, 0.29879705810546875, 0.29891278076171873, 0.29888101196289063, 0.2989956970214844, 0.29885440063476565, 0.29872845458984376, 0.29881036376953124, 0.29893939208984377, 0.2989168701171875, 0.29879092407226565, 0.2989588623046875, 0.2989014892578125, 0.29880218505859374, 0.29904794311523436, 0.2987724914550781, 0.29912576293945314, 0.29889022827148437, 0.2988779602050781, 0.2988830871582031, 0.2989260864257812, 0.2988697509765625, 0.29885235595703125, 0.2987653198242187, 0.29883392333984377, 0.29885952758789064, 0.29876837158203123, 0.29888922119140626, 0.298998779296875, 0.2989137878417969, 0.2988769226074219, 0.2987448425292969, 0.6411837158203125, 0.2987673645019531, 0.2987755432128906, 0.2987161560058594, 0.2988308410644531, 0.2989434814453125, 0.29867929077148436, 0.2986700744628906, 0.29879296875, 0.29884622192382815, 0.29911654663085935, 0.2989598693847656, 0.2990049133300781, 0.2988994445800781, 0.2992814025878906, 0.29890252685546875, 0.2989742126464844, 0.29876019287109373, 0.29890252685546875, 0.2987694091796875, 0.29888922119140626, 0.2989137878417969, 0.29885440063476565, 0.2989066162109375, 0.2988206176757813, 0.29879296875, 0.2987888488769531, 0.29874893188476564, 0.298829833984375, 0.29893734741210937, 0.29940017700195315, 0.29889434814453125, 0.298829833984375, 0.2988830871582031, 0.2988994445800781, 0.2990162048339844, 0.2989823913574219, 0.29884622192382815, 0.29896395874023435, 0.2990172119140625, 0.29889022827148437, 0.2988503112792969, 0.298745849609375, 0.2989895629882813, 0.2990100402832031, 0.29893734741210937, 0.2988861389160156, 0.29892913818359373, 0.29891583251953124, 0.29912371826171874, 0.298935302734375, 0.29889739990234376, 0.29890969848632815, 0.2988707885742187, 0.2987857971191406, 0.298977294921875, 0.29891278076171873, 0.2990110778808594, 0.2991124572753906, 0.2989363098144531, 0.2988584899902344, 0.2989465637207031, 0.29895065307617186, 0.641522705078125, 0.298787841796875, 0.29876837158203123, 0.29884722900390626, 0.29865164184570314, 0.2988779602050781, 0.29880117797851563, 0.298661865234375, 0.29881036376953124, 0.2990540771484375, 0.29873663330078126, 0.2987796630859375, 0.2988769226074219, 0.2989619140625, 0.2989916076660156, 0.298925048828125, 0.29885235595703125, 0.298861572265625, 0.29878680419921877, 0.29885235595703125, 0.2988216247558594, 0.2989045715332031, 0.29874893188476564, 0.29875711059570315, 0.29881036376953124, 0.29895269775390626, 0.29883392333984377, 0.29884518432617185, 0.2988687438964844, 0.2988963928222656, 0.29879092407226565, 0.29900390625, 0.29879705810546875, 0.2992803955078125, 0.298956787109375, 0.29879910278320315, 0.29885336303710935, 0.29882366943359373, 0.298935302734375, 0.2987694091796875, 0.298777587890625, 0.29879092407226565, 0.29875201416015623, 0.29888204956054687, 0.29894757080078127, 0.29877862548828127, 0.29879705810546875, 0.29888922119140626, 0.2988441467285156, 0.29900698852539065, 0.29907659912109374, 0.2989619140625, 0.29929470825195315, 0.2988861389160156, 0.2989270935058594, 0.29891278076171873, 0.29884210205078127, 0.2989056091308594, 0.29896600341796875, 0.2988984375, 0.29889022827148437, 0.29891583251953124, 0.2989547424316406]",tokens/s,3.2920160420021825,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3746,7 +3746,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,4118.597632,5287.44448,0.0,4640.997376,4330.668544,s,10,3.1970594482421872,0.31970594482421877,0.0012811478039207278,0.31938397216796877,0.32123822631835935,0.32139851684570314,0.3215267492675781,"[0.31836029052734377, 0.3215588073730469, 0.31788543701171873, 0.319045654296875, 0.3190616149902344, 0.31840402221679687, 0.31970632934570314, 0.32065835571289064, 0.32120260620117186, 0.32117633056640627]",tokens/s,800.7358140955256,kWh,3.7591551740964256e-06,2.0592643818247325e-06,1.8986638453191707e-05,2.4805058009112864e-05,tokens/kWh,10320475.763691055,MB,4118.597632,5287.44448,0.0,4640.997376,4515.667456,s,10,175.13562109375,17.513562109375,0.0017768259462071251,17.513850585937497,17.5153654296875,17.51584091796875,17.51622130859375,"[17.5152421875, 17.5134375, 17.51631640625, 17.5116796875, 17.514263671875, 17.511447265625, 17.51483984375, 17.511279296875, 17.515259765625, 17.51185546875]",tokens/s,3.59721224080829,kWh,0.0002068085063662794,0.0001133488492668755,0.0010180742467503973,0.0013382316023835524,tokens/kWh,47077.05294643273,,s,629,177.73623620605477,0.2825695329190059,0.038042917370424166,0.27798220825195313,0.27814030151367186,0.2782353515625,0.5972927026367187,"[0.27811532592773436, 0.2778900451660156, 0.27781427001953124, 0.278012939453125, 0.27799038696289063, 0.2779176940917969, 0.278012939453125, 0.27791973876953124, 0.27834060668945315, 0.27785113525390626, 0.277939208984375, 0.278029296875, 0.27790643310546875, 0.2777313232421875, 0.277923828125, 0.27794125366210937, 0.2779473876953125, 0.2780712890625, 0.2780887145996094, 0.27796990966796875, 0.27798834228515623, 0.27780609130859374, 0.27821261596679686, 0.27799859619140627, 0.2780968933105469, 0.277833740234375, 0.27810302734375, 0.2780989379882812, 0.27795455932617186, 0.2779525146484375, 0.278108154296875, 0.27808154296875, 0.2780149841308594, 0.27790234375, 0.27806515502929685, 0.2780364685058594, 0.2780139465332031, 0.277865478515625, 0.277928955078125, 0.27805490112304687, 0.27810406494140627, 0.27789926147460936, 0.2781460571289063, 0.2780682373046875, 0.2780078125, 0.2779381713867187, 0.27819622802734373, 0.2781051025390625, 0.2781255798339844, 0.27789825439453125, 0.2779596862792969, 0.2780364685058594, 0.27797607421875, 0.2783692932128906, 0.27815115356445314, 0.27802420043945314, 0.27813067626953125, 0.27788595581054687, 0.2780682373046875, 0.2780968933105469, 0.278118408203125, 0.27799859619140627, 0.5999585571289062, 0.2779914245605469, 0.2784163818359375, 0.2778818664550781, 0.27786648559570315, 0.278034423828125, 0.2778347473144531, 0.2779862976074219, 0.2777989196777344, 0.27785830688476565, 0.2777702331542969, 0.2777927551269531, 0.27792691040039064, 0.278029296875, 0.27775796508789063, 0.2779228210449219, 0.2779402160644531, 0.27796173095703125, 0.27776101684570315, 0.2779862976074219, 0.27793304443359373, 0.2779596862792969, 0.27787982177734377, 0.2779207763671875, 0.2779576416015625, 0.2780917663574219, 0.27807537841796875, 0.27791256713867185, 0.27790130615234376, 0.2779842529296875, 0.2779504699707031, 0.2783477783203125, 0.2779279479980469, 0.27805490112304687, 0.2778204040527344, 0.2778542175292969, 0.2779811706542969, 0.27787774658203124, 0.2782310485839844, 0.27790438842773435, 0.2780006408691406, 0.27790130615234376, 0.2778408813476563, 0.2780907592773437, 0.2779299926757812, 0.277939208984375, 0.27783578491210936, 0.27826174926757813, 0.27810098266601563, 0.27833343505859376, 0.277970947265625, 0.27794842529296876, 0.278012939453125, 0.2779811706542969, 0.2778101806640625, 0.27804364013671873, 0.27798016357421873, 0.27800473022460936, 0.2778818664550781, 0.278002685546875, 0.2779033508300781, 0.2780405883789063, 0.2782843017578125, 0.5971548461914062, 0.2779535217285156, 0.27780197143554686, 0.278213623046875, 0.27794943237304687, 0.2779422607421875, 0.27781427001953124, 0.2779504699707031, 0.2779320373535156, 0.2779535217285156, 0.2779095153808594, 0.27794943237304687, 0.27799755859375, 0.27791156005859374, 0.27797402954101563, 0.2780139465332031, 0.2778623962402344, 0.2780057678222656, 0.27779483032226565, 0.27794943237304687, 0.2780641174316406, 0.2779781188964844, 0.27794329833984377, 0.27810714721679686, 0.2781757507324219, 0.2780313720703125, 0.27787673950195313, 0.27807537841796875, 0.27811431884765625, 0.27794842529296876, 0.2779955139160156, 0.2781460571289063, 0.27803237915039064, 0.27800167846679685, 0.2780262451171875, 0.27801803588867186, 0.2779781188964844, 0.2780405883789063, 0.27790438842773435, 0.2783918151855469, 0.2780979309082031, 0.2779996032714844, 0.2779207763671875, 0.27818597412109375, 0.27806207275390626, 0.2780201110839844, 0.27812966918945314, 0.27813479614257813, 0.278192138671875, 0.27825457763671874, 0.2780149841308594, 0.2780989379882812, 0.2780907592773437, 0.27817166137695315, 0.2779586486816406, 0.2780498046875, 0.2779791259765625, 0.2781583251953125, 0.27810202026367187, 0.2781614074707031, 0.2780334167480469, 0.278002685546875, 0.278392822265625, 0.5973463134765625, 0.2779453430175781, 0.27787774658203124, 0.277865478515625, 0.278029296875, 0.27785009765625, 0.2778736572265625, 0.2779053955078125, 0.27785009765625, 0.27782962036132813, 0.2777774047851562, 0.27799755859375, 0.27794943237304687, 0.27787060546875, 0.2777999267578125, 0.27798220825195313, 0.2778961791992188, 0.2779381713867187, 0.2778531799316406, 0.2780928039550781, 0.2779719543457031, 0.27791668701171873, 0.2777712707519531, 0.27799859619140627, 0.27791256713867185, 0.2778951721191406, 0.27785830688476565, 0.2779781188964844, 0.27801089477539065, 0.278044677734375, 0.27773849487304686, 0.2779535217285156, 0.2779648132324219, 0.2780498046875, 0.2778531799316406, 0.2779576416015625, 0.2779156494140625, 0.278012939453125, 0.27791461181640625, 0.27799038696289063, 0.27808154296875, 0.2779852905273438, 0.2778204040527344, 0.27817779541015625, 0.2780508117675781, 0.27812353515625, 0.2779873352050781, 0.27802215576171874, 0.27810202026367187, 0.27819110107421874, 0.27801190185546876, 0.2780487670898438, 0.2780610656738281, 0.27813479614257813, 0.2779105224609375, 0.2779576416015625, 0.27790643310546875, 0.27801190185546876, 0.27795455932617186, 0.2779381713867187, 0.2780231628417969, 0.2780712890625, 0.27800985717773435, 0.59753369140625, 0.27792486572265623, 0.2778367919921875, 0.278245361328125, 0.27796173095703125, 0.2780078125, 0.2778153076171875, 0.27805902099609375, 0.27785626220703125, 0.2779832458496094, 0.27780813598632814, 0.2779053955078125, 0.2779320373535156, 0.2779648132324219, 0.2777978820800781, 0.27794638061523436, 0.2779893798828125, 0.27795660400390626, 0.2779258728027344, 0.2780149841308594, 0.2778623962402344, 0.27793508911132814, 0.27788803100585935, 0.27812042236328127, 0.2779473876953125, 0.2780477294921875, 0.27788491821289063, 0.27803750610351563, 0.2779207763671875, 0.2780426330566406, 0.2778951721191406, 0.27804672241210937, 0.2779781188964844, 0.27805902099609375, 0.277796875, 0.2779637756347656, 0.27799859619140627, 0.27796990966796875, 0.2780088195800781, 0.2779504699707031, 0.2780159912109375, 0.27801702880859375, 0.277865478515625, 0.2781685791015625, 0.27800677490234377, 0.2781051025390625, 0.2779996032714844, 0.27807437133789065, 0.27806207275390626, 0.278139892578125, 0.2779299926757812, 0.278066162109375, 0.2780835876464844, 0.2781214599609375, 0.27794125366210937, 0.2781419372558594, 0.2781163635253906, 0.2781460571289063, 0.27796786499023435, 0.27823822021484373, 0.27809381103515624, 0.27807949829101564, 0.2781194152832031, 0.5966243896484374, 0.2779586486816406, 0.27771597290039063, 0.2778900451660156, 0.2777845764160156, 0.27794842529296876, 0.27777432250976564, 0.27786138916015624, 0.2779832458496094, 0.27784295654296876, 0.27791360473632815, 0.2778265686035156, 0.27825048828125, 0.2778839111328125, 0.277760009765625, 0.2779811706542969, 0.2778316650390625, 0.277928955078125, 0.2776801147460938, 0.27795147705078127, 0.27797503662109374, 0.27823822021484373, 0.2777794494628906, 0.2780354614257812, 0.2778675231933594, 0.278034423828125, 0.27775796508789063, 0.2779596862792969, 0.2779453430175781, 0.27806515502929685, 0.27786444091796875, 0.2780897216796875, 0.2780426330566406, 0.27793508911132814, 0.2780201110839844, 0.2778951721191406, 0.2779228210449219, 0.27799655151367186, 0.27801089477539065, 0.277960693359375, 0.27807745361328123, 0.2780201110839844, 0.2779279479980469, 0.2779719543457031, 0.2779791259765625, 0.2780078125, 0.27794329833984377, 0.27796890258789064, 0.27801190185546876, 0.27812045288085935, 0.2779719543457031, 0.27803237915039064, 0.2780354614257812, 0.27804672241210937, 0.27791973876953124, 0.27803955078125, 0.27796990966796875, 0.27832321166992186, 0.2778173522949219, 0.2779924621582031, 0.2780313720703125, 0.2779924621582031, 0.2779095153808594, 0.5980723266601562, 0.27799755859375, 0.2777897033691406, 0.2778787841796875, 0.2782607421875, 0.27835186767578124, 0.27776409912109373, 0.27780609130859374, 0.277928955078125, 0.2779156494140625, 0.27778866577148437, 0.2779658203125, 0.2778961791992188, 0.2779033508300781, 0.2778521728515625, 0.27789208984375, 0.2778408813476563, 0.2780610656738281, 0.27775180053710935, 0.277928955078125, 0.2779955139160156, 0.277886962890625, 0.27779379272460936, 0.2780897216796875, 0.27797402954101563, 0.27802215576171874, 0.277818359375, 0.2780201110839844, 0.27800473022460936, 0.27827508544921875, 0.27791256713867185, 0.2779934692382813, 0.278012939453125, 0.2779361267089844, 0.27817880249023436, 0.2780354614257812, 0.2780149841308594, 0.27794329833984377, 0.27789825439453125, 0.2780262451171875, 0.27804568481445313, 0.27850650024414064, 0.27791873168945314, 0.2781460571289063, 0.27798220825195313, 0.27811532592773436, 0.27813784790039064, 0.27810406494140627, 0.2780354614257812, 0.2781583251953125, 0.27791256713867185, 0.27816754150390627, 0.2781614074707031, 0.278108154296875, 0.2779473876953125, 0.27812045288085935, 0.27804364013671873, 0.278076416015625, 0.27797503662109374, 0.2780887145996094, 0.2781245422363281, 0.2781767578125, 0.27807232666015624, 0.5989816284179688, 0.277907470703125, 0.27786956787109374, 0.27786343383789064, 0.2780149841308594, 0.27786444091796875, 0.2781245422363281, 0.27784396362304686, 0.2778961791992188, 0.2778951721191406, 0.2777282409667969, 0.27791256713867185, 0.27789208984375, 0.2780354614257812, 0.2777200622558594, 0.2778880004882813, 0.27792691040039064, 0.2780928039550781, 0.2777774047851562, 0.2778787841796875, 0.27798220825195313, 0.2778808288574219, 0.2777569274902344, 0.27796273803710936, 0.2778173522949219, 0.2779586486816406, 0.2777763977050781, 0.27791156005859374, 0.27785830688476565, 0.2779729919433594, 0.2778050537109375, 0.2779719543457031, 0.2780149841308594, 0.27809381103515624, 0.27781427001953124, 0.27789108276367186, 0.277970947265625, 0.2780948486328125, 0.2778675231933594, 0.2779228210449219, 0.27799655151367186, 0.27811224365234377, 0.27821261596679686, 0.27802520751953125, 0.27791461181640625, 0.2779893798828125, 0.2780262756347656, 0.27805181884765623, 0.27803851318359374, 0.27812762451171874, 0.2778623962402344, 0.27810919189453126, 0.27789208984375, 0.27803237915039064, 0.27796786499023435, 0.2779924621582031, 0.27802420043945314, 0.2780498046875, 0.2778736572265625, 0.27802725219726565, 0.2780917663574219, 0.27807437133789065, 0.27804779052734374, 0.5996082763671875, 0.2778603515625, 0.27799038696289063, 0.2779002990722656, 0.2780784606933594, 0.2780487670898438, 0.27779071044921877, 0.27787060546875, 0.27787161254882814, 0.277907470703125, 0.27785830688476565, 0.2779095153808594, 0.2779945068359375, 0.27801190185546876, 0.27792178344726565, 0.2780282897949219, 0.2778726501464844, 0.27789926147460936, 0.27795147705078127, 0.27812249755859375, 0.2779299926757812, 0.2781460571289063, 0.2777681884765625, 0.27811224365234377, 0.2779893798828125, 0.2780211181640625, 0.27784396362304686, 0.27811737060546876, 0.2779924621582031, 0.2780702819824219, 0.27788595581054687, 0.27822796630859375, 0.2781614074707031, 0.27828121948242185, 0.27786444091796875, 0.27808767700195314, 0.2780159912109375, 0.2779842529296875, 0.27789312744140626, 0.27807745361328123, 0.27812863159179685, 0.27812045288085935, 0.277923828125, 0.2781317138671875, 0.2779053955078125, 0.27808050537109374, 0.2779934692382813, 0.2780682373046875, 0.27818801879882815, 0.278139892578125, 0.27804568481445313, 0.27794943237304687, 0.27806719970703125, 0.27802520751953125, 0.2779033508300781, 0.27811224365234377, 0.278108154296875, 0.2783221740722656, 0.2778388366699219, 0.27817166137695315, 0.2779535217285156, 0.27817779541015625, 0.2781368408203125, 0.5995950317382812, 0.27793508911132814, 0.2778111877441406, 0.27782962036132813, 0.277923828125, 0.2780057678222656, 0.2776893310546875, 0.27791873168945314, 0.2778787841796875, 0.27778253173828127, 0.2776565856933594, 0.27797503662109374, 0.27791668701171873, 0.27793408203125, 0.277823486328125, 0.27784909057617185, 0.2778531799316406, 0.278044677734375, 0.27778662109375, 0.278371337890625, 0.27801803588867186, 0.27795559692382815, 0.27778765869140626, 0.27804672241210937, 0.2778900451660156, 0.27810202026367187, 0.27790643310546875, 0.2779443054199219, 0.2779279479980469, 0.2781368408203125, 0.27793408203125, 0.2779914245605469, 0.2780159912109375, 0.27801702880859375, 0.2779934692382813, 0.2779084777832031, 0.2779371643066406, 0.27799038696289063, 0.2778685302734375, 0.2779504699707031, 0.2779842529296875, 0.2779535217285156, 0.27789926147460936, 0.2780508117675781, 0.2779299926757812, 0.2779852905273438, 0.27796786499023435, 0.2780845947265625, 0.278012939453125, 0.27808563232421873, 0.27792486572265623, 0.2780528564453125, 0.2779791259765625, 0.27808767700195314, 0.27787982177734377, 0.2782576599121094, 0.2780477294921875, 0.27795660400390626, 0.2779730224609375, 0.27805487060546874, 0.27805184936523436, 0.27802725219726565, 0.2779259338378906]",tokens/s,3.538951951648072,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3961,7 +3961,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4160,7 +4160,7 @@ OSError: M is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4580,7 +4580,7 @@ Traceback (most recent call last): OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4709,7 +4709,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,6439.493632,20389.036032,0.0,19742.588928,18821.508608,s,10,15.1473671875,1.5147367187499998,0.00247756470747785,1.5140679931640624,1.5185943847656251,1.5187318237304688,1.5188417749023437,"[1.5126702880859375, 1.5122486572265625, 1.5120849609375, 1.5139268798828125, 1.5142091064453125, 1.5128245849609374, 1.514491455078125, 1.5174781494140626, 1.5188692626953124, 1.5185638427734376]",tokens/s,169.00626810661714,kWh,1.786537355846829e-05,9.788986818782104e-06,8.822982058380724e-05,0.00011588418096105764,tokens/kWh,2209102.2077122647,MB,6440.05888,20389.036032,0.0,19742.588928,19408.2304,s,10,843.181140625,84.3181140625,0.002181345135820369,84.31812890625,84.32165,84.32172734375,84.32178921875,"[84.3218046875, 84.31809375, 84.3141640625, 84.317453125, 84.316203125, 84.31825, 84.3185546875, 84.3168203125, 84.3181640625, 84.3216328125]",tokens/s,0.7471704117255499,kWh,0.0009954405118525029,0.0005455906297619111,0.00497354175660798,0.006514572898222393,tokens/kWh,9670.626299567632,,s,629,855.4723406982422,1.360051416054439,0.1800573065911544,1.3383363037109375,1.3387327392578126,1.338873681640625,2.85288818359375,"[1.33827587890625, 1.338081298828125, 1.3381846923828125, 1.3384754638671874, 1.33840283203125, 1.337996337890625, 1.338251220703125, 1.33793896484375, 1.3378477783203124, 1.338167236328125, 1.337945068359375, 1.338113037109375, 1.3379625244140625, 1.3380064697265626, 1.3384990234375, 1.3382308349609375, 1.33815087890625, 1.3384324951171875, 1.3385635986328126, 1.3381785888671875, 1.3384417724609374, 1.338292236328125, 1.33815087890625, 1.3382215576171874, 1.33812841796875, 1.3385369873046875, 1.33857080078125, 1.33825537109375, 1.338660888671875, 1.338504150390625, 1.3385400390625, 1.338418212890625, 1.338197998046875, 1.339029541015625, 1.3383699951171875, 1.338683349609375, 1.3382236328125, 1.338335205078125, 1.33846728515625, 1.33833935546875, 1.3388974609375, 1.3384970703125, 1.3387171630859376, 1.33850927734375, 1.3386947021484374, 1.33789794921875, 1.3382584228515626, 1.3383731689453124, 1.3386966552734374, 1.338715087890625, 1.3383782958984376, 1.3388123779296874, 1.3381427001953126, 1.338365966796875, 1.338388427734375, 1.338744873046875, 1.338123291015625, 1.3381826171875, 1.3381212158203124, 1.338502197265625, 1.3384345703125, 1.338925048828125, 2.855078857421875, 1.3382308349609375, 1.3380341796875, 1.338641357421875, 1.3384100341796874, 1.338455078125, 1.338123291015625, 1.33812841796875, 1.338135498046875, 1.3380413818359376, 1.3386229248046875, 1.338265625, 1.3381959228515625, 1.338167236328125, 1.338260498046875, 1.3382215576171874, 1.33853076171875, 1.3382952880859376, 1.3382645263671875, 1.3385697021484375, 1.3381181640625, 1.337927734375, 1.3383004150390625, 1.3381468505859375, 1.33812841796875, 1.33810693359375, 1.33798095703125, 1.3387745361328125, 1.3384786376953124, 1.3384632568359376, 1.3380546875, 1.33858203125, 1.3382686767578125, 1.33855126953125, 1.338820556640625, 1.3380628662109375, 1.33846728515625, 1.3384324951171875, 1.3379993896484375, 1.3383935546875, 1.3383526611328125, 1.33831787109375, 1.3384007568359375, 1.338850341796875, 1.338428466796875, 1.3388790283203125, 1.3380782470703125, 1.3385390625, 1.33848681640625, 1.3382850341796875, 1.338239990234375, 1.338397705078125, 1.338281982421875, 1.3381427001953126, 1.338702880859375, 1.338449951171875, 1.3385543212890625, 1.3380904541015626, 1.3382625732421876, 1.338514404296875, 1.33867822265625, 1.338439697265625, 1.3386260986328125, 2.852787109375, 1.337776123046875, 1.3380433349609375, 1.33821240234375, 1.3381181640625, 1.3380423583984375, 1.338502197265625, 1.337923583984375, 1.3380556640625, 1.337927734375, 1.3382747802734376, 1.3379788818359375, 1.3380535888671874, 1.3380679931640624, 1.338323974609375, 1.33837109375, 1.3381468505859375, 1.3381048583984374, 1.3382010498046875, 1.338040283203125, 1.338312744140625, 1.3381253662109376, 1.338639404296875, 1.3382962646484375, 1.3380535888671874, 1.3382327880859375, 1.33817041015625, 1.338787841796875, 1.3383106689453126, 1.3387684326171876, 1.338207275390625, 1.338524658203125, 1.3381622314453125, 1.3384385986328124, 1.3382880859375, 1.3383096923828126, 1.3381539306640624, 1.3384376220703125, 1.3379154052734374, 1.3386424560546875, 1.338576904296875, 1.338471435546875, 1.3384324951171875, 1.33886767578125, 1.3384775390625, 1.3387581787109375, 1.33810693359375, 1.3385318603515626, 1.3380946044921875, 1.338239013671875, 1.338491943359375, 1.338514404296875, 1.3388123779296874, 1.338328125, 1.33863525390625, 1.3383731689453124, 1.3381826171875, 1.3379378662109376, 1.3384427490234374, 1.3381744384765626, 1.3386107177734374, 1.3386884765625, 1.3387366943359376, 2.852927490234375, 1.3382041015625, 1.33804443359375, 1.3380986328125, 1.3379327392578124, 1.33829736328125, 1.3382789306640626, 1.33785595703125, 1.3384140625, 1.338491943359375, 1.338461181640625, 1.3379215087890626, 1.3379921875, 1.3384775390625, 1.338349609375, 1.3382891845703124, 1.3381856689453124, 1.338224609375, 1.3381641845703125, 1.3382738037109374, 1.3382677001953125, 1.3385400390625, 1.3384990234375, 1.338345458984375, 1.3385574951171875, 1.3380638427734375, 1.3382021484375, 1.3386884765625, 1.3383792724609376, 1.3390386962890626, 1.3381170654296874, 1.3385206298828125, 1.3384171142578125, 1.338330078125, 1.3384376220703125, 1.338135498046875, 1.3385645751953126, 1.3382369384765624, 1.3381488037109375, 1.3385615234375, 1.338334228515625, 1.3385687255859375, 1.3384263916015624, 1.338876953125, 1.3385943603515624, 1.3389271240234375, 1.3382574462890624, 1.3385318603515626, 1.338039306640625, 1.33815087890625, 1.338440673828125, 1.338408935546875, 1.3386680908203126, 1.3384068603515624, 1.3384693603515625, 1.338323974609375, 1.3382747802734376, 1.3380362548828124, 1.3383433837890626, 1.3387017822265626, 1.338533935546875, 1.3384949951171874, 1.3386751708984375, 2.853210205078125, 1.33809765625, 1.3379481201171874, 1.33806591796875, 1.338060791015625, 1.3381856689453124, 1.338239013671875, 1.33782421875, 1.338302490234375, 1.3379266357421875, 1.338209228515625, 1.3380413818359376, 1.3381734619140624, 1.3381641845703125, 1.3383076171875, 1.338155029296875, 1.3382471923828125, 1.338187744140625, 1.338287109375, 1.3382471923828125, 1.338228759765625, 1.33800244140625, 1.3384847412109375, 1.3383424072265624, 1.3381273193359375, 1.3382645263671875, 1.3383167724609375, 1.3388646240234374, 1.3382379150390624, 1.338666015625, 1.3386270751953124, 1.33897314453125, 1.3383065185546874, 1.3384427490234374, 1.338640380859375, 1.33823388671875, 1.338455078125, 1.3382747802734376, 1.33837109375, 1.338439697265625, 1.3384366455078125, 1.3383905029296874, 1.338646484375, 1.33884521484375, 1.338330078125, 1.3389864501953126, 1.3383802490234376, 1.338328125, 1.33821337890625, 1.3384385986328124, 1.3382369384765624, 1.33838232421875, 1.338428466796875, 1.338207275390625, 1.3385430908203124, 1.338228759765625, 1.3381375732421874, 1.3383997802734375, 1.33844580078125, 1.338328125, 1.3384949951171874, 1.3387335205078126, 1.3386895751953125, 2.851811279296875, 1.3381427001953126, 1.338029052734375, 1.3379471435546875, 1.3380321044921875, 1.3382861328125, 1.3381212158203124, 1.337881591796875, 1.3381683349609375, 1.3383577880859374, 1.33817041015625, 1.338059814453125, 1.338250244140625, 1.33771875, 1.33808642578125, 1.33836181640625, 1.3384345703125, 1.3382471923828125, 1.33817138671875, 1.338324951171875, 1.3382225341796874, 1.3380731201171876, 1.3383475341796875, 1.3384796142578126, 1.338498046875, 1.338177490234375, 1.338260498046875, 1.3385491943359376, 1.33831982421875, 1.338625, 1.33840283203125, 1.338828857421875, 1.33834033203125, 1.3385369873046875, 1.338491943359375, 1.3382593994140626, 1.3384049072265625, 1.3383792724609376, 1.33875, 1.33888720703125, 1.3388912353515625, 1.3387335205078126, 1.3385902099609375, 1.3390325927734374, 1.33831884765625, 1.3388984375, 1.3380587158203125, 1.338312744140625, 1.338059814453125, 1.3385400390625, 1.3383577880859374, 1.3386107177734374, 1.338830810546875, 1.3383013916015625, 1.3383731689453124, 1.3382635498046875, 1.3384385986328124, 1.33800341796875, 1.33859326171875, 1.3383843994140625, 1.3386947021484374, 1.3384888916015625, 1.3389117431640625, 2.853399658203125, 1.3380997314453125, 1.3381314697265625, 1.3383485107421875, 1.3379093017578125, 1.3382696533203124, 1.33813037109375, 1.338008544921875, 1.3382789306640626, 1.3383157958984375, 1.33842431640625, 1.3384100341796874, 1.3381949462890625, 1.3380885009765624, 1.3384017333984375, 1.3384765625, 1.3382010498046875, 1.3380526123046874, 1.3383218994140624, 1.3381622314453125, 1.338207275390625, 1.3384222412109374, 1.3384591064453124, 1.3383270263671876, 1.3385625, 1.337927734375, 1.338203125, 1.3387591552734375, 1.33834033203125, 1.338640380859375, 1.3384735107421875, 1.338603515625, 1.3385892333984375, 1.338545166015625, 1.3386025390625, 1.3384468994140626, 1.3385206298828125, 1.338261474609375, 1.338314697265625, 1.3385943603515624, 1.33827783203125, 1.3385491943359376, 1.3389219970703126, 1.3390673828125, 1.3380843505859374, 1.33874169921875, 1.338102783203125, 1.3382635498046875, 1.33827685546875, 1.338281982421875, 1.3385645751953126, 1.33852978515625, 1.3383935546875, 1.3382266845703126, 1.33874169921875, 1.3383935546875, 1.3385174560546875, 1.3384385986328124, 1.3383218994140624, 1.3382532958984374, 1.338398681640625, 1.338607666015625, 1.3387099609375, 2.856235107421875, 1.3382266845703126, 1.3379635009765625, 1.338997802734375, 1.3381160888671875, 1.33855126953125, 1.3383414306640624, 1.33804541015625, 1.3381396484375, 1.3380484619140625, 1.3383055419921874, 1.338119140625, 1.3381806640625, 1.337987060546875, 1.3384642333984376, 1.3384119873046876, 1.338092529296875, 1.338167236328125, 1.3384990234375, 1.3382625732421876, 1.3382000732421875, 1.3384959716796876, 1.33859326171875, 1.3382962646484375, 1.3382880859375, 1.338103759765625, 1.338419189453125, 1.3385390625, 1.338028076171875, 1.3388687744140626, 1.3383526611328125, 1.3386536865234375, 1.338312744140625, 1.33836181640625, 1.3386044921875, 1.3382686767578125, 1.338302490234375, 1.3381314697265625, 1.3380546875, 1.33846728515625, 1.3385062255859375, 1.3385830078125, 1.3386895751953125, 1.338883056640625, 1.3383526611328125, 1.33880419921875, 1.3380731201171876, 1.338312744140625, 1.338076171875, 1.3381683349609375, 1.338346435546875, 1.338471435546875, 1.338356689453125, 1.3382635498046875, 1.338735595703125, 1.338377197265625, 1.338420166015625, 1.338218505859375, 1.3382369384765624, 1.33845703125, 1.338354736328125, 1.3386639404296874, 1.3384560546875, 2.8577626953125, 1.3383587646484374, 1.3380689697265624, 1.3380474853515625, 1.3381632080078125, 1.3382593994140626, 1.338166259765625, 1.338167236328125, 1.3380833740234375, 1.3378006591796876, 1.3386485595703126, 1.3379481201171874, 1.3380731201171876, 1.338176513671875, 1.3382532958984374, 1.3385789794921874, 1.3384898681640625, 1.338156005859375, 1.3380740966796876, 1.3381806640625, 1.3379296875, 1.3382943115234376, 1.33830859375, 1.338144775390625, 1.3381417236328126, 1.338250244140625, 1.3382861328125, 1.3385697021484375, 1.3382645263671875, 1.3386312255859374, 1.3383526611328125, 1.338724365234375, 1.33850732421875, 1.3385440673828124, 1.3388011474609376, 1.3382943115234376, 1.3387642822265624, 1.3382420654296876, 1.33886669921875, 1.338575927734375, 1.3384212646484375, 1.3385267333984374, 1.33864453125, 1.3386854248046876, 1.33810888671875, 1.3389066162109375, 1.3379942626953125, 1.33859228515625, 1.338541015625, 1.3379840087890624, 1.3384171142578125, 1.33873046875, 1.3383382568359374, 1.338566650390625, 1.33863427734375, 1.3383363037109375, 1.3384110107421876, 1.33809765625, 1.33823388671875, 1.3384151611328126, 1.338819580078125, 1.338630126953125, 1.33911962890625, 2.857397216796875, 1.338418212890625, 1.3382236328125, 1.3382666015625, 1.3380423583984375, 1.338387451171875, 1.338007568359375, 1.33800341796875, 1.337934814453125, 1.3378876953125, 1.3382880859375, 1.338082275390625, 1.3380628662109375, 1.33802392578125, 1.3383843994140625, 1.3384273681640626, 1.3381641845703125, 1.3382635498046875, 1.3379410400390626, 1.338555419921875, 1.3382379150390624, 1.3383065185546874, 1.33861279296875, 1.3383004150390625, 1.3380045166015626, 1.3382584228515626, 1.338186767578125, 1.3387325439453126, 1.33827685546875, 1.338945556640625, 1.3384222412109374, 1.3384488525390625, 1.338629150390625, 1.3385841064453126, 1.3385635986328126, 1.3383577880859374, 1.3383157958984375, 1.3384683837890625, 1.338165283203125, 1.3385482177734376, 1.338397705078125, 1.3388319091796874, 1.338493896484375, 1.338962890625, 1.338429443359375, 1.3388216552734375, 1.338155029296875, 1.3382799072265625, 1.338113037109375, 1.338430419921875, 1.3383331298828125, 1.338303466796875, 1.3383363037109375, 1.33825537109375, 1.338997802734375, 1.338502197265625, 1.3380904541015626, 1.3381529541015624, 1.3384119873046876, 1.33829736328125, 1.33835986328125, 1.3386004638671876, 1.3387745361328125]",tokens/s,0.7352663202256265,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4837,7 +4837,7 @@ OSError: google/recurrentgemma-7b is not a local folder and is not a valid model If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5182,7 +5182,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5285,7 +5285,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 96.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5587,7 +5587,7 @@ OSError: 0 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5846,7 +5846,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1451.06944,2426.929152,0.0,1780.482048,1621.410304,s,10,0.7041353912353515,0.07041353912353515,0.0002361853075699059,0.0705183219909668,0.07067396697998048,0.0706822479248047,0.07068887268066407,"[0.07069052886962891, 0.07028050994873047, 0.07060643005371094, 0.0705650863647461, 0.07003734588623046, 0.07016966247558594, 0.07007705688476562, 0.07050838470458984, 0.07052825927734375, 0.07067212677001954]",tokens/s,3635.664435938487,kWh,8.278217291128015e-07,4.5360747945568833e-07,3.6650922871604466e-06,4.946521495728937e-06,tokens/kWh,51753540.38611631,MB,1451.06944,2426.929152,0.0,1780.482048,1723.272192,s,10,38.0147158203125,3.80147158203125,0.0004616553351782168,3.8014595947265626,3.8020025390625,3.8021644775390624,3.8022940283203126,"[3.802326416015625, 3.801103271484375, 3.801966552734375, 3.8013779296875, 3.801707275390625, 3.801541259765625, 3.800688232421875, 3.801736328125, 3.800993896484375, 3.801274658203125]",tokens/s,16.572529516671292,kWh,4.488269805491387e-05,2.4598166829557432e-05,0.0001986788473656444,0.00026815971225011573,tokens/kWh,234934.61963905732,,s,629,38.587294712066715,0.061347050416640146,0.008392685763651334,0.06032691192626953,0.06039490661621094,0.06042992630004883,0.13075378845214844,"[0.060313568115234376, 0.06032076644897461, 0.060349441528320315, 0.060303359985351565, 0.06035148620605469, 0.0603422737121582, 0.06035763168334961, 0.06037299346923828, 0.06039756774902344, 0.06034022521972656, 0.060295169830322265, 0.06041292953491211, 0.06027775955200195, 0.06031052780151367, 0.06032793426513672, 0.060314624786376954, 0.06030131149291992, 0.06029619216918945, 0.06031052780151367, 0.060375038146972655, 0.06031769561767578, 0.060290046691894535, 0.06032998275756836, 0.0604139518737793, 0.060281856536865235, 0.06031155014038086, 0.060355583190917966, 0.06032896041870117, 0.060336128234863284, 0.06032896041870117, 0.06035456085205078, 0.06030233764648438, 0.06037811279296875, 0.06033715057373047, 0.060319744110107425, 0.06034636688232422, 0.06032588958740234, 0.060276737213134764, 0.060368896484375, 0.060331008911132813, 0.06034636688232422, 0.06036684799194336, 0.06032691192626953, 0.06030950546264648, 0.06034636688232422, 0.06035148620605469, 0.060316673278808595, 0.060295169830322265, 0.06031257629394531, 0.06030847930908203, 0.06034841537475586, 0.06037299346923828, 0.06039039993286133, 0.06031564712524414, 0.06034431838989258, 0.06038016128540039, 0.06031052780151367, 0.06034022521972656, 0.06030745697021484, 0.06039039993286133, 0.06033407974243164, 0.06071091079711914, 0.13086003112792968, 0.06027468872070312, 0.060333057403564455, 0.06030950546264648, 0.06031155014038086, 0.06030950546264648, 0.06030847930908203, 0.060276737213134764, 0.06025830459594726, 0.06025932693481445, 0.06027775955200195, 0.060268543243408204, 0.06028083038330078, 0.06025625610351563, 0.060283905029296876, 0.06025625610351563, 0.06033203125, 0.06031052780151367, 0.060375038146972655, 0.06030438232421875, 0.06034329605102539, 0.06034534454345703, 0.060276737213134764, 0.06028799819946289, 0.06030233764648438, 0.0603054084777832, 0.060316673278808595, 0.060298240661621094, 0.06029619216918945, 0.06029414367675781, 0.06036479949951172, 0.06043852615356445, 0.06035763168334961, 0.06036172866821289, 0.060450817108154295, 0.06034841537475586, 0.060365825653076174, 0.060388351440429686, 0.06029414367675781, 0.06033919906616211, 0.06032691192626953, 0.0603422737121582, 0.06036684799194336, 0.06033407974243164, 0.06032588958740234, 0.06034636688232422, 0.06038323211669922, 0.060391422271728515, 0.06032486343383789, 0.0603054084777832, 0.060316673278808595, 0.06037401580810547, 0.060386302947998044, 0.060382209777832034, 0.060336128234863284, 0.060437503814697265, 0.06042521667480469, 0.06031769561767578, 0.060365825653076174, 0.0603054084777832, 0.060407806396484375, 0.06031564712524414, 0.06038937759399414, 0.13136691284179688, 0.06031769561767578, 0.06033715057373047, 0.06035968017578125, 0.06034022521972656, 0.06029107284545898, 0.060286975860595705, 0.06029107284545898, 0.06029414367675781, 0.06024499130249023, 0.060314624786376954, 0.060262401580810546, 0.06032281494140625, 0.06025625610351563, 0.06031155014038086, 0.06026649475097656, 0.06032486343383789, 0.06039756774902344, 0.06035968017578125, 0.06030233764648438, 0.06035456085205078, 0.060295169830322265, 0.06032998275756836, 0.06029926300048828, 0.0603135986328125, 0.06033817672729492, 0.06035456085205078, 0.0603873291015625, 0.060352512359619144, 0.06028595352172852, 0.06034022521972656, 0.06030233764648438, 0.060442623138427735, 0.060336128234863284, 0.060402687072753904, 0.060331008911132813, 0.060426239013671876, 0.06041088104248047, 0.06029619216918945, 0.06032281494140625, 0.06037913513183594, 0.06038937759399414, 0.06035763168334961, 0.060439552307128906, 0.0603135986328125, 0.06057062530517578, 0.06040883255004883, 0.06034124755859375, 0.06027468872070312, 0.060290046691894535, 0.060369918823242184, 0.060333057403564455, 0.06034739303588867, 0.06037401580810547, 0.0603351058959961, 0.06033203125, 0.060369918823242184, 0.06029926300048828, 0.060440574645996094, 0.06032691192626953, 0.06042009735107422, 0.06032998275756836, 0.06038937759399414, 0.13121229553222657, 0.06032588958740234, 0.060352512359619144, 0.06033407974243164, 0.06028595352172852, 0.06032179260253906, 0.060295169830322265, 0.06027980804443359, 0.060278785705566405, 0.06031052780151367, 0.06031769561767578, 0.06029414367675781, 0.06030438232421875, 0.06028595352172852, 0.06030847930908203, 0.06031564712524414, 0.060303359985351565, 0.06029312133789062, 0.06030950546264648, 0.060303359985351565, 0.06031257629394531, 0.060319744110107425, 0.06031872177124024, 0.06030847930908203, 0.06031155014038086, 0.06036070251464844, 0.06029926300048828, 0.06032588958740234, 0.0603054084777832, 0.06032691192626953, 0.060331008911132813, 0.060352512359619144, 0.06030643081665039, 0.06039654541015625, 0.06035865783691406, 0.06033817672729492, 0.06036377716064453, 0.06035148620605469, 0.06028287887573242, 0.06038425445556641, 0.06041497421264649, 0.06032179260253906, 0.06035865783691406, 0.06032998275756836, 0.06031257629394531, 0.06040678405761719, 0.06037606430053711, 0.06036684799194336, 0.060284927368164064, 0.06030847930908203, 0.06038425445556641, 0.06036377716064453, 0.06035763168334961, 0.06036377716064453, 0.06035763168334961, 0.06036070251464844, 0.06041292953491211, 0.06029926300048828, 0.0603873291015625, 0.06032691192626953, 0.060437503814697265, 0.06032896041870117, 0.06042009735107422, 0.1306306610107422, 0.060273662567138675, 0.060333057403564455, 0.06033919906616211, 0.06030131149291992, 0.06034636688232422, 0.06030131149291992, 0.060303359985351565, 0.06026444625854492, 0.06025932693481445, 0.060267520904541017, 0.060268543243408204, 0.06032486343383789, 0.06025625610351563, 0.060273662567138675, 0.06026649475097656, 0.060267520904541017, 0.060298240661621094, 0.06034124755859375, 0.06027775955200195, 0.06034124755859375, 0.0603054084777832, 0.06027468872070312, 0.060316673278808595, 0.06037094497680664, 0.06032486343383789, 0.06032896041870117, 0.060286975860595705, 0.06031564712524414, 0.06034534454345703, 0.06036684799194336, 0.06033817672729492, 0.06033919906616211, 0.060391422271728515, 0.06032998275756836, 0.06034124755859375, 0.06038016128540039, 0.06036377716064453, 0.06034431838989258, 0.06035968017578125, 0.06033919906616211, 0.06034329605102539, 0.060375038146972655, 0.06034124755859375, 0.060728321075439455, 0.060353534698486325, 0.06036377716064453, 0.06032179260253906, 0.06030643081665039, 0.06028902435302735, 0.06030847930908203, 0.06033817672729492, 0.060355583190917966, 0.0603422737121582, 0.060846080780029295, 0.0603770866394043, 0.06039756774902344, 0.06029414367675781, 0.06039039993286133, 0.06033407974243164, 0.060375038146972655, 0.06032588958740234, 0.06038118362426758, 0.13078016662597655, 0.060283905029296876, 0.06032179260253906, 0.060455936431884766, 0.06029312133789062, 0.060260353088378904, 0.06032588958740234, 0.06031564712524414, 0.060283905029296876, 0.060262401580810546, 0.06040678405761719, 0.06024806213378906, 0.06032691192626953, 0.06025830459594726, 0.06031155014038086, 0.060240894317626956, 0.060295169830322265, 0.06030643081665039, 0.06055936050415039, 0.06036479949951172, 0.06031257629394531, 0.06029107284545898, 0.0603351058959961, 0.0603135986328125, 0.060286975860595705, 0.06032076644897461, 0.0603351058959961, 0.06029312133789062, 0.06033203125, 0.06028799819946289, 0.06037094497680664, 0.06072115325927734, 0.060316673278808595, 0.06034739303588867, 0.06034636688232422, 0.0603422737121582, 0.06034431838989258, 0.06031155014038086, 0.060303359985351565, 0.06033407974243164, 0.06033407974243164, 0.06032896041870117, 0.06034022521972656, 0.060371967315673826, 0.06029619216918945, 0.06029209518432617, 0.06038118362426758, 0.060355583190917966, 0.060298240661621094, 0.060295169830322265, 0.06033407974243164, 0.06033203125, 0.06036377716064453, 0.06036787033081055, 0.06041497421264649, 0.060402687072753904, 0.06039244842529297, 0.060268543243408204, 0.06036172866821289, 0.06028799819946289, 0.060432384490966794, 0.06033407974243164, 0.06041292953491211, 0.1306859588623047, 0.06027775955200195, 0.06031155014038086, 0.060355583190917966, 0.06030131149291992, 0.06034739303588867, 0.06032486343383789, 0.06030950546264648, 0.06028595352172852, 0.0603054084777832, 0.06030131149291992, 0.060290046691894535, 0.0603422737121582, 0.06029312133789062, 0.06030847930908203, 0.06029312133789062, 0.06029209518432617, 0.060284927368164064, 0.0603054084777832, 0.06030847930908203, 0.06030950546264648, 0.06032998275756836, 0.06029721450805664, 0.06030950546264648, 0.060270591735839846, 0.06033919906616211, 0.06032998275756836, 0.06032076644897461, 0.06032179260253906, 0.060278785705566405, 0.06030847930908203, 0.06032793426513672, 0.06032998275756836, 0.060382209777832034, 0.06034124755859375, 0.060316673278808595, 0.060333057403564455, 0.0603422737121582, 0.06031872177124024, 0.060355583190917966, 0.06029107284545898, 0.060319744110107425, 0.06037811279296875, 0.06033203125, 0.06032691192626953, 0.06035148620605469, 0.060353534698486325, 0.0603351058959961, 0.06031257629394531, 0.06040063858032227, 0.06035456085205078, 0.06031564712524414, 0.06030438232421875, 0.06034739303588867, 0.06030233764648438, 0.06033407974243164, 0.06039244842529297, 0.06027980804443359, 0.06036377716064453, 0.06030745697021484, 0.060402687072753904, 0.06032691192626953, 0.06038016128540039, 0.13107916259765626, 0.06026444625854492, 0.06031052780151367, 0.06032588958740234, 0.06027571105957031, 0.06025113677978516, 0.060268543243408204, 0.060286975860595705, 0.06029312133789062, 0.06029619216918945, 0.06031155014038086, 0.06028902435302735, 0.060295169830322265, 0.06025932693481445, 0.06028799819946289, 0.06030745697021484, 0.060284927368164064, 0.06031257629394531, 0.060368896484375, 0.060286975860595705, 0.06031052780151367, 0.0603135986328125, 0.06030745697021484, 0.06029926300048828, 0.06032588958740234, 0.060368896484375, 0.06030950546264648, 0.060424190521240234, 0.060401664733886716, 0.060333057403564455, 0.06030950546264648, 0.06031564712524414, 0.06033407974243164, 0.06035456085205078, 0.06039961624145508, 0.06041190338134766, 0.06038016128540039, 0.0603351058959961, 0.060382209777832034, 0.06035660934448242, 0.06034329605102539, 0.06033407974243164, 0.06033407974243164, 0.06030950546264648, 0.06032076644897461, 0.06036479949951172, 0.06040371322631836, 0.06059929656982422, 0.06036275100708008, 0.06029721450805664, 0.060470272064208984, 0.060352512359619144, 0.060375038146972655, 0.060352512359619144, 0.06030131149291992, 0.06033919906616211, 0.060572673797607425, 0.06031769561767578, 0.060349441528320315, 0.06039244842529297, 0.06044979095458984, 0.06032076644897461, 0.0603873291015625, 0.1311068115234375, 0.06029312133789062, 0.06030438232421875, 0.060283905029296876, 0.060257278442382815, 0.06039449691772461, 0.06032486343383789, 0.06034431838989258, 0.06030847930908203, 0.060278785705566405, 0.060375038146972655, 0.06028799819946289, 0.060298240661621094, 0.060303359985351565, 0.060300289154052736, 0.06025932693481445, 0.0603135986328125, 0.06026342391967773, 0.060295169830322265, 0.060369918823242184, 0.06033817672729492, 0.060300289154052736, 0.06032486343383789, 0.06029619216918945, 0.06030745697021484, 0.06029414367675781, 0.060355583190917966, 0.06032793426513672, 0.06035148620605469, 0.060286975860595705, 0.06029721450805664, 0.0603135986328125, 0.06036275100708008, 0.06036070251464844, 0.06037401580810547, 0.06032793426513672, 0.06034124755859375, 0.06037606430053711, 0.060303359985351565, 0.06036070251464844, 0.06030950546264648, 0.06032486343383789, 0.06034329605102539, 0.060319744110107425, 0.06032281494140625, 0.06032793426513672, 0.06038016128540039, 0.06034124755859375, 0.06031872177124024, 0.060316673278808595, 0.06035763168334961, 0.060353534698486325, 0.060355583190917966, 0.060331008911132813, 0.06034739303588867, 0.06035660934448242, 0.06042521667480469, 0.060314624786376954, 0.060398593902587894, 0.06028902435302735, 0.06045183944702148, 0.06032076644897461, 0.06038425445556641, 0.1313116149902344, 0.06028902435302735, 0.06037299346923828, 0.060316673278808595, 0.06025830459594726, 0.060295169830322265, 0.06028799819946289, 0.0603054084777832, 0.06030950546264648, 0.060303359985351565, 0.06030643081665039, 0.06028902435302735, 0.06029107284545898, 0.06029209518432617, 0.06031052780151367, 0.060303359985351565, 0.06029414367675781, 0.06030233764648438, 0.060270591735839846, 0.060273662567138675, 0.060333057403564455, 0.06029926300048828, 0.06030233764648438, 0.06034431838989258, 0.06030131149291992, 0.06032486343383789, 0.06032281494140625, 0.06031564712524414, 0.060319744110107425, 0.06029721450805664, 0.06031564712524414, 0.060295169830322265, 0.060371967315673826, 0.06037299346923828, 0.06035456085205078, 0.060455936431884766, 0.060368896484375, 0.06031564712524414, 0.06030950546264648, 0.06032691192626953, 0.06038118362426758, 0.06035763168334961, 0.06029721450805664, 0.060352512359619144, 0.060388351440429686, 0.06037299346923828, 0.06037606430053711, 0.060453887939453124, 0.06036787033081055, 0.06032691192626953, 0.0603770866394043, 0.06034431838989258, 0.06031257629394531, 0.060352512359619144, 0.060286975860595705, 0.060371967315673826, 0.06049792098999023, 0.06034739303588867, 0.060371967315673826, 0.06029312133789062, 0.06042009735107422, 0.06035763168334961, 0.0603504638671875]",tokens/s,16.300702205052616,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5888,7 +5888,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6163,7 +6163,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1841.045504,9028.763648,0.0,8382.316544,7939.283456,s,10,6.1618914184570315,0.6161891418457032,0.0013886497977201196,0.6156345825195313,0.6173467590332031,0.6185957122802734,0.6195948748779296,"[0.6153511962890625, 0.6198446655273437, 0.6149873657226562, 0.6161400146484375, 0.6167808227539062, 0.6155103149414063, 0.6170692138671875, 0.6149386596679688, 0.6156962280273437, 0.6155729370117188]",tokens/s,415.45685020217974,kWh,7.266781396336026e-06,3.981886734928754e-06,3.646037230553567e-05,4.7709040436800454e-05,tokens/kWh,5365859.335174009,MB,1841.045504,9028.763648,0.0,8382.316544,8233.862144,s,10,338.14244921875,33.814244921875,0.0010794362638467628,33.814513671875005,33.8152484375,33.815645703125,33.815963515625,"[33.8121015625, 33.81604296875, 33.81308203125, 33.81325390625, 33.81455859375, 33.81446875, 33.81516015625, 33.81460546875, 33.8148515625, 33.81432421875]",tokens/s,1.863120118327535,kWh,0.00039921343515316647,0.00021880363846992518,0.0019968464210052606,0.002614863494628352,tokens/kWh,24093.03588100079,,s,629,343.15214111328146,0.5455518936618143,0.0732745798756584,0.5367347412109374,0.5369806762695313,0.5370492797851563,1.1527956982421876,"[0.5368463134765625, 0.5365708618164062, 0.5361151733398437, 0.5366968383789062, 0.5362565307617188, 0.5366865844726563, 0.53654833984375, 0.5366886596679687, 0.5362237548828125, 0.536764404296875, 0.5362780151367188, 0.5366343383789063, 0.5364541625976562, 0.5366149291992187, 0.5362943725585938, 0.5364500732421875, 0.5362340087890625, 0.5366415405273437, 0.536511474609375, 0.5367500610351562, 0.53686376953125, 0.536848388671875, 0.5365933837890625, 0.5368043823242188, 0.5367388305664063, 0.5369036865234375, 0.5367122192382813, 0.5366353759765625, 0.5366087646484375, 0.5367654418945312, 0.536764404296875, 0.5369047241210938, 0.536553466796875, 0.5369548950195312, 0.536637451171875, 0.5367030029296875, 0.5367500610351562, 0.5368914184570313, 0.5368442993164062, 0.5367337036132812, 0.536669189453125, 0.5367767333984375, 0.5367244873046875, 0.53667431640625, 0.5367030029296875, 0.5368524780273437, 0.5369886474609376, 0.5367562255859375, 0.5367418823242187, 0.5367562255859375, 0.5367296142578125, 0.5368053588867188, 0.5372672119140625, 0.536816650390625, 0.5366978759765625, 0.5367675170898437, 0.5367500610351562, 0.5367613525390625, 0.5365473022460937, 0.536869873046875, 0.536685546875, 0.5370091552734375, 1.1570892333984375, 0.5365841674804688, 0.536964111328125, 0.536543212890625, 0.5365770263671875, 0.5371688842773438, 0.5368514404296875, 0.5365718994140625, 0.5365770263671875, 0.536542236328125, 0.5368002319335937, 0.5366016235351563, 0.5365841674804688, 0.5364766845703125, 0.536605712890625, 0.5365811157226562, 0.536616943359375, 0.5363753051757812, 0.536859619140625, 0.5365606689453125, 0.53701123046875, 0.5366098022460938, 0.5368176879882812, 0.5366456298828125, 0.536791015625, 0.5367050170898438, 0.536838134765625, 0.5368258666992187, 0.5367030029296875, 0.5365985107421875, 0.5368995971679688, 0.5368719482421875, 0.5365944213867188, 0.5365964965820312, 0.5368248291015625, 0.5365964965820312, 0.5369528198242187, 0.5366548461914062, 0.5368555297851563, 0.5367551879882813, 0.5366599731445313, 0.5366917114257812, 0.5366702270507813, 0.5368176879882812, 0.5367398681640625, 0.5368494262695312, 0.5369815063476563, 0.5369231567382813, 0.53673779296875, 0.536648681640625, 0.5367817993164062, 0.5367122192382813, 0.5366988525390625, 0.5367613525390625, 0.5369108276367187, 0.536753173828125, 0.5366026000976563, 0.536859619140625, 0.5368678588867187, 0.5366640625, 0.5368094482421875, 0.5369251708984375, 0.5369794311523437, 1.1535380859375, 0.5367613525390625, 0.5365616455078125, 0.5365206909179687, 0.5365166015625, 0.536489990234375, 0.536605712890625, 0.5365125122070312, 0.53650634765625, 0.5366702270507813, 0.5363988647460938, 0.536427490234375, 0.5364992065429688, 0.536585205078125, 0.5364838256835938, 0.5364172973632813, 0.5366087646484375, 0.5363230590820313, 0.5367664794921875, 0.5370265502929688, 0.5367152709960937, 0.5365452880859375, 0.5364879150390625, 0.536574951171875, 0.5367408447265625, 0.5365831909179688, 0.5368402099609375, 0.5366200561523438, 0.53663232421875, 0.5366671142578125, 0.5365862426757813, 0.5366865844726563, 0.5367675170898437, 0.536543212890625, 0.537218017578125, 0.5368289184570313, 0.5368893432617188, 0.5365463256835937, 0.5369180297851562, 0.5366978759765625, 0.5367122192382813, 0.5366190185546875, 0.5367244873046875, 0.5367265014648438, 0.5367787475585938, 0.5369200439453125, 0.536953857421875, 0.5367817993164062, 0.5368955078125, 0.5371924438476563, 0.5367828369140625, 0.536700927734375, 0.5367869262695313, 0.5368555297851563, 0.5369343872070312, 0.5367613525390625, 0.5368616943359374, 0.5369343872070312, 0.5368299560546875, 0.5365944213867188, 0.5369395141601563, 0.5368370971679688, 0.5369733276367188, 1.153333251953125, 0.5366835327148437, 0.5368023071289062, 0.5364889526367187, 0.5365770263671875, 0.5365227661132812, 0.536642578125, 0.5365933837890625, 0.5365155639648438, 0.5364090576171875, 0.5365176391601563, 0.536680419921875, 0.5364449462890625, 0.5364029541015625, 0.5364592895507813, 0.5364039916992187, 0.5365104370117187, 0.536364013671875, 0.536543212890625, 0.5366814575195312, 0.5367203979492188, 0.5366784057617188, 0.53673779296875, 0.5366456298828125, 0.5368237915039062, 0.5368156127929687, 0.5368576049804688, 0.5364859008789062, 0.5366353759765625, 0.5367838745117187, 0.5365975341796875, 0.5367808227539063, 0.5367982177734375, 0.5364725952148437, 0.5369876708984375, 0.5365227661132812, 0.5368074340820312, 0.5367019653320313, 0.5367756958007812, 0.536685546875, 0.5368248291015625, 0.5367306518554688, 0.537006103515625, 0.536700927734375, 0.536791015625, 0.5369712524414062, 0.536995849609375, 0.536574951171875, 0.536764404296875, 0.5368340454101562, 0.5368104858398437, 0.5366251220703125, 0.5369047241210938, 0.5368985595703125, 0.536896484375, 0.53686376953125, 0.5370675048828125, 0.5368463134765625, 0.536816650390625, 0.53661083984375, 0.5369733276367188, 0.5368729858398438, 0.5369825439453125, 1.1542476806640625, 0.5366415405273437, 0.5366640625, 0.5365626831054687, 0.536474609375, 0.536574951171875, 0.5365555419921875, 0.5365391235351562, 0.53659033203125, 0.5364735717773438, 0.5364224243164063, 0.5364786987304687, 0.5366896362304687, 0.5363251342773437, 0.5369405517578125, 0.5366445922851563, 0.5364910278320313, 0.536426513671875, 0.5368678588867187, 0.5365780639648438, 0.536853515625, 0.5366998901367187, 0.5367551879882813, 0.5364920043945313, 0.5366149291992187, 0.53673779296875, 0.5368074340820312, 0.5366876220703125, 0.5368760375976562, 0.5368248291015625, 0.5367367553710938, 0.5367664794921875, 0.5367756958007812, 0.5366128540039062, 0.5368514404296875, 0.5366548461914062, 0.5368995971679688, 0.5366292724609375, 0.5369794311523437, 0.5365975341796875, 0.5368104858398437, 0.5366077270507813, 0.5368811645507813, 0.537017333984375, 0.5368330078125, 0.5368002319335937, 0.5368453369140626, 0.53680126953125, 0.536853515625, 0.5366364135742188, 0.5368237915039062, 0.5368207397460938, 0.5370203857421875, 0.5368842163085937, 0.5367879638671875, 0.536784912109375, 0.5369517822265625, 0.536859619140625, 0.536911865234375, 0.5366599731445313, 0.5369661254882813, 0.5370623779296875, 0.536890380859375, 1.152932861328125, 0.5365995483398438, 0.5366241455078125, 0.5368361206054687, 0.5364029541015625, 0.5365759887695313, 0.5366968383789062, 0.5365770263671875, 0.5366578979492187, 0.5363333129882812, 0.536784912109375, 0.5367633666992188, 0.5364306030273438, 0.5363978271484375, 0.5364786987304687, 0.53639990234375, 0.53650634765625, 0.5363681030273437, 0.5364818115234375, 0.5364930419921875, 0.5367562255859375, 0.5365759887695313, 0.5365831909179688, 0.5367992553710937, 0.5366712036132812, 0.5368299560546875, 0.5367675170898437, 0.5364879150390625, 0.5370848999023438, 0.5367859497070312, 0.5367285766601563, 0.536700927734375, 0.5369815063476563, 0.5364613037109375, 0.5366876220703125, 0.5365463256835937, 0.5369886474609376, 0.5367203979492188, 0.5369845581054687, 0.536690673828125, 0.5368616943359374, 0.5368944702148437, 0.536853515625, 0.5368790893554688, 0.5370951538085937, 0.5369600219726562, 0.5369047241210938, 0.53686474609375, 0.5367398681640625, 0.5367030029296875, 0.5368043823242188, 0.5367684936523438, 0.536774658203125, 0.536958984375, 0.5369774169921875, 0.5366917114257812, 0.5367500610351562, 0.536859619140625, 0.536932373046875, 0.5367869262695313, 0.5369671630859375, 0.5367613525390625, 0.537069580078125, 1.154186279296875, 0.5370890502929687, 0.5366292724609375, 0.536573974609375, 0.5365975341796875, 0.5366190185546875, 0.536500244140625, 0.5365022583007812, 0.53646337890625, 0.5364029541015625, 0.5364674682617188, 0.5364838256835938, 0.536489990234375, 0.5365524291992188, 0.5364920043945313, 0.5365155639648438, 0.5368043823242188, 0.5363507080078125, 0.5365288696289062, 0.536427490234375, 0.5368023071289062, 0.536543212890625, 0.5366497192382812, 0.53667431640625, 0.536791015625, 0.5367030029296875, 0.5367736206054687, 0.5366179809570313, 0.5368545532226563, 0.5366896362304687, 0.536679443359375, 0.536700927734375, 0.5370521850585938, 0.5366261596679688, 0.5369262084960937, 0.5365770263671875, 0.537017333984375, 0.5365524291992188, 0.5369661254882813, 0.5370050659179687, 0.5367940673828125, 0.5367890014648438, 0.5367664794921875, 0.5368299560546875, 0.5367787475585938, 0.5368770751953125, 0.5369937744140625, 0.5367050170898438, 0.5367930908203125, 0.5367347412109374, 0.5366651000976562, 0.536774658203125, 0.5371709594726563, 0.5370582885742188, 0.5370501098632813, 0.5367060546875, 0.5368678588867187, 0.5369467163085937, 0.5368309936523438, 0.53673779296875, 0.5371043701171875, 0.5370818481445313, 0.53697021484375, 1.15262255859375, 0.5366098022460938, 0.5367449340820313, 0.5365330200195313, 0.5364357299804687, 0.5369825439453125, 0.5364961547851562, 0.5365463256835937, 0.5367470092773438, 0.5364203491210937, 0.5364705200195312, 0.5365596313476563, 0.536605712890625, 0.5364869384765625, 0.536479736328125, 0.53650634765625, 0.5368033447265625, 0.536511474609375, 0.5366036376953125, 0.5365176391601563, 0.536658935546875, 0.5366558837890625, 0.5366282348632813, 0.5366927490234376, 0.5367890014648438, 0.5367767333984375, 0.5369415893554688, 0.5365975341796875, 0.5366558837890625, 0.536795166015625, 0.5368033447265625, 0.5368053588867188, 0.5367357177734375, 0.5365176391601563, 0.5369917602539063, 0.5366026000976563, 0.5368033447265625, 0.5366282348632813, 0.5369937744140625, 0.5366507568359375, 0.5368033447265625, 0.5367060546875, 0.536642578125, 0.5367367553710938, 0.5369609985351562, 0.5374689331054687, 0.5369559326171875, 0.5367736206054687, 0.5367306518554688, 0.5366753540039062, 0.5367930908203125, 0.5368616943359374, 0.5366712036132812, 0.536748046875, 0.5371443481445313, 0.5367633666992188, 0.53686376953125, 0.5369989013671875, 0.5367193603515625, 0.5363732299804688, 0.536826904296875, 0.5373204345703125, 0.5370480346679688, 1.1528192138671876, 0.53673779296875, 0.5365780639648438, 0.5367572631835937, 0.5364910278320313, 0.53661083984375, 0.5366067504882812, 0.536605712890625, 0.5366712036132812, 0.5365125122070312, 0.5364674682617188, 0.5367664794921875, 0.536774658203125, 0.5365452880859375, 0.5363015747070312, 0.53698046875, 0.5363916625976562, 0.5364777221679687, 0.5366456298828125, 0.5368361206054687, 0.536827880859375, 0.5365176391601563, 0.5365032958984375, 0.5367030029296875, 0.5367664794921875, 0.5367992553710937, 0.5368186645507812, 0.5367367553710938, 0.5366241455078125, 0.536700927734375, 0.536721435546875, 0.536690673828125, 0.536585205078125, 0.5365718994140625, 0.5372119140625, 0.5367449340820313, 0.536901611328125, 0.5367623901367188, 0.5369210815429688, 0.5369815063476563, 0.536816650390625, 0.5365401611328126, 0.536647705078125, 0.5367623901367188, 0.5368576049804688, 0.536985595703125, 0.5370214233398437, 0.5368043823242188, 0.5366978759765625, 0.5367767333984375, 0.5368760375976562, 0.5366179809570313, 0.5367091064453124, 0.5368545532226563, 0.5369886474609376, 0.53673779296875, 0.5370286254882812, 0.5367930908203125, 0.5368330078125, 0.5369763793945312, 0.5369528198242187, 0.5366415405273437, 0.53699072265625, 1.1527352294921875, 0.5367654418945312, 0.5367920532226562, 0.5366527709960938, 0.5365780639648438, 0.5366466674804687, 0.5366159057617188, 0.5365032958984375, 0.5366824951171875, 0.53646337890625, 0.5366988525390625, 0.536748046875, 0.5366405029296875, 0.536431640625, 0.5365883178710937, 0.5366179809570313, 0.5364172973632813, 0.536437744140625, 0.536605712890625, 0.536542236328125, 0.5368411865234375, 0.536489990234375, 0.5365667724609375, 0.5365811157226562, 0.5368033447265625, 0.5366507568359375, 0.5368831787109375, 0.5366231079101562, 0.53659033203125, 0.5368576049804688, 0.53665380859375, 0.5366968383789062, 0.536958984375, 0.5366138916015625, 0.5369302978515625, 0.5366302490234375, 0.5368606567382812, 0.5370828857421875, 0.5368504028320312, 0.5366947631835938, 0.5367859497070312, 0.5365924072265625, 0.5367418823242187, 0.5366651000976562, 0.536711181640625, 0.5369784545898437, 0.5370706176757812, 0.536806396484375, 0.537006103515625, 0.53665380859375, 0.536748046875, 0.5366722412109375, 0.536764404296875, 0.5368176879882812, 0.5370214233398437, 0.5366753540039062, 0.5369395141601563, 0.5368197021484375, 0.537064453125, 0.5366466674804687, 0.536995849609375, 0.5367654418945312, 0.536900634765625]",tokens/s,1.8330061935774278,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6192,7 +6192,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6351,7 +6351,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpoqk1nt7p/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6749,7 +6749,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpgyjxfwoq/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7690,7 +7690,7 @@ ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please req ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1269.682176,1019.74016,0.0,373.293056,277.410816,s,10,0.37923129653930665,0.037923129653930666,0.0012152886645715089,0.038048974990844725,0.038519168853759764,0.0396308479309082,0.04052019119262695,"[0.04074252700805664, 0.03821686553955078, 0.03797974395751953, 0.03823471832275391, 0.03626841735839844, 0.03608185577392578, 0.03733708953857422, 0.03811382293701172, 0.03827212905883789, 0.03798412704467773]",tokens/s,6750.497713035296,kWh,4.301884120160883e-07,2.3572264180256752e-07,1.164744366137984e-06,1.83065541995664e-06,tokens/kWh,139840626.0453229,MB,1269.682176,1019.74016,0.0,373.293056,323.047424,s,10,23.493592773437502,2.3493592773437504,0.014632907695307939,2.3561588134765623,2.362567529296875,2.3635712402343754,2.3643742089843753,"[2.34352734375, 2.36090478515625, 2.362344482421875, 2.364574951171875, 2.338309326171875, 2.331607421875, 2.357506103515625, 2.3608046875, 2.3548115234375, 2.3192021484375]",tokens/s,26.81582191687153,kWh,2.642717914328431e-05,1.4482863030778391e-05,6.568572578086276e-05,0.00010659576795492543,tokens/kWh,591017.8350292469,,s,629,23.794291763305676,0.037828762739754634,0.004511848848474209,0.03739033508300781,0.03767849044799805,0.03791810684204102,0.07427948760986337,"[0.03534745788574219, 0.03654348754882813, 0.03639910507202149, 0.03678003311157227, 0.03678003311157227, 0.036736000061035154, 0.036125694274902344, 0.03654348754882813, 0.036519935607910156, 0.0366110725402832, 0.03650764846801758, 0.03670425415039062, 0.03646771240234375, 0.037282817840576174, 0.03782860946655273, 0.03862527847290039, 0.037749759674072264, 0.03757056045532227, 0.037465087890625, 0.0372408332824707, 0.03734630584716797, 0.03750092697143555, 0.03727872085571289, 0.03741388702392578, 0.03740262222290039, 0.03726131057739258, 0.037303295135498044, 0.03731353759765625, 0.037188640594482424, 0.037268447875976565, 0.03742617416381836, 0.03730124664306641, 0.03739136123657227, 0.037250049591064455, 0.03735039901733399, 0.03733913421630859, 0.03732787322998047, 0.03716505432128906, 0.03726233673095703, 0.0372490234375, 0.03732070541381836, 0.03729305648803711, 0.03743129730224609, 0.03723980712890625, 0.037269504547119144, 0.03735039901733399, 0.03727769470214844, 0.03724800109863281, 0.03734124755859375, 0.037339073181152344, 0.037269504547119144, 0.03736678314208984, 0.0371701774597168, 0.037266433715820314, 0.03737190246582031, 0.037318656921386716, 0.03726847839355469, 0.03745894241333008, 0.03732787322998047, 0.03729715347290039, 0.037501953125, 0.037501953125, 0.07589580535888672, 0.03728793716430664, 0.037407745361328126, 0.037207038879394534, 0.03731353759765625, 0.03784089660644531, 0.037353473663330077, 0.03739340972900391, 0.0374015998840332, 0.03751731109619141, 0.03856486511230469, 0.037754878997802735, 0.037369857788085936, 0.03734220886230469, 0.03731353759765625, 0.0373125114440918, 0.03733606338500976, 0.037425151824951174, 0.037303295135498044, 0.037362686157226564, 0.038662143707275394, 0.037748737335205076, 0.03751116943359375, 0.03752243041992188, 0.03749785614013672, 0.03739340972900391, 0.03728793716430664, 0.03726540756225586, 0.03750707244873047, 0.03739340972900391, 0.03738828659057617, 0.03735859298706055, 0.037395454406738284, 0.03731558227539063, 0.037868545532226565, 0.037456897735595705, 0.03736883163452148, 0.03723161697387695, 0.03758489608764649, 0.03761151885986328, 0.03742822265625, 0.037384193420410154, 0.03741900634765625, 0.03744976043701172, 0.03728688049316406, 0.03746815872192383, 0.03732275390625, 0.037525505065917966, 0.0375203857421875, 0.03739340972900391, 0.03733913421630859, 0.03781631851196289, 0.03749785614013672, 0.03746815872192383, 0.03744870376586914, 0.037324798583984374, 0.03752345657348633, 0.03728691101074219, 0.0374015998840332, 0.03738828659057617, 0.037454849243164064, 0.037454849243164064, 0.03749273681640625, 0.0757176284790039, 0.03736166381835938, 0.03737702560424805, 0.03760332870483398, 0.03744768142700195, 0.037354496002197264, 0.03724595260620117, 0.037343231201171875, 0.03800678253173828, 0.03777740859985351, 0.03753472137451172, 0.03747430419921875, 0.0374835205078125, 0.037585918426513674, 0.037338111877441404, 0.0374466552734375, 0.03750912094116211, 0.037379070281982424, 0.03765964889526367, 0.03735551834106445, 0.03742310333251953, 0.03738726425170898, 0.03740364837646484, 0.037495807647705076, 0.03738214492797851, 0.037310462951660156, 0.03733401489257813, 0.03747430419921875, 0.03743436813354492, 0.0373831672668457, 0.03754393768310547, 0.037370880126953124, 0.03760025787353516, 0.037601280212402347, 0.037510143280029294, 0.0373043212890625, 0.0373125114440918, 0.03747225570678711, 0.03762688064575195, 0.037282817840576174, 0.037294078826904296, 0.03753267288208008, 0.037607425689697264, 0.037400577545166014, 0.037495807647705076, 0.03745075225830078, 0.038043647766113284, 0.037773311614990236, 0.03741798400878906, 0.03741183853149414, 0.03771289443969727, 0.03767193603515625, 0.03745280075073242, 0.0372674560546875, 0.03750092697143555, 0.038141952514648435, 0.037969921112060545, 0.037384193420410154, 0.03739852905273437, 0.03727977752685547, 0.03753366470336914, 0.037572608947753904, 0.03738726425170898, 0.07624192047119141, 0.03899084854125977, 0.037817344665527344, 0.03795356750488281, 0.03747836685180664, 0.03777536010742188, 0.03754905700683594, 0.03737497711181641, 0.03731763076782227, 0.03754598236083984, 0.0381317138671875, 0.037395454406738284, 0.03748556900024414, 0.03738726425170898, 0.037493759155273435, 0.037318656921386716, 0.03763916778564453, 0.03733299255371094, 0.037407745361328126, 0.03729817581176758, 0.03746918487548828, 0.03743129730224609, 0.03740364837646484, 0.03732275390625, 0.03753267288208008, 0.03765350341796875, 0.03748863983154297, 0.03742617416381836, 0.03767708969116211, 0.03730940628051758, 0.03756748962402344, 0.03845017623901367, 0.03810201644897461, 0.03751321411132812, 0.03726847839355469, 0.03728998565673828, 0.03749273681640625, 0.03731148910522461, 0.03733606338500976, 0.03725107192993164, 0.03753881454467774, 0.03738726425170898, 0.037282817840576174, 0.03728179168701172, 0.0374917106628418, 0.037351425170898435, 0.03746819305419922, 0.037434337615966796, 0.037440513610839846, 0.03745587158203125, 0.03747532653808594, 0.03769651031494141, 0.03751935958862305, 0.037397502899169925, 0.03741491317749023, 0.037466110229492186, 0.03749478530883789, 0.037566463470458986, 0.03733401489257813, 0.03742822265625, 0.03756851196289063, 0.037501953125, 0.03759820938110352, 0.07184793853759766, 0.03510374450683594, 0.03521535873413086, 0.03516928100585937, 0.03511603164672852, 0.03513241577148438, 0.03520000076293945, 0.035148799896240236, 0.03514470291137695, 0.03511500930786133, 0.0367718391418457, 0.03750400161743164, 0.03728998565673828, 0.037482494354248046, 0.0374835205078125, 0.03756032180786133, 0.037303295135498044, 0.037424129486083986, 0.03768729782104492, 0.03739648056030274, 0.03758492660522461, 0.03751830291748047, 0.03748659133911133, 0.03748863983154297, 0.037410816192626956, 0.03736883163452148, 0.03754396820068359, 0.03754390335083008, 0.037591041564941405, 0.03752140808105469, 0.03739136123657227, 0.03755212783813477, 0.03739340972900391, 0.03741491317749023, 0.03769139099121094, 0.037408767700195314, 0.038091777801513675, 0.03745382308959961, 0.0376360969543457, 0.03739136123657227, 0.03732070541381836, 0.03746303939819336, 0.03743334579467773, 0.037364734649658206, 0.037526527404785154, 0.0374015998840332, 0.037416961669921874, 0.037416961669921874, 0.03750604629516602, 0.037427200317382815, 0.037367809295654295, 0.037498878479003905, 0.037397502899169925, 0.03749683380126953, 0.037622814178466794, 0.03773948669433594, 0.037416961669921874, 0.03744464111328125, 0.03751728057861328, 0.0374835205078125, 0.03757056045532227, 0.037477375030517575, 0.037720062255859374, 0.0716236801147461, 0.035062782287597655, 0.03518668746948242, 0.035019775390625, 0.03531980895996094, 0.035253246307373046, 0.03521843338012695, 0.034840576171875, 0.03494911956787109, 0.03526863861083984, 0.03520406341552734, 0.035253246307373046, 0.03508428955078125, 0.0372490234375, 0.037318656921386716, 0.03726335906982422, 0.03732787322998047, 0.03786342239379883, 0.037288959503173826, 0.0374835205078125, 0.03732992172241211, 0.03781631851196289, 0.038765567779541016, 0.03768012619018555, 0.03770982360839844, 0.037367809295654295, 0.03745894241333008, 0.03721318435668945, 0.03721420669555664, 0.03736678314208984, 0.03769343948364258, 0.03735756683349609, 0.03937484741210937, 0.03793203353881836, 0.037556224822998044, 0.03736883163452148, 0.03729510498046875, 0.03745177459716797, 0.03745792007446289, 0.037561344146728515, 0.037422080993652344, 0.037353473663330077, 0.037594112396240234, 0.03732070541381836, 0.03734732818603516, 0.037340160369873046, 0.037397502899169925, 0.037362686157226564, 0.03757875061035156, 0.03736064147949219, 0.037321727752685545, 0.03742822265625, 0.03740364837646484, 0.037318656921386716, 0.03724492645263672, 0.03730022430419922, 0.03753472137451172, 0.0374200325012207, 0.03728486251831055, 0.03722751998901367, 0.03739136123657227, 0.037362686157226564, 0.037375999450683595, 0.07522509002685547, 0.03725209426879883, 0.03725823974609375, 0.0374466552734375, 0.03733916854858398, 0.03731145477294922, 0.03730124664306641, 0.03746406555175781, 0.037375999450683595, 0.03750707244873047, 0.037292030334472655, 0.037459968566894535, 0.03742924880981445, 0.037405696868896485, 0.03738009643554688, 0.0372592658996582, 0.03754905700683594, 0.03736067199707031, 0.03739644622802735, 0.037394432067871096, 0.037394432067871096, 0.037392383575439454, 0.037389312744140625, 0.0372674560546875, 0.03746713638305664, 0.037416961669921874, 0.03733606338500976, 0.03751321411132812, 0.03757056045532227, 0.037351425170898435, 0.03736883163452148, 0.037335041046142575, 0.03735756683349609, 0.03738521575927734, 0.03734630584716797, 0.03721932983398438, 0.037397502899169925, 0.03740364837646484, 0.037465087890625, 0.03727974319458008, 0.03751116943359375, 0.03736883163452148, 0.03720294570922852, 0.03728179168701172, 0.03764019012451172, 0.037395454406738284, 0.037294078826904296, 0.037379070281982424, 0.03743231964111328, 0.03737705612182617, 0.03722953414916992, 0.03727360153198242, 0.03775692749023438, 0.03735039901733399, 0.03786652755737305, 0.03738211059570312, 0.03764019012451172, 0.03744255828857422, 0.03744153594970703, 0.03740671920776367, 0.038042625427246096, 0.03756032180786133, 0.03753267288208008, 0.07615692901611328, 0.03728486251831055, 0.03751628875732422, 0.03750092697143555, 0.03723161697387695, 0.03742105484008789, 0.03728384017944336, 0.03728793716430664, 0.03723369598388672, 0.037351390838623044, 0.0372592658996582, 0.03703603363037109, 0.03736576080322265, 0.03826483154296875, 0.03825356674194336, 0.037498878479003905, 0.03745075225830078, 0.03750604629516602, 0.03733606338500976, 0.0373493766784668, 0.037405696868896485, 0.037548030853271484, 0.037897216796875, 0.0374015998840332, 0.03749273681640625, 0.03787980651855469, 0.03769548797607422, 0.03733401489257813, 0.037424129486083986, 0.03724288177490234, 0.03732275390625, 0.03731353759765625, 0.037348350524902346, 0.0373043212890625, 0.03746099090576172, 0.03751731109619141, 0.037424129486083986, 0.037477375030517575, 0.03738623809814453, 0.037395454406738284, 0.0372828483581543, 0.0373001937866211, 0.03731455993652344, 0.03749683380126953, 0.037678081512451174, 0.03893964767456055, 0.037498878479003905, 0.03726847839355469, 0.03724595260620117, 0.037381118774414065, 0.03728998565673828, 0.03745177459716797, 0.03747635269165039, 0.037425151824951174, 0.03755724716186523, 0.03739033508300781, 0.03731763076782227, 0.03747532653808594, 0.03732070541381836, 0.03763302230834961, 0.0374917106628418, 0.03746303939819336, 0.037392383575439454, 0.07604838562011719, 0.03738726425170898, 0.03732275390625, 0.037763072967529294, 0.03760537719726562, 0.037493759155273435, 0.03744153594970703, 0.037367809295654295, 0.037353473663330077, 0.03748966217041016, 0.037321727752685545, 0.03739136123657227, 0.03747020721435547, 0.03746201705932617, 0.03758899307250976, 0.037288959503173826, 0.03742617416381836, 0.037359615325927735, 0.03741900634765625, 0.03733708953857422, 0.0372592658996582, 0.03737497711181641, 0.03758182525634766, 0.03737395095825195, 0.03727052688598633, 0.03736678314208984, 0.037477375030517575, 0.03746201705932617, 0.03737702560424805, 0.03690291213989258, 0.03730022430419922, 0.037541889190673826, 0.03752140808105469, 0.03749683380126953, 0.03740979385375977, 0.037354496002197264, 0.037292030334472655, 0.03724288177490234, 0.03722956848144531, 0.03722956848144531, 0.0375203857421875, 0.03734630584716797, 0.03717836761474609, 0.0371671028137207, 0.037372928619384765, 0.037341182708740234, 0.03730636978149414, 0.03713945770263672, 0.03719987106323242, 0.03745177459716797, 0.03732889556884766, 0.03736166381835938, 0.0373493766784668, 0.03739648056030274, 0.03730739212036133, 0.03730739212036133, 0.03730944061279297, 0.03724390411376953, 0.037353473663330077, 0.03740467071533203, 0.0375203857421875, 0.03734630584716797, 0.03746406555175781, 0.07565721893310547, 0.0372674560546875, 0.0373125114440918, 0.037479423522949216, 0.03725721740722656, 0.03723468780517578, 0.0374200325012207, 0.03733196640014649, 0.03729817581176758, 0.03751321411132812, 0.03730739212036133, 0.03743334579467773, 0.03721932983398438, 0.03751833724975586, 0.03735244750976562, 0.037408767700195314, 0.03565260696411133, 0.035312641143798826, 0.03527782440185547, 0.0352542724609375, 0.03772415924072266, 0.0377077751159668, 0.03799244689941406, 0.03948953628540039, 0.037106689453125, 0.03671449661254883, 0.03655372619628906, 0.03666534423828125, 0.03620249557495117, 0.03514470291137695, 0.03519692611694336, 0.035335166931152344, 0.03523788833618164, 0.03600076675415039, 0.03778662490844727, 0.03708927917480469, 0.03679846572875976, 0.036724769592285156, 0.03671651077270508, 0.0366192626953125, 0.036624416351318356, 0.03651375961303711, 0.03685990524291992, 0.03674524688720703, 0.03836924743652344, 0.03679436874389649, 0.03653734588623047, 0.03676671981811523, 0.03650559997558594, 0.03653529739379883, 0.036706302642822264, 0.03682099151611328, 0.036528129577636716, 0.03685887908935547, 0.03667865753173828, 0.036674560546875, 0.036618240356445314, 0.03663052749633789, 0.03661721420288086, 0.03660595321655274, 0.0365926399230957, 0.036544513702392575, 0.036792320251464845]",tokens/s,26.434911627418618,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2012.44672,5274.861568,0.0,4628.414464,4279.515648,s,10,3.0981137390136726,0.3098113739013672,0.0009457574536517518,0.3095596923828125,0.31112299194335935,0.3115160552978516,0.31183050598144535,"[0.3090160217285156, 0.3099482421875, 0.3095133972167969, 0.3094310302734375, 0.3096059875488281, 0.30901934814453125, 0.31002566528320313, 0.3086092834472656, 0.3119091186523438, 0.31103564453125]",tokens/s,826.3092370569365,kWh,3.6482026631181896e-06,1.9990483218862206e-06,1.7856563106788066e-05,2.3503814091792478e-05,tokens/kWh,10891849.254772445,MB,2012.44672,5274.861568,0.0,4628.414464,4463.184384,s,10,173.15158984375,17.315158984375003,0.0008435493287116899,17.3153154296875,17.316042382812498,17.31623310546875,17.31638568359375,"[17.31562890625, 17.316, 17.314966796875, 17.31573828125, 17.3152734375, 17.314123046875, 17.314578125, 17.315357421875, 17.316423828125, 17.3135]",tokens/s,3.638430352089199,kWh,0.0002044354933214308,0.00011204598750467993,0.0009878629636904115,0.0013043444445165224,tokens/kWh,48300.12521987782,,s,629,175.66576879882834,0.2792778518264358,0.036821094771226096,0.274830322265625,0.27505172119140625,0.27513446044921874,0.5842676342773437,"[0.2750351257324219, 0.27489178466796876, 0.274735107421875, 0.274735107421875, 0.2746921081542969, 0.27469003295898436, 0.2748016662597656, 0.2746644592285156, 0.2747750549316406, 0.274651123046875, 0.2747361145019531, 0.2748713073730469, 0.2746357727050781, 0.27475662231445314, 0.2746726379394531, 0.274840576171875, 0.27462655639648437, 0.27478732299804687, 0.27498086547851563, 0.27470437622070315, 0.2747474060058594, 0.2747381896972656, 0.27487026977539064, 0.27473919677734376, 0.27474533081054686, 0.27481497192382814, 0.27486822509765624, 0.27485592651367186, 0.27480474853515624, 0.27489178466796876, 0.2747463684082031, 0.27480062866210936, 0.2747781066894531, 0.2749091796875, 0.27479757690429685, 0.274808837890625, 0.2748692626953125, 0.27493991088867187, 0.27494195556640627, 0.27483444213867186, 0.275083251953125, 0.2749153137207031, 0.2749696044921875, 0.27500238037109376, 0.27485491943359375, 0.274951171875, 0.27491738891601564, 0.2748160095214844, 0.2748252258300781, 0.27477914428710937, 0.27466546630859373, 0.27490304565429685, 0.27493991088867187, 0.2749286499023437, 0.2748641357421875, 0.27500338745117187, 0.274850830078125, 0.2748538818359375, 0.2749900817871094, 0.27501055908203126, 0.2749440002441406, 0.27534439086914064, 0.5850224609375, 0.27484774780273435, 0.2746152954101562, 0.27474533081054686, 0.2746490783691406, 0.27462554931640626, 0.2749921264648437, 0.27478118896484377, 0.27476480102539064, 0.2747699279785156, 0.2747268981933594, 0.2748231811523438, 0.27470437622070315, 0.27464602661132814, 0.27498904418945314, 0.2747268981933594, 0.27469720458984376, 0.27473919677734376, 0.2747781066894531, 0.2746091613769531, 0.27474227905273435, 0.2748487548828125, 0.27456307983398437, 0.2747607116699219, 0.2747002868652344, 0.2747279357910156, 0.2747740173339844, 0.274798583984375, 0.27483444213867186, 0.27479962158203125, 0.27488357543945313, 0.2746152954101562, 0.2749706115722656, 0.275030029296875, 0.275135498046875, 0.27472897338867186, 0.27489279174804687, 0.2747115478515625, 0.2752081909179688, 0.2749573059082031, 0.2748231811523438, 0.2751098937988281, 0.274830322265625, 0.27489178466796876, 0.27508633422851564, 0.2750218200683594, 0.27489178466796876, 0.27495526123046876, 0.2749982604980469, 0.274872314453125, 0.2750330810546875, 0.2748805236816406, 0.27481805419921873, 0.27501568603515625, 0.27499725341796877, 0.27485696411132815, 0.27509146118164063, 0.2748487548828125, 0.27496652221679685, 0.2749593505859375, 0.27505868530273436, 0.27497369384765624, 0.2752174072265625, 0.5844049682617187, 0.2746695556640625, 0.274724853515625, 0.27476171875, 0.2747268981933594, 0.27471768188476564, 0.2747432861328125, 0.27462655639648437, 0.27470233154296875, 0.2746142578125, 0.27483444213867186, 0.27469619750976565, 0.274808837890625, 0.274740234375, 0.2747002868652344, 0.27517645263671875, 0.2746726379394531, 0.27463885498046875, 0.2747054138183594, 0.27489178466796876, 0.2747258911132813, 0.27476480102539064, 0.2746378173828125, 0.27478118896484377, 0.27469003295898436, 0.2747064208984375, 0.2747555847167969, 0.27474432373046875, 0.27460198974609373, 0.27469003295898436, 0.2750320739746094, 0.2746746826171875, 0.27475149536132815, 0.27477093505859373, 0.274798583984375, 0.27475762939453124, 0.27493374633789064, 0.27482623291015623, 0.27482623291015623, 0.27523379516601565, 0.2749716491699219, 0.27489178466796876, 0.27494708251953126, 0.27481805419921873, 0.2751744079589844, 0.275030029296875, 0.2748907470703125, 0.27488357543945313, 0.2750771179199219, 0.27486004638671874, 0.27486514282226565, 0.27501873779296876, 0.27495834350585935, 0.2748692626953125, 0.27510885620117187, 0.27491021728515624, 0.2748241882324219, 0.27502590942382815, 0.27491943359375, 0.2753228759765625, 0.27508428955078124, 0.27493887329101563, 0.27487026977539064, 0.5844735717773437, 0.2747054138183594, 0.2746378173828125, 0.274555908203125, 0.2748323974609375, 0.2746695556640625, 0.274724853515625, 0.27464190673828126, 0.2747678833007812, 0.2747187194824219, 0.2747412414550781, 0.274830322265625, 0.2746726379394531, 0.27480780029296875, 0.2746644592285156, 0.2751068115234375, 0.2747135925292969, 0.2747135925292969, 0.27464498901367185, 0.2746429443359375, 0.2747914123535156, 0.27477914428710937, 0.27488357543945313, 0.2748436584472656, 0.27502590942382815, 0.2748436584472656, 0.2747268981933594, 0.2746992492675781, 0.27483444213867186, 0.274693115234375, 0.2747832336425781, 0.27517236328125, 0.27475662231445314, 0.2747555847167969, 0.27486822509765624, 0.27485491943359375, 0.274951171875, 0.274946044921875, 0.2747904052734375, 0.274951171875, 0.27484161376953126, 0.27482623291015623, 0.27490509033203125, 0.27496755981445314, 0.27509963989257813, 0.2749859924316406, 0.2749716491699219, 0.274808837890625, 0.27492352294921873, 0.27492044067382815, 0.2748436584472656, 0.2750330810546875, 0.27496551513671874, 0.27488973999023436, 0.2748108825683594, 0.2751109008789063, 0.2750556030273438, 0.27504229736328123, 0.2751866760253906, 0.27491226196289065, 0.2749624328613281, 0.27511398315429686, 0.27482623291015623, 0.5845493774414062, 0.27454156494140625, 0.27462451171875, 0.2746767272949219, 0.27461221313476564, 0.274740234375, 0.2747412414550781, 0.27475762939453124, 0.2746275939941406, 0.2747412414550781, 0.27471563720703124, 0.274661376953125, 0.2746071166992187, 0.2746777648925781, 0.27463168334960936, 0.2747688903808594, 0.274735107421875, 0.27458251953125, 0.27470745849609374, 0.2746746826171875, 0.27465625, 0.2746163330078125, 0.2748231811523438, 0.2747340698242188, 0.27481396484375, 0.27505767822265625, 0.274723876953125, 0.27478421020507815, 0.2748016662597656, 0.27497882080078123, 0.2749859924316406, 0.2747699279785156, 0.2748671875, 0.2748887023925781, 0.2750474243164063, 0.27465728759765623, 0.2749900817871094, 0.27487335205078123, 0.27479962158203125, 0.2749266052246094, 0.2752122802734375, 0.274798583984375, 0.27484161376953126, 0.2749644775390625, 0.2748784790039063, 0.2749716491699219, 0.2749306945800781, 0.27479757690429685, 0.275051513671875, 0.275019775390625, 0.2748436584472656, 0.2749941711425781, 0.27496551513671874, 0.2748098449707031, 0.2751098937988281, 0.2750474243164063, 0.27502590942382815, 0.27502285766601564, 0.27510580444335936, 0.2749747314453125, 0.2753966064453125, 0.27505255126953126, 0.27491839599609375, 0.5839144897460937, 0.27459890747070315, 0.27473101806640626, 0.274634765625, 0.2746849365234375, 0.27491943359375, 0.2749634704589844, 0.27464398193359374, 0.27470950317382814, 0.274671630859375, 0.2746480712890625, 0.27475250244140625, 0.2749429626464844, 0.2747258911132813, 0.2747258911132813, 0.27462451171875, 0.27475045776367185, 0.27468084716796876, 0.2746275939941406, 0.27471563720703124, 0.27465933227539063, 0.27472384643554687, 0.27456103515625, 0.2747381896972656, 0.27501773071289065, 0.27480062866210936, 0.27477093505859373, 0.2747555847167969, 0.27470950317382814, 0.27481396484375, 0.2747105407714844, 0.27481497192382814, 0.27507302856445315, 0.27469720458984376, 0.27496652221679685, 0.274840576171875, 0.27497369384765624, 0.2749081726074219, 0.27491326904296876, 0.2748784790039063, 0.27476275634765623, 0.2748641357421875, 0.2749542541503906, 0.2748293151855469, 0.27482009887695313, 0.2748272705078125, 0.2749962158203125, 0.27489178466796876, 0.27492352294921873, 0.27513446044921874, 0.2748805236816406, 0.27491326904296876, 0.2747801513671875, 0.27504434204101563, 0.27497369384765624, 0.2749910888671875, 0.2750679016113281, 0.27495526123046876, 0.27496551513671874, 0.2748948364257813, 0.27491326904296876, 0.27494503784179686, 0.27482827758789063, 0.5851586303710937, 0.27469720458984376, 0.2747381896972656, 0.2747084655761719, 0.274798583984375, 0.27487335205078123, 0.274951171875, 0.2746634216308594, 0.2746368103027344, 0.2747054138183594, 0.274619384765625, 0.2746009521484375, 0.27468389892578127, 0.2746132507324219, 0.27507098388671875, 0.27484161376953126, 0.2746378173828125, 0.2746726379394531, 0.27468698120117185, 0.2747781066894531, 0.27464703369140625, 0.27478427124023436, 0.2746921081542969, 0.2746777648925781, 0.2748323974609375, 0.27475149536132815, 0.2747678833007812, 0.27469415283203125, 0.27480575561523435, 0.2747852783203125, 0.2747883605957031, 0.27465420532226564, 0.2747821960449219, 0.2746368103027344, 0.27494195556640627, 0.274830322265625, 0.27488153076171873, 0.27490509033203125, 0.274777099609375, 0.27503103637695314, 0.27477297973632814, 0.27513446044921874, 0.2748323974609375, 0.2748661804199219, 0.2748856201171875, 0.27483648681640627, 0.27477914428710937, 0.27497882080078123, 0.27492556762695314, 0.2749020080566406, 0.2750597229003906, 0.27492352294921873, 0.274882568359375, 0.274956298828125, 0.2749982604980469, 0.2749716491699219, 0.2750351257324219, 0.2753280029296875, 0.2750433349609375, 0.27495321655273436, 0.27487744140625, 0.2749931640625, 0.2749010009765625, 0.583531494140625, 0.2746480712890625, 0.27469003295898436, 0.27469619750976565, 0.2748661804199219, 0.27466650390625, 0.27472076416015623, 0.27455487060546874, 0.2746767272949219, 0.27469619750976565, 0.2746142578125, 0.2748641357421875, 0.2747432861328125, 0.27524609375, 0.2746132507324219, 0.27479449462890626, 0.274740234375, 0.2750003051757812, 0.27491839599609375, 0.2748252258300781, 0.27477197265625, 0.27485592651367186, 0.274724853515625, 0.27488357543945313, 0.2749091796875, 0.27475762939453124, 0.27470745849609374, 0.27477093505859373, 0.27474227905273435, 0.2748016662597656, 0.2748610534667969, 0.27486309814453125, 0.2747187194824219, 0.2747607116699219, 0.2748374938964844, 0.27483544921875, 0.27485696411132815, 0.2749071350097656, 0.27478732299804687, 0.275030029296875, 0.2747340698242188, 0.2749521789550781, 0.2749276123046875, 0.27486004638671874, 0.2748692626953125, 0.2750320739746094, 0.2748671875, 0.27484979248046876, 0.274956298828125, 0.27482623291015623, 0.27489279174804687, 0.2752983093261719, 0.27487640380859374, 0.2749521789550781, 0.2748948364257813, 0.2748518371582031, 0.2748876647949219, 0.2750433349609375, 0.2751068115234375, 0.27513446044921874, 0.27488153076171873, 0.27497268676757813, 0.27485592651367186, 0.5868973999023438, 0.2745733032226563, 0.2749358215332031, 0.2747268981933594, 0.2746163330078125, 0.2747258911132813, 0.2748098449707031, 0.2746552429199219, 0.2747821960449219, 0.2746501159667969, 0.2746081237792969, 0.2747218017578125, 0.2747586669921875, 0.2747258911132813, 0.2748395385742187, 0.27473202514648437, 0.27477297973632814, 0.27475149536132815, 0.27478118896484377, 0.27469415283203125, 0.2749081726074219, 0.2747852783203125, 0.27475967407226565, 0.2751600646972656, 0.27488973999023436, 0.27475662231445314, 0.27469830322265626, 0.27466439819335936, 0.27470745849609374, 0.2747054138183594, 0.2749347839355469, 0.2748016662597656, 0.2751201171875, 0.2747658386230469, 0.2747484130859375, 0.27480474853515624, 0.2748467102050781, 0.27473101806640626, 0.2752983093261719, 0.2748221435546875, 0.27470950317382814, 0.2752850036621094, 0.2749747314453125, 0.2748385314941406, 0.27491738891601564, 0.274850830078125, 0.2748784790039063, 0.2749634704589844, 0.27494503784179686, 0.27479244995117186, 0.27507302856445315, 0.2748968811035156, 0.27493991088867187, 0.2751795349121094, 0.2749296569824219, 0.2749214782714844, 0.27510784912109376, 0.2749081726074219, 0.27524609375, 0.2750965881347656, 0.27500338745117187, 0.2748989562988281, 0.2753546142578125, 0.5860249633789063, 0.27464190673828126, 0.27466650390625, 0.27468389892578127, 0.275125244140625, 0.27456512451171877, 0.27472280883789063, 0.27463168334960936, 0.27477197265625, 0.274661376953125, 0.27458868408203124, 0.27483648681640627, 0.27479244995117186, 0.2745937805175781, 0.27466033935546874, 0.2747381896972656, 0.27474533081054686, 0.2747330627441406, 0.2748692626953125, 0.27472384643554687, 0.2747821960449219, 0.27471563720703124, 0.2746849365234375, 0.27464498901367185, 0.2747750549316406, 0.2746859436035156, 0.27498086547851563, 0.27475662231445314, 0.27458660888671876, 0.2746746826171875, 0.27504537963867187, 0.27469003295898436, 0.27511602783203126, 0.2748887023925781, 0.2747821960449219, 0.2747197570800781, 0.2750822448730469, 0.2747740173339844, 0.2747607116699219, 0.27490509033203125, 0.27476171875, 0.27485491943359375, 0.27493170166015624, 0.27483544921875, 0.2748590087890625, 0.27491122436523435, 0.27494503784179686, 0.2748221435546875, 0.27496038818359375, 0.2748252258300781, 0.2748467102050781, 0.2750003051757812, 0.27494708251953126, 0.2748610534667969, 0.2749634704589844, 0.2748293151855469, 0.27484979248046876, 0.2749798278808594, 0.27497369384765624, 0.274904052734375, 0.275056640625, 0.2749020080566406, 0.27495321655273436]",tokens/s,3.5806634627849943,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8223,7 +8223,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1638.6048,2418.540544,0.0,1772.09344,1621.541376,s,10,0.7307771377563477,0.07307771377563477,0.00030590186697528017,0.07303974151611328,0.07345188369750977,0.07350205345153808,0.07354218925476075,"[0.07344073486328125, 0.07282319641113282, 0.0727369613647461, 0.07273868560791015, 0.07273926544189453, 0.07293727874755859, 0.07314220428466797, 0.07324988555908203, 0.07355222320556641, 0.07341670227050781]",tokens/s,3503.119990671552,kWh,8.589871899613245e-07,4.704820193973832e-07,3.7392864052756902e-06,5.068755614634398e-06,tokens/kWh,50505492.76056682,MB,1638.6048,2418.540544,0.0,1772.09344,1723.273216,s,10,38.20281567382813,3.8202815673828128,0.0022956512254473085,3.8195750732421874,3.82337958984375,3.8243252929687497,3.82508185546875,"[3.82527099609375, 3.81770703125, 3.819302978515625, 3.818822509765625, 3.82106005859375, 3.82316943359375, 3.821086669921875, 3.81984716796875, 3.817920166015625, 3.818628662109375]",tokens/s,16.49093107112517,kWh,4.511295209232352e-05,2.472463132112629e-05,0.00019686329074332517,0.000266700874156775,tokens/kWh,236219.69818879056,,s,629,38.79888072204587,0.06168343517018425,0.008717620254388004,0.06061260986328125,0.06073487243652344,0.060805118560791016,0.1336954461669922,"[0.06117068862915039, 0.061228031158447264, 0.0611962890625, 0.06114406585693359, 0.06115327835083008, 0.06097817611694336, 0.06119424057006836, 0.06067814254760742, 0.060709888458251954, 0.06057062530517578, 0.060631038665771485, 0.060595199584960936, 0.06056959915161133, 0.06056959915161133, 0.06066790390014649, 0.06054707336425781, 0.0605849609375, 0.06076313781738281, 0.06055321502685547, 0.06061670303344727, 0.06071705627441406, 0.06055321502685547, 0.060527614593505856, 0.06062387084960937, 0.060576766967773435, 0.060677120208740234, 0.0607006721496582, 0.06059929656982422, 0.0606033935546875, 0.06082559967041016, 0.060639232635498044, 0.060633087158203126, 0.06068326568603516, 0.060611583709716796, 0.06060236740112305, 0.06072524642944336, 0.060627967834472656, 0.060657665252685546, 0.06071603012084961, 0.06063513565063477, 0.06063616180419922, 0.060747776031494144, 0.06059212875366211, 0.06063513565063477, 0.06071603012084961, 0.06056243133544922, 0.06069452667236328, 0.06072012710571289, 0.06063411331176758, 0.06065868759155273, 0.06072012710571289, 0.06061670303344727, 0.06061670303344727, 0.06073446273803711, 0.06052249526977539, 0.06070476913452148, 0.06071807861328125, 0.06060851287841797, 0.06081433486938476, 0.06072115325927734, 0.06060134506225586, 0.06065971374511719, 0.13372006225585936, 0.06056345748901367, 0.06059929656982422, 0.060644351959228515, 0.06058291244506836, 0.06054195022583008, 0.06054297637939453, 0.06056140899658203, 0.06055731201171875, 0.06055731201171875, 0.06055936050415039, 0.06073241424560547, 0.060556289672851565, 0.06056755065917969, 0.06058700942993164, 0.06049792098999023, 0.06058393478393555, 0.06057984161376953, 0.06057984161376953, 0.06055219268798828, 0.06055219268798828, 0.06055833435058594, 0.060554241180419924, 0.060556289672851565, 0.060524543762207034, 0.06058700942993164, 0.060572673797607425, 0.06054297637939453, 0.060660736083984375, 0.06058291244506836, 0.06063820648193359, 0.06058291244506836, 0.06065868759155273, 0.06058905410766602, 0.06061363220214844, 0.06056959915161133, 0.06060851287841797, 0.060611583709716796, 0.06063820648193359, 0.060581886291503906, 0.06066790390014649, 0.060590080261230465, 0.06060543823242188, 0.0605849609375, 0.06063820648193359, 0.06059724807739258, 0.06060748672485351, 0.06061568069458008, 0.060606464385986325, 0.06057574462890625, 0.06063820648193359, 0.06063820648193359, 0.06061056137084961, 0.06062080001831055, 0.06065151977539063, 0.06050611114501953, 0.06061670303344727, 0.06059724807739258, 0.06063411331176758, 0.06061056137084961, 0.06062387084960937, 0.06065151977539063, 0.06062694549560547, 0.1336197052001953, 0.06059212875366211, 0.06061260986328125, 0.06053887939453125, 0.06051327896118164, 0.06058905410766602, 0.06058086395263672, 0.060821502685546876, 0.06053887939453125, 0.06054912185668945, 0.060526592254638675, 0.06054502487182617, 0.060572673797607425, 0.060606464385986325, 0.06061260986328125, 0.06055014419555664, 0.0605665283203125, 0.0607191047668457, 0.0607825927734375, 0.060576766967773435, 0.060581886291503906, 0.06057369613647461, 0.06056959915161133, 0.060537857055664064, 0.06054297637939453, 0.060614654541015625, 0.060593151092529295, 0.06062694549560547, 0.06073446273803711, 0.06060851287841797, 0.06065356826782226, 0.06064025497436523, 0.06066483306884766, 0.06061056137084961, 0.06058905410766602, 0.06059929656982422, 0.060627967834472656, 0.060805118560791016, 0.06068838500976562, 0.0606033935546875, 0.06065459060668945, 0.060609535217285154, 0.060649471282958986, 0.06057369613647461, 0.060609535217285154, 0.0607825927734375, 0.06077030563354492, 0.06063411331176758, 0.06058291244506836, 0.060590080261230465, 0.06062899017333984, 0.060633087158203126, 0.06063718414306641, 0.06060543823242188, 0.06073344039916992, 0.06053376007080078, 0.06064640045166016, 0.060606464385986325, 0.06062694549560547, 0.0605849609375, 0.06060851287841797, 0.060642303466796874, 0.060631038665771485, 0.13370060729980468, 0.06058803176879883, 0.06058905410766602, 0.060625919342041014, 0.060572673797607425, 0.06060134506225586, 0.060593151092529295, 0.06061363220214844, 0.06056857681274414, 0.06059212875366211, 0.06057984161376953, 0.06055321502685547, 0.0605849609375, 0.06060543823242188, 0.060572673797607425, 0.06061568069458008, 0.06055833435058594, 0.06058291244506836, 0.06060748672485351, 0.06059724807739258, 0.0606033935546875, 0.06055731201171875, 0.060611583709716796, 0.06058905410766602, 0.060537857055664064, 0.06057984161376953, 0.06060031890869141, 0.06059417724609375, 0.06063820648193359, 0.06058598327636719, 0.06059622573852539, 0.06056345748901367, 0.06068121719360352, 0.060641281127929686, 0.060593151092529295, 0.06062694549560547, 0.06065459060668945, 0.060606464385986325, 0.06061056137084961, 0.06062182235717774, 0.060677120208740234, 0.06061772918701172, 0.06061772918701172, 0.06057369613647461, 0.06061260986328125, 0.060611583709716796, 0.06065868759155273, 0.06058598327636719, 0.06074367904663086, 0.06069145584106445, 0.060614654541015625, 0.06061670303344727, 0.06061056137084961, 0.06064025497436523, 0.060668926239013675, 0.06055116653442383, 0.06069657516479492, 0.06063616180419922, 0.06061670303344727, 0.060606464385986325, 0.060649471282958986, 0.06063820648193359, 0.06067814254760742, 0.1336821746826172, 0.06062182235717774, 0.06055321502685547, 0.0606484489440918, 0.06058291244506836, 0.060598270416259765, 0.060590080261230465, 0.06055936050415039, 0.06054604721069336, 0.06061568069458008, 0.06059622573852539, 0.06057164764404297, 0.06057164764404297, 0.060581886291503906, 0.06061875152587891, 0.06058086395263672, 0.06057881546020508, 0.06066790390014649, 0.0607303695678711, 0.06062899017333984, 0.06068940734863281, 0.06056857681274414, 0.06057779312133789, 0.060598270416259765, 0.06055526351928711, 0.0607006721496582, 0.06061670303344727, 0.06058803176879883, 0.06073651123046875, 0.06060543823242188, 0.06059724807739258, 0.060723201751708984, 0.06062182235717774, 0.060598270416259765, 0.06078976058959961, 0.06057779312133789, 0.06063820648193359, 0.0607303695678711, 0.06069760131835938, 0.06062387084960937, 0.06078976058959961, 0.06061260986328125, 0.060715007781982425, 0.0607191047668457, 0.06060748672485351, 0.06065356826782226, 0.06072012710571289, 0.06056345748901367, 0.06062899017333984, 0.0607303695678711, 0.0606668815612793, 0.06066995239257812, 0.060747776031494144, 0.060622848510742185, 0.06086963272094727, 0.06071603012084961, 0.06065459060668945, 0.06058803176879883, 0.06078566360473633, 0.060780544281005856, 0.060609535217285154, 0.06074367904663086, 0.060682239532470705, 0.13460275268554686, 0.060777473449707034, 0.06060748672485351, 0.06059212875366211, 0.0605665283203125, 0.06059212875366211, 0.06088499069213867, 0.06057369613647461, 0.060593151092529295, 0.060865535736083984, 0.060672000885009764, 0.06060748672485351, 0.06066790390014649, 0.0605747184753418, 0.060598270416259765, 0.06067097473144531, 0.060554241180419924, 0.06058700942993164, 0.060897281646728516, 0.06061875152587891, 0.060625919342041014, 0.06071603012084961, 0.06054912185668945, 0.061172737121582034, 0.060655616760253904, 0.060606464385986325, 0.06061568069458008, 0.06068326568603516, 0.060595199584960936, 0.06058086395263672, 0.0607088623046875, 0.0605747184753418, 0.060647422790527344, 0.0607723503112793, 0.060680191040039064, 0.06061363220214844, 0.06075699234008789, 0.06065663909912109, 0.06072012710571289, 0.060821502685546876, 0.06069145584106445, 0.06064025497436523, 0.06078976058959961, 0.060649471282958986, 0.0606453742980957, 0.06069760131835938, 0.06060543823242188, 0.0606033935546875, 0.060731391906738284, 0.060761089324951174, 0.06069247817993164, 0.060706817626953125, 0.06063206481933594, 0.060622848510742185, 0.06078668975830078, 0.06051532745361328, 0.060649471282958986, 0.0607088623046875, 0.0605849609375, 0.06058393478393555, 0.060750846862792966, 0.060598270416259765, 0.06064025497436523, 0.1340211181640625, 0.06060441589355469, 0.06063206481933594, 0.0605665283203125, 0.06056140899658203, 0.06060031890869141, 0.06067097473144531, 0.06052864074707031, 0.06059417724609375, 0.06070169448852539, 0.06054297637939453, 0.06055526351928711, 0.0607303695678711, 0.0605665283203125, 0.06057881546020508, 0.060715007781982425, 0.06057062530517578, 0.06058905410766602, 0.060709888458251954, 0.06054604721069336, 0.06056447982788086, 0.06082048034667969, 0.06061056137084961, 0.06072012710571289, 0.06068326568603516, 0.06057164764404297, 0.06063616180419922, 0.06069145584106445, 0.0606300163269043, 0.06065663909912109, 0.060805118560791016, 0.06057881546020508, 0.060631038665771485, 0.060690433502197265, 0.06058905410766602, 0.060611583709716796, 0.060727294921875, 0.06058598327636719, 0.060622848510742185, 0.06071603012084961, 0.06065049743652344, 0.060642303466796874, 0.06084198379516602, 0.06067609786987305, 0.06059622573852539, 0.06077951812744141, 0.060639232635498044, 0.06058291244506836, 0.06075699234008789, 0.06057164764404297, 0.0606484489440918, 0.06079590225219727, 0.06061056137084961, 0.060631038665771485, 0.0608092155456543, 0.060526592254638675, 0.06073651123046875, 0.06082355117797852, 0.060633087158203126, 0.06062080001831055, 0.060767230987548826, 0.06055219268798828, 0.06063206481933594, 0.1341685791015625, 0.060556289672851565, 0.06056345748901367, 0.06055833435058594, 0.06054707336425781, 0.060593151092529295, 0.06062182235717774, 0.060539905548095706, 0.06055219268798828, 0.06054092788696289, 0.06060134506225586, 0.06057574462890625, 0.060593151092529295, 0.06060031890869141, 0.06060134506225586, 0.0605296630859375, 0.06058700942993164, 0.06057164764404297, 0.06063616180419922, 0.06056857681274414, 0.06062080001831055, 0.06058086395263672, 0.06061260986328125, 0.06055833435058594, 0.06055731201171875, 0.06060748672485351, 0.06112870407104492, 0.060614654541015625, 0.06075904083251953, 0.06056447982788086, 0.060665855407714846, 0.06062387084960937, 0.06066790390014649, 0.06061977767944336, 0.060639232635498044, 0.06057881546020508, 0.060633087158203126, 0.06067302322387695, 0.06075596618652344, 0.06059622573852539, 0.06068735885620117, 0.06061977767944336, 0.06073651123046875, 0.060614654541015625, 0.060633087158203126, 0.06059212875366211, 0.06072422409057617, 0.06060031890869141, 0.06061875152587891, 0.06058905410766602, 0.0606484489440918, 0.06058803176879883, 0.06061772918701172, 0.06061670303344727, 0.06084096145629883, 0.060682239532470705, 0.06066790390014649, 0.06055833435058594, 0.06060441589355469, 0.060657665252685546, 0.06071705627441406, 0.06069145584106445, 0.060665855407714846, 0.13447782897949218, 0.06059622573852539, 0.06055731201171875, 0.06055936050415039, 0.0605296630859375, 0.060560382843017575, 0.06056243133544922, 0.060537857055664064, 0.06055526351928711, 0.06057574462890625, 0.06055116653442383, 0.06061670303344727, 0.060581886291503906, 0.06056243133544922, 0.060593151092529295, 0.06055321502685547, 0.060593151092529295, 0.06059417724609375, 0.06057574462890625, 0.06056140899658203, 0.06055731201171875, 0.060560382843017575, 0.06057574462890625, 0.06057369613647461, 0.06055014419555664, 0.060593151092529295, 0.06060134506225586, 0.060560382843017575, 0.060625919342041014, 0.060581886291503906, 0.06063820648193359, 0.06060748672485351, 0.06061568069458008, 0.06059622573852539, 0.060590080261230465, 0.06055014419555664, 0.0606033935546875, 0.060624897003173826, 0.060633087158203126, 0.060611583709716796, 0.060706817626953125, 0.06060748672485351, 0.06061772918701172, 0.060598270416259765, 0.060639232635498044, 0.06060748672485351, 0.060593151092529295, 0.06059110260009765, 0.06060134506225586, 0.06065151977539063, 0.060631038665771485, 0.06060031890869141, 0.06061875152587891, 0.06061977767944336, 0.06063718414306641, 0.06055731201171875, 0.060641281127929686, 0.060609535217285154, 0.060657665252685546, 0.06061260986328125, 0.06058291244506836, 0.06075904083251953, 0.06065049743652344, 0.13431602478027344, 0.0606300163269043, 0.06057164764404297, 0.06059110260009765, 0.06057881546020508, 0.06055833435058594, 0.060598270416259765, 0.06052249526977539, 0.06056140899658203, 0.060581886291503906, 0.060590080261230465, 0.06055116653442383, 0.06054092788696289, 0.060680191040039064, 0.060581886291503906, 0.06056345748901367, 0.0606033935546875, 0.060633087158203126, 0.060631038665771485, 0.06058803176879883, 0.06057574462890625, 0.06054195022583008, 0.06056243133544922, 0.060539905548095706, 0.06053683090209961, 0.06066175842285156, 0.06061670303344727, 0.060625919342041014, 0.06065868759155273, 0.06060134506225586, 0.06064332962036133, 0.06061056137084961, 0.060655616760253904, 0.06057984161376953, 0.06073855972290039, 0.06058905410766602, 0.060609535217285154, 0.060590080261230465, 0.06063206481933594, 0.06060543823242188, 0.06066790390014649, 0.060614654541015625, 0.060639232635498044, 0.060641281127929686, 0.06064640045166016, 0.06058700942993164, 0.06070272064208984, 0.06057062530517578, 0.0606033935546875, 0.06059929656982422, 0.060644351959228515, 0.060641281127929686, 0.060624897003173826, 0.06062387084960937, 0.06069760131835938, 0.06053171157836914, 0.06066483306884766, 0.06056755065917969, 0.06059724807739258, 0.060598270416259765, 0.06061363220214844, 0.0606484489440918, 0.060665855407714846]",tokens/s,16.211807874204894,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8265,7 +8265,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8540,7 +8540,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,3047.75168,9028.763648,0.0,8382.316544,7939.414528,s,10,6.222475708007812,0.6222475708007812,0.0004898261264812622,0.6223728332519531,0.6229043151855469,0.6229362579345703,0.622961812133789,"[0.6215844116210938, 0.622897216796875, 0.6224472045898437, 0.622341796875, 0.6226093139648438, 0.6215834350585937, 0.6229682006835937, 0.6217582397460938, 0.6224038696289063, 0.6218820190429688]",tokens/s,411.41181101044583,kWh,7.33948267168469e-06,4.021746173983795e-06,3.6795045775871255e-05,4.815627462153974e-05,tokens/kWh,5316025.834886617,MB,3048.873984,9028.763648,0.0,8382.316544,8233.863168,s,10,338.50034765624997,33.850034765625004,0.0009082101814251832,33.849732421875004,33.851029296875005,33.851565429687504,33.851994335937505,"[33.8521015625, 33.848640625, 33.85091015625, 33.8501171875, 33.84971484375, 33.84975, 33.8495, 33.84959765625, 33.8505546875, 33.8494609375]",tokens/s,1.8611502303086858,kWh,0.0003996493489709166,0.00021904255237639625,0.002000050750366157,0.00261874265171347,tokens/kWh,24057.34674187189,,s,629,343.5634073486327,0.5462057350534704,0.0739750793316962,0.5372897338867187,0.5376120849609375,0.5377253540039063,1.1596667919921875,"[0.5369999389648438, 0.5371217651367187, 0.53671728515625, 0.537164794921875, 0.5368248291015625, 0.537206787109375, 0.53724365234375, 0.5371791381835938, 0.5368934326171875, 0.5374484252929688, 0.5368237915039062, 0.5371473999023437, 0.5371371459960937, 0.5370214233398437, 0.5369026489257812, 0.5371494140625, 0.5369077758789063, 0.53736962890625, 0.5368944702148437, 0.5374730224609375, 0.5372733154296875, 0.53732763671875, 0.5372559204101562, 0.5374003295898437, 0.53711669921875, 0.53743408203125, 0.5371146240234375, 0.5374095458984375, 0.53713818359375, 0.5375703125, 0.5371351318359375, 0.5376665649414063, 0.537343994140625, 0.537660400390625, 0.537238525390625, 0.5394493408203125, 0.5372672119140625, 0.5375682373046875, 0.5374033813476562, 0.5375723266601562, 0.5372518310546875, 0.5374627685546876, 0.5372958984375, 0.5376112670898437, 0.53732763671875, 0.5375897827148437, 0.5374566650390625, 0.537591796875, 0.5373368530273438, 0.537617431640625, 0.5372743530273437, 0.5374218139648438, 0.5373358154296874, 0.5373521728515624, 0.5374238891601563, 0.5377003784179688, 0.5373890380859375, 0.5374044189453125, 0.5368565673828125, 0.5374822387695313, 0.5371740112304687, 0.5375590209960938, 1.160731689453125, 0.5372897338867187, 0.5373317260742188, 0.536974365234375, 0.5370726318359375, 0.537296875, 0.5373296508789063, 0.5368442993164062, 0.5371084594726563, 0.5366640625, 0.5370921020507813, 0.5370153198242188, 0.5370511474609375, 0.5370552368164062, 0.5369251708984375, 0.5369098510742187, 0.537049072265625, 0.5368402099609375, 0.537354248046875, 0.5370050659179687, 0.5376522216796875, 0.5371463623046875, 0.5372323608398437, 0.5373798217773438, 0.5375703125, 0.5374054565429688, 0.5375672607421875, 0.5371392211914062, 0.5372927856445312, 0.5372262573242188, 0.5374105834960937, 0.5371617431640625, 0.5374464111328126, 0.5370101928710938, 0.5375160522460938, 0.5369467163085937, 0.5375999755859375, 0.5370654907226563, 0.5373839111328125, 0.537290771484375, 0.5371781005859375, 0.5372406005859375, 0.5375191040039062, 0.5373685913085937, 0.5378119506835938, 0.5372958984375, 0.5376890869140625, 0.5371627807617188, 0.5374525146484375, 0.5372999877929687, 0.5371555786132812, 0.5372078247070312, 0.5375774536132812, 0.5374822387695313, 0.5375621337890625, 0.5373931274414062, 0.5377433471679688, 0.5372989501953125, 0.5375375366210937, 0.5370828857421875, 0.5374310302734375, 0.5372927856445312, 0.5374443359375, 1.1605975341796875, 0.5371514892578125, 0.53732763671875, 0.5370419311523438, 0.5369763793945312, 0.537027587890625, 0.5371105346679688, 0.53686376953125, 0.5371781005859375, 0.5369999389648438, 0.5372098388671875, 0.5370951538085937, 0.5372415771484375, 0.5368361206054687, 0.53697021484375, 0.5368361206054687, 0.5373788452148438, 0.5369077758789063, 0.5374474487304688, 0.5372006225585938, 0.5376256103515625, 0.5371401977539062, 0.5372794799804688, 0.5370992431640625, 0.5374248657226562, 0.5371719970703125, 0.5375221557617188, 0.5370470581054687, 0.5375098876953125, 0.5372866821289063, 0.5376030883789062, 0.5372876586914063, 0.5373480834960938, 0.5370255126953125, 0.5380044555664063, 0.5371494140625, 0.5373460693359375, 0.5372190551757813, 0.5374095458984375, 0.5371156616210937, 0.5373409423828125, 0.53713818359375, 0.537439208984375, 0.5372620849609375, 0.5375928344726563, 0.5374689331054687, 0.5375150146484375, 0.5372815551757812, 0.5375293579101562, 0.5374801635742188, 0.5372764282226562, 0.5374781494140625, 0.53758154296875, 0.5375764770507813, 0.537491455078125, 0.5374320678710938, 0.537807861328125, 0.537385986328125, 0.537585693359375, 0.537101318359375, 0.5378580322265625, 0.5371678466796875, 0.5375958862304687, 1.160279052734375, 0.5369609985351562, 0.537275390625, 0.5370265502929688, 0.5371607055664063, 0.5369251708984375, 0.5370347290039063, 0.5371637573242187, 0.5371658325195312, 0.5367777099609375, 0.53724365234375, 0.536953857421875, 0.5370439453125, 0.5368780517578124, 0.5370480346679688, 0.536875, 0.537122802734375, 0.5370419311523438, 0.5371986083984375, 0.5371566162109375, 0.5373511962890625, 0.5370081176757813, 0.5376532592773438, 0.5372518310546875, 0.537491455078125, 0.5375631103515625, 0.5374771118164062, 0.537164794921875, 0.5374412841796875, 0.5372344360351563, 0.5374207763671875, 0.537302001953125, 0.5374576416015625, 0.5371043701171875, 0.5376644897460937, 0.5371770629882813, 0.537439208984375, 0.5372200927734375, 0.537670654296875, 0.5372406005859375, 0.5374464111328126, 0.5371873168945313, 0.5375457153320312, 0.5373306884765625, 0.5375682373046875, 0.5374586791992187, 0.5378385620117188, 0.5372211303710938, 0.5372948608398438, 0.5371371459960937, 0.537417724609375, 0.5371791381835938, 0.5375631103515625, 0.537343994140625, 0.5376399536132812, 0.53737060546875, 0.5378252563476562, 0.5372784423828125, 0.5375928344726563, 0.5371002807617188, 0.5377566528320312, 0.5371740112304687, 0.5375252685546875, 1.1604622802734375, 0.53707470703125, 0.5371064453125, 0.5369815063476563, 0.5370859375, 0.537218017578125, 0.5371986083984375, 0.5368463134765625, 0.5371791381835938, 0.5366578979492187, 0.5372927856445312, 0.5372723388671875, 0.5371627807617188, 0.5369517822265625, 0.537027587890625, 0.5370296020507812, 0.53722314453125, 0.53665380859375, 0.5373429565429687, 0.53697021484375, 0.5379092407226562, 0.5369476928710938, 0.5372815551757812, 0.5371320190429687, 0.5373726806640625, 0.5372579956054687, 0.5376419677734375, 0.5369989013671875, 0.537333740234375, 0.5371514892578125, 0.5376256103515625, 0.5371617431640625, 0.5374033813476562, 0.5371043701171875, 0.53758154296875, 0.5370869750976562, 0.53758154296875, 0.5371371459960937, 0.5374607543945312, 0.5373399047851563, 0.53734912109375, 0.5372211303710938, 0.5374893798828125, 0.5373450317382813, 0.5374873657226562, 0.537511962890625, 0.537697265625, 0.5372620849609375, 0.5374576416015625, 0.5373378295898438, 0.5374310302734375, 0.5373788452148438, 0.5375825805664063, 0.537449462890625, 0.5375846557617188, 0.5375979614257812, 0.5378068237304687, 0.5373235473632813, 0.5375047607421874, 0.5369569091796875, 0.5375068359375, 0.5371873168945313, 0.5376583862304688, 1.1593800048828125, 0.5369968872070312, 0.5371678466796875, 0.5368790893554688, 0.5370142822265624, 0.5371074829101562, 0.536890380859375, 0.5370245361328125, 0.5371945190429688, 0.5371904296875, 0.5371555786132812, 0.5369866333007812, 0.5370675048828125, 0.5370357666015625, 0.53704296875, 0.5369476928710938, 0.5370439453125, 0.5368975219726563, 0.5371740112304687, 0.5370562744140625, 0.5374412841796875, 0.5371197509765625, 0.5372723388671875, 0.5372692260742188, 0.53777099609375, 0.5373419799804687, 0.537575439453125, 0.5372262573242188, 0.537449462890625, 0.537206787109375, 0.5375191040039062, 0.537218017578125, 0.5374361572265625, 0.5370224609375, 0.5376153564453126, 0.5371289672851562, 0.5373737182617188, 0.537380859375, 0.53743408203125, 0.5372569580078125, 0.537427978515625, 0.5372313842773437, 0.5373604125976562, 0.5372620849609375, 0.537870361328125, 0.5373726806640625, 0.5377362060546875, 0.5370101928710938, 0.5374207763671875, 0.5372886962890625, 0.5375262451171875, 0.5372723388671875, 0.5374781494140625, 0.5374996337890625, 0.5376716918945312, 0.53737060546875, 0.5378887939453125, 0.5372467041015625, 0.5373880615234375, 0.5372190551757813, 0.5374843139648438, 0.5374290161132812, 0.5376194458007812, 1.1607940673828125, 0.5370654907226563, 0.5372876586914063, 0.537017333984375, 0.53704296875, 0.5369180297851562, 0.5370848999023438, 0.5368524780273437, 0.5371135864257812, 0.5366937866210938, 0.5370224609375, 0.5369159545898438, 0.5372426147460938, 0.5372047119140625, 0.5370357666015625, 0.53686376953125, 0.5372938232421876, 0.5368125610351563, 0.5374832763671875, 0.5372518310546875, 0.5377505493164062, 0.5370930786132813, 0.5374719848632813, 0.5372303466796875, 0.5375682373046875, 0.5374269409179687, 0.5374423217773437, 0.5369999389648438, 0.5372743530273437, 0.53722216796875, 0.5375057983398438, 0.5372692260742188, 0.5375211791992187, 0.5371504516601563, 0.5373726806640625, 0.5371248779296875, 0.5375283203125, 0.5371975708007812, 0.5374423217773437, 0.5372303466796875, 0.5373818969726563, 0.5371207885742187, 0.5374259033203125, 0.5370664672851563, 0.5375396118164063, 0.5372713012695313, 0.5377269897460938, 0.5370449829101562, 0.5375191040039062, 0.5372846069335937, 0.5377986450195312, 0.5370890502929687, 0.5374740600585938, 0.537313232421875, 0.5375160522460938, 0.5373552856445313, 0.53772900390625, 0.53762353515625, 0.5375580444335938, 0.537280517578125, 0.5377146606445312, 0.5373737182617188, 0.5375047607421874, 1.15934716796875, 0.5369712524414062, 0.537196533203125, 0.536880126953125, 0.5370664672851563, 0.5370029907226562, 0.537049072265625, 0.5369415893554688, 0.5371279296875, 0.536838134765625, 0.5372139282226562, 0.5371422729492188, 0.5371709594726563, 0.5370296020507812, 0.5370101928710938, 0.5370931396484375, 0.536943603515625, 0.5370654907226563, 0.5372897338867187, 0.5371084594726563, 0.5374290161132812, 0.5371207885742187, 0.5373665161132812, 0.5372057495117187, 0.5374658813476563, 0.5373778076171875, 0.5378252563476562, 0.5370029907226562, 0.5374115600585937, 0.5372979125976562, 0.5375590209960938, 0.5370470581054687, 0.5373511962890625, 0.5372149658203125, 0.5375795288085937, 0.5372313842773437, 0.537481201171875, 0.5371986083984375, 0.5374371948242187, 0.5373419799804687, 0.5374535522460937, 0.5371392211914062, 0.53747509765625, 0.5372406005859375, 0.5374719848632813, 0.5375580444335938, 0.537575439453125, 0.5373143310546875, 0.537759765625, 0.5374566650390625, 0.5375201416015625, 0.53722216796875, 0.537439208984375, 0.5372764282226562, 0.537491455078125, 0.5373030395507813, 0.5375774536132812, 0.5373572998046875, 0.5374269409179687, 0.5370941162109375, 0.537722900390625, 0.537449462890625, 0.537548828125, 1.1603404541015625, 0.5369415893554688, 0.5372897338867187, 0.5370214233398437, 0.537112548828125, 0.5370777587890625, 0.5371781005859375, 0.5371637573242187, 0.5372262573242188, 0.5365135498046875, 0.5372323608398437, 0.5368197021484375, 0.5370265502929688, 0.5368811645507813, 0.5369989013671875, 0.53697021484375, 0.5371770629882813, 0.5369098510742187, 0.537143310546875, 0.5370941162109375, 0.53747607421875, 0.5370531616210937, 0.5374371948242187, 0.5373184204101562, 0.5373921508789062, 0.5372119140625, 0.5374556274414063, 0.5370972290039062, 0.5374115600585937, 0.537291748046875, 0.537585693359375, 0.5373604125976562, 0.5374586791992187, 0.5370582885742188, 0.537618408203125, 0.5372846069335937, 0.5375836791992188, 0.5370541381835937, 0.5375037231445312, 0.5372927856445312, 0.5374207763671875, 0.5376399536132812, 0.5375559692382812, 0.537228271484375, 0.537797607421875, 0.5371484375, 0.5375344848632813, 0.5373040771484375, 0.5375641479492187, 0.5371904296875, 0.537469970703125, 0.5375662231445313, 0.5374935302734375, 0.537575439453125, 0.5377136840820312, 0.537512939453125, 0.537786376953125, 0.5373778076171875, 0.5373972778320313, 0.537248779296875, 0.5377208251953125, 0.537248779296875, 0.5375057983398438, 1.1597783203125, 0.5370828857421875, 0.5371627807617188, 0.536975341796875, 0.536900634765625, 0.5369651489257813, 0.5374464111328126, 0.5368780517578124, 0.537112548828125, 0.5368463134765625, 0.5373921508789062, 0.5370316772460938, 0.5370029907226562, 0.5370009765625, 0.5370685424804688, 0.5369661254882813, 0.5370408935546875, 0.5369763793945312, 0.5371197509765625, 0.5369815063476563, 0.5374320678710938, 0.5372119140625, 0.5373112182617188, 0.5372886962890625, 0.5373265991210937, 0.5372948608398438, 0.5376635131835937, 0.537280517578125, 0.5374115600585937, 0.537296875, 0.5374033813476562, 0.5372446899414063, 0.53747607421875, 0.5371268920898438, 0.5376399536132812, 0.5370029907226562, 0.537449462890625, 0.5372672119140625, 0.5375283203125, 0.5371187133789063, 0.5373040771484375, 0.5373184204101562, 0.5375262451171875, 0.53732763671875, 0.5376163940429688, 0.537343994140625, 0.5375949096679687, 0.5370869750976562, 0.5373829345703125, 0.5372406005859375, 0.537692138671875, 0.537375732421875, 0.5375538940429687, 0.5374884033203124, 0.5375672607421875, 0.537333740234375, 0.5376993408203125, 0.5373265991210937, 0.5372713012695313, 0.5371146240234375, 0.53762353515625, 0.5374884033203124, 0.537591796875]",tokens/s,1.830811973993839,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8569,7 +8569,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8700,7 +8700,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GP ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1280.995328,1017.643008,0.0,371.195904,315.446272,s,10,0.19319747543334959,0.01931974754333496,0.0010467105269549052,0.01894753646850586,0.019827580642700193,0.021101390457153316,0.022120438308715823,"[0.022375200271606447, 0.01901312065124512, 0.018929792404174806, 0.018751968383789064, 0.019248479843139647, 0.018965280532836915, 0.019544511795043944, 0.018915008544921875, 0.018772640228271485, 0.018681472778320312]",tokens/s,13250.690746645725,kWh,2.25981147787691e-07,1.238256752321744e-07,5.593843438318166e-07,9.091911668516819e-07,tokens/kWh,281568947.58059365,MB,1281.323008,1017.643008,0.0,371.195904,340.902912,s,10,11.583821777343749,1.1583821777343748,0.009627584898594657,1.1564221801757812,1.16910693359375,1.1740381469726562,1.1779831176757813,"[1.1789693603515625, 1.1680111083984375, 1.1652188720703125, 1.14978955078125, 1.1497177734375, 1.1605277099609375, 1.14636669921875, 1.15310693359375, 1.1523763427734375, 1.1597374267578124]",tokens/s,54.38619586086756,kWh,1.370763023194169e-05,7.511351751950323e-06,3.022333472675601e-05,5.144231671064802e-05,tokens/kWh,1224672.6824991468,,s,629,11.732324380874624,0.018652344007749816,0.002332081824858889,0.018243648529052733,0.01894564514160156,0.01906585578918457,0.03724599380493164,"[0.021345279693603517, 0.02056822395324707, 0.018622304916381834, 0.018869312286376953, 0.018895904541015626, 0.01879747200012207, 0.01884569549560547, 0.018906112670898437, 0.01901055908203125, 0.01845964813232422, 0.018405376434326173, 0.018534400939941405, 0.018283519744873047, 0.018355199813842774, 0.018320383071899413, 0.018226175308227538, 0.01817398452758789, 0.017952735900878907, 0.018078720092773438, 0.018127872467041017, 0.018137088775634767, 0.018027519226074217, 0.018348031997680665, 0.01796403121948242, 0.01798041534423828, 0.017889280319213868, 0.01796505546569824, 0.018174976348876954, 0.01842995262145996, 0.01819647979736328, 0.018330623626708984, 0.018799615859985352, 0.018984960556030273, 0.019307519912719725, 0.019362815856933592, 0.018956287384033203, 0.01905971145629883, 0.018949119567871094, 0.018910207748413087, 0.01904128074645996, 0.018909183502197266, 0.018541568756103514, 0.019268608093261717, 0.02164851188659668, 0.019189632415771485, 0.0190699520111084, 0.018994176864624023, 0.019017728805541992, 0.01874028778076172, 0.018768831253051756, 0.01828659248352051, 0.01800294494628906, 0.018103391647338866, 0.018309024810791014, 0.018288639068603514, 0.018249727249145507, 0.018296831130981444, 0.01827020835876465, 0.018159616470336915, 0.018174976348876954, 0.018149375915527344, 0.018314239501953124, 0.03768937683105469, 0.018380767822265626, 0.01881395149230957, 0.018351104736328124, 0.018108415603637695, 0.018318336486816408, 0.018051071166992186, 0.01820979118347168, 0.018351104736328124, 0.018297855377197265, 0.01827436828613281, 0.018240447998046874, 0.018114559173583983, 0.018911359786987304, 0.018926464080810546, 0.018877439498901367, 0.018922496795654296, 0.01882624053955078, 0.018942975997924806, 0.019012672424316406, 0.019114944458007814, 0.018929664611816405, 0.018882560729980468, 0.018909183502197266, 0.018893823623657227, 0.01883545684814453, 0.018930688858032226, 0.019140640258789064, 0.018999263763427733, 0.018945024490356444, 0.01903104019165039, 0.01900032043457031, 0.01904332733154297, 0.018966527938842775, 0.019555328369140625, 0.019763200759887696, 0.019224576950073242, 0.018750463485717773, 0.018890752792358398, 0.018163711547851562, 0.018000896453857423, 0.017994752883911135, 0.01820159912109375, 0.018122751235961913, 0.018229248046875, 0.018174976348876954, 0.018365440368652345, 0.018112512588500978, 0.01814630317687988, 0.018116607666015624, 0.01814630317687988, 0.01823846435546875, 0.01822003173828125, 0.01822822380065918, 0.01809715270996094, 0.01807257652282715, 0.01807360076904297, 0.01816985511779785, 0.018189311981201172, 0.018229248046875, 0.0181790714263916, 0.018077695846557617, 0.018066432952880858, 0.03762995147705078, 0.018239488601684572, 0.018173952102661133, 0.018126911163330078, 0.018182079315185548, 0.01829478454589844, 0.018241535186767577, 0.01802444839477539, 0.018283519744873047, 0.01821696090698242, 0.01819443130493164, 0.018512895584106445, 0.01808076858520508, 0.018274303436279296, 0.01821183967590332, 0.018199552536010744, 0.018395135879516602, 0.018157567977905274, 0.018331743240356444, 0.018150304794311522, 0.01825382423400879, 0.01842790412902832, 0.018110464096069336, 0.018165760040283203, 0.018052095413208007, 0.018263200759887695, 0.018966367721557617, 0.018345983505249023, 0.018093055725097656, 0.018099199295043944, 0.017991680145263672, 0.018292736053466797, 0.018103296279907227, 0.018127872467041017, 0.01846272087097168, 0.018257919311523436, 0.018172927856445312, 0.018128896713256838, 0.018699264526367186, 0.019384319305419923, 0.018778112411499022, 0.019094528198242186, 0.018982912063598634, 0.018769920349121092, 0.018692096710205077, 0.018780160903930664, 0.018891775131225585, 0.018884607315063476, 0.019056640625, 0.018760704040527345, 0.01886412811279297, 0.018914304733276367, 0.018719743728637696, 0.018747392654418944, 0.01879449653625488, 0.018683904647827147, 0.018913280487060546, 0.01885798454284668, 0.019105791091918945, 0.018896896362304686, 0.018958335876464845, 0.018868288040161132, 0.018834367752075195, 0.03720499038696289, 0.017994752883911135, 0.0180633602142334, 0.01810246467590332, 0.018297664642333983, 0.018044927597045898, 0.01820979118347168, 0.01815449523925781, 0.018136064529418947, 0.018220064163208007, 0.01806947135925293, 0.018532352447509767, 0.018326528549194337, 0.01820467185974121, 0.01804902458190918, 0.018113567352294923, 0.018221023559570313, 0.018256895065307616, 0.01828556823730469, 0.018092031478881835, 0.02020966339111328, 0.019123199462890626, 0.01904025650024414, 0.018780160903930664, 0.01865830421447754, 0.018556928634643553, 0.018364416122436524, 0.01815260887145996, 0.01814313507080078, 0.018269119262695314, 0.018164735794067383, 0.018215936660766603, 0.017910911560058595, 0.01787481689453125, 0.01821286392211914, 0.01820057678222656, 0.018172927856445312, 0.018020351409912108, 0.01806540870666504, 0.018712575912475587, 0.018587648391723634, 0.017889280319213868, 0.01821696090698242, 0.01816985511779785, 0.018085887908935547, 0.018095104217529297, 0.01798761558532715, 0.018116575241088867, 0.01822003173828125, 0.0180849609375, 0.017928096771240236, 0.01814227294921875, 0.018178016662597656, 0.01807561683654785, 0.01817087936401367, 0.0182476806640625, 0.018051071166992186, 0.017951744079589844, 0.018018304824829103, 0.01820476722717285, 0.017925024032592773, 0.01824665641784668, 0.018471935272216796, 0.038201343536376955, 0.019075071334838867, 0.019009536743164062, 0.018894975662231445, 0.018753408432006836, 0.018868223190307617, 0.01863065528869629, 0.018965503692626954, 0.018865152359008788, 0.018891775131225585, 0.01866444778442383, 0.018408447265625, 0.018060287475585936, 0.017885183334350584, 0.017976320266723633, 0.017949695587158202, 0.01811155128479004, 0.01834592056274414, 0.018148351669311523, 0.017930240631103517, 0.017978368759155275, 0.0182609920501709, 0.01817190361022949, 0.018254911422729492, 0.018199487686157225, 0.018199583053588868, 0.017978336334228514, 0.01819340705871582, 0.018198528289794923, 0.01797340774536133, 0.0180928955078125, 0.018215936660766603, 0.01821286392211914, 0.018172927856445312, 0.018160640716552736, 0.018123775482177733, 0.01808076858520508, 0.018069503784179687, 0.018173952102661133, 0.01807257652282715, 0.01820979118347168, 0.018010112762451173, 0.01825382423400879, 0.01825279998779297, 0.01800601577758789, 0.018143232345581056, 0.018050048828125, 0.01795686340332031, 0.01807257652282715, 0.018137088775634767, 0.018118656158447266, 0.018051071166992186, 0.017949695587158202, 0.018066495895385743, 0.018194368362426758, 0.01843404769897461, 0.01803775978088379, 0.017875999450683595, 0.01807766342163086, 0.01803059196472168, 0.01820467185974121, 0.018163711547851562, 0.018373632431030275, 0.037416961669921874, 0.018315263748168945, 0.01838489532470703, 0.018367584228515626, 0.018121631622314453, 0.018227231979370116, 0.01852310371398926, 0.018521087646484375, 0.019975168228149414, 0.01922355270385742, 0.018580480575561522, 0.018808832168579103, 0.018928640365600585, 0.01879347229003906, 0.018755712509155274, 0.01847488021850586, 0.018259967803955078, 0.018225151062011717, 0.01827737617492676, 0.018280448913574218, 0.018578432083129884, 0.018981887817382814, 0.01879449653625488, 0.018777088165283205, 0.018954240798950195, 0.018790399551391602, 0.01884262466430664, 0.018937856674194335, 0.018939903259277344, 0.018922496795654296, 0.01882009506225586, 0.018137088775634767, 0.01798963165283203, 0.018159616470336915, 0.018094079971313477, 0.018282495498657226, 0.01803264045715332, 0.017977344512939454, 0.01803980827331543, 0.01808076858520508, 0.018320383071899413, 0.01829478454589844, 0.018257919311523436, 0.018310144424438478, 0.017977344512939454, 0.018115583419799804, 0.01820876884460449, 0.018452640533447265, 0.017978208541870117, 0.018363391876220703, 0.018113536834716795, 0.018053119659423827, 0.018121728897094725, 0.01816985511779785, 0.018192384719848635, 0.018215936660766603, 0.018172927856445312, 0.01824460792541504, 0.018326528549194337, 0.01824460792541504, 0.018246816635131835, 0.018262880325317383, 0.018293760299682618, 0.03905228805541992, 0.018197504043579102, 0.018062335968017578, 0.01804083251953125, 0.01804902458190918, 0.018176000595092775, 0.018295808792114256, 0.017986560821533205, 0.018333696365356447, 0.018181119918823242, 0.01824358367919922, 0.018193536758422852, 0.018217920303344726, 0.018208703994750976, 0.018136064529418947, 0.018336864471435548, 0.018207647323608397, 0.018132991790771484, 0.017987583160400392, 0.01828659248352051, 0.01802342414855957, 0.01806540870666504, 0.01798041534423828, 0.01827123260498047, 0.018158592224121094, 0.01825279998779297, 0.018174976348876954, 0.01807155227661133, 0.018310144424438478, 0.01829478454589844, 0.01805414390563965, 0.01807155227661133, 0.018274303436279296, 0.01821900749206543, 0.01820569610595703, 0.018152448654174806, 0.018167903900146484, 0.018721696853637695, 0.01816268730163574, 0.018124799728393554, 0.017999872207641602, 0.018158592224121094, 0.01822412872314453, 0.01817190361022949, 0.018019327163696287, 0.017936544418334963, 0.017980255126953126, 0.018255903244018556, 0.017989599227905273, 0.018241632461547853, 0.01825987243652344, 0.018539520263671876, 0.018529279708862305, 0.018309247970581054, 0.01804889678955078, 0.018128896713256838, 0.01839411163330078, 0.01815350341796875, 0.01822319984436035, 0.01798643112182617, 0.01789030456542969, 0.01817094421386719, 0.018261951446533205, 0.0376627197265625, 0.01943654441833496, 0.01903308868408203, 0.018584575653076172, 0.018528255462646484, 0.018967552185058592, 0.01887539291381836, 0.01881804847717285, 0.018651136398315428, 0.01829478454589844, 0.01781760025024414, 0.01783296012878418, 0.017933311462402343, 0.018304000854492186, 0.018899967193603515, 0.018166784286499024, 0.018104352951049806, 0.017926111221313476, 0.01844633674621582, 0.01827737617492676, 0.018300928115844727, 0.018282495498657226, 0.018314239501953124, 0.018314239501953124, 0.01830297660827637, 0.018214975357055664, 0.018388927459716795, 0.018646015167236327, 0.01819443130493164, 0.017946624755859376, 0.018058240890502928, 0.017987583160400392, 0.018267135620117187, 0.01808896064758301, 0.01830297660827637, 0.01825382423400879, 0.01807257652282715, 0.018144256591796876, 0.01823846435546875, 0.018239488601684572, 0.01830297660827637, 0.01802239990234375, 0.018226175308227538, 0.01816166305541992, 0.01818435287475586, 0.018342784881591797, 0.018327520370483397, 0.018268320083618166, 0.018170719146728517, 0.018379776000976563, 0.018341888427734376, 0.018197568893432617, 0.018314176559448243, 0.018314239501953124, 0.018317312240600587, 0.018132991790771484, 0.01810758399963379, 0.018072383880615234, 0.018077695846557617, 0.018281471252441405, 0.018387968063354493, 0.018068479537963866, 0.017932287216186525, 0.037222400665283206, 0.018480127334594726, 0.018143232345581056, 0.018041856765747072, 0.018266111373901366, 0.01818726348876953, 0.01836953544616699, 0.018258943557739257, 0.01836953544616699, 0.018282495498657226, 0.018149375915527344, 0.018415615081787108, 0.018137088775634767, 0.01807360076904297, 0.01810534477233887, 0.018094079971313477, 0.018118656158447266, 0.01843097686767578, 0.01861529541015625, 0.01822719955444336, 0.01824051284790039, 0.01816985511779785, 0.01846681594848633, 0.018869247436523438, 0.018137088775634767, 0.018048063278198242, 0.018235328674316407, 0.018192384719848635, 0.01823232078552246, 0.018298879623413086, 0.018140159606933593, 0.01807360076904297, 0.01808793640136719, 0.018168832778930662, 0.018542591094970702, 0.018525184631347655, 0.018155519485473632, 0.01807161521911621, 0.017998783111572266, 0.018185216903686522, 0.017945600509643556, 0.018742271423339844, 0.018907136917114258, 0.01886310386657715, 0.01881088066101074, 0.01864816093444824, 0.01880668830871582, 0.01886412811279297, 0.018655231475830078, 0.017969215393066406, 0.018292671203613282, 0.017846368789672853, 0.018097055435180663, 0.018155519485473632, 0.018289663314819335, 0.01828659248352051, 0.018322431564331054, 0.018144256591796876, 0.018168832778930662, 0.018215936660766603, 0.01809715270996094, 0.018126848220825196, 0.018100223541259765, 0.03725516891479492, 0.018487295150756835, 0.018094079971313477, 0.018165760040283203, 0.018243648529052733, 0.018313152313232422, 0.018050048828125, 0.017924095153808595, 0.01808291244506836, 0.018131872177124024, 0.01863680076599121, 0.0184453125, 0.01882009506225586, 0.019468448638916017, 0.01891004753112793, 0.018972671508789063, 0.018988031387329102, 0.01880678367614746, 0.018761728286743166, 0.018769920349121092, 0.018964479446411133, 0.018885631561279297, 0.01905356788635254, 0.01882828712463379, 0.01894812774658203, 0.018978784561157227, 0.018922496795654296, 0.018893888473510742, 0.018397119522094725, 0.018312192916870116, 0.01841049575805664, 0.018309120178222657, 0.018355199813842774, 0.018215967178344727, 0.018113504409790038, 0.018373632431030275, 0.017963008880615236, 0.01823641586303711, 0.018147327423095702, 0.01879859161376953, 0.018380800247192384, 0.018332672119140626, 0.018139135360717772, 0.018008064270019532, 0.018236448287963867, 0.01829680061340332, 0.01824051284790039, 0.01799577522277832, 0.01802444839477539, 0.01838591957092285, 0.018339839935302735, 0.01810534477233887, 0.01785651206970215, 0.017928192138671875, 0.018020511627197266, 0.01837552070617676, 0.01826304054260254, 0.018059263229370116, 0.01823744010925293, 0.01820159912109375, 0.018020351409912108, 0.018190336227416993, 0.018371583938598633]",tokens/s,53.61256470417404,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-16B-nl,Salesforce/codegen-16B-nl,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1769.34912,11891.376128,0.0,11244.929024,10437.174272,s,10,7.440513244628905,0.7440513244628907,0.0008051333046437837,0.7439403991699218,0.7448412414550781,0.7453177459716798,0.745698949584961,"[0.7457942504882813, 0.7435556030273438, 0.7447269287109375, 0.7431126708984375, 0.7447353515625, 0.7433367919921875, 0.7431660766601562, 0.7437805786132813, 0.7441002197265625, 0.7442047729492187]",tokens/s,344.0622865429331,kWh,8.784825413946122e-06,4.813721496793732e-06,4.2452196660142895e-05,5.605074357088275e-05,tokens/kWh,4567289.9892266,MB,1769.34912,11891.376128,0.0,11244.929024,10730.075136,s,10,411.34482031249996,41.134482031249995,0.004974161508721268,41.133431640625,41.14023203125,41.142840625,41.1449275,"[41.13965234375, 41.14544921875, 41.12939453125, 41.127875, 41.13319140625, 41.135828125, 41.13083984375, 41.1370546875, 41.13186328125, 41.133671875]",tokens/s,1.5315617673789768,kWh,0.0004856886855832168,0.00026619995115832355,0.0023717616791538577,0.003123650315895398,tokens/kWh,20168.710844299796,,s,629,417.38696643066413,0.6635722836735518,0.08835475259072602,0.652906494140625,0.6533408935546875,0.6534932250976563,1.3962570458984376,"[0.6527467651367187, 0.65278466796875, 0.652885009765625, 0.65274365234375, 0.6526473999023438, 0.6523862915039063, 0.652621826171875, 0.6530610961914063, 0.6526361694335937, 0.65248974609375, 0.6527897338867188, 0.6530089111328125, 0.6527088623046875, 0.6529024047851563, 0.6527999877929688, 0.6530201416015625, 0.65272216796875, 0.6532402954101563, 0.6529054565429687, 0.6530672607421875, 0.6527948608398437, 0.6526781616210937, 0.6528972778320312, 0.6534553833007812, 0.6526719970703125, 0.6530232543945312, 0.653286376953125, 0.6527815551757813, 0.6528092041015625, 0.652990478515625, 0.6529351806640625, 0.6529915161132812, 0.653464599609375, 0.6528297119140625, 0.65297509765625, 0.6530723876953125, 0.6528040771484375, 0.6527938842773438, 0.6534625244140625, 0.6530365600585938, 0.6535577392578125, 0.6534000854492188, 0.6530560302734375, 0.6531502685546875, 0.6528839111328125, 0.6528389282226562, 0.6530703125, 0.6533990478515626, 0.6529915161132812, 0.6528317260742188, 0.6531737670898438, 0.6530518798828125, 0.6527283325195312, 0.653496337890625, 0.6533304443359375, 0.6533273315429687, 0.6535270385742188, 0.6535403442382812, 0.653475830078125, 0.6533560791015625, 0.6531522216796875, 0.6534563598632812, 1.3965128173828125, 0.6527662353515625, 0.6527958984375, 0.6524364624023438, 0.6525255737304687, 0.6524774169921875, 0.6529208374023437, 0.652748779296875, 0.65314404296875, 0.6535782470703125, 0.6530027465820313, 0.6528818969726562, 0.6526925048828125, 0.6529013671875, 0.653107177734375, 0.6531942138671875, 0.6530672607421875, 0.6530396118164062, 0.6529607543945313, 0.6527958984375, 0.65289013671875, 0.6534953002929688, 0.6527354736328125, 0.6534369506835938, 0.6532301025390626, 0.6531849975585937, 0.653216796875, 0.6527754516601563, 0.6526781616210937, 0.6528737182617188, 0.6533191528320312, 0.653137939453125, 0.6527938842773438, 0.6530057983398437, 0.653048828125, 0.6530293579101563, 0.6533253173828125, 0.6530191650390625, 0.6530826416015625, 0.6537021484375, 0.6532034301757812, 0.6531942138671875, 0.6530181274414063, 0.6530048217773438, 0.6531112670898438, 0.653338623046875, 0.653370361328125, 0.6532393188476563, 0.6533990478515626, 0.6531123046875, 0.6532874145507812, 0.653106201171875, 0.6534471435546875, 0.6532372436523437, 0.6535567626953125, 0.6534932250976563, 0.6531686401367187, 0.6535045166015625, 0.6532474975585938, 0.6531184692382812, 0.6533038330078125, 0.6533877563476562, 0.6537277221679687, 1.398035400390625, 0.6525358276367188, 0.6527037353515625, 0.6530211791992188, 0.652685302734375, 0.6527416381835938, 0.6529177856445313, 0.6525440063476563, 0.65275390625, 0.6524580078125, 0.6523504638671875, 0.652569580078125, 0.6526484375, 0.6522930908203125, 0.6524733276367187, 0.6524334106445312, 0.65293310546875, 0.6525020141601563, 0.6529607543945313, 0.65254296875, 0.6530191650390625, 0.653095947265625, 0.6526617431640624, 0.6529495239257812, 0.6527191162109375, 0.652727294921875, 0.6527446899414062, 0.6529566650390625, 0.6530468139648438, 0.6528563232421875, 0.6526658325195313, 0.652663818359375, 0.6530242309570312, 0.6528040771484375, 0.6530416870117187, 0.6527467651367187, 0.6534533081054688, 0.6530631713867188, 0.6527078247070313, 0.652959716796875, 0.652927001953125, 0.6528020629882813, 0.6531942138671875, 0.65305908203125, 0.652990478515625, 0.652885009765625, 0.6527457275390625, 0.6529597778320313, 0.6528880615234375, 0.6532208862304687, 0.6528031005859375, 0.6532218627929688, 0.6532781982421875, 0.6527908325195313, 0.6528562622070313, 0.6530416870117187, 0.652859375, 0.6529771728515625, 0.6530160522460937, 0.65293310546875, 0.653043701171875, 0.6527866821289062, 0.6528706665039062, 1.3960069580078125, 0.6525593872070312, 0.6528092041015625, 0.6527579956054688, 0.6525819091796875, 0.6522286376953125, 0.6525307006835938, 0.652564453125, 0.652389404296875, 0.6527446899414062, 0.6526300048828125, 0.6527713012695312, 0.6525562744140625, 0.652705810546875, 0.6525819091796875, 0.6526351318359375, 0.65310205078125, 0.6526033935546875, 0.6533038330078125, 0.6530365600585938, 0.652516357421875, 0.65240576171875, 0.6525009765625, 0.65260546875, 0.6524815063476562, 0.6527232055664063, 0.6527365112304687, 0.6526914672851563, 0.6528818969726562, 0.6526607055664062, 0.6528573608398438, 0.65282763671875, 0.6525972290039063, 0.6530949096679688, 0.6533079223632813, 0.6527703247070312, 0.6528665771484375, 0.6527570190429688, 0.6526085205078125, 0.65305908203125, 0.6531604614257812, 0.6529658813476562, 0.65286962890625, 0.6528440551757813, 0.652632080078125, 0.6529740600585937, 0.6530457763671875, 0.6529075317382812, 0.6527887573242187, 0.6530682983398437, 0.6527017211914062, 0.65301708984375, 0.6527385864257812, 0.6529976196289062, 0.65310205078125, 0.6535454711914063, 0.6528880615234375, 0.6529730834960937, 0.6530068359375, 0.6530252685546875, 0.652854248046875, 0.6531235961914063, 0.6537083129882812, 1.398119384765625, 0.65293310546875, 0.65303857421875, 0.6526300048828125, 0.6527467651367187, 0.6525368041992188, 0.6528123168945312, 0.652611572265625, 0.6528132934570312, 0.6527293701171875, 0.6526064453125, 0.6525624389648438, 0.653022216796875, 0.6527446899414062, 0.6527017211914062, 0.653359130859375, 0.6527334594726563, 0.6524825439453125, 0.6526340942382812, 0.6526945190429687, 0.6525634765625, 0.6526668701171875, 0.6528880615234375, 0.6529412841796876, 0.6527098999023437, 0.6525542602539063, 0.652727294921875, 0.6529341430664063, 0.65318603515625, 0.6526863403320312, 0.6527262573242187, 0.6531788940429688, 0.6528440551757813, 0.6530509033203125, 0.6528675537109375, 0.6529351806640625, 0.6528031005859375, 0.65316552734375, 0.653005859375, 0.6529402465820312, 0.6529024047851563, 0.6528389282226562, 0.6527754516601563, 0.6529351806640625, 0.6533416748046875, 0.6532628784179687, 0.6529669189453124, 0.6530703125, 0.65284814453125, 0.65274365234375, 0.6526975708007813, 0.653169677734375, 0.6534840087890625, 0.652959716796875, 0.6528972778320312, 0.6531133422851563, 0.6529915161132812, 0.6528440551757813, 0.6532638549804688, 0.6533826293945313, 0.653318115234375, 0.6532761840820313, 0.653032470703125, 1.396304931640625, 0.6526607055664062, 0.6528460693359375, 0.6529412841796876, 0.6527395629882813, 0.6526822509765625, 0.652837890625, 0.6526986083984375, 0.6526095581054687, 0.6526044311523438, 0.6528993530273437, 0.6527232055664063, 0.6532402954101563, 0.6530662231445312, 0.6526239013671875, 0.6524539184570313, 0.6527129516601563, 0.6527109375, 0.65282763671875, 0.6532833251953125, 0.653000732421875, 0.6526310424804688, 0.652737548828125, 0.65278564453125, 0.65257470703125, 0.653043701171875, 0.6532413330078125, 0.6531768188476562, 0.6532413330078125, 0.6526648559570313, 0.6527979736328124, 0.6527559814453125, 0.6525931396484375, 0.6530263061523438, 0.6533355712890625, 0.652837890625, 0.6529290161132812, 0.652854248046875, 0.6528942260742188, 0.6532843627929688, 0.6530672607421875, 0.6533038330078125, 0.6530570678710937, 0.6535946044921875, 0.6530897827148437, 0.6533273315429687, 0.6531296997070313, 0.6530109252929688, 0.6527180786132812, 0.6536857299804687, 0.6528952026367187, 0.65280615234375, 0.6531348266601562, 0.6529710083007813, 0.6529607543945313, 0.6530099487304688, 0.653043701171875, 0.6529556274414062, 0.6532618408203125, 0.6531204833984375, 0.6531849975585937, 0.6530713500976563, 0.6529403076171875, 1.3961339111328126, 0.6530078735351562, 0.6526279907226562, 0.6524651489257812, 0.65278564453125, 0.6523176879882813, 0.6522777709960937, 0.652837890625, 0.6526074829101562, 0.65278564453125, 0.6526760864257812, 0.6527836303710938, 0.65282763671875, 0.6525460205078125, 0.6526873779296875, 0.6528081665039063, 0.6531000366210937, 0.6525419311523437, 0.6525849609375, 0.6524835815429687, 0.6529423217773438, 0.6528983154296875, 0.6524999389648437, 0.6530252685546875, 0.6525706176757813, 0.6533919067382813, 0.652591064453125, 0.6528839721679688, 0.6527866821289062, 0.6525234985351562, 0.6525593872070312, 0.6529781494140625, 0.6528173828125, 0.65263818359375, 0.6528880615234375, 0.652843017578125, 0.65278564453125, 0.6529412841796876, 0.6531727294921875, 0.652859375, 0.6534010620117188, 0.65289013671875, 0.6531348266601562, 0.6529403076171875, 0.6528173828125, 0.652906494140625, 0.6529945678710938, 0.6536222534179688, 0.6529812622070312, 0.6531532592773438, 0.6526484375, 0.6530867309570313, 0.6529454345703125, 0.6528583984375, 0.6533949584960937, 0.6535220336914063, 0.6530999145507812, 0.6530130004882813, 0.65289013671875, 0.6530621337890625, 0.6529996948242187, 0.6529505004882813, 0.6532177734375, 1.3966458740234375, 0.6525204467773438, 0.6525778198242187, 0.652927978515625, 0.6524282836914063, 0.6527579956054688, 0.6527825927734375, 0.6535157470703125, 0.65275390625, 0.6528399658203125, 0.6525614013671875, 0.6527999877929688, 0.6529197998046875, 0.652885986328125, 0.6527518920898437, 0.6526412963867188, 0.6525245361328125, 0.6526566162109375, 0.6527129516601563, 0.6526525268554687, 0.6529853515625, 0.6530518798828125, 0.6533673095703125, 0.6529362182617188, 0.65265869140625, 0.6525859985351562, 0.6527385864257812, 0.65299560546875, 0.6531829833984375, 0.6527109375, 0.6531573486328125, 0.6529669189453124, 0.6529403076171875, 0.6528173828125, 0.6531840209960937, 0.65339599609375, 0.6529423217773438, 0.6534061889648437, 0.6537799682617188, 0.6529351806640625, 0.6527119140625, 0.6531481323242188, 0.6529300537109375, 0.6530877685546875, 0.6532301025390626, 0.652896240234375, 0.653011962890625, 0.6526361694335937, 0.6528839721679688, 0.6531204833984375, 0.65299560546875, 0.6533201904296875, 0.6530570068359375, 0.65350146484375, 0.65318603515625, 0.6534410400390624, 0.6530877685546875, 0.6532843627929688, 0.6531849975585937, 0.653676513671875, 0.653001708984375, 0.65317578125, 0.6527201538085937, 1.397074951171875, 0.6525890502929688, 0.6529771728515625, 0.6529843139648438, 0.6529228515625, 0.6524180297851563, 0.6524927978515624, 0.65261669921875, 0.652579833984375, 0.6524293212890625, 0.6530365600585938, 0.6528051147460937, 0.6528235473632813, 0.6525880126953125, 0.65275390625, 0.6524876708984375, 0.652864501953125, 0.6530672607421875, 0.65260546875, 0.6535639038085937, 0.65280615234375, 0.652812255859375, 0.6528092041015625, 0.652943359375, 0.6526351318359375, 0.6530078735351562, 0.652906494140625, 0.6527969360351562, 0.652626953125, 0.6527314453125, 0.6525040283203125, 0.65293310546875, 0.6526105346679687, 0.6528685913085938, 0.6529197998046875, 0.6529515380859375, 0.652821533203125, 0.6530089111328125, 0.6527928466796875, 0.6528409423828125, 0.6529515380859375, 0.653402099609375, 0.6530344848632812, 0.6527191162109375, 0.6530764770507812, 0.65318603515625, 0.6530682983398437, 0.6530089111328125, 0.6533406982421875, 0.6529822998046875, 0.6532177734375, 0.6534010620117188, 0.6528839721679688, 0.6531163940429687, 0.6527764282226562, 0.6531154174804688, 0.6530263061523438, 0.6531287231445313, 0.6528675537109375, 0.6529710083007813, 0.6530610961914063, 0.6528153686523438, 0.652843017578125, 1.397327880859375, 0.6529044189453125, 0.6529126586914062, 0.6528338012695313, 0.6528132934570312, 0.6527037353515625, 0.6528184204101563, 0.6531676025390625, 0.6526679077148437, 0.6525787963867188, 0.6527026977539062, 0.652801025390625, 0.6524569702148437, 0.6524016723632813, 0.6524241943359375, 0.6528737182617188, 0.6528952026367187, 0.6526658325195313, 0.6528194580078125, 0.6528071899414063, 0.6525542602539063, 0.6527293701171875, 0.652859375, 0.6531840209960937, 0.6529013671875, 0.6527825927734375, 0.6534932250976563, 0.6527160034179688, 0.6529791870117188, 0.652732421875, 0.6533007202148438, 0.6534481811523437, 0.6530836181640625, 0.6526351318359375, 0.6527877197265625, 0.6526402587890625, 0.6525419311523437, 0.652747802734375, 0.6533416748046875, 0.6531266479492187, 0.65276416015625, 0.6529249267578126, 0.6530160522460937, 0.6528460693359375, 0.6529403076171875, 0.653253662109375, 0.6531358642578124, 0.6536539916992188, 0.653033447265625, 0.6529884033203125, 0.653033447265625, 0.6527457275390625, 0.6529013671875, 0.6532136840820313, 0.6529832763671874, 0.6531993408203125, 0.6529832763671874, 0.6528511962890625, 0.6531082153320312, 0.6527078857421875, 0.6528624877929687, 0.6531368408203125, 0.653422607421875]",tokens/s,1.506994828753209,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9070,7 +9070,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1856.221184,3368.5504,0.0,2722.103296,2585.357312,s,10,0.7366494750976563,0.07366494750976563,0.00039249671689789645,0.0736723518371582,0.07419290008544921,0.07419320220947266,0.07419344390869141,"[0.0741935043334961, 0.07335641479492187, 0.07308262634277343, 0.07368495941162109, 0.07417286682128907, 0.07325129699707031, 0.07419283294677734, 0.07332640075683594, 0.07372882843017578, 0.07365974426269531]",tokens/s,3475.1942226805027,kWh,8.657666441394562e-07,4.743867474013832e-07,4.119346361168371e-06,5.4594997527092105e-06,tokens/kWh,46890743.03427949,MB,1856.221184,3368.5504,0.0,2722.103296,2628.769792,s,10,38.86898974609376,3.8868989746093745,0.002142874021588634,3.886506591796875,3.8892139404296877,3.890647985839844,3.891795222167969,"[3.89208203125, 3.884625, 3.88758349609375, 3.88604248046875, 3.886436767578125, 3.885375, 3.888895263671875, 3.886576416015625, 3.88693505859375, 3.884438232421875]",tokens/s,16.20829365814206,kWh,4.582713047034936e-05,2.5112934640191582e-05,0.00021524919024462234,0.0002861892553551633,tokens/kWh,220134.050531759,,s,629,39.468826690673836,0.06274853209964042,0.008805392613180215,0.061639678955078124,0.06181806030273437,0.06226268157958985,0.1355356140136719,"[0.0625162239074707, 0.06246809768676758, 0.062304256439208984, 0.06238105773925781, 0.06236467361450195, 0.06246092987060547, 0.062456832885742185, 0.06226432037353516, 0.061895679473876954, 0.06161510467529297, 0.061515777587890626, 0.06170009613037109, 0.06151987075805664, 0.06168473434448242, 0.06160793685913086, 0.061649921417236325, 0.061518848419189455, 0.061574142456054685, 0.062078079223632815, 0.06155251312255859, 0.061625343322753906, 0.061636608123779295, 0.06158028793334961, 0.061625343322753906, 0.06154547119140625, 0.06160076904296875, 0.06159564971923828, 0.06160076904296875, 0.061606910705566405, 0.06164582443237305, 0.061652992248535154, 0.061623294830322264, 0.06156185531616211, 0.06175743865966797, 0.061557758331298826, 0.061669376373291014, 0.06161100769042969, 0.061769729614257814, 0.061658111572265625, 0.06219878387451172, 0.0616703987121582, 0.061666305541992185, 0.061620223999023435, 0.0616734733581543, 0.06161510467529297, 0.06164684677124024, 0.061663230895996096, 0.06168985748291016, 0.06160179138183594, 0.06177382278442383, 0.06160179138183594, 0.06164582443237305, 0.06184550476074219, 0.06161100769042969, 0.061652992248535154, 0.06158848190307617, 0.06170828628540039, 0.06169190216064453, 0.061652992248535154, 0.061575233459472654, 0.061744064331054685, 0.06192236709594726, 0.13562771606445312, 0.06159257507324219, 0.061623294830322264, 0.06195609664916992, 0.06166732788085937, 0.06157209777832031, 0.06155878448486328, 0.06166527938842774, 0.06160793685913086, 0.06156083297729492, 0.061618209838867184, 0.06155465698242187, 0.061639678955078124, 0.06154035186767578, 0.06164787292480469, 0.061722625732421876, 0.06158950424194336, 0.06161305618286133, 0.06162435150146484, 0.06246294403076172, 0.0619683837890625, 0.0616099853515625, 0.06166220855712891, 0.06152601623535156, 0.06164889526367188, 0.06153734588623047, 0.06161401748657227, 0.06158233642578125, 0.06157004928588867, 0.06159769439697266, 0.06160179138183594, 0.06163455963134765, 0.061584384918212894, 0.061551616668701174, 0.061639678955078124, 0.06158848190307617, 0.061601951599121095, 0.06156272125244141, 0.06161612701416016, 0.06156697463989258, 0.06195609664916992, 0.06161510467529297, 0.06171657562255859, 0.061641632080078126, 0.061723648071289064, 0.061636608123779295, 0.06167552185058594, 0.06159564971923828, 0.0615997428894043, 0.061645889282226564, 0.061612991333007815, 0.06159667205810547, 0.06162227249145508, 0.061636608123779295, 0.06173286437988281, 0.061914112091064455, 0.06161305618286133, 0.06163455963134765, 0.06170214462280273, 0.06166425704956055, 0.06158643341064453, 0.06162944030761719, 0.06158540725708008, 0.13527247619628907, 0.061581279754638674, 0.06155878448486328, 0.06152499389648437, 0.06166835021972656, 0.061590526580810545, 0.061517822265625, 0.06159872055053711, 0.061587455749511716, 0.06171955108642578, 0.06164070510864258, 0.061568000793457034, 0.06157721710205078, 0.06150348663330078, 0.06161203384399414, 0.06168169784545898, 0.06163759994506836, 0.061608959197998046, 0.06161407852172852, 0.06154240036010742, 0.06161920166015625, 0.062476287841796874, 0.06166835021972656, 0.06163046264648438, 0.06195814514160156, 0.06177587127685547, 0.0623902702331543, 0.061707263946533204, 0.06162124633789062, 0.0615731201171875, 0.06163251113891602, 0.061565025329589844, 0.061674400329589846, 0.06159564971923828, 0.0616519660949707, 0.06163763046264648, 0.06163455963134765, 0.06162944030761719, 0.0616734733581543, 0.06159667205810547, 0.0616703987121582, 0.061571071624755856, 0.06169702529907226, 0.06159462356567383, 0.06166425704956055, 0.06162636947631836, 0.06164889526367188, 0.06272614288330078, 0.06170521545410156, 0.06171955108642578, 0.06169702529907226, 0.06160281753540039, 0.061656063079833984, 0.06173286437988281, 0.062260223388671876, 0.061813758850097655, 0.061811710357666014, 0.061819969177246095, 0.06171744155883789, 0.061764606475830076, 0.06167244720458984, 0.061655040740966796, 0.061584384918212894, 0.13583973693847656, 0.06158643341064453, 0.06156083297729492, 0.06164070510864258, 0.06155263900756836, 0.06152703857421875, 0.06155059051513672, 0.061587455749511716, 0.06175641632080078, 0.061603839874267576, 0.06168678283691406, 0.061625343322753906, 0.06214553451538086, 0.0615813102722168, 0.06157926559448242, 0.06165913772583008, 0.06169401550292969, 0.06158636856079101, 0.061620223999023435, 0.06160588836669922, 0.06167244720458984, 0.061709312438964846, 0.06159769439697266, 0.061652992248535154, 0.06162124633789062, 0.061636608123779295, 0.061606910705566405, 0.06158950424194336, 0.06163251113891602, 0.061591552734375, 0.06161203384399414, 0.061639678955078124, 0.06177280044555664, 0.06168988800048828, 0.06166422271728516, 0.06160588836669922, 0.06171443176269531, 0.061841407775878904, 0.06170521545410156, 0.061625343322753906, 0.06257766342163086, 0.06173388671875, 0.061682689666748045, 0.061551616668701174, 0.0617154541015625, 0.06164787292480469, 0.06163763046264648, 0.061587455749511716, 0.06166732788085937, 0.06187228775024414, 0.06169481658935547, 0.06167552185058594, 0.06157209777832031, 0.06177996826171875, 0.06175129699707031, 0.06170214462280273, 0.061555744171142575, 0.06177276611328125, 0.06169497680664063, 0.061618175506591794, 0.061625343322753906, 0.06185369491577149, 0.06161920166015625, 0.13637632751464843, 0.062094337463378904, 0.061722625732421876, 0.06158950424194336, 0.06156083297729492, 0.061636608123779295, 0.06153113555908203, 0.061567039489746095, 0.06163654327392578, 0.0615813102722168, 0.0616703987121582, 0.06159360122680664, 0.061620223999023435, 0.061590526580810545, 0.06161305618286133, 0.06163763046264648, 0.06162227249145508, 0.06240870285034179, 0.06161920166015625, 0.06155878448486328, 0.06172671890258789, 0.0615599365234375, 0.06163238525390625, 0.06171136093139649, 0.06158540725708008, 0.06151987075805664, 0.06161407852172852, 0.06159564971923828, 0.0616099853515625, 0.0615997428894043, 0.0616827507019043, 0.061606849670410156, 0.06166527938842774, 0.06172774505615235, 0.06173183822631836, 0.06160793685913086, 0.06163455963134765, 0.061603839874267576, 0.061603839874267576, 0.06159360122680664, 0.06163148880004883, 0.0615731201171875, 0.061639678955078124, 0.06152601623535156, 0.06165302276611328, 0.06166934585571289, 0.06161407852172852, 0.061620223999023435, 0.06162636947631836, 0.06263603210449219, 0.06173593521118164, 0.06169295883178711, 0.061610977172851564, 0.06161305618286133, 0.06162124633789062, 0.06166835021972656, 0.061700126647949216, 0.06224483108520508, 0.06181785583496094, 0.06161612701416016, 0.06189977645874024, 0.061739105224609375, 0.061621150970458984, 0.13536972045898438, 0.06163455963134765, 0.06161731338500977, 0.06172252655029297, 0.06154950332641602, 0.061603839874267576, 0.06166016006469727, 0.061590526580810545, 0.06159769439697266, 0.06160281753540039, 0.06174310302734375, 0.06155878448486328, 0.0615813102722168, 0.06159769439697266, 0.06164582443237305, 0.06159257507324219, 0.061636608123779295, 0.06157209777832031, 0.06174720001220703, 0.06168166351318359, 0.061584384918212894, 0.06155878448486328, 0.06153948974609375, 0.06161494445800781, 0.06175129699707031, 0.06155878448486328, 0.06163251113891602, 0.06156492614746094, 0.061671424865722656, 0.06155878448486328, 0.06164889526367188, 0.0617625617980957, 0.06167859268188477, 0.06308761596679688, 0.061780990600585936, 0.061655040740966796, 0.061669376373291014, 0.06156185531616211, 0.06179635238647461, 0.06155263900756836, 0.061687808990478515, 0.06156595230102539, 0.0617523193359375, 0.061723648071289064, 0.061685760498046874, 0.06160588836669922, 0.0616049919128418, 0.061594497680664065, 0.061568000793457034, 0.061666305541992185, 0.06168371200561523, 0.061661182403564455, 0.061633537292480466, 0.061652992248535154, 0.06165094375610351, 0.06170111846923828, 0.06174515151977539, 0.06168166351318359, 0.06168985748291016, 0.06164684677124024, 0.061617153167724606, 0.06170320129394531, 0.06162019348144531, 0.13634867858886718, 0.061669376373291014, 0.06157619094848633, 0.061677566528320314, 0.06161100769042969, 0.06153526306152344, 0.06168368148803711, 0.06178713607788086, 0.06175539016723633, 0.061590526580810545, 0.061674495697021485, 0.06164377593994141, 0.06158643341064453, 0.06159564971923828, 0.06164377593994141, 0.06164787292480469, 0.061636608123779295, 0.06153420639038086, 0.06174105453491211, 0.06166835021972656, 0.06171443176269531, 0.06157823944091797, 0.061693950653076174, 0.0633589744567871, 0.06182400131225586, 0.06165708923339844, 0.06162124633789062, 0.061742080688476565, 0.06158233642578125, 0.061774848937988285, 0.061881343841552736, 0.061818878173828126, 0.06159360122680664, 0.06161612701416016, 0.06182809448242187, 0.06172876739501953, 0.06173183822631836, 0.061664287567138674, 0.061742080688476565, 0.06162838363647461, 0.061633537292480466, 0.06159462356567383, 0.06164070510864258, 0.061852737426757814, 0.061847488403320314, 0.0616099853515625, 0.06166220855712891, 0.06165708923339844, 0.06167552185058594, 0.061625343322753906, 0.061797374725341796, 0.06176358413696289, 0.06170828628540039, 0.061644798278808595, 0.06157823944091797, 0.061695999145507815, 0.061557758331298826, 0.06290022277832032, 0.06170828628540039, 0.06167552185058594, 0.06157004928588867, 0.061685760498046874, 0.061639678955078124, 0.13560012817382813, 0.06154956817626953, 0.06159257507324219, 0.061664321899414065, 0.06175942230224609, 0.061708446502685546, 0.061510494232177734, 0.061656063079833984, 0.06162636947631836, 0.06158441543579102, 0.061638622283935546, 0.0615546875, 0.06163455963134765, 0.06155168151855469, 0.06161196899414063, 0.06161407852172852, 0.06238719940185547, 0.06159769439697266, 0.06166016006469727, 0.061658111572265625, 0.06163455963134765, 0.06155059051513672, 0.06164684677124024, 0.061532161712646485, 0.061644798278808595, 0.061692928314208986, 0.061917182922363284, 0.061932544708251956, 0.061709312438964846, 0.06181990432739258, 0.06227046585083008, 0.06178303909301758, 0.06163455963134765, 0.06159257507324219, 0.061641727447509766, 0.061695999145507815, 0.061706432342529295, 0.06160879898071289, 0.06169392013549805, 0.061598751068115236, 0.06168060684204101, 0.061641727447509766, 0.06170009613037109, 0.061723648071289064, 0.061636608123779295, 0.061666305541992185, 0.06164582443237305, 0.061813758850097655, 0.06163148880004883, 0.06166220855712891, 0.06170009613037109, 0.06167859268188477, 0.06173491287231445, 0.06168678283691406, 0.061613086700439454, 0.06166217422485352, 0.06156185531616211, 0.061712383270263675, 0.061704193115234375, 0.06167660903930664, 0.06163347244262695, 0.06173388671875, 0.06157926559448242, 0.1358909454345703, 0.061585472106933596, 0.06161913681030273, 0.06158643341064453, 0.06191007995605469, 0.06163654327392578, 0.06151379013061523, 0.06163552093505859, 0.06162739181518555, 0.06161305618286133, 0.061639678955078124, 0.061499393463134766, 0.06164582443237305, 0.06164070510864258, 0.06162944030761719, 0.061644798278808595, 0.06170009613037109, 0.06242816162109375, 0.06164787292480469, 0.06159878540039063, 0.06164985656738281, 0.06165094375610351, 0.06169702529907226, 0.06156595230102539, 0.06164889526367188, 0.06171750259399414, 0.06167244720458984, 0.06157209777832031, 0.06165200042724609, 0.06158233642578125, 0.061629409790039065, 0.06162124633789062, 0.06156288146972656, 0.061585472106933596, 0.06165599822998047, 0.06158540725708008, 0.06170111846923828, 0.06161616134643555, 0.06170723342895508, 0.06167244720458984, 0.06189977645874024, 0.06173183822631836, 0.061659168243408204, 0.06166115188598633, 0.06170431900024414, 0.06161600112915039, 0.061656063079833984, 0.061663230895996096, 0.06173081588745117, 0.06170828628540039, 0.0616734733581543, 0.06315827178955079, 0.06174924850463867, 0.061710334777832034, 0.06160486221313476, 0.06162944030761719, 0.06163558578491211, 0.06182092666625977, 0.06171955108642578, 0.061692928314208986, 0.06157823944091797, 0.061656063079833984, 0.06159769439697266, 0.13598208618164062, 0.06162636947631836, 0.06159462356567383, 0.06156492614746094, 0.06158643341064453, 0.061563934326171875, 0.06156079864501953, 0.06168678283691406, 0.06158339309692383, 0.06161507034301758, 0.06163558578491211, 0.0615618896484375, 0.06162019348144531, 0.061518848419189455, 0.06167244720458984, 0.06163046264648438, 0.06163763046264648, 0.06160588836669922, 0.06163148880004883, 0.06158540725708008, 0.06164889526367188, 0.06164892959594727, 0.061594593048095704, 0.06151168060302734, 0.06163251113891602, 0.06157926559448242, 0.0629678077697754, 0.06161305618286133, 0.06163558578491211, 0.06159360122680664, 0.06161203384399414, 0.061655040740966796, 0.061587455749511716, 0.06164275360107422, 0.06164889526367188, 0.06159872055053711, 0.061656063079833984, 0.06163455963134765, 0.061687808990478515, 0.0616099853515625, 0.061692928314208986, 0.061608959197998046, 0.06170111846923828, 0.06162739181518555, 0.06164889526367188, 0.06160179138183594, 0.06163251113891602, 0.06161407852172852, 0.061652030944824215, 0.061599681854248044, 0.06173183822631836, 0.06154444885253906, 0.06162636947631836, 0.061669376373291014, 0.06157622528076172, 0.06165910339355469, 0.06154550552368164, 0.06167855834960938, 0.06169702529907226, 0.061652992248535154, 0.061699073791503904, 0.06180044937133789, 0.06164179229736328]",tokens/s,15.936627782974552,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1479.114752,1841.823744,0.0,1195.37664,1022.155264,s,10,0.6871012496948242,0.06871012496948242,0.00033873627735752553,0.06860185623168946,0.06884490509033203,0.06926672515869141,0.06960418121337891,"[0.06968854522705079, 0.06850118255615234, 0.06852729797363281, 0.06862016296386719, 0.06869097900390625, 0.06858354949951172, 0.06851074981689453, 0.06875116729736327, 0.06849430084228515, 0.06873331451416016]",tokens/s,3725.797327740305,kWh,8.094942381483598e-07,4.4356150458715745e-07,3.836808624999982e-06,5.089864367735499e-06,tokens/kWh,50296035.710259095,MB,1479.114752,1841.823744,0.0,1195.37664,1091.597312,s,10,35.17510473632813,3.517510473632812,0.0019858794372678286,3.5167731933593753,3.5191181640625,3.52109130859375,3.5226698242187497,"[3.523064453125, 3.5170556640625, 3.51595458984375, 3.516420166015625, 3.516550048828125, 3.5174658203125, 3.51656103515625, 3.516367919921875, 3.5186796875, 3.5169853515625]",tokens/s,17.91039443158641,kWh,4.1521001642561397e-05,2.275550190917111e-05,0.0001961486569188012,0.00026042516047053367,tokens/kWh,241912.10974458922,,s,629,35.73639368057251,0.05681461634431242,0.008170033415171557,0.05581414413452149,0.05589258193969727,0.055985150909423825,0.12453187377929688,"[0.05623807907104492, 0.056237056732177736, 0.056357887268066405, 0.056323070526123044, 0.05624934387207031, 0.05632819366455078, 0.05631078338623047, 0.056376319885253906, 0.05601279830932617, 0.05586739349365234, 0.05577318572998047, 0.05581107330322266, 0.055946239471435545, 0.05584998321533203, 0.05613875198364258, 0.05595852661132812, 0.05590220642089844, 0.05599334335327148, 0.055934974670410156, 0.05595238494873047, 0.05628211212158203, 0.05583871841430664, 0.05575987243652344, 0.055757823944091796, 0.055818241119384764, 0.055790592193603515, 0.0557946891784668, 0.05578649520874023, 0.05578035354614258, 0.05583871841430664, 0.055836673736572265, 0.055803905487060546, 0.055874561309814455, 0.0558919677734375, 0.05581721496582031, 0.05585919952392578, 0.055828479766845705, 0.05583871841430664, 0.055795711517333986, 0.05586636734008789, 0.055826431274414064, 0.055774208068847655, 0.055801856994628904, 0.05579776000976563, 0.055820289611816405, 0.05580083084106445, 0.0557496337890625, 0.05584281539916992, 0.0558766098022461, 0.05587865447998047, 0.05584281539916992, 0.055782398223876956, 0.055809024810791016, 0.055788543701171874, 0.05576499176025391, 0.05582950210571289, 0.055911422729492184, 0.0558551025390625, 0.05586739349365234, 0.05582745742797852, 0.055890945434570315, 0.05582438278198242, 0.12451123046875, 0.05576396942138672, 0.05577523040771484, 0.05576499176025391, 0.05576806259155274, 0.05575475311279297, 0.05576499176025391, 0.0557916145324707, 0.05580799865722656, 0.05574655914306641, 0.05579673767089844, 0.05575987243652344, 0.055812095642089846, 0.055739391326904295, 0.05584384155273438, 0.0557209587097168, 0.05575372695922851, 0.055790592193603515, 0.05578342437744141, 0.05578035354614258, 0.05590323257446289, 0.055731201171875, 0.05590630340576172, 0.055772159576416014, 0.05581721496582031, 0.05579673767089844, 0.05584998321533203, 0.055812095642089846, 0.0557946891784668, 0.05591449737548828, 0.05588684844970703, 0.055831550598144535, 0.05584998321533203, 0.05613363265991211, 0.05586227035522461, 0.055790592193603515, 0.05580799865722656, 0.05578649520874023, 0.05587046432495117, 0.0557946891784668, 0.05585715103149414, 0.0557916145324707, 0.05584691238403321, 0.05583257675170898, 0.05586329650878906, 0.05585919952392578, 0.05588275146484375, 0.055853057861328125, 0.05580083084106445, 0.05595340728759766, 0.05588275146484375, 0.0557844467163086, 0.05586739349365234, 0.05587046432495117, 0.055825408935546876, 0.05580595016479492, 0.05588172912597656, 0.05586841583251953, 0.055831550598144535, 0.055839744567871094, 0.05582233428955078, 0.05582131195068359, 0.05584076690673828, 0.12454399871826172, 0.055744510650634765, 0.055806976318359375, 0.05572710418701172, 0.055826431274414064, 0.05572608184814453, 0.05575987243652344, 0.05573222351074219, 0.055831550598144535, 0.055772159576416014, 0.055736320495605465, 0.05578649520874023, 0.055790592193603515, 0.055723007202148435, 0.05581414413452149, 0.05574348831176758, 0.055772159576416014, 0.05571583938598633, 0.05583462524414062, 0.05577830505371094, 0.055772159576416014, 0.05574655914306641, 0.05588479995727539, 0.055787521362304686, 0.055787521362304686, 0.055798782348632815, 0.05585715103149414, 0.05578956985473633, 0.05578342437744141, 0.055831550598144535, 0.055812095642089846, 0.05575065612792969, 0.05581107330322266, 0.05577318572998047, 0.05585100936889648, 0.055803905487060546, 0.055812095642089846, 0.0557916145324707, 0.05586227035522461, 0.05580287933349609, 0.055769088745117185, 0.055812095642089846, 0.05584998321533203, 0.055766014099121096, 0.05582438278198242, 0.05583769607543945, 0.055818241119384764, 0.05590118408203125, 0.05585715103149414, 0.055828479766845705, 0.05595033645629883, 0.055844863891601565, 0.05581107330322266, 0.05580083084106445, 0.0558653450012207, 0.05584281539916992, 0.05595443344116211, 0.05584998321533203, 0.05580799865722656, 0.055853057861328125, 0.05577318572998047, 0.055803905487060546, 0.055818241119384764, 0.12458598327636719, 0.05576806259155274, 0.05575680160522461, 0.0557916145324707, 0.05617663955688477, 0.05577523040771484, 0.0557496337890625, 0.055771137237548826, 0.0557844467163086, 0.05573427200317383, 0.05581926345825195, 0.055929855346679686, 0.0558551025390625, 0.055806976318359375, 0.055798782348632815, 0.055728126525878906, 0.0557946891784668, 0.05577318572998047, 0.055744510650634765, 0.0557762565612793, 0.055772159576416014, 0.0557496337890625, 0.055841793060302736, 0.055755775451660154, 0.055769088745117185, 0.055782398223876956, 0.05583257675170898, 0.05581619262695312, 0.05586022567749024, 0.05580595016479492, 0.05581926345825195, 0.05580083084106445, 0.05580595016479492, 0.05582950210571289, 0.055815166473388675, 0.055844863891601565, 0.05583462524414062, 0.05582745742797852, 0.05579776000976563, 0.05574553680419922, 0.0557844467163086, 0.055785472869873044, 0.055795711517333986, 0.0557844467163086, 0.05582233428955078, 0.05586227035522461, 0.05587148666381836, 0.05582131195068359, 0.0559370231628418, 0.055825408935546876, 0.05587148666381836, 0.05581414413452149, 0.05578649520874023, 0.055856128692626954, 0.05586739349365234, 0.055809024810791016, 0.055826431274414064, 0.05580799865722656, 0.05578956985473633, 0.055820289611816405, 0.05583359909057617, 0.055810047149658204, 0.05586329650878906, 0.12467711639404297, 0.055820289611816405, 0.055785472869873044, 0.05572403335571289, 0.055782398223876956, 0.05574758529663086, 0.055757823944091796, 0.05576294326782227, 0.05584281539916992, 0.05578035354614258, 0.055801856994628904, 0.05575884628295898, 0.05578342437744141, 0.05572403335571289, 0.055812095642089846, 0.0557496337890625, 0.0562718734741211, 0.05581926345825195, 0.055861248016357425, 0.05574041748046875, 0.055828479766845705, 0.05575167846679688, 0.0558919677734375, 0.055742462158203124, 0.05580083084106445, 0.055785472869873044, 0.05584793472290039, 0.055769088745117185, 0.055836673736572265, 0.05579673767089844, 0.05579264068603516, 0.05577318572998047, 0.05582233428955078, 0.05581926345825195, 0.055810047149658204, 0.0557762565612793, 0.05582438278198242, 0.05582438278198242, 0.05586739349365234, 0.05574860763549805, 0.055815166473388675, 0.05576294326782227, 0.05587558364868164, 0.05578137588500977, 0.055825408935546876, 0.055771137237548826, 0.05580595016479492, 0.0557946891784668, 0.055907329559326174, 0.05586329650878906, 0.05581721496582031, 0.055972862243652347, 0.05586227035522461, 0.05586636734008789, 0.055795711517333986, 0.0557916145324707, 0.05586636734008789, 0.055795711517333986, 0.05584588623046875, 0.05582233428955078, 0.05583052825927735, 0.05580083084106445, 0.05589503860473633, 0.12453990173339843, 0.05574041748046875, 0.05575680160522461, 0.055809024810791016, 0.0557844467163086, 0.05577011108398437, 0.056180736541748044, 0.05583257675170898, 0.05579673767089844, 0.05576704025268555, 0.0558551025390625, 0.05578137588500977, 0.05581107330322266, 0.055779327392578126, 0.055820289611816405, 0.055755775451660154, 0.0557844467163086, 0.05576294326782227, 0.055777278900146485, 0.05581619262695312, 0.05575680160522461, 0.05578956985473633, 0.05586329650878906, 0.055803905487060546, 0.05586022567749024, 0.055777278900146485, 0.055766014099121096, 0.05580492782592773, 0.055858177185058595, 0.055806976318359375, 0.055831550598144535, 0.055782398223876956, 0.055772159576416014, 0.05580083084106445, 0.05586943817138672, 0.055809024810791016, 0.05583769607543945, 0.05579264068603516, 0.0557946891784668, 0.055731201171875, 0.05578137588500977, 0.055809024810791016, 0.05582131195068359, 0.05595238494873047, 0.05582745742797852, 0.055853057861328125, 0.05588787078857422, 0.05581619262695312, 0.055787521362304686, 0.05583257675170898, 0.05588787078857422, 0.05584691238403321, 0.055831550598144535, 0.055856128692626954, 0.05627699279785156, 0.0558766098022461, 0.05584588623046875, 0.055831550598144535, 0.05580799865722656, 0.05583052825927735, 0.055809024810791016, 0.05583564758300781, 0.05584691238403321, 0.12479897308349609, 0.05577830505371094, 0.05582438278198242, 0.05576192092895508, 0.05574655914306641, 0.05575987243652344, 0.0557916145324707, 0.05576192092895508, 0.05581619262695312, 0.05576499176025391, 0.05581107330322266, 0.05576499176025391, 0.055831550598144535, 0.055787521362304686, 0.05582950210571289, 0.05572710418701172, 0.05584076690673828, 0.05575680160522461, 0.055825408935546876, 0.055723007202148435, 0.055841793060302736, 0.05574860763549805, 0.055815166473388675, 0.055744510650634765, 0.055820289611816405, 0.05578342437744141, 0.055858177185058595, 0.055798782348632815, 0.05581107330322266, 0.055782398223876956, 0.05587251281738281, 0.0558551025390625, 0.05586431884765625, 0.05585715103149414, 0.055812095642089846, 0.05580287933349609, 0.05588275146484375, 0.05590323257446289, 0.05586431884765625, 0.05576294326782227, 0.05586022567749024, 0.055760894775390625, 0.05585203170776367, 0.05575884628295898, 0.05586431884765625, 0.055777278900146485, 0.055823360443115234, 0.05583052825927735, 0.055874561309814455, 0.05587148666381836, 0.055790592193603515, 0.05584588623046875, 0.05584588623046875, 0.05587558364868164, 0.05578342437744141, 0.055825408935546876, 0.05590220642089844, 0.05594828796386719, 0.05585919952392578, 0.055787521362304686, 0.05580595016479492, 0.05581107330322266, 0.05586943817138672, 0.12475801849365234, 0.0557496337890625, 0.05581926345825195, 0.05589811325073242, 0.05580287933349609, 0.05578137588500977, 0.055772159576416014, 0.05578137588500977, 0.05575475311279297, 0.05574758529663086, 0.05581414413452149, 0.055793663024902344, 0.05579776000976563, 0.055806976318359375, 0.05585203170776367, 0.05576704025268555, 0.05576499176025391, 0.055812095642089846, 0.05581926345825195, 0.05579776000976563, 0.05581619262695312, 0.0557844467163086, 0.05584588623046875, 0.055787521362304686, 0.05582438278198242, 0.05579980850219726, 0.055790592193603515, 0.055790592193603515, 0.05583052825927735, 0.05582438278198242, 0.05583769607543945, 0.05579980850219726, 0.05581414413452149, 0.05580595016479492, 0.055826431274414064, 0.05581619262695312, 0.05584691238403321, 0.05578035354614258, 0.05590630340576172, 0.055806976318359375, 0.05579264068603516, 0.05581107330322266, 0.055844863891601565, 0.0557946891784668, 0.05581107330322266, 0.055809024810791016, 0.05588275146484375, 0.055815166473388675, 0.055785472869873044, 0.05583462524414062, 0.0558837776184082, 0.05583871841430664, 0.055788543701171874, 0.05585203170776367, 0.055810047149658204, 0.055790592193603515, 0.05580287933349609, 0.055844863891601565, 0.05579673767089844, 0.055831550598144535, 0.055782398223876956, 0.05583052825927735, 0.05590016174316406, 0.12446822357177735, 0.055798782348632815, 0.055782398223876956, 0.05577011108398437, 0.05578035354614258, 0.05575167846679688, 0.05574655914306641, 0.05576396942138672, 0.05580287933349609, 0.05576192092895508, 0.055820289611816405, 0.05574758529663086, 0.05579673767089844, 0.055725055694580077, 0.055809024810791016, 0.055728126525878906, 0.055841793060302736, 0.055744510650634765, 0.05581619262695312, 0.05592166519165039, 0.05584998321533203, 0.05577830505371094, 0.055877632141113284, 0.05575680160522461, 0.05580492782592773, 0.05580492782592773, 0.05587865447998047, 0.055777278900146485, 0.05586431884765625, 0.05670809555053711, 0.05588684844970703, 0.05614080047607422, 0.055812095642089846, 0.05583359909057617, 0.05583769607543945, 0.05582950210571289, 0.05588582229614258, 0.05592268753051758, 0.056118270874023435, 0.05578342437744141, 0.055861248016357425, 0.056174591064453126, 0.05588172912597656, 0.055752704620361325, 0.05583564758300781, 0.05581414413452149, 0.05596979141235352, 0.05585408020019531, 0.05584384155273438, 0.05587558364868164, 0.05581721496582031, 0.05582950210571289, 0.05588479995727539, 0.05593804931640625, 0.05578956985473633, 0.0557844467163086, 0.05588684844970703, 0.05579980850219726, 0.05584691238403321, 0.05581926345825195, 0.0558131217956543, 0.055779327392578126, 0.055839744567871094, 0.12470374298095703, 0.055769088745117185, 0.05576396942138672, 0.055777278900146485, 0.05578342437744141, 0.05576806259155274, 0.05575372695922851, 0.055731201171875, 0.055752704620361325, 0.05572403335571289, 0.05583769607543945, 0.05580083084106445, 0.055795711517333986, 0.05576396942138672, 0.05582233428955078, 0.05577011108398437, 0.0557916145324707, 0.05578035354614258, 0.05583052825927735, 0.05581619262695312, 0.05581414413452149, 0.05578956985473633, 0.055841793060302736, 0.05577830505371094, 0.055798782348632815, 0.055806976318359375, 0.05582745742797852, 0.05582131195068359, 0.05592575836181641, 0.055820289611816405, 0.05586022567749024, 0.055810047149658204, 0.0557844467163086, 0.055815166473388675, 0.0558551025390625, 0.05585203170776367, 0.055874561309814455, 0.055790592193603515, 0.055785472869873044, 0.05573222351074219, 0.05581721496582031, 0.05588479995727539, 0.05584076690673828, 0.05581107330322266, 0.05585203170776367, 0.05581414413452149, 0.0559370231628418, 0.05584384155273438, 0.05582438278198242, 0.05584588623046875, 0.05587148666381836, 0.05585203170776367, 0.055803905487060546, 0.05607219314575195, 0.05584076690673828, 0.055806976318359375, 0.055844863891601565, 0.05585919952392578, 0.05580799865722656, 0.05607731246948242, 0.055861248016357425, 0.05581619262695312, 0.05583564758300781]",tokens/s,17.601104510496405,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9983,7 +9983,7 @@ If this is a private repository, make sure to pass a token having permission to 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1506.42688,1579.679744,0.0,933.23264,856.69632,s,10,0.43564921951293945,0.043564921951293946,0.0031757046407324095,0.04331248092651367,0.04437063980102539,0.0485523281097412,0.051897678756713875,"[0.052734016418457035, 0.04333430480957031, 0.04164771270751953, 0.043380638122558594, 0.0414692497253418, 0.04140153503417969, 0.04329619216918945, 0.04344137573242188, 0.04332876968383789, 0.04161542510986328]",tokens/s,5876.287355368403,kWh,4.969496102560134e-07,2.72305043992154e-07,1.379101243335609e-06,2.1483558975837764e-06,tokens/kWh,119160889.63095888,MB,1506.42688,1579.679744,0.0,933.23264,899.62496,s,10,24.820336181640624,2.4820336181640625,0.03505314468873128,2.4815734863281254,2.524392333984375,2.52884619140625,2.53240927734375,"[2.533300048828125, 2.512392822265625, 2.510242431640625, 2.523402587890625, 2.427984130859375, 2.441885986328125, 2.461017822265625, 2.4780390625, 2.48510791015625, 2.44696337890625]",tokens/s,25.382412042670286,kWh,2.9558689719627776e-05,1.6199240900285592e-05,7.306973609346903e-05,0.00011882766671338238,tokens/kWh,530179.5595461687,,s,629,25.16415895462036,0.04000661201052522,0.005123259677494279,0.03999129486083984,0.040338432312011716,0.04051681213378906,0.07998427032470704,"[0.04113715362548828, 0.041027584075927735, 0.041578495025634765, 0.04084121704101563, 0.04151398468017578, 0.041063423156738284, 0.04110131072998047, 0.0394700813293457, 0.039803905487060545, 0.03899084854125977, 0.03885260772705078, 0.04075110244750976, 0.04097536087036133, 0.04030361557006836, 0.04015513610839844, 0.04008652877807617, 0.0400261116027832, 0.04022272109985352, 0.04007628631591797, 0.040078399658203125, 0.04018272018432617, 0.040068096160888675, 0.040422401428222655, 0.04001484680175781, 0.04015411376953125, 0.04017356872558594, 0.04025958251953125, 0.04034048080444336, 0.04012134552001953, 0.04011212921142578, 0.04016537475585937, 0.040008705139160154, 0.040130561828613284, 0.04007731246948242, 0.04001484680175781, 0.04027699279785156, 0.03999129486083984, 0.04007731246948242, 0.040084480285644535, 0.0402503662109375, 0.04009983825683594, 0.04006399917602539, 0.04001279830932617, 0.040207359313964845, 0.04010092926025391, 0.04016838455200195, 0.04013158416748047, 0.040172542572021484, 0.040498241424560544, 0.04003936004638672, 0.040325119018554685, 0.04010291290283203, 0.040225791931152347, 0.04005376052856445, 0.04014796829223633, 0.039678016662597654, 0.04023392105102539, 0.04024422454833984, 0.03986227035522461, 0.04005376052856445, 0.040235008239746094, 0.04025446319580078, 0.08356147003173828, 0.04013158416748047, 0.040151039123535154, 0.04020537567138672, 0.04016428756713867, 0.03999334335327148, 0.040237056732177735, 0.0403394546508789, 0.04013260650634766, 0.04007219314575195, 0.04012236785888672, 0.04005785751342773, 0.03998822402954102, 0.040136703491210936, 0.040052734375, 0.040161312103271486, 0.03975164794921875, 0.040303680419921876, 0.04009260940551758, 0.04006707382202149, 0.04007628631591797, 0.04026265716552734, 0.04016128158569336, 0.040051712036132815, 0.040156158447265625, 0.04031795120239258, 0.04012748718261719, 0.04016230392456055, 0.03998617553710938, 0.04028518295288086, 0.04020633697509766, 0.04005068969726563, 0.0399288330078125, 0.040256511688232424, 0.039979007720947264, 0.04008550262451172, 0.04009164810180664, 0.04038860702514648, 0.040223743438720705, 0.04016844940185547, 0.04011724853515625, 0.04027699279785156, 0.04012441635131836, 0.04024729537963867, 0.040101886749267575, 0.04006399917602539, 0.04013158416748047, 0.04007526397705078, 0.039964672088623046, 0.04025958251953125, 0.04006707382202149, 0.040025089263916014, 0.04005068969726563, 0.03987968063354492, 0.03824639892578125, 0.038312961578369144, 0.03836006546020508, 0.038322208404541015, 0.03826992034912109, 0.03886899185180664, 0.03874611282348633, 0.0382751350402832, 0.038287296295166015, 0.08017305755615234, 0.03826483154296875, 0.038435840606689455, 0.03826073455810547, 0.03853107070922852, 0.04071014404296875, 0.03881062316894531, 0.03902771377563476, 0.040190975189208986, 0.04033331298828125, 0.04007526397705078, 0.04009471893310547, 0.0400711669921875, 0.04029747009277344, 0.03996364974975586, 0.040139774322509765, 0.04016128158569336, 0.04015206527709961, 0.04011315155029297, 0.040158206939697266, 0.039951358795166016, 0.04007526397705078, 0.04027699279785156, 0.03998720169067383, 0.04024729537963867, 0.03996672058105469, 0.0404398078918457, 0.040035327911376956, 0.040134654998779294, 0.04028108978271484, 0.03998822402954102, 0.039959552764892575, 0.040130561828613284, 0.04015411376953125, 0.04040499114990234, 0.04033433532714844, 0.04011929702758789, 0.040313854217529296, 0.04006614303588867, 0.039992225646972655, 0.04012543869018555, 0.04003942489624023, 0.04003635025024414, 0.04030976104736328, 0.04017561721801758, 0.03996364974975586, 0.04004048156738281, 0.04014076614379883, 0.040002559661865236, 0.04034969711303711, 0.04014284896850586, 0.04008345413208008, 0.04016844940185547, 0.03999948883056641, 0.039951358795166016, 0.03837644958496094, 0.03825766372680664, 0.03821158218383789, 0.038324222564697266, 0.039932926177978514, 0.040046592712402344, 0.04010291290283203, 0.04031283187866211, 0.08356454467773437, 0.040120319366455076, 0.040197120666503904, 0.04001484680175781, 0.04011520004272461, 0.0399738883972168, 0.04001587295532227, 0.040022014617919925, 0.040030208587646485, 0.03988684844970703, 0.040529918670654294, 0.04023910522460938, 0.039923713684082034, 0.0401797103881836, 0.04036505508422852, 0.040136703491210936, 0.039977985382080077, 0.04037222290039062, 0.040443904876708986, 0.04031999969482422, 0.04006092834472656, 0.04047359848022461, 0.040342529296875, 0.04001279830932617, 0.040022014617919925, 0.040134654998779294, 0.04007219314575195, 0.04011212921142578, 0.040000511169433595, 0.04015209579467773, 0.04024316787719726, 0.0400711669921875, 0.04026265716552734, 0.04001792144775391, 0.04007526397705078, 0.0404398078918457, 0.04011724853515625, 0.04022784042358398, 0.040081409454345705, 0.04112998580932617, 0.04065484619140625, 0.04024524688720703, 0.04050841522216797, 0.04051763153076172, 0.04026163101196289, 0.040182785034179686, 0.04026572799682617, 0.04014080047607422, 0.040338432312011716, 0.04013363265991211, 0.04012236785888672, 0.039981056213378906, 0.04031488037109375, 0.04002304077148437, 0.04011520004272461, 0.040033279418945314, 0.04029849624633789, 0.040041473388671874, 0.03832729721069336, 0.038545406341552735, 0.03859763336181641, 0.03841331100463867, 0.03827199935913086, 0.07982694244384765, 0.03820544052124023, 0.0380579833984375, 0.03835903930664063, 0.038199295043945314, 0.03845939254760742, 0.04019404983520508, 0.04046745681762695, 0.03990630340576172, 0.04013363265991211, 0.03996364974975586, 0.03836006546020508, 0.03817062377929688, 0.038171646118164065, 0.0381921272277832, 0.03843379211425781, 0.03821670532226563, 0.03819007873535156, 0.03816755294799805, 0.038141952514648435, 0.038299648284912106, 0.038242305755615234, 0.038491134643554685, 0.038255615234375, 0.03834982299804687, 0.038201343536376955, 0.03825151824951172, 0.038152191162109376, 0.03830886459350586, 0.03816755294799805, 0.038171646118164065, 0.03821263885498047, 0.03831600189208984, 0.03889561462402344, 0.03994112014770508, 0.04019302368164063, 0.03997183990478516, 0.04004761505126953, 0.03841843032836914, 0.03847679901123047, 0.038371326446533204, 0.038234111785888675, 0.03837849426269531, 0.038258689880371094, 0.03819007873535156, 0.03830886459350586, 0.03834982299804687, 0.03825254440307617, 0.03839590454101562, 0.038346752166748044, 0.03817779159545898, 0.03827199935913086, 0.03827609634399414, 0.03823311996459961, 0.03817987060546875, 0.03817059326171875, 0.0383221435546875, 0.038166526794433595, 0.03820646286010742, 0.03827609634399414, 0.03827507019042969, 0.038204414367675785, 0.03846553421020508, 0.07984639739990235, 0.03822284698486328, 0.03814604949951172, 0.038381568908691405, 0.038329345703125, 0.03812659072875976, 0.0381030387878418, 0.0381921272277832, 0.03803033447265625, 0.03826483154296875, 0.03832831954956055, 0.038163455963134765, 0.038128639221191404, 0.03808256149291992, 0.03830476760864258, 0.038152191162109376, 0.03814912033081055, 0.03819724655151367, 0.03816150283813476, 0.03834255981445313, 0.038214656829833986, 0.0388587532043457, 0.03990937423706055, 0.03986841583251953, 0.04010291290283203, 0.040019966125488284, 0.03821670532226563, 0.038198272705078126, 0.03840409469604492, 0.038588417053222655, 0.03825664138793945, 0.038371326446533204, 0.03809075164794922, 0.03829862213134766, 0.03820236968994141, 0.038201343536376955, 0.038217727661132815, 0.038234111785888675, 0.038209537506103515, 0.03817574310302734, 0.038247425079345705, 0.0382371826171875, 0.03834470367431641, 0.03826892852783203, 0.03811840057373047, 0.03825459289550781, 0.03821363067626953, 0.039229438781738284, 0.04001279830932617, 0.040089599609375, 0.0400076789855957, 0.040134654998779294, 0.040187904357910156, 0.040088607788085935, 0.039887840270996094, 0.04031283187866211, 0.04015411376953125, 0.04027084732055664, 0.04039273452758789, 0.03996566390991211, 0.03829145431518555, 0.03822079849243164, 0.04009471893310547, 0.08352470397949219, 0.0401376953125, 0.040349632263183596, 0.04077977752685547, 0.04024524688720703, 0.04009676742553711, 0.040515583038330076, 0.040248321533203124, 0.04014796829223633, 0.04009574508666992, 0.04011929702758789, 0.040210430145263674, 0.040223743438720705, 0.0399738883972168, 0.04003839874267578, 0.0382371826171875, 0.038629375457763675, 0.038491134643554685, 0.03825766372680664, 0.03826483154296875, 0.03846656036376953, 0.038354942321777344, 0.038373374938964845, 0.03835289764404297, 0.03834265518188477, 0.03841535949707031, 0.03825664138793945, 0.038163455963134765, 0.038193153381347655, 0.038416385650634766, 0.03832217788696289, 0.03832320022583008, 0.0391464958190918, 0.040323070526123043, 0.04012851333618164, 0.04029849624633789, 0.04007526397705078, 0.040306751251220706, 0.03839788818359375, 0.0381399040222168, 0.038373374938964845, 0.03852492904663086, 0.038386688232421876, 0.038383617401123046, 0.03821670532226563, 0.03838054275512695, 0.03832012939453125, 0.038214656829833986, 0.03820032119750977, 0.038507518768310545, 0.03825664138793945, 0.03829145431518555, 0.038245376586914064, 0.03831193542480469, 0.038317054748535154, 0.0382740478515625, 0.038214656829833986, 0.038345729827880856, 0.03892124938964844, 0.03984790420532226, 0.04002918243408203, 0.040172542572021484, 0.04021247863769531, 0.08194969940185547, 0.03841331100463867, 0.038258689880371094, 0.038373374938964845, 0.038317054748535154, 0.038228992462158204, 0.03824127960205078, 0.038424575805664066, 0.03839487838745117, 0.038250495910644534, 0.038561790466308594, 0.03838873672485352, 0.04093439865112305, 0.040123390197753905, 0.04016742324829101, 0.040235008239746094, 0.04011212921142578, 0.04035276794433594, 0.04093132781982422, 0.03868569564819336, 0.038564895629882814, 0.0386354866027832, 0.03827199935913086, 0.038193153381347655, 0.038217727661132815, 0.038247425079345705, 0.03840518569946289, 0.038314945220947264, 0.03830988693237305, 0.03851468658447266, 0.03837952041625976, 0.03823308944702149, 0.03839184188842774, 0.0384859848022461, 0.0382105598449707, 0.0382740478515625, 0.03916595077514649, 0.04033740615844727, 0.038247425079345705, 0.03826176071166992, 0.039144447326660156, 0.041452545166015625, 0.040622081756591794, 0.04063641738891602, 0.040891391754150394, 0.04014591979980469, 0.040700927734375, 0.04041318511962891, 0.040308734893798825, 0.04019404983520508, 0.03999129486083984, 0.04006707382202149, 0.040828929901123044, 0.04010905456542969, 0.04021145629882812, 0.04004249572753906, 0.04000460815429688, 0.04015718460083008, 0.04007014465332031, 0.04011315155029297, 0.03992473602294922, 0.040338432312011716, 0.04014899063110351, 0.08364339447021485, 0.03997491073608399, 0.04009267044067383, 0.03998207855224609, 0.040376319885253906, 0.04017766571044922, 0.04027494430541992, 0.04038451385498047, 0.0402503662109375, 0.04045414352416992, 0.04019814300537109, 0.0397916145324707, 0.03818700790405274, 0.0382105598449707, 0.03836108779907227, 0.03888435363769531, 0.04011929702758789, 0.040153087615966795, 0.039880702972412106, 0.03999846267700195, 0.04047872161865235, 0.040325119018554685, 0.040395774841308595, 0.040190975189208986, 0.040041473388671874, 0.040049663543701174, 0.03856281661987305, 0.0383375358581543, 0.038228992462158204, 0.038348800659179685, 0.03831398391723633, 0.03893657684326172, 0.040150016784667966, 0.04014796829223633, 0.03996876907348633, 0.040226814270019534, 0.03987865447998047, 0.04001792144775391, 0.040079360961914064, 0.04005376052856445, 0.040172542572021484, 0.04024422454833984, 0.04010291290283203, 0.04010905456542969, 0.0401162223815918, 0.040218624114990234, 0.03831808090209961, 0.03854848098754883, 0.038324222564697266, 0.038430721282958984, 0.038348800659179685, 0.03827302551269531, 0.03822796630859375, 0.039934974670410156, 0.04017766571044922, 0.038391807556152346, 0.038354942321777344, 0.03833446502685547, 0.03852799987792969, 0.03838771057128906, 0.038245376586914064, 0.03821875381469726, 0.03834163284301758, 0.08003788757324219, 0.0383631362915039, 0.03848601531982422, 0.03860172653198242, 0.038424575805664066, 0.03824127960205078, 0.03828224182128906, 0.03827814483642578, 0.03832217788696289, 0.038422527313232424, 0.03840921783447265, 0.0383559684753418, 0.03847782516479492, 0.038231040954589846, 0.03817375946044922, 0.038246337890625, 0.03848294448852539, 0.038414337158203124, 0.038351871490478515, 0.03842764663696289, 0.038779937744140625, 0.0384532470703125, 0.038394847869873044, 0.038340606689453126, 0.03857612609863281, 0.0383375358581543, 0.03840716934204102, 0.03837855911254883, 0.03835078430175781, 0.038474750518798825, 0.03840921783447265, 0.038712318420410154, 0.0397844467163086, 0.04009881591796875, 0.04036812973022461, 0.0402606086730957, 0.04012441635131836, 0.04048793411254883, 0.04021350479125976, 0.04022886276245117, 0.04016025543212891, 0.03828326416015625, 0.03829350280761719, 0.038520832061767575, 0.03837747192382813, 0.038302719116210936, 0.03847884750366211, 0.03830476760864258, 0.038267902374267575, 0.038665214538574216, 0.0383744010925293, 0.038373374938964845, 0.038384639739990234, 0.038225921630859375, 0.03869081497192383, 0.038553600311279294, 0.03964313507080078, 0.04035276794433594, 0.03975884628295898, 0.04010700988769531, 0.040389633178710936, 0.03976704025268555, 0.04012748718261719]",tokens/s,24.995868176413264,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1281.970176,1017.643008,0.0,371.195904,277.410816,s,10,0.40465283584594725,0.04046528358459473,0.0012749464270020827,0.040339471817016595,0.041213354873657224,0.04256080570220947,0.04363876636505127,"[0.04390825653076172, 0.039191009521484375, 0.04020032119750976, 0.04062038421630859, 0.04001875305175781, 0.04091392135620117, 0.04051433563232422, 0.03929123306274414, 0.03951599884033203, 0.040478622436523434]",tokens/s,6326.4106246733445,kWh,4.743588816529138e-07,2.5992338141804797e-07,1.2597421947630295e-06,1.994024457833991e-06,tokens/kWh,128383580.7500977,MB,1281.970176,1017.643008,0.0,371.195904,323.048448,s,10,24.280882568359374,2.428088256835937,0.01672270399851807,2.436138427734375,2.4378572509765624,2.4385178588867187,2.4390463452148436,"[2.426797119140625, 2.4161357421875, 2.43644775390625, 2.436491455078125, 2.4358291015625, 2.439178466796875, 2.381922119140625, 2.4365849609375, 2.433785400390625, 2.43771044921875]",tokens/s,25.94633857423941,kWh,2.789970010339496e-05,1.5289993593269173e-05,6.758836843923807e-05,0.0001107780621359022,tokens/kWh,568704.65853349,,s,629,24.602083320617677,0.03911301004867675,0.004804808080024372,0.03870412826538086,0.0389910530090332,0.03930890274047851,0.07752167083740236,"[0.039656448364257815, 0.038916095733642575, 0.03911679840087891, 0.038441982269287106, 0.03871846389770508, 0.03863347244262695, 0.03846553421020508, 0.038747135162353515, 0.038983680725097655, 0.038621185302734375, 0.039468032836914066, 0.03973324966430664, 0.040008705139160154, 0.03956224060058594, 0.038986751556396484, 0.03861708831787109, 0.03892736053466797, 0.038964225769042966, 0.03865497589111328, 0.03835289764404297, 0.03786240005493164, 0.03894784164428711, 0.03873689651489258, 0.038623233795166016, 0.03889459228515625, 0.03893350219726562, 0.038599681854248044, 0.038431808471679686, 0.03866719818115234, 0.03852288055419922, 0.03881881713867188, 0.037525505065917966, 0.03853823852539062, 0.03861196899414063, 0.03871027374267578, 0.0388218879699707, 0.03872870254516601, 0.038642688751220705, 0.03880550384521484, 0.0386693115234375, 0.037972991943359374, 0.03834265518188477, 0.038626304626464845, 0.03872358322143555, 0.03853311920166016, 0.038572032928466796, 0.03861913681030273, 0.03886796951293945, 0.039011329650878904, 0.038809600830078124, 0.03837952041625976, 0.037607425689697264, 0.03740979385375977, 0.03724185562133789, 0.03723164749145508, 0.03744559860229492, 0.037335041046142575, 0.037253120422363284, 0.03730022430419922, 0.037302272796630856, 0.03739648056030274, 0.037308414459228514, 0.0768358383178711, 0.03727462387084961, 0.03721932983398438, 0.03744255828857422, 0.03722444915771484, 0.03723263931274414, 0.03723468780517578, 0.037351425170898435, 0.03726028823852539, 0.037222400665283206, 0.03719782257080078, 0.037326847076416016, 0.03750092697143555, 0.038975486755371096, 0.03851468658447266, 0.038324222564697266, 0.03836415863037109, 0.03865702438354492, 0.03859251022338867, 0.03796889495849609, 0.038454273223876956, 0.03858124923706055, 0.03827609634399414, 0.039785472869873044, 0.03924787139892578, 0.03859251022338867, 0.03877068710327149, 0.03864780807495117, 0.038196224212646485, 0.03844095993041992, 0.03861913681030273, 0.038670337677001954, 0.038452224731445314, 0.037917697906494144, 0.03871744155883789, 0.038529022216796875, 0.038520832061767575, 0.038778881072998046, 0.03852799987792969, 0.038389759063720705, 0.03772422409057617, 0.03858425521850586, 0.03871641540527344, 0.03861094284057617, 0.03887411117553711, 0.03873996734619141, 0.03890790557861328, 0.03827711868286133, 0.038902782440185545, 0.03904307174682617, 0.0387922248840332, 0.038702049255371095, 0.03786137771606445, 0.038809600830078124, 0.03896934509277344, 0.03885977554321289, 0.038787071228027346, 0.038757377624511716, 0.03866009521484375, 0.038470657348632815, 0.038882305145263675, 0.03882393646240234, 0.038712318420410154, 0.07939891052246094, 0.03869286346435547, 0.038795265197753906, 0.03887513732910156, 0.03879423904418945, 0.039034881591796876, 0.03891302490234375, 0.038836223602294925, 0.03861708831787109, 0.03800678253173828, 0.03891097640991211, 0.03804876708984375, 0.038882305145263675, 0.038168575286865236, 0.037981182098388674, 0.03827097702026367, 0.03884646224975586, 0.038865921020507815, 0.0377077751159668, 0.03886284637451172, 0.03846460723876953, 0.0388955192565918, 0.0387583999633789, 0.038612991333007815, 0.038899711608886715, 0.038785022735595705, 0.03849932861328125, 0.037820415496826174, 0.038653953552246094, 0.03862220764160156, 0.038809600830078124, 0.03883724975585937, 0.0387512321472168, 0.03888844680786133, 0.03887411117553711, 0.038833152770996096, 0.03888947296142578, 0.03813888168334961, 0.03830380630493164, 0.038789054870605466, 0.038876190185546874, 0.038718433380126954, 0.03883520126342774, 0.03773132705688476, 0.03853110504150391, 0.03896828842163086, 0.03829248046875, 0.037789695739746096, 0.038752254486083985, 0.03887206268310547, 0.03885158538818359, 0.038711296081542966, 0.038781951904296875, 0.03914342498779297, 0.03878400039672852, 0.038970367431640625, 0.038795265197753906, 0.03885772705078125, 0.03893145751953125, 0.03871027374267578, 0.038882305145263675, 0.03935232162475586, 0.03893657684326172, 0.0801976318359375, 0.040052734375, 0.039428096771240234, 0.038870014190673825, 0.038866943359375, 0.03884646224975586, 0.03879219055175781, 0.038886398315429685, 0.038757377624511716, 0.04078496170043945, 0.039140289306640624, 0.03867238235473633, 0.03890687942504883, 0.038779903411865234, 0.03772723388671875, 0.03863865661621094, 0.038687679290771486, 0.03894681549072266, 0.03893964767456055, 0.03887411117553711, 0.03868371200561523, 0.03804358291625977, 0.03752959823608398, 0.03767091369628906, 0.03757567977905273, 0.03858227157592774, 0.038712318420410154, 0.038975486755371096, 0.03836723327636719, 0.03893862533569336, 0.038735870361328126, 0.038797374725341796, 0.038236095428466794, 0.03860889434814453, 0.03884236907958984, 0.03882495880126953, 0.038738975524902346, 0.03886486434936524, 0.03886489486694336, 0.038742015838623044, 0.03873894500732422, 0.03862015914916992, 0.038100990295410156, 0.03866624069213867, 0.03854131317138672, 0.038796287536621094, 0.0379607048034668, 0.03861708831787109, 0.038790145874023435, 0.03870515060424805, 0.03848908615112305, 0.03781836700439453, 0.0387665901184082, 0.038441982269287106, 0.037978111267089845, 0.038556671142578124, 0.03888127899169922, 0.03868569564819336, 0.03868876647949219, 0.03892428970336914, 0.037594112396240234, 0.038866943359375, 0.038779903411865234, 0.07831346893310547, 0.03814604949951172, 0.03882700729370117, 0.038877185821533204, 0.039695358276367186, 0.03878604888916016, 0.0389826545715332, 0.038836223602294925, 0.038972415924072266, 0.03871744155883789, 0.03882393646240234, 0.03850649642944336, 0.03882495880126953, 0.0381921272277832, 0.03886796951293945, 0.038828033447265625, 0.039008255004882815, 0.0389222412109375, 0.038781951904296875, 0.03867750549316406, 0.038834175109863284, 0.03868876647949219, 0.03847577667236328, 0.03872358322143555, 0.03866726303100586, 0.03881062316894531, 0.03909939193725586, 0.038662143707275394, 0.038950912475585936, 0.038830078125, 0.038580257415771486, 0.038977504730224606, 0.03872870254516601, 0.038736961364746095, 0.03875628662109375, 0.038795265197753906, 0.038596607208251955, 0.03853414535522461, 0.03887206268310547, 0.03856076812744141, 0.03790950393676758, 0.038529022216796875, 0.03870412826538086, 0.03858432006835937, 0.037351425170898435, 0.03753881454467774, 0.038594558715820314, 0.038742015838623044, 0.038214656829833986, 0.038109184265136715, 0.03885363388061523, 0.03869696044921875, 0.038675457000732424, 0.038441982269287106, 0.03867750549316406, 0.03830476760864258, 0.0388587532043457, 0.038814720153808595, 0.038749183654785156, 0.038161407470703124, 0.03876147079467773, 0.03878297424316406, 0.03930521774291992, 0.07981977844238282, 0.03822796630859375, 0.03854131317138672, 0.03871027374267578, 0.03872153472900391, 0.03880550384521484, 0.038801406860351564, 0.0385873908996582, 0.03875635147094727, 0.03867852783203125, 0.03861606216430664, 0.038575103759765625, 0.03855052947998047, 0.03796489715576172, 0.03806198501586914, 0.03942092895507812, 0.04028108978271484, 0.038964225769042966, 0.03882393646240234, 0.03866419219970703, 0.03899084854125977, 0.03872051239013672, 0.037768192291259765, 0.03894169616699219, 0.0388935661315918, 0.03912704086303711, 0.03989913558959961, 0.038953983306884765, 0.03872972869873047, 0.03805184173583984, 0.037454849243164064, 0.03857100677490234, 0.038653953552246094, 0.03874508666992187, 0.03873689651489258, 0.038778881072998046, 0.03867340850830078, 0.03877580642700195, 0.03796480178833008, 0.03889152145385742, 0.0387665901184082, 0.03850137710571289, 0.03892633438110352, 0.03906969451904297, 0.0388403205871582, 0.03877171325683594, 0.03902873611450195, 0.03871744155883789, 0.03862015914916992, 0.03813683319091797, 0.038904830932617186, 0.03887308883666992, 0.038991870880126955, 0.03880243301391602, 0.03870207977294922, 0.03907174301147461, 0.03927449417114258, 0.038986751556396484, 0.0388218879699707, 0.03859763336181641, 0.03830579376220703, 0.03789209747314453, 0.03869081497192383, 0.07953612518310547, 0.03946086502075195, 0.03886796951293945, 0.03824127960205078, 0.037698593139648434, 0.037787647247314454, 0.03757564926147461, 0.03742924880981445, 0.037425151824951174, 0.03743027114868164, 0.037174270629882815, 0.037351425170898435, 0.0374200325012207, 0.037800960540771485, 0.037493759155273435, 0.037550079345703126, 0.03741593551635742, 0.037563392639160156, 0.037478462219238284, 0.037314495086669924, 0.037454849243164064, 0.037424129486083986, 0.03731763076782227, 0.03743129730224609, 0.03770265579223633, 0.03746201705932617, 0.03763411331176758, 0.03753977584838867, 0.037443584442138675, 0.03744255828857422, 0.0373196792602539, 0.03894784164428711, 0.03799244689941406, 0.03865497589111328, 0.03867750549316406, 0.038719486236572266, 0.03785830307006836, 0.038100990295410156, 0.03767705535888672, 0.03843379211425781, 0.038870014190673825, 0.03885055923461914, 0.0385054702758789, 0.03881881713867188, 0.03787366485595703, 0.03811635208129883, 0.03767193603515625, 0.03761459350585938, 0.03747123336791992, 0.037459968566894535, 0.03746105575561524, 0.037547969818115236, 0.037591041564941405, 0.03749990463256836, 0.03776716613769531, 0.037759998321533206, 0.03740467071533203, 0.037466110229492186, 0.03745587158203125, 0.03762176132202148, 0.03740467071533203, 0.03756851196289063, 0.037424129486083986, 0.07716556549072266, 0.03832524871826172, 0.03913113784790039, 0.03907583999633789, 0.03930316925048828, 0.04006604766845703, 0.03922431945800781, 0.03863961410522461, 0.03853311920166016, 0.03850035095214844, 0.038386688232421876, 0.03758899307250976, 0.0377077751159668, 0.03966054534912109, 0.03961654281616211, 0.039029727935791014, 0.038882305145263675, 0.03892940902709961, 0.038949951171875, 0.038799297332763674, 0.03883827209472656, 0.03873177719116211, 0.038470657348632815, 0.03772825622558594, 0.03750096130371094, 0.03739542388916016, 0.03849523162841797, 0.038932479858398435, 0.03892736053466797, 0.038724609375, 0.03894784164428711, 0.03870003128051758, 0.03879731369018555, 0.03885772705078125, 0.039008255004882815, 0.03877580642700195, 0.03863449478149414, 0.03875020980834961, 0.038730751037597655, 0.038609920501708986, 0.03889152145385742, 0.03879935836791992, 0.03799859237670898, 0.03885055923461914, 0.03883827209472656, 0.03868364715576172, 0.0378419189453125, 0.03827097702026367, 0.038882366180419924, 0.03889350509643555, 0.03880550384521484, 0.03875328063964844, 0.0384983024597168, 0.03806719970703125, 0.03889766311645508, 0.038747135162353515, 0.03880652618408203, 0.03879731369018555, 0.038317054748535154, 0.03869696044921875, 0.038836223602294925, 0.038951934814453124, 0.03869081497192383, 0.07766015625, 0.03868057632446289, 0.03878092956542969, 0.03884236907958984, 0.03863961410522461, 0.03917420959472656, 0.03880441665649414, 0.03875942230224609, 0.03831193542480469, 0.03883724975585937, 0.03871334457397461, 0.038863872528076174, 0.03870003128051758, 0.0387061767578125, 0.03883929443359375, 0.03847679901123047, 0.0389939193725586, 0.039126014709472655, 0.03876249694824219, 0.03862835311889649, 0.038225921630859375, 0.037928958892822266, 0.038201343536376955, 0.038752254486083985, 0.03788288116455078, 0.038316032409667966, 0.03892838287353516, 0.03883827209472656, 0.038575103759765625, 0.03866624069213867, 0.038742015838623044, 0.03868569564819336, 0.038523902893066404, 0.03863142395019531, 0.03812351989746094, 0.03848396682739258, 0.0384266242980957, 0.0380579833984375, 0.03916595077514649, 0.03902873611450195, 0.03880038452148438, 0.03798527908325195, 0.03868979263305664, 0.03873484802246094, 0.03874303817749023, 0.03897139358520508, 0.03858534240722656, 0.03883724975585937, 0.03884134292602539, 0.03871027374267578, 0.038441982269287106, 0.03866726303100586, 0.038435840606689455, 0.037664768218994144, 0.03777433776855469, 0.03874611282348633, 0.03870412826538086, 0.039122943878173826, 0.03954585647583008, 0.0387061767578125, 0.03868569564819336, 0.038558719635009765, 0.03892326354980469, 0.0795125732421875, 0.03869388961791992, 0.038509567260742186, 0.03903078460693359, 0.03869900894165039, 0.03888844680786133, 0.03870924758911133, 0.03882495880126953, 0.03866828918457031, 0.03871334457397461, 0.03749068832397461, 0.0382371826171875, 0.038708225250244144, 0.03870412826538086, 0.038577152252197267, 0.039311359405517575, 0.03893862533569336, 0.03851878356933594, 0.03875328063964844, 0.03947212982177734, 0.03978035354614258, 0.038626304626464845, 0.03877171325683594, 0.038569984436035154, 0.0387512321472168, 0.038697982788085936, 0.03871027374267578, 0.038152191162109376, 0.03879423904418945, 0.0386324462890625, 0.03985612869262695, 0.038866943359375, 0.038002784729003904, 0.03856582260131836, 0.03872560119628906, 0.038675457000732424, 0.038768638610839845, 0.03867750549316406, 0.03897651290893555, 0.03865599822998047, 0.03887104034423828, 0.03846963119506836, 0.038752254486083985, 0.0383109130859375, 0.039003135681152344, 0.03877273559570313, 0.03880755233764648, 0.038665214538574216, 0.0388853759765625, 0.038697982788085936, 0.03744563293457031, 0.038452224731445314, 0.03876761627197266, 0.038814720153808595, 0.03867750549316406, 0.03863961410522461, 0.03885977554321289, 0.03878604888916016, 0.03861920166015625, 0.039180225372314456, 0.038986751556396484, 0.03813785552978516, 0.037392383575439454]",tokens/s,25.56694048234806,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2029.686784,5274.861568,0.0,4628.414464,4279.64672,s,10,3.174773986816406,0.3174773986816406,0.0009221487872443522,0.3173043212890625,0.31849222106933595,0.31901463470458985,0.319432565612793,"[0.31953704833984375, 0.31614645385742185, 0.31721881103515626, 0.3171442565917969, 0.31738983154296874, 0.3167364807128906, 0.3175615539550781, 0.31837612915039065, 0.31800189208984375, 0.3166615295410156]",tokens/s,806.356613299302,kWh,3.736858639038272e-06,2.0476319108638563e-06,1.8145144724437956e-05,2.392963527434008e-05,tokens/kWh,10698031.836469762,MB,2029.686784,5274.861568,0.0,4628.414464,4463.185408,s,10,176.95450000000002,17.69545,0.0031517961497735305,17.6950830078125,17.699231054687502,17.70004228515625,17.700691269531248,"[17.69897265625, 17.69905078125, 17.700853515625, 17.69500390625, 17.695162109375, 17.69576171875, 17.693421875, 17.69222265625, 17.693814453125, 17.690236328125]",tokens/s,3.5602372361256704,kWh,0.00020892207810655237,0.00011450671390637579,0.0009965308336823604,0.0013199596256952884,tokens/kWh,47728.732586661325,,s,629,179.5270961608886,0.2854166870602364,0.03770346789831642,0.28084429931640625,0.2811535339355469,0.2813114379882813,0.5979956518554688,"[0.2810900573730469, 0.28075418090820314, 0.2808637390136719, 0.28086273193359373, 0.2813634643554688, 0.2807715759277344, 0.2808360900878906, 0.28084429931640625, 0.28089447021484376, 0.28069683837890624, 0.2808883056640625, 0.28103372192382814, 0.28076544189453123, 0.28069683837890624, 0.2809046936035156, 0.28098458862304687, 0.2806855773925781, 0.28079000854492187, 0.2813388671875, 0.2808258666992188, 0.2808616943359375, 0.28091802978515623, 0.2808821716308594, 0.28074496459960935, 0.28068453979492186, 0.28067019653320313, 0.28085964965820315, 0.28079205322265627, 0.2806651000976563, 0.28080743408203124, 0.2807859191894531, 0.2809405517578125, 0.28102041625976565, 0.28240179443359376, 0.28073675537109377, 0.2808053894042969, 0.28098355102539063, 0.28075518798828125, 0.2811678771972656, 0.2814289855957031, 0.28092312622070315, 0.2808555603027344, 0.28077566528320314, 0.2810408935546875, 0.28091595458984375, 0.2808340454101563, 0.28091189575195313, 0.2809783935546875, 0.28075726318359373, 0.28088116455078127, 0.2808842163085937, 0.28079718017578126, 0.2806927490234375, 0.28096408081054686, 0.28156112670898437, 0.28095278930664064, 0.28083203125, 0.281001953125, 0.28085964965820315, 0.28099172973632813, 0.28073165893554686, 0.28094259643554687, 0.59812353515625, 0.2807715759277344, 0.28079205322265627, 0.2807306213378906, 0.28069989013671875, 0.28109515380859373, 0.280732666015625, 0.2807408752441406, 0.28080743408203124, 0.28074188232421876, 0.2810009460449219, 0.2807080993652344, 0.28133273315429685, 0.28084326171875, 0.28075518798828125, 0.28082382202148437, 0.28070196533203123, 0.2806036376953125, 0.28080230712890625, 0.28076544189453123, 0.28069989013671875, 0.28076031494140624, 0.28054937744140623, 0.28094772338867186, 0.2809251708984375, 0.28090264892578126, 0.2812037048339844, 0.2809169921875, 0.2808821716308594, 0.28092312622070315, 0.2808821716308594, 0.2807029724121094, 0.28090673828125, 0.2810491027832031, 0.28080126953125, 0.2807142333984375, 0.28088626098632813, 0.2810163269042969, 0.28109619140625, 0.2811351013183594, 0.28115045166015623, 0.2812467346191406, 0.2808637390136719, 0.2808412170410156, 0.2813102111816406, 0.28072039794921877, 0.28143719482421875, 0.2813716430664063, 0.2811709289550781, 0.28095489501953125, 0.280875, 0.2809036865234375, 0.28120269775390627, 0.2809466857910156, 0.2809661560058594, 0.28132965087890627, 0.28103067016601563, 0.28087808227539063, 0.28130508422851563, 0.2808392028808594, 0.28099169921875, 0.28102554321289064, 0.2814218139648437, 0.5983283081054688, 0.280806396484375, 0.28079000854492187, 0.28083917236328126, 0.28077362060546873, 0.28097125244140625, 0.2809374694824219, 0.28083096313476563, 0.2809302978515625, 0.2808637390136719, 0.28075418090820314, 0.28077362060546873, 0.28114227294921873, 0.2807705688476563, 0.28075314331054685, 0.28077362060546873, 0.28073675537109377, 0.280890380859375, 0.28105523681640626, 0.28109722900390627, 0.28098458862304687, 0.28080126953125, 0.28088626098632813, 0.2811689453125, 0.2808769836425781, 0.2808493957519531, 0.28083917236328126, 0.28073370361328126, 0.280880126953125, 0.2808493957519531, 0.2811535339355469, 0.28096511840820315, 0.28092108154296874, 0.2810409240722656, 0.281107421875, 0.28079000854492187, 0.2810408935546875, 0.2809927673339844, 0.2809743347167969, 0.28115866088867186, 0.28111154174804687, 0.2810992736816406, 0.28116583251953126, 0.28090573120117185, 0.2810828857421875, 0.2810828857421875, 0.2809743347167969, 0.2810828857421875, 0.281702392578125, 0.28085964965820315, 0.2809661560058594, 0.2809671630859375, 0.2809518127441406, 0.2811023254394531, 0.281312255859375, 0.281206787109375, 0.2811064453125, 0.2809405517578125, 0.28098355102539063, 0.2809098205566406, 0.28094259643554687, 0.2807930908203125, 0.2810654602050781, 0.5982269287109375, 0.28070199584960936, 0.2806517333984375, 0.2808186950683594, 0.2806579284667969, 0.2811094970703125, 0.2807715759277344, 0.2807982177734375, 0.28059237670898435, 0.2807500915527344, 0.28073675537109377, 0.2805841979980469, 0.2806558837890625, 0.2806814575195313, 0.2806732788085938, 0.2806876220703125, 0.28078387451171877, 0.28072659301757813, 0.2806312255859375, 0.28082278442382813, 0.28084326171875, 0.28062823486328126, 0.28076544189453123, 0.280748046875, 0.28071014404296873, 0.2810214538574219, 0.2813992919921875, 0.28084530639648436, 0.28090264892578126, 0.2808401794433594, 0.2808299560546875, 0.2809927673339844, 0.28081253051757815, 0.28067636108398436, 0.2807152709960937, 0.2807992248535156, 0.2808186950683594, 0.2809610290527344, 0.2807982177734375, 0.280875, 0.280995849609375, 0.2812569580078125, 0.2808719482421875, 0.2807510986328125, 0.2810491027832031, 0.28073165893554686, 0.280985595703125, 0.28093548583984373, 0.2809241027832031, 0.280806396484375, 0.28093646240234377, 0.2808463439941406, 0.2809794616699219, 0.28087808227539063, 0.2808883056640625, 0.2814617614746094, 0.28086782836914065, 0.28073165893554686, 0.2811760559082031, 0.2809938049316406, 0.28129177856445314, 0.28095187377929687, 0.2817914123535156, 0.5984921875, 0.28099789428710936, 0.28095077514648437, 0.281122802734375, 0.2807726135253906, 0.2808883056640625, 0.28075418090820314, 0.28093646240234377, 0.28081253051757815, 0.28075213623046874, 0.2807726135253906, 0.28081253051757815, 0.28140032958984373, 0.28087808227539063, 0.28151806640625, 0.2811494445800781, 0.2811781005859375, 0.2809241638183594, 0.2809599914550781, 0.28065176391601565, 0.2806476745605469, 0.28059033203125, 0.28050125122070313, 0.2807224426269531, 0.28059039306640626, 0.2806568298339844, 0.2806446228027344, 0.2807173156738281, 0.280627197265625, 0.28063540649414065, 0.2808340454101563, 0.2807080993652344, 0.2810368041992187, 0.2810357666015625, 0.28083712768554686, 0.2806732788085938, 0.2807879638671875, 0.2809333801269531, 0.2808832092285156, 0.280869873046875, 0.2805882873535156, 0.28103067016601563, 0.2808279113769531, 0.2807982177734375, 0.2809129638671875, 0.28075820922851563, 0.28076953125, 0.2810152893066406, 0.28108184814453124, 0.2806855773925781, 0.2812528686523437, 0.2812149658203125, 0.28091494750976564, 0.28111154174804687, 0.28087908935546874, 0.2808330383300781, 0.2809251708984375, 0.28099172973632813, 0.28070401000976564, 0.2810091552734375, 0.2808401794433594, 0.2807623901367188, 0.2810029907226563, 0.5976668090820313, 0.28066201782226563, 0.280774658203125, 0.28086578369140625, 0.28072857666015627, 0.2810163879394531, 0.28058001708984376, 0.2806947937011719, 0.28056781005859377, 0.2807726135253906, 0.28080230712890625, 0.280553466796875, 0.28076544189453123, 0.280690673828125, 0.2808268737792969, 0.280875, 0.28081048583984375, 0.28086477661132814, 0.28081048583984375, 0.2810245056152344, 0.2807193603515625, 0.2807080993652344, 0.2806855773925781, 0.2810071105957031, 0.2810624084472656, 0.28088525390625, 0.28093646240234377, 0.2807060546875, 0.28088116455078127, 0.2806405029296875, 0.28068658447265626, 0.2808965148925781, 0.28075930786132813, 0.2807490539550781, 0.28122213745117186, 0.280764404296875, 0.28102655029296875, 0.28092825317382814, 0.28138290405273436, 0.28117196655273435, 0.28144537353515625, 0.2809876403808594, 0.281027587890625, 0.28075424194335935, 0.2808401184082031, 0.28082278442382813, 0.28102655029296875, 0.28107366943359374, 0.28075314331054685, 0.28084735107421877, 0.28093438720703123, 0.28076544189453123, 0.28106341552734376, 0.28107879638671873, 0.28101837158203125, 0.2814535827636719, 0.28100811767578127, 0.28086477661132814, 0.2809016418457031, 0.28101119995117185, 0.2808832092285156, 0.28088528442382815, 0.28095895385742187, 0.598540283203125, 0.2806599731445312, 0.2808401794433594, 0.2807408752441406, 0.28097842407226564, 0.2807705688476563, 0.28113101196289064, 0.28089547729492187, 0.28100607299804686, 0.28068453979492186, 0.2810439758300781, 0.28067019653320313, 0.28120269775390627, 0.2807859191894531, 0.280658935546875, 0.2809333801269531, 0.28086578369140625, 0.2808279113769531, 0.28084530639648436, 0.2808913879394531, 0.28075314331054685, 0.28059237670898435, 0.28081561279296874, 0.28066815185546873, 0.28071832275390624, 0.2806855773925781, 0.28078182983398436, 0.2810224914550781, 0.2808012390136719, 0.28073370361328126, 0.28072857666015627, 0.2806651000976563, 0.2808084411621094, 0.280742919921875, 0.2807459716796875, 0.2807408752441406, 0.28083096313476563, 0.280958984375, 0.2808985595703125, 0.2808555603027344, 0.2810992736816406, 0.2811351013183594, 0.2806947937011719, 0.28073370361328126, 0.2807859191894531, 0.2808637390136719, 0.2812098693847656, 0.2812231750488281, 0.28102041625976565, 0.2809169921875, 0.2809938049316406, 0.28092007446289063, 0.2809825134277344, 0.2808616943359375, 0.28068453979492186, 0.28086782836914065, 0.2807234497070312, 0.28062619018554685, 0.2807685241699219, 0.28067123413085937, 0.2807296142578125, 0.2808002624511719, 0.28117300415039065, 0.5992601318359375, 0.2805698547363281, 0.2806220703125, 0.28086578369140625, 0.2806927490234375, 0.28068658447265626, 0.28074188232421876, 0.28050225830078124, 0.28068658447265626, 0.2806220703125, 0.28081460571289063, 0.2807357482910156, 0.28070706176757815, 0.28100311279296875, 0.28070492553710935, 0.2808493957519531, 0.2806118469238281, 0.2809036865234375, 0.2807828369140625, 0.2809046936035156, 0.2806640625, 0.28067636108398436, 0.28069888305664065, 0.28089959716796875, 0.2806876220703125, 0.28073165893554686, 0.28076544189453123, 0.2807173156738281, 0.2807705688476563, 0.2804951171875, 0.2809938049316406, 0.2808279113769531, 0.28095693969726565, 0.28122726440429685, 0.2810163269042969, 0.28069989013671875, 0.2808842163085937, 0.28082278442382813, 0.28080126953125, 0.28084429931640625, 0.28069580078125, 0.2806753234863281, 0.28080230712890625, 0.2806804504394531, 0.2807798461914062, 0.2811206970214844, 0.28091802978515623, 0.28079718017578126, 0.28101937866210935, 0.2810828857421875, 0.2809466857910156, 0.2809190368652344, 0.2809415588378906, 0.28108389282226565, 0.28117913818359375, 0.28109515380859373, 0.2810357666015625, 0.2808545227050781, 0.280890380859375, 0.2808606872558594, 0.28100811767578127, 0.28085760498046874, 0.28095489501953125, 0.5989181518554687, 0.28075518798828125, 0.28063436889648435, 0.2807193603515625, 0.2805350341796875, 0.2809671630859375, 0.2804613037109375, 0.28068453979492186, 0.28063641357421876, 0.2805770263671875, 0.2807633972167969, 0.28062924194335936, 0.2807562255859375, 0.2807296142578125, 0.2805565490722656, 0.2806087646484375, 0.2806476745605469, 0.2805893249511719, 0.28073983764648436, 0.28069171142578125, 0.2810378112792969, 0.28060671997070313, 0.28059442138671875, 0.28090673828125, 0.28125799560546877, 0.28063436889648435, 0.28085861206054685, 0.280511474609375, 0.28059442138671875, 0.2806200256347656, 0.2808002624511719, 0.2806661071777344, 0.280890380859375, 0.28054629516601565, 0.2807828369140625, 0.28066302490234374, 0.280732666015625, 0.2807224426269531, 0.28073165893554686, 0.2807142333984375, 0.28113101196289064, 0.280958984375, 0.28107366943359374, 0.2807439270019531, 0.2809518127441406, 0.28086578369140625, 0.28080435180664065, 0.2811535339355469, 0.2808248291015625, 0.28094064331054686, 0.2808831176757812, 0.2809313354492188, 0.2811975708007812, 0.281122802734375, 0.2807705688476563, 0.2809292907714844, 0.281385986328125, 0.2810009460449219, 0.2827796630859375, 0.28104193115234377, 0.2813173828125, 0.2812630920410156, 0.2812651672363281, 0.597570556640625, 0.2809108581542969, 0.28094259643554687, 0.28096307373046875, 0.280585205078125, 0.2806241149902344, 0.28071218872070314, 0.28067840576171876, 0.2805749816894531, 0.2805186767578125, 0.28062612915039065, 0.280669189453125, 0.2806609802246094, 0.2807214050292969, 0.28059954833984374, 0.28069888305664065, 0.28063540649414065, 0.28063128662109377, 0.2805401611328125, 0.28073983764648436, 0.28051763916015626, 0.28082278442382813, 0.2806200256347656, 0.2806640625, 0.28071218872070314, 0.2807879638671875, 0.28067123413085937, 0.2807459716796875, 0.2807715759277344, 0.2807060546875, 0.28080743408203124, 0.28084326171875, 0.28085247802734375, 0.2809169921875, 0.2808770446777344, 0.28057907104492186, 0.28081048583984375, 0.28083712768554686, 0.28081765747070314, 0.28080743408203124, 0.28084326171875, 0.2808401794433594, 0.28075314331054685, 0.2806026306152344, 0.28084640502929686, 0.28075103759765624, 0.2807500915527344, 0.2811566162109375, 0.2810798034667969, 0.28084429931640625, 0.2807982177734375, 0.28074188232421876, 0.28069989013671875, 0.28078695678710935, 0.280806396484375, 0.2809016418457031, 0.28089447021484376, 0.2807982177734375, 0.28089547729492187, 0.28093438720703123, 0.2809241638183594, 0.2807132568359375, 0.28090365600585937]",tokens/s,3.50364938469401,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10016,7 +10016,7 @@ ImportError: This modeling file requires the following packages that were not fo 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,3918.962688,5287.44448,0.0,4640.997376,4330.799616,s,10,3.2259157409667965,0.32259157409667966,0.0005529953640440514,0.3225971221923828,0.3233306610107422,0.32335841522216796,0.3233806185913086,"[0.32232916259765626, 0.32165042114257814, 0.3231456298828125, 0.32269686889648436, 0.3224973754882812, 0.3220177917480469, 0.32332449340820313, 0.32209555053710937, 0.3233861694335938, 0.3227722778320313]",tokens/s,793.5731139812028,kWh,3.802365012880829e-06,2.0835388033233454e-06,1.894839189480954e-05,2.4834295711013717e-05,tokens/kWh,10308325.348903172,MB,3918.962688,5287.44448,0.0,4640.997376,4515.66848,s,10,175.46737109375,17.546737109375,0.001177635321808986,17.546580078125,17.548339062500002,17.548616796875,17.548838984375,"[17.54512890625, 17.5464765625, 17.54570703125, 17.54668359375, 17.54525, 17.54827734375, 17.54630078125, 17.547208984375, 17.54889453125, 17.547443359375]",tokens/s,3.590411117878998,kWh,0.00020716305749490858,0.000113541591573937,0.0010045058435346013,0.0013252104926034468,tokens/kWh,47539.61755632732,,s,629,178.09222891235356,0.28313549906574487,0.03836657912459716,0.27851162719726563,0.27869696044921877,0.2787557373046875,0.6009740600585938,"[0.2784675903320312, 0.2784378967285156, 0.2781951904296875, 0.2783365173339844, 0.27832830810546877, 0.2783682556152344, 0.2783068237304688, 0.27829656982421874, 0.27835595703125, 0.2783907775878906, 0.27824639892578124, 0.2782801818847656, 0.27823822021484373, 0.2784276428222656, 0.27843072509765626, 0.27827508544921875, 0.27839898681640624, 0.2785382385253906, 0.2784798583984375, 0.2784778137207031, 0.2784051208496094, 0.2784163818359375, 0.2784532470703125, 0.2783907775878906, 0.27848089599609377, 0.27852593994140623, 0.27842868041992186, 0.2785771484375, 0.27849215698242186, 0.27850341796875, 0.27848907470703127, 0.27865191650390625, 0.278687744140625, 0.27855462646484375, 0.27841741943359377, 0.278445068359375, 0.2784675903320312, 0.27849932861328125, 0.2784747619628906, 0.2785249328613281, 0.27833856201171875, 0.27862322998046873, 0.2785730590820312, 0.27850341796875, 0.27848089599609377, 0.2786488342285156, 0.27850137329101565, 0.27865396118164065, 0.2785515441894531, 0.27858328247070313, 0.27858123779296873, 0.2786918334960938, 0.2785577087402344, 0.2786856994628906, 0.27846246337890623, 0.278719482421875, 0.2786211853027344, 0.2786273193359375, 0.2784972839355469, 0.27865191650390625, 0.27861709594726564, 0.27877069091796874, 0.6005718994140625, 0.2785904541015625, 0.27841741943359377, 0.278265869140625, 0.27840103149414064, 0.278761474609375, 0.27836517333984373, 0.278371337890625, 0.27834674072265625, 0.27851263427734374, 0.2784460754394531, 0.27822695922851565, 0.27832321166992186, 0.27824026489257814, 0.27833956909179686, 0.27833856201171875, 0.2783477783203125, 0.2783631286621094, 0.2784829406738281, 0.27833139038085936, 0.278445068359375, 0.27841741943359377, 0.27845632934570314, 0.278518798828125, 0.2784481201171875, 0.27845428466796873, 0.27843994140625, 0.278403076171875, 0.2785658874511719, 0.27850546264648435, 0.27858123779296873, 0.27847064208984373, 0.278645751953125, 0.2785935363769531, 0.27854437255859377, 0.278513671875, 0.2785935363769531, 0.27843890380859376, 0.27848703002929687, 0.27840716552734374, 0.278572021484375, 0.27874200439453123, 0.27856179809570314, 0.27862014770507815, 0.27857510375976563, 0.27847271728515627, 0.2787010498046875, 0.2785290222167969, 0.27869287109375, 0.2786938781738281, 0.27897445678710936, 0.2785535888671875, 0.2785587158203125, 0.2785372009277344, 0.27866827392578125, 0.27850750732421875, 0.27867852783203123, 0.2785105895996094, 0.27856179809570314, 0.27850546264648435, 0.2785904541015625, 0.27853619384765627, 0.27867852783203123, 0.602208251953125, 0.278476806640625, 0.27831808471679687, 0.2782474365234375, 0.27843173217773437, 0.2783057861328125, 0.27829556274414063, 0.2782945251464844, 0.27835186767578124, 0.278297607421875, 0.2783129577636719, 0.2783918151855469, 0.2784460754394531, 0.278329345703125, 0.27832321166992186, 0.27822283935546877, 0.2788935546875, 0.2784245910644531, 0.27843585205078125, 0.27840921020507814, 0.2784132995605469, 0.27833242797851565, 0.278619140625, 0.2783498229980469, 0.27841537475585937, 0.278392822265625, 0.2784245910644531, 0.27871743774414065, 0.27856997680664064, 0.27848193359375, 0.27866009521484375, 0.2784880676269531, 0.2784696350097656, 0.27849932861328125, 0.2784460754394531, 0.27853515625, 0.2785382385253906, 0.2783631286621094, 0.2785986633300781, 0.2784296875, 0.2785321044921875, 0.27840615844726563, 0.2784778137207031, 0.2785730590820312, 0.2785024108886719, 0.2784245910644531, 0.27858740234375, 0.2785238952636719, 0.27861505126953123, 0.27867034912109373, 0.2786283569335937, 0.2786007080078125, 0.2790328369140625, 0.27862527465820314, 0.2787256469726562, 0.2784747619628906, 0.27861196899414065, 0.278619140625, 0.278724609375, 0.2785935363769531, 0.27863348388671877, 0.278540283203125, 0.27875839233398436, 0.6016409301757812, 0.27838360595703127, 0.2785607604980469, 0.2782525329589844, 0.27845733642578124, 0.2782443542480469, 0.2783703002929688, 0.2783139953613281, 0.27840408325195315, 0.2783375244140625, 0.27838772583007815, 0.27835903930664063, 0.278392822265625, 0.27826995849609376, 0.27838055419921875, 0.2785269775390625, 0.2785331115722656, 0.27831808471679687, 0.27850958251953123, 0.2784471130371094, 0.27848089599609377, 0.2784481201171875, 0.2783477783203125, 0.2785208435058594, 0.27845016479492185, 0.27845016479492185, 0.27848602294921876, 0.27848397827148436, 0.27852801513671877, 0.2785433654785156, 0.27850750732421875, 0.27853927612304685, 0.27850650024414064, 0.2785740661621094, 0.2785577087402344, 0.27850546264648435, 0.2784829406738281, 0.27839385986328125, 0.27899493408203124, 0.27845120239257815, 0.2785689697265625, 0.27843585205078125, 0.27867544555664064, 0.2785535888671875, 0.2785884094238281, 0.2786611328125, 0.27860275268554685, 0.27862322998046873, 0.278677490234375, 0.2786437072753906, 0.2785607604980469, 0.2785710144042969, 0.27888128662109374, 0.27861093139648435, 0.2785730590820312, 0.278513671875, 0.2786631774902344, 0.27858740234375, 0.2785740661621094, 0.278498291015625, 0.27868057250976563, 0.2785986633300781, 0.27873382568359373, 0.6014218139648437, 0.27839385986328125, 0.2783211669921875, 0.278245361328125, 0.2784296875, 0.27821466064453126, 0.2783621215820313, 0.2782853088378906, 0.2784296875, 0.2782576599121094, 0.27834368896484374, 0.2782033996582031, 0.2784378967285156, 0.2782167053222656, 0.27833856201171875, 0.2784532470703125, 0.2783006591796875, 0.2784163818359375, 0.27851263427734374, 0.27829656982421874, 0.2784829406738281, 0.2782607421875, 0.27854745483398435, 0.2784122924804687, 0.2784747619628906, 0.27844403076171875, 0.27850546264648435, 0.27843994140625, 0.27860379028320315, 0.27854745483398435, 0.27849932861328125, 0.2784655456542969, 0.27863961791992187, 0.2785003662109375, 0.27850137329101565, 0.27850341796875, 0.27838156127929686, 0.2785167236328125, 0.2785791931152344, 0.2784675903320312, 0.27858944702148436, 0.2783692932128906, 0.2785310668945313, 0.2785157165527344, 0.2785218505859375, 0.27848907470703127, 0.27869696044921877, 0.2786590576171875, 0.27863348388671877, 0.2786990051269531, 0.278697998046875, 0.2785228881835938, 0.27875326538085937, 0.27864474487304686, 0.2785310668945313, 0.27853619384765627, 0.27875225830078126, 0.27861709594726564, 0.278640625, 0.27856280517578125, 0.2788239440917969, 0.2785249328613281, 0.2788229064941406, 0.6009180297851563, 0.27875738525390625, 0.27840408325195315, 0.2787317810058594, 0.27844198608398435, 0.2784952392578125, 0.2783703002929688, 0.2784204711914062, 0.27834878540039065, 0.2785208435058594, 0.27833139038085936, 0.2782863464355469, 0.27851980590820313, 0.27829144287109375, 0.2784184265136719, 0.27834060668945315, 0.27849114990234375, 0.27847885131835937, 0.2785331115722656, 0.27837542724609377, 0.2785218505859375, 0.27835595703125, 0.27865805053710935, 0.27854232788085936, 0.27851776123046873, 0.27856280517578125, 0.2784716796875, 0.2784163818359375, 0.2786263122558594, 0.27854232788085936, 0.27852801513671877, 0.2785372009277344, 0.2785607604980469, 0.2785638427734375, 0.27857818603515627, 0.27848193359375, 0.2784655456542969, 0.2784696350097656, 0.27858944702148436, 0.27878091430664065, 0.2787389526367188, 0.278572021484375, 0.2785587158203125, 0.2785300598144531, 0.278513671875, 0.2785515441894531, 0.27863961791992187, 0.2785525817871094, 0.2786160583496094, 0.27866622924804685, 0.27864678955078126, 0.2784737243652344, 0.27865191650390625, 0.2786570129394531, 0.2786816101074219, 0.2785464172363281, 0.2786365356445312, 0.2785863647460938, 0.2786990051269531, 0.27853619384765627, 0.2786836547851563, 0.27861505126953123, 0.27869696044921877, 0.60225537109375, 0.2784429931640625, 0.2783365173339844, 0.2782105712890625, 0.27842868041992186, 0.27831195068359377, 0.27833242797851565, 0.2782791748046875, 0.27833447265625, 0.27844403076171875, 0.2783211669921875, 0.27834674072265625, 0.278361083984375, 0.27834573364257814, 0.27851162719726563, 0.27837542724609377, 0.27849319458007815, 0.27838360595703127, 0.2784901123046875, 0.2784737243652344, 0.2785658874511719, 0.27838772583007815, 0.27850750732421875, 0.2784481201171875, 0.27856179809570314, 0.27863961791992187, 0.27845632934570314, 0.2784112548828125, 0.2784471130371094, 0.27855563354492185, 0.27854745483398435, 0.27850341796875, 0.2786283569335937, 0.27844403076171875, 0.27850341796875, 0.2784655456542969, 0.27848089599609377, 0.27845428466796873, 0.2786590576171875, 0.2785300598144531, 0.2785771484375, 0.2784051208496094, 0.2785986633300781, 0.2785863647460938, 0.278540283203125, 0.27851162719726563, 0.27865191650390625, 0.2786641845703125, 0.27860888671875, 0.2785986633300781, 0.2785607604980469, 0.27842355346679687, 0.2787317810058594, 0.27872869873046874, 0.2786344909667969, 0.278513671875, 0.2787471313476563, 0.27864984130859377, 0.2786918334960938, 0.27854437255859377, 0.278729736328125, 0.2786068420410156, 0.27870822143554685, 0.600995849609375, 0.2783549499511719, 0.27836007690429687, 0.278371337890625, 0.27842355346679687, 0.2783150024414062, 0.2783477783203125, 0.2783150024414062, 0.278302734375, 0.278265869140625, 0.27850650024414064, 0.27838772583007815, 0.27848397827148436, 0.27830374145507814, 0.2784696350097656, 0.2783365173339844, 0.278434814453125, 0.27846246337890623, 0.2786211853027344, 0.27852801513671877, 0.2786058349609375, 0.2782320556640625, 0.27850958251953123, 0.27844198608398435, 0.27843994140625, 0.27850445556640624, 0.27860479736328125, 0.2783682556152344, 0.2785986633300781, 0.27849215698242186, 0.278518798828125, 0.27844403076171875, 0.2786017150878906, 0.2785689697265625, 0.27850750732421875, 0.27848193359375, 0.278582275390625, 0.2784532470703125, 0.27861810302734374, 0.27844403076171875, 0.27855667114257815, 0.2784184265136719, 0.278572021484375, 0.2786222229003906, 0.2785904541015625, 0.2785658874511719, 0.27858944702148436, 0.27860479736328125, 0.2786621398925781, 0.2785904541015625, 0.2787461242675781, 0.2785249328613281, 0.278729736328125, 0.27870513916015627, 0.2786385803222656, 0.27873690795898437, 0.2787164306640625, 0.278656005859375, 0.2786437072753906, 0.27850546264648435, 0.27900827026367186, 0.2785638427734375, 0.278751220703125, 0.6023884887695312, 0.2784880676269531, 0.2782607421875, 0.27817062377929686, 0.2784378967285156, 0.27832421875, 0.27845733642578124, 0.27837850952148435, 0.27836517333984373, 0.2786263122558594, 0.2783846435546875, 0.2783068237304688, 0.2785535888671875, 0.27822796630859375, 0.278466552734375, 0.27838668823242185, 0.2784409484863281, 0.27841537475585937, 0.2784962463378906, 0.27847885131835937, 0.2784583740234375, 0.2783365173339844, 0.27845220947265625, 0.2784747619628906, 0.2787461242675781, 0.2785228881835938, 0.27894476318359374, 0.27845733642578124, 0.27854132080078126, 0.2785638427734375, 0.27855563354492185, 0.27856997680664064, 0.27877786254882814, 0.2786375732421875, 0.27846450805664064, 0.278371337890625, 0.27854849243164065, 0.27875839233398436, 0.27850650024414064, 0.2785464172363281, 0.27856793212890624, 0.27850341796875, 0.2785372009277344, 0.2785372009277344, 0.2785802307128906, 0.2786058349609375, 0.2786017150878906, 0.2786488342285156, 0.27872357177734375, 0.27849114990234375, 0.27882699584960935, 0.2785935363769531, 0.2786324462890625, 0.27864166259765627, 0.27864678955078126, 0.278508544921875, 0.2786897888183594, 0.27858123779296873, 0.2786375732421875, 0.27848193359375, 0.2791280517578125, 0.2786713562011719, 0.27919769287109375, 0.601754638671875, 0.27856484985351565, 0.27835699462890623, 0.2782443542480469, 0.27838772583007815, 0.27829861450195315, 0.2783529052734375, 0.27829861450195315, 0.27851263427734374, 0.27839898681640624, 0.27828839111328124, 0.27840615844726563, 0.2784962463378906, 0.2782300109863281, 0.27842868041992186, 0.2783846435546875, 0.27839486694335935, 0.27840716552734374, 0.27844198608398435, 0.2783999938964844, 0.2784880676269531, 0.27841537475585937, 0.27849114990234375, 0.278540283203125, 0.27850650024414064, 0.27849420166015626, 0.27862322998046873, 0.27851162719726563, 0.2787327880859375, 0.2785658874511719, 0.2785986633300781, 0.27849215698242186, 0.27883621215820314, 0.27854437255859377, 0.2786068420410156, 0.2784635009765625, 0.2785802307128906, 0.27857510375976563, 0.27845120239257815, 0.278466552734375, 0.27857818603515627, 0.27845938110351565, 0.27869491577148436, 0.27854132080078126, 0.2785986633300781, 0.2784737243652344, 0.278582275390625, 0.2786631774902344, 0.27878091430664065, 0.27861196899414065, 0.278540283203125, 0.27854949951171876, 0.27867239379882813, 0.2786641845703125, 0.27863961791992187, 0.27857614135742187, 0.27870925903320315, 0.27857614135742187, 0.2787891235351562, 0.2785321044921875, 0.278697998046875, 0.2785003662109375, 0.278793212890625]",tokens/s,3.531877858126851,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1290.141696,893.91104,0.0,247.463936,189.067776,s,10,0.17335734748840334,0.017335734748840335,0.0006986572357890928,0.017282976150512695,0.017675709533691405,0.01844868698120117,0.019067068939208984,"[0.019221664428710938, 0.01659644889831543, 0.017503936767578124, 0.01656208038330078, 0.01697455978393555, 0.017311616897583006, 0.01723129653930664, 0.01738979148864746, 0.017264448165893554, 0.017301504135131835]",tokens/s,14767.184876148673,kWh,1.9831383332754588e-07,1.0866644055836607e-07,4.57388975157767e-07,7.643692490436789e-07,tokens/kWh,334916665.3685871,MB,1290.141696,893.91104,0.0,247.463936,209.72544,s,10,10.436258666992186,1.0436258666992186,0.010111977487859941,1.0485770874023437,1.0494609252929687,1.050062371826172,1.0505435290527345,"[1.0226334228515626, 1.0244058837890626, 1.050663818359375, 1.046568603515625, 1.048908203125, 1.04754736328125, 1.0486011962890625, 1.0490499267578124, 1.0493272705078125, 1.048552978515625]",tokens/s,60.366460826863644,kWh,1.2123845675536542e-05,6.643003766727856e-06,2.4958020246038906e-05,4.3724869688303304e-05,tokens/kWh,1440827.621651047,,s,629,10.57114009952544,0.016806264069197854,0.002093122475857818,0.0166246395111084,0.016762265777587892,0.01684336700439453,0.03366842239379886,"[0.0176629753112793, 0.017868799209594728, 0.01615564727783203, 0.01593958377838135, 0.01593958377838135, 0.01584332847595215, 0.01585049629211426, 0.015973376274108886, 0.015890432357788087, 0.015871999740600586, 0.01590272045135498, 0.015851519584655763, 0.01586176013946533, 0.015874048233032227, 0.0158341121673584, 0.01598464012145996, 0.015884287834167482, 0.01589145565032959, 0.015879167556762695, 0.01635430335998535, 0.016563199996948243, 0.01657142448425293, 0.016606176376342773, 0.016712703704833985, 0.016513023376464844, 0.016590848922729492, 0.01646080017089844, 0.016466943740844727, 0.016539648056030275, 0.016448511123657226, 0.016493568420410155, 0.016873472213745116, 0.017076223373413087, 0.017175552368164062, 0.017912832260131836, 0.016334848403930666, 0.016071680068969727, 0.016046079635620117, 0.016097280502319337, 0.01587712001800537, 0.01589964771270752, 0.015855615615844726, 0.015866880416870118, 0.016, 0.015978495597839357, 0.01580851173400879, 0.015851519584655763, 0.01581670379638672, 0.01593958377838135, 0.01594265556335449, 0.015951871871948242, 0.015827967643737794, 0.015838208198547362, 0.015857664108276368, 0.015928319931030274, 0.015857664108276368, 0.015780863761901857, 0.015846400260925292, 0.015830016136169432, 0.01585868835449219, 0.01580031967163086, 0.01600716781616211, 0.03281612777709961, 0.015912960052490235, 0.015828991889953615, 0.01584537601470947, 0.01580031967163086, 0.015820832252502443, 0.015781855583190917, 0.015782912254333495, 0.01584537601470947, 0.015941632270812987, 0.015981568336486816, 0.015823871612548827, 0.015827967643737794, 0.015830016136169432, 0.015791104316711425, 0.015830016136169432, 0.01580851173400879, 0.015788031578063966, 0.01579417610168457, 0.01577779197692871, 0.015791104316711425, 0.015927295684814453, 0.01576857566833496, 0.015823904037475585, 0.015766495704650878, 0.01576857566833496, 0.01576038360595703, 0.015887359619140624, 0.01580953598022461, 0.015804415702819825, 0.015751168251037596, 0.01620992088317871, 0.01657651138305664, 0.016671743392944336, 0.01661030387878418, 0.01659699249267578, 0.016531455993652345, 0.01659699249267578, 0.01657241630554199, 0.016523263931274415, 0.01658880043029785, 0.01663385581970215, 0.016664575576782227, 0.016541696548461913, 0.016542720794677734, 0.01662566375732422, 0.01663692855834961, 0.01662873649597168, 0.01658470344543457, 0.01660211181640625, 0.01660108757019043, 0.01682329559326172, 0.016644096374511717, 0.016726015090942382, 0.016631807327270508, 0.01664204788208008, 0.01677824020385742, 0.016652288436889647, 0.01665023994445801, 0.01662668800354004, 0.016664575576782227, 0.016706560134887697, 0.018386943817138672, 0.03456409454345703, 0.016761856079101564, 0.0165980167388916, 0.016696319580078126, 0.01661337661743164, 0.016709632873535156, 0.01657344055175781, 0.016691200256347655, 0.01660927963256836, 0.0168089599609375, 0.01662259292602539, 0.01663692855834961, 0.016693248748779296, 0.016867328643798828, 0.016676864624023437, 0.016776191711425782, 0.016703487396240235, 0.016655359268188476, 0.01661030387878418, 0.01677107238769531, 0.016653312683105468, 0.01677824020385742, 0.01663692855834961, 0.016663551330566406, 0.016681983947753908, 0.01680486488342285, 0.016677888870239257, 0.016865280151367186, 0.01675468826293945, 0.016883712768554687, 0.016670719146728515, 0.016719871520996094, 0.016915456771850586, 0.016685056686401366, 0.016695295333862305, 0.0172042236328125, 0.01613209533691406, 0.016484352111816408, 0.0166430721282959, 0.01661337661743164, 0.01660825538635254, 0.016729087829589845, 0.01662259292602539, 0.016673791885375978, 0.01658470344543457, 0.01659699249267578, 0.01658060836791992, 0.016684032440185546, 0.016668672561645507, 0.016849920272827147, 0.016669696807861328, 0.01657344055175781, 0.0166430721282959, 0.01660620880126953, 0.016679935455322266, 0.01681817626953125, 0.016763904571533202, 0.016678911209106445, 0.016291839599609375, 0.016350208282470705, 0.01640140724182129, 0.016780288696289062, 0.01658982467651367, 0.032780288696289066, 0.015847423553466796, 0.01586995220184326, 0.016199680328369142, 0.016530431747436524, 0.01659903907775879, 0.016520191192626953, 0.016704511642456055, 0.01655705642700195, 0.01670966339111328, 0.01667580795288086, 0.01656729507446289, 0.016672767639160157, 0.016649215698242188, 0.01657756805419922, 0.01667683219909668, 0.01662054443359375, 0.0166297607421875, 0.016652288436889647, 0.016715776443481444, 0.016690208435058595, 0.01671881675720215, 0.016662527084350585, 0.01659187126159668, 0.016660480499267577, 0.016702463150024414, 0.016734207153320312, 0.016631807327270508, 0.016649215698242188, 0.01662668800354004, 0.01657651138305664, 0.0165928955078125, 0.01660211181640625, 0.01664102363586426, 0.016755712509155272, 0.01660620880126953, 0.01657753562927246, 0.016578559875488282, 0.016538623809814454, 0.01663795280456543, 0.016549888610839843, 0.016681983947753908, 0.01657344055175781, 0.01660723114013672, 0.01663488006591797, 0.016661504745483398, 0.0166748161315918, 0.01660416030883789, 0.016704511642456055, 0.016594944000244142, 0.016739328384399413, 0.016781312942504883, 0.016680959701538087, 0.016887807846069337, 0.016635904312133788, 0.016587776184082033, 0.01660416030883789, 0.017123327255249024, 0.016730112075805666, 0.016656383514404297, 0.01661952018737793, 0.016679935455322266, 0.01663488006591797, 0.033999870300292966, 0.016693248748779296, 0.0166246395111084, 0.01663385581970215, 0.016559104919433593, 0.016686080932617187, 0.01657344055175781, 0.016861183166503906, 0.016689151763916017, 0.01661440086364746, 0.01660211181640625, 0.016712703704833985, 0.016664575576782227, 0.016639999389648438, 0.016566272735595702, 0.01666662406921387, 0.016730112075805666, 0.016750591278076172, 0.016558080673217773, 0.016583711624145507, 0.016715744018554687, 0.0166246395111084, 0.016611328125, 0.016748544692993163, 0.01682841682434082, 0.016564224243164064, 0.01659596824645996, 0.016544767379760742, 0.016571392059326173, 0.0166297607421875, 0.01678233528137207, 0.01658572769165039, 0.016587776184082033, 0.01660723114013672, 0.01682841682434082, 0.016683008193969725, 0.01680384063720703, 0.016679935455322266, 0.016562175750732423, 0.016552959442138672, 0.01654374313354492, 0.016672767639160157, 0.016587776184082033, 0.016647167205810547, 0.016582656860351562, 0.01658060836791992, 0.016675840377807616, 0.01665843200683594, 0.016724992752075195, 0.016664575576782227, 0.01662566375732422, 0.01662668800354004, 0.016733184814453125, 0.01666662406921387, 0.016645120620727538, 0.0165928955078125, 0.01658880043029785, 0.01663488006591797, 0.016555007934570314, 0.016547840118408205, 0.01661337661743164, 0.01660620880126953, 0.01662054443359375, 0.03440639877319336, 0.01659903907775879, 0.016659456253051756, 0.016648191452026367, 0.016559104919433593, 0.01657651138305664, 0.016587776184082033, 0.01661644744873047, 0.01676697540283203, 0.01660313606262207, 0.01660416030883789, 0.016558080673217773, 0.016709632873535156, 0.01660211181640625, 0.016653312683105468, 0.016664575576782227, 0.016570367813110352, 0.016704511642456055, 0.016742399215698242, 0.01682636833190918, 0.016648191452026367, 0.016582656860351562, 0.01655193519592285, 0.016547840118408205, 0.016590848922729492, 0.01660006332397461, 0.016673791885375978, 0.01665126419067383, 0.016548864364624022, 0.016526336669921874, 0.016540672302246092, 0.016578559875488282, 0.016707584381103514, 0.016541696548461913, 0.01659903907775879, 0.016635904312133788, 0.016540672302246092, 0.01664204788208008, 0.016648191452026367, 0.0166430721282959, 0.01658060836791992, 0.01659187126159668, 0.016561151504516602, 0.016546815872192384, 0.016673791885375978, 0.01660211181640625, 0.016761856079101564, 0.01660313606262207, 0.01661337661743164, 0.01665843200683594, 0.016727039337158203, 0.016615423202514648, 0.016548864364624022, 0.01660927963256836, 0.01663692855834961, 0.016594944000244142, 0.016563199996948243, 0.016691200256347655, 0.016559104919433593, 0.016562175750732423, 0.01664204788208008, 0.016623615264892578, 0.016647167205810547, 0.034320384979248046, 0.01662668800354004, 0.01662873649597168, 0.016719871520996094, 0.01658060836791992, 0.01659391975402832, 0.016541696548461913, 0.016769023895263673, 0.01661952018737793, 0.016732160568237304, 0.016673791885375978, 0.016645120620727538, 0.016639999389648438, 0.01678335952758789, 0.016663551330566406, 0.016677888870239257, 0.01665843200683594, 0.016671743392944336, 0.016571392059326173, 0.0166297607421875, 0.0165928955078125, 0.0165980167388916, 0.016716800689697265, 0.016712703704833985, 0.01656934356689453, 0.0165928955078125, 0.016582656860351562, 0.0165928955078125, 0.016675840377807616, 0.016615423202514648, 0.01664102363586426, 0.016548864364624022, 0.016684032440185546, 0.016732160568237304, 0.01683865547180176, 0.016715776443481444, 0.016680959701538087, 0.01660825538635254, 0.016655359268188476, 0.016570367813110352, 0.016696319580078126, 0.01660620880126953, 0.01664204788208008, 0.01656524848937988, 0.01658060836791992, 0.016925695419311524, 0.016958463668823243, 0.016635904312133788, 0.0166246395111084, 0.01660723114013672, 0.016542720794677734, 0.015812607765197755, 0.016108543395996093, 0.016539648056030275, 0.016955392837524414, 0.01660108757019043, 0.016710655212402344, 0.01660518455505371, 0.016663551330566406, 0.01662566375732422, 0.01680998420715332, 0.01680384063720703, 0.01657753562927246, 0.034348033905029295, 0.016601119995117188, 0.01664508819580078, 0.01661440086364746, 0.016578559875488282, 0.016726015090942382, 0.01659187126159668, 0.016656383514404297, 0.016692224502563476, 0.016729087829589845, 0.01663795280456543, 0.016776191711425782, 0.01664204788208008, 0.016749631881713866, 0.016670656204223634, 0.0166430721282959, 0.016963584899902344, 0.016668672561645507, 0.01663488006591797, 0.01660825538635254, 0.01659903907775879, 0.01660006332397461, 0.01660620880126953, 0.016537599563598633, 0.01657548713684082, 0.01660313606262207, 0.016751615524291993, 0.016710655212402344, 0.01659903907775879, 0.016712703704833985, 0.01664204788208008, 0.01659391975402832, 0.016664575576782227, 0.01669327926635742, 0.016682975769042967, 0.016630783081054687, 0.01677724838256836, 0.01666761589050293, 0.01660108757019043, 0.0165928955078125, 0.0166297607421875, 0.01656729507446289, 0.01660723114013672, 0.016558080673217773, 0.01661644744873047, 0.016696319580078126, 0.01658367919921875, 0.01661952018737793, 0.01663283157348633, 0.016555007934570314, 0.016644096374511717, 0.01663488006591797, 0.01663385581970215, 0.01677926445007324, 0.016644096374511717, 0.016756736755371093, 0.0166297607421875, 0.016649215698242188, 0.01656729507446289, 0.016653312683105468, 0.01655705642700195, 0.01661030387878418, 0.016627712249755858, 0.03435212707519531, 0.01662156867980957, 0.01663795280456543, 0.016692224502563476, 0.016611328125, 0.01661235237121582, 0.0166246395111084, 0.01662054443359375, 0.016759807586669923, 0.01684480094909668, 0.01637887954711914, 0.016338943481445312, 0.016481279373168945, 0.016728063583374024, 0.01662668800354004, 0.016659456253051756, 0.016670719146728515, 0.016752639770507814, 0.01661440086364746, 0.01664204788208008, 0.016842752456665038, 0.01684377670288086, 0.016652288436889647, 0.016669696807861328, 0.016548864364624022, 0.016702463150024414, 0.016708608627319335, 0.01660518455505371, 0.016676864624023437, 0.01662054443359375, 0.016660480499267577, 0.016559104919433593, 0.016678911209106445, 0.016579584121704103, 0.016738304138183592, 0.016615423202514648, 0.016837631225585938, 0.016656383514404297, 0.016630783081054687, 0.01664102363586426, 0.01663795280456543, 0.016648191452026367, 0.01661030387878418, 0.0166246395111084, 0.01662668800354004, 0.016688127517700196, 0.01663795280456543, 0.016702463150024414, 0.01684889602661133, 0.016638975143432617, 0.016704511642456055, 0.01662259292602539, 0.016684032440185546, 0.01656012725830078, 0.01660825538635254, 0.01663692855834961, 0.01660620880126953, 0.016578559875488282, 0.01658163261413574, 0.01662054443359375, 0.01682431983947754, 0.01662054443359375, 0.016586751937866212, 0.034361343383789066, 0.01664614486694336, 0.016710655212402344, 0.016623615264892578, 0.016571392059326173, 0.016660480499267577, 0.0166297607421875, 0.016574464797973632, 0.01662566375732422, 0.01658982467651367, 0.01659596824645996, 0.01680179214477539, 0.016695295333862305, 0.016713727951049806, 0.016659456253051756, 0.01664614486694336, 0.016721920013427736, 0.016632863998413086, 0.0166409912109375, 0.016559104919433593, 0.01660620880126953, 0.016704511642456055, 0.01681612777709961, 0.016563199996948243, 0.016623615264892578, 0.0166297607421875, 0.016761856079101564, 0.016652288436889647, 0.01665126419067383, 0.01660211181640625, 0.01662873649597168, 0.016561151504516602, 0.016660480499267577, 0.016574464797973632, 0.016579584121704103, 0.01662156867980957, 0.01659596824645996, 0.01676697540283203, 0.01654374313354492, 0.016647167205810547, 0.016544767379760742, 0.01659596824645996, 0.01655705642700195, 0.01661235237121582, 0.01660927963256836, 0.016630783081054687, 0.0166430721282959, 0.016574464797973632, 0.016720895767211915, 0.016700416564941405, 0.016671743392944336, 0.016638975143432617, 0.01660620880126953, 0.01660211181640625, 0.01657344055175781, 0.016545791625976563, 0.01658880043029785, 0.016717824935913086, 0.016690176010131837, 0.016655359268188476, 0.016661504745483398, 0.0166430721282959, 0.01665433692932129]",tokens/s,59.50162367332891,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,3689.066496,5211.947008,0.0,4565.499904,4315.13088,s,10,2.8045252990722656,0.2804525299072266,0.00062484693400047,0.28048956298828126,0.2810666198730469,0.2813779449462891,0.2816270050048828,"[0.2808427429199219, 0.27973049926757815, 0.2806927490234375, 0.28070648193359377, 0.2809974365234375, 0.2796501770019531, 0.28168927001953126, 0.2797050476074219, 0.28022451782226565, 0.280286376953125]",tokens/s,912.8104498992559,kWh,3.3089337525544346e-06,1.8130049195709337e-06,1.64667107042228e-05,2.1588649376348166e-05,tokens/kWh,11858083.177749202,MB,3689.066496,5211.947008,0.0,4565.499904,4383.673344,s,10,151.17197265625,15.117197265625,0.001595355493039014,15.11665625,15.1191974609375,15.12004208984375,15.120717792968751,"[15.12088671875, 15.1161455078125, 15.117458984375, 15.116115234375, 15.11771484375, 15.1162314453125, 15.1165712890625, 15.1167412109375, 15.11509765625, 15.119009765625]",tokens/s,4.167439168320951,kWh,0.0001783895491781058,9.777224420422193e-05,0.0008703265265075908,0.0011464883198899188,tokens/kWh,54950.40717558205,,s,629,153.45473150634766,0.24396618681454318,0.03336199246270589,0.23990885925292968,0.24012165222167967,0.24032051086425782,0.5203096752929688,"[0.2403778533935547, 0.2398525695800781, 0.23982998657226562, 0.2397706298828125, 0.2398392333984375, 0.23967234802246093, 0.23975010681152345, 0.2397655029296875, 0.24011878967285155, 0.24025497436523438, 0.23989144897460937, 0.23997952270507814, 0.2401607666015625, 0.24013926696777343, 0.2398750762939453, 0.23990885925292968, 0.2398269500732422, 0.2398586883544922, 0.23992626953125, 0.23986483764648436, 0.23992735290527345, 0.23978182983398438, 0.23988531494140625, 0.24013722229003906, 0.24008090209960936, 0.2400860137939453, 0.23997235107421874, 0.23982182312011718, 0.23994367980957032, 0.23997235107421874, 0.23994265747070312, 0.23988742065429688, 0.24012896728515626, 0.23988223266601563, 0.23986892700195311, 0.23992422485351564, 0.24000306701660157, 0.2398924865722656, 0.239994873046875, 0.2410373077392578, 0.2400675811767578, 0.2399129638671875, 0.23988429260253907, 0.2399201354980469, 0.24005427551269531, 0.2403461151123047, 0.24001434326171875, 0.2399201354980469, 0.24022221374511718, 0.23999282836914063, 0.24007986450195312, 0.24014335632324219, 0.24005938720703124, 0.24000408935546874, 0.23999693298339844, 0.24004095458984376, 0.2400010223388672, 0.2400849914550781, 0.24011570739746094, 0.23999795532226562, 0.23993344116210938, 0.23986175537109375, 0.5203558349609375, 0.23969587707519532, 0.2398719940185547, 0.23987814331054688, 0.23991194152832032, 0.2397716827392578, 0.23996514892578125, 0.24009933471679687, 0.239783935546875, 0.239931396484375, 0.23986892700195311, 0.2399078369140625, 0.2397890625, 0.23978291320800782, 0.23977984619140624, 0.23988018798828126, 0.2398935089111328, 0.23988429260253907, 0.24012083435058593, 0.23986688232421874, 0.23980953979492187, 0.23979212951660156, 0.2397962188720703, 0.23999795532226562, 0.23985664367675782, 0.239963134765625, 0.23987306213378906, 0.23991395568847657, 0.24004095458984376, 0.24035328674316406, 0.239815673828125, 0.23985049438476563, 0.2398709716796875, 0.23984434509277344, 0.240005126953125, 0.23993241882324218, 0.24054681396484376, 0.23979519653320314, 0.23975117492675782, 0.2398013458251953, 0.2398709716796875, 0.23993959045410157, 0.2398719940185547, 0.23984332275390624, 0.23986073303222658, 0.24000921630859376, 0.24060928344726562, 0.24000717163085938, 0.23999591064453124, 0.24008294677734374, 0.2398709716796875, 0.23996517944335938, 0.23989561462402345, 0.2399784393310547, 0.23985356140136718, 0.2398924865722656, 0.24010342407226562, 0.23997541809082032, 0.23987814331054688, 0.23996723937988282, 0.240005126953125, 0.24007168579101562, 0.23989453125, 0.5209968872070313, 0.23973683166503906, 0.23968972778320313, 0.2400696563720703, 0.2396651153564453, 0.23987916564941406, 0.23976039123535156, 0.23983616638183594, 0.23984127807617187, 0.23996109008789063, 0.23985049438476563, 0.24, 0.23975833129882812, 0.23992320251464844, 0.2399805450439453, 0.2399303741455078, 0.23978086853027344, 0.23998873901367188, 0.23990885925292968, 0.24003071594238282, 0.2398750762939453, 0.24009727478027343, 0.2400245819091797, 0.24043827819824218, 0.23983103942871092, 0.24015359497070313, 0.24030924987792968, 0.24007577514648437, 0.23983103942871092, 0.2399498291015625, 0.24010751342773437, 0.239889404296875, 0.23988636779785155, 0.23984637451171875, 0.2397655029296875, 0.24004095458984376, 0.2399139862060547, 0.23993446350097655, 0.23987814331054688, 0.2399692840576172, 0.2399129638671875, 0.23998976135253905, 0.23990272521972655, 0.2400747528076172, 0.23984538269042968, 0.24001741027832033, 0.23984640502929688, 0.24032051086425782, 0.2399457244873047, 0.24007069396972655, 0.23992726135253906, 0.240005126953125, 0.23995391845703126, 0.23994367980957032, 0.23995802307128905, 0.23991500854492187, 0.24000819396972656, 0.2399805450439453, 0.24000306701660157, 0.24007373046875, 0.2401116180419922, 0.2400511932373047, 0.23992732238769532, 0.5211043701171875, 0.23985052490234374, 0.23974703979492187, 0.2398177947998047, 0.23999276733398436, 0.23998873901367188, 0.23975628662109374, 0.23989453125, 0.23987405395507813, 0.23979008483886718, 0.2395924530029297, 0.23982595825195313, 0.24075465393066406, 0.23995603942871094, 0.23985760498046876, 0.23985971069335937, 0.23988735961914062, 0.2398719940185547, 0.23996517944335938, 0.23984844970703126, 0.2397696075439453, 0.23981158447265624, 0.2398392333984375, 0.23985664367675782, 0.23992831420898436, 0.239994873046875, 0.23993959045410157, 0.2399139862060547, 0.24002560424804686, 0.23980339050292968, 0.23986483764648436, 0.23979315185546876, 0.23978291320800782, 0.23999591064453124, 0.23983616638183594, 0.23977880859375, 0.2398023681640625, 0.23979110717773439, 0.23977267456054688, 0.2399252471923828, 0.23989657592773436, 0.24022425842285156, 0.24046182250976564, 0.2399764404296875, 0.23983821105957032, 0.24021197509765624, 0.23990988159179688, 0.23981158447265624, 0.23983001708984375, 0.24002867126464844, 0.23998565673828126, 0.23988838195800782, 0.24006040954589844, 0.24036863708496095, 0.23994371032714842, 0.24007778930664062, 0.23996723937988282, 0.24000306701660157, 0.24007986450195312, 0.23987712097167968, 0.239994873046875, 0.239963134765625, 0.23997952270507814, 0.5213009643554688, 0.23983001708984375, 0.23982899475097658, 0.23997541809082032, 0.23994061279296874, 0.23986073303222658, 0.239857666015625, 0.23986688232421874, 0.23975935363769532, 0.2398955535888672, 0.23973785400390624, 0.23990476989746093, 0.23977574157714843, 0.2398924865722656, 0.2398760986328125, 0.23988018798828126, 0.23983821105957032, 0.24103424072265625, 0.2399866943359375, 0.2399938507080078, 0.23992320251464844, 0.2398935089111328, 0.2399498291015625, 0.24020684814453125, 0.23982797241210937, 0.23987303161621093, 0.23994265747070312, 0.2401648712158203, 0.2398904266357422, 0.2397777862548828, 0.23993650817871093, 0.2398709716796875, 0.23988018798828126, 0.23994163513183595, 0.23988531494140625, 0.23996109008789063, 0.23981471252441405, 0.23982687377929687, 0.23970611572265624, 0.23990885925292968, 0.23984844970703126, 0.23992428588867187, 0.2398299560546875, 0.23997337341308594, 0.24029286193847657, 0.2405806121826172, 0.2399815673828125, 0.23986790466308594, 0.239972412109375, 0.23992518615722655, 0.23981875610351563, 0.2398802185058594, 0.23986070251464844, 0.23986892700195311, 0.2398760986328125, 0.23995802307128905, 0.24007481384277343, 0.24067578125, 0.23984434509277344, 0.2399129638671875, 0.24007986450195312, 0.2402672576904297, 0.2400849914550781, 0.5201909790039062, 0.23994776916503907, 0.2396968994140625, 0.2398023681640625, 0.23982899475097658, 0.23996517944335938, 0.23963749694824218, 0.23974400329589843, 0.23974911499023438, 0.23987712097167968, 0.23980035400390626, 0.23988429260253907, 0.23991087341308595, 0.23987405395507813, 0.23973683166503906, 0.23989657592773436, 0.23971737670898438, 0.23987814331054688, 0.23985356140136718, 0.2399272918701172, 0.23986277770996095, 0.23987814331054688, 0.2400133056640625, 0.2400235595703125, 0.23996723937988282, 0.24009523010253905, 0.2400133056640625, 0.24005223083496094, 0.23984742736816406, 0.23986277770996095, 0.24005941772460937, 0.23992521667480468, 0.23985459899902345, 0.23984947204589843, 0.24013209533691407, 0.2402590789794922, 0.23980953979492187, 0.23992933654785156, 0.24046694946289063, 0.23999897766113282, 0.23994674682617187, 0.240216064453125, 0.24031744384765624, 0.24002867126464844, 0.23994061279296874, 0.23988429260253907, 0.24000723266601562, 0.23993440246582032, 0.2398197784423828, 0.23996620178222655, 0.2398392333984375, 0.2399385528564453, 0.23983206176757813, 0.23992115783691406, 0.23999591064453124, 0.23996723937988282, 0.23992422485351564, 0.24003890991210938, 0.23988223266601563, 0.24004197692871093, 0.24005427551269531, 0.23986688232421874, 0.23987303161621093, 0.5225390014648438, 0.23981875610351563, 0.23979315185546876, 0.23980953979492187, 0.2397081604003906, 0.23985562133789062, 0.2397112274169922, 0.23981158447265624, 0.23976556396484375, 0.23979922485351562, 0.23977471923828125, 0.23988735961914062, 0.23971328735351563, 0.23974400329589843, 0.2397337646484375, 0.23986688232421874, 0.2403031005859375, 0.24, 0.23979212951660156, 0.2399518737792969, 0.24010035705566407, 0.2398167724609375, 0.23980947875976563, 0.2414950408935547, 0.24010546875, 0.23994061279296874, 0.23981875610351563, 0.23991194152832032, 0.23982182312011718, 0.24000819396972656, 0.24010240173339845, 0.2399550018310547, 0.23980230712890624, 0.23984844970703126, 0.23968255615234374, 0.23978291320800782, 0.23979827880859375, 0.2398924865722656, 0.23988223266601563, 0.2397696075439453, 0.23988327026367187, 0.24003379821777343, 0.2399252471923828, 0.23987303161621093, 0.23980339050292968, 0.24008908081054686, 0.23982797241210937, 0.2397532196044922, 0.23991909790039062, 0.24007986450195312, 0.23986277770996095, 0.2398719940185547, 0.23978291320800782, 0.24007270812988282, 0.24017100524902343, 0.23988735961914062, 0.23990885925292968, 0.2400911407470703, 0.24032768249511718, 0.2399805450439453, 0.24003173828125, 0.24032051086425782, 0.24008090209960936, 0.5199175415039062, 0.23988735961914062, 0.23981056213378907, 0.23968768310546876, 0.23967027282714845, 0.23986892700195311, 0.23981056213378907, 0.23983512878417967, 0.2398085174560547, 0.23995494079589844, 0.23984538269042968, 0.23982386779785156, 0.23981875610351563, 0.23989759826660156, 0.24007781982421875, 0.24029592895507812, 0.2409441223144531, 0.23982080078125, 0.23992115783691406, 0.2403461151123047, 0.24001126098632813, 0.23998976135253905, 0.23987405395507813, 0.23995391845703126, 0.23986892700195311, 0.23998873901367188, 0.23998873901367188, 0.23987916564941406, 0.23979315185546876, 0.23976345825195314, 0.23986076354980468, 0.24014128112792968, 0.2398392333984375, 0.23983103942871092, 0.23984844970703126, 0.2400552978515625, 0.2399006652832031, 0.23979827880859375, 0.23985049438476563, 0.24005427551269531, 0.23995391845703126, 0.23992115783691406, 0.23981875610351563, 0.24004710388183595, 0.23992115783691406, 0.23981260681152344, 0.23980339050292968, 0.23988429260253907, 0.23983616638183594, 0.24006962585449218, 0.2398924865722656, 0.2398719940185547, 0.23988327026367187, 0.2399550018310547, 0.24008082580566406, 0.24018328857421875, 0.24004812622070312, 0.24004812622070312, 0.239963134765625, 0.23990579223632813, 0.2403225555419922, 0.2400430145263672, 0.23988223266601563, 0.5207296142578125, 0.23983206176757813, 0.23972761535644532, 0.2398638153076172, 0.23969692993164063, 0.23994569396972656, 0.23972557067871095, 0.23981875610351563, 0.2398515167236328, 0.23985049438476563, 0.23972659301757812, 0.23985459899902345, 0.24005938720703124, 0.24018328857421875, 0.23991091918945312, 0.23990885925292968, 0.24021197509765624, 0.24111308288574218, 0.2398085174560547, 0.23978291320800782, 0.23995391845703126, 0.2401853485107422, 0.23982591247558593, 0.23990374755859376, 0.23981465148925782, 0.23993241882324218, 0.23984127807617187, 0.2397665252685547, 0.2398341064453125, 0.23991603088378907, 0.23988531494140625, 0.23983718872070312, 0.2398638153076172, 0.24002047729492187, 0.23981260681152344, 0.23993446350097655, 0.23970098876953125, 0.23996109008789063, 0.2399518737792969, 0.24001536560058595, 0.2397644805908203, 0.23991500854492187, 0.23984538269042968, 0.23986585998535156, 0.2396743621826172, 0.23986994934082032, 0.23976856994628906, 0.23987303161621093, 0.2397706298828125, 0.23988018798828126, 0.23983103942871092, 0.23991705322265625, 0.23990272521972655, 0.23994163513183595, 0.24011468505859376, 0.23999078369140625, 0.24000408935546874, 0.2400511932373047, 0.24006349182128905, 0.23988432312011718, 0.23994979858398438, 0.239889404296875, 0.23994061279296874, 0.5206292724609375, 0.2397777862548828, 0.2398085174560547, 0.23981773376464843, 0.23988429260253907, 0.23986178588867188, 0.23979417419433594, 0.2399385223388672, 0.23975833129882812, 0.2398525390625, 0.2398935089111328, 0.24011776733398438, 0.2398935089111328, 0.2399488067626953, 0.24001536560058595, 0.24023347473144532, 0.24003788757324218, 0.23994265747070312, 0.23981056213378907, 0.24012492370605468, 0.24007168579101562, 0.23998258972167968, 0.23982489013671876, 0.23999078369140625, 0.23994776916503907, 0.24004812622070312, 0.241833984375, 0.23985049438476563, 0.23989964294433594, 0.23984947204589843, 0.23972761535644532, 0.24009523010253905, 0.24031642150878907, 0.2398760986328125, 0.23976345825195314, 0.23986277770996095, 0.23985459899902345, 0.23983001708984375, 0.2398197784423828, 0.23988838195800782, 0.2399938507080078, 0.24, 0.23984742736816406, 0.240110595703125, 0.24010855102539064, 0.24039321899414062, 0.2399385528564453, 0.23988633728027345, 0.2398709716796875, 0.23997747802734376, 0.23985356140136718, 0.23993650817871093, 0.2399447021484375, 0.24004205322265626, 0.24003884887695312, 0.23985562133789062, 0.24004608154296875, 0.24009420776367188, 0.23995802307128905, 0.24000306701660157, 0.24003993225097656, 0.24007577514648437, 0.2399006652832031]",tokens/s,4.098928679654178,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10516,7 +10516,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1451.905024,2418.540544,0.0,1772.09344,1621.541376,s,10,0.7316847381591797,0.07316847381591798,0.0003457241418479124,0.07326449584960938,0.07357357177734375,0.07357862548828126,0.07358266845703125,"[0.07346591949462891, 0.0728661117553711, 0.07270166778564453, 0.07324726104736329, 0.07263625335693359, 0.07288240051269532, 0.07358367919921875, 0.07328173065185548, 0.073447265625, 0.07357244873046875]",tokens/s,3498.7746313263483,kWh,8.589118671110095e-07,4.706412848463514e-07,3.7715146919278033e-06,5.101067843885164e-06,tokens/kWh,50185570.51870551,MB,1451.905024,2418.540544,0.0,1772.09344,1723.273216,s,10,38.208432373046875,3.8208432373046874,0.0009523784305681953,3.8212119140625003,3.821659497070313,3.821850988769531,3.822004182128906,"[3.8209521484375, 3.821502685546875, 3.821594482421875, 3.8205712890625, 3.82204248046875, 3.819373291015625, 3.8214716796875, 3.819123291015625, 3.82018408203125, 3.821616943359375]",tokens/s,16.488506878508232,kWh,4.511216866125423e-05,2.4723927950708465e-05,0.0001974284518234696,0.0002672645484354323,tokens/kWh,235721.49904954564,,s,629,38.80548554611203,0.06169393568539279,0.008717092816584612,0.06063616180419922,0.060711936950683595,0.060762318420410154,0.1337114239501953,"[0.060548095703125, 0.06056447982788086, 0.06056243133544922, 0.06060441589355469, 0.06070169448852539, 0.06058393478393555, 0.06068428802490235, 0.06065151977539063, 0.060698623657226565, 0.060622848510742185, 0.06061056137084961, 0.06058598327636719, 0.06065971374511719, 0.060631038665771485, 0.06055321502685547, 0.060627967834472656, 0.06062080001831055, 0.06059212875366211, 0.06057062530517578, 0.06062694549560547, 0.06059212875366211, 0.06070169448852539, 0.060641281127929686, 0.06050099182128906, 0.06061056137084961, 0.060644351959228515, 0.06061363220214844, 0.06065459060668945, 0.060680191040039064, 0.06059929656982422, 0.0606033935546875, 0.06075494384765625, 0.06057984161376953, 0.060627967834472656, 0.06067609786987305, 0.06060851287841797, 0.060625919342041014, 0.06067507171630859, 0.06063820648193359, 0.06066790390014649, 0.0607375373840332, 0.06063718414306641, 0.06057164764404297, 0.06067814254760742, 0.0606033935546875, 0.06069452667236328, 0.06069247817993164, 0.06072934341430664, 0.060672000885009764, 0.06068121719360352, 0.06060441589355469, 0.060665855407714846, 0.06066995239257812, 0.06066995239257812, 0.0605849609375, 0.06070272064208984, 0.060647422790527344, 0.06064025497436523, 0.06067302322387695, 0.060631038665771485, 0.060611583709716796, 0.06069145584106445, 0.1337518005371094, 0.060679168701171876, 0.060565502166748046, 0.060598270416259765, 0.06051942443847656, 0.06059212875366211, 0.06055731201171875, 0.06060748672485351, 0.06057062530517578, 0.06058700942993164, 0.060539905548095706, 0.06059212875366211, 0.060524543762207034, 0.06062694549560547, 0.0605296630859375, 0.06064025497436523, 0.060556289672851565, 0.0607088623046875, 0.06061772918701172, 0.0605849609375, 0.060622848510742185, 0.06062387084960937, 0.060554241180419924, 0.06061875152587891, 0.06054912185668945, 0.060633087158203126, 0.06063718414306641, 0.06061670303344727, 0.06066175842285156, 0.06064025497436523, 0.06060236740112305, 0.06065663909912109, 0.06065151977539063, 0.06146867370605469, 0.060672000885009764, 0.06063513565063477, 0.06062899017333984, 0.06063513565063477, 0.06063513565063477, 0.06066175842285156, 0.06074060821533203, 0.0606453742980957, 0.06065663909912109, 0.0606484489440918, 0.06070374298095703, 0.0606453742980957, 0.06063411331176758, 0.06067814254760742, 0.060641281127929686, 0.06077132797241211, 0.060819454193115234, 0.06069964981079102, 0.06061772918701172, 0.0607375373840332, 0.06071091079711914, 0.06056857681274414, 0.06065868759155273, 0.060695552825927736, 0.06065151977539063, 0.06061363220214844, 0.06063616180419922, 0.06097817611694336, 0.060883968353271485, 0.13362074279785155, 0.06075699234008789, 0.060611583709716796, 0.06068838500976562, 0.06058086395263672, 0.06061977767944336, 0.06060031890869141, 0.06062182235717774, 0.06057881546020508, 0.0607006721496582, 0.06058598327636719, 0.060609535217285154, 0.06054604721069336, 0.060641281127929686, 0.06057779312133789, 0.060614654541015625, 0.060682239532470705, 0.06065049743652344, 0.060854270935058595, 0.060627967834472656, 0.060611583709716796, 0.06075289535522461, 0.060668926239013675, 0.06060236740112305, 0.060581886291503906, 0.060715007781982425, 0.06063616180419922, 0.06065356826782226, 0.06065459060668945, 0.060662784576416016, 0.060657665252685546, 0.060644351959228515, 0.06066483306884766, 0.06063513565063477, 0.060652542114257815, 0.060660736083984375, 0.060652542114257815, 0.060655616760253904, 0.06063206481933594, 0.06065151977539063, 0.06070272064208984, 0.060677120208740234, 0.060706817626953125, 0.06065151977539063, 0.06060748672485351, 0.06067609786987305, 0.060674049377441405, 0.06074982452392578, 0.060652542114257815, 0.060711936950683595, 0.06064640045166016, 0.06088806533813477, 0.060649471282958986, 0.06063820648193359, 0.06074060821533203, 0.06055014419555664, 0.06067097473144531, 0.06065868759155273, 0.060693504333496094, 0.060663806915283204, 0.06061977767944336, 0.060641281127929686, 0.06065151977539063, 0.13421568298339845, 0.061047809600830075, 0.06058086395263672, 0.06061363220214844, 0.06059417724609375, 0.060660736083984375, 0.06055219268798828, 0.06056959915161133, 0.06055833435058594, 0.06059110260009765, 0.060598270416259765, 0.06062080001831055, 0.0606033935546875, 0.060644351959228515, 0.06060134506225586, 0.06059212875366211, 0.060576766967773435, 0.06063616180419922, 0.06062387084960937, 0.060624897003173826, 0.060870655059814455, 0.06065971374511719, 0.06057369613647461, 0.06063411331176758, 0.060572673797607425, 0.06064640045166016, 0.06059110260009765, 0.06067814254760742, 0.06068428802490235, 0.060662784576416016, 0.06063411331176758, 0.06064640045166016, 0.060652542114257815, 0.06065663909912109, 0.060611583709716796, 0.060641281127929686, 0.06062182235717774, 0.060668926239013675, 0.06061977767944336, 0.060657665252685546, 0.06068838500976562, 0.060642303466796874, 0.060652542114257815, 0.060644351959228515, 0.060609535217285154, 0.060624897003173826, 0.06063206481933594, 0.06063616180419922, 0.06063820648193359, 0.06064332962036133, 0.06060748672485351, 0.0606300163269043, 0.06061363220214844, 0.06065459060668945, 0.06070579147338867, 0.06060441589355469, 0.06068633651733398, 0.06062694549560547, 0.06064025497436523, 0.06063513565063477, 0.06063616180419922, 0.06065356826782226, 0.060655616760253904, 0.13350604248046874, 0.060693504333496094, 0.06060851287841797, 0.06065663909912109, 0.06060031890869141, 0.06069247817993164, 0.06064332962036133, 0.06060441589355469, 0.06056345748901367, 0.06063616180419922, 0.060576766967773435, 0.06068121719360352, 0.06066483306884766, 0.06058598327636719, 0.06065868759155273, 0.060622848510742185, 0.060655616760253904, 0.060633087158203126, 0.06061056137084961, 0.06066995239257812, 0.06062694549560547, 0.06065356826782226, 0.06086963272094727, 0.060609535217285154, 0.060611583709716796, 0.060919807434082034, 0.06071398544311524, 0.060652542114257815, 0.060660736083984375, 0.0606484489440918, 0.060668926239013675, 0.06059929656982422, 0.060655616760253904, 0.0608092155456543, 0.06080204772949219, 0.06074879837036133, 0.06071398544311524, 0.060622848510742185, 0.060660736083984375, 0.06069452667236328, 0.06069964981079102, 0.06062080001831055, 0.060665855407714846, 0.060649471282958986, 0.06063718414306641, 0.06069145584106445, 0.06063616180419922, 0.0606484489440918, 0.06067302322387695, 0.06059110260009765, 0.06071398544311524, 0.06069145584106445, 0.06063616180419922, 0.060655616760253904, 0.0607303695678711, 0.060576766967773435, 0.06069657516479492, 0.060682239532470705, 0.060644351959228515, 0.06062387084960937, 0.06065151977539063, 0.0606300163269043, 0.06076006317138672, 0.13374668884277344, 0.060639232635498044, 0.06056959915161133, 0.06061568069458008, 0.06057164764404297, 0.060706817626953125, 0.06055731201171875, 0.060556289672851565, 0.06061670303344727, 0.06055321502685547, 0.06053376007080078, 0.06062899017333984, 0.060548095703125, 0.06059929656982422, 0.060622848510742185, 0.06057369613647461, 0.06059110260009765, 0.06065868759155273, 0.06058291244506836, 0.06058905410766602, 0.06061056137084961, 0.06056755065917969, 0.06058291244506836, 0.06062694549560547, 0.06063616180419922, 0.06063513565063477, 0.06063718414306641, 0.06055526351928711, 0.060622848510742185, 0.060649471282958986, 0.06062899017333984, 0.060614654541015625, 0.06070476913452148, 0.060560382843017575, 0.06061875152587891, 0.0606453742980957, 0.06059724807739258, 0.06061670303344727, 0.060728321075439455, 0.060606464385986325, 0.060657665252685546, 0.06066175842285156, 0.06058598327636719, 0.06061260986328125, 0.06068940734863281, 0.06057984161376953, 0.06063206481933594, 0.060663806915283204, 0.06060543823242188, 0.06058905410766602, 0.060711936950683595, 0.06060031890869141, 0.06067097473144531, 0.06065151977539063, 0.060677120208740234, 0.06054604721069336, 0.06071091079711914, 0.060649471282958986, 0.06063718414306641, 0.06063820648193359, 0.06069964981079102, 0.06061875152587891, 0.06069760131835938, 0.1345075225830078, 0.06063411331176758, 0.06067097473144531, 0.060625919342041014, 0.060560382843017575, 0.06066790390014649, 0.060548095703125, 0.06058291244506836, 0.060598270416259765, 0.060606464385986325, 0.06056140899658203, 0.0605665283203125, 0.06058803176879883, 0.06055321502685547, 0.06061260986328125, 0.060576766967773435, 0.06076313781738281, 0.06071295928955078, 0.060576766967773435, 0.060572673797607425, 0.060627967834472656, 0.060576766967773435, 0.06057164764404297, 0.06062080001831055, 0.06053887939453125, 0.06092595291137695, 0.0607006721496582, 0.06059110260009765, 0.06064332962036133, 0.06065049743652344, 0.06057779312133789, 0.060633087158203126, 0.06071295928955078, 0.060647422790527344, 0.060639232635498044, 0.060668926239013675, 0.06061363220214844, 0.06061056137084961, 0.060639232635498044, 0.06059212875366211, 0.060693504333496094, 0.06067097473144531, 0.06083379364013672, 0.06059417724609375, 0.0606668815612793, 0.06058393478393555, 0.06063820648193359, 0.06070579147338867, 0.06062694549560547, 0.06063206481933594, 0.06072115325927734, 0.06065663909912109, 0.06077951812744141, 0.06069452667236328, 0.06077030563354492, 0.06057779312133789, 0.06068326568603516, 0.06058598327636719, 0.06092390441894531, 0.060906494140625, 0.060660736083984375, 0.06083686447143555, 0.060761089324951174, 0.13417062377929687, 0.060611583709716796, 0.060647422790527344, 0.06059622573852539, 0.060543998718261716, 0.06064025497436523, 0.06065868759155273, 0.06062899017333984, 0.060560382843017575, 0.06058598327636719, 0.060523521423339846, 0.06059417724609375, 0.06057369613647461, 0.06057984161376953, 0.06057164764404297, 0.06060031890869141, 0.06056857681274414, 0.060624897003173826, 0.0606453742980957, 0.06063513565063477, 0.06058700942993164, 0.060609535217285154, 0.06055833435058594, 0.060572673797607425, 0.06055116653442383, 0.060662784576416016, 0.0606453742980957, 0.06065971374511719, 0.0606300163269043, 0.06062387084960937, 0.06058700942993164, 0.060649471282958986, 0.06067097473144531, 0.060679168701171876, 0.060624897003173826, 0.060611583709716796, 0.060622848510742185, 0.060665855407714846, 0.060625919342041014, 0.06063411331176758, 0.06066790390014649, 0.060622848510742185, 0.0606300163269043, 0.060641281127929686, 0.06063820648193359, 0.06062080001831055, 0.06061670303344727, 0.06062080001831055, 0.06059929656982422, 0.0606300163269043, 0.06063206481933594, 0.06067609786987305, 0.06058803176879883, 0.06060543823242188, 0.0606484489440918, 0.060543998718261716, 0.060644351959228515, 0.06063411331176758, 0.060581886291503906, 0.06059212875366211, 0.060672000885009764, 0.06064025497436523, 0.060680191040039064, 0.1343651885986328, 0.06059212875366211, 0.06057164764404297, 0.06062080001831055, 0.06056755065917969, 0.060633087158203126, 0.06053683090209961, 0.06060134506225586, 0.06061670303344727, 0.060595199584960936, 0.06065049743652344, 0.060642303466796874, 0.06053171157836914, 0.0606033935546875, 0.06058700942993164, 0.06060441589355469, 0.060576766967773435, 0.060662784576416016, 0.06059110260009765, 0.06068940734863281, 0.06059929656982422, 0.060624897003173826, 0.06051942443847656, 0.060624897003173826, 0.06059110260009765, 0.060644351959228515, 0.06058393478393555, 0.06063820648193359, 0.06062899017333984, 0.06070169448852539, 0.060633087158203126, 0.060674049377441405, 0.060647422790527344, 0.06063411331176758, 0.06060236740112305, 0.060633087158203126, 0.060649471282958986, 0.06068428802490235, 0.060614654541015625, 0.060641281127929686, 0.06069657516479492, 0.06066175842285156, 0.06067302322387695, 0.060644351959228515, 0.06066995239257812, 0.06064025497436523, 0.060655616760253904, 0.06065049743652344, 0.060660736083984375, 0.06073241424560547, 0.060633087158203126, 0.060674049377441405, 0.06063411331176758, 0.060680191040039064, 0.06072217559814453, 0.060614654541015625, 0.060624897003173826, 0.060649471282958986, 0.06065151977539063, 0.06063616180419922, 0.0607088623046875, 0.06064025497436523, 0.06068735885620117, 0.13449215698242187, 0.060625919342041014, 0.06061260986328125, 0.060611583709716796, 0.060521472930908204, 0.06062694549560547, 0.06057881546020508, 0.06061670303344727, 0.0606033935546875, 0.06058598327636719, 0.06053887939453125, 0.06060031890869141, 0.06055526351928711, 0.06065151977539063, 0.06101196670532227, 0.06065459060668945, 0.0605849609375, 0.06064025497436523, 0.060624897003173826, 0.060652542114257815, 0.06062899017333984, 0.060672000885009764, 0.06060441589355469, 0.060598270416259765, 0.06057369613647461, 0.06062899017333984, 0.06068428802490235, 0.060649471282958986, 0.06072012710571289, 0.06067609786987305, 0.06069452667236328, 0.060657665252685546, 0.06086656188964844, 0.06068633651733398, 0.06073241424560547, 0.06067507171630859, 0.060631038665771485, 0.060680191040039064, 0.06071091079711914, 0.060662784576416016, 0.060674049377441405, 0.06072217559814453, 0.06072012710571289, 0.06066995239257812, 0.06067609786987305, 0.060609535217285154, 0.06061363220214844, 0.060674049377441405, 0.06062182235717774, 0.060662784576416016, 0.0607191047668457, 0.060662784576416016, 0.06069145584106445, 0.060709888458251954, 0.06071295928955078, 0.060631038665771485, 0.06070169448852539, 0.060624897003173826, 0.060647422790527344, 0.0607191047668457, 0.060633087158203126, 0.060622848510742185, 0.060674049377441405]",tokens/s,16.209048570016403,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10558,7 +10558,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10833,7 +10833,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2107.244544,9028.763648,0.0,8382.316544,7939.414528,s,10,6.2175233154296885,0.6217523315429688,0.0006060610905763154,0.6219114074707031,0.622343017578125,0.6224655151367188,0.6225635131835937,"[0.6211746826171874, 0.6221463012695313, 0.62206103515625, 0.6223157958984376, 0.6221713256835938, 0.6207510375976563, 0.6225880126953125, 0.620806396484375, 0.6217617797851562, 0.6217469482421875]",tokens/s,411.7395094678596,kWh,7.3365436659918894e-06,4.020108750067747e-06,3.64477579163554e-05,4.780441033241504e-05,tokens/kWh,5355154.434912305,MB,2107.244544,9028.763648,0.0,8382.316544,8233.863168,s,10,338.38613671875004,33.838613671874995,0.0013682233241316137,33.838173828125,33.839473828125,33.8409205078125,33.8420778515625,"[33.8423671875, 33.83902734375, 33.83762890625, 33.8383828125, 33.83859765625, 33.8376875, 33.837484375, 33.83784375, 33.83796484375, 33.83915234375]",tokens/s,1.8617783994018209,kWh,0.00039950132826964066,0.00021896144419255187,0.0019980115918720536,0.002616474364334246,tokens/kWh,24078.20266033073,,s,629,343.44471649169924,0.5460170373476936,0.07393127168051418,0.53711669921875,0.5374224243164063,0.5375160522460938,1.1593396875,"[0.5371217651367187, 0.5371709594726563, 0.5366651000976562, 0.537037841796875, 0.536721435546875, 0.5372631225585938, 0.5367900390625, 0.5371453247070312, 0.5368391723632813, 0.5368822021484375, 0.5365924072265625, 0.53690673828125, 0.5367193603515625, 0.5368309936523438, 0.53667431640625, 0.536953857421875, 0.53671630859375, 0.537275390625, 0.53671728515625, 0.5374269409179687, 0.5368370971679688, 0.5372139282226562, 0.5369415893554688, 0.5373849487304687, 0.5370521850585938, 0.5373521728515624, 0.53697021484375, 0.5371996459960937, 0.53709619140625, 0.5371986083984375, 0.5375160522460938, 0.53732861328125, 0.5369405517578125, 0.5373890380859375, 0.5368350830078125, 0.539198486328125, 0.5370634155273437, 0.5373604125976562, 0.5372958984375, 0.5371893920898437, 0.5370286254882812, 0.5372313842773437, 0.5371893920898437, 0.53741259765625, 0.5372507934570312, 0.5379686279296875, 0.5368719482421875, 0.5375682373046875, 0.5370552368164062, 0.5373081665039062, 0.537153564453125, 0.537343994140625, 0.5372528686523438, 0.5376102294921875, 0.53725390625, 0.537523193359375, 0.5369896850585938, 0.537259033203125, 0.5368565673828125, 0.5373132934570313, 0.5371412353515626, 0.53785498046875, 1.159841796875, 0.537038818359375, 0.537017333984375, 0.53684326171875, 0.5368760375976562, 0.537006103515625, 0.5370521850585938, 0.5366651000976562, 0.5371259155273438, 0.53661181640625, 0.5370153198242188, 0.53686474609375, 0.5367869262695313, 0.5367296142578125, 0.5369548950195312, 0.5367306518554688, 0.5370265502929688, 0.537007080078125, 0.5371494140625, 0.5369845581054687, 0.5373634643554688, 0.5367992553710937, 0.5371412353515626, 0.5370480346679688, 0.5374586791992187, 0.5371259155273438, 0.5373778076171875, 0.5375211791992187, 0.5374515380859375, 0.5371729736328125, 0.5374013671875, 0.5369978637695313, 0.53728564453125, 0.5368043823242188, 0.5373153076171875, 0.5369825439453125, 0.537270263671875, 0.5371094970703125, 0.53711669921875, 0.5369528198242187, 0.5371340942382813, 0.5368822021484375, 0.5371781005859375, 0.5371904296875, 0.5374402465820313, 0.5372866821289063, 0.5374013671875, 0.536911865234375, 0.5372446899414063, 0.53709619140625, 0.5371760864257813, 0.5372518310546875, 0.5371791381835938, 0.5371320190429687, 0.5374412841796875, 0.537218017578125, 0.5374197998046875, 0.5370767211914063, 0.53718017578125, 0.5369876708984375, 0.5377146606445312, 0.5373378295898438, 0.5373132934570313, 1.15976708984375, 0.5368463134765625, 0.537006103515625, 0.5365985107421875, 0.5367725830078125, 0.5366599731445313, 0.5368688354492187, 0.5365924072265625, 0.5370224609375, 0.53661083984375, 0.536900634765625, 0.5370091552734375, 0.53690673828125, 0.5368411865234375, 0.5370726318359375, 0.5367490844726562, 0.536974365234375, 0.5368411865234375, 0.5371555786132812, 0.5371617431640625, 0.5373317260742188, 0.5367808227539063, 0.5371094970703125, 0.5370316772460938, 0.5373429565429687, 0.537143310546875, 0.5374402465820313, 0.5370582885742188, 0.5370162963867188, 0.5369845581054687, 0.5371781005859375, 0.537079833984375, 0.53718017578125, 0.5371064453125, 0.537385986328125, 0.5369876708984375, 0.53728564453125, 0.5370982666015625, 0.5374218139648438, 0.5371422729492188, 0.5370470581054687, 0.5369978637695313, 0.5371453247070312, 0.5370398559570313, 0.5374299926757813, 0.5371023559570313, 0.5374505004882812, 0.5369210815429688, 0.5371842651367188, 0.5370501098632813, 0.5371821899414062, 0.5369190673828125, 0.5373470458984375, 0.5372252197265625, 0.53749658203125, 0.5371637573242187, 0.5375364990234375, 0.5372713012695313, 0.5375160522460938, 0.5368955078125, 0.5375631103515625, 0.5373101806640626, 0.5374607543945312, 1.1603077392578125, 0.5368985595703125, 0.53696923828125, 0.5368391723632813, 0.5368319702148437, 0.5371904296875, 0.5369384765625, 0.5367337036132812, 0.5369794311523437, 0.5366517944335938, 0.537037841796875, 0.5368688354492187, 0.5368319702148437, 0.5369343872070312, 0.5368125610351563, 0.536690673828125, 0.53686376953125, 0.5367347412109374, 0.53703271484375, 0.536848388671875, 0.5371566162109375, 0.536901611328125, 0.5371791381835938, 0.5370736694335938, 0.53732763671875, 0.5371719970703125, 0.5374044189453125, 0.5371207885742187, 0.5372057495117187, 0.5371105346679688, 0.5373675537109375, 0.5369600219726562, 0.5372620849609375, 0.5369722900390625, 0.537396240234375, 0.5369630737304687, 0.5375784912109375, 0.5370101928710938, 0.5372200927734375, 0.5372108764648438, 0.537186279296875, 0.5369098510742187, 0.5371688842773438, 0.5370357666015625, 0.5374678955078125, 0.5371658325195312, 0.5376061401367187, 0.53692724609375, 0.5371392211914062, 0.5370582885742188, 0.5371289672851562, 0.5370029907226562, 0.5373409423828125, 0.537228271484375, 0.5373900756835938, 0.5372395629882812, 0.5374474487304688, 0.5372374877929688, 0.5372262573242188, 0.5368739624023438, 0.5374719848632813, 0.5371740112304687, 0.5377188110351563, 1.159825439453125, 0.536869873046875, 0.5369948120117187, 0.5368688354492187, 0.5370460205078125, 0.536690673828125, 0.536985595703125, 0.5366886596679687, 0.5370675048828125, 0.5364070434570313, 0.53718017578125, 0.5367019653320313, 0.5370408935546875, 0.5369384765625, 0.5369620361328125, 0.5367838745117187, 0.5370101928710938, 0.536869873046875, 0.537027587890625, 0.5369927978515625, 0.5370398559570313, 0.5367500610351562, 0.5371637573242187, 0.5370224609375, 0.5372241821289062, 0.5369630737304687, 0.5373501586914062, 0.5369835815429688, 0.5371658325195312, 0.5370562744140625, 0.5373265991210937, 0.5370470581054687, 0.5372631225585938, 0.5368872680664063, 0.5373521728515624, 0.5371197509765625, 0.5373317260742188, 0.5371627807617188, 0.5372139282226562, 0.5371361083984375, 0.5373163452148437, 0.5369282836914062, 0.5371760864257813, 0.5369364624023437, 0.5373532104492188, 0.5373030395507813, 0.5374054565429688, 0.5369262084960937, 0.5374248657226562, 0.5371453247070312, 0.5373746948242187, 0.537185302734375, 0.537692138671875, 0.5372098388671875, 0.5374781494140625, 0.5371146240234375, 0.5374945068359375, 0.5370982666015625, 0.5371719970703125, 0.537037841796875, 0.5375795288085937, 0.5371279296875, 0.5373051147460938, 1.1591044921875, 0.5368514404296875, 0.5373880615234375, 0.5367183227539063, 0.5368985595703125, 0.5367828369140625, 0.5369292602539063, 0.5367296142578125, 0.5370357666015625, 0.5368729858398438, 0.5370439453125, 0.5367787475585938, 0.5368627319335938, 0.5367992553710937, 0.5369036865234375, 0.5367838745117187, 0.5368934326171875, 0.5367030029296875, 0.53711669921875, 0.5368422241210937, 0.5371361083984375, 0.5369579467773438, 0.5373798217773438, 0.5369467163085937, 0.5371658325195312, 0.537038818359375, 0.5374607543945312, 0.5370890502929687, 0.5372354736328125, 0.53711669921875, 0.5372088623046875, 0.5368576049804688, 0.5372252197265625, 0.5368668212890625, 0.5373583374023437, 0.5369968872070312, 0.5375211791992187, 0.5370316772460938, 0.5372211303710938, 0.537091064453125, 0.5372241821289062, 0.5368074340820312, 0.5371074829101562, 0.537154541015625, 0.5373409423828125, 0.5372426147460938, 0.5373746948242187, 0.53696923828125, 0.5373778076171875, 0.5370316772460938, 0.5371064453125, 0.5371883544921875, 0.5374105834960937, 0.5373143310546875, 0.5374822387695313, 0.5371658325195312, 0.537438232421875, 0.5372293090820313, 0.5373399047851563, 0.5369476928710938, 0.5372938232421876, 0.53732861328125, 0.5371525268554688, 1.1606148681640625, 0.5368709106445313, 0.537049072265625, 0.5368524780273437, 0.5368616943359374, 0.5371586303710938, 0.5369528198242187, 0.5366599731445313, 0.537049072265625, 0.5364971313476562, 0.5371094970703125, 0.536690673828125, 0.5367777099609375, 0.5367633666992188, 0.53699072265625, 0.5371033325195312, 0.5367920532226562, 0.5366876220703125, 0.5370706176757812, 0.5370224609375, 0.5373143310546875, 0.5367756958007812, 0.5372948608398438, 0.5370153198242188, 0.5372548828125, 0.536932373046875, 0.5373552856445313, 0.5369825439453125, 0.5371002807617188, 0.537217041015625, 0.537332763671875, 0.5370439453125, 0.5372569580078125, 0.5368811645507813, 0.5373009643554687, 0.5369774169921875, 0.5372927856445312, 0.53722216796875, 0.5372958984375, 0.536995849609375, 0.5370992431640625, 0.5370265502929688, 0.5373112182617188, 0.5370603637695313, 0.5374248657226562, 0.5372364501953125, 0.537511962890625, 0.5369671630859375, 0.5372569580078125, 0.5368422241210937, 0.5370869750976562, 0.5370029907226562, 0.53732861328125, 0.5371924438476563, 0.5372190551757813, 0.537006103515625, 0.5375999755859375, 0.5372948608398438, 0.5373081665039062, 0.5367767333984375, 0.53734912109375, 0.537365478515625, 0.5374453735351562, 1.1584112548828125, 0.536911865234375, 0.5371248779296875, 0.5366497192382812, 0.53728564453125, 0.5368248291015625, 0.536859619140625, 0.536838134765625, 0.5368760375976562, 0.5366333618164062, 0.5369682006835937, 0.5368125610351563, 0.5369620361328125, 0.5368678588867187, 0.5370029907226562, 0.536616943359375, 0.5369866333007812, 0.5368657836914063, 0.537059326171875, 0.536795166015625, 0.53722216796875, 0.5368893432617188, 0.5371883544921875, 0.5369292602539063, 0.5370715942382812, 0.5371924438476563, 0.537575439453125, 0.5368770751953125, 0.537238525390625, 0.5371658325195312, 0.5372815551757812, 0.537037841796875, 0.5371259155273438, 0.5369180297851562, 0.537354248046875, 0.5370726318359375, 0.537275390625, 0.5371340942382813, 0.537291748046875, 0.537017333984375, 0.5371401977539062, 0.53724365234375, 0.5372682495117187, 0.5370900268554688, 0.5373521728515624, 0.5370685424804688, 0.5372989501953125, 0.536911865234375, 0.5372057495117187, 0.5371248779296875, 0.5373235473632813, 0.5369395141601563, 0.537427978515625, 0.5376696166992188, 0.537343994140625, 0.5369886474609376, 0.537523193359375, 0.5372200927734375, 0.5373306884765625, 0.5369682006835937, 0.5373040771484375, 0.5371627807617188, 0.537533447265625, 1.15943115234375, 0.5369047241210938, 0.5374474487304688, 0.5369364624023437, 0.5368299560546875, 0.53684326171875, 0.5372139282226562, 0.5367030029296875, 0.5370921020507813, 0.5367357177734375, 0.5369508056640625, 0.536848388671875, 0.5369784545898437, 0.5365545043945312, 0.5367613525390625, 0.536711181640625, 0.5370614013671875, 0.5371105346679688, 0.5371883544921875, 0.5368330078125, 0.5372927856445312, 0.5367603149414063, 0.5372252197265625, 0.53699072265625, 0.5373624267578125, 0.5371207885742187, 0.537406494140625, 0.5369968872070312, 0.5372323608398437, 0.537059326171875, 0.5373685913085937, 0.5369948120117187, 0.5372723388671875, 0.5368565673828125, 0.5371392211914062, 0.5369763793945312, 0.5374299926757813, 0.5370664672851563, 0.5371392211914062, 0.5373184204101562, 0.5369876708984375, 0.5368934326171875, 0.5372743530273437, 0.5370818481445313, 0.537354248046875, 0.537133056640625, 0.537364501953125, 0.5369241333007813, 0.5372211303710938, 0.537133056640625, 0.537238525390625, 0.5370582885742188, 0.5374187622070312, 0.5373501586914062, 0.5373060913085937, 0.5370634155273437, 0.5375375366210937, 0.5371637573242187, 0.5374136352539063, 0.5369292602539063, 0.5373798217773438, 0.5370337524414063, 0.5373726806640625, 1.159437255859375, 0.53688525390625, 0.5370245361328125, 0.5368156127929687, 0.5369190673828125, 0.5368115234375, 0.5373972778320313, 0.5369129028320313, 0.5371268920898438, 0.5366988525390625, 0.5368770751953125, 0.5367101440429688, 0.5367337036132812, 0.536869873046875, 0.5367859497070312, 0.5366343383789063, 0.5368739624023438, 0.5368955078125, 0.5373378295898438, 0.5368309936523438, 0.5371996459960937, 0.5370316772460938, 0.53713818359375, 0.5369682006835937, 0.5373306884765625, 0.5371463623046875, 0.5374033813476562, 0.5371002807617188, 0.5371340942382813, 0.537112548828125, 0.5374371948242187, 0.5369467163085937, 0.537270263671875, 0.5368545532226563, 0.5374730224609375, 0.5369088134765625, 0.5374105834960937, 0.5371248779296875, 0.5373726806640625, 0.537154541015625, 0.53719140625, 0.5370562744140625, 0.537333740234375, 0.5370787963867187, 0.5373941650390625, 0.53711767578125, 0.5374453735351562, 0.5370869750976562, 0.537343994140625, 0.5371422729492188, 0.5372979125976562, 0.5372733154296875, 0.5372272338867188, 0.53725390625, 0.5373798217773438, 0.5372610473632813, 0.5374996337890625, 0.5372794799804688, 0.537186279296875, 0.5369476928710938, 0.5374187622070312, 0.5371514892578125, 0.53749658203125]",tokens/s,1.8314446832237186,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10862,7 +10862,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10993,7 +10993,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GP ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1271.554048,1017.643008,0.0,371.195904,315.446272,s,10,0.19529705619812013,0.019529705619812012,0.0009140443366206495,0.019363104820251463,0.021000485801696778,0.021201970767974854,0.021363158740997316,"[0.020955711364746093, 0.018462751388549806, 0.01938128089904785, 0.019328096389770507, 0.02140345573425293, 0.018568992614746094, 0.018725887298583984, 0.01934492874145508, 0.019606687545776366, 0.019519264221191407]",tokens/s,13108.236497958242,kWh,2.261678843151771e-07,1.2392898202703704e-07,5.473570344435316e-07,8.974539007857457e-07,tokens/kWh,285251420.4638978,MB,1271.554048,1017.643008,0.0,371.195904,340.902912,s,10,11.830128784179688,1.1830128784179688,0.010220015064036666,1.1847022705078125,1.1948863647460937,1.1950671447753907,1.1952117687988282,"[1.1664869384765626, 1.1741630859375, 1.19484619140625, 1.1935650634765624, 1.1794437255859376, 1.188765869140625, 1.1682054443359375, 1.1879569091796875, 1.1814476318359375, 1.1952479248046874]",tokens/s,53.25385813571976,kWh,1.3598944398914649e-05,7.451885775429064e-06,2.9480500321153233e-05,5.0531330495496944e-05,tokens/kWh,1246751.2606978395,,s,629,11.982314514160153,0.019049784601208517,0.002366055636612425,0.01886207962036133,0.019090431213378906,0.01922559928894043,0.037088911590576173,"[0.020017152786254884, 0.01830297660827637, 0.01788211250305176, 0.017999872207641602, 0.017724416732788087, 0.017986560821533205, 0.01797324752807617, 0.017954816818237306, 0.018149375915527344, 0.01821286392211914, 0.017949695587158202, 0.01780019187927246, 0.017811519622802734, 0.01791276741027832, 0.01782374382019043, 0.018069503784179687, 0.018145280838012694, 0.018121728897094725, 0.017817728042602538, 0.017745792388916017, 0.017811456680297853, 0.017704959869384765, 0.017649663925170898, 0.01785139274597168, 0.017945600509643556, 0.017889280319213868, 0.017904640197753906, 0.017810432434082032, 0.017958911895751953, 0.01818009567260742, 0.018121728897094725, 0.018191360473632814, 0.01867366409301758, 0.019141664505004884, 0.019366880416870118, 0.01925324821472168, 0.019109952926635743, 0.018839487075805662, 0.018568191528320312, 0.019009536743164062, 0.018899967193603515, 0.01883647918701172, 0.0190382080078125, 0.01876479911804199, 0.018717695236206054, 0.018765823364257812, 0.01883545684814453, 0.018886688232421876, 0.018552799224853516, 0.01900748825073242, 0.018927616119384767, 0.018916351318359375, 0.018776063919067384, 0.018911231994628908, 0.018926591873168946, 0.019090431213378906, 0.018936832427978514, 0.018993152618408202, 0.018960384368896483, 0.018997247695922852, 0.019090431213378906, 0.018817119598388672, 0.03665296173095703, 0.017954816818237306, 0.018086912155151368, 0.01783193588256836, 0.01798860740661621, 0.018092031478881835, 0.018083839416503905, 0.018058240890502928, 0.018153472900390624, 0.017795072555541993, 0.0180316162109375, 0.018076671600341796, 0.018059263229370116, 0.01798860740661621, 0.01844428825378418, 0.01783500862121582, 0.01799065589904785, 0.018164863586425783, 0.019538816452026368, 0.019020927429199218, 0.018961280822753907, 0.01883750343322754, 0.01900851249694824, 0.01863577651977539, 0.018566144943237304, 0.018934783935546876, 0.018930688858032226, 0.01903104019165039, 0.019366912841796875, 0.019103744506835937, 0.01902592086791992, 0.018909183502197266, 0.018679807662963867, 0.018682880401611326, 0.01884876823425293, 0.01885593605041504, 0.018932735443115235, 0.018775039672851563, 0.018779136657714843, 0.01899929618835449, 0.018562047958374024, 0.018780160903930664, 0.01865830421447754, 0.018711551666259766, 0.018930688858032226, 0.018962560653686525, 0.018713472366333007, 0.018715648651123046, 0.018820127487182616, 0.01867670440673828, 0.018869247436523438, 0.018967552185058592, 0.0190248966217041, 0.01882316780090332, 0.01884979248046875, 0.018946048736572265, 0.018947071075439453, 0.018792448043823243, 0.018743295669555664, 0.018677759170532226, 0.018784255981445314, 0.019042303085327148, 0.018932895660400392, 0.03863228988647461, 0.01879654312133789, 0.019182592391967773, 0.01887948799133301, 0.01880985641479492, 0.018988031387329102, 0.018840576171875, 0.018990079879760743, 0.01983897590637207, 0.01920102310180664, 0.019095552444458007, 0.019406848907470704, 0.019109888076782225, 0.018985984802246093, 0.01922559928894043, 0.018942975997924806, 0.01883443260192871, 0.019072000503540038, 0.019217439651489258, 0.019035104751586915, 0.018884607315063476, 0.01901260757446289, 0.018982912063598634, 0.01926976013183594, 0.018795391082763672, 0.01906380844116211, 0.018922496795654296, 0.018951168060302736, 0.018964479446411133, 0.018975744247436522, 0.01904742431640625, 0.018957311630249024, 0.018982912063598634, 0.01903206443786621, 0.01886617660522461, 0.01887948799133301, 0.01885491180419922, 0.01879654312133789, 0.018924543380737305, 0.018731008529663085, 0.018837663650512697, 0.018797407150268553, 0.01885081672668457, 0.018994176864624023, 0.01904025650024414, 0.01883750343322754, 0.018844671249389648, 0.01892777633666992, 0.018924383163452147, 0.019001344680786132, 0.01899212837219238, 0.019002368927001953, 0.01883852767944336, 0.018927616119384767, 0.0189040641784668, 0.018844671249389648, 0.01922777557373047, 0.018747264862060548, 0.01879756736755371, 0.01862860870361328, 0.01885491180419922, 0.018626560211181642, 0.018851839065551757, 0.03850870513916015, 0.018795360565185548, 0.01887027168273926, 0.018887744903564454, 0.018889663696289062, 0.018948095321655273, 0.01903206443786621, 0.01909350395202637, 0.01902592086791992, 0.018903039932250978, 0.019079168319702147, 0.019013631820678712, 0.0190515193939209, 0.018990079879760743, 0.01904640007019043, 0.019207296371459962, 0.01885683250427246, 0.018952192306518553, 0.018957311630249024, 0.019281919479370118, 0.019170303344726563, 0.01906483268737793, 0.018900064468383788, 0.018952096939086914, 0.01886310386657715, 0.01868083190917969, 0.01889587211608887, 0.018905088424682616, 0.018905088424682616, 0.018985984802246093, 0.018840576171875, 0.01866035270690918, 0.018889728546142577, 0.01878118324279785, 0.018689023971557618, 0.018696191787719727, 0.018860031127929687, 0.01886207962036133, 0.0188272647857666, 0.01899519920349121, 0.018921472549438476, 0.018953216552734374, 0.018710687637329103, 0.01884553527832031, 0.018985984802246093, 0.019359743118286133, 0.018723840713500976, 0.018914304733276367, 0.01888768005371094, 0.019006464004516603, 0.01885491180419922, 0.018954240798950195, 0.018852863311767578, 0.01925734329223633, 0.019005439758300782, 0.019078144073486326, 0.018884607315063476, 0.018915327072143554, 0.018993152618408202, 0.018742271423339844, 0.019376127243041993, 0.0190515193939209, 0.01882316780090332, 0.04119039916992188, 0.018949119567871094, 0.019070976257324217, 0.018922496795654296, 0.018876415252685547, 0.019170303344726563, 0.019082239151000976, 0.018942975997924806, 0.018954240798950195, 0.018775039672851563, 0.018718719482421875, 0.018762752532958983, 0.018670591354370117, 0.018731039047241212, 0.019002336502075196, 0.018882560729980468, 0.018729984283447267, 0.01904947280883789, 0.018869247436523438, 0.018756704330444338, 0.0187554874420166, 0.018935808181762694, 0.01886617660522461, 0.018966527938842775, 0.018799615859985352, 0.01905766487121582, 0.018744319915771485, 0.01900851249694824, 0.019042303085327148, 0.018704383850097657, 0.018779136657714843, 0.018898944854736328, 0.018880512237548826, 0.019350528717041016, 0.018659328460693358, 0.018634752273559572, 0.018785280227661134, 0.018497535705566406, 0.018531360626220704, 0.018791391372680665, 0.018892799377441406, 0.019111936569213867, 0.018805856704711913, 0.01858550453186035, 0.018798656463623047, 0.018900928497314454, 0.01904844856262207, 0.018962432861328125, 0.018732032775878905, 0.01922559928894043, 0.018960384368896483, 0.018158592224121094, 0.017957887649536132, 0.017901632308959962, 0.017839040756225586, 0.017967103958129883, 0.0178606071472168, 0.01805619239807129, 0.018120704650878908, 0.018111488342285157, 0.018066432952880858, 0.017943552017211914, 0.01808076858520508, 0.036969470977783206, 0.018720767974853517, 0.018741312026977538, 0.018897855758666992, 0.018877439498901367, 0.018736127853393555, 0.01883955192565918, 0.018892799377441406, 0.019025951385498046, 0.01890915107727051, 0.01892972755432129, 0.018944992065429687, 0.018826208114624023, 0.018890752792358398, 0.018911231994628908, 0.018881536483764647, 0.018629728317260744, 0.018756511688232422, 0.018934783935546876, 0.018855968475341798, 0.018990047454833986, 0.01881907272338867, 0.019137535095214844, 0.01884569549560547, 0.01884160041809082, 0.018712608337402344, 0.018834400177001952, 0.018743295669555664, 0.018499584197998048, 0.01885491180419922, 0.018917375564575196, 0.018942975997924806, 0.01904640007019043, 0.018973695755004884, 0.018658336639404298, 0.01889686393737793, 0.018727935791015626, 0.019357696533203125, 0.019466239929199217, 0.01920614433288574, 0.01903001594543457, 0.018939903259277344, 0.01880268859863281, 0.01885081672668457, 0.01865727996826172, 0.0187064323425293, 0.01901568031311035, 0.018894847869873048, 0.01869004821777344, 0.01888768005371094, 0.018707456588745116, 0.018958335876464845, 0.01881507110595703, 0.018725791931152345, 0.018954240798950195, 0.018803712844848632, 0.018970624923706055, 0.018896896362304686, 0.01886617660522461, 0.018924575805664062, 0.018801631927490233, 0.01897881507873535, 0.01885081672668457, 0.03713536071777344, 0.01801420783996582, 0.01782681655883789, 0.01780121612548828, 0.017937408447265626, 0.01788313674926758, 0.018086912155151368, 0.017979488372802735, 0.018089887619018554, 0.01799679946899414, 0.01784217643737793, 0.01804800033569336, 0.018077695846557617, 0.018420736312866212, 0.01813811111450195, 0.018081792831420897, 0.01803264045715332, 0.017977344512939454, 0.01801215934753418, 0.017757247924804688, 0.01795167922973633, 0.01799679946899414, 0.01804083251953125, 0.01806438446044922, 0.01840640068054199, 0.018937856674194335, 0.01873523139953613, 0.018714591979980467, 0.0187554874420166, 0.0188723201751709, 0.018910207748413087, 0.01945292854309082, 0.01900032043457031, 0.018687999725341797, 0.018775039672851563, 0.01882111930847168, 0.018571264266967775, 0.01883558464050293, 0.01885580825805664, 0.018915327072143554, 0.018856960296630858, 0.018940959930419922, 0.01891119956970215, 0.018887712478637696, 0.01914569664001465, 0.01885491180419922, 0.01862246322631836, 0.01860403251647949, 0.018769920349121092, 0.018704383850097657, 0.01893587112426758, 0.019063743591308593, 0.018961408615112304, 0.019014656066894533, 0.018930688858032226, 0.01883033561706543, 0.018955263137817382, 0.018968576431274413, 0.01900748825073242, 0.018897920608520507, 0.019014656066894533, 0.018824192047119142, 0.01883135986328125, 0.038749183654785156, 0.01903923225402832, 0.01925433540344238, 0.018804672241210938, 0.01879449653625488, 0.018709503173828124, 0.018998271942138673, 0.0188538875579834, 0.01887846374511719, 0.018940927505493164, 0.018928640365600585, 0.0188221435546875, 0.018743295669555664, 0.01880575942993164, 0.01882019233703613, 0.019021728515625, 0.01899929618835449, 0.018715648651123046, 0.019109888076782225, 0.01904640007019043, 0.01886412811279297, 0.01902796745300293, 0.01884671974182129, 0.018951168060302736, 0.01845248031616211, 0.01866649627685547, 0.018948095321655273, 0.0188221435546875, 0.018982912063598634, 0.018701311111450195, 0.018494464874267577, 0.018754560470581053, 0.018967552185058592, 0.018783231735229493, 0.018856960296630858, 0.01887539291381836, 0.018757696151733397, 0.018967487335205077, 0.018883583068847656, 0.01875660705566406, 0.018545663833618165, 0.01883647918701172, 0.01883135986328125, 0.01861631965637207, 0.018696191787719727, 0.018667520523071288, 0.018711551666259766, 0.018933759689331055, 0.018709503173828124, 0.018785280227661134, 0.01859891128540039, 0.018540576934814455, 0.01872483253479004, 0.01882521629333496, 0.018964479446411133, 0.0194652156829834, 0.018729984283447267, 0.019041311264038085, 0.01892348861694336, 0.01902592086791992, 0.018899967193603515, 0.018945024490356444, 0.01891744041442871, 0.039, 0.01886207962036133, 0.018724863052368163, 0.018851839065551757, 0.01878835105895996, 0.018738208770751955, 0.01889481544494629, 0.018662399291992187, 0.01781760025024414, 0.017797119140625, 0.017918975830078124, 0.017838079452514647, 0.01800601577758789, 0.018058240890502928, 0.017975296020507812, 0.018059263229370116, 0.01868815994262695, 0.01871446418762207, 0.0186562557220459, 0.01878953552246094, 0.01872265625, 0.018795551300048827, 0.018784223556518556, 0.01888768005371094, 0.018791423797607423, 0.018795520782470702, 0.01904844856262207, 0.0188221435546875, 0.018909183502197266, 0.01878835105895996, 0.01879654312133789, 0.0188723201751709, 0.018998271942138673, 0.01898806381225586, 0.01884873580932617, 0.018703359603881836, 0.018964479446411133, 0.019092479705810548, 0.01896361541748047, 0.01868988800048828, 0.018832384109497072, 0.018837535858154297, 0.018735200881958007, 0.018717567443847657, 0.019216384887695313, 0.01887948799133301, 0.018950143814086915, 0.018972671508789063, 0.019096576690673828, 0.01882009506225586, 0.018949119567871094, 0.01900851249694824, 0.018808832168579103, 0.018893823623657227, 0.018735103607177735, 0.018786304473876952, 0.018913280487060546, 0.018896896362304686, 0.019161088943481445, 0.01902796745300293, 0.01904435157775879, 0.019018943786621095, 0.018677568435668944, 0.038816768646240236, 0.018958335876464845, 0.019017728805541992, 0.018979839324951172, 0.01882521629333496, 0.018888704299926756, 0.01883340835571289, 0.01893174362182617, 0.018801631927490233, 0.018751487731933594, 0.018884607315063476, 0.01901875114440918, 0.01905254364013672, 0.018852863311767578, 0.01902796745300293, 0.01903104019165039, 0.019095552444458007, 0.018968576431274413, 0.019164159774780275, 0.018955263137817382, 0.018951168060302736, 0.018985984802246093, 0.01905254364013672, 0.018993152618408202, 0.018977792739868164, 0.01875660705566406, 0.018876415252685547, 0.018720895767211913, 0.01888140869140625, 0.018778112411499022, 0.018736127853393555, 0.018723840713500976, 0.018611200332641603, 0.019019775390625, 0.019397632598876953, 0.019515392303466796, 0.01984102439880371, 0.01907721519470215, 0.01902275276184082, 0.019108928680419923, 0.018827199935913086, 0.018966527938842775, 0.01899929618835449, 0.018892799377441406, 0.01909760093688965, 0.01885593605041504, 0.018971647262573242, 0.019215360641479492, 0.018757631301879883, 0.019134464263916014, 0.01903923225402832, 0.019009536743164062, 0.01903308868408203, 0.019096576690673828, 0.01880985641479492, 0.018685951232910156, 0.01884671974182129, 0.01868499183654785, 0.019015615463256835, 0.01898700714111328, 0.018881599426269532, 0.018994112014770508, 0.019103744506835937]",tokens/s,52.49403187144489,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Salesforce/codegen-16B-nl,Salesforce/codegen-16B-nl,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1812.414464,11891.376128,0.0,11244.929024,10436.256768,s,10,7.439572692871093,0.7439572692871095,0.0008714544962476244,0.7437487487792969,0.7443923461914063,0.7453685363769531,0.7461494885253905,"[0.7463447265625, 0.74344580078125, 0.7431934204101562, 0.7440980834960937, 0.7432208862304688, 0.7434606323242188, 0.743556396484375, 0.7439411010742187, 0.7441754150390625, 0.74413623046875]",tokens/s,344.10578479233055,kWh,8.78386018531663e-06,4.811645483667754e-06,4.263015315171338e-05,5.622565882069777e-05,tokens/kWh,4553081.375469119,MB,1813.250048,11891.376128,0.0,11244.929024,10730.124288,s,10,411.32077734375,41.132077734375,0.0035682239044481146,41.1322109375,41.137268750000004,41.13731796875,41.13735734375,"[41.1373671875, 41.132734375, 41.1316875, 41.12786328125, 41.12597265625, 41.13050390625, 41.1372578125, 41.13343359375, 41.12947265625, 41.134484375]",tokens/s,1.5316512918905987,kWh,0.00048559434207422404,0.00026614820952837497,0.002371767778365088,0.003123510329967687,tokens/kWh,20169.614742606514,,s,629,417.3603321533206,0.663529939830398,0.08832732816621393,0.652874755859375,0.653243408203125,0.6533930908203125,1.39622314453125,"[0.6530560302734375, 0.65274267578125, 0.6525982666015625, 0.652653564453125, 0.652601318359375, 0.6530682983398437, 0.6528573608398438, 0.6528583984375, 0.6530365600585938, 0.6527334594726563, 0.6526627807617188, 0.6529054565429687, 0.6528788452148437, 0.6528194580078125, 0.6533068237304688, 0.6530610961914063, 0.6530549926757813, 0.6526453857421874, 0.6527969360351562, 0.65291162109375, 0.65301708984375, 0.6531358642578124, 0.6527825927734375, 0.6527999877929688, 0.6528389282226562, 0.6530457763671875, 0.65295361328125, 0.6529638671875, 0.6528235473632813, 0.6529945678710938, 0.653243408203125, 0.6532218627929688, 0.652948486328125, 0.6528409423828125, 0.6529187622070313, 0.6527262573242187, 0.653053955078125, 0.6529075317382812, 0.65293310546875, 0.6529146728515625, 0.6530498657226562, 0.653000732421875, 0.6531154174804688, 0.652927001953125, 0.6532239379882813, 0.6535833740234375, 0.6530744018554687, 0.653233154296875, 0.652674072265625, 0.6528614501953125, 0.6530457763671875, 0.6528726806640625, 0.6530406494140625, 0.6529894409179687, 0.653138916015625, 0.6526207885742188, 0.6533232421875, 0.6529812622070312, 0.6528389282226562, 0.6529556274414062, 0.6532802734375, 0.653212646484375, 1.396294677734375, 0.65265869140625, 0.652442626953125, 0.6525101928710938, 0.65289013671875, 0.6526863403320312, 0.6524282836914063, 0.6526934814453125, 0.6525409545898437, 0.6526239013671875, 0.6528583984375, 0.652990478515625, 0.6530744018554687, 0.6530631713867188, 0.6530795288085938, 0.652579833984375, 0.6526627807617188, 0.6525849609375, 0.6526340942382812, 0.6529996948242187, 0.6528450317382812, 0.6529177856445313, 0.6525849609375, 0.6525440063476563, 0.6526607055664062, 0.6527150268554688, 0.6526709594726563, 0.6531112670898438, 0.65360791015625, 0.6530764770507812, 0.6530979614257812, 0.6530426635742187, 0.6529505004882813, 0.652748779296875, 0.6531727294921875, 0.653011962890625, 0.6526146850585938, 0.6529310913085937, 0.6529720458984375, 0.6528348388671875, 0.6533939208984375, 0.6528767700195313, 0.6530232543945312, 0.65400732421875, 0.6527682495117187, 0.6532024536132812, 0.6528726806640625, 0.6528675537109375, 0.6528809204101562, 0.6525962524414063, 0.6530969848632813, 0.65284814453125, 0.6528972778320312, 0.65278466796875, 0.6529822998046875, 0.6528809204101562, 0.6528071899414063, 0.6527999877929688, 0.6533150634765625, 0.65329150390625, 0.6535679931640626, 0.6529740600585937, 0.6530877685546875, 1.3959822998046876, 0.6529853515625, 0.6526556396484375, 0.6527498168945313, 0.6532198486328125, 0.6525972290039063, 0.6524252319335937, 0.6525859985351562, 0.6526023559570312, 0.6529525756835938, 0.6530232543945312, 0.6528706665039062, 0.6525993041992187, 0.6526576538085938, 0.6524313354492187, 0.65303759765625, 0.65289013671875, 0.6526351318359375, 0.6526730346679688, 0.6527744140625, 0.65242724609375, 0.6525890502929688, 0.6524979248046875, 0.652642333984375, 0.6526842651367187, 0.6535157470703125, 0.6526822509765625, 0.652727294921875, 0.6526473999023438, 0.6525808715820313, 0.6529013671875, 0.6527549438476562, 0.6526904907226563, 0.6529207763671875, 0.6530181274414063, 0.6531665649414062, 0.652938232421875, 0.6530048217773438, 0.6527549438476562, 0.65276416015625, 0.65320654296875, 0.6533499145507813, 0.65331201171875, 0.6528256225585938, 0.6529412841796876, 0.6526904296875, 0.652916748046875, 0.6531583862304687, 0.6532269897460937, 0.6527897338867188, 0.6528358154296875, 0.65314404296875, 0.6529362182617188, 0.65278466796875, 0.6528409423828125, 0.6531635131835938, 0.6535239868164062, 0.65337548828125, 0.6532198486328125, 0.6530682983398437, 0.6528880615234375, 0.65310107421875, 0.6533202514648437, 1.397123046875, 0.6522962036132812, 0.6524856567382813, 0.6524671630859376, 0.6524682006835938, 0.6523873291015625, 0.652527587890625, 0.6525296630859375, 0.6522194213867187, 0.65272216796875, 0.6525880126953125, 0.6523699340820313, 0.6523811645507812, 0.6526883544921875, 0.6524754028320312, 0.65251123046875, 0.653138916015625, 0.6526986083984375, 0.6525542602539063, 0.652579833984375, 0.6525941772460937, 0.6528235473632813, 0.6532802734375, 0.6527744140625, 0.6530477905273437, 0.6530468139648438, 0.652590087890625, 0.6526617431640624, 0.6527385864257812, 0.653243408203125, 0.6531707153320313, 0.65314306640625, 0.6529208374023437, 0.6529556274414062, 0.6531307373046875, 0.6527764282226562, 0.6528614501953125, 0.6530078735351562, 0.6530938720703126, 0.6532454223632812, 0.6528409423828125, 0.65280615234375, 0.6527600708007812, 0.6526914672851563, 0.6530109252929688, 0.6528010864257813, 0.652791748046875, 0.6528409423828125, 0.652916748046875, 0.6529423217773438, 0.65280615234375, 0.6530509033203125, 0.6532987060546875, 0.6532352294921875, 0.6532269897460937, 0.6532689819335937, 0.6528071899414063, 0.6527559814453125, 0.6529658813476562, 0.6530488891601562, 0.6527354736328125, 0.6528317260742188, 0.6532055053710938, 1.3962239990234375, 0.6522808227539062, 0.6524036865234375, 0.6524794921875, 0.6530989990234375, 0.6525030517578125, 0.6528368530273437, 0.652431396484375, 0.652458984375, 0.6524169921875, 0.6527979736328124, 0.6531522827148437, 0.6529720458984375, 0.6528245849609375, 0.6527119140625, 0.6529351806640625, 0.6526033935546875, 0.6526863403320312, 0.6525675659179687, 0.6530775146484376, 0.65268017578125, 0.6526934814453125, 0.6526156616210937, 0.65263720703125, 0.652642333984375, 0.6526822509765625, 0.6527999877929688, 0.6526300048828125, 0.6526033935546875, 0.6526187744140625, 0.6527191162109375, 0.6527139892578125, 0.6527334594726563, 0.6525234985351562, 0.6529003295898438, 0.6528358154296875, 0.652747802734375, 0.6531891479492188, 0.65274267578125, 0.6529146728515625, 0.6527396240234375, 0.6529453735351562, 0.65310107421875, 0.6528511962890625, 0.6529003295898438, 0.6527980346679687, 0.6525654296875, 0.65280615234375, 0.6527918090820313, 0.6529085693359375, 0.6532034301757812, 0.6529791870117188, 0.6532505493164062, 0.6529495239257812, 0.6528460693359375, 0.6527334594726563, 0.6529054565429687, 0.6528081665039063, 0.6530529174804688, 0.6530703125, 0.6531522827148437, 0.653106201171875, 0.6527354736328125, 1.396220947265625, 0.6531154174804688, 0.652801025390625, 0.653065185546875, 0.652675048828125, 0.652779541015625, 0.65251025390625, 0.6525501098632812, 0.6528460693359375, 0.6527385864257812, 0.6529986572265625, 0.6522655029296875, 0.6526504516601562, 0.6524989624023437, 0.65251123046875, 0.6524682006835938, 0.6526351318359375, 0.6526658325195313, 0.6525296630859375, 0.6525736694335937, 0.6526228637695313, 0.6524119262695313, 0.6525603637695313, 0.6528204956054687, 0.6526351318359375, 0.6527252197265625, 0.6525183715820313, 0.6526402587890625, 0.6525993041992187, 0.6524119262695313, 0.6528368530273437, 0.6529505004882813, 0.652843017578125, 0.6528265991210938, 0.6528552856445312, 0.6527774658203125, 0.6529669189453124, 0.653106201171875, 0.6532741088867188, 0.6532218627929688, 0.6529915161132812, 0.6530150756835937, 0.6528972778320312, 0.6527344360351562, 0.6526689453125, 0.6529771728515625, 0.6529228515625, 0.6532310791015625, 0.6531707153320313, 0.6531215209960938, 0.6532188110351562, 0.6529607543945313, 0.6528624877929687, 0.6531921997070312, 0.6533816528320312, 0.6530037841796875, 0.653359130859375, 0.6530211791992188, 0.6531962890625, 0.6532526245117187, 0.6531983642578125, 0.6532720336914063, 0.6536365966796875, 1.3966376953125, 0.6525726928710938, 0.6523207397460937, 0.6524825439453125, 0.6524968872070313, 0.6534379272460937, 0.6525757446289062, 0.6527897338867188, 0.6527119140625, 0.652526611328125, 0.6523760986328125, 0.65272314453125, 0.6526085205078125, 0.6529075317382812, 0.6529884033203125, 0.6531532592773438, 0.6531287231445313, 0.65270068359375, 0.652705810546875, 0.6528317260742188, 0.653264892578125, 0.6530426635742187, 0.6530498657226562, 0.65275390625, 0.6526044311523438, 0.6528818969726562, 0.65260546875, 0.6527570190429688, 0.6528655395507813, 0.6533887939453125, 0.6528471069335937, 0.653391845703125, 0.6528583984375, 0.653075439453125, 0.6527744140625, 0.653095947265625, 0.6532362060546875, 0.6528706665039062, 0.6531512451171875, 0.6531102905273437, 0.6528952026367187, 0.6531512451171875, 0.6529986572265625, 0.6529577026367187, 0.653117431640625, 0.6531163940429687, 0.653154296875, 0.65341748046875, 0.6533274536132813, 0.6534061279296876, 0.6530570068359375, 0.65308056640625, 0.6530672607421875, 0.6530928344726562, 0.6531143798828125, 0.6529689331054688, 0.6531450805664063, 0.653233154296875, 0.6529024047851563, 0.65328125, 0.6534799194335937, 0.6530867309570313, 0.6534922485351562, 1.3970196533203125, 0.6528092041015625, 0.6530826416015625, 0.65263720703125, 0.65261669921875, 0.6526617431640624, 0.65263818359375, 0.652643310546875, 0.6525614013671875, 0.6526331176757812, 0.6529290161132812, 0.6531737670898438, 0.6526464233398438, 0.6530057983398437, 0.6527815551757813, 0.652675048828125, 0.652821533203125, 0.6530089111328125, 0.6530693359375, 0.6528256225585938, 0.652811279296875, 0.653000732421875, 0.652779541015625, 0.6528706665039062, 0.6528286743164062, 0.6527283325195312, 0.6534901733398437, 0.6527928466796875, 0.6526658325195313, 0.6537155151367188, 0.6528931274414063, 0.65274365234375, 0.6532055053710938, 0.6529832763671874, 0.6527283325195312, 0.6526566162109375, 0.652643310546875, 0.6526505126953125, 0.652853271484375, 0.65282763671875, 0.6527262573242187, 0.6531266479492187, 0.6530938720703126, 0.653048828125, 0.6534696655273438, 0.6529351806640625, 0.6527446899414062, 0.6529126586914062, 0.6531901245117188, 0.6530068359375, 0.6527866821289062, 0.6527037963867187, 0.6526924438476562, 0.6527160034179688, 0.6528348388671875, 0.6528471069335937, 0.65350146484375, 0.6534850463867188, 0.65312255859375, 0.65337548828125, 0.6529249267578126, 0.6529208374023437, 0.6531655883789063, 1.397223388671875, 0.6524876708984375, 0.6524805297851562, 0.6524385375976562, 0.6525757446289062, 0.6523709716796875, 0.6523740234375, 0.6525859985351562, 0.6527559814453125, 0.65253173828125, 0.652495849609375, 0.6527897338867188, 0.6526248779296875, 0.6524805297851562, 0.6527744140625, 0.65310205078125, 0.6527662353515625, 0.6525819091796875, 0.6526607055664062, 0.652695556640625, 0.6524405517578125, 0.6527406005859375, 0.6526505126953125, 0.6531716918945313, 0.6526863403320312, 0.6526044311523438, 0.6529617919921875, 0.6528932495117188, 0.6525480346679687, 0.6525726928710938, 0.6531266479492187, 0.6526771240234375, 0.6526719970703125, 0.6527150268554688, 0.6526453857421874, 0.65263818359375, 0.652927001953125, 0.65286962890625, 0.65310107421875, 0.65362841796875, 0.6530979614257812, 0.653264892578125, 0.652790771484375, 0.6529525756835938, 0.6529863891601563, 0.6530662231445312, 0.6527334594726563, 0.65296484375, 0.652843017578125, 0.652853271484375, 0.6530109252929688, 0.653211669921875, 0.6529710083007813, 0.6534041748046875, 0.6534133911132812, 0.6530693359375, 0.6532454223632812, 0.6532249755859375, 0.6530191650390625, 0.6529628295898438, 0.6533099365234375, 0.6530693359375, 0.6531338500976562, 1.397011474609375, 0.652874755859375, 0.6528573608398438, 0.6530682983398437, 0.6525573120117187, 0.6529822998046875, 0.6531246337890625, 0.6528839721679688, 0.6529187622070313, 0.6528256225585938, 0.6527866821289062, 0.6528993530273437, 0.6527129516601563, 0.652927978515625, 0.6529617919921875, 0.6528993530273437, 0.6526627807617188, 0.6527836303710938, 0.6525388793945313, 0.6524876708984375, 0.6530293579101563, 0.6528767700195313, 0.6527989501953125, 0.6527918090820313, 0.6526709594726563, 0.65257470703125, 0.652927001953125, 0.6529924926757813, 0.65281640625, 0.6529515380859375, 0.6529577026367187, 0.653001708984375, 0.6528297119140625, 0.6529638671875, 0.6530775146484376, 0.6529700317382813, 0.6532412719726562, 0.6529720458984375, 0.65293310546875, 0.6531235961914063, 0.6527662353515625, 0.6526484375, 0.6531204833984375, 0.6531041259765625, 0.652822509765625, 0.6528655395507813, 0.6528143310546874, 0.6528634643554687, 0.6527395629882813, 0.65272216796875, 0.6529495849609375, 0.6534317626953124, 0.6528675537109375, 0.6528818969726562, 0.6532792358398437, 0.6530078735351562, 0.6531532592773438, 0.6536857299804687, 0.6533027954101562, 0.6531686401367187, 0.653180908203125, 0.6531204833984375, 0.6529126586914062]",tokens/s,1.507090999172706,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11363,7 +11363,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1852.063744,3368.5504,0.0,2722.103296,2585.357312,s,10,0.7361118698120117,0.07361118698120116,0.0003642046363895853,0.07361937713623047,0.07414464721679688,0.07420198898315429,0.07424786239624023,"[0.07425933074951171, 0.07323929595947265, 0.07386649322509765, 0.07366835021972656, 0.07318646240234375, 0.07325574493408203, 0.07413190460205078, 0.07326553344726562, 0.07365193939208985, 0.0735868148803711]",tokens/s,3477.732264599637,kWh,8.660967405115021e-07,4.745775110883439e-07,4.11042299636542e-06,5.451097247965265e-06,tokens/kWh,46963022.00360804,MB,1852.452864,3368.5504,0.0,2722.103296,2628.769792,s,10,38.912985107421875,3.8912985107421876,0.002125353517438014,3.89062548828125,3.8938447753906247,3.895329248046875,3.896516826171875,"[3.896813720703125, 3.89010009765625, 3.890788330078125, 3.890443603515625, 3.88959716796875, 3.890462646484375, 3.890949951171875, 3.890897216796875, 3.893514892578125, 3.88941748046875]",tokens/s,16.189968419560802,kWh,4.5877188203519475e-05,2.5143288266442475e-05,0.00021536239147063786,0.00028638286794059977,tokens/kWh,219985.22625685547,,s,629,39.51227389144898,0.06281760555079327,0.008801190571361054,0.061709312438964846,0.06187274169921875,0.06235340805053711,0.13552583007812502,"[0.06250291061401367, 0.0624824333190918, 0.06248860931396484, 0.06241788864135742, 0.06238412857055664, 0.062355457305908205, 0.06268928146362304, 0.06286848068237305, 0.061916160583496097, 0.061690879821777345, 0.06164377593994141, 0.06167552185058594, 0.06169702529907226, 0.061725696563720706, 0.06180147171020508, 0.06190796661376953, 0.06164275360107422, 0.06171340942382812, 0.06235033416748047, 0.06176665496826172, 0.06160793685913086, 0.06171750259399414, 0.061628414154052735, 0.06169497680664063, 0.06164275360107422, 0.061661182403564455, 0.06167244720458984, 0.0616806411743164, 0.061723648071289064, 0.061810817718505856, 0.06176448059082031, 0.06172876739501953, 0.06162124633789062, 0.06174105453491211, 0.06160793685913086, 0.06174924850463867, 0.06167859268188477, 0.0617625617980957, 0.06171340942382812, 0.062069759368896485, 0.06172172927856445, 0.06168972778320313, 0.06166220855712891, 0.06177689743041992, 0.06163558578491211, 0.06175948715209961, 0.061769729614257814, 0.06181683349609375, 0.06176665496826172, 0.061758464813232425, 0.06166425704956055, 0.061740032196044924, 0.061720577239990235, 0.06168371200561523, 0.06170640182495117, 0.06162825775146484, 0.061742080688476565, 0.06170009613037109, 0.06172774505615235, 0.061704193115234375, 0.061769729614257814, 0.06176870346069336, 0.13554074096679689, 0.06162432098388672, 0.06162432098388672, 0.06162134552001953, 0.06155356979370117, 0.06167654418945313, 0.06153932952880859, 0.06170521545410156, 0.06169804763793945, 0.06160486221313476, 0.06173798370361328, 0.061633537292480466, 0.06171964645385742, 0.06159964752197265, 0.061695999145507815, 0.061778942108154294, 0.06163148880004883, 0.061780990600585936, 0.061868030548095705, 0.06245068740844727, 0.06177689743041992, 0.06273433685302734, 0.06171750259399414, 0.061677566528320314, 0.06168678283691406, 0.061638656616210936, 0.061682689666748045, 0.06182297515869141, 0.061917182922363284, 0.06169702529907226, 0.06170624160766602, 0.06173286437988281, 0.061692928314208986, 0.061625343322753906, 0.06172774505615235, 0.06163455963134765, 0.061687808990478515, 0.061682815551757815, 0.06168051147460937, 0.061712383270263675, 0.06174822235107422, 0.061663230895996096, 0.06169001770019531, 0.061602657318115234, 0.06268313598632813, 0.06171955108642578, 0.061693984985351565, 0.061719520568847656, 0.06171955108642578, 0.061679615020751956, 0.06182809448242187, 0.06170521545410156, 0.061693950653076174, 0.06171340942382812, 0.06160179138183594, 0.06178508758544922, 0.06174822235107422, 0.061707393646240234, 0.061671295166015626, 0.06169721603393555, 0.06164358520507812, 0.06166527938842774, 0.061709312438964846, 0.13617971801757814, 0.06163046264648438, 0.06165401458740234, 0.06161407852172852, 0.06157516860961914, 0.061656063079833984, 0.06166016006469727, 0.061707263946533204, 0.061707263946533204, 0.06162944030761719, 0.06173388671875, 0.06161932754516602, 0.06166515350341797, 0.061652992248535154, 0.06172979354858398, 0.06173798370361328, 0.06173183822631836, 0.06171340942382812, 0.06171750259399414, 0.061639678955078124, 0.06162432098388672, 0.06173183822631836, 0.06170828628540039, 0.06276300811767578, 0.061687808990478515, 0.061636608123779295, 0.06173183822631836, 0.06161612701416016, 0.0616806411743164, 0.06161203384399414, 0.06170009613037109, 0.06170316696166992, 0.06180966567993164, 0.061895679473876954, 0.06174105453491211, 0.061710334777832034, 0.06173081588745117, 0.06171660614013672, 0.061677440643310544, 0.061674495697021485, 0.0617441291809082, 0.061692928314208986, 0.0617625617980957, 0.06161612701416016, 0.06178611373901367, 0.06169926452636719, 0.061711166381835936, 0.0628408317565918, 0.06211075210571289, 0.062286815643310546, 0.06192435073852539, 0.061740032196044924, 0.0617534065246582, 0.06165497589111328, 0.061690879821777345, 0.06172671890258789, 0.06164889526367188, 0.06174310302734375, 0.06171340942382812, 0.061661182403564455, 0.061671424865722656, 0.061740032196044924, 0.06165708923339844, 0.1359667205810547, 0.06163763046264648, 0.06169497680664063, 0.061628414154052735, 0.061633537292480466, 0.061764606475830076, 0.06156185531616211, 0.061704193115234375, 0.0617891845703125, 0.06193459320068359, 0.06180044937133789, 0.061663230895996096, 0.0622295036315918, 0.06165913772583008, 0.06168166351318359, 0.061690879821777345, 0.06174924850463867, 0.06168371200561523, 0.061770751953125, 0.0616519660949707, 0.06170111846923828, 0.06165913772583008, 0.061707263946533204, 0.0617441291809082, 0.06170009613037109, 0.06163148880004883, 0.06163558578491211, 0.061644798278808595, 0.06166425704956055, 0.06169190216064453, 0.061710334777832034, 0.06169497680664063, 0.061644798278808595, 0.06169497680664063, 0.061671520233154295, 0.06164572906494141, 0.06277836990356446, 0.06170009613037109, 0.06170009613037109, 0.06167859268188477, 0.06167660903930664, 0.06171231842041015, 0.06174924850463867, 0.06165094375610351, 0.06181478500366211, 0.061690879821777345, 0.06172671890258789, 0.06169190216064453, 0.06179020690917969, 0.06173388671875, 0.06178508758544922, 0.061685760498046874, 0.061876224517822265, 0.062241790771484375, 0.06170828628540039, 0.06173081588745117, 0.06157823944091797, 0.061774848937988285, 0.06169497680664063, 0.06177382278442383, 0.06168678283691406, 0.06180556869506836, 0.06197145462036133, 0.13539430236816405, 0.06167552185058594, 0.061671424865722656, 0.06164070510864258, 0.061591552734375, 0.06165401458740234, 0.06159667205810547, 0.06172876739501953, 0.06168678283691406, 0.06160281753540039, 0.061707263946533204, 0.06159667205810547, 0.0617154541015625, 0.06167244720458984, 0.061693950653076174, 0.0616519660949707, 0.06165094375610351, 0.06250291061401367, 0.06174720001220703, 0.06172159957885742, 0.06170009613037109, 0.061649921417236325, 0.061625343322753906, 0.06166732788085937, 0.06170111846923828, 0.061608959197998046, 0.06172774505615235, 0.06161305618286133, 0.06164889526367188, 0.061679615020751956, 0.06164070510864258, 0.06172265625, 0.061671390533447265, 0.061677566528320314, 0.061658111572265625, 0.061710334777832034, 0.06173183822631836, 0.061690879821777345, 0.061764606475830076, 0.06166835021972656, 0.06195308685302734, 0.06167443084716797, 0.06182809448242187, 0.06171244812011719, 0.061844417572021484, 0.06171750259399414, 0.06164582443237305, 0.06171340942382812, 0.06173593521118164, 0.06290636825561523, 0.06180249786376953, 0.06168678283691406, 0.061709312438964846, 0.06173183822631836, 0.061699073791503904, 0.061742080688476565, 0.061656063079833984, 0.061811710357666014, 0.061736961364746094, 0.061709312438964846, 0.061710334777832034, 0.061693950653076174, 0.06191308975219727, 0.13563392639160157, 0.0618526725769043, 0.061897727966308595, 0.06170316696166992, 0.06156803131103516, 0.061701087951660155, 0.06166128158569336, 0.06172867202758789, 0.0617154541015625, 0.061690879821777345, 0.06177996826171875, 0.06166527938842774, 0.06173183822631836, 0.061636608123779295, 0.0617523193359375, 0.06169702529907226, 0.0616673583984375, 0.061728736877441404, 0.061797374725341796, 0.06189875030517578, 0.0617441291809082, 0.06164275360107422, 0.061674495697021485, 0.06159360122680664, 0.06164796829223633, 0.06164368057250977, 0.061792255401611325, 0.06162739181518555, 0.061645023345947264, 0.06166812896728516, 0.06170115280151367, 0.06169699096679687, 0.061623294830322264, 0.06330368041992188, 0.06182092666625977, 0.06176051330566406, 0.06183628845214844, 0.06172774505615235, 0.06193971252441406, 0.061797374725341796, 0.0616888313293457, 0.06170329666137695, 0.06172966384887695, 0.06168166351318359, 0.06176371383666992, 0.061695873260498045, 0.061770751953125, 0.061723648071289064, 0.06177689743041992, 0.06173491287231445, 0.06175641632080078, 0.06165708923339844, 0.06165401458740234, 0.06177382278442383, 0.06164684677124024, 0.06175641632080078, 0.06164377593994141, 0.06168166351318359, 0.06175027084350586, 0.061699073791503904, 0.06170009613037109, 0.061710334777832034, 0.06166527938842774, 0.13643980407714842, 0.061716545104980466, 0.061607872009277344, 0.06166732788085937, 0.0616099853515625, 0.06168371200561523, 0.06158540725708008, 0.06174105453491211, 0.061677566528320314, 0.061644798278808595, 0.061795326232910154, 0.06170624160766602, 0.06171648025512695, 0.06165919876098633, 0.06173894500732422, 0.061732929229736326, 0.06168876647949219, 0.06168377685546875, 0.0616693115234375, 0.06164275360107422, 0.06164377593994141, 0.06178201675415039, 0.06199398422241211, 0.06298828887939453, 0.06168985748291016, 0.06167871856689453, 0.061652862548828125, 0.06170316696166992, 0.061723648071289064, 0.06167244720458984, 0.06170934295654297, 0.061823966979980466, 0.06195199966430664, 0.061687808990478515, 0.06180352020263672, 0.061764606475830076, 0.061685760498046874, 0.061674495697021485, 0.061693950653076174, 0.06170624160766602, 0.06166835021972656, 0.06171443176269531, 0.06177382278442383, 0.06163148880004883, 0.061774848937988285, 0.0616734733581543, 0.06170624160766602, 0.061707263946533204, 0.06180044937133789, 0.06171340942382812, 0.06172671890258789, 0.061739009857177736, 0.06166220855712891, 0.06182505416870117, 0.06167343902587891, 0.061732929229736326, 0.06168364715576172, 0.06287052917480469, 0.06172166442871094, 0.06172563171386719, 0.06172159957885742, 0.0617441291809082, 0.061709312438964846, 0.13548748779296876, 0.061685760498046874, 0.06184550476074219, 0.06171340942382812, 0.061636608123779295, 0.06169295883178711, 0.06171337509155273, 0.06171340942382812, 0.06171443176269531, 0.06175743865966797, 0.06177996826171875, 0.06164684677124024, 0.0617410888671875, 0.06167958450317383, 0.06168371200561523, 0.06167859268188477, 0.06231347274780273, 0.061669376373291014, 0.06177791976928711, 0.06170111846923828, 0.06169190216064453, 0.06170214462280273, 0.06165401458740234, 0.06170009613037109, 0.06169190216064453, 0.06161612701416016, 0.061682689666748045, 0.06170521545410156, 0.06170316696166992, 0.06166220855712891, 0.06175436782836914, 0.0616888313293457, 0.06170624160766602, 0.0616734733581543, 0.061725696563720706, 0.061830142974853515, 0.06189158248901367, 0.06173798370361328, 0.06172876739501953, 0.06174310302734375, 0.061740032196044924, 0.06168371200561523, 0.06173183822631836, 0.06170316696166992, 0.06178201675415039, 0.061722686767578125, 0.06168569564819336, 0.06171852874755859, 0.061712383270263675, 0.06175955200195313, 0.061817790985107424, 0.062246910095214845, 0.06236262512207031, 0.06189673614501953, 0.06172975921630859, 0.061764606475830076, 0.06163455963134765, 0.06180681610107422, 0.06167631912231445, 0.06176870346069336, 0.06178611373901367, 0.06176563262939453, 0.0617891845703125, 0.13594931030273438, 0.06170828628540039, 0.061732894897460935, 0.061699039459228514, 0.061656063079833984, 0.0616376953125, 0.06158432006835937, 0.06172467041015625, 0.06168473434448242, 0.06169497680664063, 0.06175641632080078, 0.06178620910644531, 0.061687713623046876, 0.0616806411743164, 0.06187007904052735, 0.06170214462280273, 0.06175641632080078, 0.06246201705932617, 0.06178400039672852, 0.061663230895996096, 0.06180352020263672, 0.06171750259399414, 0.06183731079101563, 0.061625343322753906, 0.061712383270263675, 0.061710334777832034, 0.06167244720458984, 0.06171443176269531, 0.06174723052978515, 0.06172361755371094, 0.061871135711669925, 0.061722591400146486, 0.06174617767333984, 0.061699073791503904, 0.061830142974853515, 0.06177996826171875, 0.061927425384521485, 0.06182195281982422, 0.061843456268310545, 0.061658206939697265, 0.06172662353515625, 0.06186086273193359, 0.06174310302734375, 0.06176153564453125, 0.06186905670166016, 0.06178815841674805, 0.06171340942382812, 0.06176563262939453, 0.061828289031982425, 0.0616824951171875, 0.061835262298583986, 0.0634439697265625, 0.06188544082641602, 0.06187519836425781, 0.06173081588745117, 0.06184243011474609, 0.06188544082641602, 0.06187212753295898, 0.06179020690917969, 0.06177587127685547, 0.0616888313293457, 0.06170521545410156, 0.0617441291809082, 0.13600973510742187, 0.062129150390625, 0.06186905670166016, 0.0616888313293457, 0.06164889526367188, 0.061658111572265625, 0.06158335876464844, 0.06171852874755859, 0.06171340942382812, 0.06164684677124024, 0.061730880737304684, 0.06165702438354492, 0.06167552185058594, 0.061641727447509766, 0.061739009857177736, 0.06163455963134765, 0.06174617767333984, 0.061677566528320314, 0.06170828628540039, 0.061725696563720706, 0.061671424865722656, 0.06169702529907226, 0.06174515151977539, 0.061661182403564455, 0.06163046264648438, 0.06156185531616211, 0.06298624038696289, 0.06177280044555664, 0.06197350311279297, 0.061710334777832034, 0.06174105453491211, 0.061628414154052735, 0.06164787292480469, 0.0615997428894043, 0.061770751953125, 0.06168678283691406, 0.06174515151977539, 0.06170111846923828, 0.06171955108642578, 0.061655040740966796, 0.06164377593994141, 0.06171852874755859, 0.061695999145507815, 0.061661182403564455, 0.06176153564453125, 0.06170316696166992, 0.061671424865722656, 0.061693023681640625, 0.061720481872558595, 0.06173593521118164, 0.06173081588745117, 0.061682689666748045, 0.061692928314208986, 0.06175539016723633, 0.061644798278808595, 0.06165708923339844, 0.06165404891967773, 0.06168368148803711, 0.06167552185058594, 0.061703231811523436, 0.061744064331054685, 0.06178303909301758, 0.06170316696166992]",tokens/s,15.919104066954871,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1490.198528,1841.823744,0.0,1195.37664,1022.155264,s,10,0.6871841278076172,0.06871841278076171,0.00033427755700674906,0.06862833786010741,0.06882823104858399,0.06926301155090332,0.06961083595275878,"[0.06969779205322266, 0.06856297302246094, 0.06858707427978515, 0.06850640106201172, 0.0686982421875, 0.06862300872802735, 0.0686336669921875, 0.06864895629882813, 0.0687316131591797, 0.06849440002441406]",tokens/s,3725.347976483958,kWh,8.097485796501648e-07,4.4370366405853753e-07,3.761924622931496e-06,5.0153768666401985e-06,tokens/kWh,51043023.64649507,MB,1490.198528,1841.823744,0.0,1195.37664,1091.597312,s,10,35.169312255859374,3.5169312255859375,0.00128683836079331,3.516599853515625,3.5173736328125,3.5190165283203125,3.5203308447265624,"[3.520659423828125, 3.51630859375, 3.515842041015625, 3.516979248046875, 3.5166650390625, 3.51653466796875, 3.516714599609375, 3.516245361328125, 3.517008544921875, 3.516354736328125]",tokens/s,17.91334432179688,kWh,4.1521653622797086e-05,2.2756081263433605e-05,0.0001967916524864682,0.0002610693873726989,tokens/kWh,241315.15622726805,,s,629,35.7307012748718,0.056805566414740584,0.008168059089268582,0.05580799865722656,0.055887052917480466,0.055967333984375,0.12452544342041016,"[0.05626367950439453, 0.05630361557006836, 0.05626367950439453, 0.056338432312011716, 0.05623910522460938, 0.05629337692260742, 0.05628313446044922, 0.056346622467041016, 0.05584896087646484, 0.05588991928100586, 0.05574758529663086, 0.055757823944091796, 0.05575884628295898, 0.055839744567871094, 0.05575680160522461, 0.055742462158203124, 0.0557209587097168, 0.05613260650634765, 0.05585203170776367, 0.0557844467163086, 0.055812095642089846, 0.05583769607543945, 0.05573529434204102, 0.05577523040771484, 0.05584588623046875, 0.05580595016479492, 0.0557946891784668, 0.055825408935546876, 0.05578137588500977, 0.05588684844970703, 0.055812095642089846, 0.055757823944091796, 0.05581414413452149, 0.05580595016479492, 0.05578137588500977, 0.05587353515625, 0.05582745742797852, 0.055779327392578126, 0.05571379089355469, 0.05586943817138672, 0.055806976318359375, 0.05578137588500977, 0.05583871841430664, 0.055760894775390625, 0.05582233428955078, 0.05581721496582031, 0.05575372695922851, 0.05586022567749024, 0.055790592193603515, 0.0557946891784668, 0.05580492782592773, 0.055793663024902344, 0.055809024810791016, 0.055818241119384764, 0.05578342437744141, 0.05583462524414062, 0.0558653450012207, 0.055798782348632815, 0.05585715103149414, 0.05593804931640625, 0.055790592193603515, 0.055820289611816405, 0.12452658843994141, 0.055755775451660154, 0.055787521362304686, 0.05579776000976563, 0.05577830505371094, 0.05574758529663086, 0.05581414413452149, 0.055787521362304686, 0.05573017501831055, 0.05575680160522461, 0.05580492782592773, 0.05572915267944336, 0.05578649520874023, 0.05576806259155274, 0.0557946891784668, 0.05575167846679688, 0.05578956985473633, 0.05579673767089844, 0.05581926345825195, 0.05581721496582031, 0.05590323257446289, 0.05579776000976563, 0.05584588623046875, 0.055739391326904295, 0.055790592193603515, 0.05578137588500977, 0.055836673736572265, 0.05580083084106445, 0.055825408935546876, 0.05577011108398437, 0.055809024810791016, 0.05585715103149414, 0.055793663024902344, 0.05574758529663086, 0.055787521362304686, 0.05576499176025391, 0.055809024810791016, 0.055766014099121096, 0.05583462524414062, 0.05594828796386719, 0.05584896087646484, 0.05576294326782227, 0.05584281539916992, 0.05577830505371094, 0.055858177185058595, 0.0560076789855957, 0.05585408020019531, 0.05578342437744141, 0.05575065612792969, 0.055859233856201174, 0.05579670333862305, 0.055782398223876956, 0.05586739349365234, 0.05582950210571289, 0.055809024810791016, 0.05578342437744141, 0.0558551025390625, 0.0561715202331543, 0.055932926177978515, 0.05579776000976563, 0.05574553680419922, 0.0558131217956543, 0.05583052825927735, 0.12455526733398438, 0.0557916145324707, 0.05582950210571289, 0.05575065612792969, 0.05580799865722656, 0.05575987243652344, 0.0558131217956543, 0.055723007202148435, 0.05578137588500977, 0.055793663024902344, 0.05577523040771484, 0.05576806259155274, 0.055795711517333986, 0.05574655914306641, 0.05582745742797852, 0.05574860763549805, 0.05578137588500977, 0.05572608184814453, 0.05576806259155274, 0.055809024810791016, 0.0557916145324707, 0.05582438278198242, 0.05580083084106445, 0.05575475311279297, 0.05576192092895508, 0.05580083084106445, 0.055844863891601565, 0.05580492782592773, 0.055823360443115234, 0.05579673767089844, 0.05576294326782227, 0.055785472869873044, 0.05581619262695312, 0.05576806259155274, 0.05587251281738281, 0.05579776000976563, 0.0557844467163086, 0.05576294326782227, 0.055839744567871094, 0.05573017501831055, 0.055815166473388675, 0.0558766098022461, 0.05590220642089844, 0.05579776000976563, 0.055818241119384764, 0.055841793060302736, 0.05584588623046875, 0.05583564758300781, 0.05581619262695312, 0.05581619262695312, 0.05587148666381836, 0.05586636734008789, 0.055820289611816405, 0.05580595016479492, 0.05584998321533203, 0.055793663024902344, 0.055785472869873044, 0.05583052825927735, 0.055785472869873044, 0.0557916145324707, 0.05583564758300781, 0.05590323257446289, 0.05586636734008789, 0.12449382019042969, 0.05575372695922851, 0.05576499176025391, 0.05578649520874023, 0.05577318572998047, 0.05608038330078125, 0.0558766098022461, 0.055744510650634765, 0.055757823944091796, 0.05576704025268555, 0.05586739349365234, 0.0557762565612793, 0.05626163101196289, 0.05574860763549805, 0.05581107330322266, 0.05573836898803711, 0.0557916145324707, 0.05576396942138672, 0.05574348831176758, 0.05576499176025391, 0.0557844467163086, 0.05575475311279297, 0.05585203170776367, 0.05574041748046875, 0.05575884628295898, 0.05578342437744141, 0.05584998321533203, 0.0558131217956543, 0.05589913558959961, 0.05588684844970703, 0.05580492782592773, 0.0557762565612793, 0.05579776000976563, 0.05576806259155274, 0.05586431884765625, 0.0557916145324707, 0.05583052825927735, 0.055790592193603515, 0.05582438278198242, 0.05580799865722656, 0.055820289611816405, 0.05578137588500977, 0.055798782348632815, 0.0558551025390625, 0.055766014099121096, 0.055825408935546876, 0.055982078552246094, 0.055766014099121096, 0.055793663024902344, 0.05594214248657227, 0.05587251281738281, 0.055806976318359375, 0.05583257675170898, 0.055965694427490234, 0.05590630340576172, 0.0557946891784668, 0.05584588623046875, 0.05582438278198242, 0.05581926345825195, 0.05581414413452149, 0.055809024810791016, 0.055823360443115234, 0.05583564758300781, 0.12459212493896485, 0.05578956985473633, 0.05584588623046875, 0.055777278900146485, 0.055769088745117185, 0.05573734283447265, 0.055818241119384764, 0.055785472869873044, 0.05583052825927735, 0.0557844467163086, 0.05582438278198242, 0.055782398223876956, 0.05585100936889648, 0.05584281539916992, 0.05582131195068359, 0.05576499176025391, 0.0557844467163086, 0.055757823944091796, 0.05580287933349609, 0.055757823944091796, 0.055841793060302736, 0.05579673767089844, 0.055818241119384764, 0.055736320495605465, 0.055801856994628904, 0.05581721496582031, 0.05588582229614258, 0.05578649520874023, 0.05585919952392578, 0.055790592193603515, 0.055836673736572265, 0.05586329650878906, 0.055874561309814455, 0.0558653450012207, 0.055793663024902344, 0.05587046432495117, 0.05577523040771484, 0.05574860763549805, 0.055839744567871094, 0.05576396942138672, 0.05587558364868164, 0.05575884628295898, 0.055861248016357425, 0.055810047149658204, 0.056005630493164066, 0.055798782348632815, 0.055818241119384764, 0.055913471221923826, 0.05584691238403321, 0.05581721496582031, 0.05577830505371094, 0.05587865447998047, 0.05581414413452149, 0.05581414413452149, 0.0557916145324707, 0.05582339096069336, 0.05581923294067383, 0.05581107330322266, 0.055856128692626954, 0.05586636734008789, 0.05587148666381836, 0.055801856994628904, 0.05587558364868164, 0.12469657897949218, 0.055790592193603515, 0.055731201171875, 0.05574860763549805, 0.055801856994628904, 0.055787521362304686, 0.0557844467163086, 0.0557496337890625, 0.055774208068847655, 0.0557844467163086, 0.05589913558959961, 0.055757823944091796, 0.055774208068847655, 0.05576806259155274, 0.05582745742797852, 0.05572710418701172, 0.055798782348632815, 0.055744510650634765, 0.055760894775390625, 0.05584281539916992, 0.05578035354614258, 0.05584384155273438, 0.05585919952392578, 0.05575680160522461, 0.055766014099121096, 0.05579673767089844, 0.055812095642089846, 0.0557946891784668, 0.05587353515625, 0.055825408935546876, 0.055828479766845705, 0.055782398223876956, 0.055831550598144535, 0.055787521362304686, 0.055877632141113284, 0.055788543701171874, 0.05577318572998047, 0.0557916145324707, 0.055841793060302736, 0.055771137237548826, 0.055823360443115234, 0.05577318572998047, 0.05582131195068359, 0.055801856994628904, 0.05582950210571289, 0.055844863891601565, 0.05596672058105469, 0.055787521362304686, 0.05583052825927735, 0.055897087097167966, 0.055853057861328125, 0.05585100936889648, 0.05584896087646484, 0.05590220642089844, 0.055806976318359375, 0.056011775970458984, 0.05590118408203125, 0.05583564758300781, 0.055809024810791016, 0.055826431274414064, 0.055774208068847655, 0.055810047149658204, 0.055856128692626954, 0.12463922882080078, 0.05575680160522461, 0.055785472869873044, 0.05579673767089844, 0.05581107330322266, 0.05583257675170898, 0.055858177185058595, 0.055795711517333986, 0.055752704620361325, 0.055806976318359375, 0.055839744567871094, 0.05576294326782227, 0.05579980850219726, 0.05575065612792969, 0.055801856994628904, 0.05574860763549805, 0.05582438278198242, 0.055736320495605465, 0.05581107330322266, 0.05579673767089844, 0.055809024810791016, 0.055801856994628904, 0.05583769607543945, 0.05578956985473633, 0.05581619262695312, 0.05584998321533203, 0.05587046432495117, 0.05583359909057617, 0.055967742919921876, 0.055815166473388675, 0.05589503860473633, 0.055825408935546876, 0.05581619262695312, 0.05591551971435547, 0.055801856994628904, 0.05583769607543945, 0.05586329650878906, 0.05574860763549805, 0.05582131195068359, 0.05574041748046875, 0.05583769607543945, 0.055839744567871094, 0.05586841583251953, 0.05581721496582031, 0.05584588623046875, 0.05578137588500977, 0.055820289611816405, 0.055841793060302736, 0.05583462524414062, 0.0557916145324707, 0.055788543701171874, 0.05590118408203125, 0.055795711517333986, 0.05579673767089844, 0.0557844467163086, 0.05580492782592773, 0.055831550598144535, 0.05579264068603516, 0.0558551025390625, 0.05587251281738281, 0.055856128692626954, 0.05582032012939453, 0.05584585571289063, 0.12463410949707031, 0.05575065612792969, 0.05573734283447265, 0.05575475311279297, 0.05578342437744141, 0.055742462158203124, 0.055790592193603515, 0.05573734283447265, 0.05578649520874023, 0.05577830505371094, 0.0558551025390625, 0.055788543701171874, 0.0557946891784668, 0.055787521362304686, 0.055820289611816405, 0.055739391326904295, 0.055820289611816405, 0.05573427200317383, 0.05605068969726563, 0.055803905487060546, 0.05579776000976563, 0.055766014099121096, 0.055853057861328125, 0.055741439819335936, 0.055752704620361325, 0.05580083084106445, 0.05584076690673828, 0.0558131217956543, 0.05589606475830078, 0.05578137588500977, 0.05597081756591797, 0.055836673736572265, 0.05578956985473633, 0.05579980850219726, 0.05587353515625, 0.05579673767089844, 0.055774208068847655, 0.05579980850219726, 0.055815166473388675, 0.05575372695922851, 0.05580492782592773, 0.05586022567749024, 0.055844863891601565, 0.055806976318359375, 0.055806976318359375, 0.055812095642089846, 0.05589606475830078, 0.055772159576416014, 0.05577318572998047, 0.05588684844970703, 0.05584793472290039, 0.05581721496582031, 0.055810047149658204, 0.05586329650878906, 0.05579673767089844, 0.055772159576416014, 0.05586227035522461, 0.05584281539916992, 0.055785472869873044, 0.05581619262695312, 0.0557946891784668, 0.05583257675170898, 0.05588787078857422, 0.1247088623046875, 0.05576806259155274, 0.05583462524414062, 0.055766014099121096, 0.055772159576416014, 0.05578956985473633, 0.05581414413452149, 0.05579673767089844, 0.055742462158203124, 0.0557762565612793, 0.05581619262695312, 0.05573529434204102, 0.05583564758300781, 0.05576704025268555, 0.055815166473388675, 0.05571583938598633, 0.05580595016479492, 0.05576704025268555, 0.055790592193603515, 0.055782398223876956, 0.05589811325073242, 0.0557946891784668, 0.055809024810791016, 0.05574758529663086, 0.055798782348632815, 0.05577830505371094, 0.05583052825927735, 0.05578137588500977, 0.05585408020019531, 0.055790592193603515, 0.05580799865722656, 0.05587251281738281, 0.0557844467163086, 0.055795711517333986, 0.05579673767089844, 0.05584588623046875, 0.055782398223876956, 0.055806976318359375, 0.05589606475830078, 0.055777278900146485, 0.055844863891601565, 0.0557762565612793, 0.056202239990234375, 0.05586431884765625, 0.05582438278198242, 0.05581414413452149, 0.055820289611816405, 0.05581926345825195, 0.055769088745117185, 0.05581619262695312, 0.05580595016479492, 0.05586943817138672, 0.05609164810180664, 0.055853057861328125, 0.0558551025390625, 0.055856128692626954, 0.056139774322509765, 0.05579980850219726, 0.05579264068603516, 0.05577318572998047, 0.05584384155273438, 0.055788543701171874, 0.05585408020019531, 0.12452249908447266, 0.055742462158203124, 0.05573734283447265, 0.055742462158203124, 0.055787521362304686, 0.05576806259155274, 0.0557946891784668, 0.05572608184814453, 0.055795711517333986, 0.055779327392578126, 0.055890945434570315, 0.055760894775390625, 0.05575475311279297, 0.05573222351074219, 0.055815166473388675, 0.055723007202148435, 0.055858177185058595, 0.05573529434204102, 0.055725055694580077, 0.055790592193603515, 0.05591961669921875, 0.05581619262695312, 0.05587865447998047, 0.05575884628295898, 0.05578956985473633, 0.05578956985473633, 0.05581414413452149, 0.055790592193603515, 0.05590937423706055, 0.05582438278198242, 0.055806976318359375, 0.05578342437744141, 0.055907329559326174, 0.055831550598144535, 0.05587558364868164, 0.055785472869873044, 0.05579264068603516, 0.055812095642089846, 0.05582950210571289, 0.05575372695922851, 0.05583871841430664, 0.0558551025390625, 0.055823360443115234, 0.05580287933349609, 0.0557946891784668, 0.05583359909057617, 0.05591756820678711, 0.055779327392578126, 0.055826431274414064, 0.05596160125732422, 0.055982078552246094, 0.055801856994628904, 0.05582233428955078, 0.05587251281738281, 0.05585100936889648, 0.0557762565612793, 0.055826431274414064, 0.055815166473388675, 0.05581926345825195, 0.05585919952392578, 0.05578137588500977, 0.05581619262695312, 0.05582950210571289]",tokens/s,17.60390861520409,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12276,7 +12276,7 @@ If this is a private repository, make sure to pass a token having permission to 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1488.359424,1579.679744,0.0,933.23264,856.69632,s,10,0.4187196159362792,0.041871961593627924,0.0014808966097482894,0.04126091194152832,0.043131636810302736,0.0445920425415039,0.04576036712646484,"[0.04605244827270508, 0.04131564712524414, 0.0411776008605957, 0.04146476745605469, 0.04118272018432617, 0.04280710220336914, 0.0412061767578125, 0.0415596809387207, 0.04101833724975586, 0.04093513488769531]",tokens/s,6113.876452326468,kWh,4.934550676908758e-07,2.7027020769764927e-07,1.3917580578494868e-06,2.155483333238012e-06,tokens/kWh,118766865.90539834,MB,1488.359424,1579.679744,0.0,933.23264,899.62496,s,10,24.019910400390625,2.4019910400390625,0.019429569356012314,2.394274169921875,2.414185791015625,2.436057373046875,2.453554638671875,"[2.457928955078125, 2.409325439453125, 2.399566650390625, 2.391402587890625, 2.39863720703125, 2.391660400390625, 2.391578369140625, 2.391282958984375, 2.391639892578125, 2.396887939453125]",tokens/s,26.228241050797365,kWh,2.945148775767949e-05,1.6140594842595416e-05,7.348580184415957e-05,0.00011907788444443448,tokens/kWh,529065.4960316985,,s,629,24.35256624984743,0.03871632154188778,0.004963175047223502,0.037943294525146484,0.03889213485717773,0.03936890869140625,0.079117353515625,"[0.040226814270019534, 0.04027084732055664, 0.040318977355957034, 0.03974041748046875, 0.040825855255126955, 0.04090367889404297, 0.03897753524780274, 0.039828479766845705, 0.04030361557006836, 0.03983871841430664, 0.04004761505126953, 0.04024524688720703, 0.03980287933349609, 0.039346176147460936, 0.038795265197753906, 0.039277568817138675, 0.03892326354980469, 0.03963289642333984, 0.039182334899902346, 0.038902782440185545, 0.039244800567626956, 0.03907788848876953, 0.0392806396484375, 0.03914035034179687, 0.03907788848876953, 0.03934003067016602, 0.03828940963745117, 0.038212608337402344, 0.03859251022338867, 0.038125568389892575, 0.03932876968383789, 0.03894169616699219, 0.03829043197631836, 0.03913216018676758, 0.03880038452148438, 0.038793216705322264, 0.039011329650878904, 0.039003135681152344, 0.039147518157958985, 0.038547454833984376, 0.03806617736816406, 0.03796889495849609, 0.03785420989990235, 0.03794432067871094, 0.03905023956298828, 0.03965235137939453, 0.03999129486083984, 0.039384063720703126, 0.039397377014160156, 0.0378419189453125, 0.03808563232421875, 0.037797889709472655, 0.037797889709472655, 0.03783987045288086, 0.037891071319580076, 0.03788083267211914, 0.03912499237060547, 0.038760448455810545, 0.03786342239379883, 0.038950912475585936, 0.0378419189453125, 0.03783270263671875, 0.0792985610961914, 0.03790643310546875, 0.03781017684936523, 0.038370304107666016, 0.03831193542480469, 0.03775078582763672, 0.03785420989990235, 0.038435840606689455, 0.038365184783935545, 0.03781017684936523, 0.037905406951904294, 0.03861094284057617, 0.03815731048583984, 0.037917697906494144, 0.037776382446289065, 0.03875532913208008, 0.03827711868286133, 0.03894784164428711, 0.0380313606262207, 0.038877185821533204, 0.037923839569091795, 0.03808870315551758, 0.03812351989746094, 0.03917004776000976, 0.0390645751953125, 0.03829862213134766, 0.03843379211425781, 0.038542335510253906, 0.03800166320800781, 0.0378869743347168, 0.038740993499755856, 0.03850239944458008, 0.03803955078125, 0.038438911437988284, 0.03892838287353516, 0.03817267227172851, 0.038012928009033206, 0.03801190567016602, 0.039065601348876954, 0.03803033447265625, 0.03791462326049805, 0.03789619064331055, 0.03887923049926758, 0.038371326446533204, 0.03790848159790039, 0.037904384613037106, 0.03786956787109375, 0.037988353729248046, 0.037943294525146484, 0.03816755294799805, 0.03861503982543945, 0.03803340911865234, 0.037966846466064456, 0.03835084915161133, 0.03848396682739258, 0.03804159927368164, 0.03789926528930664, 0.038539264678955076, 0.038340606689453126, 0.03802521514892578, 0.03793203353881836, 0.038679550170898434, 0.03826892852783203, 0.07922994995117187, 0.03890790557861328, 0.03783679962158203, 0.03792281723022461, 0.03792588806152344, 0.03827507019042969, 0.03808153533935547, 0.03808768081665039, 0.03786137771606445, 0.03799552154541016, 0.0380313606262207, 0.03805388641357422, 0.03804569625854492, 0.03840409469604492, 0.03806412887573242, 0.037905406951904294, 0.03795455932617187, 0.03794636917114258, 0.03813478469848633, 0.03800985717773438, 0.03793203353881836, 0.03811123275756836, 0.038125568389892575, 0.03792793655395508, 0.03794739151000977, 0.03824435043334961, 0.0380847053527832, 0.0379257926940918, 0.03787161636352539, 0.037819393157958986, 0.03803647994995117, 0.037956607818603515, 0.039212032318115236, 0.03805184173583984, 0.0379504623413086, 0.0379156494140625, 0.04073574447631836, 0.038391807556152346, 0.03786649703979492, 0.03888947296142578, 0.0400711669921875, 0.03781222534179687, 0.03783679962158203, 0.037981182098388674, 0.03775078582763672, 0.03779891204833984, 0.03775590515136719, 0.0378419189453125, 0.03782144165039063, 0.037781505584716796, 0.03789311981201172, 0.03784396743774414, 0.03780710220336914, 0.038095870971679685, 0.037797889709472655, 0.03782451248168945, 0.03784601593017578, 0.03794124984741211, 0.037806079864501956, 0.037781505584716796, 0.037884929656982425, 0.03780710220336914, 0.03832627105712891, 0.07955046081542969, 0.03795251083374023, 0.03794947052001953, 0.03773539352416992, 0.03782860946655273, 0.03780710220336914, 0.037874687194824216, 0.037789695739746096, 0.037867584228515626, 0.03773331069946289, 0.038191104888916014, 0.03783679962158203, 0.037872638702392575, 0.03788185501098633, 0.037850112915039064, 0.0378869743347168, 0.03800371170043945, 0.037870655059814455, 0.038172607421875, 0.037921791076660154, 0.03788390350341797, 0.03808256149291992, 0.03796377563476563, 0.037819393157958986, 0.03782553482055664, 0.03784806442260742, 0.038128639221191404, 0.038128639221191404, 0.03777228927612305, 0.03794944000244141, 0.03797196960449219, 0.03817062377929688, 0.03783168029785156, 0.037855232238769534, 0.037928958892822266, 0.03805286407470703, 0.0382371826171875, 0.03802828979492188, 0.037988353729248046, 0.037961727142333986, 0.038179840087890625, 0.03786240005493164, 0.037978111267089845, 0.0379156494140625, 0.038059009552001956, 0.038112255096435545, 0.037910526275634765, 0.03792588806152344, 0.03811942291259766, 0.03815935897827148, 0.037978111267089845, 0.037874687194824216, 0.03809996795654297, 0.03806924819946289, 0.03789311981201172, 0.037894176483154296, 0.037800926208496094, 0.038100990295410156, 0.037956607818603515, 0.037953536987304685, 0.037889022827148434, 0.03792899322509766, 0.03823203277587891, 0.07917670440673828, 0.03791462326049805, 0.03787366485595703, 0.03793305587768555, 0.03789926528930664, 0.03786240005493164, 0.038117374420166016, 0.03861913681030273, 0.03791974258422852, 0.038007808685302735, 0.03791257476806641, 0.03793817520141601, 0.03802521514892578, 0.0379422721862793, 0.03785830307006836, 0.038046718597412106, 0.037959678649902344, 0.037958656311035156, 0.03790028762817383, 0.037817344665527344, 0.03794124984741211, 0.03805184173583984, 0.03784294509887695, 0.03805388641357422, 0.03809894561767578, 0.03805286407470703, 0.03784499359130859, 0.03791974258422852, 0.03803955078125, 0.03802828979492188, 0.038161407470703124, 0.03802214431762695, 0.037894142150878905, 0.03818598556518555, 0.037917697906494144, 0.03795455932617187, 0.037975040435791016, 0.03828224182128906, 0.038002689361572264, 0.037910526275634765, 0.03793100738525391, 0.03803852844238281, 0.03797708892822266, 0.03793612670898437, 0.0378869743347168, 0.0380313606262207, 0.03825971221923828, 0.037953536987304685, 0.03790643310546875, 0.038029312133789066, 0.03808153533935547, 0.03808870315551758, 0.0379791374206543, 0.03979673767089844, 0.03898777770996094, 0.03879116821289062, 0.037969921112060545, 0.03789926528930664, 0.03830579376220703, 0.038193153381347655, 0.038182910919189454, 0.03849523162841797, 0.03826278305053711, 0.08257740783691406, 0.03822489547729492, 0.03793920135498047, 0.03778559875488281, 0.03784499359130859, 0.037872638702392575, 0.038141952514648435, 0.037921791076660154, 0.03786342239379883, 0.03808256149291992, 0.03808051300048828, 0.0378419189453125, 0.03813273620605469, 0.038711296081542966, 0.038114303588867186, 0.03786240005493164, 0.03781119918823242, 0.03850137710571289, 0.03799347305297852, 0.0378152961730957, 0.0378869743347168, 0.037989376068115234, 0.03785728073120117, 0.037850112915039064, 0.0379607048034668, 0.03787782287597656, 0.03834259033203125, 0.03791462326049805, 0.03792588806152344, 0.03778047943115234, 0.038484992980957033, 0.03798015975952149, 0.03783782577514649, 0.03778559875488281, 0.03781017684936523, 0.0379791374206543, 0.03785625457763672, 0.03782553482055664, 0.0378603515625, 0.037923839569091795, 0.0378152961730957, 0.03780505752563477, 0.03792998504638672, 0.03794124984741211, 0.03778355026245117, 0.03781119918823242, 0.03775385665893555, 0.03779891204833984, 0.03803238296508789, 0.03786956787109375, 0.03816960144042969, 0.037814273834228515, 0.03785625457763672, 0.038043647766113284, 0.03777849578857422, 0.037824447631835935, 0.037713920593261716, 0.03772108840942383, 0.03790233612060547, 0.0377784309387207, 0.037776382446289065, 0.0378152961730957, 0.037921791076660154, 0.07919411468505859, 0.03782758331298828, 0.03773747253417969, 0.03775385665893555, 0.03801804733276367, 0.037800960540771485, 0.03785728073120117, 0.03783679962158203, 0.037719039916992186, 0.038024192810058595, 0.03775590515136719, 0.037768192291259765, 0.037823486328125, 0.037910526275634765, 0.03777433776855469, 0.03780505752563477, 0.03785113525390625, 0.0377968635559082, 0.037874687194824216, 0.03781222534179687, 0.03783475112915039, 0.03781119918823242, 0.037754878997802735, 0.03797094345092773, 0.03784499359130859, 0.03780710220336914, 0.03785728073120117, 0.03789209747314453, 0.0379791374206543, 0.03779891204833984, 0.03774771118164062, 0.03775078582763672, 0.03789926528930664, 0.03783987045288086, 0.037738529205322266, 0.037819358825683595, 0.03775795364379883, 0.037889022827148434, 0.037773311614990236, 0.03777740859985351, 0.03778662490844727, 0.03790950393676758, 0.03822079849243164, 0.04011008071899414, 0.03838054275512695, 0.0379576301574707, 0.03801599884033203, 0.03980595016479492, 0.03857920074462891, 0.03786956787109375, 0.03796384048461914, 0.03790534210205078, 0.03787059020996094, 0.037891071319580076, 0.037838848114013675, 0.03804774475097656, 0.03783270263671875, 0.03785113525390625, 0.038819839477539066, 0.038046718597412106, 0.03789823913574219, 0.03790233612060547, 0.03783679962158203, 0.08027750396728515, 0.038125568389892575, 0.03789619064331055, 0.03792588806152344, 0.03793100738525391, 0.03866726303100586, 0.03827916717529297, 0.03803443145751953, 0.03789619064331055, 0.03781222534179687, 0.03788185501098633, 0.03791257476806641, 0.03783065414428711, 0.03784505462646484, 0.03773846435546875, 0.03788899230957031, 0.03809280014038086, 0.03783475112915039, 0.03798323059082031, 0.03775385665893555, 0.037966846466064456, 0.038214656829833986, 0.03794124984741211, 0.03781119918823242, 0.037868545532226565, 0.03785113525390625, 0.03793100738525391, 0.0380211181640625, 0.03793817520141601, 0.037855232238769534, 0.03802828979492188, 0.0378081283569336, 0.03788185501098633, 0.03851264190673828, 0.03827097702026367, 0.03790643310546875, 0.03789619064331055, 0.03783065414428711, 0.037907455444335936, 0.03800985717773438, 0.03793920135498047, 0.03778867340087891, 0.03791872024536133, 0.037806079864501956, 0.03793920135498047, 0.03792793655395508, 0.037855232238769534, 0.03791974258422852, 0.03787161636352539, 0.03796582412719727, 0.038024192810058595, 0.03782860946655273, 0.03795251083374023, 0.03783782577514649, 0.037956607818603515, 0.03788083267211914, 0.03803964614868164, 0.03777833557128906, 0.037766143798828124, 0.03796275329589844, 0.03784703826904297, 0.03799859237670898, 0.03801804733276367, 0.07892582702636719, 0.03798527908325195, 0.037719039916992186, 0.03776921463012695, 0.03804467010498047, 0.037907455444335936, 0.03788288116455078, 0.03789516830444336, 0.03778867340087891, 0.037966846466064456, 0.03824844741821289, 0.03791155242919922, 0.03777228927612305, 0.03783065414428711, 0.03782758331298828, 0.03794739151000977, 0.03860070419311523, 0.0380948486328125, 0.037868545532226565, 0.03780505752563477, 0.0378152961730957, 0.037822463989257815, 0.03793407821655274, 0.03778771209716797, 0.03782342529296875, 0.037855232238769534, 0.03776921463012695, 0.038046718597412106, 0.037803009033203126, 0.03780198287963867, 0.037838848114013675, 0.03789516830444336, 0.03798015975952149, 0.0378603515625, 0.03887923049926758, 0.03785625457763672, 0.03788390350341797, 0.03847884750366211, 0.03800064086914062, 0.03786342239379883, 0.03780710220336914, 0.037826591491699216, 0.03794736099243164, 0.0378787841796875, 0.03783065414428711, 0.03801702499389648, 0.03784601593017578, 0.03804569625854492, 0.03787571334838867, 0.03788288116455078, 0.03784806442260742, 0.03797094345092773, 0.038019073486328124, 0.03792281723022461, 0.037956607818603515, 0.03784601593017578, 0.03790848159790039, 0.03802316665649414, 0.037868545532226565, 0.037868545532226565, 0.03931238555908203, 0.03824435043334961, 0.038171646118164065, 0.07896473693847657, 0.03796480178833008, 0.038031391143798825, 0.038429664611816405, 0.03807231903076172, 0.03814297485351562, 0.04003839874267578, 0.038354942321777344, 0.038171646118164065, 0.03779481506347656, 0.03799347305297852, 0.03781631851196289, 0.037822463989257815, 0.03803852844238281, 0.037795841217041014, 0.03802009582519531, 0.03805184173583984, 0.037812255859375, 0.038042591094970706, 0.0378787841796875, 0.03813785552978516, 0.03794636917114258, 0.038004737854003906, 0.03808153533935547, 0.03786240005493164, 0.03785318374633789, 0.03793817520141601, 0.0379156494140625, 0.03786137771606445, 0.03794124984741211, 0.037972991943359374, 0.03801599884033203, 0.03807743835449219, 0.038757377624511716, 0.038007808685302735, 0.03791155242919922, 0.03793407821655274, 0.03837849426269531, 0.038100990295410156, 0.037975040435791016, 0.037891071319580076, 0.03786751937866211, 0.03800678253173828, 0.03780505752563477, 0.03791360092163086, 0.037907455444335936, 0.03789209747314453, 0.03788288116455078, 0.037894142150878905, 0.03790335845947266, 0.0380497932434082, 0.0378787841796875, 0.0379504623413086, 0.03782963180541992, 0.037920768737792966, 0.03815731048583984, 0.03894476699829102, 0.03841228866577148, 0.0381102066040039, 0.03795455932617187, 0.03790848159790039, 0.03784396743774414, 0.03802521514892578]",tokens/s,25.828900065262776,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1279.995904,1017.643008,0.0,371.195904,277.410816,s,10,0.4175716514587402,0.04175716514587402,0.001329291485763711,0.04133342361450196,0.042088745498657223,0.04389397258758545,0.04533815425872803,"[0.045699199676513674, 0.04128799819946289, 0.04128377532958984, 0.04097267150878906, 0.04157187271118164, 0.04107036972045899, 0.04121916961669922, 0.04137884902954102, 0.04168758392333984, 0.04140016174316406]",tokens/s,6130.684377296505,kWh,4.643787057311447e-07,2.5445859858069414e-07,1.253729761153002e-06,1.972567065464841e-06,tokens/kWh,129780124.83427167,MB,1279.995904,1017.643008,0.0,371.195904,323.048448,s,10,24.960871826171868,2.496087182617187,0.007120166670465486,2.4940638427734374,2.4991390625000003,2.5077969482421874,2.5147232568359374,"[2.516454833984375, 2.49401806640625, 2.495659423828125, 2.49662109375, 2.49185595703125, 2.493477294921875, 2.4905234375, 2.49093701171875, 2.494109619140625, 2.497215087890625]",tokens/s,25.239503026470214,kWh,2.763858284111377e-05,1.5146819142039876e-05,6.740021405084962e-05,0.00011018561603400326,tokens/kWh,571762.4701626954,,s,629,25.29028705596926,0.04020713363429131,0.0049559405951164134,0.03952540969848633,0.039936614990234376,0.040598938751220705,0.08109604888916017,"[0.04100912094116211, 0.04127743911743164, 0.04177305603027344, 0.04085964965820312, 0.041427967071533206, 0.04092620849609375, 0.04130918502807617, 0.040700927734375, 0.042106880187988284, 0.04071731185913086, 0.040460289001464846, 0.04082995223999023, 0.03992473602294922, 0.039828479766845705, 0.03968511962890625, 0.039523326873779296, 0.03983257675170898, 0.039553024291992187, 0.039913471221923826, 0.039607295989990236, 0.03934515380859375, 0.03961139297485351, 0.03963187026977539, 0.039706623077392575, 0.03935334396362305, 0.03937484741210937, 0.03951923370361328, 0.039556095123291016, 0.03966873550415039, 0.03936972808837891, 0.03945574569702148, 0.039840831756591796, 0.0411473274230957, 0.039992321014404295, 0.03950080108642578, 0.0393994255065918, 0.03961958312988281, 0.039554046630859374, 0.03956838226318359, 0.039792640686035156, 0.03939740753173828, 0.03990217590332031, 0.03950284957885742, 0.03952537536621094, 0.039384063720703126, 0.039395328521728515, 0.0393809928894043, 0.03959091186523438, 0.03945676803588867, 0.039395328521728515, 0.03952947235107422, 0.03961139297485351, 0.03948441696166992, 0.0394967041015625, 0.03940147018432617, 0.039337982177734376, 0.04001484680175781, 0.03948953628540039, 0.03933695983886719, 0.039452671051025394, 0.03969023895263672, 0.03961958312988281, 0.08116633605957031, 0.03934003067016602, 0.03947724914550781, 0.03954687881469727, 0.03953664016723633, 0.039618560791015625, 0.039354366302490236, 0.039488510131835936, 0.039809024810791016, 0.03934310531616211, 0.039357440948486325, 0.03944243240356445, 0.03957657623291016, 0.03940761566162109, 0.039419902801513675, 0.03942399978637695, 0.0395489273071289, 0.039482368469238284, 0.03993907165527344, 0.03930419158935547, 0.0394967041015625, 0.03933491134643555, 0.039373825073242184, 0.03941785430908203, 0.039444480895996094, 0.039467041015625, 0.039427040100097656, 0.03946086502075195, 0.039478271484375, 0.039577598571777346, 0.03945471954345703, 0.039298046112060545, 0.039298046112060545, 0.03948339080810547, 0.039370750427246096, 0.03942502212524414, 0.03943833541870117, 0.039602176666259765, 0.03958988952636719, 0.03945881652832031, 0.039532543182373044, 0.04032716751098633, 0.04127542495727539, 0.040283103942871094, 0.03961139297485351, 0.03957657623291016, 0.03948953628540039, 0.03951615905761719, 0.039367679595947266, 0.03977833557128906, 0.039686111450195315, 0.04003123092651367, 0.04087910461425781, 0.040207359313964845, 0.03972403335571289, 0.03946393585205078, 0.039387134552001955, 0.039554046630859374, 0.03936972808837891, 0.03947724914550781, 0.0394005126953125, 0.03951507186889648, 0.03940662384033203, 0.08121955108642578, 0.039504894256591795, 0.039790592193603515, 0.03975270462036133, 0.03982131195068359, 0.039556129455566406, 0.039650272369384766, 0.03966259384155273, 0.03948339080810547, 0.03957145690917969, 0.03962265777587891, 0.039591934204101564, 0.03957452774047852, 0.03968716812133789, 0.03951103973388672, 0.039772159576416014, 0.03950592041015625, 0.039828479766845705, 0.03968819046020508, 0.039564289093017575, 0.039532543182373044, 0.03951615905761719, 0.03973324966430664, 0.03973222351074219, 0.03958476638793945, 0.03959807968139648, 0.03997183990478516, 0.03969740676879883, 0.03954380798339844, 0.0401080322265625, 0.03958784103393555, 0.03946188735961914, 0.03951718521118164, 0.03965542221069336, 0.03964416122436523, 0.039585792541503906, 0.039599105834960936, 0.039634944915771485, 0.0393891830444336, 0.03955712127685547, 0.03965542221069336, 0.03983161544799805, 0.03946899032592773, 0.039577598571777346, 0.039572479248046875, 0.03957555389404297, 0.039698463439941406, 0.03968406295776367, 0.039599105834960936, 0.03945779037475586, 0.03973529434204102, 0.038989822387695314, 0.03940147018432617, 0.03963904190063477, 0.03952435302734375, 0.03980287933349609, 0.039634944915771485, 0.03947724914550781, 0.03947417449951172, 0.0395786247253418, 0.039591934204101564, 0.03949465560913086, 0.03950899124145508, 0.08075161743164062, 0.03947520065307617, 0.03970457458496094, 0.03956329727172852, 0.03949052810668945, 0.03964108657836914, 0.03945369720458984, 0.03959091186523438, 0.03958476638793945, 0.03949465560913086, 0.03940147018432617, 0.03942502212524414, 0.039354366302490236, 0.039539710998535156, 0.03950387191772461, 0.03942297744750976, 0.039923713684082034, 0.03959603118896484, 0.03958169555664062, 0.03936870574951172, 0.039532543182373044, 0.039547904968261716, 0.039572479248046875, 0.03954687881469727, 0.03948748779296875, 0.03956531143188476, 0.04043366241455078, 0.040976383209228515, 0.039962623596191404, 0.039689216613769535, 0.039373825073242184, 0.03946188735961914, 0.03932364654541016, 0.03958476638793945, 0.039539710998535156, 0.03945471954345703, 0.039446529388427735, 0.039629825592041014, 0.039375873565673826, 0.03932467269897461, 0.039597057342529295, 0.040233985900878906, 0.04049612808227539, 0.03984896087646484, 0.039357440948486325, 0.03942195129394531, 0.03954483032226563, 0.039456832885742185, 0.03950073623657226, 0.03953049468994141, 0.03962060928344727, 0.039444480895996094, 0.039754753112792966, 0.03976806259155274, 0.04006911849975586, 0.039629825592041014, 0.039419902801513675, 0.03955199813842773, 0.03977830505371094, 0.039602176666259765, 0.040025089263916014, 0.03968819046020508, 0.03958169555664062, 0.08158719635009766, 0.03952844619750977, 0.03982643127441406, 0.03968000030517578, 0.03958476638793945, 0.03935334396362305, 0.039572479248046875, 0.040210430145263674, 0.03953561782836914, 0.039654399871826174, 0.03932057571411133, 0.03947417449951172, 0.03926323318481445, 0.039636993408203126, 0.039689216613769535, 0.03954278564453125, 0.039504894256591795, 0.0395601921081543, 0.03949055862426758, 0.0393963508605957, 0.03964518356323242, 0.039634944915771485, 0.039430145263671876, 0.03942297744750976, 0.03927347183227539, 0.039599105834960936, 0.039367679595947266, 0.039485439300537106, 0.039806976318359374, 0.040528896331787106, 0.039532543182373044, 0.039362560272216796, 0.03942502212524414, 0.039651329040527344, 0.03954483032226563, 0.03994521713256836, 0.040018943786621096, 0.03948748779296875, 0.039367679595947266, 0.039653377532958986, 0.03947110366821289, 0.03956633758544922, 0.03923660659790039, 0.039523326873779296, 0.03932876968383789, 0.039406593322753904, 0.039462913513183595, 0.03944140625, 0.03933184051513672, 0.039863296508789066, 0.03953664016723633, 0.039504894256591795, 0.039191551208496093, 0.040030208587646485, 0.03945881652832031, 0.03943324661254883, 0.03956835174560547, 0.0393891830444336, 0.03934822463989258, 0.039390209197998044, 0.03933292770385742, 0.03960108947753906, 0.03945881652832031, 0.08104959869384766, 0.03950592041015625, 0.03941891098022461, 0.039413726806640625, 0.03958988952636719, 0.03938102340698242, 0.03919769668579102, 0.03935843276977539, 0.03926323318481445, 0.03933184051513672, 0.03955507278442383, 0.039395328521728515, 0.0393963508605957, 0.039449600219726565, 0.03967795181274414, 0.039828479766845705, 0.039708671569824217, 0.0393994255065918, 0.03942195129394531, 0.03932160186767578, 0.039772159576416014, 0.0416184310913086, 0.03949363327026367, 0.03941273498535156, 0.0393114242553711, 0.039466945648193356, 0.03957555389404297, 0.039329792022705076, 0.039397377014160156, 0.03927859115600586, 0.039390239715576175, 0.039481311798095706, 0.03924889755249023, 0.039397377014160156, 0.03929600143432617, 0.039414783477783204, 0.039723007202148435, 0.039465984344482424, 0.04114329528808594, 0.04011625671386719, 0.039448543548583986, 0.03947315216064453, 0.03952537536621094, 0.039370750427246096, 0.03939123153686523, 0.03942604827880859, 0.039684097290039064, 0.03952742385864258, 0.040645633697509766, 0.039919647216796875, 0.03952329635620117, 0.03967078399658203, 0.03952540969848633, 0.03946083068847656, 0.039446529388427735, 0.03928268814086914, 0.03962879943847656, 0.039470142364501956, 0.03950073623657226, 0.03953868865966797, 0.03960422515869141, 0.03969945526123047, 0.03978854370117187, 0.08111411285400391, 0.040289279937744144, 0.0393072624206543, 0.03955199813842773, 0.039501823425292966, 0.0393891830444336, 0.0393359375, 0.03960934448242188, 0.03942604827880859, 0.039554046630859374, 0.03937279891967774, 0.03945369720458984, 0.03937177658081055, 0.039611423492431644, 0.03978031921386719, 0.0395335693359375, 0.039449600219726565, 0.03945574569702148, 0.03949158477783203, 0.039651329040527344, 0.039408641815185545, 0.039446529388427735, 0.039370750427246096, 0.039618560791015625, 0.03969126510620117, 0.03963391876220703, 0.039793663024902344, 0.03998822402954102, 0.03951513671875, 0.03942502212524414, 0.03936460876464844, 0.03952844619750977, 0.03944550323486328, 0.039468032836914066, 0.03973836898803711, 0.039488510131835936, 0.039398399353027344, 0.03951103973388672, 0.039433216094970705, 0.03953049468994141, 0.039384063720703126, 0.039610366821289066, 0.03942195129394531, 0.03937484741210937, 0.039444480895996094, 0.039605247497558595, 0.0394700813293457, 0.0394598388671875, 0.03948134231567383, 0.03951103973388672, 0.039702529907226565, 0.03990528106689453, 0.03952844619750977, 0.03959807968139648, 0.039534591674804685, 0.03949772644042969, 0.03940966415405273, 0.039390209197998044, 0.03936153411865234, 0.03949260711669922, 0.03947724914550781, 0.039478271484375, 0.03961964797973633, 0.08132498931884766, 0.039577598571777346, 0.039392257690429686, 0.039580673217773435, 0.039521278381347655, 0.03939328002929687, 0.039190528869628906, 0.03920588684082031, 0.039411712646484375, 0.0393891830444336, 0.03947110366821289, 0.03920588684082031, 0.03942911911010742, 0.03978137588500977, 0.03956531143188476, 0.03957452774047852, 0.039624702453613284, 0.03947110366821289, 0.03937996673583984, 0.03949676895141602, 0.03960723114013672, 0.03957350540161133, 0.03959299087524414, 0.04049200057983399, 0.040079360961914064, 0.03965235137939453, 0.03947315216064453, 0.039523326873779296, 0.03956633758544922, 0.03941785430908203, 0.03943219375610352, 0.03935539245605469, 0.03939328002929687, 0.03963596725463867, 0.03949772644042969, 0.039539710998535156, 0.03936460876464844, 0.039382015228271484, 0.03985203170776367, 0.04112588882446289, 0.03990835189819336, 0.039303199768066406, 0.03935740661621094, 0.03930828857421875, 0.039629825592041014, 0.039370750427246096, 0.03955507278442383, 0.039414783477783204, 0.039392257690429686, 0.03963391876220703, 0.039408641815185545, 0.03947727966308594, 0.03938198471069336, 0.03943116760253906, 0.039287841796875, 0.03957244873046875, 0.03941686248779297, 0.03938608169555664, 0.03933695983886719, 0.03938816070556641, 0.03966566467285156, 0.03951001739501953, 0.03967795181274414, 0.0816015396118164, 0.03950899124145508, 0.03951308822631836, 0.039567359924316405, 0.03965951919555664, 0.03955820846557617, 0.03948230361938477, 0.03946086502075195, 0.03948339080810547, 0.03967385482788086, 0.039572479248046875, 0.03956326293945313, 0.03950387191772461, 0.03951103973388672, 0.03965446472167969, 0.03958572769165039, 0.03975372695922851, 0.039676959991455076, 0.03963081741333008, 0.03968307113647461, 0.039874561309814455, 0.03963596725463867, 0.039359489440917966, 0.03943936157226562, 0.03984998321533203, 0.03992166519165039, 0.039825408935546876, 0.03975987243652344, 0.03933695983886719, 0.03944755172729492, 0.03970355224609375, 0.03947520065307617, 0.03924070358276367, 0.039408641815185545, 0.03939328002929687, 0.0395786247253418, 0.039427135467529295, 0.039470016479492186, 0.03935641479492188, 0.039585792541503906, 0.03967488098144531, 0.039550975799560545, 0.039375873565673826, 0.039498752593994144, 0.03950387191772461, 0.03968307113647461, 0.0395601921081543, 0.03951513671875, 0.039567359924316405, 0.03954278564453125, 0.040025089263916014, 0.03950694274902344, 0.03945369720458984, 0.03981414413452149, 0.03958169555664062, 0.03971583938598633, 0.040190975189208986, 0.039646209716796874, 0.03933184051513672, 0.03954380798339844, 0.03963289642333984, 0.03962777709960937, 0.039547904968261716, 0.081259521484375, 0.03940454483032227, 0.03937996673583984, 0.03957452774047852, 0.039414783477783204, 0.039419902801513675, 0.03935644912719727, 0.03931337738037109, 0.03946700668334961, 0.0394117431640625, 0.03942396926879883, 0.039762943267822266, 0.039406593322753904, 0.039550975799560545, 0.039481407165527345, 0.03942700958251953, 0.0395489273071289, 0.03993600082397461, 0.04035379028320313, 0.04098361587524414, 0.03966969680786133, 0.03950080108642578, 0.039616512298583983, 0.03977830505371094, 0.039462913513183595, 0.03949260711669922, 0.039669761657714846, 0.03949977493286133, 0.03933388900756836, 0.039332862854003905, 0.039398399353027344, 0.03946700668334961, 0.039672832489013675, 0.039769088745117184, 0.040182785034179686, 0.04107059097290039, 0.04016742324829101, 0.03948441696166992, 0.03967078399658203, 0.03942604827880859, 0.03934207916259766, 0.03936460876464844, 0.03937279891967774, 0.03931545639038086, 0.04087807846069336, 0.04114636611938476, 0.039907329559326174, 0.03947110366821289, 0.03962879943847656, 0.03942707061767578, 0.03959609603881836, 0.03944236755371094, 0.039609375, 0.03951305770874024, 0.03944857788085938, 0.03932160186767578, 0.03956531143188476, 0.03942399978637695, 0.03932160186767578, 0.03932057571411133, 0.039365631103515625, 0.03976499176025391, 0.03956326293945313]",tokens/s,24.871208405344607,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2019.69664,5274.861568,0.0,4628.414464,4279.64672,s,10,3.1774854736328124,0.31774854736328123,0.0007817722146641278,0.3174602966308594,0.3187982666015625,0.3189695983886719,0.3191066638183594,"[0.31848922729492185, 0.3174812927246094, 0.31692730712890627, 0.31720437622070313, 0.3174393005371094, 0.31661932373046875, 0.3174072570800781, 0.31876019287109375, 0.31914093017578127, 0.3180162658691406]",tokens/s,805.6685140634671,kWh,3.737864001757569e-06,2.0481879523572387e-06,1.80164467048112e-05,2.3802498658926007e-05,tokens/kWh,10755173.38193397,MB,2019.69664,5274.861568,0.0,4628.414464,4463.185408,s,10,176.9617734375,17.69617734375,0.0029728792469339506,17.696136718749997,17.6995603515625,17.699568261718753,17.699574589843753,"[17.69631640625, 17.695939453125, 17.698544921875, 17.693357421875, 17.690775390625, 17.69253125, 17.699216796875, 17.69955859375, 17.699576171875, 17.69595703125]",tokens/s,3.5600909041664055,kWh,0.00020889662316689894,0.00011449271197744792,0.0009934635569229903,0.0013168528920673372,tokens/kWh,47841.33473033258,,s,629,179.53831649780273,0.2854345254337086,0.0377383923728263,0.28085964965820315,0.28117113647460934,0.2812770324707031,0.5980694702148438,"[0.2809251708984375, 0.280669189453125, 0.2807091064453125, 0.2809108581542969, 0.28111666870117186, 0.28064971923828125, 0.2810654602050781, 0.28062106323242186, 0.2806558837890625, 0.2807910461425781, 0.28068453979492186, 0.2808770446777344, 0.2806640625, 0.2807613525390625, 0.28071832275390624, 0.28078387451171877, 0.2807152709960937, 0.28067431640625, 0.28070501708984374, 0.2807490539550781, 0.28121600341796876, 0.2806056823730469, 0.28095077514648437, 0.28073471069335937, 0.28082278442382813, 0.2807633972167969, 0.2808299560546875, 0.2809405517578125, 0.28083712768554686, 0.28074700927734375, 0.28070401000976564, 0.28072857666015627, 0.2808084411621094, 0.2823638916015625, 0.2806558837890625, 0.2807715759277344, 0.2807767028808594, 0.2810624084472656, 0.28082278442382813, 0.2810224609375, 0.28100811767578127, 0.2810163269042969, 0.2808545227050781, 0.28095077514648437, 0.2809190368652344, 0.28101324462890626, 0.2807828369140625, 0.2812323913574219, 0.2809671630859375, 0.28106137084960936, 0.2808330383300781, 0.2810439758300781, 0.28064971923828125, 0.28111770629882815, 0.2812231750488281, 0.2810408935546875, 0.28118118286132815, 0.28099072265625, 0.2808832092285156, 0.2812590026855469, 0.2809385070800781, 0.2808832092285156, 0.5983795166015625, 0.28064871215820314, 0.2806937561035156, 0.2806876220703125, 0.2805964660644531, 0.28077362060546873, 0.28084735107421877, 0.2808135681152344, 0.28077978515625, 0.2806302795410156, 0.2807633972167969, 0.2808637390136719, 0.281501708984375, 0.2808135681152344, 0.28127743530273436, 0.2808412170410156, 0.2809886779785156, 0.2807439270019531, 0.2807940979003906, 0.28082382202148437, 0.28080126953125, 0.28092620849609373, 0.2807767028808594, 0.28085964965820315, 0.2809610290527344, 0.2808668212890625, 0.28077978515625, 0.2808545227050781, 0.2811832275390625, 0.28089959716796875, 0.28109515380859373, 0.28077566528320314, 0.2809292907714844, 0.28073983764648436, 0.28083712768554686, 0.2807214050292969, 0.28081561279296874, 0.28083096313476563, 0.28084735107421877, 0.2808258666992188, 0.28108901977539064, 0.28077566528320314, 0.28080230712890625, 0.2808279113769531, 0.28091494750976564, 0.280700927734375, 0.28091494750976564, 0.28119143676757813, 0.28127642822265625, 0.2812569580078125, 0.2810091552734375, 0.2808279113769531, 0.28108697509765623, 0.2808934326171875, 0.28086273193359373, 0.28114431762695313, 0.28093438720703123, 0.28095489501953125, 0.2810040283203125, 0.2808350830078125, 0.28096307373046875, 0.2807848815917969, 0.2808197021484375, 0.5980743408203125, 0.28070401000976564, 0.2806609802246094, 0.28080743408203124, 0.280690673828125, 0.2807357482910156, 0.28078900146484376, 0.28069888305664065, 0.28073883056640625, 0.28077566528320314, 0.280637451171875, 0.2809395141601562, 0.28084225463867185, 0.280669189453125, 0.28070196533203123, 0.28089242553710936, 0.2807613525390625, 0.28102554321289064, 0.280958984375, 0.28090777587890625, 0.28084530639648436, 0.2811146240234375, 0.2807060546875, 0.2808821716308594, 0.28075314331054685, 0.2806824951171875, 0.2807562255859375, 0.2807132263183594, 0.2805975036621094, 0.280627197265625, 0.2807439270019531, 0.280848388671875, 0.2807879638671875, 0.2806599731445312, 0.2806077575683594, 0.28076031494140624, 0.28118426513671874, 0.28088525390625, 0.2811043701171875, 0.281143310546875, 0.2812733459472656, 0.2809036865234375, 0.28103372192382814, 0.28068453979492186, 0.28123956298828123, 0.28111566162109375, 0.28099072265625, 0.28118118286132815, 0.2812323913574219, 0.28117300415039065, 0.2811197509765625, 0.2810224609375, 0.28101937866210935, 0.28089447021484376, 0.28143719482421875, 0.2813931579589844, 0.28145767211914063, 0.28113409423828123, 0.2810245056152344, 0.2813102111816406, 0.2813634643554688, 0.2808913879394531, 0.2811832275390625, 0.5980569458007813, 0.280806396484375, 0.28072549438476563, 0.28071218872070314, 0.2806876220703125, 0.2807807922363281, 0.28076235961914064, 0.2810705871582031, 0.2806241149902344, 0.2806855773925781, 0.28086578369140625, 0.2805698547363281, 0.2807582702636719, 0.28055859375, 0.28054937744140623, 0.28078387451171877, 0.2807562255859375, 0.2807173156738281, 0.28059033203125, 0.2807162780761719, 0.2807500915527344, 0.28070196533203123, 0.2806824951171875, 0.2808340454101563, 0.28079718017578126, 0.28069171142578125, 0.2811125793457031, 0.2808883056640625, 0.28106341552734376, 0.28073471069335937, 0.28064254760742186, 0.2807705688476563, 0.2809989013671875, 0.28072857666015627, 0.28084225463867185, 0.28058624267578125, 0.28073370361328126, 0.28079513549804686, 0.28098458862304687, 0.28078900146484376, 0.2809722900390625, 0.2808965148925781, 0.2811125793457031, 0.28095693969726565, 0.281196533203125, 0.28067636108398436, 0.28092312622070315, 0.2809385070800781, 0.28081561279296874, 0.28086477661132814, 0.2808913879394531, 0.28067431640625, 0.28132455444335935, 0.280922119140625, 0.28108697509765623, 0.2814136352539062, 0.2810091552734375, 0.28069989013671875, 0.280838134765625, 0.28130712890625, 0.2808555603027344, 0.2809016418457031, 0.2813931579589844, 0.5983590698242187, 0.2807306213378906, 0.28087808227539063, 0.2805196838378906, 0.2805125122070313, 0.2804551696777344, 0.2805780334472656, 0.2806200256347656, 0.2805002136230469, 0.28068453979492186, 0.28058010864257815, 0.2806302795410156, 0.2806098022460938, 0.2805196838378906, 0.2809016418457031, 0.28069989013671875, 0.28064971923828125, 0.28055755615234373, 0.28064666748046874, 0.28067019653320313, 0.2806671447753906, 0.28063641357421876, 0.28048590087890624, 0.2806886291503906, 0.2806722412109375, 0.2805166015625, 0.2806087646484375, 0.2805964660644531, 0.28062106323242186, 0.2810009460449219, 0.2807500915527344, 0.2806753234863281, 0.28076235961914064, 0.28126412963867187, 0.28134194946289065, 0.281164794921875, 0.28111666870117186, 0.2810347595214844, 0.28092620849609373, 0.2809241638183594, 0.28083096313476563, 0.28105523681640626, 0.2809354248046875, 0.2808258666992188, 0.281017333984375, 0.2810296325683594, 0.2807910461425781, 0.2810521545410156, 0.28102859497070315, 0.2807940979003906, 0.2811197509765625, 0.28095590209960936, 0.2808842163085937, 0.2810460205078125, 0.2808913879394531, 0.2809292907714844, 0.2807777404785156, 0.28074188232421876, 0.28088525390625, 0.280995849609375, 0.28099172973632813, 0.28075213623046874, 0.28117196655273435, 0.5974200439453125, 0.28051559448242186, 0.28070501708984374, 0.28070501708984374, 0.280669189453125, 0.28124365234375, 0.2808965148925781, 0.2807500915527344, 0.2808545227050781, 0.2804951171875, 0.280748046875, 0.28090264892578126, 0.28080435180664065, 0.2806118469238281, 0.28069580078125, 0.2808616943359375, 0.28078387451171877, 0.2806773681640625, 0.28072039794921877, 0.2808463439941406, 0.28071014404296873, 0.2810408935546875, 0.28070501708984374, 0.28084530639648436, 0.2805821533203125, 0.2807767028808594, 0.2807459716796875, 0.2806599731445312, 0.28074188232421876, 0.28069580078125, 0.2806732788085938, 0.2806507568359375, 0.2808084411621094, 0.2809466857910156, 0.2810572814941406, 0.2805780334472656, 0.28073883056640625, 0.28082891845703123, 0.2810071105957031, 0.28080331420898436, 0.28094976806640626, 0.28130303955078123, 0.28097842407226564, 0.2809968566894531, 0.2808493957519531, 0.2807193603515625, 0.2809456787109375, 0.2811535339355469, 0.2808330383300781, 0.28075418090820314, 0.280853515625, 0.2809169921875, 0.28076953125, 0.280890380859375, 0.2809374694824219, 0.281280517578125, 0.28133172607421875, 0.28101324462890626, 0.28093438720703123, 0.28075314331054685, 0.2808053894042969, 0.28079205322265627, 0.2809333801269531, 0.5982566528320312, 0.2806384582519531, 0.28069989013671875, 0.28062823486328126, 0.2805821533203125, 0.28079205322265627, 0.2808053894042969, 0.28074496459960935, 0.28080743408203124, 0.2807490539550781, 0.2807940979003906, 0.2807173156738281, 0.28145458984375, 0.281248779296875, 0.28092825317382814, 0.2808186950683594, 0.280995849609375, 0.2806609802246094, 0.2807244873046875, 0.2807807922363281, 0.28073779296875, 0.28084326171875, 0.2809169921875, 0.2808883056640625, 0.2809671630859375, 0.2806528015136719, 0.28088525390625, 0.2807490539550781, 0.28113714599609374, 0.2809948120117188, 0.28080331420898436, 0.28099993896484377, 0.28106137084960936, 0.2807500915527344, 0.28100302124023435, 0.280774658203125, 0.28079718017578126, 0.28105319213867186, 0.2808555603027344, 0.2810224609375, 0.2812200927734375, 0.28123544311523435, 0.2811412353515625, 0.28102554321289064, 0.28113201904296875, 0.2809661560058594, 0.280995849609375, 0.2812538757324219, 0.2814095458984375, 0.2813102111816406, 0.28100506591796875, 0.28089447021484376, 0.28107672119140625, 0.28094464111328127, 0.28103884887695313, 0.28113613891601563, 0.2810777587890625, 0.28092312622070315, 0.2810982360839844, 0.280875, 0.2811453552246094, 0.28104193115234377, 0.28097842407226564, 0.6000148315429688, 0.28073370361328126, 0.2808883056640625, 0.28085964965820315, 0.2808688659667969, 0.28098355102539063, 0.280806396484375, 0.28103988647460937, 0.2807173156738281, 0.2808053894042969, 0.28107879638671873, 0.2807080993652344, 0.2809938049316406, 0.28080947875976564, 0.28094259643554687, 0.28093646240234377, 0.2807214050292969, 0.2808412170410156, 0.28081048583984375, 0.2815068054199219, 0.28095590209960936, 0.28094976806640626, 0.28077978515625, 0.2809466857910156, 0.2811985778808594, 0.2808299560546875, 0.28084530639648436, 0.28066201782226563, 0.2807275390625, 0.280890380859375, 0.28104702758789063, 0.28076748657226563, 0.2809395141601562, 0.2808934326171875, 0.2809876403808594, 0.28116683959960936, 0.28092007446289063, 0.2807060546875, 0.2810091552734375, 0.28097024536132814, 0.280890380859375, 0.28086782836914065, 0.28100811767578127, 0.28064358520507815, 0.2809088134765625, 0.2808115234375, 0.2807173156738281, 0.28098150634765623, 0.28097537231445313, 0.2808688659667969, 0.2809374694824219, 0.28094976806640626, 0.2810654602050781, 0.28078387451171877, 0.2811709289550781, 0.2815887451171875, 0.2812333984375, 0.28107879638671873, 0.2810439758300781, 0.281069580078125, 0.28113818359375, 0.28109515380859373, 0.2812569580078125, 0.6002237548828125, 0.280869873046875, 0.2811043701171875, 0.28084225463867185, 0.2808985595703125, 0.28092007446289063, 0.28068658447265626, 0.28079513549804686, 0.28076953125, 0.28095281982421877, 0.2808002624511719, 0.28057089233398436, 0.28077978515625, 0.28081460571289063, 0.2807982177734375, 0.2810357666015625, 0.28085760498046874, 0.28081561279296874, 0.28080947875976564, 0.2812375183105469, 0.28106341552734376, 0.2807726135253906, 0.2809108581542969, 0.2809692077636719, 0.2808115234375, 0.28083712768554686, 0.2811781005859375, 0.28099172973632813, 0.2809948120117188, 0.2808340454101563, 0.28090264892578126, 0.28101324462890626, 0.2809876403808594, 0.2817751159667969, 0.2810347595214844, 0.2808299560546875, 0.281091064453125, 0.28080435180664065, 0.28114227294921873, 0.2809620361328125, 0.2809395141601562, 0.28100811767578127, 0.28102041625976565, 0.2808565673828125, 0.28097024536132814, 0.2810491027832031, 0.281064453125, 0.281091064453125, 0.2807859191894531, 0.2808688659667969, 0.28093646240234377, 0.28071832275390624, 0.2809395141601562, 0.2809599914550781, 0.2808821716308594, 0.28085247802734375, 0.28103372192382814, 0.2807777404785156, 0.2810357666015625, 0.28092108154296874, 0.28099789428710936, 0.28096820068359374, 0.28156927490234374, 0.5991085815429688, 0.280742919921875, 0.2808258666992188, 0.2806661071777344, 0.28075418090820314, 0.2806241149902344, 0.2807306213378906, 0.2808197021484375, 0.2806886291503906, 0.2807244873046875, 0.28067019653320313, 0.2805350341796875, 0.2807224426269531, 0.280869873046875, 0.28091390991210935, 0.2807828369140625, 0.2808197021484375, 0.28051046752929687, 0.2806558837890625, 0.280774658203125, 0.28073165893554686, 0.2806384582519531, 0.280711181640625, 0.28101223754882815, 0.28087091064453124, 0.2806732788085938, 0.28076748657226563, 0.28072653198242187, 0.28075418090820314, 0.28080230712890625, 0.2811269226074219, 0.28098046875, 0.28079513549804686, 0.281059326171875, 0.2811566162109375, 0.2811617431640625, 0.2810163269042969, 0.28065689086914064, 0.2807767028808594, 0.2808002624511719, 0.28075726318359373, 0.28086782836914065, 0.28077566528320314, 0.2806456298828125, 0.28085964965820315, 0.28072039794921877, 0.2807767028808594, 0.2809374694824219, 0.28083096313476563, 0.2811514892578125, 0.2812262268066406, 0.2808821716308594, 0.2809948120117188, 0.2811064453125, 0.28104193115234377, 0.2811760559082031, 0.28112997436523435, 0.28082891845703123, 0.2809241638183594, 0.28104806518554687, 0.2811535339355469, 0.2811269226074219, 0.28101223754882815]",tokens/s,3.5034304223728086,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12309,7 +12309,7 @@ ImportError: This modeling file requires the following packages that were not fo 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,4111.491072,5287.44448,0.0,4640.997376,4330.799616,s,10,3.226784301757813,0.3226784301757813,0.00036527946974059694,0.3228235473632812,0.3229854919433594,0.3231664367675781,0.3233111926269531,"[0.3222513732910156, 0.3229452819824219, 0.3228663024902344, 0.32282205200195313, 0.3228476257324219, 0.32209072875976563, 0.3233473815917969, 0.32249490356445315, 0.3228250427246094, 0.3222936096191406]",tokens/s,793.3595061205122,kWh,3.802613670834237e-06,2.0836708957028095e-06,1.8872671348126036e-05,2.475895591466308e-05,tokens/kWh,10339692.872444117,MB,4111.491072,5287.44448,0.0,4640.997376,4515.66848,s,10,175.44622851562502,17.544622851562497,0.0012475331034878638,17.5441357421875,17.546382031249998,17.546624609374998,17.546818671875,"[17.5441171875, 17.545755859375, 17.543529296875, 17.545173828125, 17.546328125, 17.54321484375, 17.543701171875, 17.5468671875, 17.544154296875, 17.54338671875]",tokens/s,3.5908437891777947,kWh,0.00020715286491128305,0.0001135369823019525,0.0010055265370594742,0.0013262163842727099,tokens/kWh,47503.560314215894,,s,629,178.07182846069333,0.28310306591525175,0.03837524534756707,0.278476806640625,0.2786437072753906,0.2787579833984375,0.6014762182617188,"[0.27898776245117185, 0.2789980163574219, 0.27833856201171875, 0.2784163818359375, 0.27828939819335935, 0.2783426513671875, 0.2782156677246094, 0.2783068237304688, 0.278192138671875, 0.27826995849609376, 0.2781388854980469, 0.2783426513671875, 0.27819931030273437, 0.27850750732421875, 0.27829043579101564, 0.2783999938964844, 0.2782494812011719, 0.2784798583984375, 0.2784337768554688, 0.2785310668945313, 0.2782771301269531, 0.27843585205078125, 0.2784327697753906, 0.2784798583984375, 0.2783631286621094, 0.27856793212890624, 0.27837542724609377, 0.2785208435058594, 0.27838668823242185, 0.2784635009765625, 0.2784471130371094, 0.2785535888671875, 0.2784409484863281, 0.2785791931152344, 0.278408203125, 0.27850958251953123, 0.2783365173339844, 0.27850958251953123, 0.2784204711914062, 0.27854437255859377, 0.27840203857421875, 0.27851263427734374, 0.2784604187011719, 0.2785433654785156, 0.27850546264648435, 0.27857510375976563, 0.2786293640136719, 0.278572021484375, 0.2785208435058594, 0.27859967041015626, 0.27844403076171875, 0.27852291870117185, 0.27841943359375, 0.2788905029296875, 0.2784184265136719, 0.2785710144042969, 0.27854232788085936, 0.2785515441894531, 0.27840408325195315, 0.27864984130859377, 0.27850958251953123, 0.2785853576660156, 0.6017566528320313, 0.27841537475585937, 0.27831195068359377, 0.27905227661132814, 0.27840103149414064, 0.27819110107421874, 0.27843685913085936, 0.2782791748046875, 0.2782801818847656, 0.2781614074707031, 0.27826483154296877, 0.2782443542480469, 0.2783846435546875, 0.2782863464355469, 0.27843585205078125, 0.27829351806640623, 0.278255615234375, 0.2783743896484375, 0.2784132995605469, 0.2783170471191406, 0.2789324951171875, 0.2784429931640625, 0.2784604187011719, 0.2784757690429687, 0.2785218505859375, 0.27847271728515627, 0.2788792419433594, 0.27871539306640625, 0.27856484985351565, 0.27845632934570314, 0.278508544921875, 0.2784757690429687, 0.2786160583496094, 0.27841741943359377, 0.2785157165527344, 0.2784747619628906, 0.2785228881835938, 0.2784194641113281, 0.27849420166015626, 0.2787901306152344, 0.27852801513671877, 0.2783682861328125, 0.2785689392089844, 0.27856692504882813, 0.27852593994140623, 0.27843072509765626, 0.278687744140625, 0.27848703002929687, 0.2785269775390625, 0.278518798828125, 0.27858740234375, 0.27840103149414064, 0.2785638427734375, 0.2785321044921875, 0.27859661865234375, 0.2784675903320312, 0.2790205383300781, 0.2785515441894531, 0.2786007080078125, 0.278513671875, 0.2786293640136719, 0.2785740661621094, 0.27877886962890625, 0.6017741088867188, 0.27854849243164065, 0.278319091796875, 0.27832830810546877, 0.2783969421386719, 0.2782054443359375, 0.27833242797851565, 0.2782443542480469, 0.2784460754394531, 0.27819827270507813, 0.27826174926757813, 0.27834878540039065, 0.2784337768554688, 0.2783303527832031, 0.27842559814453127, 0.2782597045898437, 0.278403076171875, 0.2785167236328125, 0.27843685913085936, 0.2784327697753906, 0.2784460754394531, 0.27831195068359377, 0.27852593994140623, 0.278476806640625, 0.27842251586914063, 0.2785218505859375, 0.2786344909667969, 0.2785024108886719, 0.27854949951171876, 0.27850137329101565, 0.27848397827148436, 0.27858123779296873, 0.27854132080078126, 0.27842559814453127, 0.27854132080078126, 0.27832830810546877, 0.27854745483398435, 0.2783006591796875, 0.27839794921875, 0.27840615844726563, 0.27851467895507814, 0.27842251586914063, 0.2784471130371094, 0.27854849243164065, 0.2784737243652344, 0.2784901123046875, 0.27870513916015627, 0.27853927612304685, 0.2786344909667969, 0.2785372009277344, 0.2786078796386719, 0.278498291015625, 0.27864474487304686, 0.2784655456542969, 0.27859457397460935, 0.2785228881835938, 0.27856280517578125, 0.2785730590820312, 0.27854541015625, 0.2784122924804687, 0.27856484985351565, 0.2784122924804687, 0.27866827392578125, 0.601596923828125, 0.27889459228515623, 0.2783109130859375, 0.27825048828125, 0.27840716552734374, 0.2782853088378906, 0.2783498229980469, 0.2783273010253906, 0.2783160400390625, 0.2782740478515625, 0.27838156127929686, 0.2782013549804688, 0.2784880676269531, 0.27822900390625, 0.27846142578125, 0.2783918151855469, 0.2784266357421875, 0.27835186767578124, 0.2784952392578125, 0.2784378967285156, 0.27846450805664064, 0.27835391235351564, 0.27842868041992186, 0.27840203857421875, 0.2784296875, 0.27854745483398435, 0.27856793212890624, 0.2784204711914062, 0.27891302490234376, 0.278434814453125, 0.27845428466796873, 0.2784532470703125, 0.27840615844726563, 0.2785607604980469, 0.2784849853515625, 0.2785587158203125, 0.27897445678710936, 0.2788044738769531, 0.278498291015625, 0.2784757690429687, 0.278513671875, 0.27842251586914063, 0.2785238952636719, 0.27860275268554685, 0.27848089599609377, 0.27847271728515627, 0.27856484985351565, 0.2786160583496094, 0.27856484985351565, 0.27850546264648435, 0.27857510375976563, 0.2784552917480469, 0.27852291870117185, 0.27853717041015624, 0.27864678955078126, 0.27850546264648435, 0.2786437072753906, 0.27860479736328125, 0.27857510375976563, 0.2783743896484375, 0.27860888671875, 0.27841741943359377, 0.27879220581054687, 0.6017177734375, 0.27843173217773437, 0.27856484985351565, 0.27820645141601563, 0.2784532470703125, 0.27828121948242185, 0.2783703002929688, 0.27863961791992187, 0.2783273010253906, 0.27872869873046874, 0.27829144287109375, 0.27829248046875, 0.27839385986328125, 0.278645751953125, 0.2785607604980469, 0.2783549499511719, 0.27834573364257814, 0.27826278686523437, 0.2784409484863281, 0.2784675903320312, 0.27854949951171876, 0.27830374145507814, 0.278476806640625, 0.27845428466796873, 0.2783498229980469, 0.27835699462890623, 0.2788239440917969, 0.27835699462890623, 0.2785310668945313, 0.278550537109375, 0.27856484985351565, 0.27840103149414064, 0.27856793212890624, 0.2785863647460938, 0.2785269775390625, 0.27840716552734374, 0.2784952392578125, 0.27840615844726563, 0.27848703002929687, 0.27833856201171875, 0.27859762573242186, 0.27845632934570314, 0.2785884094238281, 0.27898675537109374, 0.278550537109375, 0.27880551147460936, 0.27859762573242186, 0.27865396118164065, 0.27867034912109373, 0.27876043701171876, 0.27874508666992187, 0.27836517333984373, 0.27855975341796874, 0.2786437072753906, 0.2785218811035156, 0.2784992980957031, 0.2785587158203125, 0.27855563354492185, 0.27860992431640624, 0.27846142578125, 0.2785710144042969, 0.2784337768554688, 0.27867544555664064, 0.6009231567382812, 0.27844915771484374, 0.27832626342773437, 0.27822796630859375, 0.278319091796875, 0.27819314575195314, 0.2783150024414062, 0.2781388854980469, 0.27829556274414063, 0.2782443542480469, 0.27844198608398435, 0.2782054443359375, 0.2784204711914062, 0.2782515258789062, 0.27840103149414064, 0.2783088684082031, 0.27841433715820313, 0.2784051208496094, 0.27846450805664064, 0.27850137329101565, 0.2785863647460938, 0.2783273010253906, 0.2784471130371094, 0.27842251586914063, 0.2783897705078125, 0.2784798583984375, 0.2786458129882812, 0.2784869689941406, 0.2785003662109375, 0.2784901123046875, 0.27853927612304685, 0.27848703002929687, 0.27861810302734374, 0.2784532470703125, 0.27848907470703127, 0.27840203857421875, 0.2785525817871094, 0.27850137329101565, 0.2785290222167969, 0.27842251586914063, 0.27850445556640624, 0.27838772583007815, 0.27854232788085936, 0.2785658874511719, 0.2785986633300781, 0.2783918151855469, 0.2785904541015625, 0.2784696350097656, 0.27860275268554685, 0.27858944702148436, 0.27858944702148436, 0.2784378967285156, 0.2785228881835938, 0.27860479736328125, 0.2785290222167969, 0.27843685913085936, 0.2785802307128906, 0.2785003662109375, 0.2785658874511719, 0.27848397827148436, 0.2786344909667969, 0.2784583740234375, 0.2787481689453125, 0.602113037109375, 0.2783365173339844, 0.2783211669921875, 0.27824127197265625, 0.2783375244140625, 0.2782105712890625, 0.27836721801757813, 0.27829144287109375, 0.2783723449707031, 0.27835903930664063, 0.27832421875, 0.2782996520996094, 0.2784122924804687, 0.27834878540039065, 0.2786160583496094, 0.2783375244140625, 0.27824127197265625, 0.278371337890625, 0.2785658874511719, 0.27838772583007815, 0.27851980590820313, 0.27827096557617187, 0.27854949951171876, 0.2785321044921875, 0.2784276428222656, 0.27848703002929687, 0.2785791931152344, 0.2783416442871094, 0.2785105895996094, 0.2784778137207031, 0.27854232788085936, 0.2784430541992188, 0.27862832641601565, 0.27843585205078125, 0.2785218505859375, 0.27845733642578124, 0.27856179809570314, 0.27849932861328125, 0.2785587158203125, 0.2783477783203125, 0.27854541015625, 0.27844915771484374, 0.2785433654785156, 0.2784583740234375, 0.27859457397460935, 0.27845016479492185, 0.27858944702148436, 0.27850958251953123, 0.27855667114257815, 0.2786160583496094, 0.27856997680664064, 0.27856179809570314, 0.27869696044921877, 0.27845120239257815, 0.2784655456542969, 0.27840103149414064, 0.278550537109375, 0.27846142578125, 0.2785904541015625, 0.2785884094238281, 0.27869696044921877, 0.27850137329101565, 0.2786795654296875, 0.6017638549804688, 0.2783743896484375, 0.27835699462890623, 0.2782576599121094, 0.2783549499511719, 0.2782853088378906, 0.2783774719238281, 0.2782945251464844, 0.27829043579101564, 0.27835391235351564, 0.278423583984375, 0.27823715209960936, 0.27836517333984373, 0.2787543029785156, 0.2783446960449219, 0.27829144287109375, 0.27843685913085936, 0.27847064208984373, 0.27850958251953123, 0.2783897705078125, 0.27845733642578124, 0.2784163818359375, 0.2784163818359375, 0.27838156127929686, 0.2784635009765625, 0.2784245910644531, 0.27858432006835937, 0.27845016479492185, 0.27850650024414064, 0.27848602294921876, 0.27913421630859375, 0.27850445556640624, 0.2785167236328125, 0.2786078796386719, 0.2784952392578125, 0.27845733642578124, 0.2789427185058594, 0.27852593994140623, 0.27844915771484374, 0.2784583740234375, 0.27856692504882813, 0.278466552734375, 0.27858328247070313, 0.2785433654785156, 0.27850546264648435, 0.27851776123046873, 0.27864678955078126, 0.27859967041015626, 0.27850650024414064, 0.2791065673828125, 0.27865396118164065, 0.2785607604980469, 0.27861810302734374, 0.2784532470703125, 0.27860992431640624, 0.2784962463378906, 0.2786344909667969, 0.2785310668945313, 0.278687744140625, 0.2785935363769531, 0.27867340087890624, 0.27854745483398435, 0.2786990051269531, 0.601702392578125, 0.278392822265625, 0.2786713562011719, 0.2782300109863281, 0.2784583740234375, 0.2782576599121094, 0.2783621215820313, 0.2783365173339844, 0.2783354797363281, 0.27830374145507814, 0.27835699462890623, 0.2782668762207031, 0.27834060668945315, 0.27823513793945315, 0.27846142578125, 0.27838668823242185, 0.2784194641113281, 0.27830374145507814, 0.27834878540039065, 0.2784184265136719, 0.27842355346679687, 0.27863348388671877, 0.27850546264648435, 0.2784122924804687, 0.27835391235351564, 0.2784122924804687, 0.27847271728515627, 0.27833856201171875, 0.27849932861328125, 0.27848602294921876, 0.2785218505859375, 0.27842251586914063, 0.27849420166015626, 0.2784583740234375, 0.2783856506347656, 0.27835186767578124, 0.27848501586914065, 0.27843887329101563, 0.2786611328125, 0.27842355346679687, 0.27854849243164065, 0.2784798583984375, 0.27856280517578125, 0.2785218505859375, 0.27852801513671877, 0.2784655456542969, 0.2786488342285156, 0.2786211853027344, 0.27856280517578125, 0.27854437255859377, 0.2785904541015625, 0.27843072509765626, 0.2785986633300781, 0.2788648986816406, 0.27861709594726564, 0.27849114990234375, 0.27876455688476565, 0.2785853576660156, 0.27858944702148436, 0.27855462646484375, 0.2785802307128906, 0.2784532470703125, 0.2786324462890625, 0.6011658325195313, 0.2784409484863281, 0.2784972839355469, 0.278302734375, 0.27838668823242185, 0.27827096557617187, 0.27833447265625, 0.27829556274414063, 0.2782597045898437, 0.27825357055664063, 0.2784184265136719, 0.27841537475585937, 0.27831808471679687, 0.278408203125, 0.2784194641113281, 0.27828121948242185, 0.2784747619628906, 0.2783160400390625, 0.2783692932128906, 0.2782843017578125, 0.2784716796875, 0.27823410034179685, 0.278403076171875, 0.27845120239257815, 0.2784215087890625, 0.27848907470703127, 0.27865805053710935, 0.2783918151855469, 0.27849319458007815, 0.2785710144042969, 0.27845733642578124, 0.27843685913085936, 0.2785300598144531, 0.2784215087890625, 0.27843072509765626, 0.27834878540039065, 0.2785382385253906, 0.2784409484863281, 0.2785535888671875, 0.27840615844726563, 0.2786293640136719, 0.27847064208984373, 0.27848193359375, 0.27861505126953123, 0.2784409484863281, 0.27836007690429687, 0.27853515625, 0.2786437072753906, 0.2786375732421875, 0.27863143920898437, 0.27857818603515627, 0.278403076171875, 0.2786017150878906, 0.2784962463378906, 0.27854849243164065, 0.27848193359375, 0.27860479736328125, 0.27847885131835937, 0.2786693115234375, 0.2785024108886719, 0.2786283569335937, 0.27850546264648435, 0.2787266540527344]",tokens/s,3.532282480823979,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1254.72768,893.91104,0.0,247.463936,189.067776,s,10,0.1715008964538574,0.01715008964538574,0.0004626842891741546,0.016930991172790524,0.01763515911102295,0.01799036340713501,0.01827452684402466,"[0.018345567703247072, 0.016922784805297852, 0.01693615913391113, 0.01692582321166992, 0.017556224822998047, 0.017033056259155275, 0.016823040008544923, 0.0168306884765625, 0.016779743194580077, 0.017347808837890624]",tokens/s,14927.035676975438,kWh,2.0434204854217232e-07,1.1196999001085051e-07,4.699686574049221e-07,7.862806959579449e-07,tokens/kWh,325583473.3270527,MB,1254.72768,893.91104,0.0,247.463936,209.72544,s,10,10.341356201171875,1.0341356201171874,0.011260867081140505,1.0309447021484375,1.0374079345703124,1.052245959472656,1.0641163793945312,"[1.067083984375, 1.034110595703125, 1.0270076904296874, 1.03406640625, 1.02809033203125, 1.0301700439453125, 1.02668896484375, 1.0310333251953125, 1.0308560791015624, 1.032248779296875]",tokens/s,60.92044290366953,kWh,1.2578433529029486e-05,6.8925208202391405e-06,2.5497357671990878e-05,4.4968312021259504e-05,tokens/kWh,1400986.5429286233,,s,629,10.476059639930726,0.01665510276618557,0.0020868862873047386,0.016270336151123048,0.016829439163208008,0.017095475006103516,0.03337338790893555,"[0.018183168411254884, 0.01777561569213867, 0.017458175659179686, 0.017114112854003907, 0.017076255798339844, 0.016827360153198242, 0.0170700798034668, 0.01721241569519043, 0.017140735626220704, 0.01682636833190918, 0.016756736755371093, 0.01745408058166504, 0.018266111373901366, 0.01863884735107422, 0.017704959869384765, 0.017714176177978515, 0.017732608795166017, 0.017265663146972657, 0.01683046340942383, 0.016926719665527345, 0.017978368759155275, 0.01718272018432617, 0.016821247100830078, 0.01663283157348633, 0.016713727951049806, 0.016701440811157226, 0.01680588722229004, 0.020279296875, 0.01758515167236328, 0.017098751068115235, 0.016845823287963867, 0.016728063583374024, 0.016270336151123048, 0.016235519409179687, 0.01621708869934082, 0.01623040008544922, 0.016331775665283203, 0.016232448577880858, 0.01620070457458496, 0.016192512512207033, 0.01634611129760742, 0.016282623291015624, 0.01660211181640625, 0.01742438316345215, 0.01684480094909668, 0.016737279891967775, 0.01683148765563965, 0.016470016479492186, 0.016519168853759765, 0.01620070457458496, 0.016315391540527344, 0.01623347282409668, 0.016220159530639648, 0.016683008193969725, 0.0166748161315918, 0.016706560134887697, 0.0163450870513916, 0.01617203140258789, 0.01619865608215332, 0.01680281639099121, 0.0162795524597168, 0.01620172882080078, 0.033454078674316406, 0.01616383934020996, 0.01723494338989258, 0.01641881561279297, 0.01637068748474121, 0.01679052734375, 0.016773120880126953, 0.01701068878173828, 0.01681715202331543, 0.0165130558013916, 0.016794591903686523, 0.016244735717773438, 0.016242687225341796, 0.016236543655395508, 0.01623347282409668, 0.016317440032958985, 0.016303104400634767, 0.01640447998046875, 0.01636249542236328, 0.01620889663696289, 0.016381952285766603, 0.01680793571472168, 0.016550912857055664, 0.01620889663696289, 0.016312320709228514, 0.016145408630371092, 0.01641472053527832, 0.016720895767211915, 0.01656831932067871, 0.01627136039733887, 0.01623142433166504, 0.01623347282409668, 0.016322559356689453, 0.016323583602905273, 0.01621299171447754, 0.016236543655395508, 0.016145408630371092, 0.0161474552154541, 0.01613926315307617, 0.016295936584472655, 0.01623961639404297, 0.01637990379333496, 0.016274431228637695, 0.016232448577880858, 0.016153600692749022, 0.01624678421020508, 0.01622118377685547, 0.01621299171447754, 0.01620582389831543, 0.01618636894226074, 0.016265247344970705, 0.016647136688232422, 0.016829439163208008, 0.016760831832885743, 0.01677107238769531, 0.016228351593017578, 0.016390144348144533, 0.01621401596069336, 0.01658367919921875, 0.016865280151367186, 0.016723968505859374, 0.016656383514404297, 0.016207872390747072, 0.033452030181884765, 0.016235519409179687, 0.016303104400634767, 0.016270336151123048, 0.01616281509399414, 0.01623040008544922, 0.01621401596069336, 0.016240640640258788, 0.016286720275878908, 0.016234495162963866, 0.01625497627258301, 0.01621401596069336, 0.016351232528686522, 0.016365568161010743, 0.016229408264160156, 0.01626825523376465, 0.016292863845825196, 0.016194559097290038, 0.01617817687988281, 0.01620070457458496, 0.016151552200317384, 0.016302080154418946, 0.01620582389831543, 0.016250879287719726, 0.016088064193725587, 0.016062463760375977, 0.016493568420410155, 0.01620889663696289, 0.016194559097290038, 0.016242687225341796, 0.016199680328369142, 0.016165887832641602, 0.016288768768310546, 0.01618943977355957, 0.016261119842529297, 0.01624678421020508, 0.01620377540588379, 0.01644339179992676, 0.016234495162963866, 0.01626316833496094, 0.01657241630554199, 0.01700044822692871, 0.016633888244628907, 0.016202720642089843, 0.01618739128112793, 0.01640550422668457, 0.01618124771118164, 0.016285696029663087, 0.016357376098632814, 0.01644441604614258, 0.01678643226623535, 0.016695295333862305, 0.016562175750732423, 0.016128000259399415, 0.01621504020690918, 0.016323583602905273, 0.016528383255004882, 0.016735231399536133, 0.016228351593017578, 0.016149503707885742, 0.01615667152404785, 0.016160768508911134, 0.016157695770263672, 0.03344486236572266, 0.016183296203613282, 0.016316415786743164, 0.01622425651550293, 0.016158720016479493, 0.01620172882080078, 0.01617817687988281, 0.016249856948852538, 0.016202751159667968, 0.016242687225341796, 0.01624678421020508, 0.016199680328369142, 0.01655193519592285, 0.01623142433166504, 0.01621401596069336, 0.01622323226928711, 0.01620172882080078, 0.01617715263366699, 0.016143360137939454, 0.016316415786743164, 0.016213024139404297, 0.016451583862304688, 0.016706527709960936, 0.01660006332397461, 0.016229408264160156, 0.016191455841064455, 0.016210943222045898, 0.016304128646850585, 0.016183296203613282, 0.01617919921875, 0.0162293758392334, 0.016151552200317384, 0.016207872390747072, 0.0161976318359375, 0.016304128646850585, 0.01639833641052246, 0.01638604736328125, 0.01639219284057617, 0.01617715263366699, 0.01623040008544922, 0.016100351333618163, 0.01621299171447754, 0.01626316833496094, 0.016219135284423827, 0.016536575317382812, 0.016694271087646484, 0.01665126419067383, 0.016476160049438478, 0.016942079544067384, 0.016932863235473633, 0.01703731155395508, 0.016926719665527345, 0.016442367553710938, 0.01680281639099121, 0.016696319580078126, 0.016767999649047852, 0.01623756790161133, 0.016317440032958985, 0.016234495162963866, 0.01638604736328125, 0.0184453125, 0.016962560653686523, 0.016866304397583007, 0.03458867263793945, 0.01624678421020508, 0.016277503967285157, 0.01623859214782715, 0.01623961639404297, 0.016559104919433593, 0.016515071868896485, 0.016265216827392577, 0.016252927780151367, 0.016472063064575194, 0.016501760482788085, 0.01621401596069336, 0.016342016220092775, 0.016307199478149414, 0.016207872390747072, 0.016251903533935547, 0.016854015350341797, 0.016319488525390623, 0.01617407989501953, 0.016660480499267577, 0.016258047103881835, 0.016157695770263672, 0.016145408630371092, 0.016306175231933593, 0.016184320449829103, 0.016192512512207033, 0.016154624938964843, 0.016185375213623048, 0.016234464645385742, 0.01627136039733887, 0.016261119842529297, 0.016166912078857423, 0.01618124771118164, 0.016166912078857423, 0.01624575996398926, 0.016261119842529297, 0.016321535110473632, 0.016261119842529297, 0.016310272216796876, 0.0161474552154541, 0.016143360137939454, 0.016207872390747072, 0.016141311645507812, 0.01663795280456543, 0.0162795524597168, 0.016729087829589845, 0.01680179214477539, 0.016707584381103514, 0.01662259292602539, 0.016207872390747072, 0.016175104141235352, 0.01617100715637207, 0.01624678421020508, 0.016265216827392577, 0.016506879806518555, 0.01615667152404785, 0.01634611129760742, 0.016226303100585936, 0.01621196746826172, 0.016289791107177733, 0.01619865608215332, 0.016295936584472655, 0.016232448577880858, 0.033576961517333984, 0.01683148765563965, 0.016562175750732423, 0.016425983428955078, 0.01615667152404785, 0.01635327911376953, 0.016531455993652345, 0.01615564727783203, 0.016305152893066405, 0.01620479965209961, 0.016259071350097656, 0.01619660758972168, 0.016322559356689453, 0.016505855560302735, 0.01617919921875, 0.016327680587768553, 0.01620377540588379, 0.016330751419067382, 0.016325632095336915, 0.0163450870513916, 0.016166912078857423, 0.016281600952148437, 0.016344064712524413, 0.01598464012145996, 0.01600614356994629, 0.01620684814453125, 0.016269311904907227, 0.01624166488647461, 0.01621401596069336, 0.016166912078857423, 0.01643110466003418, 0.0162478084564209, 0.01620070457458496, 0.016368640899658202, 0.016277503967285157, 0.016306175231933593, 0.016323583602905273, 0.016470016479492186, 0.01639116859436035, 0.01621708869934082, 0.016556032180786134, 0.016355327606201172, 0.016416767120361327, 0.016706560134887697, 0.01679769515991211, 0.01676697540283203, 0.01638502311706543, 0.016867328643798828, 0.016672767639160157, 0.016691200256347655, 0.016285696029663087, 0.016282623291015624, 0.016662527084350585, 0.01625395202636719, 0.016311296463012694, 0.016342016220092775, 0.016273408889770507, 0.01618943977355957, 0.01617203140258789, 0.01622118377685547, 0.016335872650146483, 0.016292863845825196, 0.016184320449829103, 0.033347583770751955, 0.01620172882080078, 0.01618739128112793, 0.016285696029663087, 0.016144384384155275, 0.016135168075561524, 0.016236543655395508, 0.01626316833496094, 0.01616998481750488, 0.01621811294555664, 0.01620992088317871, 0.01637068748474121, 0.016356351852416993, 0.016292863845825196, 0.016167936325073243, 0.016332799911499024, 0.01623859214782715, 0.01622323226928711, 0.016237600326538086, 0.01645257568359375, 0.016720895767211915, 0.016760831832885743, 0.016701440811157226, 0.01623347282409668, 0.016333824157714845, 0.01618636894226074, 0.01623961639404297, 0.016364543914794923, 0.01621401596069336, 0.016232448577880858, 0.016207872390747072, 0.016158720016479493, 0.016219135284423827, 0.016154624938964843, 0.016368640899658202, 0.01618227195739746, 0.01625395202636719, 0.016333824157714845, 0.016546815872192384, 0.016202751159667968, 0.016363519668579102, 0.016688127517700196, 0.016680959701538087, 0.016829439163208008, 0.01664204788208008, 0.016312320709228514, 0.01618739128112793, 0.016199680328369142, 0.01619046401977539, 0.0162293758392334, 0.01616896057128906, 0.016118783950805664, 0.016310272216796876, 0.01618435287475586, 0.016152544021606444, 0.01616281509399414, 0.016235519409179687, 0.01613926315307617, 0.016184320449829103, 0.01615564727783203, 0.016267263412475585, 0.016306175231933593, 0.01613926315307617, 0.0333834228515625, 0.01620889663696289, 0.016227327346801757, 0.016262144088745118, 0.016135168075561524, 0.01619353675842285, 0.016143360137939454, 0.016118783950805664, 0.01636761665344238, 0.016312320709228514, 0.016297983169555663, 0.016742399215698242, 0.016307199478149414, 0.0162293758392334, 0.016267263412475585, 0.01617203140258789, 0.016176128387451173, 0.01621196746826172, 0.01616796875, 0.01610441589355469, 0.016270336151123048, 0.016269311904907227, 0.016498687744140626, 0.016684032440185546, 0.016690176010131837, 0.016681983947753908, 0.016674848556518556, 0.016639968872070313, 0.016927743911743166, 0.016721920013427736, 0.016792575836181642, 0.016303104400634767, 0.016249856948852538, 0.01620479965209961, 0.01639321517944336, 0.016719871520996094, 0.016694271087646484, 0.01622220802307129, 0.016495616912841796, 0.016726015090942382, 0.01678950309753418, 0.01657344055175781, 0.016466943740844727, 0.01618534469604492, 0.01622118377685547, 0.016153600692749022, 0.016136192321777345, 0.016167936325073243, 0.01618534469604492, 0.016325632095336915, 0.016228351593017578, 0.01622323226928711, 0.016192512512207033, 0.01617305564880371, 0.016251903533935547, 0.016216064453125, 0.016151552200317384, 0.016194559097290038, 0.0161474552154541, 0.016720928192138673, 0.016764896392822266, 0.016318464279174806, 0.01621708869934082, 0.033334270477294925, 0.01618739128112793, 0.01618534469604492, 0.016296960830688476, 0.016485376358032225, 0.01620684814453125, 0.016141311645507812, 0.01618943977355957, 0.016322559356689453, 0.016303104400634767, 0.016207872390747072, 0.016191488265991212, 0.016226367950439455, 0.016618431091308592, 0.01618739128112793, 0.01620479965209961, 0.016159744262695314, 0.016256000518798826, 0.01614847946166992, 0.01617919921875, 0.015962112426757814, 0.01598259162902832, 0.016281600952148437, 0.0163768310546875, 0.01622220802307129, 0.01624678421020508, 0.016262144088745118, 0.016157695770263672, 0.01617001533508301, 0.016210912704467773, 0.016457727432250976, 0.016157695770263672, 0.01640755271911621, 0.01617817687988281, 0.016266239166259765, 0.01618534469604492, 0.016772096633911132, 0.017090560913085938, 0.01696460723876953, 0.0174335994720459, 0.016855039596557618, 0.016877567291259766, 0.016698368072509767, 0.016594944000244142, 0.01618841552734375, 0.016291839599609375, 0.016662527084350585, 0.016294912338256837, 0.016352256774902343, 0.016389120101928712, 0.016326656341552736, 0.016390144348144533, 0.016296960830688476, 0.0164270076751709, 0.01634611129760742, 0.01620377540588379, 0.016167936325073243, 0.016143360137939454, 0.0162795524597168, 0.016259071350097656, 0.01618739128112793, 0.016858112335205077, 0.016731136322021483, 0.03470848083496094, 0.01662259292602539, 0.016287744522094725, 0.01625497627258301, 0.01618332862854004, 0.01624060821533203, 0.01616486358642578, 0.016199680328369142, 0.01616383934020996, 0.016667648315429686, 0.0164270076751709, 0.01636249542236328, 0.016160768508911134, 0.016291839599609375, 0.016152576446533205, 0.01622528076171875, 0.01619660758972168, 0.01622118377685547, 0.016161792755126952, 0.016153600692749022, 0.01617715263366699, 0.01619353675842285, 0.01620889663696289, 0.016124927520751953, 0.015997952461242675, 0.0164270076751709, 0.016098304748535155, 0.01613209533691406, 0.01616383934020996, 0.01646899223327637, 0.016352256774902343, 0.01616486358642578, 0.01621196746826172, 0.01618124771118164, 0.016833536148071288, 0.016727039337158203, 0.016483327865600587, 0.01638400077819824, 0.016275455474853515, 0.016220159530639648, 0.01618022346496582, 0.01618227195739746, 0.01613209533691406, 0.01617203140258789, 0.01623961639404297, 0.016344064712524413, 0.016881664276123046, 0.017074207305908203, 0.017734624862670897, 0.016886783599853517, 0.01680588722229004, 0.01680076789855957, 0.01656934356689453, 0.016191488265991212, 0.016295936584472655, 0.01617407989501953, 0.016850944519042968, 0.016727039337158203, 0.01665843200683594, 0.016336896896362304, 0.016302080154418946, 0.016258047103881835, 0.01617817687988281]",tokens/s,60.041658946126375,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-6B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,3691.425792,5211.947008,0.0,4565.499904,4315.13088,s,10,2.8053516845703124,0.2805351684570313,0.0004480646012681654,0.28061317443847655,0.28096784057617186,0.2810714141845703,0.28115427307128904,"[0.2804563903808594, 0.27977456665039063, 0.28117498779296873, 0.28078695678710935, 0.2808992919921875, 0.28094482421875, 0.27981097412109374, 0.2802847290039063, 0.28076995849609376, 0.28044900512695314]",tokens/s,912.5415590780402,kWh,3.3092777670165643e-06,1.813301800711997e-06,1.6326147320167017e-05,2.1448726887895576e-05,tokens/kWh,11935440.33349931,MB,3691.425792,5211.947008,0.0,4565.499904,4383.673344,s,10,151.18848046875,15.118848046874998,0.0027821151656467562,15.11984619140625,15.12149365234375,15.121812255859375,15.122067138671875,"[15.1142509765625, 15.114376953125, 15.116474609375, 15.1179326171875, 15.1195205078125, 15.120171875, 15.1214228515625, 15.122130859375, 15.121267578125, 15.120931640625]",tokens/s,4.166984138253961,kWh,0.0001784252902791824,9.779173102294405e-05,0.0008596400256744463,0.001135857046976573,tokens/kWh,55464.72610060708,,s,629,153.47323796081554,0.24399560884072405,0.03338408494625594,0.23994061279296874,0.24014356079101562,0.24033689270019531,0.5206728076171875,"[0.23988018798828126, 0.23966412353515626, 0.2399303741455078, 0.23972662353515625, 0.23980029296875, 0.23966720581054687, 0.23980441284179688, 0.23968870544433593, 0.23981260681152344, 0.23969381713867188, 0.2398197784423828, 0.23980441284179688, 0.23979417419433594, 0.23982591247558593, 0.23978297424316405, 0.23968966674804687, 0.2397460479736328, 0.23982591247558593, 0.24002867126464844, 0.23977267456054688, 0.2398719940185547, 0.2399139862060547, 0.2398586883544922, 0.23982899475097658, 0.23979212951660156, 0.23972659301757812, 0.23980441284179688, 0.2397460479736328, 0.23980645751953125, 0.23991807556152345, 0.23991194152832032, 0.23983103942871092, 0.24007168579101562, 0.2401658935546875, 0.23983314514160156, 0.23976441955566405, 0.2398760986328125, 0.2399006652832031, 0.2398525390625, 0.23990988159179688, 0.2398392333984375, 0.239815673828125, 0.24020889282226562, 0.24050277709960938, 0.24006040954589844, 0.2399898223876953, 0.24000096130371093, 0.23991807556152345, 0.2399252471923828, 0.23990170288085938, 0.24004920959472656, 0.23989138793945314, 0.23997439575195312, 0.2399938507080078, 0.23991500854492187, 0.23999897766113282, 0.2399692840576172, 0.23991909790039062, 0.24029183959960937, 0.2402109375, 0.24019148254394532, 0.24002252197265625, 0.52034765625, 0.23986277770996095, 0.23982386779785156, 0.2398269500732422, 0.2397389373779297, 0.23981254577636718, 0.23979008483886718, 0.24015155029296875, 0.23990476989746093, 0.24000921630859376, 0.23982182312011718, 0.23983103942871092, 0.23984742736816406, 0.23975628662109374, 0.23986892700195311, 0.23984335327148437, 0.23982691955566407, 0.23977369689941405, 0.23979212951660156, 0.23984742736816406, 0.23972352600097657, 0.2398638153076172, 0.23997337341308594, 0.2398587188720703, 0.23978697204589844, 0.23989356994628908, 0.23981459045410156, 0.23986994934082032, 0.2397532196044922, 0.23984742736816406, 0.2397286376953125, 0.23983103942871092, 0.23970611572265624, 0.23982591247558593, 0.2397081604003906, 0.23984844970703126, 0.23969279479980468, 0.23982803344726564, 0.23986784362792968, 0.239815673828125, 0.24011468505859376, 0.240505859375, 0.23988946533203126, 0.23987193298339843, 0.23994367980957032, 0.24006451416015626, 0.23992422485351564, 0.23987303161621093, 0.2399805450439453, 0.2400358428955078, 0.23991807556152345, 0.23994163513183595, 0.23998258972167968, 0.23992831420898436, 0.23990272521972655, 0.2401300506591797, 0.24004914855957032, 0.24002969360351561, 0.239963134765625, 0.24010342407226562, 0.24004914855957032, 0.2398638153076172, 0.24026931762695314, 0.521860107421875, 0.23985459899902345, 0.239724609375, 0.2399866180419922, 0.23972557067871095, 0.2398586883544922, 0.23977165222167968, 0.23977471923828125, 0.23968666076660156, 0.23982797241210937, 0.2397081604003906, 0.2398085174560547, 0.23981465148925782, 0.23982284545898438, 0.2398586883544922, 0.23999282836914063, 0.23994061279296874, 0.23976345825195314, 0.23983616638183594, 0.23996620178222655, 0.23977880859375, 0.23979519653320314, 0.23982386779785156, 0.24098713684082032, 0.2398402557373047, 0.23995391845703126, 0.23997337341308594, 0.2400184326171875, 0.23989663696289062, 0.23991600036621094, 0.23995695495605468, 0.23978598022460937, 0.23989964294433594, 0.2398423309326172, 0.24002351379394532, 0.239815673828125, 0.2398638153076172, 0.23993753051757813, 0.23985971069335937, 0.23986790466308594, 0.23990579223632813, 0.24008090209960936, 0.2399550018310547, 0.24002554321289063, 0.24000204467773437, 0.239994873046875, 0.23996006774902343, 0.23991705322265625, 0.23987405395507813, 0.23994265747070312, 0.23987814331054688, 0.23996109008789063, 0.23988327026367187, 0.23988633728027345, 0.23988018798828126, 0.23990684509277344, 0.24002761840820314, 0.2401781768798828, 0.23998565673828126, 0.24023143005371095, 0.24086528015136718, 0.24008607482910158, 0.24003680419921875, 0.5212160034179687, 0.23997850036621093, 0.23983718872070312, 0.23975935363769532, 0.23993446350097655, 0.2399569854736328, 0.2398321228027344, 0.23992825317382813, 0.23975526428222657, 0.2398904266357422, 0.23975526428222657, 0.23980441284179688, 0.24054579162597656, 0.23996826171875, 0.23982591247558593, 0.23983512878417967, 0.23973068237304687, 0.2400921630859375, 0.2404997100830078, 0.2399621124267578, 0.23977984619140624, 0.23988223266601563, 0.23981773376464843, 0.23980953979492187, 0.23986073303222658, 0.23993959045410157, 0.23980953979492187, 0.23989657592773436, 0.23987405395507813, 0.23991705322265625, 0.23988327026367187, 0.2399385528564453, 0.23993959045410157, 0.2399385528564453, 0.2399692840576172, 0.2400358428955078, 0.2399518737792969, 0.23998873901367188, 0.24005734252929686, 0.24006246948242188, 0.23999897766113282, 0.24004608154296875, 0.2399252471923828, 0.24001434326171875, 0.24022425842285156, 0.24002969360351561, 0.24000614929199218, 0.23993446350097655, 0.2399006652832031, 0.24005836486816406, 0.23999180603027342, 0.24003379821777343, 0.2398709716796875, 0.2400133056640625, 0.24013618469238282, 0.24014028930664064, 0.23999078369140625, 0.239931396484375, 0.24003482055664063, 0.24005427551269531, 0.24012185668945313, 0.2399006652832031, 0.2399764404296875, 0.5212764282226563, 0.2398955535888672, 0.2397972412109375, 0.23982080078125, 0.23980543518066405, 0.23985466003417968, 0.239880126953125, 0.23983718872070312, 0.2398330841064453, 0.2399272918701172, 0.23977880859375, 0.24003482055664063, 0.24019148254394532, 0.23990988159179688, 0.24018637084960937, 0.2408284149169922, 0.23992218017578126, 0.24102093505859376, 0.2400010223388672, 0.24018022155761717, 0.23982489013671876, 0.23983206176757813, 0.2399518737792969, 0.23996517944335938, 0.2401443786621094, 0.24013926696777343, 0.23987916564941406, 0.2399303741455078, 0.23980441284179688, 0.23997439575195312, 0.23980543518066405, 0.23988633728027345, 0.23987916564941406, 0.2398402557373047, 0.23995289611816406, 0.2399518737792969, 0.23989759826660156, 0.2400358428955078, 0.23986895751953125, 0.23993545532226562, 0.23991194152832032, 0.2400430145263672, 0.2400870361328125, 0.23996109008789063, 0.23996931457519532, 0.23996617126464845, 0.23985459899902345, 0.23999180603027342, 0.23983821105957032, 0.23998771667480467, 0.2400552978515625, 0.24013516235351562, 0.24004608154296875, 0.24011776733398438, 0.24005427551269531, 0.239951904296875, 0.24000405883789064, 0.24009829711914063, 0.23997850036621093, 0.24004095458984376, 0.23992115783691406, 0.24001023864746093, 0.23995802307128905, 0.5214299926757813, 0.2398904266357422, 0.23992626953125, 0.23979417419433594, 0.23976039123535156, 0.23995494079589844, 0.23974195861816405, 0.23989759826660156, 0.24016390991210937, 0.24046278381347655, 0.23983616638183594, 0.24008908081054686, 0.24008090209960936, 0.23978189086914062, 0.23984230041503907, 0.24015469360351563, 0.23977261352539062, 0.23997030639648437, 0.23999078369140625, 0.23992218017578126, 0.23984538269042968, 0.23995802307128905, 0.23983206176757813, 0.2399621124267578, 0.23989453125, 0.2399488067626953, 0.23983821105957032, 0.2400245819091797, 0.23998976135253905, 0.23988327026367187, 0.24005433654785155, 0.2403552703857422, 0.23997541809082032, 0.23996415710449218, 0.239815673828125, 0.24005535888671875, 0.23999891662597655, 0.24001638793945312, 0.24002867126464844, 0.24005938720703124, 0.2399518737792969, 0.23995802307128905, 0.23997235107421874, 0.2401730499267578, 0.24023960876464845, 0.24008192443847656, 0.23993959045410157, 0.24002252197265625, 0.24006144714355468, 0.2401239013671875, 0.23988327026367187, 0.23995596313476564, 0.24006655883789063, 0.24059504699707032, 0.23992515563964845, 0.24001126098632813, 0.23989144897460937, 0.24023660278320313, 0.2400654754638672, 0.23997439575195312, 0.24005938720703124, 0.23999795532226562, 0.2399447021484375, 0.5203138427734375, 0.23985049438476563, 0.23983512878417967, 0.24011878967285155, 0.23979110717773439, 0.24002560424804686, 0.2399498291015625, 0.23987303161621093, 0.23984640502929688, 0.2400205078125, 0.23974703979492187, 0.23997541809082032, 0.24000921630859376, 0.24002867126464844, 0.23995802307128905, 0.23987712097167968, 0.23981158447265624, 0.2400184326171875, 0.2400358428955078, 0.24000921630859376, 0.24026521301269532, 0.24031948852539062, 0.2399692840576172, 0.24197427368164062, 0.23986585998535156, 0.239857666015625, 0.23988429260253907, 0.24009727478027343, 0.2399252471923828, 0.23986688232421874, 0.23998573303222656, 0.23984736633300782, 0.23986483764648436, 0.2402037811279297, 0.24042501831054688, 0.23991084289550782, 0.23989144897460937, 0.23988838195800782, 0.23997030639648437, 0.23996415710449218, 0.23990885925292968, 0.23982899475097658, 0.240005126953125, 0.24034098815917967, 0.23998565673828126, 0.23986994934082032, 0.2398402557373047, 0.23994985961914062, 0.24000611877441405, 0.23993653869628906, 0.23998358154296875, 0.24002867126464844, 0.23993344116210938, 0.24010137939453124, 0.24018739318847657, 0.24002969360351561, 0.24010855102539064, 0.240005126953125, 0.24002969360351561, 0.24001741027832033, 0.23994473266601563, 0.24000918579101563, 0.2401126403808594, 0.5207992553710937, 0.2400235595703125, 0.23988531494140625, 0.23978802490234374, 0.23990374755859376, 0.23999897766113282, 0.24003890991210938, 0.23992626953125, 0.2399498291015625, 0.24010240173339845, 0.2398023681640625, 0.23996620178222655, 0.23992115783691406, 0.23988018798828126, 0.23982591247558593, 0.2398607940673828, 0.24102188110351563, 0.23994163513183595, 0.23991705322265625, 0.23988429260253907, 0.23991807556152345, 0.23994061279296874, 0.23998361206054689, 0.23994061279296874, 0.24013926696777343, 0.24000204467773437, 0.23999282836914063, 0.24006349182128905, 0.23995802307128905, 0.23995596313476564, 0.24011672973632814, 0.24034815979003907, 0.24002662658691407, 0.23992422485351564, 0.23982591247558593, 0.2399447021484375, 0.2398587188720703, 0.2399190673828125, 0.23990988159179688, 0.23986997985839845, 0.23992828369140626, 0.24009011840820313, 0.23994163513183595, 0.24005938720703124, 0.23999078369140625, 0.2400245819091797, 0.23992218017578126, 0.24194047546386718, 0.24000306701660157, 0.24009625244140625, 0.2400122833251953, 0.23997439575195312, 0.24015155029296875, 0.24044338989257813, 0.24007577514648437, 0.24002252197265625, 0.23997850036621093, 0.23997850036621093, 0.24000408935546874, 0.239889404296875, 0.23992626953125, 0.23996006774902343, 0.24005734252929686, 0.521218017578125, 0.2397398986816406, 0.23977369689941405, 0.2399139862060547, 0.23973683166503906, 0.23983718872070312, 0.23993548583984375, 0.23984538269042968, 0.23984742736816406, 0.24001434326171875, 0.23988735961914062, 0.2399201354980469, 0.23986994934082032, 0.24014335632324219, 0.23980645751953125, 0.23989453125, 0.2400552978515625, 0.24133631896972657, 0.2404874267578125, 0.2400880584716797, 0.23997132873535157, 0.23975529479980467, 0.23980848693847656, 0.2399385528564453, 0.23981263732910157, 0.23990882873535158, 0.2397962188720703, 0.23988018798828126, 0.23988531494140625, 0.23995085144042969, 0.23991091918945312, 0.2400870361328125, 0.2400860137939453, 0.23994163513183595, 0.2398719940185547, 0.23988018798828126, 0.23981362915039062, 0.239931396484375, 0.23982284545898438, 0.23989144897460937, 0.2398904266357422, 0.239963134765625, 0.23995802307128905, 0.23987712097167968, 0.2399252471923828, 0.23997030639648437, 0.23991909790039062, 0.24007577514648437, 0.23990579223632813, 0.2402529296875, 0.240247802734375, 0.24222515869140626, 0.24009420776367188, 0.24000408935546874, 0.24003173828125, 0.23993344116210938, 0.23996826171875, 0.240110595703125, 0.23997541809082032, 0.2399447021484375, 0.24019664001464844, 0.24008087158203126, 0.240184326171875, 0.5212047119140625, 0.23991091918945312, 0.23979417419433594, 0.24003071594238282, 0.23989453125, 0.2399252471923828, 0.2398935089111328, 0.24010240173339845, 0.24033074951171876, 0.23999897766113282, 0.23993650817871093, 0.23987916564941406, 0.23988633728027345, 0.23987712097167968, 0.23985459899902345, 0.2398197784423828, 0.23977676391601563, 0.2399447021484375, 0.23983929443359375, 0.23985658264160156, 0.2397972412109375, 0.23987303161621093, 0.23984950256347656, 0.23986376953125, 0.23986790466308594, 0.2400245819091797, 0.24176025390625, 0.24003382873535156, 0.2399610595703125, 0.23991091918945312, 0.23988018798828126, 0.24049766540527343, 0.23994367980957032, 0.23988633728027345, 0.2399447021484375, 0.24010342407226562, 0.2399447021484375, 0.2399764404296875, 0.23998565673828126, 0.24010855102539064, 0.2400625, 0.24000918579101563, 0.23999180603027342, 0.23997850036621093, 0.23994163513183595, 0.23994265747070312, 0.2400911407470703, 0.23993959045410157, 0.23996723937988282, 0.24002149963378908, 0.24012393188476563, 0.24004707336425782, 0.2400911407470703, 0.2402406463623047, 0.24006655883789063, 0.240005126953125, 0.24013618469238282, 0.24016793823242188, 0.239857666015625, 0.23995904541015625, 0.23997132873535157, 0.2398935089111328, 0.23992320251464844]",tokens/s,4.098434413435621,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen2-beta-72B,Qwen/Qwen2-beta-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12706,7 +12706,7 @@ OSError: 8 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12834,7 +12834,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12893,7 +12893,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1443.536896,2426.929152,0.0,1780.482048,1621.410304,s,10,0.7068105316162109,0.07068105316162109,0.00033069583366084575,0.07070081710815429,0.07103355484008789,0.07120412101745605,0.07134057395935059,"[0.07069222259521485, 0.0702221450805664, 0.07099565124511718, 0.07054032135009766, 0.07070941162109375, 0.07019967651367187, 0.07137468719482422, 0.07050953674316406, 0.07078886413574219, 0.07077801513671875]",tokens/s,3621.904153219448,kWh,8.280014542525558e-07,4.5352803657666435e-07,3.715746950839271e-06,4.997276441668491e-06,tokens/kWh,51227904.43718713,MB,1443.536896,2426.929152,0.0,1780.482048,1723.272192,s,10,38.065419677734376,3.8065419677734376,0.002289567625927924,3.8060361328125003,3.8076869140625003,3.8103506591796874,3.8124816552734373,"[3.813014404296875, 3.806056640625, 3.804952880859375, 3.806771484375, 3.8046533203125, 3.805589599609375, 3.80634765625, 3.806015625, 3.807094970703125, 3.804923095703125]",tokens/s,16.55045459458066,kWh,4.493906886566492e-05,2.4628168189011863e-05,0.00019878599838376527,0.000268353235438442,tokens/kWh,234765.19631697034,,s,629,38.640658420562765,0.061431889380862865,0.008435353672886576,0.06039961624145508,0.06047948837280273,0.06056448059082031,0.13117497192382813,"[0.06107852935791016, 0.06100889587402344, 0.06097612762451172, 0.06086963272094727, 0.061018112182617185, 0.06080614471435547, 0.06082867050170898, 0.06069145584106445, 0.06044672012329102, 0.060388351440429686, 0.06043033599853516, 0.06039449691772461, 0.06051430511474609, 0.060368896484375, 0.06042828750610352, 0.06034124755859375, 0.06040063858032227, 0.0605747184753418, 0.06040063858032227, 0.06043340682983398, 0.06040063858032227, 0.06036070251464844, 0.06043648147583008, 0.06034329605102539, 0.060440574645996094, 0.060382209777832034, 0.060477439880371096, 0.060386302947998044, 0.060402687072753904, 0.06038937759399414, 0.06038425445556641, 0.060371967315673826, 0.06056140899658203, 0.060434432983398435, 0.06046003341674805, 0.0605665283203125, 0.0605030403137207, 0.06035968017578125, 0.06042828750610352, 0.060386302947998044, 0.06044672012329102, 0.060416000366210934, 0.060462078094482424, 0.060398593902587894, 0.06042316818237305, 0.060455936431884766, 0.06044876861572265, 0.06036070251464844, 0.06043135833740235, 0.06038323211669922, 0.060483585357666014, 0.06042316818237305, 0.060598270416259765, 0.06038118362426758, 0.06044569778442383, 0.06175641632080078, 0.060464126586914066, 0.060442623138427735, 0.06043648147583008, 0.06043545532226562, 0.060418048858642576, 0.060424190521240234, 0.13181951904296876, 0.060440574645996094, 0.060472320556640625, 0.060391422271728515, 0.06032588958740234, 0.0603955192565918, 0.060331008911132813, 0.060371967315673826, 0.06035148620605469, 0.06034431838989258, 0.06032179260253906, 0.060355583190917966, 0.060352512359619144, 0.060385280609130856, 0.0603422737121582, 0.06037913513183594, 0.06034124755859375, 0.06037606430053711, 0.06037401580810547, 0.06042214584350586, 0.060349441528320315, 0.06038118362426758, 0.06040063858032227, 0.06038118362426758, 0.06034534454345703, 0.06042009735107422, 0.06041190338134766, 0.06039961624145508, 0.0603770866394043, 0.06039244842529297, 0.06035660934448242, 0.060388351440429686, 0.06036172866821289, 0.06043135833740235, 0.06046515274047851, 0.06044364929199219, 0.06042316818237305, 0.060483585357666014, 0.06034841537475586, 0.06043648147583008, 0.06038016128540039, 0.060409854888916016, 0.06041497421264649, 0.06037606430053711, 0.06034739303588867, 0.06041190338134766, 0.060437503814697265, 0.06042009735107422, 0.060368896484375, 0.060365825653076174, 0.06037094497680664, 0.06043545532226562, 0.06042316818237305, 0.06043852615356445, 0.060368896484375, 0.060437503814697265, 0.06047436904907227, 0.060375038146972655, 0.060437503814697265, 0.06037606430053711, 0.06049280166625977, 0.060416000366210934, 0.060424190521240234, 0.13166899108886718, 0.06033919906616211, 0.06037094497680664, 0.0603873291015625, 0.060402687072753904, 0.06045798492431641, 0.060355583190917966, 0.06040678405761719, 0.0603504638671875, 0.06037401580810547, 0.06034636688232422, 0.06035865783691406, 0.060352512359619144, 0.06037299346923828, 0.06032588958740234, 0.060369918823242184, 0.06034841537475586, 0.060391422271728515, 0.06036787033081055, 0.060404735565185545, 0.060355583190917966, 0.0603873291015625, 0.06036070251464844, 0.06039756774902344, 0.06034636688232422, 0.06039244842529297, 0.06037094497680664, 0.060412960052490236, 0.060467166900634764, 0.060371967315673826, 0.06035456085205078, 0.0603873291015625, 0.06039449691772461, 0.06044672012329102, 0.06039654541015625, 0.06040063858032227, 0.060439552307128906, 0.06043135833740235, 0.06035148620605469, 0.060416000366210934, 0.06039039993286133, 0.06038425445556641, 0.060385280609130856, 0.06039654541015625, 0.06032896041870117, 0.06044364929199219, 0.06042316818237305, 0.06042931365966797, 0.060483585357666014, 0.06036377716064453, 0.060401664733886716, 0.06041907119750976, 0.06040576171875, 0.060402687072753904, 0.06036787033081055, 0.06038425445556641, 0.06044467163085938, 0.060409854888916016, 0.060437503814697265, 0.060369918823242184, 0.06047641754150391, 0.060407806396484375, 0.06041702270507812, 0.13118156433105468, 0.06036684799194336, 0.06035763168334961, 0.0603770866394043, 0.06031155014038086, 0.06035968017578125, 0.06035148620605469, 0.06037606430053711, 0.06033817672729492, 0.060862464904785155, 0.06047129440307617, 0.060401664733886716, 0.0603504638671875, 0.06036684799194336, 0.060369918823242184, 0.060418048858642576, 0.06037606430053711, 0.06039039993286133, 0.06034534454345703, 0.06039244842529297, 0.06033203125, 0.06042316818237305, 0.06031872177124024, 0.06051737594604492, 0.06032896041870117, 0.06043852615356445, 0.06038937759399414, 0.0605296630859375, 0.060375038146972655, 0.06068735885620117, 0.06043033599853516, 0.06040678405761719, 0.0603504638671875, 0.060437503814697265, 0.06040883255004883, 0.06041497421264649, 0.060470272064208984, 0.060434432983398435, 0.060523521423339846, 0.06047334289550781, 0.06040576171875, 0.06041190338134766, 0.060385280609130856, 0.06039756774902344, 0.0603770866394043, 0.06044876861572265, 0.06042726516723633, 0.06041907119750976, 0.06034124755859375, 0.060368896484375, 0.06036479949951172, 0.06041292953491211, 0.060382209777832034, 0.06074265670776367, 0.06057984161376953, 0.06045183944702148, 0.060477439880371096, 0.06036787033081055, 0.0604139518737793, 0.06040883255004883, 0.06046515274047851, 0.060416000366210934, 0.06042828750610352, 0.13133517456054689, 0.06037299346923828, 0.060365825653076174, 0.06038937759399414, 0.060333057403564455, 0.060365825653076174, 0.06033919906616211, 0.060352512359619144, 0.06033407974243164, 0.06035763168334961, 0.06035456085205078, 0.06034124755859375, 0.060336128234863284, 0.06036684799194336, 0.06036479949951172, 0.06038425445556641, 0.06032691192626953, 0.060369918823242184, 0.06036070251464844, 0.06038118362426758, 0.06033203125, 0.06039039993286133, 0.06033407974243164, 0.06036275100708008, 0.060355583190917966, 0.060418048858642576, 0.060385280609130856, 0.06042009735107422, 0.060331008911132813, 0.0603955192565918, 0.060355583190917966, 0.060385280609130856, 0.06039244842529297, 0.06041702270507812, 0.060426239013671876, 0.060424190521240234, 0.06043033599853516, 0.06041497421264649, 0.060355583190917966, 0.06041292953491211, 0.06036275100708008, 0.060407806396484375, 0.06040678405761719, 0.06038323211669922, 0.060437503814697265, 0.06044364929199219, 0.06044569778442383, 0.06041907119750976, 0.060353534698486325, 0.060418048858642576, 0.060385280609130856, 0.060416000366210934, 0.06038937759399414, 0.0604139518737793, 0.060385280609130856, 0.06043648147583008, 0.06046822357177734, 0.0603504638671875, 0.06047129440307617, 0.06038425445556641, 0.06048051071166992, 0.06044876861572265, 0.0604139518737793, 0.130845703125, 0.06035456085205078, 0.06035968017578125, 0.0603955192565918, 0.0603238410949707, 0.06036479949951172, 0.060352512359619144, 0.06036787033081055, 0.060355583190917966, 0.06032896041870117, 0.0603873291015625, 0.06034534454345703, 0.06035865783691406, 0.06034739303588867, 0.06038118362426758, 0.0603504638671875, 0.06037094497680664, 0.060385280609130856, 0.06036377716064453, 0.060401664733886716, 0.06036377716064453, 0.060391422271728515, 0.06037811279296875, 0.06038937759399414, 0.06030950546264648, 0.060393470764160156, 0.060416000366210934, 0.060475391387939455, 0.06039756774902344, 0.06038425445556641, 0.06041497421264649, 0.06039654541015625, 0.06035865783691406, 0.06042316818237305, 0.06040883255004883, 0.06039449691772461, 0.06043545532226562, 0.06064025497436523, 0.060382209777832034, 0.060442623138427735, 0.06040063858032227, 0.060409854888916016, 0.06037606430053711, 0.06038425445556641, 0.06032896041870117, 0.06041702270507812, 0.060424190521240234, 0.06051839828491211, 0.060369918823242184, 0.06036787033081055, 0.06040576171875, 0.06050099182128906, 0.060424190521240234, 0.060402687072753904, 0.060368896484375, 0.060486656188964844, 0.06061670303344727, 0.06039654541015625, 0.06042214584350586, 0.060404735565185545, 0.060560382843017575, 0.06044467163085938, 0.06042009735107422, 0.13205914306640626, 0.06037299346923828, 0.06082252883911133, 0.06047436904907227, 0.06030950546264648, 0.06037811279296875, 0.060336128234863284, 0.060368896484375, 0.06032281494140625, 0.06037811279296875, 0.06031564712524414, 0.06044876861572265, 0.0604661750793457, 0.06039039993286133, 0.060353534698486325, 0.060386302947998044, 0.06032896041870117, 0.060382209777832034, 0.06035865783691406, 0.06039756774902344, 0.06032998275756836, 0.06042931365966797, 0.06038425445556641, 0.060442623138427735, 0.06044876861572265, 0.060391422271728515, 0.0603422737121582, 0.06039654541015625, 0.0604139518737793, 0.06036479949951172, 0.060353534698486325, 0.06038937759399414, 0.060440574645996094, 0.06048051071166992, 0.06043135833740235, 0.06040063858032227, 0.06046822357177734, 0.06043545532226562, 0.060393470764160156, 0.060421119689941405, 0.06042521667480469, 0.060450817108154295, 0.060391422271728515, 0.06040063858032227, 0.060365825653076174, 0.060416000366210934, 0.06047641754150391, 0.06045183944702148, 0.06034124755859375, 0.06060441589355469, 0.060424190521240234, 0.060447742462158206, 0.06042828750610352, 0.06047948837280273, 0.06037299346923828, 0.060418048858642576, 0.06048972702026367, 0.060391422271728515, 0.06043135833740235, 0.060369918823242184, 0.06054195022583008, 0.06044876861572265, 0.06041088104248047, 0.13115802001953125, 0.06042828750610352, 0.06035763168334961, 0.06038323211669922, 0.06030847930908203, 0.06037913513183594, 0.06053887939453125, 0.06036684799194336, 0.06036070251464844, 0.06045695877075195, 0.060365825653076174, 0.060375038146972655, 0.06034431838989258, 0.060369918823242184, 0.06037299346923828, 0.060369918823242184, 0.06036172866821289, 0.060494846343994144, 0.060402687072753904, 0.060462078094482424, 0.06046105575561524, 0.06041907119750976, 0.060421119689941405, 0.06037913513183594, 0.06034636688232422, 0.06039449691772461, 0.06035456085205078, 0.06039247894287109, 0.06034223937988281, 0.060368896484375, 0.060440574645996094, 0.06048153686523437, 0.06046105575561524, 0.060442623138427735, 0.060488704681396485, 0.06050201416015625, 0.060416000366210934, 0.06039449691772461, 0.060382209777832034, 0.06044569778442383, 0.06039654541015625, 0.06040371322631836, 0.06044569778442383, 0.060369918823242184, 0.060401664733886716, 0.0604139518737793, 0.06041497421264649, 0.06039654541015625, 0.06044364929199219, 0.0603873291015625, 0.06037606430053711, 0.060434432983398435, 0.06038425445556641, 0.060424190521240234, 0.06037913513183594, 0.060407806396484375, 0.060470272064208984, 0.06036070251464844, 0.06043648147583008, 0.060527614593505856, 0.060483585357666014, 0.06039961624145508, 0.06042009735107422, 0.1314693145751953, 0.060409854888916016, 0.06036377716064453, 0.060426273345947266, 0.06032380676269531, 0.06034636688232422, 0.06032998275756836, 0.06037811279296875, 0.06034841537475586, 0.06038118362426758, 0.060472320556640625, 0.060371967315673826, 0.06034534454345703, 0.060349441528320315, 0.06040063858032227, 0.06034329605102539, 0.06041907119750976, 0.06053376007080078, 0.06042931365966797, 0.060382209777832034, 0.06047948837280273, 0.060418048858642576, 0.06037094497680664, 0.06039756774902344, 0.06042214584350586, 0.06038118362426758, 0.06034329605102539, 0.06052556610107422, 0.06041190338134766, 0.06060851287841797, 0.06038937759399414, 0.06039961624145508, 0.06040371322631836, 0.06044160079956055, 0.06049792098999023, 0.0604139518737793, 0.060505088806152345, 0.06044160079956055, 0.06043135833740235, 0.06070169448852539, 0.060440574645996094, 0.06040576171875, 0.06042828750610352, 0.060391422271728515, 0.060369918823242184, 0.060447742462158206, 0.06047641754150391, 0.06041907119750976, 0.060447742462158206, 0.06072217559814453, 0.060470272064208984, 0.060469249725341796, 0.06040063858032227, 0.060439552307128906, 0.060385280609130856, 0.06046310424804688, 0.06047436904907227, 0.06038937759399414, 0.060418048858642576, 0.06040576171875, 0.060464126586914066, 0.06039039993286133, 0.06039244842529297, 0.13141401672363281, 0.06035456085205078, 0.060371967315673826, 0.06040883255004883, 0.0603422737121582, 0.060402687072753904, 0.06033715057373047, 0.06034022521972656, 0.06035660934448242, 0.06035763168334961, 0.06033817672729492, 0.060368896484375, 0.060333057403564455, 0.060388351440429686, 0.060368896484375, 0.060393470764160156, 0.06032896041870117, 0.06041292953491211, 0.06034739303588867, 0.06040063858032227, 0.06032793426513672, 0.06037913513183594, 0.06035456085205078, 0.0603770866394043, 0.06034124755859375, 0.06042828750610352, 0.06037401580810547, 0.060442623138427735, 0.06034739303588867, 0.060388351440429686, 0.06035968017578125, 0.060409854888916016, 0.060355583190917966, 0.060442623138427735, 0.06044364929199219, 0.060437503814697265, 0.06044876861572265, 0.06044467163085938, 0.0603770866394043, 0.06041292953491211, 0.060386302947998044, 0.06039961624145508, 0.06038118362426758, 0.06043340682983398, 0.060369918823242184, 0.06043033599853516, 0.06046515274047851, 0.06043033599853516, 0.0603351058959961, 0.060391422271728515, 0.06037401580810547, 0.06042521667480469, 0.0603955192565918, 0.06042009735107422, 0.060371967315673826, 0.06050406265258789, 0.0604661750793457, 0.06038425445556641, 0.06045183944702148, 0.060386302947998044, 0.06047129440307617, 0.060393470764160156, 0.06040678405761719]",tokens/s,16.278190530658136,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12935,7 +12935,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13354,7 +13354,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2418.307072,9028.763648,0.0,8382.316544,7939.283456,s,10,6.167790039062499,0.61677900390625,0.0015724819730622269,0.6163188171386719,0.6186426635742188,0.6196823547363282,0.6205141076660157,"[0.6154488525390625, 0.6207220458984375, 0.6163983764648437, 0.616369384765625, 0.6161586303710938, 0.6162682495117188, 0.6171541137695312, 0.61841162109375, 0.6153932495117187, 0.6154655151367188]",tokens/s,415.059524365573,kWh,7.267181906435225e-06,3.982036919842357e-06,3.6222087801172345e-05,4.747130662744993e-05,tokens/kWh,5392731.276791313,MB,2418.307072,9028.763648,0.0,8382.316544,8233.862144,s,10,338.13460546874995,33.81346054687499,0.001160851078231068,33.813755859375,33.814566015625,33.8146169921875,33.8146577734375,"[33.81088671875, 33.81180859375, 33.81329296875, 33.8145546875, 33.81344140625, 33.814203125, 33.81466796875, 33.81398828125, 33.81423828125, 33.8135234375]",tokens/s,1.8631633373538987,kWh,0.0003992094929516316,0.00021880157859535757,0.0019938841196158203,0.0026118951911628095,tokens/kWh,24120.4165516123,,s,629,343.1500421142575,0.5455485566204417,0.0733377889745091,0.5367131958007813,0.5369653442382812,0.5370488891601563,1.1532696044921875,"[0.5365380859375, 0.5367070922851562, 0.5365258178710938, 0.5365678100585938, 0.5363978271484375, 0.5368893432617188, 0.5364459228515625, 0.5367337036132812, 0.53631591796875, 0.5366466674804687, 0.5364623413085937, 0.5367838745117187, 0.5364408569335938, 0.5364674682617188, 0.536364013671875, 0.5366456298828125, 0.5363200073242187, 0.5367244873046875, 0.5366456298828125, 0.5368350830078125, 0.5363937377929687, 0.53671728515625, 0.5367060546875, 0.5366753540039062, 0.536754150390625, 0.5368975219726563, 0.5367050170898438, 0.5365545043945312, 0.5365575561523438, 0.5367234497070312, 0.5365892944335937, 0.5366651000976562, 0.536605712890625, 0.536995849609375, 0.5365831909179688, 0.5367439575195313, 0.5365278930664062, 0.5368299560546875, 0.5366886596679687, 0.536685546875, 0.536537109375, 0.5367859497070312, 0.5366128540039062, 0.5368207397460938, 0.5366558837890625, 0.5369497680664063, 0.5365565185546874, 0.5367982177734375, 0.5366558837890625, 0.5367808227539063, 0.5366763305664063, 0.5368923950195312, 0.5366241455078125, 0.536774658203125, 0.53675927734375, 0.5367725830078125, 0.5367039794921875, 0.536869873046875, 0.5366456298828125, 0.537069580078125, 0.5366548461914062, 0.5369712524414062, 1.158197265625, 0.5365196533203125, 0.5368145751953125, 0.536332275390625, 0.536406005859375, 0.5366067504882812, 0.5366845703125, 0.5364172973632813, 0.5368176879882812, 0.5368688354492187, 0.536642578125, 0.536406005859375, 0.5366333618164062, 0.5364541625976562, 0.5366098022460938, 0.5361817626953125, 0.53676953125, 0.5361551513671875, 0.536685546875, 0.5366466674804687, 0.5368422241210937, 0.5364224243164063, 0.5367869262695313, 0.5369600219726562, 0.5367777099609375, 0.5367060546875, 0.5369262084960937, 0.536658935546875, 0.5365780639648438, 0.5367879638671875, 0.5365667724609375, 0.5368043823242188, 0.53661181640625, 0.5367131958007813, 0.5368822021484375, 0.5366190185546875, 0.53648486328125, 0.5366988525390625, 0.5368545532226563, 0.5366681518554688, 0.5364121704101562, 0.5366712036132812, 0.5368248291015625, 0.5367808227539063, 0.5365872802734375, 0.5368678588867187, 0.536943603515625, 0.5366456298828125, 0.5367275390625, 0.5368524780273437, 0.5368688354492187, 0.5366353759765625, 0.536722412109375, 0.5367449340820313, 0.5368883056640625, 0.5366231079101562, 0.5369866333007812, 0.5368473510742188, 0.5367838745117187, 0.5364449462890625, 0.5368104858398437, 0.5368524780273437, 0.5368975219726563, 1.1535626220703126, 0.5364807739257812, 0.5365350341796875, 0.5364326171875, 0.5363937377929687, 0.5365145874023437, 0.5367337036132812, 0.5366046752929687, 0.5366548461914062, 0.5365155639648438, 0.5366333618164062, 0.5364786987304687, 0.5366466674804687, 0.5365094604492188, 0.5365811157226562, 0.5364182739257812, 0.5364715576171875, 0.536364013671875, 0.53667431640625, 0.536796142578125, 0.5367265014648438, 0.5366466674804687, 0.536521728515625, 0.5366763305664063, 0.5366128540039062, 0.5367654418945312, 0.536900634765625, 0.5364971313476562, 0.5365094604492188, 0.5370931396484375, 0.536764404296875, 0.5365831909179688, 0.5366036376953125, 0.5365330200195313, 0.5368893432617188, 0.5366026000976563, 0.5368606567382812, 0.536816650390625, 0.5369476928710938, 0.5367490844726562, 0.536848388671875, 0.5366753540039062, 0.53680126953125, 0.5367080688476562, 0.5367828369140625, 0.5370787963867187, 0.5374157104492188, 0.536700927734375, 0.5368831787109375, 0.5369343872070312, 0.5367613525390625, 0.536642578125, 0.537175048828125, 0.5367551879882813, 0.5368934326171875, 0.5366619873046875, 0.5368678588867187, 0.536922119140625, 0.5367337036132812, 0.536595458984375, 0.5369989013671875, 0.5368309936523438, 0.5367244873046875, 1.1537080078125, 0.5367060546875, 0.5366784057617188, 0.5363886108398438, 0.5365841674804688, 0.5365995483398438, 0.536474609375, 0.5364500732421875, 0.5364602661132812, 0.5364244384765625, 0.53661083984375, 0.5364818115234375, 0.5368258666992187, 0.5365626831054687, 0.5365032958984375, 0.5363486938476563, 0.5364080810546875, 0.536258544921875, 0.5372548828125, 0.5367050170898438, 0.5369456787109375, 0.5366630249023437, 0.5366046752929687, 0.5366466674804687, 0.53688525390625, 0.53661083984375, 0.5367572631835937, 0.5366036376953125, 0.5366927490234376, 0.5367080688476562, 0.5365616455078125, 0.5367736206054687, 0.5366415405273437, 0.5366763305664063, 0.5369446411132812, 0.53675927734375, 0.536859619140625, 0.53688525390625, 0.5369733276367188, 0.5365831909179688, 0.5367142333984375, 0.5367131958007813, 0.536679443359375, 0.5372713012695313, 0.5369129028320313, 0.5368627319335938, 0.5370460205078125, 0.53677978515625, 0.5368248291015625, 0.5368831787109375, 0.5366865844726563, 0.5365606689453125, 0.5368023071289062, 0.5370009765625, 0.536958984375, 0.536795166015625, 0.5366497192382812, 0.5369476928710938, 0.5372333984375, 0.5365616455078125, 0.5370224609375, 0.5368002319335937, 0.53704296875, 1.153555419921875, 0.536501220703125, 0.5367725830078125, 0.5364859008789062, 0.5365391235351562, 0.53650634765625, 0.536647705078125, 0.5364469604492188, 0.5366046752929687, 0.536479736328125, 0.5366087646484375, 0.5368074340820312, 0.5366292724609375, 0.5366722412109375, 0.53648486328125, 0.5362606201171874, 0.5367900390625, 0.53659033203125, 0.5366394653320312, 0.536732666015625, 0.5366343383789063, 0.5365166015625, 0.536553466796875, 0.536722412109375, 0.5367357177734375, 0.5368555297851563, 0.536791015625, 0.5365360717773437, 0.5365933837890625, 0.5367091064453124, 0.5368463134765625, 0.53677978515625, 0.5366712036132812, 0.5366599731445313, 0.5368442993164062, 0.5365278930664062, 0.536806396484375, 0.53661181640625, 0.5368319702148437, 0.5365330200195313, 0.5365309448242187, 0.5367255249023437, 0.5366333618164062, 0.5368831787109375, 0.5366343383789063, 0.5369200439453125, 0.5369671630859375, 0.5370470581054687, 0.5367920532226562, 0.5366865844726563, 0.536700927734375, 0.5368156127929687, 0.536958984375, 0.5368831787109375, 0.5370194091796875, 0.5367992553710937, 0.5370132446289062, 0.5367459716796875, 0.536838134765625, 0.536479736328125, 0.5374484252929688, 0.53677978515625, 0.5368411865234375, 1.153471435546875, 0.5364951171875, 0.5366415405273437, 0.5368473510742188, 0.5363834838867187, 0.5366210327148437, 0.536531982421875, 0.5365493774414063, 0.536679443359375, 0.536647705078125, 0.5366435546875, 0.5364019165039062, 0.5366824951171875, 0.5364306030273438, 0.5364357299804687, 0.536369140625, 0.536427490234375, 0.5365278930664062, 0.5366179809570313, 0.5367347412109374, 0.5368822021484375, 0.5366978759765625, 0.5366753540039062, 0.536853515625, 0.53673779296875, 0.536754150390625, 0.5367623901367188, 0.5366763305664063, 0.5367183227539063, 0.5367817993164062, 0.5367337036132812, 0.5367459716796875, 0.5368053588867188, 0.5364777221679687, 0.5369343872070312, 0.5366507568359375, 0.5367244873046875, 0.536754150390625, 0.536985595703125, 0.5369190673828125, 0.5368668212890625, 0.5366292724609375, 0.536880126953125, 0.5368176879882812, 0.5366988525390625, 0.5369866333007812, 0.5370194091796875, 0.5367459716796875, 0.5368370971679688, 0.5365135498046875, 0.53667431640625, 0.53671630859375, 0.5368494262695312, 0.53669580078125, 0.53698046875, 0.5367992553710937, 0.536774658203125, 0.5369190673828125, 0.5369292602539063, 0.5370726318359375, 0.5369497680664063, 0.5366865844726563, 0.5370203857421875, 1.154618408203125, 0.5365596313476563, 0.5364193115234375, 0.5363773193359375, 0.5362728881835938, 0.5365473022460937, 0.5363855590820312, 0.5364039916992187, 0.5364664306640625, 0.5363722534179688, 0.536680419921875, 0.5364766845703125, 0.5371268920898438, 0.5366917114257812, 0.5365166015625, 0.5365975341796875, 0.5365575561523438, 0.5364418334960938, 0.5368084716796875, 0.5367070922851562, 0.5367449340820313, 0.5367603149414063, 0.5367551879882813, 0.536754150390625, 0.5367510986328125, 0.5367684936523438, 0.5367675170898437, 0.5366599731445313, 0.5364961547851562, 0.5366578979492187, 0.536806396484375, 0.5367562255859375, 0.5367408447265625, 0.536627197265625, 0.5369661254882813, 0.5367470092773438, 0.5369343872070312, 0.5368074340820312, 0.5368576049804688, 0.536795166015625, 0.5367285766601563, 0.5365309448242187, 0.5366722412109375, 0.5369343872070312, 0.5367562255859375, 0.5368893432617188, 0.5369651489257813, 0.5368514404296875, 0.5369467163085937, 0.5367101440429688, 0.5367296142578125, 0.5370715942382812, 0.53705419921875, 0.5369313354492188, 0.5368729858398438, 0.5368361206054687, 0.5369497680664063, 0.5369282836914062, 0.5368883056640625, 0.5366251220703125, 0.5370183715820313, 0.5368709106445313, 0.5369528198242187, 1.1556219482421874, 0.5364859008789062, 0.5367736206054687, 0.5364080810546875, 0.536553466796875, 0.5366824951171875, 0.5366098022460938, 0.5365892944335937, 0.53646337890625, 0.536279052734375, 0.5366405029296875, 0.5365493774414063, 0.53646337890625, 0.5365104370117187, 0.5364838256835938, 0.5365084228515625, 0.5366610107421875, 0.5364357299804687, 0.536647705078125, 0.5364889526367187, 0.536748046875, 0.5367930908203125, 0.5366671142578125, 0.5367449340820313, 0.5367019653320313, 0.5370091552734375, 0.536827880859375, 0.536573974609375, 0.5367787475585938, 0.5366098022460938, 0.536838134765625, 0.5366077270507813, 0.5368043823242188, 0.5365616455078125, 0.5371207885742187, 0.5365278930664062, 0.5368494262695312, 0.5366507568359375, 0.5369343872070312, 0.5367572631835937, 0.5367551879882813, 0.5369886474609376, 0.5369292602539063, 0.5371954956054688, 0.536795166015625, 0.5366947631835938, 0.536932373046875, 0.5366435546875, 0.5369630737304687, 0.5365084228515625, 0.537143310546875, 0.5367654418945312, 0.536848388671875, 0.5367890014648438, 0.5368955078125, 0.5366302490234375, 0.5367152709960937, 0.5368729858398438, 0.5369559326171875, 0.5366732788085937, 0.5370153198242188, 0.536826904296875, 0.536880126953125, 1.1527025146484375, 0.536489990234375, 0.5366835327148437, 0.5366138916015625, 0.5365176391601563, 0.5364889526367187, 0.536453125, 0.5365125122070312, 0.5366312866210937, 0.5364080810546875, 0.5364920043945313, 0.5364111328125, 0.5365258178710938, 0.5365238037109376, 0.5367664794921875, 0.5364930419921875, 0.5364602661132812, 0.5364971313476562, 0.536774658203125, 0.5366835327148437, 0.5369774169921875, 0.5367152709960937, 0.53650634765625, 0.53692724609375, 0.5366179809570313, 0.5366988525390625, 0.5368176879882812, 0.5365545043945312, 0.5367337036132812, 0.536700927734375, 0.5366466674804687, 0.536774658203125, 0.5367050170898438, 0.53663232421875, 0.5369866333007812, 0.5367091064453124, 0.5367388305664063, 0.5366220703125, 0.5368125610351563, 0.5369733276367188, 0.5369313354492188, 0.53682177734375, 0.53686474609375, 0.5370634155273437, 0.5370572509765625, 0.536827880859375, 0.536806396484375, 0.536573974609375, 0.5369467163085937, 0.5368350830078125, 0.5367930908203125, 0.536605712890625, 0.5369088134765625, 0.536859619140625, 0.5371924438476563, 0.5367551879882813, 0.5368657836914063, 0.5368893432617188, 0.5367039794921875, 0.5366220703125, 0.5370398559570313, 0.5369682006835937, 0.5367817993164062, 1.1527506103515626, 0.5365135498046875, 0.5366722412109375, 0.536648681640625, 0.5365913696289063, 0.536416259765625, 0.5366036376953125, 0.5365339965820313, 0.5371463623046875, 0.53639990234375, 0.536574951171875, 0.5365350341796875, 0.5370501098632813, 0.5364541625976562, 0.5363189697265625, 0.5364039916992187, 0.5364520874023437, 0.5365821533203124, 0.5366282348632813, 0.5366784057617188, 0.5367418823242187, 0.5366732788085937, 0.536711181640625, 0.5367203979492188, 0.5366394653320312, 0.5366917114257812, 0.5369784545898437, 0.53665380859375, 0.5367470092773438, 0.5368893432617188, 0.536816650390625, 0.5366619873046875, 0.5365606689453125, 0.5364520874023437, 0.536932373046875, 0.5365616455078125, 0.5367787475585938, 0.5365401611328126, 0.5368002319335937, 0.5364520874023437, 0.5367603149414063, 0.5366005859375, 0.536791015625, 0.5371688842773438, 0.53677978515625, 0.5367255249023437, 0.5369937744140625, 0.536795166015625, 0.5367992553710937, 0.5370582885742188, 0.5368258666992187, 0.536658935546875, 0.5370296020507812, 0.5368555297851563, 0.5368770751953125, 0.5367736206054687, 0.5367725830078125, 0.5368186645507812, 0.536853515625, 0.5365596313476563, 0.5368340454101562, 0.5368145751953125, 0.5369313354492188]",tokens/s,1.8330174058103814,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13383,7 +13383,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13425,7 +13425,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 96.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13496,7 +13496,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13568,7 +13568,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13626,7 +13626,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13683,7 +13683,7 @@ ChildProcessError: Traceback (most recent call last): RuntimeError: The size of tensor a (60) must match the size of tensor b (32) at non-singleton dimension 2 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13712,7 +13712,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13774,7 +13774,7 @@ ValueError: OPTForCausalLM does not support an attention implementation through ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5086.445568,5956.435968,0.0,5309.988864,5047.4112,s,10,3.3951274108886715,0.33951274108886714,0.00038503905864011734,0.33945068359375,0.33990344848632814,0.34012059326171873,0.34029430908203123,"[0.33982321166992185, 0.3391507568359375, 0.33938748168945315, 0.33922314453125, 0.34033773803710937, 0.3395138854980469, 0.33896197509765624, 0.3398551940917969, 0.3395834045410156, 0.3392906188964844]",tokens/s,754.0217759692035,kWh,4.00638022908458e-06,2.1953130347537806e-06,1.9920034454532867e-05,2.6121727718371226e-05,tokens/kWh,9800270.593126083,MB,5086.445568,5956.435968,0.0,5309.988864,5098.333184,s,10,188.221390625,18.822139062500003,0.0011505271698466547,18.8222666015625,18.8231517578125,18.82387861328125,18.82446009765625,"[18.821546875, 18.82124609375, 18.822708984375, 18.82460546875, 18.822560546875, 18.82055859375, 18.8222265625, 18.822306640625, 18.820640625, 18.822990234375]",tokens/s,3.3471222261616944,kWh,0.00022222505159952025,0.00012179794381854662,0.0010774225934188715,0.0014214455888369383,tokens/kWh,44321.077426219425,,s,629,190.97790853881838,0.3036214762143376,0.0403576770019219,0.29874688720703124,0.2989168701171875,0.2990350280761719,0.6382961645507812,"[0.2991718444824219, 0.29857177734375, 0.2985912170410156, 0.2988011474609375, 0.29874789428710935, 0.2987694091796875, 0.29878680419921877, 0.29858917236328125, 0.29871923828125, 0.29875506591796874, 0.29862911987304686, 0.29869976806640625, 0.29868133544921877, 0.29870693969726564, 0.2988308410644531, 0.2986854248046875, 0.2986752014160156, 0.2989803466796875, 0.29859634399414064, 0.2987222900390625, 0.2987008056640625, 0.2986854248046875, 0.2986875, 0.2986332092285156, 0.2987663269042969, 0.29869259643554685, 0.2985953369140625, 0.29869976806640625, 0.298566650390625, 0.29867929077148436, 0.29865780639648437, 0.29856668090820315, 0.29851031494140623, 0.2987509765625, 0.2991912841796875, 0.2989486083984375, 0.29886465454101563, 0.2987008056640625, 0.29902334594726565, 0.2987386779785156, 0.29886257934570315, 0.29884622192382815, 0.29862298583984376, 0.2987376708984375, 0.29878475952148437, 0.2987100219726562, 0.2987274169921875, 0.29880218505859374, 0.2987960205078125, 0.29913088989257813, 0.2988216247558594, 0.29875405883789063, 0.2986885070800781, 0.29871206665039063, 0.2987673645019531, 0.29861785888671877, 0.29866598510742187, 0.2988287963867188, 0.2988287963867188, 0.29879705810546875, 0.2987857971191406, 0.2987950134277344, 0.6382120971679688, 0.29854412841796873, 0.2985953369140625, 0.2985646057128906, 0.29854104614257815, 0.298724365234375, 0.29859634399414064, 0.29881240844726564, 0.29854617309570314, 0.29873358154296875, 0.2988380126953125, 0.29890969848632815, 0.2988410949707031, 0.29858612060546874, 0.2986250305175781, 0.2987253723144531, 0.29870693969726564, 0.29874893188476564, 0.2987386779785156, 0.2988083190917969, 0.29867315673828126, 0.2986280822753906, 0.29873358154296875, 0.2987376708984375, 0.2986905517578125, 0.298777587890625, 0.29880523681640625, 0.29867828369140625, 0.2988226623535156, 0.2988072814941406, 0.2987427978515625, 0.29883905029296876, 0.29878680419921877, 0.2985922546386719, 0.29876837158203123, 0.29872332763671877, 0.29868646240234376, 0.2986690673828125, 0.2988482666015625, 0.29870181274414065, 0.29881857299804687, 0.29884518432617185, 0.2989219970703125, 0.2987427978515625, 0.2989219970703125, 0.2988748779296875, 0.2987110290527344, 0.29886566162109374, 0.2988779602050781, 0.29880523681640625, 0.29869464111328126, 0.298745849609375, 0.2987427978515625, 0.29859942626953123, 0.2986700744628906, 0.29881857299804687, 0.2987734985351562, 0.29879806518554686, 0.298703857421875, 0.2987110290527344, 0.298756103515625, 0.2991769714355469, 0.2987376708984375, 0.6383523559570312, 0.2986219482421875, 0.29856051635742187, 0.29854617309570314, 0.2987694091796875, 0.29871923828125, 0.298534912109375, 0.2986680297851563, 0.29889434814453125, 0.2986250305175781, 0.29877044677734377, 0.29878375244140626, 0.2986680297851563, 0.2991124572753906, 0.2987796630859375, 0.2987530212402344, 0.2987222900390625, 0.2987888488769531, 0.2988287963867188, 0.2987950134277344, 0.2988482666015625, 0.29875405883789063, 0.29862603759765627, 0.2985953369140625, 0.2988308410644531, 0.2987632751464844, 0.2986342468261719, 0.299082763671875, 0.2986588134765625, 0.2988001403808594, 0.29873358154296875, 0.29896295166015624, 0.29876123046875, 0.2986895446777344, 0.298682373046875, 0.2987694091796875, 0.29866494750976563, 0.298819580078125, 0.2986956787109375, 0.2985513000488281, 0.29881240844726564, 0.2988011474609375, 0.298814453125, 0.29863833618164065, 0.29881753540039063, 0.2988830871582031, 0.29885440063476565, 0.29881036376953124, 0.29866290283203123, 0.298682373046875, 0.29867315673828126, 0.29896395874023435, 0.29883700561523435, 0.2987694091796875, 0.2988328857421875, 0.29885748291015624, 0.2988564453125, 0.29895269775390626, 0.2987806701660156, 0.29891278076171873, 0.2988830871582031, 0.29890765380859374, 0.2989168701171875, 0.6389586181640625, 0.29872128295898437, 0.2986536865234375, 0.29879910278320315, 0.2986854248046875, 0.29872128295898437, 0.29872845458984376, 0.2987356262207031, 0.29869464111328126, 0.29878375244140626, 0.2991523742675781, 0.2986168212890625, 0.29872640991210936, 0.2986588134765625, 0.29912472534179685, 0.29872845458984376, 0.2986956787109375, 0.29874688720703124, 0.2986332092285156, 0.29874789428710935, 0.29879806518554686, 0.29864346313476564, 0.29869876098632814, 0.29862399291992187, 0.29865676879882813, 0.298787841796875, 0.2986106872558594, 0.29867724609375, 0.2988134460449219, 0.298756103515625, 0.29874072265625, 0.2988482666015625, 0.2986065979003906, 0.29880422973632814, 0.298756103515625, 0.2998691711425781, 0.2986905517578125, 0.298714111328125, 0.2987028503417969, 0.2987509765625, 0.2988083190917969, 0.2988206176757813, 0.2987182006835937, 0.298703857421875, 0.29875201416015623, 0.29893121337890627, 0.2988267517089844, 0.2989168701171875, 0.2986741638183594, 0.2989014892578125, 0.29890866088867185, 0.2989270935058594, 0.29876837158203123, 0.29879910278320315, 0.29877862548828127, 0.29884622192382815, 0.29892095947265623, 0.2989322204589844, 0.298756103515625, 0.2989404296875, 0.29872845458984376, 0.2988206176757813, 0.29878375244140626, 0.63944189453125, 0.29866494750976563, 0.29873458862304686, 0.29941351318359377, 0.2987222900390625, 0.2986700744628906, 0.298650634765625, 0.29863015747070315, 0.29849395751953123, 0.2986536865234375, 0.29863729858398436, 0.29870489501953124, 0.2986311950683594, 0.2986546936035156, 0.298714111328125, 0.298745849609375, 0.29884210205078127, 0.29883187866210936, 0.2987581481933594, 0.2987724914550781, 0.2988584899902344, 0.298640380859375, 0.2987530212402344, 0.29866598510742187, 0.2987734985351562, 0.2988441467285156, 0.29874993896484375, 0.2986393737792969, 0.2987663269042969, 0.2986833801269531, 0.29886053466796875, 0.29871206665039063, 0.29884622192382815, 0.2987632751464844, 0.2985902099609375, 0.29863116455078126, 0.29862911987304686, 0.29875506591796874, 0.2986752014160156, 0.2986854248046875, 0.29865472412109373, 0.298703857421875, 0.2987304992675781, 0.29888204956054687, 0.2987386779785156, 0.2986557312011719, 0.2988277893066406, 0.29886465454101563, 0.2989363098144531, 0.298903564453125, 0.2989363098144531, 0.2989404296875, 0.29923941040039065, 0.29895782470703125, 0.298714111328125, 0.2987028503417969, 0.2989168701171875, 0.2988216247558594, 0.29880935668945313, 0.29886465454101563, 0.29877658081054687, 0.2988164978027344, 0.29866598510742187, 0.6385172729492188, 0.2987161560058594, 0.29870181274414065, 0.29873458862304686, 0.2985277404785156, 0.29870181274414065, 0.29862399291992187, 0.2987222900390625, 0.29852978515625, 0.2984530029296875, 0.29863833618164065, 0.2985779113769531, 0.2986393737792969, 0.2987673645019531, 0.2987304992675781, 0.2987028503417969, 0.2987581481933594, 0.298576904296875, 0.2986188659667969, 0.29863116455078126, 0.29871307373046874, 0.2985830383300781, 0.29869259643554685, 0.29868441772460935, 0.29859634399414064, 0.29861785888671877, 0.2987222900390625, 0.2986536865234375, 0.2988011474609375, 0.2985912170410156, 0.2987427978515625, 0.29869259643554685, 0.2987960205078125, 0.2987673645019531, 0.2987694091796875, 0.29872946166992187, 0.29888204956054687, 0.29878475952148437, 0.2987581481933594, 0.29880422973632814, 0.2987530212402344, 0.29877044677734377, 0.298819580078125, 0.29879910278320315, 0.29877862548828127, 0.29886465454101563, 0.29878271484375, 0.29871923828125, 0.298576904296875, 0.2992762756347656, 0.29873663330078126, 0.2988359680175781, 0.29874688720703124, 0.29913394165039064, 0.2987376708984375, 0.29887387084960937, 0.29878988647460936, 0.29888101196289063, 0.29867724609375, 0.29872128295898437, 0.2988380126953125, 0.2989219970703125, 0.29882366943359373, 0.6379100341796875, 0.29849395751953123, 0.2987642822265625, 0.2986956787109375, 0.2986536865234375, 0.2986342468261719, 0.29868646240234376, 0.29869259643554685, 0.2986905517578125, 0.298967041015625, 0.2986536865234375, 0.2986956787109375, 0.29867111206054686, 0.2986670227050781, 0.29867724609375, 0.29880218505859374, 0.2986833801269531, 0.2987028503417969, 0.29863116455078126, 0.29869363403320315, 0.2986977233886719, 0.29881753540039063, 0.2988114013671875, 0.2987325439453125, 0.2990878601074219, 0.2988226623535156, 0.2988031921386719, 0.29867623901367185, 0.2988410949707031, 0.29869158935546875, 0.2987182006835937, 0.29871514892578127, 0.29878988647460936, 0.2987304992675781, 0.29869363403320315, 0.29874789428710935, 0.298756103515625, 0.2987591552734375, 0.2987161560058594, 0.29865676879882813, 0.29869876098632814, 0.29875711059570315, 0.2986875, 0.29886566162109374, 0.29875711059570315, 0.2986752014160156, 0.2987008056640625, 0.29882571411132813, 0.2986741638183594, 0.29876019287109373, 0.29889434814453125, 0.2987161560058594, 0.29932748413085936, 0.29880422973632814, 0.2988687438964844, 0.29856051635742187, 0.29914111328125, 0.29886566162109374, 0.29886566162109374, 0.298956787109375, 0.29882366943359373, 0.29876123046875, 0.29887387084960937, 0.6389237670898438, 0.2986998291015625, 0.29864031982421874, 0.2987253723144531, 0.2986250305175781, 0.2985779113769531, 0.2987386779785156, 0.2985850830078125, 0.29863629150390625, 0.29863729858398436, 0.29869158935546875, 0.29864141845703124, 0.2990602111816406, 0.2987724914550781, 0.2986608581542969, 0.2989906005859375, 0.2988503112792969, 0.29879705810546875, 0.29870489501953124, 0.298724365234375, 0.29857073974609377, 0.29866494750976563, 0.29862405395507813, 0.29871405029296877, 0.2985922546386719, 0.29864346313476564, 0.29879296875, 0.299109375, 0.2994646911621094, 0.2987386779785156, 0.29884722900390626, 0.29866598510742187, 0.29875711059570315, 0.2986332092285156, 0.29882571411132813, 0.2988114013671875, 0.29872332763671877, 0.2986680297851563, 0.2987653198242187, 0.29900082397460936, 0.2987509765625, 0.29862399291992187, 0.2986475524902344, 0.29878170776367186, 0.2986905517578125, 0.29885336303710935, 0.298756103515625, 0.29877044677734377, 0.2987724914550781, 0.29862911987304686, 0.2986526794433594, 0.298903564453125, 0.298745849609375, 0.29878988647460936, 0.2988226623535156, 0.2988164978027344, 0.2988072814941406, 0.29877862548828127, 0.29900390625, 0.29874072265625, 0.29883392333984377, 0.2988072814941406, 0.298956787109375, 0.6386237182617187, 0.29877658081054687, 0.2986854248046875, 0.29864651489257815, 0.2986045532226563, 0.2988031921386719, 0.2985267333984375, 0.2986618957519531, 0.29867315673828126, 0.29864651489257815, 0.29855230712890624, 0.29865472412109373, 0.2986639404296875, 0.2986680297851563, 0.2987386779785156, 0.2985472106933594, 0.2987100219726562, 0.2987182006835937, 0.2986219482421875, 0.29865164184570314, 0.29867724609375, 0.2986342468261719, 0.2985973815917969, 0.2986598510742188, 0.29906228637695315, 0.2986956787109375, 0.29896600341796875, 0.29861785888671877, 0.2990868530273437, 0.298724365234375, 0.2988769226074219, 0.2987591552734375, 0.29865164184570314, 0.298787841796875, 0.29884005737304686, 0.298682373046875, 0.29856768798828126, 0.2986956787109375, 0.29869158935546875, 0.2987182006835937, 0.2987448425292969, 0.2987745361328125, 0.2987950134277344, 0.29871923828125, 0.29873663330078126, 0.29879193115234376, 0.29870489501953124, 0.2987960205078125, 0.2988062744140625, 0.29881549072265623, 0.29877658081054687, 0.2988441467285156, 0.2987806701660156, 0.2988707885742187, 0.29862399291992187, 0.2986117248535156, 0.2987694091796875, 0.29873663330078126, 0.29883905029296876, 0.29911654663085935, 0.29874176025390625, 0.29878375244140626, 0.29891787719726565, 0.638328857421875, 0.2986250305175781, 0.2985748596191406, 0.29873663330078126, 0.29864346313476564, 0.29865472412109373, 0.29872845458984376, 0.2986977233886719, 0.29866598510742187, 0.29880935668945313, 0.29860147094726563, 0.29870693969726564, 0.29879193115234376, 0.2987386779785156, 0.29876019287109373, 0.2987950134277344, 0.29864141845703124, 0.29870693969726564, 0.29863629150390625, 0.29877044677734377, 0.29883905029296876, 0.29858917236328125, 0.2987734985351562, 0.29867623901367185, 0.29884518432617185, 0.2988001403808594, 0.298745849609375, 0.29894451904296876, 0.29879910278320315, 0.298703857421875, 0.29870489501953124, 0.29902334594726565, 0.2988717956542969, 0.298745849609375, 0.29904281616210937, 0.29879193115234376, 0.2988031921386719, 0.2987653198242187, 0.2987325439453125, 0.2988206176757813, 0.29881036376953124, 0.29879092407226565, 0.298650634765625, 0.29869158935546875, 0.29877862548828127, 0.29889739990234376, 0.29879705810546875, 0.29879193115234376, 0.29879296875, 0.29889022827148437, 0.29869876098632814, 0.2989056091308594, 0.2989189147949219, 0.2988441467285156, 0.29881857299804687, 0.29889022827148437, 0.29874789428710935, 0.29886257934570315, 0.2987694091796875, 0.298851318359375, 0.2988492736816406, 0.29875711059570315, 0.2988687438964844]",tokens/s,3.293574659040466,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13949,7 +13949,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,4110.757888,5287.44448,0.0,4640.997376,4330.668544,s,10,3.1969297485351564,0.31969297485351567,0.001348702205900169,0.31914218139648437,0.32145095520019534,0.32196441497802736,0.322375182800293,"[0.3186290283203125, 0.3224778747558594, 0.31885498046875, 0.31913238525390625, 0.3191519775390625, 0.31846505737304687, 0.3200827941894531, 0.3207144470214844, 0.32133685302734377, 0.3180843505859375]",tokens/s,800.7683000144123,kWh,3.758179541263316e-06,2.0593209790774835e-06,1.8989615886123856e-05,2.4807116406464656e-05,tokens/kWh,10319619.411036715,MB,4110.757888,5287.44448,0.0,4640.997376,4515.667456,s,10,175.15052539062503,17.515052539062502,0.0018447663069760805,17.515354492187498,17.517081640624998,17.517545703125,17.517916953125,"[17.516314453125, 17.513849609375, 17.516134765625, 17.511609375, 17.515056640625, 17.513658203125, 17.516978515625, 17.51565234375, 17.518009765625, 17.51326171875]",tokens/s,3.596906138848048,kWh,0.00020681544852339562,0.0001133520352390417,0.0010167293237994686,0.001336896807561906,tokens/kWh,47124.055980725156,,s,629,177.750416656494,0.2825920773553166,0.03806230539059325,0.278002685546875,0.2781634521484375,0.27826237182617186,0.5973247680664062,"[0.27860992431640624, 0.27796890258789064, 0.2778388366699219, 0.2779924621582031, 0.278066162109375, 0.27786444091796875, 0.2777835388183594, 0.27789825439453125, 0.2779504699707031, 0.27782144165039063, 0.2779309997558594, 0.27789825439453125, 0.27791461181640625, 0.27783782958984377, 0.27793408203125, 0.27778662109375, 0.2780508117675781, 0.27777227783203123, 0.2779862976074219, 0.27807537841796875, 0.2779053955078125, 0.27794329833984377, 0.2780487670898438, 0.2779258728027344, 0.2780487670898438, 0.27790643310546875, 0.27796786499023435, 0.2779258728027344, 0.2781552734375, 0.2780569458007813, 0.2780641174316406, 0.27803851318359374, 0.2781112365722656, 0.2779862976074219, 0.27808154296875, 0.27809381103515624, 0.27819314575195314, 0.2780559387207031, 0.2779658203125, 0.27812966918945314, 0.2781440124511719, 0.27802215576171874, 0.2781583251953125, 0.2780712890625, 0.2780262451171875, 0.27805184936523436, 0.2781255798339844, 0.27800167846679685, 0.278181884765625, 0.2779955139160156, 0.2780979309082031, 0.2779791259765625, 0.2781634521484375, 0.27801702880859375, 0.27805902099609375, 0.27803033447265624, 0.2780702819824219, 0.2781051025390625, 0.2779955139160156, 0.27840716552734374, 0.27825457763671874, 0.27805902099609375, 0.6008422241210938, 0.278034423828125, 0.27797402954101563, 0.2779924621582031, 0.27792486572265623, 0.27854949951171876, 0.27780300903320315, 0.27780300903320315, 0.27782962036132813, 0.27787673950195313, 0.27778866577148437, 0.2779053955078125, 0.27800677490234377, 0.27789926147460936, 0.2778347473144531, 0.2777978820800781, 0.2779156494140625, 0.27812353515625, 0.27781939697265623, 0.27812249755859375, 0.2779033508300781, 0.27796173095703125, 0.2778306579589844, 0.2780006408691406, 0.2777835388183594, 0.2780692443847656, 0.277818359375, 0.2778603515625, 0.27797503662109374, 0.27809588623046877, 0.27791360473632815, 0.27798016357421873, 0.27808050537109374, 0.2780610656738281, 0.27797708129882814, 0.2779371643066406, 0.27804672241210937, 0.27804672241210937, 0.2779084777832031, 0.2779504699707031, 0.2779473876953125, 0.27842559814453127, 0.27789825439453125, 0.27808050537109374, 0.27795660400390626, 0.2780354614257812, 0.27798016357421873, 0.2781337585449219, 0.278034423828125, 0.2782637939453125, 0.27796685791015624, 0.27807745361328123, 0.2779402160644531, 0.27826995849609376, 0.27799859619140627, 0.27812863159179685, 0.2779176940917969, 0.2780999755859375, 0.2779873352050781, 0.2780979309082031, 0.27799755859375, 0.27807949829101564, 0.2780159912109375, 0.5972192993164063, 0.2780426330566406, 0.27779583740234376, 0.2778480529785156, 0.2779473876953125, 0.2780057678222656, 0.2779320373535156, 0.2778316650390625, 0.2777558898925781, 0.27795559692382815, 0.2778880004882813, 0.27797607421875, 0.2779033508300781, 0.2783354797363281, 0.2778040466308594, 0.27798220825195313, 0.2778470458984375, 0.27807537841796875, 0.2782801818847656, 0.2779371643066406, 0.2779781188964844, 0.2780334167480469, 0.2780231628417969, 0.2780211181640625, 0.27808154296875, 0.2785587158203125, 0.2779914245605469, 0.27797607421875, 0.2779873352050781, 0.27805902099609375, 0.2780835876464844, 0.2781051025390625, 0.27797708129882814, 0.2781767578125, 0.27785830688476565, 0.277865478515625, 0.27793304443359373, 0.2779924621582031, 0.27794638061523436, 0.2780702819824219, 0.2779484558105469, 0.27794427490234375, 0.2778828735351562, 0.27815115356445314, 0.27796173095703125, 0.2780702819824219, 0.2779371643066406, 0.2781368408203125, 0.2780641174316406, 0.2786897888183594, 0.277907470703125, 0.27796685791015624, 0.2779893798828125, 0.2781440124511719, 0.2782853088378906, 0.27813479614257813, 0.2779914245605469, 0.27810302734375, 0.27808255004882815, 0.2780641174316406, 0.2779453430175781, 0.2784460754394531, 0.27808563232421873, 0.5973657836914062, 0.27796890258789064, 0.27786138916015624, 0.27774566650390625, 0.27787982177734377, 0.2779381713867187, 0.2778122253417969, 0.2779596862792969, 0.2777999267578125, 0.27791668701171873, 0.2776084594726563, 0.2777763977050781, 0.27786444091796875, 0.277960693359375, 0.2777794494628906, 0.2778050537109375, 0.27780914306640625, 0.27791873168945314, 0.2777989196777344, 0.278118408203125, 0.2779361267089844, 0.27799655151367186, 0.2777415771484375, 0.27795559692382815, 0.277928955078125, 0.27811328125, 0.2781317138671875, 0.27800473022460936, 0.278029296875, 0.2780559387207031, 0.27793408203125, 0.2779279479980469, 0.2780426330566406, 0.2780579833984375, 0.2779381713867187, 0.27791668701171873, 0.27802215576171874, 0.2780968933105469, 0.2780088195800781, 0.2779258728027344, 0.2781665344238281, 0.2780159912109375, 0.2779873352050781, 0.2779945068359375, 0.2779811706542969, 0.27802215576171874, 0.27794125366210937, 0.27793304443359373, 0.27801190185546876, 0.27817166137695315, 0.2779422607421875, 0.2779924621582031, 0.2779320373535156, 0.27813784790039064, 0.2779095153808594, 0.2779996032714844, 0.27801089477539065, 0.2781603698730469, 0.27780609130859374, 0.278012939453125, 0.27797402954101563, 0.2780262451171875, 0.27819622802734373, 0.5975726318359375, 0.2780139465332031, 0.2778470458984375, 0.27806719970703125, 0.27787161254882814, 0.27796890258789064, 0.2779207763671875, 0.2778746948242187, 0.27775079345703124, 0.27796273803710936, 0.27774053955078126, 0.27780096435546875, 0.27782861328125, 0.27800372314453126, 0.2778880004882813, 0.2778480529785156, 0.27787982177734377, 0.27799038696289063, 0.2778757019042969, 0.278012939453125, 0.27793304443359373, 0.2780631103515625, 0.27790234375, 0.2780334167480469, 0.27802520751953125, 0.2781419372558594, 0.27788595581054687, 0.27802725219726565, 0.27800677490234377, 0.2781614074707031, 0.2779781188964844, 0.2780211181640625, 0.27803237915039064, 0.2782033996582031, 0.27809381103515624, 0.2779648132324219, 0.27801702880859375, 0.27804568481445313, 0.27800372314453126, 0.2782637939453125, 0.27806002807617186, 0.27815936279296877, 0.27799655151367186, 0.2780784606933594, 0.2779934692382813, 0.2781634521484375, 0.27805389404296876, 0.2781481018066406, 0.2780784606933594, 0.27816754150390627, 0.278044677734375, 0.27812762451171874, 0.27803033447265624, 0.27824127197265625, 0.27803851318359374, 0.2779924621582031, 0.2779473876953125, 0.2782156677246094, 0.27799859619140627, 0.277923828125, 0.2781163635253906, 0.27806207275390626, 0.2780477294921875, 0.596780029296875, 0.27795660400390626, 0.27787673950195313, 0.27799655151367186, 0.27799655151367186, 0.2779637756347656, 0.27791256713867185, 0.2778070983886719, 0.2779002990722656, 0.2779228210449219, 0.27778561401367186, 0.2780426330566406, 0.27786956787109374, 0.2779729919433594, 0.2781552734375, 0.277907470703125, 0.27780096435546875, 0.2779596862792969, 0.2777630615234375, 0.27805184936523436, 0.27790643310546875, 0.2779781188964844, 0.27776614379882814, 0.2779596862792969, 0.27791256713867185, 0.2780692443847656, 0.2778828735351562, 0.2778746948242187, 0.27790234375, 0.27818801879882815, 0.2778746948242187, 0.2779228210449219, 0.2780006408691406, 0.27844403076171875, 0.27789108276367186, 0.27794638061523436, 0.2780211181640625, 0.27810714721679686, 0.27806515502929685, 0.2781481018066406, 0.2779986267089844, 0.27808969116210935, 0.2780999755859375, 0.27813479614257813, 0.2778173522949219, 0.2779811706542969, 0.277970947265625, 0.2781951904296875, 0.2780405883789063, 0.27821978759765625, 0.2779084777832031, 0.2779719543457031, 0.27793408203125, 0.27815936279296877, 0.2778623962402344, 0.27813479614257813, 0.2779402160644531, 0.27821261596679686, 0.27808154296875, 0.2780405883789063, 0.2779207763671875, 0.2780631103515625, 0.2781051025390625, 0.5984942016601562, 0.2780979309082031, 0.277855224609375, 0.27779483032226565, 0.27799038696289063, 0.278066162109375, 0.27786956787109374, 0.2779422607421875, 0.27826174926757813, 0.2779832458496094, 0.277796875, 0.27789413452148437, 0.27803955078125, 0.27804672241210937, 0.2779361267089844, 0.2780231628417969, 0.27792486572265623, 0.27805184936523436, 0.2779658203125, 0.2780712890625, 0.2781317138671875, 0.2780928039550781, 0.2779443054199219, 0.2780149841308594, 0.2779504699707031, 0.2780702819824219, 0.2779525146484375, 0.27796990966796875, 0.27806719970703125, 0.278034423828125, 0.2780405883789063, 0.278002685546875, 0.27802215576171874, 0.27820645141601563, 0.2779361267089844, 0.2779596862792969, 0.2780201110839844, 0.2780692443847656, 0.277960693359375, 0.2780354614257812, 0.2779791259765625, 0.2780498046875, 0.2780415954589844, 0.278066162109375, 0.27822695922851565, 0.2780313720703125, 0.27807232666015624, 0.27818701171875, 0.2780415954589844, 0.2782801818847656, 0.27808767700195314, 0.2781061096191406, 0.2780712890625, 0.278044677734375, 0.27798834228515623, 0.27814093017578123, 0.27845428466796873, 0.27822592163085935, 0.2780159912109375, 0.2780559387207031, 0.2779934692382813, 0.2782310485839844, 0.2781583251953125, 0.5990051879882813, 0.27789312744140626, 0.2778961791992188, 0.2781265869140625, 0.27796786499023435, 0.2780559387207031, 0.27769650268554685, 0.27773849487304686, 0.2780559387207031, 0.27800372314453126, 0.2776801147460938, 0.27787060546875, 0.27782144165039063, 0.27800473022460936, 0.27786444091796875, 0.2779299926757812, 0.2782156677246094, 0.27800167846679685, 0.2778316650390625, 0.27791668701171873, 0.2779422607421875, 0.2783897705078125, 0.2778204040527344, 0.2779105224609375, 0.27800372314453126, 0.2780231628417969, 0.277970947265625, 0.277939208984375, 0.27832012939453127, 0.2780641174316406, 0.27799755859375, 0.27813067626953125, 0.27803851318359374, 0.278150146484375, 0.27794638061523436, 0.2779176940917969, 0.27802215576171874, 0.2780631103515625, 0.27795455932617186, 0.27806207275390626, 0.2780682373046875, 0.27805490112304687, 0.2779576416015625, 0.2780088195800781, 0.277907470703125, 0.278139892578125, 0.27790130615234376, 0.27797607421875, 0.2780682373046875, 0.27826278686523437, 0.277939208984375, 0.27811431884765625, 0.2783999938964844, 0.27817166137695315, 0.2779648132324219, 0.27796685791015624, 0.27807537841796875, 0.27856692504882813, 0.27819827270507813, 0.2779945068359375, 0.278118408203125, 0.2781051025390625, 0.2781685791015625, 0.5997485961914063, 0.27802215576171874, 0.2779996032714844, 0.27782962036132813, 0.27801089477539065, 0.2780426330566406, 0.2778408813476563, 0.2778839111328125, 0.277833740234375, 0.27784600830078127, 0.27785726928710935, 0.27782962036132813, 0.2780631103515625, 0.27799038696289063, 0.2778828735351562, 0.27795147705078127, 0.2778327026367188, 0.27798016357421873, 0.2779279479980469, 0.2780078125, 0.2779719543457031, 0.2780078125, 0.2779309997558594, 0.278034423828125, 0.27794943237304687, 0.278086669921875, 0.277823486328125, 0.2779084777832031, 0.2780201110839844, 0.27817062377929686, 0.27802520751953125, 0.27811328125, 0.2780989379882812, 0.27811532592773436, 0.27788595581054687, 0.27801803588867186, 0.27810202026367187, 0.27812249755859375, 0.27801702880859375, 0.27796173095703125, 0.27799038696289063, 0.27811737060546876, 0.27811019897460937, 0.27808154296875, 0.2779719543457031, 0.2780282897949219, 0.2780211181640625, 0.278066162109375, 0.27813479614257813, 0.2783088684082031, 0.27800677490234377, 0.2780928039550781, 0.2781051025390625, 0.27819314575195314, 0.27791668701171873, 0.278002685546875, 0.27808255004882815, 0.27803955078125, 0.2780610656738281, 0.2780631103515625, 0.2780364685058594, 0.2781470642089844, 0.2780149841308594, 0.5994915771484375, 0.2779402160644531, 0.27785113525390626, 0.2778450012207031, 0.2779842529296875, 0.27790130615234376, 0.27784396362304686, 0.2778245239257813, 0.277923828125, 0.2779576416015625, 0.2778316650390625, 0.27778253173828127, 0.2778204040527344, 0.2779279479980469, 0.27787982177734377, 0.2778787841796875, 0.277796875, 0.277907470703125, 0.27791973876953124, 0.27798220825195313, 0.27790438842773435, 0.27808050537109374, 0.27791156005859374, 0.27804364013671873, 0.27786138916015624, 0.27808050537109374, 0.27792178344726565, 0.2780641174316406, 0.27806719970703125, 0.2780426330566406, 0.2778623962402344, 0.278029296875, 0.278066162109375, 0.2781890563964844, 0.27787982177734377, 0.27800677490234377, 0.2780415954589844, 0.2781695861816406, 0.2779719543457031, 0.2780477294921875, 0.27800985717773435, 0.2780211181640625, 0.2778736572265625, 0.27810714721679686, 0.27792486572265623, 0.27814093017578123, 0.2779535217285156, 0.2780579833984375, 0.2779596862792969, 0.2782637939453125, 0.27796786499023435, 0.27802215576171874, 0.2780498046875, 0.2780979309082031, 0.27786138916015624, 0.2779873352050781, 0.27800473022460936, 0.27812045288085935, 0.2780364685058594, 0.27805490112304687, 0.2780139465332031, 0.2782054443359375, 0.27814910888671873]",tokens/s,3.5386696235742376,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14164,7 +14164,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14363,7 +14363,7 @@ OSError: M is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14783,7 +14783,7 @@ Traceback (most recent call last): OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14912,7 +14912,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,6436.08576,20389.036032,0.0,19742.588928,18821.508608,s,10,15.157101196289062,1.5157101196289062,0.003170114166599588,1.5143783569335938,1.5197610107421875,1.5208548095703125,1.5217298486328126,"[1.5126845703125, 1.512470458984375, 1.5143804931640625, 1.5141666259765625, 1.514376220703125, 1.512878173828125, 1.515317626953125, 1.5193604736328126, 1.5195179443359375, 1.5219486083984375]",tokens/s,168.89773096103423,kWh,1.786541634135776e-05,9.79020476837144e-06,8.850873747360267e-05,0.00011616435858333187,tokens/kWh,2203774.058773418,MB,6438.002688,20389.036032,0.0,19742.588928,19408.2304,s,10,843.2092499999999,84.320925,0.002661197598122423,84.32021093750001,84.32392734375,84.325639453125,84.327009140625,"[84.323546875, 84.3200234375, 84.3196015625, 84.321078125, 84.3183359375, 84.3178828125, 84.3218359375, 84.3203984375, 84.3191953125, 84.3273515625]",tokens/s,0.7471455039185113,kWh,0.000995496139741606,0.0005456199585125352,0.004966448584266792,0.006507564682520933,tokens/kWh,9681.040922914766,,s,629,855.5100844726549,1.3601114220550974,0.18017856913180233,1.33838134765625,1.33879111328125,1.3389119140624999,2.8532972265625003,"[1.33838330078125, 1.338135498046875, 1.338103759765625, 1.337954345703125, 1.338292236328125, 1.338123291015625, 1.3379686279296874, 1.337951171875, 1.33804443359375, 1.3383045654296875, 1.3380516357421874, 1.337988037109375, 1.3382154541015625, 1.3380885009765624, 1.3382850341796875, 1.3384017333984375, 1.338282958984375, 1.33829833984375, 1.338177490234375, 1.3382943115234376, 1.338187744140625, 1.33867822265625, 1.338323974609375, 1.338177490234375, 1.3380833740234375, 1.33817236328125, 1.3384478759765626, 1.3384119873046876, 1.3385123291015626, 1.338522705078125, 1.338946533203125, 1.3383382568359374, 1.3386168212890626, 1.33869775390625, 1.3384273681640626, 1.3386065673828125, 1.3384437255859376, 1.33874072265625, 1.33851953125, 1.3385645751953126, 1.33836083984375, 1.3386669921875, 1.3389434814453125, 1.338682373046875, 1.3389168701171874, 1.3387059326171875, 1.338524658203125, 1.338112060546875, 1.3382738037109374, 1.338470458984375, 1.33846630859375, 1.338481689453125, 1.338217529296875, 1.3385523681640625, 1.3385206298828125, 1.3382369384765624, 1.337903076171875, 1.338323974609375, 1.338281982421875, 1.338577880859375, 1.33874072265625, 1.3389075927734375, 2.85612548828125, 1.3380362548828124, 1.3384171142578125, 1.338639404296875, 1.337964599609375, 1.33838134765625, 1.3382850341796875, 1.3384775390625, 1.3385257568359374, 1.3377781982421875, 1.338303466796875, 1.337987060546875, 1.3382584228515626, 1.338038330078125, 1.33821240234375, 1.3383648681640625, 1.338181640625, 1.33842431640625, 1.3387275390625, 1.3381058349609376, 1.3383424072265624, 1.338270751953125, 1.3384775390625, 1.3381263427734376, 1.338206298828125, 1.33825732421875, 1.3382103271484376, 1.3385584716796874, 1.338059814453125, 1.3389271240234375, 1.33803515625, 1.3390428466796875, 1.3384058837890624, 1.3389639892578125, 1.3388267822265625, 1.338429443359375, 1.33853076171875, 1.338356689453125, 1.338428466796875, 1.338724365234375, 1.3384376220703125, 1.338541015625, 1.3387283935546874, 1.3388216552734375, 1.3385850830078125, 1.3389854736328124, 1.3382000732421875, 1.3383629150390626, 1.3383331298828125, 1.3383916015625, 1.3385103759765624, 1.3383638916015625, 1.3384805908203126, 1.338123291015625, 1.3385267333984374, 1.3388809814453124, 1.33817041015625, 1.3378948974609375, 1.33848779296875, 1.33816015625, 1.33851953125, 1.33850927734375, 1.3387950439453125, 2.853568603515625, 1.338101806640625, 1.33815087890625, 1.3382379150390624, 1.33825439453125, 1.3381744384765626, 1.3384129638671876, 1.3381632080078125, 1.3381375732421874, 1.3378150634765624, 1.3382593994140626, 1.33808740234375, 1.337988037109375, 1.3381693115234374, 1.3381324462890625, 1.3380936279296876, 1.338429443359375, 1.338197021484375, 1.33809765625, 1.3381221923828126, 1.338328125, 1.3383106689453126, 1.3385123291015626, 1.338229736328125, 1.338282958984375, 1.3386322021484376, 1.338462158203125, 1.338744873046875, 1.338181640625, 1.3385010986328125, 1.3385123291015626, 1.3384263916015624, 1.3385533447265625, 1.3387633056640624, 1.3384591064453124, 1.338155029296875, 1.338788818359375, 1.338418212890625, 1.3386004638671876, 1.338249267578125, 1.3388902587890625, 1.3383546142578124, 1.3384744873046874, 1.3390633544921875, 1.338366943359375, 1.3388748779296875, 1.3383424072265624, 1.338250244140625, 1.3385625, 1.3382236328125, 1.338692626953125, 1.338902587890625, 1.33848779296875, 1.338251220703125, 1.33867626953125, 1.3384058837890624, 1.33861376953125, 1.3383055419921874, 1.338281982421875, 1.338312744140625, 1.338599365234375, 1.33861376953125, 1.3388585205078125, 2.85312109375, 1.3381170654296874, 1.338102783203125, 1.3384007568359375, 1.3381734619140624, 1.3384765625, 1.3382646484375, 1.337976806640625, 1.3380946044921875, 1.3380577392578126, 1.338217529296875, 1.3380413818359376, 1.3382645263671875, 1.3380198974609374, 1.338167236328125, 1.3382799072265625, 1.3384786376953124, 1.3382943115234376, 1.3386229248046875, 1.3382103271484376, 1.337924560546875, 1.3382840576171875, 1.33852978515625, 1.3382205810546874, 1.3383424072265624, 1.3382379150390624, 1.3384171142578125, 1.33878173828125, 1.3382154541015625, 1.3386751708984375, 1.3386956787109374, 1.338599365234375, 1.3385687255859375, 1.3383782958984376, 1.338692626953125, 1.3380567626953126, 1.338788818359375, 1.3382574462890624, 1.3384222412109374, 1.3383055419921874, 1.338651611328125, 1.3384990234375, 1.3384417724609374, 1.3387919921875, 1.338693603515625, 1.339273193359375, 1.3382420654296876, 1.3387283935546874, 1.3383782958984376, 1.3382728271484374, 1.3388123779296874, 1.3383935546875, 1.338482666015625, 1.3383321533203125, 1.33850732421875, 1.338587158203125, 1.3388779296875, 1.33827685546875, 1.3387684326171876, 1.3382061767578124, 1.3386629638671874, 1.33879296875, 1.3388123779296874, 2.85336572265625, 1.3381539306640624, 1.3380556640625, 1.338523681640625, 1.33804443359375, 1.3385030517578125, 1.338166259765625, 1.3381856689453124, 1.337996337890625, 1.3377669677734374, 1.3383956298828126, 1.33794921875, 1.3381846923828125, 1.337997314453125, 1.3380955810546875, 1.3382861328125, 1.3384765625, 1.3382677001953125, 1.3384693603515625, 1.338175537109375, 1.3383792724609376, 1.3383802490234376, 1.3384591064453124, 1.3381478271484375, 1.33817041015625, 1.33816015625, 1.33810888671875, 1.3385738525390625, 1.3384058837890624, 1.338608642578125, 1.338224609375, 1.338683349609375, 1.33831982421875, 1.338618896484375, 1.3387069091796875, 1.3382994384765625, 1.33880419921875, 1.338345458984375, 1.3385123291015626, 1.33853076171875, 1.338577880859375, 1.338240966796875, 1.33871826171875, 1.3390069580078126, 1.33830859375, 1.3391697998046874, 1.3380157470703125, 1.33829736328125, 1.338114013671875, 1.338156005859375, 1.3383414306640624, 1.3386527099609375, 1.3385123291015626, 1.3385001220703125, 1.338565673828125, 1.338566650390625, 1.338281005859375, 1.33831884765625, 1.3384908447265624, 1.3383905029296874, 1.3385390625, 1.3388421630859375, 1.3386884765625, 2.851640380859375, 1.338324951171875, 1.3379962158203125, 1.3380853271484374, 1.3381488037109375, 1.3381253662109376, 1.3382379150390624, 1.338017822265625, 1.338619873046875, 1.337976806640625, 1.33848681640625, 1.337903076171875, 1.338588134765625, 1.3379849853515624, 1.3383536376953125, 1.3382103271484376, 1.33810888671875, 1.3383853759765625, 1.3383260498046874, 1.3380946044921875, 1.338181640625, 1.3383946533203126, 1.3384130859375, 1.3385594482421874, 1.3382113037109375, 1.338124267578125, 1.338328125, 1.3386270751953124, 1.3382451171875, 1.338799072265625, 1.3383802490234376, 1.3385257568359374, 1.338417236328125, 1.33832080078125, 1.3384765625, 1.3382215576171874, 1.3384263916015624, 1.338396728515625, 1.33853076171875, 1.3386475830078124, 1.3387509765625, 1.3384683837890625, 1.3387899169921875, 1.3387919921875, 1.3385533447265625, 1.3391268310546875, 1.3381468505859375, 1.338397705078125, 1.3383485107421875, 1.33824609375, 1.33825634765625, 1.338388427734375, 1.3384888916015625, 1.33812841796875, 1.3388359375, 1.338244140625, 1.3383956298828126, 1.338101806640625, 1.3383690185546875, 1.3383741455078124, 1.3385155029296876, 1.3386322021484376, 1.33859228515625, 2.854215576171875, 1.338356689453125, 1.3385379638671875, 1.3381529541015624, 1.33810791015625, 1.3381365966796874, 1.3386229248046875, 1.3378641357421874, 1.338134521484375, 1.3378743896484375, 1.3383321533203125, 1.3382052001953124, 1.3383260498046874, 1.338207275390625, 1.3384837646484375, 1.338671142578125, 1.338228759765625, 1.3382061767578124, 1.338640380859375, 1.338203125, 1.3383311767578125, 1.3386884765625, 1.338472412109375, 1.33840380859375, 1.3382052001953124, 1.33817138671875, 1.3381181640625, 1.338577880859375, 1.3382061767578124, 1.338861572265625, 1.3389066162109375, 1.338660888671875, 1.3386341552734375, 1.3385523681640625, 1.3385430908203124, 1.338102783203125, 1.3386177978515625, 1.3383690185546875, 1.338588134765625, 1.3384847412109375, 1.3383792724609376, 1.3389332275390624, 1.3387847900390626, 1.3388912353515625, 1.338240966796875, 1.3393377685546874, 1.3380147705078125, 1.338346435546875, 1.3382327880859375, 1.33857177734375, 1.338472412109375, 1.33872021484375, 1.338575927734375, 1.33836181640625, 1.3385318603515626, 1.338082275390625, 1.3387274169921874, 1.3383782958984376, 1.3383505859375, 1.338428466796875, 1.3386341552734375, 1.3385738525390625, 1.3386802978515624, 2.85828515625, 1.33825439453125, 1.338598388671875, 1.338102783203125, 1.3382010498046875, 1.3382891845703124, 1.338208251953125, 1.337956298828125, 1.3384530029296875, 1.3379942626953125, 1.338576904296875, 1.3380638427734375, 1.338123291015625, 1.3382757568359376, 1.33821240234375, 1.338217529296875, 1.33817041015625, 1.3384171142578125, 1.338460205078125, 1.338281005859375, 1.33812841796875, 1.3382943115234376, 1.3385491943359376, 1.338522705078125, 1.338440673828125, 1.338386474609375, 1.338228759765625, 1.3383526611328125, 1.3381888427734374, 1.338619873046875, 1.3384171142578125, 1.3388001708984374, 1.33842431640625, 1.3388319091796874, 1.3386240234375, 1.3381939697265626, 1.33859228515625, 1.3382901611328124, 1.338334228515625, 1.3385902099609375, 1.33840380859375, 1.3385789794921874, 1.338714111328125, 1.3387325439453126, 1.338462158203125, 1.3388338623046876, 1.3383720703125, 1.33831884765625, 1.338462158203125, 1.33825537109375, 1.3387796630859374, 1.3383209228515625, 1.3384263916015624, 1.3384837646484375, 1.338556396484375, 1.338260498046875, 1.338534912109375, 1.3380771484375, 1.338418212890625, 1.338197021484375, 1.3385902099609375, 1.3388994140625, 1.33888818359375, 2.8586865234375, 1.338609619140625, 1.338398681640625, 1.338145751953125, 1.338203125, 1.338523681640625, 1.3383199462890625, 1.33819482421875, 1.33796142578125, 1.3378519287109376, 1.338119140625, 1.338113037109375, 1.338135498046875, 1.3380516357421874, 1.33864453125, 1.33827587890625, 1.338181640625, 1.3382677001953125, 1.3383045654296875, 1.337975830078125, 1.338377197265625, 1.3383536376953125, 1.3386004638671876, 1.33821435546875, 1.338138671875, 1.3382799072265625, 1.3388739013671875, 1.33859228515625, 1.33829833984375, 1.3384765625, 1.338504150390625, 1.3385277099609374, 1.338335205078125, 1.33824609375, 1.3385943603515624, 1.33834033203125, 1.338545166015625, 1.3384049072265625, 1.3381488037109375, 1.338460205078125, 1.3383106689453126, 1.3387652587890626, 1.3386905517578125, 1.338962890625, 1.338398681640625, 1.338982421875, 1.3382379150390624, 1.3384539794921875, 1.3381795654296875, 1.3384007568359375, 1.33836083984375, 1.338472412109375, 1.338291259765625, 1.338218505859375, 1.338640380859375, 1.338345458984375, 1.3385645751953126, 1.3381519775390625, 1.3383629150390626, 1.3388759765625, 1.3386044921875, 1.33867724609375, 1.3387591552734375, 2.86118408203125, 1.3389434814453125, 1.33874072265625, 1.338156005859375, 1.338240966796875, 1.3385543212890625, 1.3379993896484375, 1.3379625244140625, 1.33812841796875, 1.33773828125, 1.3383751220703124, 1.3379461669921875, 1.3382154541015625, 1.338399658203125, 1.338345458984375, 1.3383638916015625, 1.338197998046875, 1.3385594482421874, 1.3383792724609376, 1.33813037109375, 1.3387049560546875, 1.338354736328125, 1.3387908935546875, 1.338270751953125, 1.338091552734375, 1.33819287109375, 1.3384530029296875, 1.338455078125, 1.338661865234375, 1.338914794921875, 1.33852978515625, 1.33894140625, 1.3384273681640626, 1.33825634765625, 1.3386331787109376, 1.3382471923828125, 1.338392578125, 1.338302490234375, 1.3384324951171875, 1.33825537109375, 1.3385830078125, 1.3388687744140626, 1.3387059326171875, 1.338957763671875, 1.33882470703125, 1.33918408203125, 1.3382420654296876, 1.3383424072265624, 1.3381488037109375, 1.3381571044921876, 1.338545166015625, 1.338265625, 1.33840283203125, 1.338265625, 1.3387376708984375, 1.338366943359375, 1.3383372802734375, 1.338070068359375, 1.3385430908203124, 1.3384478759765626, 1.3390029296875, 1.3386383056640625, 1.3390069580078126]",tokens/s,0.7352338814190846,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15040,7 +15040,7 @@ OSError: google/recurrentgemma-7b is not a local folder and is not a valid model If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15385,7 +15385,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15488,7 +15488,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 96.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15790,7 +15790,7 @@ OSError: 0 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch