diff --git "a/perf-df-gptq-1xA10.csv" "b/perf-df-gptq-1xA10.csv" --- "a/perf-df-gptq-1xA10.csv" +++ "b/perf-df-gptq-1xA10.csv" @@ -229,7 +229,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1452.27776,2426.929152,0.0,1780.482048,1621.410304,s,10,0.7058684158325197,0.07058684158325197,0.0002928328251085697,0.07070625686645508,0.07090748825073243,0.07092075080871582,0.07093136085510254,"[0.07084793853759766, 0.07022029113769532, 0.07070374298095704, 0.07075682830810547, 0.07070877075195313, 0.07024800109863281, 0.07009436798095703, 0.07044992065429688, 0.07093401336669922, 0.070904541015625]",tokens/s,3626.738273847639,kWh,8.283197671383412e-07,4.537640673683759e-07,3.7133809815803766e-06,4.995464816087093e-06,tokens/kWh,51246482.444555126,MB,1452.27776,2426.929152,0.0,1780.482048,1723.272192,s,10,38.030801269531246,3.8030801269531245,0.0020719084747050052,3.802459106445313,3.8038822265625,3.806487744140625,3.808572158203125,"[3.80909326171875, 3.80165771484375, 3.8017744140625, 3.80289404296875, 3.802064697265625, 3.80330322265625, 3.80299072265625, 3.802145751953125, 3.8027724609375, 3.80210498046875]",tokens/s,16.565520025073223,kWh,4.4899062949067045e-05,2.4607307695079302e-05,0.00019991478192082245,0.0002694211525649688,tokens/kWh,233834.6466126413,,s,629,38.604979228973406,0.06137516570584005,0.008407749083102415,0.06034739303588867,0.06043340682983398,0.060483585357666014,0.13084835754394533,"[0.06085631942749024, 0.06110108947753906, 0.06107440185546875, 0.06086963272094727, 0.06092902374267578, 0.060903423309326174, 0.06097305679321289, 0.06068940734863281, 0.060516353607177734, 0.06042521667480469, 0.06034124755859375, 0.06029312133789062, 0.06031155014038086, 0.06031769561767578, 0.060355583190917966, 0.060298240661621094, 0.06034534454345703, 0.06035148620605469, 0.060369918823242184, 0.060316673278808595, 0.06033817672729492, 0.060298240661621094, 0.06036377716064453, 0.06029312133789062, 0.06034534454345703, 0.06041497421264649, 0.060398593902587894, 0.06032076644897461, 0.060527614593505856, 0.06036479949951172, 0.06035148620605469, 0.06034329605102539, 0.06039961624145508, 0.06036172866821289, 0.06039654541015625, 0.06041088104248047, 0.06041190338134766, 0.06058393478393555, 0.060483585357666014, 0.06034431838989258, 0.06034841537475586, 0.06036070251464844, 0.060386302947998044, 0.0604139518737793, 0.06040681457519531, 0.06037091064453125, 0.06038323211669922, 0.06031872177124024, 0.06034739303588867, 0.0603135986328125, 0.0604068489074707, 0.060338111877441404, 0.060382209777832034, 0.06051123046875, 0.06040883255004883, 0.060437503814697265, 0.06033407974243164, 0.06041190338134766, 0.060483585357666014, 0.06047129440307617, 0.060355583190917966, 0.060355583190917966, 0.13078938293457032, 0.060290046691894535, 0.06030131149291992, 0.060345375061035156, 0.06027977752685547, 0.06031769561767578, 0.060283905029296876, 0.06029721450805664, 0.06027775955200195, 0.06029414367675781, 0.06029926300048828, 0.06034022521972656, 0.06031872177124024, 0.06030131149291992, 0.06027468872070312, 0.060336128234863284, 0.06028902435302735, 0.06032281494140625, 0.06031257629394531, 0.06031564712524414, 0.06030438232421875, 0.06032998275756836, 0.06030438232421875, 0.06031155014038086, 0.06029107284545898, 0.06038016128540039, 0.06032179260253906, 0.06031872177124024, 0.060393470764160156, 0.06030847930908203, 0.06031564712524414, 0.06033715057373047, 0.06030131149291992, 0.06035763168334961, 0.06037811279296875, 0.06036787033081055, 0.06038019180297852, 0.06038729476928711, 0.06032998275756836, 0.060353534698486325, 0.06034841537475586, 0.0603351058959961, 0.060355583190917966, 0.06032998275756836, 0.06028083038330078, 0.06039961624145508, 0.06038016128540039, 0.06034841537475586, 0.06039449691772461, 0.06032793426513672, 0.06034841537475586, 0.060371967315673826, 0.06034431838989258, 0.06035865783691406, 0.06034636688232422, 0.06037401580810547, 0.060424190521240234, 0.06042726516723633, 0.060418048858642576, 0.06036377716064453, 0.060437503814697265, 0.060439552307128906, 0.060385280609130856, 0.13130752563476564, 0.060295169830322265, 0.060316673278808595, 0.0603238410949707, 0.060283905029296876, 0.0603135986328125, 0.06028902435302735, 0.060281856536865235, 0.06031872177124024, 0.06030438232421875, 0.06033407974243164, 0.060260353088378904, 0.06028799819946289, 0.060267520904541017, 0.06033919906616211, 0.06032998275756836, 0.06028287887573242, 0.060295169830322265, 0.060303359985351565, 0.060314624786376954, 0.0603054084777832, 0.06033919906616211, 0.06035763168334961, 0.06036070251464844, 0.06036684799194336, 0.06033715057373047, 0.06031052780151367, 0.06033817672729492, 0.060349441528320315, 0.06032076644897461, 0.06037299346923828, 0.06037299346923828, 0.060331008911132813, 0.0604139518737793, 0.060386302947998044, 0.06040576171875, 0.060388351440429686, 0.060371967315673826, 0.06029926300048828, 0.060388351440429686, 0.06034739303588867, 0.060368896484375, 0.060353534698486325, 0.0603351058959961, 0.06035148620605469, 0.060365825653076174, 0.060401664733886716, 0.060434432983398435, 0.06033407974243164, 0.06032998275756836, 0.06036275100708008, 0.06036684799194336, 0.06034124755859375, 0.060365825653076174, 0.06031257629394531, 0.06035968017578125, 0.060437503814697265, 0.06032588958740234, 0.06041088104248047, 0.060349441528320315, 0.060385280609130856, 0.060316673278808595, 0.06039244842529297, 0.13137100219726563, 0.06030847930908203, 0.0605747184753418, 0.060353534698486325, 0.06028902435302735, 0.0603351058959961, 0.060286975860595705, 0.06034739303588867, 0.06035865783691406, 0.06034534454345703, 0.060290046691894535, 0.060437503814697265, 0.060319744110107425, 0.06032076644897461, 0.06028595352172852, 0.06031872177124024, 0.06034841537475586, 0.060665855407714846, 0.06037811279296875, 0.06034329605102539, 0.06041292953491211, 0.0603351058959961, 0.0603422737121582, 0.060314624786376954, 0.06029619216918945, 0.06047129440307617, 0.06029107284545898, 0.06037811279296875, 0.06033203125, 0.06035968017578125, 0.0603422737121582, 0.0603504638671875, 0.0603770866394043, 0.060416000366210934, 0.06037094497680664, 0.0603422737121582, 0.06034431838989258, 0.06035865783691406, 0.0603351058959961, 0.06036380767822266, 0.06032892990112305, 0.060375072479248046, 0.060357601165771484, 0.06030131149291992, 0.060316673278808595, 0.060349441528320315, 0.06036070251464844, 0.06037811279296875, 0.0603054084777832, 0.06037913513183594, 0.06035968017578125, 0.060391422271728515, 0.06034739303588867, 0.06038016128540039, 0.06030847930908203, 0.06037299346923828, 0.06040371322631836, 0.060298240661621094, 0.060401664733886716, 0.06032691192626953, 0.06044160079956055, 0.06035456085205078, 0.06043340682983398, 0.13126860046386718, 0.060421119689941405, 0.060418048858642576, 0.060375038146972655, 0.06035148620605469, 0.06029926300048828, 0.06027264022827149, 0.06029107284545898, 0.06031564712524414, 0.06032691192626953, 0.0603238410949707, 0.060284927368164064, 0.060268543243408204, 0.0603504638671875, 0.06031872177124024, 0.060319744110107425, 0.060298240661621094, 0.0603135986328125, 0.060462078094482424, 0.06032998275756836, 0.06028595352172852, 0.060331008911132813, 0.060298240661621094, 0.06032896041870117, 0.0602716178894043, 0.06033203125, 0.060316673278808595, 0.06044160079956055, 0.06032179260253906, 0.06032896041870117, 0.06029619216918945, 0.06037606430053711, 0.06034739303588867, 0.06036377716064453, 0.06037606430053711, 0.060398593902587894, 0.060385280609130856, 0.060369918823242184, 0.06032793426513672, 0.060407806396484375, 0.06034534454345703, 0.06032076644897461, 0.060368896484375, 0.06032588958740234, 0.060336128234863284, 0.06036070251464844, 0.06034329605102539, 0.060352512359619144, 0.0603238410949707, 0.060331008911132813, 0.06036172866821289, 0.06038323211669922, 0.06035456085205078, 0.06036684799194336, 0.06034841537475586, 0.0603422737121582, 0.060440574645996094, 0.06031052780151367, 0.060365825653076174, 0.06039039993286133, 0.06045798492431641, 0.060355583190917966, 0.060402687072753904, 0.13087129211425783, 0.06030643081665039, 0.0603504638671875, 0.06032486343383789, 0.06032486343383789, 0.06033407974243164, 0.06032691192626953, 0.06029721450805664, 0.060313663482666015, 0.060308414459228514, 0.06031155014038086, 0.06032588958740234, 0.060319744110107425, 0.060314624786376954, 0.06033203125, 0.060319744110107425, 0.06032896041870117, 0.06031769561767578, 0.06030745697021484, 0.06030131149291992, 0.06047436904907227, 0.060437503814697265, 0.06032076644897461, 0.060355583190917966, 0.06031872177124024, 0.06036275100708008, 0.06032179260253906, 0.06032793426513672, 0.0603054084777832, 0.060298240661621094, 0.060365825653076174, 0.06033407974243164, 0.060319744110107425, 0.060388351440429686, 0.06036172866821289, 0.060455936431884766, 0.060388351440429686, 0.060401664733886716, 0.060369918823242184, 0.06038323211669922, 0.06042828750610352, 0.06038937759399414, 0.06034124755859375, 0.06034739303588867, 0.060298240661621094, 0.060333057403564455, 0.06038118362426758, 0.060369918823242184, 0.060900352478027345, 0.06035865783691406, 0.060401664733886716, 0.06035968017578125, 0.0603351058959961, 0.06038118362426758, 0.06034431838989258, 0.06064025497436523, 0.06043545532226562, 0.060369918823242184, 0.06040883255004883, 0.060340255737304685, 0.06042825698852539, 0.06035660934448242, 0.06040063858032227, 0.1306808319091797, 0.06031257629394531, 0.06034636688232422, 0.06033407974243164, 0.0605665283203125, 0.06033715057373047, 0.060300289154052736, 0.06031052780151367, 0.060319744110107425, 0.06033203125, 0.06029312133789062, 0.06029312133789062, 0.06032691192626953, 0.06031155014038086, 0.06032691192626953, 0.0603238410949707, 0.06026137542724609, 0.06029312133789062, 0.060290046691894535, 0.06035660934448242, 0.06030950546264648, 0.06033715057373047, 0.06032896041870117, 0.06042214584350586, 0.06032179260253906, 0.06034431838989258, 0.06031465530395508, 0.06036374282836914, 0.060314624786376954, 0.06031052780151367, 0.06032281494140625, 0.060391422271728515, 0.06036172866821289, 0.06047334289550781, 0.06042726516723633, 0.06035865783691406, 0.060375038146972655, 0.06036479949951172, 0.06031769561767578, 0.06042931365966797, 0.06032799911499023, 0.060305343627929685, 0.06035968017578125, 0.060355583190917966, 0.06034022521972656, 0.06037606430053711, 0.060352512359619144, 0.06036684799194336, 0.06035456085205078, 0.06041088104248047, 0.06037811279296875, 0.0605849609375, 0.060386302947998044, 0.06039654541015625, 0.06033817672729492, 0.06036275100708008, 0.060475391387939455, 0.06033817672729492, 0.06041907119750976, 0.06040371322631836, 0.06048255920410156, 0.060510208129882816, 0.06039449691772461, 0.1310750732421875, 0.06031872177124024, 0.06032793426513672, 0.0603770866394043, 0.06030643081665039, 0.060368896484375, 0.06032179260253906, 0.06029926300048828, 0.06029209518432617, 0.06032998275756836, 0.06029209518432617, 0.06027980804443359, 0.06030131149291992, 0.06029721450805664, 0.060278785705566405, 0.06032793426513672, 0.060319744110107425, 0.06037094497680664, 0.0603422737121582, 0.06032896041870117, 0.060290046691894535, 0.06036070251464844, 0.06032588958740234, 0.0603351058959961, 0.060300289154052736, 0.06034739303588867, 0.0603054084777832, 0.060368896484375, 0.06031564712524414, 0.06034431838989258, 0.06032076644897461, 0.06033919906616211, 0.06032691192626953, 0.06046105575561524, 0.060409854888916016, 0.060349441528320315, 0.060369918823242184, 0.06035865783691406, 0.06032588958740234, 0.06040576171875, 0.060365825653076174, 0.06039244842529297, 0.06039039993286133, 0.060307487487792966, 0.060313568115234376, 0.06037606430053711, 0.06039961624145508, 0.06036377716064453, 0.06030847930908203, 0.0603054084777832, 0.06034534454345703, 0.06039039993286133, 0.06034534454345703, 0.06036684799194336, 0.0603135986328125, 0.06032486343383789, 0.06041292953491211, 0.06047948837280273, 0.06050099182128906, 0.06032076644897461, 0.06044467163085938, 0.06034329605102539, 0.060402687072753904, 0.13153587341308592, 0.060286975860595705, 0.06035148620605469, 0.06040576171875, 0.060283905029296876, 0.060349441528320315, 0.06030847930908203, 0.060300289154052736, 0.06027980804443359, 0.06028287887573242, 0.060319744110107425, 0.06028595352172852, 0.060298240661621094, 0.06046003341674805, 0.06037404632568359, 0.06031766510009766, 0.06031872177124024, 0.060281856536865235, 0.060303359985351565, 0.060319744110107425, 0.06031872177124024, 0.06056959915161133, 0.06044979095458984, 0.06042009735107422, 0.06029619216918945, 0.06031872177124024, 0.06032896041870117, 0.06034841537475586, 0.06034534454345703, 0.060450817108154295, 0.06033615875244141, 0.060324832916259764, 0.060365825653076174, 0.06036684799194336, 0.06040063858032227, 0.06036070251464844, 0.06043340682983398, 0.06036684799194336, 0.06036073684692383, 0.06044566345214844, 0.06033919906616211, 0.060365825653076174, 0.06036377716064453, 0.060349441528320315, 0.06034841537475586, 0.060369918823242184, 0.060381214141845704, 0.06036067199707031, 0.06034431838989258, 0.06032793426513672, 0.060352512359619144, 0.06038118362426758, 0.06040063858032227, 0.06036377716064453, 0.06029721450805664, 0.060442623138427735, 0.06050201416015625, 0.06030643081665039, 0.060416000366210934, 0.06033817672729492, 0.060416000366210934, 0.06034841537475586, 0.06035763168334961, 0.13149183654785157, 0.06031257629394531, 0.06034431838989258, 0.06033817672729492, 0.060267520904541017, 0.06034124755859375, 0.06029209518432617, 0.06033203125, 0.06028799819946289, 0.060331008911132813, 0.06029926300048828, 0.06027980804443359, 0.06034739303588867, 0.06031564712524414, 0.06030745697021484, 0.06032486343383789, 0.0602716178894043, 0.060336128234863284, 0.0603504638671875, 0.06035456085205078, 0.06035763168334961, 0.06036479949951172, 0.06032691192626953, 0.06033715057373047, 0.06032486343383789, 0.06039244842529297, 0.060295169830322265, 0.060368896484375, 0.06030847930908203, 0.06031872177124024, 0.0603351058959961, 0.06036479949951172, 0.06035865783691406, 0.06043033599853516, 0.06038016128540039, 0.0603422737121582, 0.060401664733886716, 0.060371967315673826, 0.0603135986328125, 0.06046003341674805, 0.060333057403564455, 0.060369918823242184, 0.06035456085205078, 0.060352512359619144, 0.0603422737121582, 0.06037094497680664, 0.060349441528320315, 0.060386302947998044, 0.06028902435302735, 0.06037913513183594, 0.06036172866821289, 0.06035865783691406, 0.06035763168334961, 0.060386302947998044, 0.06034124755859375, 0.06039039993286133, 0.06043135833740235, 0.06030131149291992, 0.06038425445556641, 0.06040678405761719, 0.0603873291015625, 0.06038118362426758, 0.06035660934448242]",tokens/s,16.29323503243669,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -271,7 +271,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -546,7 +546,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2472.198144,9028.763648,0.0,8382.316544,7939.283456,s,10,6.1620845947265614,0.6162084594726561,0.0011803163390921698,0.6157510986328125,0.6171693908691406,0.6182771697998047,0.619163392944336,"[0.6155607299804687, 0.6193849487304688, 0.6151373901367188, 0.61642626953125, 0.6163339233398437, 0.6153935546875, 0.6169232177734375, 0.61542236328125, 0.6157410278320312, 0.6157611694335937]",tokens/s,415.4438259725966,kWh,7.26740989420149e-06,3.982230299933529e-06,3.644055202823391e-05,4.769019222236893e-05,tokens/kWh,5367980.040976309,MB,2472.779776,9028.763648,0.0,8382.316544,8233.862144,s,10,338.1937578125,33.81937578125,0.0011103055076484762,33.81973828125,33.820403125000006,33.820582421875,33.820725859375,"[33.81766796875, 33.82027734375, 33.8172265625, 33.82005078125, 33.81898828125, 33.8189453125, 33.82076171875, 33.8196171875, 33.819859375, 33.82036328125]",tokens/s,1.8628374576602387,kWh,0.0003992788065804376,0.0002188394560563585,0.0020008438000201846,0.0026189620626569806,tokens/kWh,24055.331269703638,,s,629,343.20291455078103,0.5456326145481419,0.07326185412723402,0.5368145751953125,0.53704130859375,0.5371090820312501,1.15295984375,"[0.5366138916015625, 0.5368842163085937, 0.5365032958984375, 0.5366651000976562, 0.5367828369140625, 0.5368319702148437, 0.536500244140625, 0.5367971801757813, 0.5363753051757812, 0.5366179809570313, 0.5364674682617188, 0.5371392211914062, 0.5367183227539063, 0.5364090576171875, 0.5364602661132812, 0.5366578979492187, 0.5366630249023437, 0.536791015625, 0.5366445922851563, 0.5370265502929688, 0.5366231079101562, 0.5366497192382812, 0.5368258666992187, 0.5368289184570313, 0.53686376953125, 0.5370091552734375, 0.5367439575195313, 0.5367941284179687, 0.5368914184570313, 0.5367971801757813, 0.5367203979492188, 0.5368668212890625, 0.5365339965820313, 0.5368442993164062, 0.5367408447265625, 0.5368688354492187, 0.5367408447265625, 0.536742919921875, 0.5369190673828125, 0.5368074340820312, 0.5366405029296875, 0.53675830078125, 0.5370582885742188, 0.536890380859375, 0.5368985595703125, 0.5370951538085937, 0.5367521362304688, 0.5369446411132812, 0.5367613525390625, 0.5367725830078125, 0.5367357177734375, 0.5368411865234375, 0.5369948120117187, 0.53707470703125, 0.5366937866210938, 0.5367838745117187, 0.536826904296875, 0.536911865234375, 0.5365247802734375, 0.5369384765625, 0.5366364135742188, 0.5370501098632813, 1.15660498046875, 0.5366784057617188, 0.5368135375976563, 0.5366763305664063, 0.5366251220703125, 0.5368370971679688, 0.5370634155273437, 0.5365442504882812, 0.5366190185546875, 0.5368156127929687, 0.5367030029296875, 0.5364469604492188, 0.5368719482421875, 0.5366988525390625, 0.5366445922851563, 0.5365545043945312, 0.5366353759765625, 0.5364398193359375, 0.5369077758789063, 0.5365125122070312, 0.5368678588867187, 0.5366210327148437, 0.5368207397460938, 0.536680419921875, 0.5368309936523438, 0.536911865234375, 0.5369497680664063, 0.5366302490234375, 0.53686376953125, 0.5368207397460938, 0.536680419921875, 0.5367817993164062, 0.5367613525390625, 0.5367193603515625, 0.5368514404296875, 0.5367982177734375, 0.5366886596679687, 0.536848388671875, 0.5369354248046875, 0.5368606567382812, 0.5369999389648438, 0.5371566162109375, 0.53701123046875, 0.5369251708984375, 0.5368104858398437, 0.53696923828125, 0.5372661743164062, 0.536816650390625, 0.537080810546875, 0.5368729858398438, 0.5366435546875, 0.536722412109375, 0.5367203979492188, 0.5367930908203125, 0.5370296020507812, 0.5369251708984375, 0.5368258666992187, 0.5370675048828125, 0.53682177734375, 0.53661083984375, 0.5368616943359374, 0.5369835815429688, 0.5369682006835937, 1.153555419921875, 0.5368104858398437, 0.536627197265625, 0.536426513671875, 0.5363701782226562, 0.5369129028320313, 0.5365360717773437, 0.5366179809570313, 0.5368780517578124, 0.5365504150390625, 0.5364623413085937, 0.5367234497070312, 0.5365247802734375, 0.5367357177734375, 0.5363619995117187, 0.536642578125, 0.5365043334960937, 0.5364766845703125, 0.5367828369140625, 0.5367777099609375, 0.5369682006835937, 0.53671728515625, 0.5365309448242187, 0.5366927490234376, 0.5366732788085937, 0.536748046875, 0.5368842163085937, 0.536711181640625, 0.5367275390625, 0.53675927734375, 0.5366507568359375, 0.5366527709960938, 0.537112548828125, 0.5368504028320312, 0.5369364624023437, 0.5365678100585938, 0.53705419921875, 0.5368473510742188, 0.536774658203125, 0.5369088134765625, 0.5367490844726562, 0.5367777099609375, 0.5368207397460938, 0.5368125610351563, 0.5366998901367187, 0.536975341796875, 0.5371187133789063, 0.5369088134765625, 0.5368340454101562, 0.5368309936523438, 0.53690673828125, 0.536796142578125, 0.53704296875, 0.5369108276367187, 0.5367992553710937, 0.5370162963867188, 0.536953857421875, 0.5370245361328125, 0.5368985595703125, 0.5370009765625, 0.5369139404296875, 0.5368319702148437, 0.5369620361328125, 1.1536087646484374, 0.53665380859375, 0.5366210327148437, 0.5366343383789063, 0.5365514526367188, 0.5365883178710937, 0.5368463134765625, 0.5365125122070312, 0.536474609375, 0.5365883178710937, 0.5365862426757813, 0.5367296142578125, 0.536585205078125, 0.5366619873046875, 0.5366722412109375, 0.5366896362304687, 0.536859619140625, 0.5366026000976563, 0.5365985107421875, 0.536827880859375, 0.5368822021484375, 0.5369682006835937, 0.5367449340820313, 0.5368606567382812, 0.5368719482421875, 0.5370050659179687, 0.5367684936523438, 0.5367459716796875, 0.5366732788085937, 0.5369620361328125, 0.5367920532226562, 0.536858642578125, 0.5368411865234375, 0.5368186645507812, 0.5369609985351562, 0.5366333618164062, 0.5371084594726563, 0.5371924438476563, 0.5368709106445313, 0.5368985595703125, 0.5369241333007813, 0.5369467163085937, 0.5369282836914062, 0.5369384765625, 0.5369241333007813, 0.5369927978515625, 0.5370040283203125, 0.5367838745117187, 0.5370081176757813, 0.5368955078125, 0.5368207397460938, 0.5365821533203124, 0.537007080078125, 0.5369712524414062, 0.5370511474609375, 0.5367930908203125, 0.5369210815429688, 0.5370777587890625, 0.5367992553710937, 0.5366517944335938, 0.5372211303710938, 0.5368811645507813, 0.5370286254882812, 1.1536322021484375, 0.5366702270507813, 0.53663232421875, 0.5365944213867188, 0.5365933837890625, 0.536669189453125, 0.536585205078125, 0.5366456298828125, 0.5366261596679688, 0.5365391235351562, 0.53661181640625, 0.5368197021484375, 0.5366292724609375, 0.5367183227539063, 0.5363814697265625, 0.5366312866210937, 0.5366026000976563, 0.5365043334960937, 0.5368678588867187, 0.5367070922851562, 0.5367982177734375, 0.5367142333984375, 0.5365944213867188, 0.5368822021484375, 0.536764404296875, 0.5367572631835937, 0.5367725830078125, 0.5366507568359375, 0.5366978759765625, 0.5369057006835938, 0.5367367553710938, 0.5368442993164062, 0.5371207885742187, 0.5368688354492187, 0.536975341796875, 0.5372374877929688, 0.5367255249023437, 0.5367490844726562, 0.536848388671875, 0.5366968383789062, 0.5368555297851563, 0.5369047241210938, 0.53682275390625, 0.5368104858398437, 0.536953857421875, 0.537143310546875, 0.5368975219726563, 0.536890380859375, 0.5368473510742188, 0.5369108276367187, 0.5374085083007812, 0.5368156127929687, 0.5369896850585938, 0.5370736694335938, 0.53701123046875, 0.536953857421875, 0.5369763793945312, 0.5369876708984375, 0.536796142578125, 0.5366702270507813, 0.5370020141601562, 0.5367982177734375, 0.5368473510742188, 1.152858154296875, 0.5366998901367187, 0.536784912109375, 0.5366947631835938, 0.5364510498046875, 0.5367101440429688, 0.536543212890625, 0.5367490844726562, 0.5370357666015625, 0.5365555419921875, 0.5364735717773438, 0.5366886596679687, 0.5366702270507813, 0.5367971801757813, 0.53650634765625, 0.5364786987304687, 0.5364551391601563, 0.53650537109375, 0.5366241455078125, 0.5367337036132812, 0.536774658203125, 0.536574951171875, 0.5367716064453125, 0.537091064453125, 0.5366610107421875, 0.5366507568359375, 0.5369815063476563, 0.53661083984375, 0.5366763305664063, 0.5367521362304688, 0.5368361206054687, 0.5368678588867187, 0.5367817993164062, 0.5368463134765625, 0.53707470703125, 0.5366835327148437, 0.536911865234375, 0.536911865234375, 0.5369528198242187, 0.5372190551757813, 0.5367623901367188, 0.5369282836914062, 0.5367654418945312, 0.5368145751953125, 0.5367992553710937, 0.5374668579101562, 0.5371975708007812, 0.5368463134765625, 0.5368524780273437, 0.5368186645507812, 0.53682177734375, 0.5366507568359375, 0.5368616943359374, 0.5369609985351562, 0.53703271484375, 0.53690673828125, 0.5366845703125, 0.5369559326171875, 0.5369241333007813, 0.5368340454101562, 0.5370408935546875, 0.5368411865234375, 0.5369733276367188, 1.1542640380859375, 0.5367930297851562, 0.5365872802734375, 0.5365022583007812, 0.5364807739257812, 0.5368463134765625, 0.5367828369140625, 0.53667431640625, 0.5366415405273437, 0.536479736328125, 0.5365933837890625, 0.5367357788085938, 0.5367592163085938, 0.5366712036132812, 0.5365125122070312, 0.53663232421875, 0.5365278930664062, 0.5366466064453125, 0.5368186645507812, 0.5369241333007813, 0.5368115234375, 0.5370562744140625, 0.5365463256835937, 0.53686376953125, 0.5368514404296875, 0.5370029907226562, 0.5367296142578125, 0.5366886596679687, 0.5367869262695313, 0.53690673828125, 0.5365933837890625, 0.5370020141601562, 0.5367613525390625, 0.53675830078125, 0.5370337524414063, 0.5367316284179687, 0.5368453369140626, 0.5372139282226562, 0.5369917602539063, 0.5368442993164062, 0.5370337524414063, 0.5368677978515625, 0.5368289184570313, 0.5368361206054687, 0.5369886474609376, 0.5370162963867188, 0.5369282836914062, 0.5369343872070312, 0.53701220703125, 0.53686474609375, 0.536791015625, 0.5366722412109375, 0.5372876586914063, 0.5370562744140625, 0.536880126953125, 0.5368606567382812, 0.5368473510742188, 0.537079833984375, 0.53690673828125, 0.5369467163085937, 0.537111572265625, 0.5369978637695313, 0.5370552368164062, 1.152716796875, 0.5368033447265625, 0.5367306518554688, 0.5365186767578125, 0.5366087646484375, 0.5367869262695313, 0.5367767333984375, 0.5365708618164062, 0.536700927734375, 0.536627197265625, 0.5366988525390625, 0.5369722900390625, 0.5366445922851563, 0.5371094970703125, 0.5364469604492188, 0.5367030029296875, 0.5365135498046875, 0.5366927490234376, 0.5367418823242187, 0.53669580078125, 0.536995849609375, 0.53682275390625, 0.536711181640625, 0.5367275390625, 0.5368361206054687, 0.5368811645507813, 0.5368923950195312, 0.5364756469726563, 0.5366835327148437, 0.5368811645507813, 0.5367879638671875, 0.5370316772460938, 0.5366599731445313, 0.5367091064453124, 0.5370531616210937, 0.5366466674804687, 0.5371033325195312, 0.5367869262695313, 0.5368330078125, 0.536774658203125, 0.5366927490234376, 0.5368402099609375, 0.53682275390625, 0.5367613525390625, 0.5369722900390625, 0.5369609985351562, 0.5371791381835938, 0.5367408447265625, 0.536826904296875, 0.5367736206054687, 0.5367817993164062, 0.5369354248046875, 0.53711669921875, 0.5369241333007813, 0.5370623779296875, 0.5369609985351562, 0.5368463134765625, 0.5369671630859375, 0.536806396484375, 0.5367019653320313, 0.5371135864257812, 0.5369528198242187, 0.53698046875, 1.1529993896484374, 0.5368299560546875, 0.5370091552734375, 0.5366353759765625, 0.5365616455078125, 0.5366558837890625, 0.536468505859375, 0.5366865844726563, 0.5366210327148437, 0.5364827880859375, 0.5366784057617188, 0.5368084716796875, 0.536574951171875, 0.5367470092773438, 0.5364807739257812, 0.5367275390625, 0.5365463256835937, 0.5368893432617188, 0.5367490844726562, 0.5366538696289063, 0.5368145751953125, 0.5367122192382813, 0.536791015625, 0.5369139404296875, 0.5368699340820312, 0.5368401489257812, 0.5368053588867188, 0.5367551879882813, 0.536791015625, 0.5368115234375, 0.5367808227539063, 0.5368504028320312, 0.5366098022460938, 0.536838134765625, 0.5369794311523437, 0.5365626831054687, 0.5368678588867187, 0.5369231567382813, 0.5369620361328125, 0.5371443481445313, 0.5370357666015625, 0.5368053588867188, 0.5367316284179687, 0.5368494262695312, 0.5369876708984375, 0.5369149169921875, 0.5369937744140625, 0.5367244873046875, 0.53688525390625, 0.5367900390625, 0.5368668212890625, 0.5369569091796875, 0.536943603515625, 0.5370767211914063, 0.53701123046875, 0.5369712524414062, 0.5369476928710938, 0.53696923828125, 0.53707568359375, 0.53675830078125, 0.5370572509765625, 0.5368934326171875, 0.5369927978515625, 1.1530455322265625, 0.5370398559570313, 0.536827880859375, 0.53667431640625, 0.5366241455078125, 0.5367562255859375, 0.5369866333007812, 0.5366569213867187, 0.536543212890625, 0.5366067504882812, 0.5366558837890625, 0.536690673828125, 0.5370921020507813, 0.5367521362304688, 0.5363927001953125, 0.5365883178710937, 0.5365883178710937, 0.5364572143554688, 0.5367060546875, 0.5369026489257812, 0.5367357177734375, 0.5368033447265625, 0.5366507568359375, 0.5368565673828125, 0.5367767333984375, 0.536933349609375, 0.5369292602539063, 0.5367490844726562, 0.5366527709960938, 0.53698046875, 0.536901611328125, 0.5368790893554688, 0.5367060546875, 0.5367142333984375, 0.5368207397460938, 0.5366681518554688, 0.536848388671875, 0.5368248291015625, 0.536901611328125, 0.5367930908203125, 0.5368125610351563, 0.5368084716796875, 0.5369190673828125, 0.5370562744140625, 0.5367808227539063, 0.53699072265625, 0.5371361083984375, 0.5368914184570313, 0.5368616943359374, 0.5369302978515625, 0.5368442993164062, 0.536732666015625, 0.5369579467773438, 0.5370081176757813, 0.5371074829101562, 0.5370603637695313, 0.5368606567382812, 0.5370623779296875, 0.5368893432617188, 0.5367756958007812, 0.5370880126953125, 0.5369036865234375, 0.5369425659179687]",tokens/s,1.8327350186501152,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -575,7 +575,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1132,7 +1132,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpginfv8wt/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2073,7 +2073,7 @@ ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please req ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1279.791104,1019.74016,0.0,373.293056,277.410816,s,10,0.3806483840942383,0.03806483840942383,0.0010374051807092283,0.03839825630187988,0.038631001281738284,0.03909522857666015,0.03946661041259766,"[0.03955945587158203, 0.03852783966064453, 0.03837411117553711, 0.038422401428222654, 0.03847126388549805, 0.038292991638183595, 0.03834467315673828, 0.03843852615356445, 0.036084991455078125, 0.03613212966918945]",tokens/s,6725.366787229584,kWh,4.3617786259962035e-07,2.3900503699099743e-07,1.1732389459702955e-06,1.8484218455609134e-06,tokens/kWh,138496523.73174343,MB,1280.086016,1019.74016,0.0,373.293056,323.047424,s,10,23.712255126953128,2.371225512695313,0.003882883813537052,2.3718930664062503,2.376249291992188,2.3762570922851562,2.3762633325195313,"[2.362728515625, 2.37624755859375, 2.373291748046875, 2.367230224609375, 2.371970947265625, 2.371815185546875, 2.376264892578125, 2.37161865234375, 2.37223681640625, 2.3688505859375]",tokens/s,26.56854004931377,kWh,2.7777078536237154e-05,1.5222735153430559e-05,6.781520513363374e-05,0.00011081501882330147,tokens/kWh,568514.9961527847,,s,629,24.01549103546141,0.03818043089898478,0.004509392475924156,0.03761459350585938,0.03786444854736328,0.03806208114624023,0.07505829895019543,"[0.03638784027099609, 0.03659775924682617, 0.036880382537841795, 0.03663974380493164, 0.03670937728881836, 0.03717529678344727, 0.036736000061035154, 0.03655987167358398, 0.03629260635375976, 0.03701657485961914, 0.03663974380493164, 0.0363612174987793, 0.03665100860595703, 0.038844417572021485, 0.03839078521728516, 0.038013950347900394, 0.03768627166748047, 0.03772825622558594, 0.037766143798828124, 0.03749990463256836, 0.037585918426513674, 0.03749683380126953, 0.03755929565429687, 0.03760537719726562, 0.037743614196777346, 0.03752140808105469, 0.0375551986694336, 0.03762995147705078, 0.037495807647705076, 0.03751935958862305, 0.03787571334838867, 0.03770880126953125, 0.0376360969543457, 0.03753472137451172, 0.037599231719970705, 0.03759718322753906, 0.03856595230102539, 0.03801900863647461, 0.037601280212402347, 0.03762073516845703, 0.0376545295715332, 0.03752243041992188, 0.03753881454467774, 0.03760435104370117, 0.03775795364379883, 0.037550079345703126, 0.03786649703979492, 0.03831398391723633, 0.03769036865234375, 0.037601280212402347, 0.03755929565429687, 0.037498878479003905, 0.03746303939819336, 0.03752345657348633, 0.03794432067871094, 0.037713920593261716, 0.03760435104370117, 0.037495807647705076, 0.037591041564941405, 0.03748454284667969, 0.03763916778564453, 0.03788083267211914, 0.07681126403808594, 0.03763711929321289, 0.03831808090209961, 0.03748966217041016, 0.03763507080078125, 0.03758489608764649, 0.03759308624267578, 0.037579776763916016, 0.03772419357299805, 0.03755926513671875, 0.03763097763061524, 0.03747020721435547, 0.03756748962402344, 0.03761663818359375, 0.037901313781738284, 0.03767398452758789, 0.03793817520141601, 0.03772415924072266, 0.037601280212402347, 0.03751321411132812, 0.03760332870483398, 0.03770470428466797, 0.03760435104370117, 0.03767705535888672, 0.03810611343383789, 0.0376995849609375, 0.03758694458007812, 0.03785728073120117, 0.03757878494262695, 0.03765244674682617, 0.03786956787109375, 0.03756032180786133, 0.037838848114013675, 0.037773311614990236, 0.037525505065917966, 0.037884929656982425, 0.03775692749023438, 0.03761151885986328, 0.0378869743347168, 0.03772518539428711, 0.037797889709472655, 0.03774566268920899, 0.037759998321533206, 0.03785113525390625, 0.038040576934814455, 0.03770470428466797, 0.0377077751159668, 0.03755929565429687, 0.037599231719970705, 0.037661697387695314, 0.037553150177001955, 0.037823486328125, 0.03759820938110352, 0.03769651031494141, 0.037648384094238284, 0.03772313690185547, 0.03791667175292969, 0.03774259185791016, 0.03760435104370117, 0.03763711929321289, 0.03762483215332031, 0.03786444854736328, 0.03789516830444336, 0.07634636688232421, 0.03765350341796875, 0.03746713638305664, 0.03770470428466797, 0.03758796691894531, 0.03751116943359375, 0.03785318374633789, 0.037889022827148434, 0.03752345657348633, 0.0374774398803711, 0.037649345397949216, 0.03765760040283203, 0.037743614196777346, 0.0377149429321289, 0.03764329528808594, 0.03746198272705078, 0.037561344146728515, 0.037763072967529294, 0.03768832015991211, 0.03759823989868164, 0.037550048828125, 0.03745382308959961, 0.03772825622558594, 0.03768012619018555, 0.03773952102661133, 0.03763097763061524, 0.03759820938110352, 0.03765657424926758, 0.0381921272277832, 0.037905406951904294, 0.03770675277709961, 0.03752345657348633, 0.03779072189331055, 0.0376627197265625, 0.03760844802856445, 0.03748761749267578, 0.03757670211791992, 0.0375654411315918, 0.03750713729858399, 0.03755820846557617, 0.0375551986694336, 0.03763302230834961, 0.037645313262939455, 0.03746099090576172, 0.03763507080078125, 0.0376258544921875, 0.03769651031494141, 0.0375203857421875, 0.037610496520996094, 0.037612545013427735, 0.03768627166748047, 0.03968307113647461, 0.03810201644897461, 0.03744870376586914, 0.03748761749267578, 0.03760844802856445, 0.037705726623535156, 0.0375654411315918, 0.037533695220947266, 0.037482494354248046, 0.037713920593261716, 0.037579776763916016, 0.037495807647705076, 0.07634534454345702, 0.03747840118408203, 0.037424129486083986, 0.03769753646850586, 0.03747020721435547, 0.03760639953613281, 0.03746815872192383, 0.03759308624267578, 0.03759001541137695, 0.037582847595214845, 0.037572608947753904, 0.037525505065917966, 0.03752959823608398, 0.03775590515136719, 0.03758899307250976, 0.037459968566894535, 0.0377446403503418, 0.037495807647705076, 0.03755929565429687, 0.037438465118408204, 0.03756851196289063, 0.037602302551269534, 0.037515262603759765, 0.03751321411132812, 0.03760639953613281, 0.037541889190673826, 0.03759206390380859, 0.03758694458007812, 0.03766681671142578, 0.037548030853271484, 0.037512191772460936, 0.037582847595214845, 0.037493759155273435, 0.037449726104736326, 0.03746815872192383, 0.03757056045532227, 0.037823486328125, 0.037482494354248046, 0.037579776763916016, 0.03767091369628906, 0.03766886520385742, 0.03751424026489258, 0.03748147201538086, 0.037495807647705076, 0.03765555191040039, 0.03773132705688476, 0.03749273681640625, 0.03760639953613281, 0.03768217468261719, 0.037533695220947266, 0.037394432067871096, 0.03765350341796875, 0.03762278366088867, 0.03748966217041016, 0.03768729782104492, 0.0376514892578125, 0.03766985702514648, 0.03745280075073242, 0.037541889190673826, 0.03762073516845703, 0.03751731109619141, 0.037572608947753904, 0.03764223861694336, 0.07663206481933593, 0.037482494354248046, 0.03758694458007812, 0.03754598236083984, 0.03749785614013672, 0.03743641662597656, 0.03762790298461914, 0.03742105484008789, 0.03755929565429687, 0.037628929138183595, 0.03765350341796875, 0.03744768142700195, 0.03746918487548828, 0.03767500686645508, 0.03757670211791992, 0.03751222229003906, 0.03749884796142578, 0.03762995147705078, 0.03771596908569336, 0.03759718322753906, 0.03778252792358398, 0.03771596908569336, 0.03758182525634766, 0.03766886520385742, 0.0377262077331543, 0.03749683380126953, 0.0374917106628418, 0.037571582794189456, 0.03759513473510742, 0.03750912094116211, 0.037544960021972655, 0.0376627197265625, 0.03753574371337891, 0.037466110229492186, 0.03755212783813477, 0.03748044967651367, 0.03741798400878906, 0.03748761749267578, 0.03759414291381836, 0.03739030456542969, 0.03777740859985351, 0.03743129730224609, 0.037466110229492186, 0.037466110229492186, 0.03757056045532227, 0.03745177459716797, 0.03782451248168945, 0.03930624008178711, 0.03780198287963867, 0.037553150177001955, 0.037572608947753904, 0.037628929138183595, 0.03765350341796875, 0.03766579055786133, 0.03758796691894531, 0.03759718322753906, 0.03784396743774414, 0.038711296081542966, 0.03820236968994141, 0.037833728790283204, 0.03772931289672852, 0.03775075149536133, 0.03758796691894531, 0.07656038665771485, 0.03744153594970703, 0.037580799102783204, 0.03758489608764649, 0.03763507080078125, 0.03759308624267578, 0.03766681671142578, 0.0375470085144043, 0.0374835205078125, 0.03772313690185547, 0.037703678131103514, 0.037541889190673826, 0.0375551986694336, 0.037652511596679684, 0.037497825622558593, 0.03801190567016602, 0.037644287109375, 0.03753881454467774, 0.03759718322753906, 0.03752345657348633, 0.037541889190673826, 0.03757363128662109, 0.0374835205078125, 0.03762176132202148, 0.03808870315551758, 0.03783782577514649, 0.03767705535888672, 0.03754086303710937, 0.03700121688842774, 0.03764223861694336, 0.03752959823608398, 0.037765121459960936, 0.03765862274169922, 0.03768012619018555, 0.03785116958618164, 0.03752854537963867, 0.03746201705932617, 0.037591041564941405, 0.03764022445678711, 0.03770057678222656, 0.03760435104370117, 0.037569534301757815, 0.03757875061035156, 0.037550079345703126, 0.03768320083618164, 0.03784806442260742, 0.037806079864501956, 0.037579776763916016, 0.03758796691894531, 0.03765248107910156, 0.03765862274169922, 0.03774566268920899, 0.03764121627807617, 0.03772313690185547, 0.03784294509887695, 0.037817344665527344, 0.03783987045288086, 0.03761971282958984, 0.03763814544677734, 0.03765862274169922, 0.0377077751159668, 0.03765862274169922, 0.03762995147705078, 0.07627571105957032, 0.03767603302001953, 0.03746713638305664, 0.03762278366088867, 0.03759001541137695, 0.037689342498779296, 0.03778559875488281, 0.03764019012451172, 0.03788390350341797, 0.03751424026489258, 0.03749273681640625, 0.03770880126953125, 0.03754291152954101, 0.037703678131103514, 0.03769036865234375, 0.037574657440185545, 0.03785420989990235, 0.03765657424926758, 0.03755929565429687, 0.037599231719970705, 0.03741593551635742, 0.03765555191040039, 0.0375838737487793, 0.038059009552001956, 0.03752243041992188, 0.03781324768066406, 0.03762790298461914, 0.03782761764526367, 0.03764118576049805, 0.03759513473510742, 0.03786444854736328, 0.037617664337158206, 0.03751731109619141, 0.037566463470458986, 0.03778252792358398, 0.037596160888671876, 0.03751321411132812, 0.03806412887573242, 0.03936972808837891, 0.03786342239379883, 0.03775897598266602, 0.037601280212402347, 0.03762483215332031, 0.03776921463012695, 0.03782860946655273, 0.037449726104736326, 0.037705726623535156, 0.0375654411315918, 0.03755929565429687, 0.03753267288208008, 0.03763302230834961, 0.03753881454467774, 0.037580799102783204, 0.03768832015991211, 0.037574657440185545, 0.037531646728515625, 0.03764940643310547, 0.03754393768310547, 0.03761151885986328, 0.037515262603759765, 0.03761459350585938, 0.03788288116455078, 0.03977830505371094, 0.07646412658691407, 0.03738828659057617, 0.03792588806152344, 0.037748737335205076, 0.03746713638305664, 0.037897216796875, 0.037553150177001955, 0.03762176132202148, 0.037343231201171875, 0.037651454925537106, 0.037405696868896485, 0.0374835205078125, 0.03782758331298828, 0.03766681671142578, 0.03759001541137695, 0.0375838737487793, 0.0378081283569336, 0.03773235321044922, 0.03765760040283203, 0.03766886520385742, 0.03765964889526367, 0.03771187210083008, 0.03750707244873047, 0.03753472137451172, 0.037531646728515625, 0.03755110549926758, 0.03769343948364258, 0.03748659133911133, 0.037596160888671876, 0.03750707244873047, 0.03753881454467774, 0.037820415496826174, 0.03772415924072266, 0.03755929565429687, 0.037553150177001955, 0.03769036865234375, 0.03775897598266602, 0.037594112396240234, 0.03757875061035156, 0.03759513473510742, 0.03752755355834961, 0.03780505752563477, 0.03753267288208008, 0.03759820938110352, 0.03762790298461914, 0.03850447845458985, 0.03777225494384766, 0.03763916778564453, 0.03750297546386719, 0.037561344146728515, 0.037645313262939455, 0.03790848159790039, 0.037582847595214845, 0.037553150177001955, 0.037664768218994144, 0.03753574371337891, 0.03770675277709961, 0.03749075317382813, 0.03756332778930664, 0.03781324768066406, 0.03759718322753906, 0.037773311614990236, 0.037571582794189456, 0.07186124420166015, 0.037190654754638675, 0.03747225570678711, 0.03754908752441406, 0.037579742431640625, 0.03760639953613281, 0.03770982360839844, 0.03739136123657227, 0.03773440170288086, 0.0376360969543457, 0.03769241714477539, 0.03769241714477539, 0.03752345657348633, 0.03768627166748047, 0.03762688064575195, 0.03752140808105469, 0.037733375549316404, 0.03770675277709961, 0.037754878997802735, 0.03753472137451172, 0.03745792007446289, 0.03804876708984375, 0.03773747253417969, 0.037572608947753904, 0.03754291152954101, 0.03757056045532227, 0.03814297485351562, 0.03765350341796875, 0.03766988754272461, 0.03759001541137695, 0.03777536010742188, 0.03753472137451172, 0.037574657440185545, 0.03759718322753906, 0.03750707244873047, 0.03766886520385742, 0.03754393768310547, 0.03745280075073242, 0.03773952102661133, 0.03783168029785156, 0.03754086303710937, 0.03762688064575195, 0.037483551025390624, 0.03754902267456055, 0.03769036865234375, 0.03759718322753906, 0.037456897735595705, 0.03748761749267578, 0.038013950347900394, 0.037768192291259765, 0.03762073516845703, 0.03767705535888672, 0.037928958892822266, 0.03753779220581055, 0.03802828979492188, 0.037585918426513674, 0.039975936889648435, 0.03809791946411133, 0.03769343948364258, 0.037564414978027344, 0.03762278366088867, 0.03766988754272461, 0.03770470428466797, 0.07192781066894531, 0.03523993682861328, 0.03559731292724609, 0.038373374938964845, 0.03784908676147461, 0.03781017684936523, 0.03761663818359375, 0.037765121459960936, 0.037564414978027344, 0.037615615844726565, 0.03755417633056641, 0.038040576934814455, 0.03775692749023438, 0.03775692749023438, 0.037582847595214845, 0.037716991424560545, 0.03759308624267578, 0.03775283050537109, 0.03753062438964844, 0.03760435104370117, 0.03752447891235351, 0.037569534301757815, 0.03765964889526367, 0.03779891204833984, 0.03758182525634766, 0.037698558807373043, 0.037585918426513674, 0.03753062438964844, 0.03767398452758789, 0.037664768218994144, 0.0376627197265625, 0.037787647247314454, 0.03765657424926758, 0.037833728790283204, 0.037781505584716796, 0.037684223175048825, 0.037719039916992186, 0.03794944000244141, 0.037571582794189456, 0.03768627166748047, 0.037579776763916016, 0.037574657440185545, 0.03738521575927734, 0.03818188858032227, 0.0376627197265625, 0.03749990463256836, 0.037795841217041014, 0.03761151885986328, 0.037803009033203126, 0.0377077751159668, 0.03760435104370117, 0.03778867340087891, 0.037615615844726565, 0.03751731109619141, 0.03779379272460937, 0.03823820877075195, 0.03774054336547852, 0.03762688064575195, 0.03750400161743164, 0.037553150177001955, 0.037956607818603515, 0.03771289443969727, 0.037722110748291016]",tokens/s,26.191427819286087,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2018.93888,5274.861568,0.0,4628.414464,4279.515648,s,10,3.0976935119628908,0.30976935119628907,0.0007532381367795811,0.30955081176757815,0.3109149108886719,0.3109400146484375,0.31096009765625,"[0.3092184143066406, 0.3087049560546875, 0.3094501647949219, 0.3094185485839844, 0.30965145874023436, 0.3089674072265625, 0.30985198974609374, 0.31055612182617187, 0.31090933227539064, 0.3109651184082031]",tokens/s,826.4213325539185,kWh,3.650658585206427e-06,1.999819851334905e-06,1.778628527278799e-05,2.3436763709329325e-05,tokens/kWh,10923009.813769454,MB,2023.018496,5274.861568,0.0,4628.414464,4463.184384,s,10,173.1631875,17.31631875,0.0012339063464567417,17.316303710937497,17.317707031250002,17.318111328125,17.318434765625,"[17.316962890625, 17.318515625, 17.315822265625, 17.315650390625, 17.3176171875, 17.314611328125, 17.314958984375, 17.315001953125, 17.31678515625, 17.31726171875]",tokens/s,3.6381866671286587,kWh,0.00020445280675334164,0.00011205687635914728,0.0009877879636304137,0.0013042976467429026,tokens/kWh,48301.8582125973,,s,629,175.6766339416501,0.27929512550341873,0.0368162753573841,0.2748590087890625,0.2750474243164063,0.27512094726562497,0.584229677734375,"[0.27480780029296875, 0.27519488525390623, 0.2747535400390625, 0.27477914428710937, 0.2746726379394531, 0.2748118896484375, 0.2747586669921875, 0.2748160095214844, 0.2747463684082031, 0.2747105407714844, 0.27477197265625, 0.27491226196289065, 0.274830322265625, 0.2747412414550781, 0.27507815551757814, 0.2747678833007812, 0.2746982421875, 0.2746357727050781, 0.2747279357910156, 0.27481292724609374, 0.2747586669921875, 0.27477093505859373, 0.27480267333984376, 0.2746726379394531, 0.2747781066894531, 0.2748037109375, 0.27487026977539064, 0.27482623291015623, 0.2746746826171875, 0.27491021728515624, 0.2748866577148438, 0.27482110595703124, 0.27484774780273435, 0.2748170166015625, 0.2750279541015625, 0.2748395385742187, 0.2748538818359375, 0.2749214782714844, 0.27490509033203125, 0.27485696411132815, 0.2751037292480469, 0.27481396484375, 0.2748989562988281, 0.2748948364257813, 0.2748395385742187, 0.2748098449707031, 0.2748395385742187, 0.2748293151855469, 0.2749644775390625, 0.27495526123046876, 0.27507815551757814, 0.274988037109375, 0.27512933349609375, 0.2749081726074219, 0.2749347839355469, 0.2750074768066406, 0.2749327392578125, 0.27496551513671874, 0.2750074768066406, 0.27491226196289065, 0.27496139526367186, 0.2749798278808594, 0.5839052734375, 0.2747381896972656, 0.2746685485839844, 0.27480780029296875, 0.27462655639648437, 0.2747535400390625, 0.274850830078125, 0.27504537963867187, 0.274882568359375, 0.2748190612792969, 0.27480575561523435, 0.2750689392089844, 0.2746798095703125, 0.27480575561523435, 0.2747381896972656, 0.2746982421875, 0.27474533081054686, 0.27501773071289065, 0.27489996337890626, 0.27472384643554687, 0.2748252258300781, 0.2747166748046875, 0.27489996337890626, 0.2747637634277344, 0.274777099609375, 0.27480679321289064, 0.27467572021484377, 0.2747699279785156, 0.2752634887695313, 0.274724853515625, 0.27481497192382814, 0.27493170166015624, 0.27494503784179686, 0.2750330810546875, 0.27488153076171873, 0.2748272705078125, 0.27500338745117187, 0.27493991088867187, 0.27490609741210936, 0.2749163513183594, 0.274882568359375, 0.2751744079589844, 0.2748958740234375, 0.27488461303710937, 0.27520510864257813, 0.2751129455566406, 0.2748856201171875, 0.27490609741210936, 0.275019775390625, 0.27484979248046876, 0.2750382080078125, 0.27491021728515624, 0.2749378662109375, 0.2749491271972656, 0.27492352294921873, 0.2749931640625, 0.27498086547851563, 0.2750474243164063, 0.27502899169921874, 0.27504434204101563, 0.2750597229003906, 0.27496856689453125, 0.27508428955078124, 0.584416259765625, 0.274740234375, 0.27477093505859373, 0.27474432373046875, 0.2746582946777344, 0.274671630859375, 0.27458456420898436, 0.2746163330078125, 0.27472384643554687, 0.2748641357421875, 0.27479244995117186, 0.27462451171875, 0.2746695556640625, 0.27464703369140625, 0.27479449462890626, 0.2748590087890625, 0.27491738891601564, 0.27486004638671874, 0.2748610534667969, 0.27494503784179686, 0.27507815551757814, 0.27484979248046876, 0.27474533081054686, 0.2746705932617188, 0.2747832336425781, 0.27472897338867186, 0.27475762939453124, 0.27486209106445314, 0.27463372802734376, 0.27475762939453124, 0.27485491943359375, 0.27470437622070315, 0.2752000122070313, 0.27496139526367186, 0.27483444213867186, 0.27475149536132815, 0.274956298828125, 0.2748334045410156, 0.2748170166015625, 0.27502590942382815, 0.27488973999023436, 0.2748323974609375, 0.2750013427734375, 0.27480575561523435, 0.27495834350585935, 0.27506585693359376, 0.27492556762695314, 0.27496139526367186, 0.2750126037597656, 0.274946044921875, 0.27499725341796877, 0.2750269470214844, 0.27491943359375, 0.274872314453125, 0.2750965881347656, 0.27492556762695314, 0.2749634704589844, 0.27498699951171873, 0.27491943359375, 0.2748856201171875, 0.27493887329101563, 0.2748241882324219, 0.2750003051757812, 0.5843558349609375, 0.2747125854492187, 0.27458660888671876, 0.2746224670410156, 0.27471463012695313, 0.2747678833007812, 0.27459890747070315, 0.274724853515625, 0.2747064208984375, 0.2748293151855469, 0.2745948181152344, 0.27471768188476564, 0.2746695556640625, 0.274840576171875, 0.2746921081542969, 0.27465625, 0.27479345703125, 0.27493170166015624, 0.27479962158203125, 0.27482110595703124, 0.27469515991210935, 0.2747832336425781, 0.27483544921875, 0.2746767272949219, 0.2749491271972656, 0.2748887023925781, 0.2751119384765625, 0.274798583984375, 0.274882568359375, 0.27497369384765624, 0.2748590087890625, 0.2748887023925781, 0.27485491943359375, 0.27481805419921873, 0.2747545471191406, 0.27473919677734376, 0.2748108825683594, 0.2749921264648437, 0.27480575561523435, 0.27500543212890627, 0.2749286499023437, 0.27485592651367186, 0.27485592651367186, 0.2749378662109375, 0.2750003051757812, 0.27502490234375, 0.274951171875, 0.2748856201171875, 0.27495013427734377, 0.27501568603515625, 0.2748528747558594, 0.2749849548339844, 0.2749573059082031, 0.27474945068359374, 0.275061767578125, 0.2749859924316406, 0.2748241882324219, 0.2749593505859375, 0.27505050659179686, 0.27484262084960936, 0.27493374633789064, 0.27519793701171874, 0.2749962158203125, 0.5846722412109375, 0.2751098937988281, 0.2747699279785156, 0.27475045776367185, 0.2746501159667969, 0.2746787719726563, 0.27471463012695313, 0.27465728759765623, 0.27470132446289064, 0.27461016845703123, 0.2745999450683594, 0.27479449462890626, 0.27462860107421877, 0.27524710083007814, 0.27474227905273435, 0.27471563720703124, 0.2747084655761719, 0.2747586669921875, 0.27493170166015624, 0.2747279357910156, 0.27475762939453124, 0.27473715209960936, 0.2749266052246094, 0.27491021728515624, 0.27474945068359374, 0.2747340698242188, 0.2748794860839844, 0.27477093505859373, 0.2746552429199219, 0.27478732299804687, 0.2748487548828125, 0.2747637634277344, 0.27487026977539064, 0.27508837890625, 0.2749296569824219, 0.2750433349609375, 0.2749542541503906, 0.2748968811035156, 0.2750556030273438, 0.2749910888671875, 0.2748293151855469, 0.27487744140625, 0.274988037109375, 0.27488973999023436, 0.2748190612792969, 0.274951171875, 0.27495321655273436, 0.2749941711425781, 0.2749327392578125, 0.27491122436523435, 0.27508428955078124, 0.275135498046875, 0.2751324157714844, 0.27496551513671874, 0.2749819030761719, 0.27534335327148435, 0.27508633422851564, 0.27493887329101563, 0.27502899169921874, 0.27484466552734377, 0.2749542541503906, 0.2749644775390625, 0.27518875122070313, 0.5838991088867187, 0.2747965393066406, 0.2747125854492187, 0.2746480712890625, 0.2746009521484375, 0.2746828918457031, 0.274661376953125, 0.2748989562988281, 0.27465625, 0.2748037109375, 0.2750279541015625, 0.2747432861328125, 0.27462042236328127, 0.27462860107421877, 0.2747166748046875, 0.2746152954101562, 0.2747555847167969, 0.27466033935546874, 0.27464602661132814, 0.27475967407226565, 0.2746582946777344, 0.27469720458984376, 0.27470233154296875, 0.27486514282226565, 0.2747135925292969, 0.2746306457519531, 0.2747412414550781, 0.2748241882324219, 0.2748016662597656, 0.27485696411132815, 0.27470745849609374, 0.2747832336425781, 0.2747484130859375, 0.2749245300292969, 0.2750597229003906, 0.27482009887695313, 0.27485696411132815, 0.27487026977539064, 0.27477914428710937, 0.2749245300292969, 0.27483544921875, 0.27486309814453125, 0.2748436584472656, 0.2752860107421875, 0.2748108825683594, 0.274850830078125, 0.27503103637695314, 0.2750986328125, 0.27488153076171873, 0.27500543212890627, 0.27479962158203125, 0.27487335205078123, 0.27490609741210936, 0.2748805236816406, 0.27498904418945314, 0.2749696044921875, 0.275061767578125, 0.27496652221679685, 0.2752071533203125, 0.27506585693359376, 0.2748989562988281, 0.2749747314453125, 0.275009521484375, 0.5848350830078125, 0.274661376953125, 0.27465728759765623, 0.274671630859375, 0.27472076416015623, 0.27464190673828126, 0.2747658386230469, 0.2749992980957031, 0.27464089965820315, 0.2746787719726563, 0.2746992492675781, 0.27476275634765623, 0.27472280883789063, 0.274703369140625, 0.2751016845703125, 0.2746326904296875, 0.2748692626953125, 0.27478427124023436, 0.27471563720703124, 0.2749327392578125, 0.274724853515625, 0.2748958740234375, 0.27469619750976565, 0.2748713073730469, 0.2746582946777344, 0.2747832336425781, 0.274766845703125, 0.2748231811523438, 0.2745599975585937, 0.274904052734375, 0.2747361145019531, 0.27491738891601564, 0.2747361145019531, 0.2747484130859375, 0.27478118896484377, 0.2747330627441406, 0.2748784790039063, 0.27495321655273436, 0.2748743591308594, 0.27503717041015624, 0.27474945068359374, 0.2748610534667969, 0.27481396484375, 0.27484161376953126, 0.274904052734375, 0.2748221435546875, 0.27497369384765624, 0.2749921264648437, 0.2749798278808594, 0.2750535583496094, 0.2748856201171875, 0.27496755981445314, 0.27506381225585935, 0.2750412902832031, 0.27498699951171873, 0.27489178466796876, 0.27523379516601565, 0.2749358215332031, 0.2749378662109375, 0.27495013427734377, 0.27487640380859374, 0.27494708251953126, 0.2748856201171875, 0.5855928344726562, 0.27480679321289064, 0.27471768188476564, 0.27463986206054686, 0.2746644592285156, 0.2746173400878906, 0.2747852783203125, 0.27456820678710936, 0.2747197570800781, 0.2749071350097656, 0.2747197570800781, 0.27468185424804686, 0.2749153137207031, 0.274634765625, 0.274661376953125, 0.2746644592285156, 0.27470950317382814, 0.2747084655761719, 0.2747883605957031, 0.2747637634277344, 0.274766845703125, 0.2748252258300781, 0.2747115478515625, 0.2748784790039063, 0.274845703125, 0.2747166748046875, 0.27474945068359374, 0.2747279357910156, 0.27474533081054686, 0.2749153137207031, 0.27497268676757813, 0.27476171875, 0.27479757690429685, 0.2747678833007812, 0.2748170166015625, 0.27481805419921873, 0.27477914428710937, 0.27481292724609374, 0.27478118896484377, 0.2751651916503906, 0.274798583984375, 0.2750699462890625, 0.27489178466796876, 0.27470437622070315, 0.2748518371582031, 0.27496551513671874, 0.27504537963867187, 0.27489279174804687, 0.2750320739746094, 0.27494503784179686, 0.2748374938964844, 0.27502386474609375, 0.27486514282226565, 0.2750382080078125, 0.2750474243164063, 0.27487026977539064, 0.27480679321289064, 0.2749706115722656, 0.2748958740234375, 0.27530035400390623, 0.2749931640625, 0.27500851440429686, 0.2750474243164063, 0.5861458129882813, 0.2749306945800781, 0.2746849365234375, 0.27467572021484377, 0.2746921081542969, 0.27488461303710937, 0.2747361145019531, 0.27468084716796876, 0.2749081726074219, 0.2747637634277344, 0.2749716491699219, 0.27516619873046877, 0.2746982421875, 0.2747187194824219, 0.27472897338867186, 0.2747330627441406, 0.27489279174804687, 0.27473202514648437, 0.2747821960449219, 0.2746531982421875, 0.27479757690429685, 0.2748590087890625, 0.2747268981933594, 0.2749153137207031, 0.2749378662109375, 0.2747852783203125, 0.2748743591308594, 0.2749091796875, 0.2748692626953125, 0.2747084655761719, 0.274798583984375, 0.2747832336425781, 0.2749624328613281, 0.27485797119140626, 0.2748938293457031, 0.2748518371582031, 0.27482827758789063, 0.2749358215332031, 0.27496038818359375, 0.27486309814453125, 0.2749624328613281, 0.27491738891601564, 0.2747637634277344, 0.2748323974609375, 0.2748641357421875, 0.27496038818359375, 0.2748948364257813, 0.27498086547851563, 0.27496038818359375, 0.274882568359375, 0.2749347839355469, 0.2749542541503906, 0.27495321655273436, 0.27502285766601564, 0.27491326904296876, 0.274830322265625, 0.27490609741210936, 0.2749593505859375, 0.27500851440429686, 0.2748876647949219, 0.2749491271972656, 0.27524505615234374, 0.2749306945800781, 0.5859583740234375, 0.2748016662597656, 0.27472384643554687, 0.2748016662597656, 0.27469619750976565, 0.27500543212890627, 0.2750474243164063, 0.27471563720703124, 0.27477297973632814, 0.27522457885742185, 0.27464190673828126, 0.27464703369140625, 0.27475250244140625, 0.2746992492675781, 0.274629638671875, 0.274724853515625, 0.27475045776367185, 0.27475250244140625, 0.2747535400390625, 0.2747340698242188, 0.2750074768066406, 0.27490509033203125, 0.27470437622070315, 0.2745856018066406, 0.27472897338867186, 0.2749440002441406, 0.27473919677734376, 0.27478631591796876, 0.27475045776367185, 0.2747699279785156, 0.274904052734375, 0.2747361145019531, 0.2749224853515625, 0.27496856689453125, 0.27500543212890627, 0.274766845703125, 0.274840576171875, 0.27487640380859374, 0.2749358215332031, 0.2749900817871094, 0.27512628173828124, 0.274840576171875, 0.2750873718261719, 0.27478936767578127, 0.2749931640625, 0.27499725341796877, 0.27488973999023436, 0.27509963989257813, 0.27493374633789064, 0.2749347839355469, 0.27490609741210936, 0.27505255126953126, 0.2748907470703125, 0.2751744079589844, 0.2749296569824219, 0.27491738891601564, 0.2752255859375, 0.2749153137207031, 0.27497268676757813, 0.2749819030761719, 0.27503616333007813, 0.2748958740234375, 0.27490509033203125]",tokens/s,3.580442008064188,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2503,7 +2503,7 @@ OSError: 8 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2631,7 +2631,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2690,7 +2690,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1443.033088,2426.929152,0.0,1780.482048,1621.410304,s,10,0.7064453201293945,0.07064453201293945,0.0003248218262012489,0.07069225692749023,0.07102691879272462,0.07108656425476075,0.07113428062438965,"[0.07067814636230468, 0.07048297882080078, 0.07070636749267578, 0.07082745361328124, 0.07072316741943359, 0.0701509780883789, 0.07005088043212891, 0.07066547393798828, 0.07101366424560547, 0.07114620971679687]",tokens/s,3623.7765713149647,kWh,8.268914728916795e-07,4.5309206087460127e-07,3.6233265194823877e-06,4.903310053248669e-06,tokens/kWh,52209629.25450497,MB,1443.033088,2426.929152,0.0,1780.482048,1723.272192,s,10,38.051747802734376,3.8051747802734375,0.001633156200721129,3.8047684326171876,3.806425,3.8079255126953124,3.8091259228515626,"[3.809426025390625, 3.806091552734375, 3.80473193359375, 3.804946533203125, 3.80474365234375, 3.8032763671875, 3.805872802734375, 3.804793212890625, 3.803895263671875, 3.803970458984375]",tokens/s,16.556401121599166,kWh,4.491550297112361e-05,2.461545215875538e-05,0.0001956961107137206,0.00026522706584359966,tokens/kWh,237532.31895703336,,s,629,38.62667061996455,0.06140965122410906,0.00841844016264897,0.0603873291015625,0.06047764434814453,0.060527001953125,0.13100806091308592,"[0.060862464904785155, 0.06109286499023438, 0.0608983039855957, 0.06086656188964844, 0.060881919860839843, 0.06083071899414062, 0.06078668975830078, 0.06044364929199219, 0.0603422737121582, 0.06038323211669922, 0.06031872177124024, 0.0603955192565918, 0.0603135986328125, 0.060398593902587894, 0.06039449691772461, 0.060418048858642576, 0.06032486343383789, 0.060388351440429686, 0.06033203125, 0.060393470764160156, 0.06033919906616211, 0.0603955192565918, 0.06034022521972656, 0.06041907119750976, 0.06042009735107422, 0.06042009735107422, 0.06048051071166992, 0.06038016128540039, 0.06032998275756836, 0.06038323211669922, 0.060336128234863284, 0.0603955192565918, 0.06038937759399414, 0.060442623138427735, 0.06038323211669922, 0.06042214584350586, 0.06044569778442383, 0.0603770866394043, 0.06039654541015625, 0.0604139518737793, 0.06038425445556641, 0.060472320556640625, 0.0603422737121582, 0.06043648147583008, 0.06041702270507812, 0.06047334289550781, 0.06039039993286133, 0.06040883255004883, 0.0603504638671875, 0.06043545532226562, 0.06041190338134766, 0.060439552307128906, 0.060388351440429686, 0.060398593902587894, 0.06063206481933594, 0.06055936050415039, 0.06030438232421875, 0.06042726516723633, 0.06035660934448242, 0.060483585357666014, 0.06033919906616211, 0.06044364929199219, 0.131093505859375, 0.060385280609130856, 0.06040576171875, 0.06040063858032227, 0.0603873291015625, 0.060388351440429686, 0.06041292953491211, 0.06032793426513672, 0.06038425445556641, 0.06032486343383789, 0.060647422790527344, 0.0603873291015625, 0.06036070251464844, 0.060319744110107425, 0.06039244842529297, 0.06034841537475586, 0.06040371322631836, 0.06038323211669922, 0.06036275100708008, 0.06034636688232422, 0.060508159637451174, 0.060393470764160156, 0.06036479949951172, 0.060565502166748046, 0.06042521667480469, 0.06041190338134766, 0.060416000366210934, 0.06036377716064453, 0.060386302947998044, 0.06051123046875, 0.06042521667480469, 0.06036172866821289, 0.060401664733886716, 0.06039961624145508, 0.060450817108154295, 0.06038425445556641, 0.06045491027832031, 0.06037401580810547, 0.06037401580810547, 0.06039449691772461, 0.060416000366210934, 0.06036275100708008, 0.060416000366210934, 0.06038016128540039, 0.060416000366210934, 0.06041497421264649, 0.060485633850097656, 0.06034431838989258, 0.060418048858642576, 0.06036479949951172, 0.060453887939453124, 0.06043648147583008, 0.060470272064208984, 0.060470272064208984, 0.06043852615356445, 0.06045183944702148, 0.06048153686523437, 0.06035660934448242, 0.06048255920410156, 0.06037606430053711, 0.06047641754150391, 0.060391422271728515, 0.06046515274047851, 0.13133619689941406, 0.06032896041870117, 0.060440574645996094, 0.060365825653076174, 0.060365825653076174, 0.060278785705566405, 0.06043135833740235, 0.060314624786376954, 0.06039654541015625, 0.060281856536865235, 0.060382209777832034, 0.06027571105957031, 0.06045491027832031, 0.06034329605102539, 0.0603955192565918, 0.06031872177124024, 0.060424190521240234, 0.06031769561767578, 0.060407806396484375, 0.060319744110107425, 0.060404735565185545, 0.06034841537475586, 0.0604139518737793, 0.06036787033081055, 0.06044160079956055, 0.06036275100708008, 0.06041907119750976, 0.0602716178894043, 0.06040576171875, 0.06030131149291992, 0.060447742462158206, 0.06032793426513672, 0.060426239013671876, 0.06038425445556641, 0.060459007263183595, 0.06040371322631836, 0.060437503814697265, 0.06034431838989258, 0.060426239013671876, 0.0604139518737793, 0.06040576171875, 0.06034022521972656, 0.060424190521240234, 0.06030950546264648, 0.06039244842529297, 0.06034739303588867, 0.060590080261230465, 0.0603504638671875, 0.060371967315673826, 0.06032281494140625, 0.060440574645996094, 0.06040371322631836, 0.06040576171875, 0.0603770866394043, 0.06047129440307617, 0.06047129440307617, 0.06056447982788086, 0.060295169830322265, 0.06043033599853516, 0.06050099182128906, 0.060493824005126956, 0.06031769561767578, 0.06044467163085938, 0.1314723815917969, 0.06027468872070312, 0.06040371322631836, 0.06033919906616211, 0.06039961624145508, 0.060336128234863284, 0.06038937759399414, 0.0603054084777832, 0.06041702270507812, 0.06029619216918945, 0.060365825653076174, 0.06028902435302735, 0.06035456085205078, 0.06032691192626953, 0.060402687072753904, 0.06038016128540039, 0.060593151092529295, 0.06032179260253906, 0.06037913513183594, 0.06032076644897461, 0.060369918823242184, 0.06031872177124024, 0.060401664733886716, 0.06030643081665039, 0.06041292953491211, 0.060353534698486325, 0.0603955192565918, 0.06035660934448242, 0.060477439880371096, 0.060314624786376954, 0.06036377716064453, 0.06034739303588867, 0.06039961624145508, 0.0603770866394043, 0.06041190338134766, 0.06040678405761719, 0.060439552307128906, 0.06037299346923828, 0.06051737594604492, 0.06039756774902344, 0.060401664733886716, 0.060365825653076174, 0.06041702270507812, 0.06029107284545898, 0.06039756774902344, 0.06039961624145508, 0.060440574645996094, 0.06035968017578125, 0.06040371322631836, 0.06081024169921875, 0.06043545532226562, 0.06036377716064453, 0.06042214584350586, 0.060398593902587894, 0.06039756774902344, 0.06041088104248047, 0.06047641754150391, 0.06032998275756836, 0.060508159637451174, 0.06032793426513672, 0.060478462219238284, 0.06045286560058594, 0.060424190521240234, 0.13134439086914063, 0.060319744110107425, 0.06065868759155273, 0.06037299346923828, 0.06037913513183594, 0.06034534454345703, 0.06036070251464844, 0.060300289154052736, 0.060336128234863284, 0.06033203125, 0.06041907119750976, 0.06032998275756836, 0.06041292953491211, 0.06030438232421875, 0.060385280609130856, 0.06034124755859375, 0.06038118362426758, 0.06038323211669922, 0.06038016128540039, 0.06030643081665039, 0.06037299346923828, 0.06034841537475586, 0.060352512359619144, 0.06032793426513672, 0.06032896041870117, 0.0603770866394043, 0.06038016128540039, 0.060447742462158206, 0.060442623138427735, 0.06034124755859375, 0.06038323211669922, 0.06034534454345703, 0.060368896484375, 0.06038323211669922, 0.06043033599853516, 0.060409854888916016, 0.060450817108154295, 0.06035660934448242, 0.06036275100708008, 0.06036787033081055, 0.060416000366210934, 0.060407806396484375, 0.06048153686523437, 0.06033817672729492, 0.06041088104248047, 0.06042521667480469, 0.06047334289550781, 0.060470272064208984, 0.060386302947998044, 0.0603238410949707, 0.0603873291015625, 0.060407806396484375, 0.060401664733886716, 0.06037913513183594, 0.06041292953491211, 0.06043033599853516, 0.060494846343994144, 0.060382209777832034, 0.06050611114501953, 0.06032691192626953, 0.06048767852783203, 0.06034124755859375, 0.06044672012329102, 0.13078834533691405, 0.06037401580810547, 0.06045798492431641, 0.06032896041870117, 0.060333057403564455, 0.0602716178894043, 0.06035660934448242, 0.06029414367675781, 0.06038937759399414, 0.06027571105957031, 0.060371967315673826, 0.06027775955200195, 0.060355583190917966, 0.06030745697021484, 0.06040576171875, 0.060273662567138675, 0.06036275100708008, 0.06027264022827149, 0.06040371322631836, 0.060298240661621094, 0.06044364929199219, 0.060303359985351565, 0.06036479949951172, 0.0603422737121582, 0.06037094497680664, 0.06032691192626953, 0.06041292953491211, 0.06036377716064453, 0.06040063858032227, 0.060319744110107425, 0.06039654541015625, 0.06034636688232422, 0.06046515274047851, 0.06034124755859375, 0.06042521667480469, 0.060331008911132813, 0.060421119689941405, 0.060319744110107425, 0.060369918823242184, 0.06033919906616211, 0.06044876861572265, 0.06032998275756836, 0.06040883255004883, 0.06030745697021484, 0.060352512359619144, 0.06034534454345703, 0.06046105575561524, 0.06034022521972656, 0.06038425445556641, 0.06030233764648438, 0.0603955192565918, 0.06034022521972656, 0.060432384490966794, 0.060401664733886716, 0.06036684799194336, 0.060398593902587894, 0.06045695877075195, 0.06030643081665039, 0.06044569778442383, 0.060319744110107425, 0.060483585357666014, 0.06034329605102539, 0.060464126586914066, 0.13067263793945313, 0.06030643081665039, 0.060472320556640625, 0.06033817672729492, 0.06035763168334961, 0.06032998275756836, 0.06036377716064453, 0.06032076644897461, 0.060353534698486325, 0.060290046691894535, 0.06033817672729492, 0.06061875152587891, 0.06047334289550781, 0.060336128234863284, 0.06037299346923828, 0.060353534698486325, 0.06040678405761719, 0.06029414367675781, 0.06038016128540039, 0.060298240661621094, 0.060369918823242184, 0.06039449691772461, 0.060434432983398435, 0.0603351058959961, 0.060401664733886716, 0.060386302947998044, 0.060424190521240234, 0.0603351058959961, 0.060398593902587894, 0.06045798492431641, 0.060388351440429686, 0.060486656188964844, 0.06044467163085938, 0.06045695877075195, 0.06081228637695312, 0.06042316818237305, 0.06042009735107422, 0.06036684799194336, 0.06051430511474609, 0.06054912185668945, 0.06043340682983398, 0.06039039993286133, 0.06043135833740235, 0.0603504638671875, 0.060386302947998044, 0.06043545532226562, 0.06046105575561524, 0.06042521667480469, 0.06042828750610352, 0.0603422737121582, 0.060382209777832034, 0.060402687072753904, 0.060439552307128906, 0.060388351440429686, 0.0604846076965332, 0.06035148620605469, 0.06046515274047851, 0.06030950546264648, 0.06046105575561524, 0.060349441528320315, 0.06053273773193359, 0.060386302947998044, 0.06046105575561524, 0.13130650329589844, 0.06029107284545898, 0.060402687072753904, 0.060393470764160156, 0.06034841537475586, 0.06032896041870117, 0.060375038146972655, 0.06038118362426758, 0.06033919906616211, 0.06034534454345703, 0.06034124755859375, 0.060286975860595705, 0.060375038146972655, 0.060331008911132813, 0.060398593902587894, 0.06032179260253906, 0.06035456085205078, 0.060369918823242184, 0.06035865783691406, 0.06031257629394531, 0.060353534698486325, 0.0603238410949707, 0.06037913513183594, 0.06042009735107422, 0.060404735565185545, 0.06040371322631836, 0.060402687072753904, 0.060303359985351565, 0.060388351440429686, 0.06038425445556641, 0.06051737594604492, 0.06032998275756836, 0.06035968017578125, 0.0603770866394043, 0.06051123046875, 0.06038118362426758, 0.06041497421264649, 0.060391422271728515, 0.060371967315673826, 0.06039449691772461, 0.06054297637939453, 0.06048972702026367, 0.060459007263183595, 0.06039449691772461, 0.06039449691772461, 0.06041497421264649, 0.06044979095458984, 0.060393470764160156, 0.06040576171875, 0.060398593902587894, 0.060398593902587894, 0.06036070251464844, 0.06041702270507812, 0.060416000366210934, 0.06040063858032227, 0.060398593902587894, 0.06043852615356445, 0.06038937759399414, 0.060440574645996094, 0.060385280609130856, 0.0604846076965332, 0.06034022521972656, 0.060485633850097656, 0.13168946838378906, 0.06034329605102539, 0.060402687072753904, 0.06033817672729492, 0.06036070251464844, 0.0602716178894043, 0.060365825653076174, 0.060290046691894535, 0.06037401580810547, 0.060283905029296876, 0.06038323211669922, 0.060278785705566405, 0.060388351440429686, 0.060284927368164064, 0.0603873291015625, 0.06031872177124024, 0.06036479949951172, 0.06031769561767578, 0.060355583190917966, 0.06030131149291992, 0.06037094497680664, 0.060290046691894535, 0.060386302947998044, 0.0603054084777832, 0.06042726516723633, 0.06039961624145508, 0.06043852615356445, 0.06031155014038086, 0.06038016128540039, 0.06030233764648438, 0.060385280609130856, 0.060349441528320315, 0.06043545532226562, 0.06037094497680664, 0.060391422271728515, 0.060353534698486325, 0.060434432983398435, 0.060365825653076174, 0.06041292953491211, 0.06035148620605469, 0.06042931365966797, 0.06034534454345703, 0.06041702270507812, 0.060352512359619144, 0.06036479949951172, 0.06036787033081055, 0.06043545532226562, 0.06035968017578125, 0.06058803176879883, 0.06035763168334961, 0.06043852615356445, 0.06034739303588867, 0.06043340682983398, 0.06037299346923828, 0.06035148620605469, 0.06041292953491211, 0.06050406265258789, 0.06034636688232422, 0.06049280166625977, 0.06036787033081055, 0.06051839828491211, 0.06036479949951172, 0.06042521667480469, 0.1317969970703125, 0.06032691192626953, 0.060418048858642576, 0.06032998275756836, 0.060472320556640625, 0.06033407974243164, 0.060442623138427735, 0.06029414367675781, 0.06045183944702148, 0.060319744110107425, 0.060365825653076174, 0.060295169830322265, 0.06056140899658203, 0.06048153686523437, 0.060404735565185545, 0.06032486343383789, 0.0603873291015625, 0.06028595352172852, 0.060391422271728515, 0.060333057403564455, 0.06036070251464844, 0.06033203125, 0.06034431838989258, 0.06033407974243164, 0.060333057403564455, 0.06036787033081055, 0.06043340682983398, 0.0603955192565918, 0.060391422271728515, 0.060278785705566405, 0.06038937759399414, 0.06032179260253906, 0.06048767852783203, 0.06038323211669922, 0.06038937759399414, 0.060386302947998044, 0.060402687072753904, 0.060371967315673826, 0.06035763168334961, 0.060353534698486325, 0.0603955192565918, 0.06033203125, 0.060382209777832034, 0.060314624786376954, 0.06036377716064453, 0.06034534454345703, 0.060485633850097656, 0.06040063858032227, 0.060355583190917966, 0.06032691192626953, 0.06041088104248047, 0.0603422737121582, 0.060416000366210934, 0.060375038146972655, 0.06036172866821289, 0.060353534698486325, 0.06043852615356445, 0.060333057403564455, 0.060437503814697265, 0.060333057403564455, 0.060472320556640625, 0.06032076644897461, 0.06042828750610352]",tokens/s,16.284085319921278,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2732,7 +2732,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3151,7 +3151,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2508.439552,9028.763648,0.0,8382.316544,7939.283456,s,10,6.160718749999999,0.616071875,0.0005587926705172899,0.6159664306640624,0.6168901062011719,0.6170025787353516,0.6170925567626954,"[0.6156204223632813, 0.6152601318359375, 0.616293212890625, 0.616349365234375, 0.6168651123046875, 0.6156136474609375, 0.6171150512695313, 0.6156689453125, 0.6160707397460937, 0.6158621215820312]",tokens/s,415.5359307710646,kWh,7.268102301491631e-06,3.982613158030504e-06,3.636417941812856e-05,4.7614894877650695e-05,tokens/kWh,5376468.868781654,MB,2508.439552,9028.763648,0.0,8382.316544,8233.862144,s,10,338.1393203125,33.813932031250005,0.0010795636913464568,33.813869140625,33.815248046875,33.8156611328125,33.8159916015625,"[33.814296875, 33.8134296875, 33.81288671875, 33.81338671875, 33.81515625, 33.81607421875, 33.81354296875, 33.8141953125, 33.81207421875, 33.81427734375]",tokens/s,1.8631373583461677,kWh,0.0003992101543976201,0.00021880183463568006,0.001996706808148448,0.002614718797181748,tokens/kWh,24094.36917954772,,s,629,343.1469031372071,0.5455435661958776,0.07325723275974945,0.5367101440429688,0.5369858032226562,0.5370556152343751,1.1530929980468752,"[0.5367275390625, 0.5368402099609375, 0.5365862426757813, 0.5366046752929687, 0.536500244140625, 0.5371934814453125, 0.536384521484375, 0.5366179809570313, 0.536248291015625, 0.5367777099609375, 0.5363753051757812, 0.5365616455078125, 0.5364766845703125, 0.536511474609375, 0.536501220703125, 0.5365964965820312, 0.5364152221679688, 0.5368176879882812, 0.5364592895507813, 0.53726513671875, 0.5365821533203124, 0.5365892944335937, 0.5366384887695312, 0.5367306518554688, 0.5367725830078125, 0.5367736206054687, 0.536537109375, 0.5368207397460938, 0.5368606567382812, 0.536796142578125, 0.5370715942382812, 0.536573974609375, 0.5366773681640625, 0.5369937744140625, 0.5366261596679688, 0.5366814575195312, 0.536542236328125, 0.536753173828125, 0.5365514526367188, 0.5366978759765625, 0.5366343383789063, 0.5366548461914062, 0.536574951171875, 0.536732666015625, 0.5368053588867188, 0.5372456665039063, 0.5365043334960937, 0.536658935546875, 0.5365442504882812, 0.5368237915039062, 0.5367091064453124, 0.5369200439453125, 0.5367900390625, 0.536974365234375, 0.536680419921875, 0.5367030029296875, 0.536838134765625, 0.5368370971679688, 0.5364818115234375, 0.5369077758789063, 0.5365545043945312, 0.537333740234375, 1.1534786376953126, 0.5365780639648438, 0.53663232421875, 0.5365493774414063, 0.53699072265625, 0.5363886108398438, 0.53667431640625, 0.5364254760742188, 0.5368883056640625, 0.536364013671875, 0.5366353759765625, 0.5364930419921875, 0.536753173828125, 0.5364725952148437, 0.5367275390625, 0.5364439086914062, 0.5366087646484375, 0.5365186767578125, 0.5367808227539063, 0.5370828857421875, 0.5370153198242188, 0.5366220703125, 0.536890380859375, 0.536553466796875, 0.5365698852539063, 0.5367623901367188, 0.536953857421875, 0.5364602661132812, 0.5366998901367187, 0.5367900390625, 0.5366876220703125, 0.536721435546875, 0.5366661376953125, 0.5365401611328126, 0.5368104858398437, 0.5369794311523437, 0.5364541625976562, 0.5367470092773438, 0.5366578979492187, 0.5365872802734375, 0.5365186767578125, 0.5366548461914062, 0.5367060546875, 0.536869873046875, 0.5368473510742188, 0.5368237915039062, 0.5370029907226562, 0.5366814575195312, 0.5366640625, 0.536711181640625, 0.5369948120117187, 0.5366640625, 0.53665380859375, 0.5369343872070312, 0.5368289184570313, 0.5369354248046875, 0.5367367553710938, 0.5368514404296875, 0.5366773681640625, 0.5364838256835938, 0.5370531616210937, 0.536764404296875, 0.5369661254882813, 1.153585205078125, 0.5365933837890625, 0.536732666015625, 0.5365339965820313, 0.5364029541015625, 0.5366087646484375, 0.536580078125, 0.536395751953125, 0.5366435546875, 0.536427490234375, 0.5366814575195312, 0.5365401611328126, 0.536468505859375, 0.5364285278320312, 0.5363650512695313, 0.5365278930664062, 0.5366292724609375, 0.536468505859375, 0.53648486328125, 0.5365299072265625, 0.536774658203125, 0.5366128540039062, 0.5364951171875, 0.5365032958984375, 0.5370203857421875, 0.5368985595703125, 0.5366384887695312, 0.53654833984375, 0.5365933837890625, 0.5366578979492187, 0.5368760375976562, 0.5367705688476563, 0.5369251708984375, 0.5367439575195313, 0.5370286254882812, 0.536648681640625, 0.5368391723632813, 0.5365596313476563, 0.5367285766601563, 0.5368002319335937, 0.5367418823242187, 0.536711181640625, 0.5370572509765625, 0.5369292602539063, 0.5370091552734375, 0.5367725830078125, 0.5368463134765625, 0.5368094482421875, 0.5370818481445313, 0.5366753540039062, 0.536648681640625, 0.5365473022460937, 0.536875, 0.5368524780273437, 0.5368463134765625, 0.5367808227539063, 0.5367828369140625, 0.5369835815429688, 0.536774658203125, 0.5366026000976563, 0.5368453369140626, 0.5368463134765625, 0.5369210815429688, 1.1540050048828125, 0.5364725952148437, 0.53673779296875, 0.5366005859375, 0.5364121704101562, 0.536585205078125, 0.5365657348632813, 0.536479736328125, 0.5365575561523438, 0.5364387817382813, 0.53650537109375, 0.5364961547851562, 0.5364869384765625, 0.5364664306640625, 0.53623193359375, 0.5365073852539063, 0.536711181640625, 0.5362216796875, 0.5367019653320313, 0.5367122192382813, 0.5370501098632813, 0.5365514526367188, 0.5365924072265625, 0.536627197265625, 0.5366241455078125, 0.5367828369140625, 0.536974365234375, 0.5365841674804688, 0.5367449340820313, 0.5366261596679688, 0.5368453369140626, 0.5365964965820312, 0.5367470092773438, 0.5365155639648438, 0.5371627807617188, 0.5366190185546875, 0.536848388671875, 0.5366978759765625, 0.5367684936523438, 0.5366149291992187, 0.53686376953125, 0.5367787475585938, 0.536605712890625, 0.536658935546875, 0.537101318359375, 0.5369169921875, 0.5370439453125, 0.5366231079101562, 0.5367838745117187, 0.5364930419921875, 0.5368565673828125, 0.5365831909179688, 0.53692724609375, 0.5368319702148437, 0.536985595703125, 0.5367777099609375, 0.5369139404296875, 0.5370009765625, 0.53667431640625, 0.5366384887695312, 0.537037841796875, 0.53671630859375, 0.5374013671875, 1.1540633544921874, 0.5367183227539063, 0.536574951171875, 0.536537109375, 0.5365288696289062, 0.5366630249023437, 0.5365299072265625, 0.5365872802734375, 0.536580078125, 0.5363568725585938, 0.5365626831054687, 0.5364346923828125, 0.5366302490234375, 0.5368023071289062, 0.5364418334960938, 0.5364080810546875, 0.5367490844726562, 0.5365657348632813, 0.5367930908203125, 0.5366405029296875, 0.5367767333984375, 0.5366978759765625, 0.5365657348632813, 0.536595458984375, 0.5366210327148437, 0.5367633666992188, 0.536784912109375, 0.5366651000976562, 0.5367859497070312, 0.5371719970703125, 0.5367050170898438, 0.5367664794921875, 0.5368033447265625, 0.536553466796875, 0.5369395141601563, 0.536732666015625, 0.53688525390625, 0.5366098022460938, 0.5369927978515625, 0.53686474609375, 0.536637451171875, 0.5367244873046875, 0.5367367553710938, 0.536932373046875, 0.5369876708984375, 0.5368678588867187, 0.5369866333007812, 0.5367203979492188, 0.536911865234375, 0.5367930908203125, 0.536795166015625, 0.53676953125, 0.5369784545898437, 0.5369251708984375, 0.5369395141601563, 0.5367275390625, 0.5368350830078125, 0.5369600219726562, 0.5368616943359374, 0.5367131958007813, 0.5370685424804688, 0.5368555297851563, 0.5369569091796875, 1.1528519287109376, 0.5366435546875, 0.5369937744140625, 0.5363619995117187, 0.5365892944335937, 0.5367408447265625, 0.5365821533203124, 0.5365094604492188, 0.5366405029296875, 0.5366814575195312, 0.5365616455078125, 0.5363660888671875, 0.5365811157226562, 0.536352783203125, 0.5364869384765625, 0.5365616455078125, 0.536542236328125, 0.5364357299804687, 0.5367992553710937, 0.5366364135742188, 0.5370408935546875, 0.5365309448242187, 0.536680419921875, 0.53680126953125, 0.5368299560546875, 0.53682177734375, 0.5368463134765625, 0.536953857421875, 0.536647705078125, 0.5366937866210938, 0.536774658203125, 0.5366599731445313, 0.5367675170898437, 0.5365504150390625, 0.5369415893554688, 0.5367234497070312, 0.5368084716796875, 0.5367572631835937, 0.5370081176757813, 0.5367817993164062, 0.5370224609375, 0.5365892944335937, 0.5371392211914062, 0.5368504028320312, 0.5367367553710938, 0.536853515625, 0.5368709106445313, 0.5366343383789063, 0.5370214233398437, 0.53682177734375, 0.5368330078125, 0.5367500610351562, 0.5370347290039063, 0.5369180297851562, 0.5368719482421875, 0.5372518310546875, 0.5368914184570313, 0.53684326171875, 0.5368125610351563, 0.5365391235351562, 0.5370531616210937, 0.5368402099609375, 0.5369620361328125, 1.1546163330078125, 0.536658935546875, 0.5366415405273437, 0.536447998046875, 0.5365227661132812, 0.536553466796875, 0.536580078125, 0.5364224243164063, 0.5365206909179687, 0.5364510498046875, 0.5366732788085937, 0.5365524291992188, 0.5365759887695313, 0.5364756469726563, 0.5366435546875, 0.5366773681640625, 0.5367859497070312, 0.5364930419921875, 0.5368688354492187, 0.536500244140625, 0.5367050170898438, 0.5366671142578125, 0.53652685546875, 0.5368053588867188, 0.5367131958007813, 0.5365862426757813, 0.536685546875, 0.5364827880859375, 0.5367152709960937, 0.536784912109375, 0.536742919921875, 0.5370839233398438, 0.5367019653320313, 0.5365166015625, 0.5370787963867187, 0.536642578125, 0.5367982177734375, 0.5366282348632813, 0.5368688354492187, 0.5366445922851563, 0.5366927490234376, 0.5366364135742188, 0.536685546875, 0.5366190185546875, 0.5369262084960937, 0.536896484375, 0.5369661254882813, 0.5370511474609375, 0.5368923950195312, 0.5365504150390625, 0.5368422241210937, 0.5365524291992188, 0.5370234985351563, 0.5368914184570313, 0.536858642578125, 0.5367828369140625, 0.537111572265625, 0.5368043823242188, 0.5368790893554688, 0.5366845703125, 0.5368555297851563, 0.5367388305664063, 0.5367633666992188, 1.15302294921875, 0.5366507568359375, 0.5366722412109375, 0.5365514526367188, 0.5365708618164062, 0.5365339965820313, 0.5365176391601563, 0.536474609375, 0.53650634765625, 0.536406005859375, 0.5365780639648438, 0.5364827880859375, 0.5365155639648438, 0.5366302490234375, 0.5365411987304688, 0.53661181640625, 0.53669580078125, 0.5366046752929687, 0.5367203979492188, 0.5367101440429688, 0.5368402099609375, 0.5367101440429688, 0.5366998901367187, 0.53671630859375, 0.536595458984375, 0.5367562255859375, 0.5368985595703125, 0.5366599731445313, 0.5367234497070312, 0.5370531616210937, 0.5366917114257812, 0.536722412109375, 0.5367152709960937, 0.5366405029296875, 0.5371094970703125, 0.5365759887695313, 0.5367122192382813, 0.536563720703125, 0.536896484375, 0.5368258666992187, 0.5366241455078125, 0.5366507568359375, 0.5368135375976563, 0.5368985595703125, 0.5373378295898438, 0.5368923950195312, 0.5369620361328125, 0.53663232421875, 0.5367705688476563, 0.5364285278320312, 0.5368125610351563, 0.5366394653320312, 0.5368565673828125, 0.5369262084960937, 0.5369302978515625, 0.53675927734375, 0.536806396484375, 0.53671630859375, 0.5367633666992188, 0.5365729370117187, 0.5371934814453125, 0.5368186645507812, 0.5369548950195312, 1.1533997802734375, 0.53671630859375, 0.53673779296875, 0.5364172973632813, 0.536574951171875, 0.5363138427734375, 0.5363886108398438, 0.5361704711914063, 0.5366302490234375, 0.5365309448242187, 0.5365596313476563, 0.5366067504882812, 0.5365155639648438, 0.5364132080078124, 0.536447998046875, 0.5364818115234375, 0.5365022583007812, 0.5364735717773438, 0.5368237915039062, 0.536616943359375, 0.5367705688476563, 0.536542236328125, 0.5366087646484375, 0.536531982421875, 0.5365647583007812, 0.5367039794921875, 0.5368289184570313, 0.5365718994140625, 0.5365913696289063, 0.5367357177734375, 0.5367091064453124, 0.53696923828125, 0.5366077270507813, 0.53646337890625, 0.5369139404296875, 0.5366016235351563, 0.5370368041992187, 0.5366947631835938, 0.5368391723632813, 0.5366251220703125, 0.5368023071289062, 0.5366968383789062, 0.5369682006835937, 0.5367828369140625, 0.536826904296875, 0.5367490844726562, 0.5370368041992187, 0.5367285766601563, 0.5367756958007812, 0.5366456298828125, 0.5367992553710937, 0.5366619873046875, 0.536922119140625, 0.5367510986328125, 0.5369384765625, 0.5367971801757813, 0.5368422241210937, 0.5369241333007813, 0.5370265502929688, 0.5366046752929687, 0.5369190673828125, 0.5368411865234375, 0.5369835815429688, 1.1531202392578126, 0.53661083984375, 0.5367684936523438, 0.5365657348632813, 0.5369896850585938, 0.5365094604492188, 0.53654833984375, 0.5365493774414063, 0.5365933837890625, 0.5366210327148437, 0.5365616455078125, 0.5365084228515625, 0.5366517944335938, 0.5366220703125, 0.5364132080078124, 0.5364838256835938, 0.5366138916015625, 0.5365933837890625, 0.53676953125, 0.536748046875, 0.5367439575195313, 0.5366619873046875, 0.5366067504882812, 0.5366651000976562, 0.5365985107421875, 0.5369415893554688, 0.5369354248046875, 0.5366578979492187, 0.5367019653320313, 0.53654833984375, 0.5365944213867188, 0.5366261596679688, 0.5367490844726562, 0.5364592895507813, 0.5370839233398438, 0.5365892944335937, 0.536669189453125, 0.536764404296875, 0.5369815063476563, 0.5365975341796875, 0.5367654418945312, 0.536732666015625, 0.5365339965820313, 0.5369415893554688, 0.5369088134765625, 0.53686376953125, 0.5369405517578125, 0.536875, 0.536875, 0.53671728515625, 0.5368309936523438, 0.53659033203125, 0.5371371459960937, 0.5367890014648438, 0.5368709106445313, 0.5366651000976562, 0.536742919921875, 0.5368914184570313, 0.5369149169921875, 0.5366353759765625, 0.537017333984375, 0.536827880859375, 0.5368944702148437]",tokens/s,1.8330341735548021,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3180,7 +3180,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3222,7 +3222,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 96.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3293,7 +3293,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3365,7 +3365,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3480,7 +3480,7 @@ ChildProcessError: Traceback (most recent call last): RuntimeError: The size of tensor a (60) must match the size of tensor b (32) at non-singleton dimension 2 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3509,7 +3509,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3571,7 +3571,7 @@ ValueError: OPTForCausalLM does not support an attention implementation through ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1712.406528,5964.824576,0.0,5318.377472,5048.328704,s,10,3.4079426879882817,0.34079426879882807,0.0009322383785557938,0.3405218048095703,0.3422028137207031,0.34224659423828124,0.34228161865234374,"[0.33942413330078125, 0.33999859619140627, 0.3406260070800781, 0.3401291809082031, 0.3404176025390625, 0.3401330871582031, 0.34098886108398435, 0.34174176025390623, 0.3421930847167969, 0.3422903747558594]",tokens/s,751.1863415494162,kWh,4.007971358520013e-06,2.195722103369917e-06,2.097056307273452e-05,2.717425653462445e-05,tokens/kWh,9420680.918126099,MB,1712.406528,5964.824576,0.0,5318.377472,5098.333184,s,10,188.29865429687501,18.8298654296875,0.0006673439432085487,18.82976953125,18.8306990234375,18.83094423828125,18.831140410156248,"[18.829521484375, 18.829875, 18.82865625, 18.829970703125, 18.8296640625, 18.83021484375, 18.83064453125, 18.829525390625, 18.831189453125, 18.829392578125]",tokens/s,3.3457488177622916,kWh,0.0002223270491041519,0.00012185299643231701,0.0011337806940608672,0.0014779607395973358,tokens/kWh,42626.30143826695,,s,629,191.06832772827136,0.3037652269129912,0.04052678174686376,0.29887591552734377,0.29901171264648435,0.2990915771484375,0.6395446215820312,"[0.29933056640625, 0.2987888488769531, 0.2987939758300781, 0.2988922729492188, 0.29877658081054687, 0.2988226623535156, 0.2988328857421875, 0.29880218505859374, 0.29885440063476565, 0.2987530212402344, 0.2989803466796875, 0.298977294921875, 0.2988349304199219, 0.2987745361328125, 0.29902029418945314, 0.29885336303710935, 0.2988748779296875, 0.2988410949707031, 0.29882571411132813, 0.2988380126953125, 0.2986956787109375, 0.2987386779785156, 0.29879092407226565, 0.29889739990234376, 0.298745849609375, 0.29895782470703125, 0.2990745544433594, 0.2989486083984375, 0.2988134460449219, 0.2987888488769531, 0.29889535522460936, 0.2988922729492188, 0.29888717651367186, 0.298935302734375, 0.29891070556640625, 0.29895269775390626, 0.2988277893066406, 0.2988584899902344, 0.29889535522460936, 0.29889739990234376, 0.2988717956542969, 0.2989383544921875, 0.2988564453125, 0.29891787719726565, 0.29884210205078127, 0.2988062744140625, 0.2989690856933594, 0.29877862548828127, 0.29881036376953124, 0.29896807861328123, 0.29892098999023436, 0.29895574951171877, 0.29893939208984377, 0.2987796630859375, 0.29892095947265623, 0.2987734985351562, 0.29884722900390626, 0.2990111083984375, 0.2989178466796875, 0.2988789672851562, 0.2989742431640625, 0.2987898559570312, 0.6390538330078125, 0.29874993896484375, 0.2987274169921875, 0.2988769226074219, 0.2988164978027344, 0.29890048217773435, 0.29877658081054687, 0.29891278076171873, 0.2988912658691406, 0.29923736572265625, 0.29886669921875, 0.29883392333984377, 0.29876019287109373, 0.29876223754882814, 0.2988769226074219, 0.29878170776367186, 0.2988779602050781, 0.2988216247558594, 0.2988441467285156, 0.29885952758789064, 0.2987755432128906, 0.29880935668945313, 0.29883392333984377, 0.2988554382324219, 0.29924658203125, 0.2988380126953125, 0.29887387084960937, 0.2988840942382813, 0.2993029174804688, 0.2988584899902344, 0.29871514892578127, 0.2991124572753906, 0.298829833984375, 0.29886053466796875, 0.29897317504882814, 0.2988963928222656, 0.29880523681640625, 0.29894247436523436, 0.29871514892578127, 0.29875201416015623, 0.29894451904296876, 0.29874789428710935, 0.29877658081054687, 0.29893017578125, 0.29889434814453125, 0.2990643310546875, 0.29894451904296876, 0.29872845458984376, 0.2989383544921875, 0.29896295166015624, 0.29883187866210936, 0.2988687438964844, 0.29901412963867186, 0.29885748291015624, 0.2988984375, 0.2988963928222656, 0.29887387084960937, 0.29880935668945313, 0.29898651123046877, 0.2988912658691406, 0.2989486083984375, 0.29897830200195313, 0.2988584899902344, 0.6398330688476562, 0.298903564453125, 0.29880218505859374, 0.2988216247558594, 0.2987315063476563, 0.298777587890625, 0.29872332763671877, 0.2986977233886719, 0.2987960205078125, 0.29879092407226565, 0.298724365234375, 0.29893121337890627, 0.2992732238769531, 0.298814453125, 0.29890866088867185, 0.2988011474609375, 0.2988769226074219, 0.2987274169921875, 0.2989219970703125, 0.29873458862304686, 0.29885235595703125, 0.2989690856933594, 0.2988492736816406, 0.2988031921386719, 0.298861572265625, 0.2988554382324219, 0.2988062744140625, 0.2989547424316406, 0.29884622192382815, 0.2987438049316406, 0.2989117431640625, 0.2988840942382813, 0.29899468994140627, 0.2988564453125, 0.298814453125, 0.298819580078125, 0.29902542114257813, 0.2989270935058594, 0.2990540771484375, 0.2987673645019531, 0.298851318359375, 0.29884518432617185, 0.29885235595703125, 0.2987008056640625, 0.29885336303710935, 0.29878170776367186, 0.2989045715332031, 0.2989260864257812, 0.2990172119140625, 0.298829833984375, 0.29888101196289063, 0.2989363098144531, 0.2988584899902344, 0.2988636169433594, 0.29881753540039063, 0.2989066162109375, 0.29882571411132813, 0.29906329345703125, 0.29887387084960937, 0.2989240417480469, 0.2989045715332031, 0.29904177856445313, 0.29881857299804687, 0.6396661987304687, 0.29890866088867185, 0.29881549072265623, 0.29886465454101563, 0.29882366943359373, 0.2986588134765625, 0.2989281311035156, 0.29879806518554686, 0.2988359680175781, 0.29883905029296876, 0.2988031921386719, 0.2987530212402344, 0.2989854736328125, 0.299030517578125, 0.2989547424316406, 0.29914111328125, 0.2987806701660156, 0.29885952758789064, 0.298814453125, 0.2987591552734375, 0.2988308410644531, 0.29881857299804687, 0.2988554382324219, 0.2988011474609375, 0.2988554382324219, 0.29885440063476565, 0.29888204956054687, 0.29879910278320315, 0.2987745361328125, 0.29876837158203123, 0.29885235595703125, 0.29886770629882814, 0.2988062744140625, 0.29895065307617186, 0.29894552612304687, 0.29887591552734377, 0.29902847290039064, 0.298745849609375, 0.29884722900390626, 0.29874176025390625, 0.29884005737304686, 0.29894247436523436, 0.29886257934570315, 0.2987888488769531, 0.2990315551757812, 0.29888717651367186, 0.29897216796875, 0.29883905029296876, 0.2989117431640625, 0.2990663757324219, 0.2989895629882813, 0.29887591552734377, 0.298925048828125, 0.2989056091308594, 0.29893939208984377, 0.29892095947265623, 0.29896395874023435, 0.2988687438964844, 0.2989609069824219, 0.29890765380859374, 0.2990243835449219, 0.2988287963867188, 0.29879092407226565, 0.6396928100585938, 0.29893939208984377, 0.29887283325195313, 0.29885440063476565, 0.2989383544921875, 0.29895269775390626, 0.2988441467285156, 0.2988482666015625, 0.2989619140625, 0.29886257934570315, 0.29881549072265623, 0.2989117431640625, 0.2988072814941406, 0.29882571411132813, 0.2989240417480469, 0.29879296875, 0.2987960205078125, 0.29870181274414065, 0.29926913452148435, 0.29865475463867186, 0.2987703857421875, 0.29870181274414065, 0.2988308410644531, 0.29877658081054687, 0.2987857971191406, 0.29879193115234376, 0.298919921875, 0.2987591552734375, 0.2987806701660156, 0.298919921875, 0.2987694091796875, 0.29903359985351563, 0.29880523681640625, 0.2990408020019531, 0.2988523254394531, 0.29895782470703125, 0.2988554382324219, 0.2988769226074219, 0.2992046203613281, 0.298925048828125, 0.29894247436523436, 0.29885336303710935, 0.2989486083984375, 0.29879296875, 0.29882571411132813, 0.2989240417480469, 0.29892913818359373, 0.29900799560546876, 0.29885952758789064, 0.29900390625, 0.29886257934570315, 0.2988062744140625, 0.29906842041015624, 0.29885336303710935, 0.29888101196289063, 0.2989056091308594, 0.2989219970703125, 0.2988503112792969, 0.2989137878417969, 0.29884005737304686, 0.2988789672851562, 0.29902029418945314, 0.29878988647460936, 0.6392319946289062, 0.29863629150390625, 0.2988984375, 0.29874688720703124, 0.29883392333984377, 0.2988584899902344, 0.29881857299804687, 0.29880218505859374, 0.29896600341796875, 0.29875506591796874, 0.2988287963867188, 0.2988226623535156, 0.2989434814453125, 0.29880218505859374, 0.2989854736328125, 0.2987653198242187, 0.29884005737304686, 0.29883187866210936, 0.2989383544921875, 0.29896295166015624, 0.298893310546875, 0.2992261047363281, 0.29879296875, 0.29881036376953124, 0.2989045715332031, 0.29901412963867186, 0.2989537353515625, 0.29883905029296876, 0.2987694091796875, 0.2988349304199219, 0.29899468994140627, 0.2988861389160156, 0.29891787719726565, 0.29884518432617185, 0.2990172119140625, 0.2990745544433594, 0.29899264526367186, 0.29902029418945314, 0.29891278076171873, 0.2989342651367187, 0.2989270935058594, 0.2989557800292969, 0.2988441467285156, 0.29887081909179686, 0.2988656311035156, 0.29873358154296875, 0.29878680419921877, 0.2989465637207031, 0.2988912658691406, 0.298819580078125, 0.29887283325195313, 0.29891583251953124, 0.29893121337890627, 0.29889022827148437, 0.2988380126953125, 0.299009033203125, 0.2988410949707031, 0.2991124572753906, 0.29884622192382815, 0.2988584899902344, 0.29896395874023435, 0.29894451904296876, 0.29879705810546875, 0.6401351928710938, 0.2988328857421875, 0.2987581481933594, 0.2988840942382813, 0.29872024536132813, 0.29893939208984377, 0.2988707885742187, 0.29880523681640625, 0.2989363098144531, 0.2989342651367187, 0.2987714538574219, 0.29879092407226565, 0.2989823913574219, 0.29887387084960937, 0.2987889099121094, 0.2989301452636719, 0.2989189147949219, 0.2988789672851562, 0.2987530212402344, 0.29921383666992185, 0.2988380126953125, 0.2988083190917969, 0.29881753540039063, 0.2987274169921875, 0.2992404479980469, 0.298977294921875, 0.2987080078125, 0.2987703857421875, 0.2988779602050781, 0.2988001403808594, 0.29884722900390626, 0.2989322509765625, 0.2988431091308594, 0.2989875183105469, 0.29900698852539065, 0.298787841796875, 0.2988380126953125, 0.2987663269042969, 0.29904177856445313, 0.2989045715332031, 0.2989270935058594, 0.2989854736328125, 0.2989793395996094, 0.2989168701171875, 0.29900286865234377, 0.2988134460449219, 0.2988492736816406, 0.2989066162109375, 0.2990909729003906, 0.29900079345703123, 0.2989189147949219, 0.2989056091308594, 0.29892913818359373, 0.29894964599609375, 0.2989598693847656, 0.29893017578125, 0.29894659423828124, 0.2988861083984375, 0.29896295166015624, 0.29888204956054687, 0.2989690856933594, 0.2989066162109375, 0.2988011474609375, 0.64087451171875, 0.2988800048828125, 0.2988359680175781, 0.29870693969726564, 0.2989414367675781, 0.2989752197265625, 0.29876223754882814, 0.2988226623535156, 0.29887591552734377, 0.2989363098144531, 0.298777587890625, 0.29906228637695315, 0.2988083190917969, 0.29909197998046877, 0.2988851318359375, 0.29884210205078127, 0.29921075439453126, 0.2988840942382813, 0.2988717956542969, 0.2987806701660156, 0.29890252685546875, 0.2987325439453125, 0.2989486083984375, 0.29873358154296875, 0.2988717956542969, 0.29869158935546875, 0.29887591552734377, 0.2990520324707031, 0.2989014892578125, 0.2989066162109375, 0.2988083190917969, 0.2989014892578125, 0.29879705810546875, 0.29891278076171873, 0.29888101196289063, 0.2989956970214844, 0.29885440063476565, 0.29872845458984376, 0.29881036376953124, 0.29893939208984377, 0.2989168701171875, 0.29879092407226565, 0.2989588623046875, 0.2989014892578125, 0.29880218505859374, 0.29904794311523436, 0.2987724914550781, 0.29912576293945314, 0.29889022827148437, 0.2988779602050781, 0.2988830871582031, 0.2989260864257812, 0.2988697509765625, 0.29885235595703125, 0.2987653198242187, 0.29883392333984377, 0.29885952758789064, 0.29876837158203123, 0.29888922119140626, 0.298998779296875, 0.2989137878417969, 0.2988769226074219, 0.2987448425292969, 0.6411837158203125, 0.2987673645019531, 0.2987755432128906, 0.2987161560058594, 0.2988308410644531, 0.2989434814453125, 0.29867929077148436, 0.2986700744628906, 0.29879296875, 0.29884622192382815, 0.29911654663085935, 0.2989598693847656, 0.2990049133300781, 0.2988994445800781, 0.2992814025878906, 0.29890252685546875, 0.2989742126464844, 0.29876019287109373, 0.29890252685546875, 0.2987694091796875, 0.29888922119140626, 0.2989137878417969, 0.29885440063476565, 0.2989066162109375, 0.2988206176757813, 0.29879296875, 0.2987888488769531, 0.29874893188476564, 0.298829833984375, 0.29893734741210937, 0.29940017700195315, 0.29889434814453125, 0.298829833984375, 0.2988830871582031, 0.2988994445800781, 0.2990162048339844, 0.2989823913574219, 0.29884622192382815, 0.29896395874023435, 0.2990172119140625, 0.29889022827148437, 0.2988503112792969, 0.298745849609375, 0.2989895629882813, 0.2990100402832031, 0.29893734741210937, 0.2988861389160156, 0.29892913818359373, 0.29891583251953124, 0.29912371826171874, 0.298935302734375, 0.29889739990234376, 0.29890969848632815, 0.2988707885742187, 0.2987857971191406, 0.298977294921875, 0.29891278076171873, 0.2990110778808594, 0.2991124572753906, 0.2989363098144531, 0.2988584899902344, 0.2989465637207031, 0.29895065307617186, 0.641522705078125, 0.298787841796875, 0.29876837158203123, 0.29884722900390626, 0.29865164184570314, 0.2988779602050781, 0.29880117797851563, 0.298661865234375, 0.29881036376953124, 0.2990540771484375, 0.29873663330078126, 0.2987796630859375, 0.2988769226074219, 0.2989619140625, 0.2989916076660156, 0.298925048828125, 0.29885235595703125, 0.298861572265625, 0.29878680419921877, 0.29885235595703125, 0.2988216247558594, 0.2989045715332031, 0.29874893188476564, 0.29875711059570315, 0.29881036376953124, 0.29895269775390626, 0.29883392333984377, 0.29884518432617185, 0.2988687438964844, 0.2988963928222656, 0.29879092407226565, 0.29900390625, 0.29879705810546875, 0.2992803955078125, 0.298956787109375, 0.29879910278320315, 0.29885336303710935, 0.29882366943359373, 0.298935302734375, 0.2987694091796875, 0.298777587890625, 0.29879092407226565, 0.29875201416015623, 0.29888204956054687, 0.29894757080078127, 0.29877862548828127, 0.29879705810546875, 0.29888922119140626, 0.2988441467285156, 0.29900698852539065, 0.29907659912109374, 0.2989619140625, 0.29929470825195315, 0.2988861389160156, 0.2989270935058594, 0.29891278076171873, 0.29884210205078127, 0.2989056091308594, 0.29896600341796875, 0.2988984375, 0.29889022827148437, 0.29891583251953124, 0.2989547424316406]",tokens/s,3.2920160420021825,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3746,7 +3746,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,4118.597632,5287.44448,0.0,4640.997376,4330.668544,s,10,3.1970594482421872,0.31970594482421877,0.0012811478039207278,0.31938397216796877,0.32123822631835935,0.32139851684570314,0.3215267492675781,"[0.31836029052734377, 0.3215588073730469, 0.31788543701171873, 0.319045654296875, 0.3190616149902344, 0.31840402221679687, 0.31970632934570314, 0.32065835571289064, 0.32120260620117186, 0.32117633056640627]",tokens/s,800.7358140955256,kWh,3.7591551740964256e-06,2.0592643818247325e-06,1.8986638453191707e-05,2.4805058009112864e-05,tokens/kWh,10320475.763691055,MB,4118.597632,5287.44448,0.0,4640.997376,4515.667456,s,10,175.13562109375,17.513562109375,0.0017768259462071251,17.513850585937497,17.5153654296875,17.51584091796875,17.51622130859375,"[17.5152421875, 17.5134375, 17.51631640625, 17.5116796875, 17.514263671875, 17.511447265625, 17.51483984375, 17.511279296875, 17.515259765625, 17.51185546875]",tokens/s,3.59721224080829,kWh,0.0002068085063662794,0.0001133488492668755,0.0010180742467503973,0.0013382316023835524,tokens/kWh,47077.05294643273,,s,629,177.73623620605477,0.2825695329190059,0.038042917370424166,0.27798220825195313,0.27814030151367186,0.2782353515625,0.5972927026367187,"[0.27811532592773436, 0.2778900451660156, 0.27781427001953124, 0.278012939453125, 0.27799038696289063, 0.2779176940917969, 0.278012939453125, 0.27791973876953124, 0.27834060668945315, 0.27785113525390626, 0.277939208984375, 0.278029296875, 0.27790643310546875, 0.2777313232421875, 0.277923828125, 0.27794125366210937, 0.2779473876953125, 0.2780712890625, 0.2780887145996094, 0.27796990966796875, 0.27798834228515623, 0.27780609130859374, 0.27821261596679686, 0.27799859619140627, 0.2780968933105469, 0.277833740234375, 0.27810302734375, 0.2780989379882812, 0.27795455932617186, 0.2779525146484375, 0.278108154296875, 0.27808154296875, 0.2780149841308594, 0.27790234375, 0.27806515502929685, 0.2780364685058594, 0.2780139465332031, 0.277865478515625, 0.277928955078125, 0.27805490112304687, 0.27810406494140627, 0.27789926147460936, 0.2781460571289063, 0.2780682373046875, 0.2780078125, 0.2779381713867187, 0.27819622802734373, 0.2781051025390625, 0.2781255798339844, 0.27789825439453125, 0.2779596862792969, 0.2780364685058594, 0.27797607421875, 0.2783692932128906, 0.27815115356445314, 0.27802420043945314, 0.27813067626953125, 0.27788595581054687, 0.2780682373046875, 0.2780968933105469, 0.278118408203125, 0.27799859619140627, 0.5999585571289062, 0.2779914245605469, 0.2784163818359375, 0.2778818664550781, 0.27786648559570315, 0.278034423828125, 0.2778347473144531, 0.2779862976074219, 0.2777989196777344, 0.27785830688476565, 0.2777702331542969, 0.2777927551269531, 0.27792691040039064, 0.278029296875, 0.27775796508789063, 0.2779228210449219, 0.2779402160644531, 0.27796173095703125, 0.27776101684570315, 0.2779862976074219, 0.27793304443359373, 0.2779596862792969, 0.27787982177734377, 0.2779207763671875, 0.2779576416015625, 0.2780917663574219, 0.27807537841796875, 0.27791256713867185, 0.27790130615234376, 0.2779842529296875, 0.2779504699707031, 0.2783477783203125, 0.2779279479980469, 0.27805490112304687, 0.2778204040527344, 0.2778542175292969, 0.2779811706542969, 0.27787774658203124, 0.2782310485839844, 0.27790438842773435, 0.2780006408691406, 0.27790130615234376, 0.2778408813476563, 0.2780907592773437, 0.2779299926757812, 0.277939208984375, 0.27783578491210936, 0.27826174926757813, 0.27810098266601563, 0.27833343505859376, 0.277970947265625, 0.27794842529296876, 0.278012939453125, 0.2779811706542969, 0.2778101806640625, 0.27804364013671873, 0.27798016357421873, 0.27800473022460936, 0.2778818664550781, 0.278002685546875, 0.2779033508300781, 0.2780405883789063, 0.2782843017578125, 0.5971548461914062, 0.2779535217285156, 0.27780197143554686, 0.278213623046875, 0.27794943237304687, 0.2779422607421875, 0.27781427001953124, 0.2779504699707031, 0.2779320373535156, 0.2779535217285156, 0.2779095153808594, 0.27794943237304687, 0.27799755859375, 0.27791156005859374, 0.27797402954101563, 0.2780139465332031, 0.2778623962402344, 0.2780057678222656, 0.27779483032226565, 0.27794943237304687, 0.2780641174316406, 0.2779781188964844, 0.27794329833984377, 0.27810714721679686, 0.2781757507324219, 0.2780313720703125, 0.27787673950195313, 0.27807537841796875, 0.27811431884765625, 0.27794842529296876, 0.2779955139160156, 0.2781460571289063, 0.27803237915039064, 0.27800167846679685, 0.2780262451171875, 0.27801803588867186, 0.2779781188964844, 0.2780405883789063, 0.27790438842773435, 0.2783918151855469, 0.2780979309082031, 0.2779996032714844, 0.2779207763671875, 0.27818597412109375, 0.27806207275390626, 0.2780201110839844, 0.27812966918945314, 0.27813479614257813, 0.278192138671875, 0.27825457763671874, 0.2780149841308594, 0.2780989379882812, 0.2780907592773437, 0.27817166137695315, 0.2779586486816406, 0.2780498046875, 0.2779791259765625, 0.2781583251953125, 0.27810202026367187, 0.2781614074707031, 0.2780334167480469, 0.278002685546875, 0.278392822265625, 0.5973463134765625, 0.2779453430175781, 0.27787774658203124, 0.277865478515625, 0.278029296875, 0.27785009765625, 0.2778736572265625, 0.2779053955078125, 0.27785009765625, 0.27782962036132813, 0.2777774047851562, 0.27799755859375, 0.27794943237304687, 0.27787060546875, 0.2777999267578125, 0.27798220825195313, 0.2778961791992188, 0.2779381713867187, 0.2778531799316406, 0.2780928039550781, 0.2779719543457031, 0.27791668701171873, 0.2777712707519531, 0.27799859619140627, 0.27791256713867185, 0.2778951721191406, 0.27785830688476565, 0.2779781188964844, 0.27801089477539065, 0.278044677734375, 0.27773849487304686, 0.2779535217285156, 0.2779648132324219, 0.2780498046875, 0.2778531799316406, 0.2779576416015625, 0.2779156494140625, 0.278012939453125, 0.27791461181640625, 0.27799038696289063, 0.27808154296875, 0.2779852905273438, 0.2778204040527344, 0.27817779541015625, 0.2780508117675781, 0.27812353515625, 0.2779873352050781, 0.27802215576171874, 0.27810202026367187, 0.27819110107421874, 0.27801190185546876, 0.2780487670898438, 0.2780610656738281, 0.27813479614257813, 0.2779105224609375, 0.2779576416015625, 0.27790643310546875, 0.27801190185546876, 0.27795455932617186, 0.2779381713867187, 0.2780231628417969, 0.2780712890625, 0.27800985717773435, 0.59753369140625, 0.27792486572265623, 0.2778367919921875, 0.278245361328125, 0.27796173095703125, 0.2780078125, 0.2778153076171875, 0.27805902099609375, 0.27785626220703125, 0.2779832458496094, 0.27780813598632814, 0.2779053955078125, 0.2779320373535156, 0.2779648132324219, 0.2777978820800781, 0.27794638061523436, 0.2779893798828125, 0.27795660400390626, 0.2779258728027344, 0.2780149841308594, 0.2778623962402344, 0.27793508911132814, 0.27788803100585935, 0.27812042236328127, 0.2779473876953125, 0.2780477294921875, 0.27788491821289063, 0.27803750610351563, 0.2779207763671875, 0.2780426330566406, 0.2778951721191406, 0.27804672241210937, 0.2779781188964844, 0.27805902099609375, 0.277796875, 0.2779637756347656, 0.27799859619140627, 0.27796990966796875, 0.2780088195800781, 0.2779504699707031, 0.2780159912109375, 0.27801702880859375, 0.277865478515625, 0.2781685791015625, 0.27800677490234377, 0.2781051025390625, 0.2779996032714844, 0.27807437133789065, 0.27806207275390626, 0.278139892578125, 0.2779299926757812, 0.278066162109375, 0.2780835876464844, 0.2781214599609375, 0.27794125366210937, 0.2781419372558594, 0.2781163635253906, 0.2781460571289063, 0.27796786499023435, 0.27823822021484373, 0.27809381103515624, 0.27807949829101564, 0.2781194152832031, 0.5966243896484374, 0.2779586486816406, 0.27771597290039063, 0.2778900451660156, 0.2777845764160156, 0.27794842529296876, 0.27777432250976564, 0.27786138916015624, 0.2779832458496094, 0.27784295654296876, 0.27791360473632815, 0.2778265686035156, 0.27825048828125, 0.2778839111328125, 0.277760009765625, 0.2779811706542969, 0.2778316650390625, 0.277928955078125, 0.2776801147460938, 0.27795147705078127, 0.27797503662109374, 0.27823822021484373, 0.2777794494628906, 0.2780354614257812, 0.2778675231933594, 0.278034423828125, 0.27775796508789063, 0.2779596862792969, 0.2779453430175781, 0.27806515502929685, 0.27786444091796875, 0.2780897216796875, 0.2780426330566406, 0.27793508911132814, 0.2780201110839844, 0.2778951721191406, 0.2779228210449219, 0.27799655151367186, 0.27801089477539065, 0.277960693359375, 0.27807745361328123, 0.2780201110839844, 0.2779279479980469, 0.2779719543457031, 0.2779791259765625, 0.2780078125, 0.27794329833984377, 0.27796890258789064, 0.27801190185546876, 0.27812045288085935, 0.2779719543457031, 0.27803237915039064, 0.2780354614257812, 0.27804672241210937, 0.27791973876953124, 0.27803955078125, 0.27796990966796875, 0.27832321166992186, 0.2778173522949219, 0.2779924621582031, 0.2780313720703125, 0.2779924621582031, 0.2779095153808594, 0.5980723266601562, 0.27799755859375, 0.2777897033691406, 0.2778787841796875, 0.2782607421875, 0.27835186767578124, 0.27776409912109373, 0.27780609130859374, 0.277928955078125, 0.2779156494140625, 0.27778866577148437, 0.2779658203125, 0.2778961791992188, 0.2779033508300781, 0.2778521728515625, 0.27789208984375, 0.2778408813476563, 0.2780610656738281, 0.27775180053710935, 0.277928955078125, 0.2779955139160156, 0.277886962890625, 0.27779379272460936, 0.2780897216796875, 0.27797402954101563, 0.27802215576171874, 0.277818359375, 0.2780201110839844, 0.27800473022460936, 0.27827508544921875, 0.27791256713867185, 0.2779934692382813, 0.278012939453125, 0.2779361267089844, 0.27817880249023436, 0.2780354614257812, 0.2780149841308594, 0.27794329833984377, 0.27789825439453125, 0.2780262451171875, 0.27804568481445313, 0.27850650024414064, 0.27791873168945314, 0.2781460571289063, 0.27798220825195313, 0.27811532592773436, 0.27813784790039064, 0.27810406494140627, 0.2780354614257812, 0.2781583251953125, 0.27791256713867185, 0.27816754150390627, 0.2781614074707031, 0.278108154296875, 0.2779473876953125, 0.27812045288085935, 0.27804364013671873, 0.278076416015625, 0.27797503662109374, 0.2780887145996094, 0.2781245422363281, 0.2781767578125, 0.27807232666015624, 0.5989816284179688, 0.277907470703125, 0.27786956787109374, 0.27786343383789064, 0.2780149841308594, 0.27786444091796875, 0.2781245422363281, 0.27784396362304686, 0.2778961791992188, 0.2778951721191406, 0.2777282409667969, 0.27791256713867185, 0.27789208984375, 0.2780354614257812, 0.2777200622558594, 0.2778880004882813, 0.27792691040039064, 0.2780928039550781, 0.2777774047851562, 0.2778787841796875, 0.27798220825195313, 0.2778808288574219, 0.2777569274902344, 0.27796273803710936, 0.2778173522949219, 0.2779586486816406, 0.2777763977050781, 0.27791156005859374, 0.27785830688476565, 0.2779729919433594, 0.2778050537109375, 0.2779719543457031, 0.2780149841308594, 0.27809381103515624, 0.27781427001953124, 0.27789108276367186, 0.277970947265625, 0.2780948486328125, 0.2778675231933594, 0.2779228210449219, 0.27799655151367186, 0.27811224365234377, 0.27821261596679686, 0.27802520751953125, 0.27791461181640625, 0.2779893798828125, 0.2780262756347656, 0.27805181884765623, 0.27803851318359374, 0.27812762451171874, 0.2778623962402344, 0.27810919189453126, 0.27789208984375, 0.27803237915039064, 0.27796786499023435, 0.2779924621582031, 0.27802420043945314, 0.2780498046875, 0.2778736572265625, 0.27802725219726565, 0.2780917663574219, 0.27807437133789065, 0.27804779052734374, 0.5996082763671875, 0.2778603515625, 0.27799038696289063, 0.2779002990722656, 0.2780784606933594, 0.2780487670898438, 0.27779071044921877, 0.27787060546875, 0.27787161254882814, 0.277907470703125, 0.27785830688476565, 0.2779095153808594, 0.2779945068359375, 0.27801190185546876, 0.27792178344726565, 0.2780282897949219, 0.2778726501464844, 0.27789926147460936, 0.27795147705078127, 0.27812249755859375, 0.2779299926757812, 0.2781460571289063, 0.2777681884765625, 0.27811224365234377, 0.2779893798828125, 0.2780211181640625, 0.27784396362304686, 0.27811737060546876, 0.2779924621582031, 0.2780702819824219, 0.27788595581054687, 0.27822796630859375, 0.2781614074707031, 0.27828121948242185, 0.27786444091796875, 0.27808767700195314, 0.2780159912109375, 0.2779842529296875, 0.27789312744140626, 0.27807745361328123, 0.27812863159179685, 0.27812045288085935, 0.277923828125, 0.2781317138671875, 0.2779053955078125, 0.27808050537109374, 0.2779934692382813, 0.2780682373046875, 0.27818801879882815, 0.278139892578125, 0.27804568481445313, 0.27794943237304687, 0.27806719970703125, 0.27802520751953125, 0.2779033508300781, 0.27811224365234377, 0.278108154296875, 0.2783221740722656, 0.2778388366699219, 0.27817166137695315, 0.2779535217285156, 0.27817779541015625, 0.2781368408203125, 0.5995950317382812, 0.27793508911132814, 0.2778111877441406, 0.27782962036132813, 0.277923828125, 0.2780057678222656, 0.2776893310546875, 0.27791873168945314, 0.2778787841796875, 0.27778253173828127, 0.2776565856933594, 0.27797503662109374, 0.27791668701171873, 0.27793408203125, 0.277823486328125, 0.27784909057617185, 0.2778531799316406, 0.278044677734375, 0.27778662109375, 0.278371337890625, 0.27801803588867186, 0.27795559692382815, 0.27778765869140626, 0.27804672241210937, 0.2778900451660156, 0.27810202026367187, 0.27790643310546875, 0.2779443054199219, 0.2779279479980469, 0.2781368408203125, 0.27793408203125, 0.2779914245605469, 0.2780159912109375, 0.27801702880859375, 0.2779934692382813, 0.2779084777832031, 0.2779371643066406, 0.27799038696289063, 0.2778685302734375, 0.2779504699707031, 0.2779842529296875, 0.2779535217285156, 0.27789926147460936, 0.2780508117675781, 0.2779299926757812, 0.2779852905273438, 0.27796786499023435, 0.2780845947265625, 0.278012939453125, 0.27808563232421873, 0.27792486572265623, 0.2780528564453125, 0.2779791259765625, 0.27808767700195314, 0.27787982177734377, 0.2782576599121094, 0.2780477294921875, 0.27795660400390626, 0.2779730224609375, 0.27805487060546874, 0.27805184936523436, 0.27802725219726565, 0.2779259338378906]",tokens/s,3.538951951648072,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3961,7 +3961,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4160,7 +4160,7 @@ OSError: M is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4580,7 +4580,7 @@ Traceback (most recent call last): OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4709,7 +4709,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,6439.493632,20389.036032,0.0,19742.588928,18821.508608,s,10,15.1473671875,1.5147367187499998,0.00247756470747785,1.5140679931640624,1.5185943847656251,1.5187318237304688,1.5188417749023437,"[1.5126702880859375, 1.5122486572265625, 1.5120849609375, 1.5139268798828125, 1.5142091064453125, 1.5128245849609374, 1.514491455078125, 1.5174781494140626, 1.5188692626953124, 1.5185638427734376]",tokens/s,169.00626810661714,kWh,1.786537355846829e-05,9.788986818782104e-06,8.822982058380724e-05,0.00011588418096105764,tokens/kWh,2209102.2077122647,MB,6440.05888,20389.036032,0.0,19742.588928,19408.2304,s,10,843.181140625,84.3181140625,0.002181345135820369,84.31812890625,84.32165,84.32172734375,84.32178921875,"[84.3218046875, 84.31809375, 84.3141640625, 84.317453125, 84.316203125, 84.31825, 84.3185546875, 84.3168203125, 84.3181640625, 84.3216328125]",tokens/s,0.7471704117255499,kWh,0.0009954405118525029,0.0005455906297619111,0.00497354175660798,0.006514572898222393,tokens/kWh,9670.626299567632,,s,629,855.4723406982422,1.360051416054439,0.1800573065911544,1.3383363037109375,1.3387327392578126,1.338873681640625,2.85288818359375,"[1.33827587890625, 1.338081298828125, 1.3381846923828125, 1.3384754638671874, 1.33840283203125, 1.337996337890625, 1.338251220703125, 1.33793896484375, 1.3378477783203124, 1.338167236328125, 1.337945068359375, 1.338113037109375, 1.3379625244140625, 1.3380064697265626, 1.3384990234375, 1.3382308349609375, 1.33815087890625, 1.3384324951171875, 1.3385635986328126, 1.3381785888671875, 1.3384417724609374, 1.338292236328125, 1.33815087890625, 1.3382215576171874, 1.33812841796875, 1.3385369873046875, 1.33857080078125, 1.33825537109375, 1.338660888671875, 1.338504150390625, 1.3385400390625, 1.338418212890625, 1.338197998046875, 1.339029541015625, 1.3383699951171875, 1.338683349609375, 1.3382236328125, 1.338335205078125, 1.33846728515625, 1.33833935546875, 1.3388974609375, 1.3384970703125, 1.3387171630859376, 1.33850927734375, 1.3386947021484374, 1.33789794921875, 1.3382584228515626, 1.3383731689453124, 1.3386966552734374, 1.338715087890625, 1.3383782958984376, 1.3388123779296874, 1.3381427001953126, 1.338365966796875, 1.338388427734375, 1.338744873046875, 1.338123291015625, 1.3381826171875, 1.3381212158203124, 1.338502197265625, 1.3384345703125, 1.338925048828125, 2.855078857421875, 1.3382308349609375, 1.3380341796875, 1.338641357421875, 1.3384100341796874, 1.338455078125, 1.338123291015625, 1.33812841796875, 1.338135498046875, 1.3380413818359376, 1.3386229248046875, 1.338265625, 1.3381959228515625, 1.338167236328125, 1.338260498046875, 1.3382215576171874, 1.33853076171875, 1.3382952880859376, 1.3382645263671875, 1.3385697021484375, 1.3381181640625, 1.337927734375, 1.3383004150390625, 1.3381468505859375, 1.33812841796875, 1.33810693359375, 1.33798095703125, 1.3387745361328125, 1.3384786376953124, 1.3384632568359376, 1.3380546875, 1.33858203125, 1.3382686767578125, 1.33855126953125, 1.338820556640625, 1.3380628662109375, 1.33846728515625, 1.3384324951171875, 1.3379993896484375, 1.3383935546875, 1.3383526611328125, 1.33831787109375, 1.3384007568359375, 1.338850341796875, 1.338428466796875, 1.3388790283203125, 1.3380782470703125, 1.3385390625, 1.33848681640625, 1.3382850341796875, 1.338239990234375, 1.338397705078125, 1.338281982421875, 1.3381427001953126, 1.338702880859375, 1.338449951171875, 1.3385543212890625, 1.3380904541015626, 1.3382625732421876, 1.338514404296875, 1.33867822265625, 1.338439697265625, 1.3386260986328125, 2.852787109375, 1.337776123046875, 1.3380433349609375, 1.33821240234375, 1.3381181640625, 1.3380423583984375, 1.338502197265625, 1.337923583984375, 1.3380556640625, 1.337927734375, 1.3382747802734376, 1.3379788818359375, 1.3380535888671874, 1.3380679931640624, 1.338323974609375, 1.33837109375, 1.3381468505859375, 1.3381048583984374, 1.3382010498046875, 1.338040283203125, 1.338312744140625, 1.3381253662109376, 1.338639404296875, 1.3382962646484375, 1.3380535888671874, 1.3382327880859375, 1.33817041015625, 1.338787841796875, 1.3383106689453126, 1.3387684326171876, 1.338207275390625, 1.338524658203125, 1.3381622314453125, 1.3384385986328124, 1.3382880859375, 1.3383096923828126, 1.3381539306640624, 1.3384376220703125, 1.3379154052734374, 1.3386424560546875, 1.338576904296875, 1.338471435546875, 1.3384324951171875, 1.33886767578125, 1.3384775390625, 1.3387581787109375, 1.33810693359375, 1.3385318603515626, 1.3380946044921875, 1.338239013671875, 1.338491943359375, 1.338514404296875, 1.3388123779296874, 1.338328125, 1.33863525390625, 1.3383731689453124, 1.3381826171875, 1.3379378662109376, 1.3384427490234374, 1.3381744384765626, 1.3386107177734374, 1.3386884765625, 1.3387366943359376, 2.852927490234375, 1.3382041015625, 1.33804443359375, 1.3380986328125, 1.3379327392578124, 1.33829736328125, 1.3382789306640626, 1.33785595703125, 1.3384140625, 1.338491943359375, 1.338461181640625, 1.3379215087890626, 1.3379921875, 1.3384775390625, 1.338349609375, 1.3382891845703124, 1.3381856689453124, 1.338224609375, 1.3381641845703125, 1.3382738037109374, 1.3382677001953125, 1.3385400390625, 1.3384990234375, 1.338345458984375, 1.3385574951171875, 1.3380638427734375, 1.3382021484375, 1.3386884765625, 1.3383792724609376, 1.3390386962890626, 1.3381170654296874, 1.3385206298828125, 1.3384171142578125, 1.338330078125, 1.3384376220703125, 1.338135498046875, 1.3385645751953126, 1.3382369384765624, 1.3381488037109375, 1.3385615234375, 1.338334228515625, 1.3385687255859375, 1.3384263916015624, 1.338876953125, 1.3385943603515624, 1.3389271240234375, 1.3382574462890624, 1.3385318603515626, 1.338039306640625, 1.33815087890625, 1.338440673828125, 1.338408935546875, 1.3386680908203126, 1.3384068603515624, 1.3384693603515625, 1.338323974609375, 1.3382747802734376, 1.3380362548828124, 1.3383433837890626, 1.3387017822265626, 1.338533935546875, 1.3384949951171874, 1.3386751708984375, 2.853210205078125, 1.33809765625, 1.3379481201171874, 1.33806591796875, 1.338060791015625, 1.3381856689453124, 1.338239013671875, 1.33782421875, 1.338302490234375, 1.3379266357421875, 1.338209228515625, 1.3380413818359376, 1.3381734619140624, 1.3381641845703125, 1.3383076171875, 1.338155029296875, 1.3382471923828125, 1.338187744140625, 1.338287109375, 1.3382471923828125, 1.338228759765625, 1.33800244140625, 1.3384847412109375, 1.3383424072265624, 1.3381273193359375, 1.3382645263671875, 1.3383167724609375, 1.3388646240234374, 1.3382379150390624, 1.338666015625, 1.3386270751953124, 1.33897314453125, 1.3383065185546874, 1.3384427490234374, 1.338640380859375, 1.33823388671875, 1.338455078125, 1.3382747802734376, 1.33837109375, 1.338439697265625, 1.3384366455078125, 1.3383905029296874, 1.338646484375, 1.33884521484375, 1.338330078125, 1.3389864501953126, 1.3383802490234376, 1.338328125, 1.33821337890625, 1.3384385986328124, 1.3382369384765624, 1.33838232421875, 1.338428466796875, 1.338207275390625, 1.3385430908203124, 1.338228759765625, 1.3381375732421874, 1.3383997802734375, 1.33844580078125, 1.338328125, 1.3384949951171874, 1.3387335205078126, 1.3386895751953125, 2.851811279296875, 1.3381427001953126, 1.338029052734375, 1.3379471435546875, 1.3380321044921875, 1.3382861328125, 1.3381212158203124, 1.337881591796875, 1.3381683349609375, 1.3383577880859374, 1.33817041015625, 1.338059814453125, 1.338250244140625, 1.33771875, 1.33808642578125, 1.33836181640625, 1.3384345703125, 1.3382471923828125, 1.33817138671875, 1.338324951171875, 1.3382225341796874, 1.3380731201171876, 1.3383475341796875, 1.3384796142578126, 1.338498046875, 1.338177490234375, 1.338260498046875, 1.3385491943359376, 1.33831982421875, 1.338625, 1.33840283203125, 1.338828857421875, 1.33834033203125, 1.3385369873046875, 1.338491943359375, 1.3382593994140626, 1.3384049072265625, 1.3383792724609376, 1.33875, 1.33888720703125, 1.3388912353515625, 1.3387335205078126, 1.3385902099609375, 1.3390325927734374, 1.33831884765625, 1.3388984375, 1.3380587158203125, 1.338312744140625, 1.338059814453125, 1.3385400390625, 1.3383577880859374, 1.3386107177734374, 1.338830810546875, 1.3383013916015625, 1.3383731689453124, 1.3382635498046875, 1.3384385986328124, 1.33800341796875, 1.33859326171875, 1.3383843994140625, 1.3386947021484374, 1.3384888916015625, 1.3389117431640625, 2.853399658203125, 1.3380997314453125, 1.3381314697265625, 1.3383485107421875, 1.3379093017578125, 1.3382696533203124, 1.33813037109375, 1.338008544921875, 1.3382789306640626, 1.3383157958984375, 1.33842431640625, 1.3384100341796874, 1.3381949462890625, 1.3380885009765624, 1.3384017333984375, 1.3384765625, 1.3382010498046875, 1.3380526123046874, 1.3383218994140624, 1.3381622314453125, 1.338207275390625, 1.3384222412109374, 1.3384591064453124, 1.3383270263671876, 1.3385625, 1.337927734375, 1.338203125, 1.3387591552734375, 1.33834033203125, 1.338640380859375, 1.3384735107421875, 1.338603515625, 1.3385892333984375, 1.338545166015625, 1.3386025390625, 1.3384468994140626, 1.3385206298828125, 1.338261474609375, 1.338314697265625, 1.3385943603515624, 1.33827783203125, 1.3385491943359376, 1.3389219970703126, 1.3390673828125, 1.3380843505859374, 1.33874169921875, 1.338102783203125, 1.3382635498046875, 1.33827685546875, 1.338281982421875, 1.3385645751953126, 1.33852978515625, 1.3383935546875, 1.3382266845703126, 1.33874169921875, 1.3383935546875, 1.3385174560546875, 1.3384385986328124, 1.3383218994140624, 1.3382532958984374, 1.338398681640625, 1.338607666015625, 1.3387099609375, 2.856235107421875, 1.3382266845703126, 1.3379635009765625, 1.338997802734375, 1.3381160888671875, 1.33855126953125, 1.3383414306640624, 1.33804541015625, 1.3381396484375, 1.3380484619140625, 1.3383055419921874, 1.338119140625, 1.3381806640625, 1.337987060546875, 1.3384642333984376, 1.3384119873046876, 1.338092529296875, 1.338167236328125, 1.3384990234375, 1.3382625732421876, 1.3382000732421875, 1.3384959716796876, 1.33859326171875, 1.3382962646484375, 1.3382880859375, 1.338103759765625, 1.338419189453125, 1.3385390625, 1.338028076171875, 1.3388687744140626, 1.3383526611328125, 1.3386536865234375, 1.338312744140625, 1.33836181640625, 1.3386044921875, 1.3382686767578125, 1.338302490234375, 1.3381314697265625, 1.3380546875, 1.33846728515625, 1.3385062255859375, 1.3385830078125, 1.3386895751953125, 1.338883056640625, 1.3383526611328125, 1.33880419921875, 1.3380731201171876, 1.338312744140625, 1.338076171875, 1.3381683349609375, 1.338346435546875, 1.338471435546875, 1.338356689453125, 1.3382635498046875, 1.338735595703125, 1.338377197265625, 1.338420166015625, 1.338218505859375, 1.3382369384765624, 1.33845703125, 1.338354736328125, 1.3386639404296874, 1.3384560546875, 2.8577626953125, 1.3383587646484374, 1.3380689697265624, 1.3380474853515625, 1.3381632080078125, 1.3382593994140626, 1.338166259765625, 1.338167236328125, 1.3380833740234375, 1.3378006591796876, 1.3386485595703126, 1.3379481201171874, 1.3380731201171876, 1.338176513671875, 1.3382532958984374, 1.3385789794921874, 1.3384898681640625, 1.338156005859375, 1.3380740966796876, 1.3381806640625, 1.3379296875, 1.3382943115234376, 1.33830859375, 1.338144775390625, 1.3381417236328126, 1.338250244140625, 1.3382861328125, 1.3385697021484375, 1.3382645263671875, 1.3386312255859374, 1.3383526611328125, 1.338724365234375, 1.33850732421875, 1.3385440673828124, 1.3388011474609376, 1.3382943115234376, 1.3387642822265624, 1.3382420654296876, 1.33886669921875, 1.338575927734375, 1.3384212646484375, 1.3385267333984374, 1.33864453125, 1.3386854248046876, 1.33810888671875, 1.3389066162109375, 1.3379942626953125, 1.33859228515625, 1.338541015625, 1.3379840087890624, 1.3384171142578125, 1.33873046875, 1.3383382568359374, 1.338566650390625, 1.33863427734375, 1.3383363037109375, 1.3384110107421876, 1.33809765625, 1.33823388671875, 1.3384151611328126, 1.338819580078125, 1.338630126953125, 1.33911962890625, 2.857397216796875, 1.338418212890625, 1.3382236328125, 1.3382666015625, 1.3380423583984375, 1.338387451171875, 1.338007568359375, 1.33800341796875, 1.337934814453125, 1.3378876953125, 1.3382880859375, 1.338082275390625, 1.3380628662109375, 1.33802392578125, 1.3383843994140625, 1.3384273681640626, 1.3381641845703125, 1.3382635498046875, 1.3379410400390626, 1.338555419921875, 1.3382379150390624, 1.3383065185546874, 1.33861279296875, 1.3383004150390625, 1.3380045166015626, 1.3382584228515626, 1.338186767578125, 1.3387325439453126, 1.33827685546875, 1.338945556640625, 1.3384222412109374, 1.3384488525390625, 1.338629150390625, 1.3385841064453126, 1.3385635986328126, 1.3383577880859374, 1.3383157958984375, 1.3384683837890625, 1.338165283203125, 1.3385482177734376, 1.338397705078125, 1.3388319091796874, 1.338493896484375, 1.338962890625, 1.338429443359375, 1.3388216552734375, 1.338155029296875, 1.3382799072265625, 1.338113037109375, 1.338430419921875, 1.3383331298828125, 1.338303466796875, 1.3383363037109375, 1.33825537109375, 1.338997802734375, 1.338502197265625, 1.3380904541015626, 1.3381529541015624, 1.3384119873046876, 1.33829736328125, 1.33835986328125, 1.3386004638671876, 1.3387745361328125]",tokens/s,0.7352663202256265,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4837,7 +4837,7 @@ OSError: google/recurrentgemma-7b is not a local folder and is not a valid model If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5182,7 +5182,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5285,7 +5285,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 96.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5587,7 +5587,7 @@ OSError: 0 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5846,7 +5846,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1451.06944,2426.929152,0.0,1780.482048,1621.410304,s,10,0.7041353912353515,0.07041353912353515,0.0002361853075699059,0.0705183219909668,0.07067396697998048,0.0706822479248047,0.07068887268066407,"[0.07069052886962891, 0.07028050994873047, 0.07060643005371094, 0.0705650863647461, 0.07003734588623046, 0.07016966247558594, 0.07007705688476562, 0.07050838470458984, 0.07052825927734375, 0.07067212677001954]",tokens/s,3635.664435938487,kWh,8.278217291128015e-07,4.5360747945568833e-07,3.6650922871604466e-06,4.946521495728937e-06,tokens/kWh,51753540.38611631,MB,1451.06944,2426.929152,0.0,1780.482048,1723.272192,s,10,38.0147158203125,3.80147158203125,0.0004616553351782168,3.8014595947265626,3.8020025390625,3.8021644775390624,3.8022940283203126,"[3.802326416015625, 3.801103271484375, 3.801966552734375, 3.8013779296875, 3.801707275390625, 3.801541259765625, 3.800688232421875, 3.801736328125, 3.800993896484375, 3.801274658203125]",tokens/s,16.572529516671292,kWh,4.488269805491387e-05,2.4598166829557432e-05,0.0001986788473656444,0.00026815971225011573,tokens/kWh,234934.61963905732,,s,629,38.587294712066715,0.061347050416640146,0.008392685763651334,0.06032691192626953,0.06039490661621094,0.06042992630004883,0.13075378845214844,"[0.060313568115234376, 0.06032076644897461, 0.060349441528320315, 0.060303359985351565, 0.06035148620605469, 0.0603422737121582, 0.06035763168334961, 0.06037299346923828, 0.06039756774902344, 0.06034022521972656, 0.060295169830322265, 0.06041292953491211, 0.06027775955200195, 0.06031052780151367, 0.06032793426513672, 0.060314624786376954, 0.06030131149291992, 0.06029619216918945, 0.06031052780151367, 0.060375038146972655, 0.06031769561767578, 0.060290046691894535, 0.06032998275756836, 0.0604139518737793, 0.060281856536865235, 0.06031155014038086, 0.060355583190917966, 0.06032896041870117, 0.060336128234863284, 0.06032896041870117, 0.06035456085205078, 0.06030233764648438, 0.06037811279296875, 0.06033715057373047, 0.060319744110107425, 0.06034636688232422, 0.06032588958740234, 0.060276737213134764, 0.060368896484375, 0.060331008911132813, 0.06034636688232422, 0.06036684799194336, 0.06032691192626953, 0.06030950546264648, 0.06034636688232422, 0.06035148620605469, 0.060316673278808595, 0.060295169830322265, 0.06031257629394531, 0.06030847930908203, 0.06034841537475586, 0.06037299346923828, 0.06039039993286133, 0.06031564712524414, 0.06034431838989258, 0.06038016128540039, 0.06031052780151367, 0.06034022521972656, 0.06030745697021484, 0.06039039993286133, 0.06033407974243164, 0.06071091079711914, 0.13086003112792968, 0.06027468872070312, 0.060333057403564455, 0.06030950546264648, 0.06031155014038086, 0.06030950546264648, 0.06030847930908203, 0.060276737213134764, 0.06025830459594726, 0.06025932693481445, 0.06027775955200195, 0.060268543243408204, 0.06028083038330078, 0.06025625610351563, 0.060283905029296876, 0.06025625610351563, 0.06033203125, 0.06031052780151367, 0.060375038146972655, 0.06030438232421875, 0.06034329605102539, 0.06034534454345703, 0.060276737213134764, 0.06028799819946289, 0.06030233764648438, 0.0603054084777832, 0.060316673278808595, 0.060298240661621094, 0.06029619216918945, 0.06029414367675781, 0.06036479949951172, 0.06043852615356445, 0.06035763168334961, 0.06036172866821289, 0.060450817108154295, 0.06034841537475586, 0.060365825653076174, 0.060388351440429686, 0.06029414367675781, 0.06033919906616211, 0.06032691192626953, 0.0603422737121582, 0.06036684799194336, 0.06033407974243164, 0.06032588958740234, 0.06034636688232422, 0.06038323211669922, 0.060391422271728515, 0.06032486343383789, 0.0603054084777832, 0.060316673278808595, 0.06037401580810547, 0.060386302947998044, 0.060382209777832034, 0.060336128234863284, 0.060437503814697265, 0.06042521667480469, 0.06031769561767578, 0.060365825653076174, 0.0603054084777832, 0.060407806396484375, 0.06031564712524414, 0.06038937759399414, 0.13136691284179688, 0.06031769561767578, 0.06033715057373047, 0.06035968017578125, 0.06034022521972656, 0.06029107284545898, 0.060286975860595705, 0.06029107284545898, 0.06029414367675781, 0.06024499130249023, 0.060314624786376954, 0.060262401580810546, 0.06032281494140625, 0.06025625610351563, 0.06031155014038086, 0.06026649475097656, 0.06032486343383789, 0.06039756774902344, 0.06035968017578125, 0.06030233764648438, 0.06035456085205078, 0.060295169830322265, 0.06032998275756836, 0.06029926300048828, 0.0603135986328125, 0.06033817672729492, 0.06035456085205078, 0.0603873291015625, 0.060352512359619144, 0.06028595352172852, 0.06034022521972656, 0.06030233764648438, 0.060442623138427735, 0.060336128234863284, 0.060402687072753904, 0.060331008911132813, 0.060426239013671876, 0.06041088104248047, 0.06029619216918945, 0.06032281494140625, 0.06037913513183594, 0.06038937759399414, 0.06035763168334961, 0.060439552307128906, 0.0603135986328125, 0.06057062530517578, 0.06040883255004883, 0.06034124755859375, 0.06027468872070312, 0.060290046691894535, 0.060369918823242184, 0.060333057403564455, 0.06034739303588867, 0.06037401580810547, 0.0603351058959961, 0.06033203125, 0.060369918823242184, 0.06029926300048828, 0.060440574645996094, 0.06032691192626953, 0.06042009735107422, 0.06032998275756836, 0.06038937759399414, 0.13121229553222657, 0.06032588958740234, 0.060352512359619144, 0.06033407974243164, 0.06028595352172852, 0.06032179260253906, 0.060295169830322265, 0.06027980804443359, 0.060278785705566405, 0.06031052780151367, 0.06031769561767578, 0.06029414367675781, 0.06030438232421875, 0.06028595352172852, 0.06030847930908203, 0.06031564712524414, 0.060303359985351565, 0.06029312133789062, 0.06030950546264648, 0.060303359985351565, 0.06031257629394531, 0.060319744110107425, 0.06031872177124024, 0.06030847930908203, 0.06031155014038086, 0.06036070251464844, 0.06029926300048828, 0.06032588958740234, 0.0603054084777832, 0.06032691192626953, 0.060331008911132813, 0.060352512359619144, 0.06030643081665039, 0.06039654541015625, 0.06035865783691406, 0.06033817672729492, 0.06036377716064453, 0.06035148620605469, 0.06028287887573242, 0.06038425445556641, 0.06041497421264649, 0.06032179260253906, 0.06035865783691406, 0.06032998275756836, 0.06031257629394531, 0.06040678405761719, 0.06037606430053711, 0.06036684799194336, 0.060284927368164064, 0.06030847930908203, 0.06038425445556641, 0.06036377716064453, 0.06035763168334961, 0.06036377716064453, 0.06035763168334961, 0.06036070251464844, 0.06041292953491211, 0.06029926300048828, 0.0603873291015625, 0.06032691192626953, 0.060437503814697265, 0.06032896041870117, 0.06042009735107422, 0.1306306610107422, 0.060273662567138675, 0.060333057403564455, 0.06033919906616211, 0.06030131149291992, 0.06034636688232422, 0.06030131149291992, 0.060303359985351565, 0.06026444625854492, 0.06025932693481445, 0.060267520904541017, 0.060268543243408204, 0.06032486343383789, 0.06025625610351563, 0.060273662567138675, 0.06026649475097656, 0.060267520904541017, 0.060298240661621094, 0.06034124755859375, 0.06027775955200195, 0.06034124755859375, 0.0603054084777832, 0.06027468872070312, 0.060316673278808595, 0.06037094497680664, 0.06032486343383789, 0.06032896041870117, 0.060286975860595705, 0.06031564712524414, 0.06034534454345703, 0.06036684799194336, 0.06033817672729492, 0.06033919906616211, 0.060391422271728515, 0.06032998275756836, 0.06034124755859375, 0.06038016128540039, 0.06036377716064453, 0.06034431838989258, 0.06035968017578125, 0.06033919906616211, 0.06034329605102539, 0.060375038146972655, 0.06034124755859375, 0.060728321075439455, 0.060353534698486325, 0.06036377716064453, 0.06032179260253906, 0.06030643081665039, 0.06028902435302735, 0.06030847930908203, 0.06033817672729492, 0.060355583190917966, 0.0603422737121582, 0.060846080780029295, 0.0603770866394043, 0.06039756774902344, 0.06029414367675781, 0.06039039993286133, 0.06033407974243164, 0.060375038146972655, 0.06032588958740234, 0.06038118362426758, 0.13078016662597655, 0.060283905029296876, 0.06032179260253906, 0.060455936431884766, 0.06029312133789062, 0.060260353088378904, 0.06032588958740234, 0.06031564712524414, 0.060283905029296876, 0.060262401580810546, 0.06040678405761719, 0.06024806213378906, 0.06032691192626953, 0.06025830459594726, 0.06031155014038086, 0.060240894317626956, 0.060295169830322265, 0.06030643081665039, 0.06055936050415039, 0.06036479949951172, 0.06031257629394531, 0.06029107284545898, 0.0603351058959961, 0.0603135986328125, 0.060286975860595705, 0.06032076644897461, 0.0603351058959961, 0.06029312133789062, 0.06033203125, 0.06028799819946289, 0.06037094497680664, 0.06072115325927734, 0.060316673278808595, 0.06034739303588867, 0.06034636688232422, 0.0603422737121582, 0.06034431838989258, 0.06031155014038086, 0.060303359985351565, 0.06033407974243164, 0.06033407974243164, 0.06032896041870117, 0.06034022521972656, 0.060371967315673826, 0.06029619216918945, 0.06029209518432617, 0.06038118362426758, 0.060355583190917966, 0.060298240661621094, 0.060295169830322265, 0.06033407974243164, 0.06033203125, 0.06036377716064453, 0.06036787033081055, 0.06041497421264649, 0.060402687072753904, 0.06039244842529297, 0.060268543243408204, 0.06036172866821289, 0.06028799819946289, 0.060432384490966794, 0.06033407974243164, 0.06041292953491211, 0.1306859588623047, 0.06027775955200195, 0.06031155014038086, 0.060355583190917966, 0.06030131149291992, 0.06034739303588867, 0.06032486343383789, 0.06030950546264648, 0.06028595352172852, 0.0603054084777832, 0.06030131149291992, 0.060290046691894535, 0.0603422737121582, 0.06029312133789062, 0.06030847930908203, 0.06029312133789062, 0.06029209518432617, 0.060284927368164064, 0.0603054084777832, 0.06030847930908203, 0.06030950546264648, 0.06032998275756836, 0.06029721450805664, 0.06030950546264648, 0.060270591735839846, 0.06033919906616211, 0.06032998275756836, 0.06032076644897461, 0.06032179260253906, 0.060278785705566405, 0.06030847930908203, 0.06032793426513672, 0.06032998275756836, 0.060382209777832034, 0.06034124755859375, 0.060316673278808595, 0.060333057403564455, 0.0603422737121582, 0.06031872177124024, 0.060355583190917966, 0.06029107284545898, 0.060319744110107425, 0.06037811279296875, 0.06033203125, 0.06032691192626953, 0.06035148620605469, 0.060353534698486325, 0.0603351058959961, 0.06031257629394531, 0.06040063858032227, 0.06035456085205078, 0.06031564712524414, 0.06030438232421875, 0.06034739303588867, 0.06030233764648438, 0.06033407974243164, 0.06039244842529297, 0.06027980804443359, 0.06036377716064453, 0.06030745697021484, 0.060402687072753904, 0.06032691192626953, 0.06038016128540039, 0.13107916259765626, 0.06026444625854492, 0.06031052780151367, 0.06032588958740234, 0.06027571105957031, 0.06025113677978516, 0.060268543243408204, 0.060286975860595705, 0.06029312133789062, 0.06029619216918945, 0.06031155014038086, 0.06028902435302735, 0.060295169830322265, 0.06025932693481445, 0.06028799819946289, 0.06030745697021484, 0.060284927368164064, 0.06031257629394531, 0.060368896484375, 0.060286975860595705, 0.06031052780151367, 0.0603135986328125, 0.06030745697021484, 0.06029926300048828, 0.06032588958740234, 0.060368896484375, 0.06030950546264648, 0.060424190521240234, 0.060401664733886716, 0.060333057403564455, 0.06030950546264648, 0.06031564712524414, 0.06033407974243164, 0.06035456085205078, 0.06039961624145508, 0.06041190338134766, 0.06038016128540039, 0.0603351058959961, 0.060382209777832034, 0.06035660934448242, 0.06034329605102539, 0.06033407974243164, 0.06033407974243164, 0.06030950546264648, 0.06032076644897461, 0.06036479949951172, 0.06040371322631836, 0.06059929656982422, 0.06036275100708008, 0.06029721450805664, 0.060470272064208984, 0.060352512359619144, 0.060375038146972655, 0.060352512359619144, 0.06030131149291992, 0.06033919906616211, 0.060572673797607425, 0.06031769561767578, 0.060349441528320315, 0.06039244842529297, 0.06044979095458984, 0.06032076644897461, 0.0603873291015625, 0.1311068115234375, 0.06029312133789062, 0.06030438232421875, 0.060283905029296876, 0.060257278442382815, 0.06039449691772461, 0.06032486343383789, 0.06034431838989258, 0.06030847930908203, 0.060278785705566405, 0.060375038146972655, 0.06028799819946289, 0.060298240661621094, 0.060303359985351565, 0.060300289154052736, 0.06025932693481445, 0.0603135986328125, 0.06026342391967773, 0.060295169830322265, 0.060369918823242184, 0.06033817672729492, 0.060300289154052736, 0.06032486343383789, 0.06029619216918945, 0.06030745697021484, 0.06029414367675781, 0.060355583190917966, 0.06032793426513672, 0.06035148620605469, 0.060286975860595705, 0.06029721450805664, 0.0603135986328125, 0.06036275100708008, 0.06036070251464844, 0.06037401580810547, 0.06032793426513672, 0.06034124755859375, 0.06037606430053711, 0.060303359985351565, 0.06036070251464844, 0.06030950546264648, 0.06032486343383789, 0.06034329605102539, 0.060319744110107425, 0.06032281494140625, 0.06032793426513672, 0.06038016128540039, 0.06034124755859375, 0.06031872177124024, 0.060316673278808595, 0.06035763168334961, 0.060353534698486325, 0.060355583190917966, 0.060331008911132813, 0.06034739303588867, 0.06035660934448242, 0.06042521667480469, 0.060314624786376954, 0.060398593902587894, 0.06028902435302735, 0.06045183944702148, 0.06032076644897461, 0.06038425445556641, 0.1313116149902344, 0.06028902435302735, 0.06037299346923828, 0.060316673278808595, 0.06025830459594726, 0.060295169830322265, 0.06028799819946289, 0.0603054084777832, 0.06030950546264648, 0.060303359985351565, 0.06030643081665039, 0.06028902435302735, 0.06029107284545898, 0.06029209518432617, 0.06031052780151367, 0.060303359985351565, 0.06029414367675781, 0.06030233764648438, 0.060270591735839846, 0.060273662567138675, 0.060333057403564455, 0.06029926300048828, 0.06030233764648438, 0.06034431838989258, 0.06030131149291992, 0.06032486343383789, 0.06032281494140625, 0.06031564712524414, 0.060319744110107425, 0.06029721450805664, 0.06031564712524414, 0.060295169830322265, 0.060371967315673826, 0.06037299346923828, 0.06035456085205078, 0.060455936431884766, 0.060368896484375, 0.06031564712524414, 0.06030950546264648, 0.06032691192626953, 0.06038118362426758, 0.06035763168334961, 0.06029721450805664, 0.060352512359619144, 0.060388351440429686, 0.06037299346923828, 0.06037606430053711, 0.060453887939453124, 0.06036787033081055, 0.06032691192626953, 0.0603770866394043, 0.06034431838989258, 0.06031257629394531, 0.060352512359619144, 0.060286975860595705, 0.060371967315673826, 0.06049792098999023, 0.06034739303588867, 0.060371967315673826, 0.06029312133789062, 0.06042009735107422, 0.06035763168334961, 0.0603504638671875]",tokens/s,16.300702205052616,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5888,7 +5888,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6163,7 +6163,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1841.045504,9028.763648,0.0,8382.316544,7939.283456,s,10,6.1618914184570315,0.6161891418457032,0.0013886497977201196,0.6156345825195313,0.6173467590332031,0.6185957122802734,0.6195948748779296,"[0.6153511962890625, 0.6198446655273437, 0.6149873657226562, 0.6161400146484375, 0.6167808227539062, 0.6155103149414063, 0.6170692138671875, 0.6149386596679688, 0.6156962280273437, 0.6155729370117188]",tokens/s,415.45685020217974,kWh,7.266781396336026e-06,3.981886734928754e-06,3.646037230553567e-05,4.7709040436800454e-05,tokens/kWh,5365859.335174009,MB,1841.045504,9028.763648,0.0,8382.316544,8233.862144,s,10,338.14244921875,33.814244921875,0.0010794362638467628,33.814513671875005,33.8152484375,33.815645703125,33.815963515625,"[33.8121015625, 33.81604296875, 33.81308203125, 33.81325390625, 33.81455859375, 33.81446875, 33.81516015625, 33.81460546875, 33.8148515625, 33.81432421875]",tokens/s,1.863120118327535,kWh,0.00039921343515316647,0.00021880363846992518,0.0019968464210052606,0.002614863494628352,tokens/kWh,24093.03588100079,,s,629,343.15214111328146,0.5455518936618143,0.0732745798756584,0.5367347412109374,0.5369806762695313,0.5370492797851563,1.1527956982421876,"[0.5368463134765625, 0.5365708618164062, 0.5361151733398437, 0.5366968383789062, 0.5362565307617188, 0.5366865844726563, 0.53654833984375, 0.5366886596679687, 0.5362237548828125, 0.536764404296875, 0.5362780151367188, 0.5366343383789063, 0.5364541625976562, 0.5366149291992187, 0.5362943725585938, 0.5364500732421875, 0.5362340087890625, 0.5366415405273437, 0.536511474609375, 0.5367500610351562, 0.53686376953125, 0.536848388671875, 0.5365933837890625, 0.5368043823242188, 0.5367388305664063, 0.5369036865234375, 0.5367122192382813, 0.5366353759765625, 0.5366087646484375, 0.5367654418945312, 0.536764404296875, 0.5369047241210938, 0.536553466796875, 0.5369548950195312, 0.536637451171875, 0.5367030029296875, 0.5367500610351562, 0.5368914184570313, 0.5368442993164062, 0.5367337036132812, 0.536669189453125, 0.5367767333984375, 0.5367244873046875, 0.53667431640625, 0.5367030029296875, 0.5368524780273437, 0.5369886474609376, 0.5367562255859375, 0.5367418823242187, 0.5367562255859375, 0.5367296142578125, 0.5368053588867188, 0.5372672119140625, 0.536816650390625, 0.5366978759765625, 0.5367675170898437, 0.5367500610351562, 0.5367613525390625, 0.5365473022460937, 0.536869873046875, 0.536685546875, 0.5370091552734375, 1.1570892333984375, 0.5365841674804688, 0.536964111328125, 0.536543212890625, 0.5365770263671875, 0.5371688842773438, 0.5368514404296875, 0.5365718994140625, 0.5365770263671875, 0.536542236328125, 0.5368002319335937, 0.5366016235351563, 0.5365841674804688, 0.5364766845703125, 0.536605712890625, 0.5365811157226562, 0.536616943359375, 0.5363753051757812, 0.536859619140625, 0.5365606689453125, 0.53701123046875, 0.5366098022460938, 0.5368176879882812, 0.5366456298828125, 0.536791015625, 0.5367050170898438, 0.536838134765625, 0.5368258666992187, 0.5367030029296875, 0.5365985107421875, 0.5368995971679688, 0.5368719482421875, 0.5365944213867188, 0.5365964965820312, 0.5368248291015625, 0.5365964965820312, 0.5369528198242187, 0.5366548461914062, 0.5368555297851563, 0.5367551879882813, 0.5366599731445313, 0.5366917114257812, 0.5366702270507813, 0.5368176879882812, 0.5367398681640625, 0.5368494262695312, 0.5369815063476563, 0.5369231567382813, 0.53673779296875, 0.536648681640625, 0.5367817993164062, 0.5367122192382813, 0.5366988525390625, 0.5367613525390625, 0.5369108276367187, 0.536753173828125, 0.5366026000976563, 0.536859619140625, 0.5368678588867187, 0.5366640625, 0.5368094482421875, 0.5369251708984375, 0.5369794311523437, 1.1535380859375, 0.5367613525390625, 0.5365616455078125, 0.5365206909179687, 0.5365166015625, 0.536489990234375, 0.536605712890625, 0.5365125122070312, 0.53650634765625, 0.5366702270507813, 0.5363988647460938, 0.536427490234375, 0.5364992065429688, 0.536585205078125, 0.5364838256835938, 0.5364172973632813, 0.5366087646484375, 0.5363230590820313, 0.5367664794921875, 0.5370265502929688, 0.5367152709960937, 0.5365452880859375, 0.5364879150390625, 0.536574951171875, 0.5367408447265625, 0.5365831909179688, 0.5368402099609375, 0.5366200561523438, 0.53663232421875, 0.5366671142578125, 0.5365862426757813, 0.5366865844726563, 0.5367675170898437, 0.536543212890625, 0.537218017578125, 0.5368289184570313, 0.5368893432617188, 0.5365463256835937, 0.5369180297851562, 0.5366978759765625, 0.5367122192382813, 0.5366190185546875, 0.5367244873046875, 0.5367265014648438, 0.5367787475585938, 0.5369200439453125, 0.536953857421875, 0.5367817993164062, 0.5368955078125, 0.5371924438476563, 0.5367828369140625, 0.536700927734375, 0.5367869262695313, 0.5368555297851563, 0.5369343872070312, 0.5367613525390625, 0.5368616943359374, 0.5369343872070312, 0.5368299560546875, 0.5365944213867188, 0.5369395141601563, 0.5368370971679688, 0.5369733276367188, 1.153333251953125, 0.5366835327148437, 0.5368023071289062, 0.5364889526367187, 0.5365770263671875, 0.5365227661132812, 0.536642578125, 0.5365933837890625, 0.5365155639648438, 0.5364090576171875, 0.5365176391601563, 0.536680419921875, 0.5364449462890625, 0.5364029541015625, 0.5364592895507813, 0.5364039916992187, 0.5365104370117187, 0.536364013671875, 0.536543212890625, 0.5366814575195312, 0.5367203979492188, 0.5366784057617188, 0.53673779296875, 0.5366456298828125, 0.5368237915039062, 0.5368156127929687, 0.5368576049804688, 0.5364859008789062, 0.5366353759765625, 0.5367838745117187, 0.5365975341796875, 0.5367808227539063, 0.5367982177734375, 0.5364725952148437, 0.5369876708984375, 0.5365227661132812, 0.5368074340820312, 0.5367019653320313, 0.5367756958007812, 0.536685546875, 0.5368248291015625, 0.5367306518554688, 0.537006103515625, 0.536700927734375, 0.536791015625, 0.5369712524414062, 0.536995849609375, 0.536574951171875, 0.536764404296875, 0.5368340454101562, 0.5368104858398437, 0.5366251220703125, 0.5369047241210938, 0.5368985595703125, 0.536896484375, 0.53686376953125, 0.5370675048828125, 0.5368463134765625, 0.536816650390625, 0.53661083984375, 0.5369733276367188, 0.5368729858398438, 0.5369825439453125, 1.1542476806640625, 0.5366415405273437, 0.5366640625, 0.5365626831054687, 0.536474609375, 0.536574951171875, 0.5365555419921875, 0.5365391235351562, 0.53659033203125, 0.5364735717773438, 0.5364224243164063, 0.5364786987304687, 0.5366896362304687, 0.5363251342773437, 0.5369405517578125, 0.5366445922851563, 0.5364910278320313, 0.536426513671875, 0.5368678588867187, 0.5365780639648438, 0.536853515625, 0.5366998901367187, 0.5367551879882813, 0.5364920043945313, 0.5366149291992187, 0.53673779296875, 0.5368074340820312, 0.5366876220703125, 0.5368760375976562, 0.5368248291015625, 0.5367367553710938, 0.5367664794921875, 0.5367756958007812, 0.5366128540039062, 0.5368514404296875, 0.5366548461914062, 0.5368995971679688, 0.5366292724609375, 0.5369794311523437, 0.5365975341796875, 0.5368104858398437, 0.5366077270507813, 0.5368811645507813, 0.537017333984375, 0.5368330078125, 0.5368002319335937, 0.5368453369140626, 0.53680126953125, 0.536853515625, 0.5366364135742188, 0.5368237915039062, 0.5368207397460938, 0.5370203857421875, 0.5368842163085937, 0.5367879638671875, 0.536784912109375, 0.5369517822265625, 0.536859619140625, 0.536911865234375, 0.5366599731445313, 0.5369661254882813, 0.5370623779296875, 0.536890380859375, 1.152932861328125, 0.5365995483398438, 0.5366241455078125, 0.5368361206054687, 0.5364029541015625, 0.5365759887695313, 0.5366968383789062, 0.5365770263671875, 0.5366578979492187, 0.5363333129882812, 0.536784912109375, 0.5367633666992188, 0.5364306030273438, 0.5363978271484375, 0.5364786987304687, 0.53639990234375, 0.53650634765625, 0.5363681030273437, 0.5364818115234375, 0.5364930419921875, 0.5367562255859375, 0.5365759887695313, 0.5365831909179688, 0.5367992553710937, 0.5366712036132812, 0.5368299560546875, 0.5367675170898437, 0.5364879150390625, 0.5370848999023438, 0.5367859497070312, 0.5367285766601563, 0.536700927734375, 0.5369815063476563, 0.5364613037109375, 0.5366876220703125, 0.5365463256835937, 0.5369886474609376, 0.5367203979492188, 0.5369845581054687, 0.536690673828125, 0.5368616943359374, 0.5368944702148437, 0.536853515625, 0.5368790893554688, 0.5370951538085937, 0.5369600219726562, 0.5369047241210938, 0.53686474609375, 0.5367398681640625, 0.5367030029296875, 0.5368043823242188, 0.5367684936523438, 0.536774658203125, 0.536958984375, 0.5369774169921875, 0.5366917114257812, 0.5367500610351562, 0.536859619140625, 0.536932373046875, 0.5367869262695313, 0.5369671630859375, 0.5367613525390625, 0.537069580078125, 1.154186279296875, 0.5370890502929687, 0.5366292724609375, 0.536573974609375, 0.5365975341796875, 0.5366190185546875, 0.536500244140625, 0.5365022583007812, 0.53646337890625, 0.5364029541015625, 0.5364674682617188, 0.5364838256835938, 0.536489990234375, 0.5365524291992188, 0.5364920043945313, 0.5365155639648438, 0.5368043823242188, 0.5363507080078125, 0.5365288696289062, 0.536427490234375, 0.5368023071289062, 0.536543212890625, 0.5366497192382812, 0.53667431640625, 0.536791015625, 0.5367030029296875, 0.5367736206054687, 0.5366179809570313, 0.5368545532226563, 0.5366896362304687, 0.536679443359375, 0.536700927734375, 0.5370521850585938, 0.5366261596679688, 0.5369262084960937, 0.5365770263671875, 0.537017333984375, 0.5365524291992188, 0.5369661254882813, 0.5370050659179687, 0.5367940673828125, 0.5367890014648438, 0.5367664794921875, 0.5368299560546875, 0.5367787475585938, 0.5368770751953125, 0.5369937744140625, 0.5367050170898438, 0.5367930908203125, 0.5367347412109374, 0.5366651000976562, 0.536774658203125, 0.5371709594726563, 0.5370582885742188, 0.5370501098632813, 0.5367060546875, 0.5368678588867187, 0.5369467163085937, 0.5368309936523438, 0.53673779296875, 0.5371043701171875, 0.5370818481445313, 0.53697021484375, 1.15262255859375, 0.5366098022460938, 0.5367449340820313, 0.5365330200195313, 0.5364357299804687, 0.5369825439453125, 0.5364961547851562, 0.5365463256835937, 0.5367470092773438, 0.5364203491210937, 0.5364705200195312, 0.5365596313476563, 0.536605712890625, 0.5364869384765625, 0.536479736328125, 0.53650634765625, 0.5368033447265625, 0.536511474609375, 0.5366036376953125, 0.5365176391601563, 0.536658935546875, 0.5366558837890625, 0.5366282348632813, 0.5366927490234376, 0.5367890014648438, 0.5367767333984375, 0.5369415893554688, 0.5365975341796875, 0.5366558837890625, 0.536795166015625, 0.5368033447265625, 0.5368053588867188, 0.5367357177734375, 0.5365176391601563, 0.5369917602539063, 0.5366026000976563, 0.5368033447265625, 0.5366282348632813, 0.5369937744140625, 0.5366507568359375, 0.5368033447265625, 0.5367060546875, 0.536642578125, 0.5367367553710938, 0.5369609985351562, 0.5374689331054687, 0.5369559326171875, 0.5367736206054687, 0.5367306518554688, 0.5366753540039062, 0.5367930908203125, 0.5368616943359374, 0.5366712036132812, 0.536748046875, 0.5371443481445313, 0.5367633666992188, 0.53686376953125, 0.5369989013671875, 0.5367193603515625, 0.5363732299804688, 0.536826904296875, 0.5373204345703125, 0.5370480346679688, 1.1528192138671876, 0.53673779296875, 0.5365780639648438, 0.5367572631835937, 0.5364910278320313, 0.53661083984375, 0.5366067504882812, 0.536605712890625, 0.5366712036132812, 0.5365125122070312, 0.5364674682617188, 0.5367664794921875, 0.536774658203125, 0.5365452880859375, 0.5363015747070312, 0.53698046875, 0.5363916625976562, 0.5364777221679687, 0.5366456298828125, 0.5368361206054687, 0.536827880859375, 0.5365176391601563, 0.5365032958984375, 0.5367030029296875, 0.5367664794921875, 0.5367992553710937, 0.5368186645507812, 0.5367367553710938, 0.5366241455078125, 0.536700927734375, 0.536721435546875, 0.536690673828125, 0.536585205078125, 0.5365718994140625, 0.5372119140625, 0.5367449340820313, 0.536901611328125, 0.5367623901367188, 0.5369210815429688, 0.5369815063476563, 0.536816650390625, 0.5365401611328126, 0.536647705078125, 0.5367623901367188, 0.5368576049804688, 0.536985595703125, 0.5370214233398437, 0.5368043823242188, 0.5366978759765625, 0.5367767333984375, 0.5368760375976562, 0.5366179809570313, 0.5367091064453124, 0.5368545532226563, 0.5369886474609376, 0.53673779296875, 0.5370286254882812, 0.5367930908203125, 0.5368330078125, 0.5369763793945312, 0.5369528198242187, 0.5366415405273437, 0.53699072265625, 1.1527352294921875, 0.5367654418945312, 0.5367920532226562, 0.5366527709960938, 0.5365780639648438, 0.5366466674804687, 0.5366159057617188, 0.5365032958984375, 0.5366824951171875, 0.53646337890625, 0.5366988525390625, 0.536748046875, 0.5366405029296875, 0.536431640625, 0.5365883178710937, 0.5366179809570313, 0.5364172973632813, 0.536437744140625, 0.536605712890625, 0.536542236328125, 0.5368411865234375, 0.536489990234375, 0.5365667724609375, 0.5365811157226562, 0.5368033447265625, 0.5366507568359375, 0.5368831787109375, 0.5366231079101562, 0.53659033203125, 0.5368576049804688, 0.53665380859375, 0.5366968383789062, 0.536958984375, 0.5366138916015625, 0.5369302978515625, 0.5366302490234375, 0.5368606567382812, 0.5370828857421875, 0.5368504028320312, 0.5366947631835938, 0.5367859497070312, 0.5365924072265625, 0.5367418823242187, 0.5366651000976562, 0.536711181640625, 0.5369784545898437, 0.5370706176757812, 0.536806396484375, 0.537006103515625, 0.53665380859375, 0.536748046875, 0.5366722412109375, 0.536764404296875, 0.5368176879882812, 0.5370214233398437, 0.5366753540039062, 0.5369395141601563, 0.5368197021484375, 0.537064453125, 0.5366466674804687, 0.536995849609375, 0.5367654418945312, 0.536900634765625]",tokens/s,1.8330061935774278,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6192,7 +6192,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6749,7 +6749,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpp77ah7ch/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7690,7 +7690,7 @@ ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please req ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1269.682176,1019.74016,0.0,373.293056,277.410816,s,10,0.37923129653930665,0.037923129653930666,0.0012152886645715089,0.038048974990844725,0.038519168853759764,0.0396308479309082,0.04052019119262695,"[0.04074252700805664, 0.03821686553955078, 0.03797974395751953, 0.03823471832275391, 0.03626841735839844, 0.03608185577392578, 0.03733708953857422, 0.03811382293701172, 0.03827212905883789, 0.03798412704467773]",tokens/s,6750.497713035296,kWh,4.301884120160883e-07,2.3572264180256752e-07,1.164744366137984e-06,1.83065541995664e-06,tokens/kWh,139840626.0453229,MB,1269.682176,1019.74016,0.0,373.293056,323.047424,s,10,23.493592773437502,2.3493592773437504,0.014632907695307939,2.3561588134765623,2.362567529296875,2.3635712402343754,2.3643742089843753,"[2.34352734375, 2.36090478515625, 2.362344482421875, 2.364574951171875, 2.338309326171875, 2.331607421875, 2.357506103515625, 2.3608046875, 2.3548115234375, 2.3192021484375]",tokens/s,26.81582191687153,kWh,2.642717914328431e-05,1.4482863030778391e-05,6.568572578086276e-05,0.00010659576795492543,tokens/kWh,591017.8350292469,,s,629,23.794291763305676,0.037828762739754634,0.004511848848474209,0.03739033508300781,0.03767849044799805,0.03791810684204102,0.07427948760986337,"[0.03534745788574219, 0.03654348754882813, 0.03639910507202149, 0.03678003311157227, 0.03678003311157227, 0.036736000061035154, 0.036125694274902344, 0.03654348754882813, 0.036519935607910156, 0.0366110725402832, 0.03650764846801758, 0.03670425415039062, 0.03646771240234375, 0.037282817840576174, 0.03782860946655273, 0.03862527847290039, 0.037749759674072264, 0.03757056045532227, 0.037465087890625, 0.0372408332824707, 0.03734630584716797, 0.03750092697143555, 0.03727872085571289, 0.03741388702392578, 0.03740262222290039, 0.03726131057739258, 0.037303295135498044, 0.03731353759765625, 0.037188640594482424, 0.037268447875976565, 0.03742617416381836, 0.03730124664306641, 0.03739136123657227, 0.037250049591064455, 0.03735039901733399, 0.03733913421630859, 0.03732787322998047, 0.03716505432128906, 0.03726233673095703, 0.0372490234375, 0.03732070541381836, 0.03729305648803711, 0.03743129730224609, 0.03723980712890625, 0.037269504547119144, 0.03735039901733399, 0.03727769470214844, 0.03724800109863281, 0.03734124755859375, 0.037339073181152344, 0.037269504547119144, 0.03736678314208984, 0.0371701774597168, 0.037266433715820314, 0.03737190246582031, 0.037318656921386716, 0.03726847839355469, 0.03745894241333008, 0.03732787322998047, 0.03729715347290039, 0.037501953125, 0.037501953125, 0.07589580535888672, 0.03728793716430664, 0.037407745361328126, 0.037207038879394534, 0.03731353759765625, 0.03784089660644531, 0.037353473663330077, 0.03739340972900391, 0.0374015998840332, 0.03751731109619141, 0.03856486511230469, 0.037754878997802735, 0.037369857788085936, 0.03734220886230469, 0.03731353759765625, 0.0373125114440918, 0.03733606338500976, 0.037425151824951174, 0.037303295135498044, 0.037362686157226564, 0.038662143707275394, 0.037748737335205076, 0.03751116943359375, 0.03752243041992188, 0.03749785614013672, 0.03739340972900391, 0.03728793716430664, 0.03726540756225586, 0.03750707244873047, 0.03739340972900391, 0.03738828659057617, 0.03735859298706055, 0.037395454406738284, 0.03731558227539063, 0.037868545532226565, 0.037456897735595705, 0.03736883163452148, 0.03723161697387695, 0.03758489608764649, 0.03761151885986328, 0.03742822265625, 0.037384193420410154, 0.03741900634765625, 0.03744976043701172, 0.03728688049316406, 0.03746815872192383, 0.03732275390625, 0.037525505065917966, 0.0375203857421875, 0.03739340972900391, 0.03733913421630859, 0.03781631851196289, 0.03749785614013672, 0.03746815872192383, 0.03744870376586914, 0.037324798583984374, 0.03752345657348633, 0.03728691101074219, 0.0374015998840332, 0.03738828659057617, 0.037454849243164064, 0.037454849243164064, 0.03749273681640625, 0.0757176284790039, 0.03736166381835938, 0.03737702560424805, 0.03760332870483398, 0.03744768142700195, 0.037354496002197264, 0.03724595260620117, 0.037343231201171875, 0.03800678253173828, 0.03777740859985351, 0.03753472137451172, 0.03747430419921875, 0.0374835205078125, 0.037585918426513674, 0.037338111877441404, 0.0374466552734375, 0.03750912094116211, 0.037379070281982424, 0.03765964889526367, 0.03735551834106445, 0.03742310333251953, 0.03738726425170898, 0.03740364837646484, 0.037495807647705076, 0.03738214492797851, 0.037310462951660156, 0.03733401489257813, 0.03747430419921875, 0.03743436813354492, 0.0373831672668457, 0.03754393768310547, 0.037370880126953124, 0.03760025787353516, 0.037601280212402347, 0.037510143280029294, 0.0373043212890625, 0.0373125114440918, 0.03747225570678711, 0.03762688064575195, 0.037282817840576174, 0.037294078826904296, 0.03753267288208008, 0.037607425689697264, 0.037400577545166014, 0.037495807647705076, 0.03745075225830078, 0.038043647766113284, 0.037773311614990236, 0.03741798400878906, 0.03741183853149414, 0.03771289443969727, 0.03767193603515625, 0.03745280075073242, 0.0372674560546875, 0.03750092697143555, 0.038141952514648435, 0.037969921112060545, 0.037384193420410154, 0.03739852905273437, 0.03727977752685547, 0.03753366470336914, 0.037572608947753904, 0.03738726425170898, 0.07624192047119141, 0.03899084854125977, 0.037817344665527344, 0.03795356750488281, 0.03747836685180664, 0.03777536010742188, 0.03754905700683594, 0.03737497711181641, 0.03731763076782227, 0.03754598236083984, 0.0381317138671875, 0.037395454406738284, 0.03748556900024414, 0.03738726425170898, 0.037493759155273435, 0.037318656921386716, 0.03763916778564453, 0.03733299255371094, 0.037407745361328126, 0.03729817581176758, 0.03746918487548828, 0.03743129730224609, 0.03740364837646484, 0.03732275390625, 0.03753267288208008, 0.03765350341796875, 0.03748863983154297, 0.03742617416381836, 0.03767708969116211, 0.03730940628051758, 0.03756748962402344, 0.03845017623901367, 0.03810201644897461, 0.03751321411132812, 0.03726847839355469, 0.03728998565673828, 0.03749273681640625, 0.03731148910522461, 0.03733606338500976, 0.03725107192993164, 0.03753881454467774, 0.03738726425170898, 0.037282817840576174, 0.03728179168701172, 0.0374917106628418, 0.037351425170898435, 0.03746819305419922, 0.037434337615966796, 0.037440513610839846, 0.03745587158203125, 0.03747532653808594, 0.03769651031494141, 0.03751935958862305, 0.037397502899169925, 0.03741491317749023, 0.037466110229492186, 0.03749478530883789, 0.037566463470458986, 0.03733401489257813, 0.03742822265625, 0.03756851196289063, 0.037501953125, 0.03759820938110352, 0.07184793853759766, 0.03510374450683594, 0.03521535873413086, 0.03516928100585937, 0.03511603164672852, 0.03513241577148438, 0.03520000076293945, 0.035148799896240236, 0.03514470291137695, 0.03511500930786133, 0.0367718391418457, 0.03750400161743164, 0.03728998565673828, 0.037482494354248046, 0.0374835205078125, 0.03756032180786133, 0.037303295135498044, 0.037424129486083986, 0.03768729782104492, 0.03739648056030274, 0.03758492660522461, 0.03751830291748047, 0.03748659133911133, 0.03748863983154297, 0.037410816192626956, 0.03736883163452148, 0.03754396820068359, 0.03754390335083008, 0.037591041564941405, 0.03752140808105469, 0.03739136123657227, 0.03755212783813477, 0.03739340972900391, 0.03741491317749023, 0.03769139099121094, 0.037408767700195314, 0.038091777801513675, 0.03745382308959961, 0.0376360969543457, 0.03739136123657227, 0.03732070541381836, 0.03746303939819336, 0.03743334579467773, 0.037364734649658206, 0.037526527404785154, 0.0374015998840332, 0.037416961669921874, 0.037416961669921874, 0.03750604629516602, 0.037427200317382815, 0.037367809295654295, 0.037498878479003905, 0.037397502899169925, 0.03749683380126953, 0.037622814178466794, 0.03773948669433594, 0.037416961669921874, 0.03744464111328125, 0.03751728057861328, 0.0374835205078125, 0.03757056045532227, 0.037477375030517575, 0.037720062255859374, 0.0716236801147461, 0.035062782287597655, 0.03518668746948242, 0.035019775390625, 0.03531980895996094, 0.035253246307373046, 0.03521843338012695, 0.034840576171875, 0.03494911956787109, 0.03526863861083984, 0.03520406341552734, 0.035253246307373046, 0.03508428955078125, 0.0372490234375, 0.037318656921386716, 0.03726335906982422, 0.03732787322998047, 0.03786342239379883, 0.037288959503173826, 0.0374835205078125, 0.03732992172241211, 0.03781631851196289, 0.038765567779541016, 0.03768012619018555, 0.03770982360839844, 0.037367809295654295, 0.03745894241333008, 0.03721318435668945, 0.03721420669555664, 0.03736678314208984, 0.03769343948364258, 0.03735756683349609, 0.03937484741210937, 0.03793203353881836, 0.037556224822998044, 0.03736883163452148, 0.03729510498046875, 0.03745177459716797, 0.03745792007446289, 0.037561344146728515, 0.037422080993652344, 0.037353473663330077, 0.037594112396240234, 0.03732070541381836, 0.03734732818603516, 0.037340160369873046, 0.037397502899169925, 0.037362686157226564, 0.03757875061035156, 0.03736064147949219, 0.037321727752685545, 0.03742822265625, 0.03740364837646484, 0.037318656921386716, 0.03724492645263672, 0.03730022430419922, 0.03753472137451172, 0.0374200325012207, 0.03728486251831055, 0.03722751998901367, 0.03739136123657227, 0.037362686157226564, 0.037375999450683595, 0.07522509002685547, 0.03725209426879883, 0.03725823974609375, 0.0374466552734375, 0.03733916854858398, 0.03731145477294922, 0.03730124664306641, 0.03746406555175781, 0.037375999450683595, 0.03750707244873047, 0.037292030334472655, 0.037459968566894535, 0.03742924880981445, 0.037405696868896485, 0.03738009643554688, 0.0372592658996582, 0.03754905700683594, 0.03736067199707031, 0.03739644622802735, 0.037394432067871096, 0.037394432067871096, 0.037392383575439454, 0.037389312744140625, 0.0372674560546875, 0.03746713638305664, 0.037416961669921874, 0.03733606338500976, 0.03751321411132812, 0.03757056045532227, 0.037351425170898435, 0.03736883163452148, 0.037335041046142575, 0.03735756683349609, 0.03738521575927734, 0.03734630584716797, 0.03721932983398438, 0.037397502899169925, 0.03740364837646484, 0.037465087890625, 0.03727974319458008, 0.03751116943359375, 0.03736883163452148, 0.03720294570922852, 0.03728179168701172, 0.03764019012451172, 0.037395454406738284, 0.037294078826904296, 0.037379070281982424, 0.03743231964111328, 0.03737705612182617, 0.03722953414916992, 0.03727360153198242, 0.03775692749023438, 0.03735039901733399, 0.03786652755737305, 0.03738211059570312, 0.03764019012451172, 0.03744255828857422, 0.03744153594970703, 0.03740671920776367, 0.038042625427246096, 0.03756032180786133, 0.03753267288208008, 0.07615692901611328, 0.03728486251831055, 0.03751628875732422, 0.03750092697143555, 0.03723161697387695, 0.03742105484008789, 0.03728384017944336, 0.03728793716430664, 0.03723369598388672, 0.037351390838623044, 0.0372592658996582, 0.03703603363037109, 0.03736576080322265, 0.03826483154296875, 0.03825356674194336, 0.037498878479003905, 0.03745075225830078, 0.03750604629516602, 0.03733606338500976, 0.0373493766784668, 0.037405696868896485, 0.037548030853271484, 0.037897216796875, 0.0374015998840332, 0.03749273681640625, 0.03787980651855469, 0.03769548797607422, 0.03733401489257813, 0.037424129486083986, 0.03724288177490234, 0.03732275390625, 0.03731353759765625, 0.037348350524902346, 0.0373043212890625, 0.03746099090576172, 0.03751731109619141, 0.037424129486083986, 0.037477375030517575, 0.03738623809814453, 0.037395454406738284, 0.0372828483581543, 0.0373001937866211, 0.03731455993652344, 0.03749683380126953, 0.037678081512451174, 0.03893964767456055, 0.037498878479003905, 0.03726847839355469, 0.03724595260620117, 0.037381118774414065, 0.03728998565673828, 0.03745177459716797, 0.03747635269165039, 0.037425151824951174, 0.03755724716186523, 0.03739033508300781, 0.03731763076782227, 0.03747532653808594, 0.03732070541381836, 0.03763302230834961, 0.0374917106628418, 0.03746303939819336, 0.037392383575439454, 0.07604838562011719, 0.03738726425170898, 0.03732275390625, 0.037763072967529294, 0.03760537719726562, 0.037493759155273435, 0.03744153594970703, 0.037367809295654295, 0.037353473663330077, 0.03748966217041016, 0.037321727752685545, 0.03739136123657227, 0.03747020721435547, 0.03746201705932617, 0.03758899307250976, 0.037288959503173826, 0.03742617416381836, 0.037359615325927735, 0.03741900634765625, 0.03733708953857422, 0.0372592658996582, 0.03737497711181641, 0.03758182525634766, 0.03737395095825195, 0.03727052688598633, 0.03736678314208984, 0.037477375030517575, 0.03746201705932617, 0.03737702560424805, 0.03690291213989258, 0.03730022430419922, 0.037541889190673826, 0.03752140808105469, 0.03749683380126953, 0.03740979385375977, 0.037354496002197264, 0.037292030334472655, 0.03724288177490234, 0.03722956848144531, 0.03722956848144531, 0.0375203857421875, 0.03734630584716797, 0.03717836761474609, 0.0371671028137207, 0.037372928619384765, 0.037341182708740234, 0.03730636978149414, 0.03713945770263672, 0.03719987106323242, 0.03745177459716797, 0.03732889556884766, 0.03736166381835938, 0.0373493766784668, 0.03739648056030274, 0.03730739212036133, 0.03730739212036133, 0.03730944061279297, 0.03724390411376953, 0.037353473663330077, 0.03740467071533203, 0.0375203857421875, 0.03734630584716797, 0.03746406555175781, 0.07565721893310547, 0.0372674560546875, 0.0373125114440918, 0.037479423522949216, 0.03725721740722656, 0.03723468780517578, 0.0374200325012207, 0.03733196640014649, 0.03729817581176758, 0.03751321411132812, 0.03730739212036133, 0.03743334579467773, 0.03721932983398438, 0.03751833724975586, 0.03735244750976562, 0.037408767700195314, 0.03565260696411133, 0.035312641143798826, 0.03527782440185547, 0.0352542724609375, 0.03772415924072266, 0.0377077751159668, 0.03799244689941406, 0.03948953628540039, 0.037106689453125, 0.03671449661254883, 0.03655372619628906, 0.03666534423828125, 0.03620249557495117, 0.03514470291137695, 0.03519692611694336, 0.035335166931152344, 0.03523788833618164, 0.03600076675415039, 0.03778662490844727, 0.03708927917480469, 0.03679846572875976, 0.036724769592285156, 0.03671651077270508, 0.0366192626953125, 0.036624416351318356, 0.03651375961303711, 0.03685990524291992, 0.03674524688720703, 0.03836924743652344, 0.03679436874389649, 0.03653734588623047, 0.03676671981811523, 0.03650559997558594, 0.03653529739379883, 0.036706302642822264, 0.03682099151611328, 0.036528129577636716, 0.03685887908935547, 0.03667865753173828, 0.036674560546875, 0.036618240356445314, 0.03663052749633789, 0.03661721420288086, 0.03660595321655274, 0.0365926399230957, 0.036544513702392575, 0.036792320251464845]",tokens/s,26.434911627418618,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2012.44672,5274.861568,0.0,4628.414464,4279.515648,s,10,3.0981137390136726,0.3098113739013672,0.0009457574536517518,0.3095596923828125,0.31112299194335935,0.3115160552978516,0.31183050598144535,"[0.3090160217285156, 0.3099482421875, 0.3095133972167969, 0.3094310302734375, 0.3096059875488281, 0.30901934814453125, 0.31002566528320313, 0.3086092834472656, 0.3119091186523438, 0.31103564453125]",tokens/s,826.3092370569365,kWh,3.6482026631181896e-06,1.9990483218862206e-06,1.7856563106788066e-05,2.3503814091792478e-05,tokens/kWh,10891849.254772445,MB,2012.44672,5274.861568,0.0,4628.414464,4463.184384,s,10,173.15158984375,17.315158984375003,0.0008435493287116899,17.3153154296875,17.316042382812498,17.31623310546875,17.31638568359375,"[17.31562890625, 17.316, 17.314966796875, 17.31573828125, 17.3152734375, 17.314123046875, 17.314578125, 17.315357421875, 17.316423828125, 17.3135]",tokens/s,3.638430352089199,kWh,0.0002044354933214308,0.00011204598750467993,0.0009878629636904115,0.0013043444445165224,tokens/kWh,48300.12521987782,,s,629,175.66576879882834,0.2792778518264358,0.036821094771226096,0.274830322265625,0.27505172119140625,0.27513446044921874,0.5842676342773437,"[0.2750351257324219, 0.27489178466796876, 0.274735107421875, 0.274735107421875, 0.2746921081542969, 0.27469003295898436, 0.2748016662597656, 0.2746644592285156, 0.2747750549316406, 0.274651123046875, 0.2747361145019531, 0.2748713073730469, 0.2746357727050781, 0.27475662231445314, 0.2746726379394531, 0.274840576171875, 0.27462655639648437, 0.27478732299804687, 0.27498086547851563, 0.27470437622070315, 0.2747474060058594, 0.2747381896972656, 0.27487026977539064, 0.27473919677734376, 0.27474533081054686, 0.27481497192382814, 0.27486822509765624, 0.27485592651367186, 0.27480474853515624, 0.27489178466796876, 0.2747463684082031, 0.27480062866210936, 0.2747781066894531, 0.2749091796875, 0.27479757690429685, 0.274808837890625, 0.2748692626953125, 0.27493991088867187, 0.27494195556640627, 0.27483444213867186, 0.275083251953125, 0.2749153137207031, 0.2749696044921875, 0.27500238037109376, 0.27485491943359375, 0.274951171875, 0.27491738891601564, 0.2748160095214844, 0.2748252258300781, 0.27477914428710937, 0.27466546630859373, 0.27490304565429685, 0.27493991088867187, 0.2749286499023437, 0.2748641357421875, 0.27500338745117187, 0.274850830078125, 0.2748538818359375, 0.2749900817871094, 0.27501055908203126, 0.2749440002441406, 0.27534439086914064, 0.5850224609375, 0.27484774780273435, 0.2746152954101562, 0.27474533081054686, 0.2746490783691406, 0.27462554931640626, 0.2749921264648437, 0.27478118896484377, 0.27476480102539064, 0.2747699279785156, 0.2747268981933594, 0.2748231811523438, 0.27470437622070315, 0.27464602661132814, 0.27498904418945314, 0.2747268981933594, 0.27469720458984376, 0.27473919677734376, 0.2747781066894531, 0.2746091613769531, 0.27474227905273435, 0.2748487548828125, 0.27456307983398437, 0.2747607116699219, 0.2747002868652344, 0.2747279357910156, 0.2747740173339844, 0.274798583984375, 0.27483444213867186, 0.27479962158203125, 0.27488357543945313, 0.2746152954101562, 0.2749706115722656, 0.275030029296875, 0.275135498046875, 0.27472897338867186, 0.27489279174804687, 0.2747115478515625, 0.2752081909179688, 0.2749573059082031, 0.2748231811523438, 0.2751098937988281, 0.274830322265625, 0.27489178466796876, 0.27508633422851564, 0.2750218200683594, 0.27489178466796876, 0.27495526123046876, 0.2749982604980469, 0.274872314453125, 0.2750330810546875, 0.2748805236816406, 0.27481805419921873, 0.27501568603515625, 0.27499725341796877, 0.27485696411132815, 0.27509146118164063, 0.2748487548828125, 0.27496652221679685, 0.2749593505859375, 0.27505868530273436, 0.27497369384765624, 0.2752174072265625, 0.5844049682617187, 0.2746695556640625, 0.274724853515625, 0.27476171875, 0.2747268981933594, 0.27471768188476564, 0.2747432861328125, 0.27462655639648437, 0.27470233154296875, 0.2746142578125, 0.27483444213867186, 0.27469619750976565, 0.274808837890625, 0.274740234375, 0.2747002868652344, 0.27517645263671875, 0.2746726379394531, 0.27463885498046875, 0.2747054138183594, 0.27489178466796876, 0.2747258911132813, 0.27476480102539064, 0.2746378173828125, 0.27478118896484377, 0.27469003295898436, 0.2747064208984375, 0.2747555847167969, 0.27474432373046875, 0.27460198974609373, 0.27469003295898436, 0.2750320739746094, 0.2746746826171875, 0.27475149536132815, 0.27477093505859373, 0.274798583984375, 0.27475762939453124, 0.27493374633789064, 0.27482623291015623, 0.27482623291015623, 0.27523379516601565, 0.2749716491699219, 0.27489178466796876, 0.27494708251953126, 0.27481805419921873, 0.2751744079589844, 0.275030029296875, 0.2748907470703125, 0.27488357543945313, 0.2750771179199219, 0.27486004638671874, 0.27486514282226565, 0.27501873779296876, 0.27495834350585935, 0.2748692626953125, 0.27510885620117187, 0.27491021728515624, 0.2748241882324219, 0.27502590942382815, 0.27491943359375, 0.2753228759765625, 0.27508428955078124, 0.27493887329101563, 0.27487026977539064, 0.5844735717773437, 0.2747054138183594, 0.2746378173828125, 0.274555908203125, 0.2748323974609375, 0.2746695556640625, 0.274724853515625, 0.27464190673828126, 0.2747678833007812, 0.2747187194824219, 0.2747412414550781, 0.274830322265625, 0.2746726379394531, 0.27480780029296875, 0.2746644592285156, 0.2751068115234375, 0.2747135925292969, 0.2747135925292969, 0.27464498901367185, 0.2746429443359375, 0.2747914123535156, 0.27477914428710937, 0.27488357543945313, 0.2748436584472656, 0.27502590942382815, 0.2748436584472656, 0.2747268981933594, 0.2746992492675781, 0.27483444213867186, 0.274693115234375, 0.2747832336425781, 0.27517236328125, 0.27475662231445314, 0.2747555847167969, 0.27486822509765624, 0.27485491943359375, 0.274951171875, 0.274946044921875, 0.2747904052734375, 0.274951171875, 0.27484161376953126, 0.27482623291015623, 0.27490509033203125, 0.27496755981445314, 0.27509963989257813, 0.2749859924316406, 0.2749716491699219, 0.274808837890625, 0.27492352294921873, 0.27492044067382815, 0.2748436584472656, 0.2750330810546875, 0.27496551513671874, 0.27488973999023436, 0.2748108825683594, 0.2751109008789063, 0.2750556030273438, 0.27504229736328123, 0.2751866760253906, 0.27491226196289065, 0.2749624328613281, 0.27511398315429686, 0.27482623291015623, 0.5845493774414062, 0.27454156494140625, 0.27462451171875, 0.2746767272949219, 0.27461221313476564, 0.274740234375, 0.2747412414550781, 0.27475762939453124, 0.2746275939941406, 0.2747412414550781, 0.27471563720703124, 0.274661376953125, 0.2746071166992187, 0.2746777648925781, 0.27463168334960936, 0.2747688903808594, 0.274735107421875, 0.27458251953125, 0.27470745849609374, 0.2746746826171875, 0.27465625, 0.2746163330078125, 0.2748231811523438, 0.2747340698242188, 0.27481396484375, 0.27505767822265625, 0.274723876953125, 0.27478421020507815, 0.2748016662597656, 0.27497882080078123, 0.2749859924316406, 0.2747699279785156, 0.2748671875, 0.2748887023925781, 0.2750474243164063, 0.27465728759765623, 0.2749900817871094, 0.27487335205078123, 0.27479962158203125, 0.2749266052246094, 0.2752122802734375, 0.274798583984375, 0.27484161376953126, 0.2749644775390625, 0.2748784790039063, 0.2749716491699219, 0.2749306945800781, 0.27479757690429685, 0.275051513671875, 0.275019775390625, 0.2748436584472656, 0.2749941711425781, 0.27496551513671874, 0.2748098449707031, 0.2751098937988281, 0.2750474243164063, 0.27502590942382815, 0.27502285766601564, 0.27510580444335936, 0.2749747314453125, 0.2753966064453125, 0.27505255126953126, 0.27491839599609375, 0.5839144897460937, 0.27459890747070315, 0.27473101806640626, 0.274634765625, 0.2746849365234375, 0.27491943359375, 0.2749634704589844, 0.27464398193359374, 0.27470950317382814, 0.274671630859375, 0.2746480712890625, 0.27475250244140625, 0.2749429626464844, 0.2747258911132813, 0.2747258911132813, 0.27462451171875, 0.27475045776367185, 0.27468084716796876, 0.2746275939941406, 0.27471563720703124, 0.27465933227539063, 0.27472384643554687, 0.27456103515625, 0.2747381896972656, 0.27501773071289065, 0.27480062866210936, 0.27477093505859373, 0.2747555847167969, 0.27470950317382814, 0.27481396484375, 0.2747105407714844, 0.27481497192382814, 0.27507302856445315, 0.27469720458984376, 0.27496652221679685, 0.274840576171875, 0.27497369384765624, 0.2749081726074219, 0.27491326904296876, 0.2748784790039063, 0.27476275634765623, 0.2748641357421875, 0.2749542541503906, 0.2748293151855469, 0.27482009887695313, 0.2748272705078125, 0.2749962158203125, 0.27489178466796876, 0.27492352294921873, 0.27513446044921874, 0.2748805236816406, 0.27491326904296876, 0.2747801513671875, 0.27504434204101563, 0.27497369384765624, 0.2749910888671875, 0.2750679016113281, 0.27495526123046876, 0.27496551513671874, 0.2748948364257813, 0.27491326904296876, 0.27494503784179686, 0.27482827758789063, 0.5851586303710937, 0.27469720458984376, 0.2747381896972656, 0.2747084655761719, 0.274798583984375, 0.27487335205078123, 0.274951171875, 0.2746634216308594, 0.2746368103027344, 0.2747054138183594, 0.274619384765625, 0.2746009521484375, 0.27468389892578127, 0.2746132507324219, 0.27507098388671875, 0.27484161376953126, 0.2746378173828125, 0.2746726379394531, 0.27468698120117185, 0.2747781066894531, 0.27464703369140625, 0.27478427124023436, 0.2746921081542969, 0.2746777648925781, 0.2748323974609375, 0.27475149536132815, 0.2747678833007812, 0.27469415283203125, 0.27480575561523435, 0.2747852783203125, 0.2747883605957031, 0.27465420532226564, 0.2747821960449219, 0.2746368103027344, 0.27494195556640627, 0.274830322265625, 0.27488153076171873, 0.27490509033203125, 0.274777099609375, 0.27503103637695314, 0.27477297973632814, 0.27513446044921874, 0.2748323974609375, 0.2748661804199219, 0.2748856201171875, 0.27483648681640627, 0.27477914428710937, 0.27497882080078123, 0.27492556762695314, 0.2749020080566406, 0.2750597229003906, 0.27492352294921873, 0.274882568359375, 0.274956298828125, 0.2749982604980469, 0.2749716491699219, 0.2750351257324219, 0.2753280029296875, 0.2750433349609375, 0.27495321655273436, 0.27487744140625, 0.2749931640625, 0.2749010009765625, 0.583531494140625, 0.2746480712890625, 0.27469003295898436, 0.27469619750976565, 0.2748661804199219, 0.27466650390625, 0.27472076416015623, 0.27455487060546874, 0.2746767272949219, 0.27469619750976565, 0.2746142578125, 0.2748641357421875, 0.2747432861328125, 0.27524609375, 0.2746132507324219, 0.27479449462890626, 0.274740234375, 0.2750003051757812, 0.27491839599609375, 0.2748252258300781, 0.27477197265625, 0.27485592651367186, 0.274724853515625, 0.27488357543945313, 0.2749091796875, 0.27475762939453124, 0.27470745849609374, 0.27477093505859373, 0.27474227905273435, 0.2748016662597656, 0.2748610534667969, 0.27486309814453125, 0.2747187194824219, 0.2747607116699219, 0.2748374938964844, 0.27483544921875, 0.27485696411132815, 0.2749071350097656, 0.27478732299804687, 0.275030029296875, 0.2747340698242188, 0.2749521789550781, 0.2749276123046875, 0.27486004638671874, 0.2748692626953125, 0.2750320739746094, 0.2748671875, 0.27484979248046876, 0.274956298828125, 0.27482623291015623, 0.27489279174804687, 0.2752983093261719, 0.27487640380859374, 0.2749521789550781, 0.2748948364257813, 0.2748518371582031, 0.2748876647949219, 0.2750433349609375, 0.2751068115234375, 0.27513446044921874, 0.27488153076171873, 0.27497268676757813, 0.27485592651367186, 0.5868973999023438, 0.2745733032226563, 0.2749358215332031, 0.2747268981933594, 0.2746163330078125, 0.2747258911132813, 0.2748098449707031, 0.2746552429199219, 0.2747821960449219, 0.2746501159667969, 0.2746081237792969, 0.2747218017578125, 0.2747586669921875, 0.2747258911132813, 0.2748395385742187, 0.27473202514648437, 0.27477297973632814, 0.27475149536132815, 0.27478118896484377, 0.27469415283203125, 0.2749081726074219, 0.2747852783203125, 0.27475967407226565, 0.2751600646972656, 0.27488973999023436, 0.27475662231445314, 0.27469830322265626, 0.27466439819335936, 0.27470745849609374, 0.2747054138183594, 0.2749347839355469, 0.2748016662597656, 0.2751201171875, 0.2747658386230469, 0.2747484130859375, 0.27480474853515624, 0.2748467102050781, 0.27473101806640626, 0.2752983093261719, 0.2748221435546875, 0.27470950317382814, 0.2752850036621094, 0.2749747314453125, 0.2748385314941406, 0.27491738891601564, 0.274850830078125, 0.2748784790039063, 0.2749634704589844, 0.27494503784179686, 0.27479244995117186, 0.27507302856445315, 0.2748968811035156, 0.27493991088867187, 0.2751795349121094, 0.2749296569824219, 0.2749214782714844, 0.27510784912109376, 0.2749081726074219, 0.27524609375, 0.2750965881347656, 0.27500338745117187, 0.2748989562988281, 0.2753546142578125, 0.5860249633789063, 0.27464190673828126, 0.27466650390625, 0.27468389892578127, 0.275125244140625, 0.27456512451171877, 0.27472280883789063, 0.27463168334960936, 0.27477197265625, 0.274661376953125, 0.27458868408203124, 0.27483648681640627, 0.27479244995117186, 0.2745937805175781, 0.27466033935546874, 0.2747381896972656, 0.27474533081054686, 0.2747330627441406, 0.2748692626953125, 0.27472384643554687, 0.2747821960449219, 0.27471563720703124, 0.2746849365234375, 0.27464498901367185, 0.2747750549316406, 0.2746859436035156, 0.27498086547851563, 0.27475662231445314, 0.27458660888671876, 0.2746746826171875, 0.27504537963867187, 0.27469003295898436, 0.27511602783203126, 0.2748887023925781, 0.2747821960449219, 0.2747197570800781, 0.2750822448730469, 0.2747740173339844, 0.2747607116699219, 0.27490509033203125, 0.27476171875, 0.27485491943359375, 0.27493170166015624, 0.27483544921875, 0.2748590087890625, 0.27491122436523435, 0.27494503784179686, 0.2748221435546875, 0.27496038818359375, 0.2748252258300781, 0.2748467102050781, 0.2750003051757812, 0.27494708251953126, 0.2748610534667969, 0.2749634704589844, 0.2748293151855469, 0.27484979248046876, 0.2749798278808594, 0.27497369384765624, 0.274904052734375, 0.275056640625, 0.2749020080566406, 0.27495321655273436]",tokens/s,3.5806634627849943,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8223,7 +8223,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1638.6048,2418.540544,0.0,1772.09344,1621.541376,s,10,0.7307771377563477,0.07307771377563477,0.00030590186697528017,0.07303974151611328,0.07345188369750977,0.07350205345153808,0.07354218925476075,"[0.07344073486328125, 0.07282319641113282, 0.0727369613647461, 0.07273868560791015, 0.07273926544189453, 0.07293727874755859, 0.07314220428466797, 0.07324988555908203, 0.07355222320556641, 0.07341670227050781]",tokens/s,3503.119990671552,kWh,8.589871899613245e-07,4.704820193973832e-07,3.7392864052756902e-06,5.068755614634398e-06,tokens/kWh,50505492.76056682,MB,1638.6048,2418.540544,0.0,1772.09344,1723.273216,s,10,38.20281567382813,3.8202815673828128,0.0022956512254473085,3.8195750732421874,3.82337958984375,3.8243252929687497,3.82508185546875,"[3.82527099609375, 3.81770703125, 3.819302978515625, 3.818822509765625, 3.82106005859375, 3.82316943359375, 3.821086669921875, 3.81984716796875, 3.817920166015625, 3.818628662109375]",tokens/s,16.49093107112517,kWh,4.511295209232352e-05,2.472463132112629e-05,0.00019686329074332517,0.000266700874156775,tokens/kWh,236219.69818879056,,s,629,38.79888072204587,0.06168343517018425,0.008717620254388004,0.06061260986328125,0.06073487243652344,0.060805118560791016,0.1336954461669922,"[0.06117068862915039, 0.061228031158447264, 0.0611962890625, 0.06114406585693359, 0.06115327835083008, 0.06097817611694336, 0.06119424057006836, 0.06067814254760742, 0.060709888458251954, 0.06057062530517578, 0.060631038665771485, 0.060595199584960936, 0.06056959915161133, 0.06056959915161133, 0.06066790390014649, 0.06054707336425781, 0.0605849609375, 0.06076313781738281, 0.06055321502685547, 0.06061670303344727, 0.06071705627441406, 0.06055321502685547, 0.060527614593505856, 0.06062387084960937, 0.060576766967773435, 0.060677120208740234, 0.0607006721496582, 0.06059929656982422, 0.0606033935546875, 0.06082559967041016, 0.060639232635498044, 0.060633087158203126, 0.06068326568603516, 0.060611583709716796, 0.06060236740112305, 0.06072524642944336, 0.060627967834472656, 0.060657665252685546, 0.06071603012084961, 0.06063513565063477, 0.06063616180419922, 0.060747776031494144, 0.06059212875366211, 0.06063513565063477, 0.06071603012084961, 0.06056243133544922, 0.06069452667236328, 0.06072012710571289, 0.06063411331176758, 0.06065868759155273, 0.06072012710571289, 0.06061670303344727, 0.06061670303344727, 0.06073446273803711, 0.06052249526977539, 0.06070476913452148, 0.06071807861328125, 0.06060851287841797, 0.06081433486938476, 0.06072115325927734, 0.06060134506225586, 0.06065971374511719, 0.13372006225585936, 0.06056345748901367, 0.06059929656982422, 0.060644351959228515, 0.06058291244506836, 0.06054195022583008, 0.06054297637939453, 0.06056140899658203, 0.06055731201171875, 0.06055731201171875, 0.06055936050415039, 0.06073241424560547, 0.060556289672851565, 0.06056755065917969, 0.06058700942993164, 0.06049792098999023, 0.06058393478393555, 0.06057984161376953, 0.06057984161376953, 0.06055219268798828, 0.06055219268798828, 0.06055833435058594, 0.060554241180419924, 0.060556289672851565, 0.060524543762207034, 0.06058700942993164, 0.060572673797607425, 0.06054297637939453, 0.060660736083984375, 0.06058291244506836, 0.06063820648193359, 0.06058291244506836, 0.06065868759155273, 0.06058905410766602, 0.06061363220214844, 0.06056959915161133, 0.06060851287841797, 0.060611583709716796, 0.06063820648193359, 0.060581886291503906, 0.06066790390014649, 0.060590080261230465, 0.06060543823242188, 0.0605849609375, 0.06063820648193359, 0.06059724807739258, 0.06060748672485351, 0.06061568069458008, 0.060606464385986325, 0.06057574462890625, 0.06063820648193359, 0.06063820648193359, 0.06061056137084961, 0.06062080001831055, 0.06065151977539063, 0.06050611114501953, 0.06061670303344727, 0.06059724807739258, 0.06063411331176758, 0.06061056137084961, 0.06062387084960937, 0.06065151977539063, 0.06062694549560547, 0.1336197052001953, 0.06059212875366211, 0.06061260986328125, 0.06053887939453125, 0.06051327896118164, 0.06058905410766602, 0.06058086395263672, 0.060821502685546876, 0.06053887939453125, 0.06054912185668945, 0.060526592254638675, 0.06054502487182617, 0.060572673797607425, 0.060606464385986325, 0.06061260986328125, 0.06055014419555664, 0.0605665283203125, 0.0607191047668457, 0.0607825927734375, 0.060576766967773435, 0.060581886291503906, 0.06057369613647461, 0.06056959915161133, 0.060537857055664064, 0.06054297637939453, 0.060614654541015625, 0.060593151092529295, 0.06062694549560547, 0.06073446273803711, 0.06060851287841797, 0.06065356826782226, 0.06064025497436523, 0.06066483306884766, 0.06061056137084961, 0.06058905410766602, 0.06059929656982422, 0.060627967834472656, 0.060805118560791016, 0.06068838500976562, 0.0606033935546875, 0.06065459060668945, 0.060609535217285154, 0.060649471282958986, 0.06057369613647461, 0.060609535217285154, 0.0607825927734375, 0.06077030563354492, 0.06063411331176758, 0.06058291244506836, 0.060590080261230465, 0.06062899017333984, 0.060633087158203126, 0.06063718414306641, 0.06060543823242188, 0.06073344039916992, 0.06053376007080078, 0.06064640045166016, 0.060606464385986325, 0.06062694549560547, 0.0605849609375, 0.06060851287841797, 0.060642303466796874, 0.060631038665771485, 0.13370060729980468, 0.06058803176879883, 0.06058905410766602, 0.060625919342041014, 0.060572673797607425, 0.06060134506225586, 0.060593151092529295, 0.06061363220214844, 0.06056857681274414, 0.06059212875366211, 0.06057984161376953, 0.06055321502685547, 0.0605849609375, 0.06060543823242188, 0.060572673797607425, 0.06061568069458008, 0.06055833435058594, 0.06058291244506836, 0.06060748672485351, 0.06059724807739258, 0.0606033935546875, 0.06055731201171875, 0.060611583709716796, 0.06058905410766602, 0.060537857055664064, 0.06057984161376953, 0.06060031890869141, 0.06059417724609375, 0.06063820648193359, 0.06058598327636719, 0.06059622573852539, 0.06056345748901367, 0.06068121719360352, 0.060641281127929686, 0.060593151092529295, 0.06062694549560547, 0.06065459060668945, 0.060606464385986325, 0.06061056137084961, 0.06062182235717774, 0.060677120208740234, 0.06061772918701172, 0.06061772918701172, 0.06057369613647461, 0.06061260986328125, 0.060611583709716796, 0.06065868759155273, 0.06058598327636719, 0.06074367904663086, 0.06069145584106445, 0.060614654541015625, 0.06061670303344727, 0.06061056137084961, 0.06064025497436523, 0.060668926239013675, 0.06055116653442383, 0.06069657516479492, 0.06063616180419922, 0.06061670303344727, 0.060606464385986325, 0.060649471282958986, 0.06063820648193359, 0.06067814254760742, 0.1336821746826172, 0.06062182235717774, 0.06055321502685547, 0.0606484489440918, 0.06058291244506836, 0.060598270416259765, 0.060590080261230465, 0.06055936050415039, 0.06054604721069336, 0.06061568069458008, 0.06059622573852539, 0.06057164764404297, 0.06057164764404297, 0.060581886291503906, 0.06061875152587891, 0.06058086395263672, 0.06057881546020508, 0.06066790390014649, 0.0607303695678711, 0.06062899017333984, 0.06068940734863281, 0.06056857681274414, 0.06057779312133789, 0.060598270416259765, 0.06055526351928711, 0.0607006721496582, 0.06061670303344727, 0.06058803176879883, 0.06073651123046875, 0.06060543823242188, 0.06059724807739258, 0.060723201751708984, 0.06062182235717774, 0.060598270416259765, 0.06078976058959961, 0.06057779312133789, 0.06063820648193359, 0.0607303695678711, 0.06069760131835938, 0.06062387084960937, 0.06078976058959961, 0.06061260986328125, 0.060715007781982425, 0.0607191047668457, 0.06060748672485351, 0.06065356826782226, 0.06072012710571289, 0.06056345748901367, 0.06062899017333984, 0.0607303695678711, 0.0606668815612793, 0.06066995239257812, 0.060747776031494144, 0.060622848510742185, 0.06086963272094727, 0.06071603012084961, 0.06065459060668945, 0.06058803176879883, 0.06078566360473633, 0.060780544281005856, 0.060609535217285154, 0.06074367904663086, 0.060682239532470705, 0.13460275268554686, 0.060777473449707034, 0.06060748672485351, 0.06059212875366211, 0.0605665283203125, 0.06059212875366211, 0.06088499069213867, 0.06057369613647461, 0.060593151092529295, 0.060865535736083984, 0.060672000885009764, 0.06060748672485351, 0.06066790390014649, 0.0605747184753418, 0.060598270416259765, 0.06067097473144531, 0.060554241180419924, 0.06058700942993164, 0.060897281646728516, 0.06061875152587891, 0.060625919342041014, 0.06071603012084961, 0.06054912185668945, 0.061172737121582034, 0.060655616760253904, 0.060606464385986325, 0.06061568069458008, 0.06068326568603516, 0.060595199584960936, 0.06058086395263672, 0.0607088623046875, 0.0605747184753418, 0.060647422790527344, 0.0607723503112793, 0.060680191040039064, 0.06061363220214844, 0.06075699234008789, 0.06065663909912109, 0.06072012710571289, 0.060821502685546876, 0.06069145584106445, 0.06064025497436523, 0.06078976058959961, 0.060649471282958986, 0.0606453742980957, 0.06069760131835938, 0.06060543823242188, 0.0606033935546875, 0.060731391906738284, 0.060761089324951174, 0.06069247817993164, 0.060706817626953125, 0.06063206481933594, 0.060622848510742185, 0.06078668975830078, 0.06051532745361328, 0.060649471282958986, 0.0607088623046875, 0.0605849609375, 0.06058393478393555, 0.060750846862792966, 0.060598270416259765, 0.06064025497436523, 0.1340211181640625, 0.06060441589355469, 0.06063206481933594, 0.0605665283203125, 0.06056140899658203, 0.06060031890869141, 0.06067097473144531, 0.06052864074707031, 0.06059417724609375, 0.06070169448852539, 0.06054297637939453, 0.06055526351928711, 0.0607303695678711, 0.0605665283203125, 0.06057881546020508, 0.060715007781982425, 0.06057062530517578, 0.06058905410766602, 0.060709888458251954, 0.06054604721069336, 0.06056447982788086, 0.06082048034667969, 0.06061056137084961, 0.06072012710571289, 0.06068326568603516, 0.06057164764404297, 0.06063616180419922, 0.06069145584106445, 0.0606300163269043, 0.06065663909912109, 0.060805118560791016, 0.06057881546020508, 0.060631038665771485, 0.060690433502197265, 0.06058905410766602, 0.060611583709716796, 0.060727294921875, 0.06058598327636719, 0.060622848510742185, 0.06071603012084961, 0.06065049743652344, 0.060642303466796874, 0.06084198379516602, 0.06067609786987305, 0.06059622573852539, 0.06077951812744141, 0.060639232635498044, 0.06058291244506836, 0.06075699234008789, 0.06057164764404297, 0.0606484489440918, 0.06079590225219727, 0.06061056137084961, 0.060631038665771485, 0.0608092155456543, 0.060526592254638675, 0.06073651123046875, 0.06082355117797852, 0.060633087158203126, 0.06062080001831055, 0.060767230987548826, 0.06055219268798828, 0.06063206481933594, 0.1341685791015625, 0.060556289672851565, 0.06056345748901367, 0.06055833435058594, 0.06054707336425781, 0.060593151092529295, 0.06062182235717774, 0.060539905548095706, 0.06055219268798828, 0.06054092788696289, 0.06060134506225586, 0.06057574462890625, 0.060593151092529295, 0.06060031890869141, 0.06060134506225586, 0.0605296630859375, 0.06058700942993164, 0.06057164764404297, 0.06063616180419922, 0.06056857681274414, 0.06062080001831055, 0.06058086395263672, 0.06061260986328125, 0.06055833435058594, 0.06055731201171875, 0.06060748672485351, 0.06112870407104492, 0.060614654541015625, 0.06075904083251953, 0.06056447982788086, 0.060665855407714846, 0.06062387084960937, 0.06066790390014649, 0.06061977767944336, 0.060639232635498044, 0.06057881546020508, 0.060633087158203126, 0.06067302322387695, 0.06075596618652344, 0.06059622573852539, 0.06068735885620117, 0.06061977767944336, 0.06073651123046875, 0.060614654541015625, 0.060633087158203126, 0.06059212875366211, 0.06072422409057617, 0.06060031890869141, 0.06061875152587891, 0.06058905410766602, 0.0606484489440918, 0.06058803176879883, 0.06061772918701172, 0.06061670303344727, 0.06084096145629883, 0.060682239532470705, 0.06066790390014649, 0.06055833435058594, 0.06060441589355469, 0.060657665252685546, 0.06071705627441406, 0.06069145584106445, 0.060665855407714846, 0.13447782897949218, 0.06059622573852539, 0.06055731201171875, 0.06055936050415039, 0.0605296630859375, 0.060560382843017575, 0.06056243133544922, 0.060537857055664064, 0.06055526351928711, 0.06057574462890625, 0.06055116653442383, 0.06061670303344727, 0.060581886291503906, 0.06056243133544922, 0.060593151092529295, 0.06055321502685547, 0.060593151092529295, 0.06059417724609375, 0.06057574462890625, 0.06056140899658203, 0.06055731201171875, 0.060560382843017575, 0.06057574462890625, 0.06057369613647461, 0.06055014419555664, 0.060593151092529295, 0.06060134506225586, 0.060560382843017575, 0.060625919342041014, 0.060581886291503906, 0.06063820648193359, 0.06060748672485351, 0.06061568069458008, 0.06059622573852539, 0.060590080261230465, 0.06055014419555664, 0.0606033935546875, 0.060624897003173826, 0.060633087158203126, 0.060611583709716796, 0.060706817626953125, 0.06060748672485351, 0.06061772918701172, 0.060598270416259765, 0.060639232635498044, 0.06060748672485351, 0.060593151092529295, 0.06059110260009765, 0.06060134506225586, 0.06065151977539063, 0.060631038665771485, 0.06060031890869141, 0.06061875152587891, 0.06061977767944336, 0.06063718414306641, 0.06055731201171875, 0.060641281127929686, 0.060609535217285154, 0.060657665252685546, 0.06061260986328125, 0.06058291244506836, 0.06075904083251953, 0.06065049743652344, 0.13431602478027344, 0.0606300163269043, 0.06057164764404297, 0.06059110260009765, 0.06057881546020508, 0.06055833435058594, 0.060598270416259765, 0.06052249526977539, 0.06056140899658203, 0.060581886291503906, 0.060590080261230465, 0.06055116653442383, 0.06054092788696289, 0.060680191040039064, 0.060581886291503906, 0.06056345748901367, 0.0606033935546875, 0.060633087158203126, 0.060631038665771485, 0.06058803176879883, 0.06057574462890625, 0.06054195022583008, 0.06056243133544922, 0.060539905548095706, 0.06053683090209961, 0.06066175842285156, 0.06061670303344727, 0.060625919342041014, 0.06065868759155273, 0.06060134506225586, 0.06064332962036133, 0.06061056137084961, 0.060655616760253904, 0.06057984161376953, 0.06073855972290039, 0.06058905410766602, 0.060609535217285154, 0.060590080261230465, 0.06063206481933594, 0.06060543823242188, 0.06066790390014649, 0.060614654541015625, 0.060639232635498044, 0.060641281127929686, 0.06064640045166016, 0.06058700942993164, 0.06070272064208984, 0.06057062530517578, 0.0606033935546875, 0.06059929656982422, 0.060644351959228515, 0.060641281127929686, 0.060624897003173826, 0.06062387084960937, 0.06069760131835938, 0.06053171157836914, 0.06066483306884766, 0.06056755065917969, 0.06059724807739258, 0.060598270416259765, 0.06061363220214844, 0.0606484489440918, 0.060665855407714846]",tokens/s,16.211807874204894,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8265,7 +8265,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8540,7 +8540,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,3047.75168,9028.763648,0.0,8382.316544,7939.414528,s,10,6.222475708007812,0.6222475708007812,0.0004898261264812622,0.6223728332519531,0.6229043151855469,0.6229362579345703,0.622961812133789,"[0.6215844116210938, 0.622897216796875, 0.6224472045898437, 0.622341796875, 0.6226093139648438, 0.6215834350585937, 0.6229682006835937, 0.6217582397460938, 0.6224038696289063, 0.6218820190429688]",tokens/s,411.41181101044583,kWh,7.33948267168469e-06,4.021746173983795e-06,3.6795045775871255e-05,4.815627462153974e-05,tokens/kWh,5316025.834886617,MB,3048.873984,9028.763648,0.0,8382.316544,8233.863168,s,10,338.50034765624997,33.850034765625004,0.0009082101814251832,33.849732421875004,33.851029296875005,33.851565429687504,33.851994335937505,"[33.8521015625, 33.848640625, 33.85091015625, 33.8501171875, 33.84971484375, 33.84975, 33.8495, 33.84959765625, 33.8505546875, 33.8494609375]",tokens/s,1.8611502303086858,kWh,0.0003996493489709166,0.00021904255237639625,0.002000050750366157,0.00261874265171347,tokens/kWh,24057.34674187189,,s,629,343.5634073486327,0.5462057350534704,0.0739750793316962,0.5372897338867187,0.5376120849609375,0.5377253540039063,1.1596667919921875,"[0.5369999389648438, 0.5371217651367187, 0.53671728515625, 0.537164794921875, 0.5368248291015625, 0.537206787109375, 0.53724365234375, 0.5371791381835938, 0.5368934326171875, 0.5374484252929688, 0.5368237915039062, 0.5371473999023437, 0.5371371459960937, 0.5370214233398437, 0.5369026489257812, 0.5371494140625, 0.5369077758789063, 0.53736962890625, 0.5368944702148437, 0.5374730224609375, 0.5372733154296875, 0.53732763671875, 0.5372559204101562, 0.5374003295898437, 0.53711669921875, 0.53743408203125, 0.5371146240234375, 0.5374095458984375, 0.53713818359375, 0.5375703125, 0.5371351318359375, 0.5376665649414063, 0.537343994140625, 0.537660400390625, 0.537238525390625, 0.5394493408203125, 0.5372672119140625, 0.5375682373046875, 0.5374033813476562, 0.5375723266601562, 0.5372518310546875, 0.5374627685546876, 0.5372958984375, 0.5376112670898437, 0.53732763671875, 0.5375897827148437, 0.5374566650390625, 0.537591796875, 0.5373368530273438, 0.537617431640625, 0.5372743530273437, 0.5374218139648438, 0.5373358154296874, 0.5373521728515624, 0.5374238891601563, 0.5377003784179688, 0.5373890380859375, 0.5374044189453125, 0.5368565673828125, 0.5374822387695313, 0.5371740112304687, 0.5375590209960938, 1.160731689453125, 0.5372897338867187, 0.5373317260742188, 0.536974365234375, 0.5370726318359375, 0.537296875, 0.5373296508789063, 0.5368442993164062, 0.5371084594726563, 0.5366640625, 0.5370921020507813, 0.5370153198242188, 0.5370511474609375, 0.5370552368164062, 0.5369251708984375, 0.5369098510742187, 0.537049072265625, 0.5368402099609375, 0.537354248046875, 0.5370050659179687, 0.5376522216796875, 0.5371463623046875, 0.5372323608398437, 0.5373798217773438, 0.5375703125, 0.5374054565429688, 0.5375672607421875, 0.5371392211914062, 0.5372927856445312, 0.5372262573242188, 0.5374105834960937, 0.5371617431640625, 0.5374464111328126, 0.5370101928710938, 0.5375160522460938, 0.5369467163085937, 0.5375999755859375, 0.5370654907226563, 0.5373839111328125, 0.537290771484375, 0.5371781005859375, 0.5372406005859375, 0.5375191040039062, 0.5373685913085937, 0.5378119506835938, 0.5372958984375, 0.5376890869140625, 0.5371627807617188, 0.5374525146484375, 0.5372999877929687, 0.5371555786132812, 0.5372078247070312, 0.5375774536132812, 0.5374822387695313, 0.5375621337890625, 0.5373931274414062, 0.5377433471679688, 0.5372989501953125, 0.5375375366210937, 0.5370828857421875, 0.5374310302734375, 0.5372927856445312, 0.5374443359375, 1.1605975341796875, 0.5371514892578125, 0.53732763671875, 0.5370419311523438, 0.5369763793945312, 0.537027587890625, 0.5371105346679688, 0.53686376953125, 0.5371781005859375, 0.5369999389648438, 0.5372098388671875, 0.5370951538085937, 0.5372415771484375, 0.5368361206054687, 0.53697021484375, 0.5368361206054687, 0.5373788452148438, 0.5369077758789063, 0.5374474487304688, 0.5372006225585938, 0.5376256103515625, 0.5371401977539062, 0.5372794799804688, 0.5370992431640625, 0.5374248657226562, 0.5371719970703125, 0.5375221557617188, 0.5370470581054687, 0.5375098876953125, 0.5372866821289063, 0.5376030883789062, 0.5372876586914063, 0.5373480834960938, 0.5370255126953125, 0.5380044555664063, 0.5371494140625, 0.5373460693359375, 0.5372190551757813, 0.5374095458984375, 0.5371156616210937, 0.5373409423828125, 0.53713818359375, 0.537439208984375, 0.5372620849609375, 0.5375928344726563, 0.5374689331054687, 0.5375150146484375, 0.5372815551757812, 0.5375293579101562, 0.5374801635742188, 0.5372764282226562, 0.5374781494140625, 0.53758154296875, 0.5375764770507813, 0.537491455078125, 0.5374320678710938, 0.537807861328125, 0.537385986328125, 0.537585693359375, 0.537101318359375, 0.5378580322265625, 0.5371678466796875, 0.5375958862304687, 1.160279052734375, 0.5369609985351562, 0.537275390625, 0.5370265502929688, 0.5371607055664063, 0.5369251708984375, 0.5370347290039063, 0.5371637573242187, 0.5371658325195312, 0.5367777099609375, 0.53724365234375, 0.536953857421875, 0.5370439453125, 0.5368780517578124, 0.5370480346679688, 0.536875, 0.537122802734375, 0.5370419311523438, 0.5371986083984375, 0.5371566162109375, 0.5373511962890625, 0.5370081176757813, 0.5376532592773438, 0.5372518310546875, 0.537491455078125, 0.5375631103515625, 0.5374771118164062, 0.537164794921875, 0.5374412841796875, 0.5372344360351563, 0.5374207763671875, 0.537302001953125, 0.5374576416015625, 0.5371043701171875, 0.5376644897460937, 0.5371770629882813, 0.537439208984375, 0.5372200927734375, 0.537670654296875, 0.5372406005859375, 0.5374464111328126, 0.5371873168945313, 0.5375457153320312, 0.5373306884765625, 0.5375682373046875, 0.5374586791992187, 0.5378385620117188, 0.5372211303710938, 0.5372948608398438, 0.5371371459960937, 0.537417724609375, 0.5371791381835938, 0.5375631103515625, 0.537343994140625, 0.5376399536132812, 0.53737060546875, 0.5378252563476562, 0.5372784423828125, 0.5375928344726563, 0.5371002807617188, 0.5377566528320312, 0.5371740112304687, 0.5375252685546875, 1.1604622802734375, 0.53707470703125, 0.5371064453125, 0.5369815063476563, 0.5370859375, 0.537218017578125, 0.5371986083984375, 0.5368463134765625, 0.5371791381835938, 0.5366578979492187, 0.5372927856445312, 0.5372723388671875, 0.5371627807617188, 0.5369517822265625, 0.537027587890625, 0.5370296020507812, 0.53722314453125, 0.53665380859375, 0.5373429565429687, 0.53697021484375, 0.5379092407226562, 0.5369476928710938, 0.5372815551757812, 0.5371320190429687, 0.5373726806640625, 0.5372579956054687, 0.5376419677734375, 0.5369989013671875, 0.537333740234375, 0.5371514892578125, 0.5376256103515625, 0.5371617431640625, 0.5374033813476562, 0.5371043701171875, 0.53758154296875, 0.5370869750976562, 0.53758154296875, 0.5371371459960937, 0.5374607543945312, 0.5373399047851563, 0.53734912109375, 0.5372211303710938, 0.5374893798828125, 0.5373450317382813, 0.5374873657226562, 0.537511962890625, 0.537697265625, 0.5372620849609375, 0.5374576416015625, 0.5373378295898438, 0.5374310302734375, 0.5373788452148438, 0.5375825805664063, 0.537449462890625, 0.5375846557617188, 0.5375979614257812, 0.5378068237304687, 0.5373235473632813, 0.5375047607421874, 0.5369569091796875, 0.5375068359375, 0.5371873168945313, 0.5376583862304688, 1.1593800048828125, 0.5369968872070312, 0.5371678466796875, 0.5368790893554688, 0.5370142822265624, 0.5371074829101562, 0.536890380859375, 0.5370245361328125, 0.5371945190429688, 0.5371904296875, 0.5371555786132812, 0.5369866333007812, 0.5370675048828125, 0.5370357666015625, 0.53704296875, 0.5369476928710938, 0.5370439453125, 0.5368975219726563, 0.5371740112304687, 0.5370562744140625, 0.5374412841796875, 0.5371197509765625, 0.5372723388671875, 0.5372692260742188, 0.53777099609375, 0.5373419799804687, 0.537575439453125, 0.5372262573242188, 0.537449462890625, 0.537206787109375, 0.5375191040039062, 0.537218017578125, 0.5374361572265625, 0.5370224609375, 0.5376153564453126, 0.5371289672851562, 0.5373737182617188, 0.537380859375, 0.53743408203125, 0.5372569580078125, 0.537427978515625, 0.5372313842773437, 0.5373604125976562, 0.5372620849609375, 0.537870361328125, 0.5373726806640625, 0.5377362060546875, 0.5370101928710938, 0.5374207763671875, 0.5372886962890625, 0.5375262451171875, 0.5372723388671875, 0.5374781494140625, 0.5374996337890625, 0.5376716918945312, 0.53737060546875, 0.5378887939453125, 0.5372467041015625, 0.5373880615234375, 0.5372190551757813, 0.5374843139648438, 0.5374290161132812, 0.5376194458007812, 1.1607940673828125, 0.5370654907226563, 0.5372876586914063, 0.537017333984375, 0.53704296875, 0.5369180297851562, 0.5370848999023438, 0.5368524780273437, 0.5371135864257812, 0.5366937866210938, 0.5370224609375, 0.5369159545898438, 0.5372426147460938, 0.5372047119140625, 0.5370357666015625, 0.53686376953125, 0.5372938232421876, 0.5368125610351563, 0.5374832763671875, 0.5372518310546875, 0.5377505493164062, 0.5370930786132813, 0.5374719848632813, 0.5372303466796875, 0.5375682373046875, 0.5374269409179687, 0.5374423217773437, 0.5369999389648438, 0.5372743530273437, 0.53722216796875, 0.5375057983398438, 0.5372692260742188, 0.5375211791992187, 0.5371504516601563, 0.5373726806640625, 0.5371248779296875, 0.5375283203125, 0.5371975708007812, 0.5374423217773437, 0.5372303466796875, 0.5373818969726563, 0.5371207885742187, 0.5374259033203125, 0.5370664672851563, 0.5375396118164063, 0.5372713012695313, 0.5377269897460938, 0.5370449829101562, 0.5375191040039062, 0.5372846069335937, 0.5377986450195312, 0.5370890502929687, 0.5374740600585938, 0.537313232421875, 0.5375160522460938, 0.5373552856445313, 0.53772900390625, 0.53762353515625, 0.5375580444335938, 0.537280517578125, 0.5377146606445312, 0.5373737182617188, 0.5375047607421874, 1.15934716796875, 0.5369712524414062, 0.537196533203125, 0.536880126953125, 0.5370664672851563, 0.5370029907226562, 0.537049072265625, 0.5369415893554688, 0.5371279296875, 0.536838134765625, 0.5372139282226562, 0.5371422729492188, 0.5371709594726563, 0.5370296020507812, 0.5370101928710938, 0.5370931396484375, 0.536943603515625, 0.5370654907226563, 0.5372897338867187, 0.5371084594726563, 0.5374290161132812, 0.5371207885742187, 0.5373665161132812, 0.5372057495117187, 0.5374658813476563, 0.5373778076171875, 0.5378252563476562, 0.5370029907226562, 0.5374115600585937, 0.5372979125976562, 0.5375590209960938, 0.5370470581054687, 0.5373511962890625, 0.5372149658203125, 0.5375795288085937, 0.5372313842773437, 0.537481201171875, 0.5371986083984375, 0.5374371948242187, 0.5373419799804687, 0.5374535522460937, 0.5371392211914062, 0.53747509765625, 0.5372406005859375, 0.5374719848632813, 0.5375580444335938, 0.537575439453125, 0.5373143310546875, 0.537759765625, 0.5374566650390625, 0.5375201416015625, 0.53722216796875, 0.537439208984375, 0.5372764282226562, 0.537491455078125, 0.5373030395507813, 0.5375774536132812, 0.5373572998046875, 0.5374269409179687, 0.5370941162109375, 0.537722900390625, 0.537449462890625, 0.537548828125, 1.1603404541015625, 0.5369415893554688, 0.5372897338867187, 0.5370214233398437, 0.537112548828125, 0.5370777587890625, 0.5371781005859375, 0.5371637573242187, 0.5372262573242188, 0.5365135498046875, 0.5372323608398437, 0.5368197021484375, 0.5370265502929688, 0.5368811645507813, 0.5369989013671875, 0.53697021484375, 0.5371770629882813, 0.5369098510742187, 0.537143310546875, 0.5370941162109375, 0.53747607421875, 0.5370531616210937, 0.5374371948242187, 0.5373184204101562, 0.5373921508789062, 0.5372119140625, 0.5374556274414063, 0.5370972290039062, 0.5374115600585937, 0.537291748046875, 0.537585693359375, 0.5373604125976562, 0.5374586791992187, 0.5370582885742188, 0.537618408203125, 0.5372846069335937, 0.5375836791992188, 0.5370541381835937, 0.5375037231445312, 0.5372927856445312, 0.5374207763671875, 0.5376399536132812, 0.5375559692382812, 0.537228271484375, 0.537797607421875, 0.5371484375, 0.5375344848632813, 0.5373040771484375, 0.5375641479492187, 0.5371904296875, 0.537469970703125, 0.5375662231445313, 0.5374935302734375, 0.537575439453125, 0.5377136840820312, 0.537512939453125, 0.537786376953125, 0.5373778076171875, 0.5373972778320313, 0.537248779296875, 0.5377208251953125, 0.537248779296875, 0.5375057983398438, 1.1597783203125, 0.5370828857421875, 0.5371627807617188, 0.536975341796875, 0.536900634765625, 0.5369651489257813, 0.5374464111328126, 0.5368780517578124, 0.537112548828125, 0.5368463134765625, 0.5373921508789062, 0.5370316772460938, 0.5370029907226562, 0.5370009765625, 0.5370685424804688, 0.5369661254882813, 0.5370408935546875, 0.5369763793945312, 0.5371197509765625, 0.5369815063476563, 0.5374320678710938, 0.5372119140625, 0.5373112182617188, 0.5372886962890625, 0.5373265991210937, 0.5372948608398438, 0.5376635131835937, 0.537280517578125, 0.5374115600585937, 0.537296875, 0.5374033813476562, 0.5372446899414063, 0.53747607421875, 0.5371268920898438, 0.5376399536132812, 0.5370029907226562, 0.537449462890625, 0.5372672119140625, 0.5375283203125, 0.5371187133789063, 0.5373040771484375, 0.5373184204101562, 0.5375262451171875, 0.53732763671875, 0.5376163940429688, 0.537343994140625, 0.5375949096679687, 0.5370869750976562, 0.5373829345703125, 0.5372406005859375, 0.537692138671875, 0.537375732421875, 0.5375538940429687, 0.5374884033203124, 0.5375672607421875, 0.537333740234375, 0.5376993408203125, 0.5373265991210937, 0.5372713012695313, 0.5371146240234375, 0.53762353515625, 0.5374884033203124, 0.537591796875]",tokens/s,1.830811973993839,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8569,7 +8569,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9070,7 +9070,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1856.221184,3368.5504,0.0,2722.103296,2585.357312,s,10,0.7366494750976563,0.07366494750976563,0.00039249671689789645,0.0736723518371582,0.07419290008544921,0.07419320220947266,0.07419344390869141,"[0.0741935043334961, 0.07335641479492187, 0.07308262634277343, 0.07368495941162109, 0.07417286682128907, 0.07325129699707031, 0.07419283294677734, 0.07332640075683594, 0.07372882843017578, 0.07365974426269531]",tokens/s,3475.1942226805027,kWh,8.657666441394562e-07,4.743867474013832e-07,4.119346361168371e-06,5.4594997527092105e-06,tokens/kWh,46890743.03427949,MB,1856.221184,3368.5504,0.0,2722.103296,2628.769792,s,10,38.86898974609376,3.8868989746093745,0.002142874021588634,3.886506591796875,3.8892139404296877,3.890647985839844,3.891795222167969,"[3.89208203125, 3.884625, 3.88758349609375, 3.88604248046875, 3.886436767578125, 3.885375, 3.888895263671875, 3.886576416015625, 3.88693505859375, 3.884438232421875]",tokens/s,16.20829365814206,kWh,4.582713047034936e-05,2.5112934640191582e-05,0.00021524919024462234,0.0002861892553551633,tokens/kWh,220134.050531759,,s,629,39.468826690673836,0.06274853209964042,0.008805392613180215,0.061639678955078124,0.06181806030273437,0.06226268157958985,0.1355356140136719,"[0.0625162239074707, 0.06246809768676758, 0.062304256439208984, 0.06238105773925781, 0.06236467361450195, 0.06246092987060547, 0.062456832885742185, 0.06226432037353516, 0.061895679473876954, 0.06161510467529297, 0.061515777587890626, 0.06170009613037109, 0.06151987075805664, 0.06168473434448242, 0.06160793685913086, 0.061649921417236325, 0.061518848419189455, 0.061574142456054685, 0.062078079223632815, 0.06155251312255859, 0.061625343322753906, 0.061636608123779295, 0.06158028793334961, 0.061625343322753906, 0.06154547119140625, 0.06160076904296875, 0.06159564971923828, 0.06160076904296875, 0.061606910705566405, 0.06164582443237305, 0.061652992248535154, 0.061623294830322264, 0.06156185531616211, 0.06175743865966797, 0.061557758331298826, 0.061669376373291014, 0.06161100769042969, 0.061769729614257814, 0.061658111572265625, 0.06219878387451172, 0.0616703987121582, 0.061666305541992185, 0.061620223999023435, 0.0616734733581543, 0.06161510467529297, 0.06164684677124024, 0.061663230895996096, 0.06168985748291016, 0.06160179138183594, 0.06177382278442383, 0.06160179138183594, 0.06164582443237305, 0.06184550476074219, 0.06161100769042969, 0.061652992248535154, 0.06158848190307617, 0.06170828628540039, 0.06169190216064453, 0.061652992248535154, 0.061575233459472654, 0.061744064331054685, 0.06192236709594726, 0.13562771606445312, 0.06159257507324219, 0.061623294830322264, 0.06195609664916992, 0.06166732788085937, 0.06157209777832031, 0.06155878448486328, 0.06166527938842774, 0.06160793685913086, 0.06156083297729492, 0.061618209838867184, 0.06155465698242187, 0.061639678955078124, 0.06154035186767578, 0.06164787292480469, 0.061722625732421876, 0.06158950424194336, 0.06161305618286133, 0.06162435150146484, 0.06246294403076172, 0.0619683837890625, 0.0616099853515625, 0.06166220855712891, 0.06152601623535156, 0.06164889526367188, 0.06153734588623047, 0.06161401748657227, 0.06158233642578125, 0.06157004928588867, 0.06159769439697266, 0.06160179138183594, 0.06163455963134765, 0.061584384918212894, 0.061551616668701174, 0.061639678955078124, 0.06158848190307617, 0.061601951599121095, 0.06156272125244141, 0.06161612701416016, 0.06156697463989258, 0.06195609664916992, 0.06161510467529297, 0.06171657562255859, 0.061641632080078126, 0.061723648071289064, 0.061636608123779295, 0.06167552185058594, 0.06159564971923828, 0.0615997428894043, 0.061645889282226564, 0.061612991333007815, 0.06159667205810547, 0.06162227249145508, 0.061636608123779295, 0.06173286437988281, 0.061914112091064455, 0.06161305618286133, 0.06163455963134765, 0.06170214462280273, 0.06166425704956055, 0.06158643341064453, 0.06162944030761719, 0.06158540725708008, 0.13527247619628907, 0.061581279754638674, 0.06155878448486328, 0.06152499389648437, 0.06166835021972656, 0.061590526580810545, 0.061517822265625, 0.06159872055053711, 0.061587455749511716, 0.06171955108642578, 0.06164070510864258, 0.061568000793457034, 0.06157721710205078, 0.06150348663330078, 0.06161203384399414, 0.06168169784545898, 0.06163759994506836, 0.061608959197998046, 0.06161407852172852, 0.06154240036010742, 0.06161920166015625, 0.062476287841796874, 0.06166835021972656, 0.06163046264648438, 0.06195814514160156, 0.06177587127685547, 0.0623902702331543, 0.061707263946533204, 0.06162124633789062, 0.0615731201171875, 0.06163251113891602, 0.061565025329589844, 0.061674400329589846, 0.06159564971923828, 0.0616519660949707, 0.06163763046264648, 0.06163455963134765, 0.06162944030761719, 0.0616734733581543, 0.06159667205810547, 0.0616703987121582, 0.061571071624755856, 0.06169702529907226, 0.06159462356567383, 0.06166425704956055, 0.06162636947631836, 0.06164889526367188, 0.06272614288330078, 0.06170521545410156, 0.06171955108642578, 0.06169702529907226, 0.06160281753540039, 0.061656063079833984, 0.06173286437988281, 0.062260223388671876, 0.061813758850097655, 0.061811710357666014, 0.061819969177246095, 0.06171744155883789, 0.061764606475830076, 0.06167244720458984, 0.061655040740966796, 0.061584384918212894, 0.13583973693847656, 0.06158643341064453, 0.06156083297729492, 0.06164070510864258, 0.06155263900756836, 0.06152703857421875, 0.06155059051513672, 0.061587455749511716, 0.06175641632080078, 0.061603839874267576, 0.06168678283691406, 0.061625343322753906, 0.06214553451538086, 0.0615813102722168, 0.06157926559448242, 0.06165913772583008, 0.06169401550292969, 0.06158636856079101, 0.061620223999023435, 0.06160588836669922, 0.06167244720458984, 0.061709312438964846, 0.06159769439697266, 0.061652992248535154, 0.06162124633789062, 0.061636608123779295, 0.061606910705566405, 0.06158950424194336, 0.06163251113891602, 0.061591552734375, 0.06161203384399414, 0.061639678955078124, 0.06177280044555664, 0.06168988800048828, 0.06166422271728516, 0.06160588836669922, 0.06171443176269531, 0.061841407775878904, 0.06170521545410156, 0.061625343322753906, 0.06257766342163086, 0.06173388671875, 0.061682689666748045, 0.061551616668701174, 0.0617154541015625, 0.06164787292480469, 0.06163763046264648, 0.061587455749511716, 0.06166732788085937, 0.06187228775024414, 0.06169481658935547, 0.06167552185058594, 0.06157209777832031, 0.06177996826171875, 0.06175129699707031, 0.06170214462280273, 0.061555744171142575, 0.06177276611328125, 0.06169497680664063, 0.061618175506591794, 0.061625343322753906, 0.06185369491577149, 0.06161920166015625, 0.13637632751464843, 0.062094337463378904, 0.061722625732421876, 0.06158950424194336, 0.06156083297729492, 0.061636608123779295, 0.06153113555908203, 0.061567039489746095, 0.06163654327392578, 0.0615813102722168, 0.0616703987121582, 0.06159360122680664, 0.061620223999023435, 0.061590526580810545, 0.06161305618286133, 0.06163763046264648, 0.06162227249145508, 0.06240870285034179, 0.06161920166015625, 0.06155878448486328, 0.06172671890258789, 0.0615599365234375, 0.06163238525390625, 0.06171136093139649, 0.06158540725708008, 0.06151987075805664, 0.06161407852172852, 0.06159564971923828, 0.0616099853515625, 0.0615997428894043, 0.0616827507019043, 0.061606849670410156, 0.06166527938842774, 0.06172774505615235, 0.06173183822631836, 0.06160793685913086, 0.06163455963134765, 0.061603839874267576, 0.061603839874267576, 0.06159360122680664, 0.06163148880004883, 0.0615731201171875, 0.061639678955078124, 0.06152601623535156, 0.06165302276611328, 0.06166934585571289, 0.06161407852172852, 0.061620223999023435, 0.06162636947631836, 0.06263603210449219, 0.06173593521118164, 0.06169295883178711, 0.061610977172851564, 0.06161305618286133, 0.06162124633789062, 0.06166835021972656, 0.061700126647949216, 0.06224483108520508, 0.06181785583496094, 0.06161612701416016, 0.06189977645874024, 0.061739105224609375, 0.061621150970458984, 0.13536972045898438, 0.06163455963134765, 0.06161731338500977, 0.06172252655029297, 0.06154950332641602, 0.061603839874267576, 0.06166016006469727, 0.061590526580810545, 0.06159769439697266, 0.06160281753540039, 0.06174310302734375, 0.06155878448486328, 0.0615813102722168, 0.06159769439697266, 0.06164582443237305, 0.06159257507324219, 0.061636608123779295, 0.06157209777832031, 0.06174720001220703, 0.06168166351318359, 0.061584384918212894, 0.06155878448486328, 0.06153948974609375, 0.06161494445800781, 0.06175129699707031, 0.06155878448486328, 0.06163251113891602, 0.06156492614746094, 0.061671424865722656, 0.06155878448486328, 0.06164889526367188, 0.0617625617980957, 0.06167859268188477, 0.06308761596679688, 0.061780990600585936, 0.061655040740966796, 0.061669376373291014, 0.06156185531616211, 0.06179635238647461, 0.06155263900756836, 0.061687808990478515, 0.06156595230102539, 0.0617523193359375, 0.061723648071289064, 0.061685760498046874, 0.06160588836669922, 0.0616049919128418, 0.061594497680664065, 0.061568000793457034, 0.061666305541992185, 0.06168371200561523, 0.061661182403564455, 0.061633537292480466, 0.061652992248535154, 0.06165094375610351, 0.06170111846923828, 0.06174515151977539, 0.06168166351318359, 0.06168985748291016, 0.06164684677124024, 0.061617153167724606, 0.06170320129394531, 0.06162019348144531, 0.13634867858886718, 0.061669376373291014, 0.06157619094848633, 0.061677566528320314, 0.06161100769042969, 0.06153526306152344, 0.06168368148803711, 0.06178713607788086, 0.06175539016723633, 0.061590526580810545, 0.061674495697021485, 0.06164377593994141, 0.06158643341064453, 0.06159564971923828, 0.06164377593994141, 0.06164787292480469, 0.061636608123779295, 0.06153420639038086, 0.06174105453491211, 0.06166835021972656, 0.06171443176269531, 0.06157823944091797, 0.061693950653076174, 0.0633589744567871, 0.06182400131225586, 0.06165708923339844, 0.06162124633789062, 0.061742080688476565, 0.06158233642578125, 0.061774848937988285, 0.061881343841552736, 0.061818878173828126, 0.06159360122680664, 0.06161612701416016, 0.06182809448242187, 0.06172876739501953, 0.06173183822631836, 0.061664287567138674, 0.061742080688476565, 0.06162838363647461, 0.061633537292480466, 0.06159462356567383, 0.06164070510864258, 0.061852737426757814, 0.061847488403320314, 0.0616099853515625, 0.06166220855712891, 0.06165708923339844, 0.06167552185058594, 0.061625343322753906, 0.061797374725341796, 0.06176358413696289, 0.06170828628540039, 0.061644798278808595, 0.06157823944091797, 0.061695999145507815, 0.061557758331298826, 0.06290022277832032, 0.06170828628540039, 0.06167552185058594, 0.06157004928588867, 0.061685760498046874, 0.061639678955078124, 0.13560012817382813, 0.06154956817626953, 0.06159257507324219, 0.061664321899414065, 0.06175942230224609, 0.061708446502685546, 0.061510494232177734, 0.061656063079833984, 0.06162636947631836, 0.06158441543579102, 0.061638622283935546, 0.0615546875, 0.06163455963134765, 0.06155168151855469, 0.06161196899414063, 0.06161407852172852, 0.06238719940185547, 0.06159769439697266, 0.06166016006469727, 0.061658111572265625, 0.06163455963134765, 0.06155059051513672, 0.06164684677124024, 0.061532161712646485, 0.061644798278808595, 0.061692928314208986, 0.061917182922363284, 0.061932544708251956, 0.061709312438964846, 0.06181990432739258, 0.06227046585083008, 0.06178303909301758, 0.06163455963134765, 0.06159257507324219, 0.061641727447509766, 0.061695999145507815, 0.061706432342529295, 0.06160879898071289, 0.06169392013549805, 0.061598751068115236, 0.06168060684204101, 0.061641727447509766, 0.06170009613037109, 0.061723648071289064, 0.061636608123779295, 0.061666305541992185, 0.06164582443237305, 0.061813758850097655, 0.06163148880004883, 0.06166220855712891, 0.06170009613037109, 0.06167859268188477, 0.06173491287231445, 0.06168678283691406, 0.061613086700439454, 0.06166217422485352, 0.06156185531616211, 0.061712383270263675, 0.061704193115234375, 0.06167660903930664, 0.06163347244262695, 0.06173388671875, 0.06157926559448242, 0.1358909454345703, 0.061585472106933596, 0.06161913681030273, 0.06158643341064453, 0.06191007995605469, 0.06163654327392578, 0.06151379013061523, 0.06163552093505859, 0.06162739181518555, 0.06161305618286133, 0.061639678955078124, 0.061499393463134766, 0.06164582443237305, 0.06164070510864258, 0.06162944030761719, 0.061644798278808595, 0.06170009613037109, 0.06242816162109375, 0.06164787292480469, 0.06159878540039063, 0.06164985656738281, 0.06165094375610351, 0.06169702529907226, 0.06156595230102539, 0.06164889526367188, 0.06171750259399414, 0.06167244720458984, 0.06157209777832031, 0.06165200042724609, 0.06158233642578125, 0.061629409790039065, 0.06162124633789062, 0.06156288146972656, 0.061585472106933596, 0.06165599822998047, 0.06158540725708008, 0.06170111846923828, 0.06161616134643555, 0.06170723342895508, 0.06167244720458984, 0.06189977645874024, 0.06173183822631836, 0.061659168243408204, 0.06166115188598633, 0.06170431900024414, 0.06161600112915039, 0.061656063079833984, 0.061663230895996096, 0.06173081588745117, 0.06170828628540039, 0.0616734733581543, 0.06315827178955079, 0.06174924850463867, 0.061710334777832034, 0.06160486221313476, 0.06162944030761719, 0.06163558578491211, 0.06182092666625977, 0.06171955108642578, 0.061692928314208986, 0.06157823944091797, 0.061656063079833984, 0.06159769439697266, 0.13598208618164062, 0.06162636947631836, 0.06159462356567383, 0.06156492614746094, 0.06158643341064453, 0.061563934326171875, 0.06156079864501953, 0.06168678283691406, 0.06158339309692383, 0.06161507034301758, 0.06163558578491211, 0.0615618896484375, 0.06162019348144531, 0.061518848419189455, 0.06167244720458984, 0.06163046264648438, 0.06163763046264648, 0.06160588836669922, 0.06163148880004883, 0.06158540725708008, 0.06164889526367188, 0.06164892959594727, 0.061594593048095704, 0.06151168060302734, 0.06163251113891602, 0.06157926559448242, 0.0629678077697754, 0.06161305618286133, 0.06163558578491211, 0.06159360122680664, 0.06161203384399414, 0.061655040740966796, 0.061587455749511716, 0.06164275360107422, 0.06164889526367188, 0.06159872055053711, 0.061656063079833984, 0.06163455963134765, 0.061687808990478515, 0.0616099853515625, 0.061692928314208986, 0.061608959197998046, 0.06170111846923828, 0.06162739181518555, 0.06164889526367188, 0.06160179138183594, 0.06163251113891602, 0.06161407852172852, 0.061652030944824215, 0.061599681854248044, 0.06173183822631836, 0.06154444885253906, 0.06162636947631836, 0.061669376373291014, 0.06157622528076172, 0.06165910339355469, 0.06154550552368164, 0.06167855834960938, 0.06169702529907226, 0.061652992248535154, 0.061699073791503904, 0.06180044937133789, 0.06164179229736328]",tokens/s,15.936627782974552,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1479.114752,1841.823744,0.0,1195.37664,1022.155264,s,10,0.6871012496948242,0.06871012496948242,0.00033873627735752553,0.06860185623168946,0.06884490509033203,0.06926672515869141,0.06960418121337891,"[0.06968854522705079, 0.06850118255615234, 0.06852729797363281, 0.06862016296386719, 0.06869097900390625, 0.06858354949951172, 0.06851074981689453, 0.06875116729736327, 0.06849430084228515, 0.06873331451416016]",tokens/s,3725.797327740305,kWh,8.094942381483598e-07,4.4356150458715745e-07,3.836808624999982e-06,5.089864367735499e-06,tokens/kWh,50296035.710259095,MB,1479.114752,1841.823744,0.0,1195.37664,1091.597312,s,10,35.17510473632813,3.517510473632812,0.0019858794372678286,3.5167731933593753,3.5191181640625,3.52109130859375,3.5226698242187497,"[3.523064453125, 3.5170556640625, 3.51595458984375, 3.516420166015625, 3.516550048828125, 3.5174658203125, 3.51656103515625, 3.516367919921875, 3.5186796875, 3.5169853515625]",tokens/s,17.91039443158641,kWh,4.1521001642561397e-05,2.275550190917111e-05,0.0001961486569188012,0.00026042516047053367,tokens/kWh,241912.10974458922,,s,629,35.73639368057251,0.05681461634431242,0.008170033415171557,0.05581414413452149,0.05589258193969727,0.055985150909423825,0.12453187377929688,"[0.05623807907104492, 0.056237056732177736, 0.056357887268066405, 0.056323070526123044, 0.05624934387207031, 0.05632819366455078, 0.05631078338623047, 0.056376319885253906, 0.05601279830932617, 0.05586739349365234, 0.05577318572998047, 0.05581107330322266, 0.055946239471435545, 0.05584998321533203, 0.05613875198364258, 0.05595852661132812, 0.05590220642089844, 0.05599334335327148, 0.055934974670410156, 0.05595238494873047, 0.05628211212158203, 0.05583871841430664, 0.05575987243652344, 0.055757823944091796, 0.055818241119384764, 0.055790592193603515, 0.0557946891784668, 0.05578649520874023, 0.05578035354614258, 0.05583871841430664, 0.055836673736572265, 0.055803905487060546, 0.055874561309814455, 0.0558919677734375, 0.05581721496582031, 0.05585919952392578, 0.055828479766845705, 0.05583871841430664, 0.055795711517333986, 0.05586636734008789, 0.055826431274414064, 0.055774208068847655, 0.055801856994628904, 0.05579776000976563, 0.055820289611816405, 0.05580083084106445, 0.0557496337890625, 0.05584281539916992, 0.0558766098022461, 0.05587865447998047, 0.05584281539916992, 0.055782398223876956, 0.055809024810791016, 0.055788543701171874, 0.05576499176025391, 0.05582950210571289, 0.055911422729492184, 0.0558551025390625, 0.05586739349365234, 0.05582745742797852, 0.055890945434570315, 0.05582438278198242, 0.12451123046875, 0.05576396942138672, 0.05577523040771484, 0.05576499176025391, 0.05576806259155274, 0.05575475311279297, 0.05576499176025391, 0.0557916145324707, 0.05580799865722656, 0.05574655914306641, 0.05579673767089844, 0.05575987243652344, 0.055812095642089846, 0.055739391326904295, 0.05584384155273438, 0.0557209587097168, 0.05575372695922851, 0.055790592193603515, 0.05578342437744141, 0.05578035354614258, 0.05590323257446289, 0.055731201171875, 0.05590630340576172, 0.055772159576416014, 0.05581721496582031, 0.05579673767089844, 0.05584998321533203, 0.055812095642089846, 0.0557946891784668, 0.05591449737548828, 0.05588684844970703, 0.055831550598144535, 0.05584998321533203, 0.05613363265991211, 0.05586227035522461, 0.055790592193603515, 0.05580799865722656, 0.05578649520874023, 0.05587046432495117, 0.0557946891784668, 0.05585715103149414, 0.0557916145324707, 0.05584691238403321, 0.05583257675170898, 0.05586329650878906, 0.05585919952392578, 0.05588275146484375, 0.055853057861328125, 0.05580083084106445, 0.05595340728759766, 0.05588275146484375, 0.0557844467163086, 0.05586739349365234, 0.05587046432495117, 0.055825408935546876, 0.05580595016479492, 0.05588172912597656, 0.05586841583251953, 0.055831550598144535, 0.055839744567871094, 0.05582233428955078, 0.05582131195068359, 0.05584076690673828, 0.12454399871826172, 0.055744510650634765, 0.055806976318359375, 0.05572710418701172, 0.055826431274414064, 0.05572608184814453, 0.05575987243652344, 0.05573222351074219, 0.055831550598144535, 0.055772159576416014, 0.055736320495605465, 0.05578649520874023, 0.055790592193603515, 0.055723007202148435, 0.05581414413452149, 0.05574348831176758, 0.055772159576416014, 0.05571583938598633, 0.05583462524414062, 0.05577830505371094, 0.055772159576416014, 0.05574655914306641, 0.05588479995727539, 0.055787521362304686, 0.055787521362304686, 0.055798782348632815, 0.05585715103149414, 0.05578956985473633, 0.05578342437744141, 0.055831550598144535, 0.055812095642089846, 0.05575065612792969, 0.05581107330322266, 0.05577318572998047, 0.05585100936889648, 0.055803905487060546, 0.055812095642089846, 0.0557916145324707, 0.05586227035522461, 0.05580287933349609, 0.055769088745117185, 0.055812095642089846, 0.05584998321533203, 0.055766014099121096, 0.05582438278198242, 0.05583769607543945, 0.055818241119384764, 0.05590118408203125, 0.05585715103149414, 0.055828479766845705, 0.05595033645629883, 0.055844863891601565, 0.05581107330322266, 0.05580083084106445, 0.0558653450012207, 0.05584281539916992, 0.05595443344116211, 0.05584998321533203, 0.05580799865722656, 0.055853057861328125, 0.05577318572998047, 0.055803905487060546, 0.055818241119384764, 0.12458598327636719, 0.05576806259155274, 0.05575680160522461, 0.0557916145324707, 0.05617663955688477, 0.05577523040771484, 0.0557496337890625, 0.055771137237548826, 0.0557844467163086, 0.05573427200317383, 0.05581926345825195, 0.055929855346679686, 0.0558551025390625, 0.055806976318359375, 0.055798782348632815, 0.055728126525878906, 0.0557946891784668, 0.05577318572998047, 0.055744510650634765, 0.0557762565612793, 0.055772159576416014, 0.0557496337890625, 0.055841793060302736, 0.055755775451660154, 0.055769088745117185, 0.055782398223876956, 0.05583257675170898, 0.05581619262695312, 0.05586022567749024, 0.05580595016479492, 0.05581926345825195, 0.05580083084106445, 0.05580595016479492, 0.05582950210571289, 0.055815166473388675, 0.055844863891601565, 0.05583462524414062, 0.05582745742797852, 0.05579776000976563, 0.05574553680419922, 0.0557844467163086, 0.055785472869873044, 0.055795711517333986, 0.0557844467163086, 0.05582233428955078, 0.05586227035522461, 0.05587148666381836, 0.05582131195068359, 0.0559370231628418, 0.055825408935546876, 0.05587148666381836, 0.05581414413452149, 0.05578649520874023, 0.055856128692626954, 0.05586739349365234, 0.055809024810791016, 0.055826431274414064, 0.05580799865722656, 0.05578956985473633, 0.055820289611816405, 0.05583359909057617, 0.055810047149658204, 0.05586329650878906, 0.12467711639404297, 0.055820289611816405, 0.055785472869873044, 0.05572403335571289, 0.055782398223876956, 0.05574758529663086, 0.055757823944091796, 0.05576294326782227, 0.05584281539916992, 0.05578035354614258, 0.055801856994628904, 0.05575884628295898, 0.05578342437744141, 0.05572403335571289, 0.055812095642089846, 0.0557496337890625, 0.0562718734741211, 0.05581926345825195, 0.055861248016357425, 0.05574041748046875, 0.055828479766845705, 0.05575167846679688, 0.0558919677734375, 0.055742462158203124, 0.05580083084106445, 0.055785472869873044, 0.05584793472290039, 0.055769088745117185, 0.055836673736572265, 0.05579673767089844, 0.05579264068603516, 0.05577318572998047, 0.05582233428955078, 0.05581926345825195, 0.055810047149658204, 0.0557762565612793, 0.05582438278198242, 0.05582438278198242, 0.05586739349365234, 0.05574860763549805, 0.055815166473388675, 0.05576294326782227, 0.05587558364868164, 0.05578137588500977, 0.055825408935546876, 0.055771137237548826, 0.05580595016479492, 0.0557946891784668, 0.055907329559326174, 0.05586329650878906, 0.05581721496582031, 0.055972862243652347, 0.05586227035522461, 0.05586636734008789, 0.055795711517333986, 0.0557916145324707, 0.05586636734008789, 0.055795711517333986, 0.05584588623046875, 0.05582233428955078, 0.05583052825927735, 0.05580083084106445, 0.05589503860473633, 0.12453990173339843, 0.05574041748046875, 0.05575680160522461, 0.055809024810791016, 0.0557844467163086, 0.05577011108398437, 0.056180736541748044, 0.05583257675170898, 0.05579673767089844, 0.05576704025268555, 0.0558551025390625, 0.05578137588500977, 0.05581107330322266, 0.055779327392578126, 0.055820289611816405, 0.055755775451660154, 0.0557844467163086, 0.05576294326782227, 0.055777278900146485, 0.05581619262695312, 0.05575680160522461, 0.05578956985473633, 0.05586329650878906, 0.055803905487060546, 0.05586022567749024, 0.055777278900146485, 0.055766014099121096, 0.05580492782592773, 0.055858177185058595, 0.055806976318359375, 0.055831550598144535, 0.055782398223876956, 0.055772159576416014, 0.05580083084106445, 0.05586943817138672, 0.055809024810791016, 0.05583769607543945, 0.05579264068603516, 0.0557946891784668, 0.055731201171875, 0.05578137588500977, 0.055809024810791016, 0.05582131195068359, 0.05595238494873047, 0.05582745742797852, 0.055853057861328125, 0.05588787078857422, 0.05581619262695312, 0.055787521362304686, 0.05583257675170898, 0.05588787078857422, 0.05584691238403321, 0.055831550598144535, 0.055856128692626954, 0.05627699279785156, 0.0558766098022461, 0.05584588623046875, 0.055831550598144535, 0.05580799865722656, 0.05583052825927735, 0.055809024810791016, 0.05583564758300781, 0.05584691238403321, 0.12479897308349609, 0.05577830505371094, 0.05582438278198242, 0.05576192092895508, 0.05574655914306641, 0.05575987243652344, 0.0557916145324707, 0.05576192092895508, 0.05581619262695312, 0.05576499176025391, 0.05581107330322266, 0.05576499176025391, 0.055831550598144535, 0.055787521362304686, 0.05582950210571289, 0.05572710418701172, 0.05584076690673828, 0.05575680160522461, 0.055825408935546876, 0.055723007202148435, 0.055841793060302736, 0.05574860763549805, 0.055815166473388675, 0.055744510650634765, 0.055820289611816405, 0.05578342437744141, 0.055858177185058595, 0.055798782348632815, 0.05581107330322266, 0.055782398223876956, 0.05587251281738281, 0.0558551025390625, 0.05586431884765625, 0.05585715103149414, 0.055812095642089846, 0.05580287933349609, 0.05588275146484375, 0.05590323257446289, 0.05586431884765625, 0.05576294326782227, 0.05586022567749024, 0.055760894775390625, 0.05585203170776367, 0.05575884628295898, 0.05586431884765625, 0.055777278900146485, 0.055823360443115234, 0.05583052825927735, 0.055874561309814455, 0.05587148666381836, 0.055790592193603515, 0.05584588623046875, 0.05584588623046875, 0.05587558364868164, 0.05578342437744141, 0.055825408935546876, 0.05590220642089844, 0.05594828796386719, 0.05585919952392578, 0.055787521362304686, 0.05580595016479492, 0.05581107330322266, 0.05586943817138672, 0.12475801849365234, 0.0557496337890625, 0.05581926345825195, 0.05589811325073242, 0.05580287933349609, 0.05578137588500977, 0.055772159576416014, 0.05578137588500977, 0.05575475311279297, 0.05574758529663086, 0.05581414413452149, 0.055793663024902344, 0.05579776000976563, 0.055806976318359375, 0.05585203170776367, 0.05576704025268555, 0.05576499176025391, 0.055812095642089846, 0.05581926345825195, 0.05579776000976563, 0.05581619262695312, 0.0557844467163086, 0.05584588623046875, 0.055787521362304686, 0.05582438278198242, 0.05579980850219726, 0.055790592193603515, 0.055790592193603515, 0.05583052825927735, 0.05582438278198242, 0.05583769607543945, 0.05579980850219726, 0.05581414413452149, 0.05580595016479492, 0.055826431274414064, 0.05581619262695312, 0.05584691238403321, 0.05578035354614258, 0.05590630340576172, 0.055806976318359375, 0.05579264068603516, 0.05581107330322266, 0.055844863891601565, 0.0557946891784668, 0.05581107330322266, 0.055809024810791016, 0.05588275146484375, 0.055815166473388675, 0.055785472869873044, 0.05583462524414062, 0.0558837776184082, 0.05583871841430664, 0.055788543701171874, 0.05585203170776367, 0.055810047149658204, 0.055790592193603515, 0.05580287933349609, 0.055844863891601565, 0.05579673767089844, 0.055831550598144535, 0.055782398223876956, 0.05583052825927735, 0.05590016174316406, 0.12446822357177735, 0.055798782348632815, 0.055782398223876956, 0.05577011108398437, 0.05578035354614258, 0.05575167846679688, 0.05574655914306641, 0.05576396942138672, 0.05580287933349609, 0.05576192092895508, 0.055820289611816405, 0.05574758529663086, 0.05579673767089844, 0.055725055694580077, 0.055809024810791016, 0.055728126525878906, 0.055841793060302736, 0.055744510650634765, 0.05581619262695312, 0.05592166519165039, 0.05584998321533203, 0.05577830505371094, 0.055877632141113284, 0.05575680160522461, 0.05580492782592773, 0.05580492782592773, 0.05587865447998047, 0.055777278900146485, 0.05586431884765625, 0.05670809555053711, 0.05588684844970703, 0.05614080047607422, 0.055812095642089846, 0.05583359909057617, 0.05583769607543945, 0.05582950210571289, 0.05588582229614258, 0.05592268753051758, 0.056118270874023435, 0.05578342437744141, 0.055861248016357425, 0.056174591064453126, 0.05588172912597656, 0.055752704620361325, 0.05583564758300781, 0.05581414413452149, 0.05596979141235352, 0.05585408020019531, 0.05584384155273438, 0.05587558364868164, 0.05581721496582031, 0.05582950210571289, 0.05588479995727539, 0.05593804931640625, 0.05578956985473633, 0.0557844467163086, 0.05588684844970703, 0.05579980850219726, 0.05584691238403321, 0.05581926345825195, 0.0558131217956543, 0.055779327392578126, 0.055839744567871094, 0.12470374298095703, 0.055769088745117185, 0.05576396942138672, 0.055777278900146485, 0.05578342437744141, 0.05576806259155274, 0.05575372695922851, 0.055731201171875, 0.055752704620361325, 0.05572403335571289, 0.05583769607543945, 0.05580083084106445, 0.055795711517333986, 0.05576396942138672, 0.05582233428955078, 0.05577011108398437, 0.0557916145324707, 0.05578035354614258, 0.05583052825927735, 0.05581619262695312, 0.05581414413452149, 0.05578956985473633, 0.055841793060302736, 0.05577830505371094, 0.055798782348632815, 0.055806976318359375, 0.05582745742797852, 0.05582131195068359, 0.05592575836181641, 0.055820289611816405, 0.05586022567749024, 0.055810047149658204, 0.0557844467163086, 0.055815166473388675, 0.0558551025390625, 0.05585203170776367, 0.055874561309814455, 0.055790592193603515, 0.055785472869873044, 0.05573222351074219, 0.05581721496582031, 0.05588479995727539, 0.05584076690673828, 0.05581107330322266, 0.05585203170776367, 0.05581414413452149, 0.0559370231628418, 0.05584384155273438, 0.05582438278198242, 0.05584588623046875, 0.05587148666381836, 0.05585203170776367, 0.055803905487060546, 0.05607219314575195, 0.05584076690673828, 0.055806976318359375, 0.055844863891601565, 0.05585919952392578, 0.05580799865722656, 0.05607731246948242, 0.055861248016357425, 0.05581619262695312, 0.05583564758300781]",tokens/s,17.601104510496405,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9983,7 +9983,7 @@ If this is a private repository, make sure to pass a token having permission to 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1506.42688,1579.679744,0.0,933.23264,856.69632,s,10,0.43564921951293945,0.043564921951293946,0.0031757046407324095,0.04331248092651367,0.04437063980102539,0.0485523281097412,0.051897678756713875,"[0.052734016418457035, 0.04333430480957031, 0.04164771270751953, 0.043380638122558594, 0.0414692497253418, 0.04140153503417969, 0.04329619216918945, 0.04344137573242188, 0.04332876968383789, 0.04161542510986328]",tokens/s,5876.287355368403,kWh,4.969496102560134e-07,2.72305043992154e-07,1.379101243335609e-06,2.1483558975837764e-06,tokens/kWh,119160889.63095888,MB,1506.42688,1579.679744,0.0,933.23264,899.62496,s,10,24.820336181640624,2.4820336181640625,0.03505314468873128,2.4815734863281254,2.524392333984375,2.52884619140625,2.53240927734375,"[2.533300048828125, 2.512392822265625, 2.510242431640625, 2.523402587890625, 2.427984130859375, 2.441885986328125, 2.461017822265625, 2.4780390625, 2.48510791015625, 2.44696337890625]",tokens/s,25.382412042670286,kWh,2.9558689719627776e-05,1.6199240900285592e-05,7.306973609346903e-05,0.00011882766671338238,tokens/kWh,530179.5595461687,,s,629,25.16415895462036,0.04000661201052522,0.005123259677494279,0.03999129486083984,0.040338432312011716,0.04051681213378906,0.07998427032470704,"[0.04113715362548828, 0.041027584075927735, 0.041578495025634765, 0.04084121704101563, 0.04151398468017578, 0.041063423156738284, 0.04110131072998047, 0.0394700813293457, 0.039803905487060545, 0.03899084854125977, 0.03885260772705078, 0.04075110244750976, 0.04097536087036133, 0.04030361557006836, 0.04015513610839844, 0.04008652877807617, 0.0400261116027832, 0.04022272109985352, 0.04007628631591797, 0.040078399658203125, 0.04018272018432617, 0.040068096160888675, 0.040422401428222655, 0.04001484680175781, 0.04015411376953125, 0.04017356872558594, 0.04025958251953125, 0.04034048080444336, 0.04012134552001953, 0.04011212921142578, 0.04016537475585937, 0.040008705139160154, 0.040130561828613284, 0.04007731246948242, 0.04001484680175781, 0.04027699279785156, 0.03999129486083984, 0.04007731246948242, 0.040084480285644535, 0.0402503662109375, 0.04009983825683594, 0.04006399917602539, 0.04001279830932617, 0.040207359313964845, 0.04010092926025391, 0.04016838455200195, 0.04013158416748047, 0.040172542572021484, 0.040498241424560544, 0.04003936004638672, 0.040325119018554685, 0.04010291290283203, 0.040225791931152347, 0.04005376052856445, 0.04014796829223633, 0.039678016662597654, 0.04023392105102539, 0.04024422454833984, 0.03986227035522461, 0.04005376052856445, 0.040235008239746094, 0.04025446319580078, 0.08356147003173828, 0.04013158416748047, 0.040151039123535154, 0.04020537567138672, 0.04016428756713867, 0.03999334335327148, 0.040237056732177735, 0.0403394546508789, 0.04013260650634766, 0.04007219314575195, 0.04012236785888672, 0.04005785751342773, 0.03998822402954102, 0.040136703491210936, 0.040052734375, 0.040161312103271486, 0.03975164794921875, 0.040303680419921876, 0.04009260940551758, 0.04006707382202149, 0.04007628631591797, 0.04026265716552734, 0.04016128158569336, 0.040051712036132815, 0.040156158447265625, 0.04031795120239258, 0.04012748718261719, 0.04016230392456055, 0.03998617553710938, 0.04028518295288086, 0.04020633697509766, 0.04005068969726563, 0.0399288330078125, 0.040256511688232424, 0.039979007720947264, 0.04008550262451172, 0.04009164810180664, 0.04038860702514648, 0.040223743438720705, 0.04016844940185547, 0.04011724853515625, 0.04027699279785156, 0.04012441635131836, 0.04024729537963867, 0.040101886749267575, 0.04006399917602539, 0.04013158416748047, 0.04007526397705078, 0.039964672088623046, 0.04025958251953125, 0.04006707382202149, 0.040025089263916014, 0.04005068969726563, 0.03987968063354492, 0.03824639892578125, 0.038312961578369144, 0.03836006546020508, 0.038322208404541015, 0.03826992034912109, 0.03886899185180664, 0.03874611282348633, 0.0382751350402832, 0.038287296295166015, 0.08017305755615234, 0.03826483154296875, 0.038435840606689455, 0.03826073455810547, 0.03853107070922852, 0.04071014404296875, 0.03881062316894531, 0.03902771377563476, 0.040190975189208986, 0.04033331298828125, 0.04007526397705078, 0.04009471893310547, 0.0400711669921875, 0.04029747009277344, 0.03996364974975586, 0.040139774322509765, 0.04016128158569336, 0.04015206527709961, 0.04011315155029297, 0.040158206939697266, 0.039951358795166016, 0.04007526397705078, 0.04027699279785156, 0.03998720169067383, 0.04024729537963867, 0.03996672058105469, 0.0404398078918457, 0.040035327911376956, 0.040134654998779294, 0.04028108978271484, 0.03998822402954102, 0.039959552764892575, 0.040130561828613284, 0.04015411376953125, 0.04040499114990234, 0.04033433532714844, 0.04011929702758789, 0.040313854217529296, 0.04006614303588867, 0.039992225646972655, 0.04012543869018555, 0.04003942489624023, 0.04003635025024414, 0.04030976104736328, 0.04017561721801758, 0.03996364974975586, 0.04004048156738281, 0.04014076614379883, 0.040002559661865236, 0.04034969711303711, 0.04014284896850586, 0.04008345413208008, 0.04016844940185547, 0.03999948883056641, 0.039951358795166016, 0.03837644958496094, 0.03825766372680664, 0.03821158218383789, 0.038324222564697266, 0.039932926177978514, 0.040046592712402344, 0.04010291290283203, 0.04031283187866211, 0.08356454467773437, 0.040120319366455076, 0.040197120666503904, 0.04001484680175781, 0.04011520004272461, 0.0399738883972168, 0.04001587295532227, 0.040022014617919925, 0.040030208587646485, 0.03988684844970703, 0.040529918670654294, 0.04023910522460938, 0.039923713684082034, 0.0401797103881836, 0.04036505508422852, 0.040136703491210936, 0.039977985382080077, 0.04037222290039062, 0.040443904876708986, 0.04031999969482422, 0.04006092834472656, 0.04047359848022461, 0.040342529296875, 0.04001279830932617, 0.040022014617919925, 0.040134654998779294, 0.04007219314575195, 0.04011212921142578, 0.040000511169433595, 0.04015209579467773, 0.04024316787719726, 0.0400711669921875, 0.04026265716552734, 0.04001792144775391, 0.04007526397705078, 0.0404398078918457, 0.04011724853515625, 0.04022784042358398, 0.040081409454345705, 0.04112998580932617, 0.04065484619140625, 0.04024524688720703, 0.04050841522216797, 0.04051763153076172, 0.04026163101196289, 0.040182785034179686, 0.04026572799682617, 0.04014080047607422, 0.040338432312011716, 0.04013363265991211, 0.04012236785888672, 0.039981056213378906, 0.04031488037109375, 0.04002304077148437, 0.04011520004272461, 0.040033279418945314, 0.04029849624633789, 0.040041473388671874, 0.03832729721069336, 0.038545406341552735, 0.03859763336181641, 0.03841331100463867, 0.03827199935913086, 0.07982694244384765, 0.03820544052124023, 0.0380579833984375, 0.03835903930664063, 0.038199295043945314, 0.03845939254760742, 0.04019404983520508, 0.04046745681762695, 0.03990630340576172, 0.04013363265991211, 0.03996364974975586, 0.03836006546020508, 0.03817062377929688, 0.038171646118164065, 0.0381921272277832, 0.03843379211425781, 0.03821670532226563, 0.03819007873535156, 0.03816755294799805, 0.038141952514648435, 0.038299648284912106, 0.038242305755615234, 0.038491134643554685, 0.038255615234375, 0.03834982299804687, 0.038201343536376955, 0.03825151824951172, 0.038152191162109376, 0.03830886459350586, 0.03816755294799805, 0.038171646118164065, 0.03821263885498047, 0.03831600189208984, 0.03889561462402344, 0.03994112014770508, 0.04019302368164063, 0.03997183990478516, 0.04004761505126953, 0.03841843032836914, 0.03847679901123047, 0.038371326446533204, 0.038234111785888675, 0.03837849426269531, 0.038258689880371094, 0.03819007873535156, 0.03830886459350586, 0.03834982299804687, 0.03825254440307617, 0.03839590454101562, 0.038346752166748044, 0.03817779159545898, 0.03827199935913086, 0.03827609634399414, 0.03823311996459961, 0.03817987060546875, 0.03817059326171875, 0.0383221435546875, 0.038166526794433595, 0.03820646286010742, 0.03827609634399414, 0.03827507019042969, 0.038204414367675785, 0.03846553421020508, 0.07984639739990235, 0.03822284698486328, 0.03814604949951172, 0.038381568908691405, 0.038329345703125, 0.03812659072875976, 0.0381030387878418, 0.0381921272277832, 0.03803033447265625, 0.03826483154296875, 0.03832831954956055, 0.038163455963134765, 0.038128639221191404, 0.03808256149291992, 0.03830476760864258, 0.038152191162109376, 0.03814912033081055, 0.03819724655151367, 0.03816150283813476, 0.03834255981445313, 0.038214656829833986, 0.0388587532043457, 0.03990937423706055, 0.03986841583251953, 0.04010291290283203, 0.040019966125488284, 0.03821670532226563, 0.038198272705078126, 0.03840409469604492, 0.038588417053222655, 0.03825664138793945, 0.038371326446533204, 0.03809075164794922, 0.03829862213134766, 0.03820236968994141, 0.038201343536376955, 0.038217727661132815, 0.038234111785888675, 0.038209537506103515, 0.03817574310302734, 0.038247425079345705, 0.0382371826171875, 0.03834470367431641, 0.03826892852783203, 0.03811840057373047, 0.03825459289550781, 0.03821363067626953, 0.039229438781738284, 0.04001279830932617, 0.040089599609375, 0.0400076789855957, 0.040134654998779294, 0.040187904357910156, 0.040088607788085935, 0.039887840270996094, 0.04031283187866211, 0.04015411376953125, 0.04027084732055664, 0.04039273452758789, 0.03996566390991211, 0.03829145431518555, 0.03822079849243164, 0.04009471893310547, 0.08352470397949219, 0.0401376953125, 0.040349632263183596, 0.04077977752685547, 0.04024524688720703, 0.04009676742553711, 0.040515583038330076, 0.040248321533203124, 0.04014796829223633, 0.04009574508666992, 0.04011929702758789, 0.040210430145263674, 0.040223743438720705, 0.0399738883972168, 0.04003839874267578, 0.0382371826171875, 0.038629375457763675, 0.038491134643554685, 0.03825766372680664, 0.03826483154296875, 0.03846656036376953, 0.038354942321777344, 0.038373374938964845, 0.03835289764404297, 0.03834265518188477, 0.03841535949707031, 0.03825664138793945, 0.038163455963134765, 0.038193153381347655, 0.038416385650634766, 0.03832217788696289, 0.03832320022583008, 0.0391464958190918, 0.040323070526123043, 0.04012851333618164, 0.04029849624633789, 0.04007526397705078, 0.040306751251220706, 0.03839788818359375, 0.0381399040222168, 0.038373374938964845, 0.03852492904663086, 0.038386688232421876, 0.038383617401123046, 0.03821670532226563, 0.03838054275512695, 0.03832012939453125, 0.038214656829833986, 0.03820032119750977, 0.038507518768310545, 0.03825664138793945, 0.03829145431518555, 0.038245376586914064, 0.03831193542480469, 0.038317054748535154, 0.0382740478515625, 0.038214656829833986, 0.038345729827880856, 0.03892124938964844, 0.03984790420532226, 0.04002918243408203, 0.040172542572021484, 0.04021247863769531, 0.08194969940185547, 0.03841331100463867, 0.038258689880371094, 0.038373374938964845, 0.038317054748535154, 0.038228992462158204, 0.03824127960205078, 0.038424575805664066, 0.03839487838745117, 0.038250495910644534, 0.038561790466308594, 0.03838873672485352, 0.04093439865112305, 0.040123390197753905, 0.04016742324829101, 0.040235008239746094, 0.04011212921142578, 0.04035276794433594, 0.04093132781982422, 0.03868569564819336, 0.038564895629882814, 0.0386354866027832, 0.03827199935913086, 0.038193153381347655, 0.038217727661132815, 0.038247425079345705, 0.03840518569946289, 0.038314945220947264, 0.03830988693237305, 0.03851468658447266, 0.03837952041625976, 0.03823308944702149, 0.03839184188842774, 0.0384859848022461, 0.0382105598449707, 0.0382740478515625, 0.03916595077514649, 0.04033740615844727, 0.038247425079345705, 0.03826176071166992, 0.039144447326660156, 0.041452545166015625, 0.040622081756591794, 0.04063641738891602, 0.040891391754150394, 0.04014591979980469, 0.040700927734375, 0.04041318511962891, 0.040308734893798825, 0.04019404983520508, 0.03999129486083984, 0.04006707382202149, 0.040828929901123044, 0.04010905456542969, 0.04021145629882812, 0.04004249572753906, 0.04000460815429688, 0.04015718460083008, 0.04007014465332031, 0.04011315155029297, 0.03992473602294922, 0.040338432312011716, 0.04014899063110351, 0.08364339447021485, 0.03997491073608399, 0.04009267044067383, 0.03998207855224609, 0.040376319885253906, 0.04017766571044922, 0.04027494430541992, 0.04038451385498047, 0.0402503662109375, 0.04045414352416992, 0.04019814300537109, 0.0397916145324707, 0.03818700790405274, 0.0382105598449707, 0.03836108779907227, 0.03888435363769531, 0.04011929702758789, 0.040153087615966795, 0.039880702972412106, 0.03999846267700195, 0.04047872161865235, 0.040325119018554685, 0.040395774841308595, 0.040190975189208986, 0.040041473388671874, 0.040049663543701174, 0.03856281661987305, 0.0383375358581543, 0.038228992462158204, 0.038348800659179685, 0.03831398391723633, 0.03893657684326172, 0.040150016784667966, 0.04014796829223633, 0.03996876907348633, 0.040226814270019534, 0.03987865447998047, 0.04001792144775391, 0.040079360961914064, 0.04005376052856445, 0.040172542572021484, 0.04024422454833984, 0.04010291290283203, 0.04010905456542969, 0.0401162223815918, 0.040218624114990234, 0.03831808090209961, 0.03854848098754883, 0.038324222564697266, 0.038430721282958984, 0.038348800659179685, 0.03827302551269531, 0.03822796630859375, 0.039934974670410156, 0.04017766571044922, 0.038391807556152346, 0.038354942321777344, 0.03833446502685547, 0.03852799987792969, 0.03838771057128906, 0.038245376586914064, 0.03821875381469726, 0.03834163284301758, 0.08003788757324219, 0.0383631362915039, 0.03848601531982422, 0.03860172653198242, 0.038424575805664066, 0.03824127960205078, 0.03828224182128906, 0.03827814483642578, 0.03832217788696289, 0.038422527313232424, 0.03840921783447265, 0.0383559684753418, 0.03847782516479492, 0.038231040954589846, 0.03817375946044922, 0.038246337890625, 0.03848294448852539, 0.038414337158203124, 0.038351871490478515, 0.03842764663696289, 0.038779937744140625, 0.0384532470703125, 0.038394847869873044, 0.038340606689453126, 0.03857612609863281, 0.0383375358581543, 0.03840716934204102, 0.03837855911254883, 0.03835078430175781, 0.038474750518798825, 0.03840921783447265, 0.038712318420410154, 0.0397844467163086, 0.04009881591796875, 0.04036812973022461, 0.0402606086730957, 0.04012441635131836, 0.04048793411254883, 0.04021350479125976, 0.04022886276245117, 0.04016025543212891, 0.03828326416015625, 0.03829350280761719, 0.038520832061767575, 0.03837747192382813, 0.038302719116210936, 0.03847884750366211, 0.03830476760864258, 0.038267902374267575, 0.038665214538574216, 0.0383744010925293, 0.038373374938964845, 0.038384639739990234, 0.038225921630859375, 0.03869081497192383, 0.038553600311279294, 0.03964313507080078, 0.04035276794433594, 0.03975884628295898, 0.04010700988769531, 0.040389633178710936, 0.03976704025268555, 0.04012748718261719]",tokens/s,24.995868176413264,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1281.970176,1017.643008,0.0,371.195904,277.410816,s,10,0.40465283584594725,0.04046528358459473,0.0012749464270020827,0.040339471817016595,0.041213354873657224,0.04256080570220947,0.04363876636505127,"[0.04390825653076172, 0.039191009521484375, 0.04020032119750976, 0.04062038421630859, 0.04001875305175781, 0.04091392135620117, 0.04051433563232422, 0.03929123306274414, 0.03951599884033203, 0.040478622436523434]",tokens/s,6326.4106246733445,kWh,4.743588816529138e-07,2.5992338141804797e-07,1.2597421947630295e-06,1.994024457833991e-06,tokens/kWh,128383580.7500977,MB,1281.970176,1017.643008,0.0,371.195904,323.048448,s,10,24.280882568359374,2.428088256835937,0.01672270399851807,2.436138427734375,2.4378572509765624,2.4385178588867187,2.4390463452148436,"[2.426797119140625, 2.4161357421875, 2.43644775390625, 2.436491455078125, 2.4358291015625, 2.439178466796875, 2.381922119140625, 2.4365849609375, 2.433785400390625, 2.43771044921875]",tokens/s,25.94633857423941,kWh,2.789970010339496e-05,1.5289993593269173e-05,6.758836843923807e-05,0.0001107780621359022,tokens/kWh,568704.65853349,,s,629,24.602083320617677,0.03911301004867675,0.004804808080024372,0.03870412826538086,0.0389910530090332,0.03930890274047851,0.07752167083740236,"[0.039656448364257815, 0.038916095733642575, 0.03911679840087891, 0.038441982269287106, 0.03871846389770508, 0.03863347244262695, 0.03846553421020508, 0.038747135162353515, 0.038983680725097655, 0.038621185302734375, 0.039468032836914066, 0.03973324966430664, 0.040008705139160154, 0.03956224060058594, 0.038986751556396484, 0.03861708831787109, 0.03892736053466797, 0.038964225769042966, 0.03865497589111328, 0.03835289764404297, 0.03786240005493164, 0.03894784164428711, 0.03873689651489258, 0.038623233795166016, 0.03889459228515625, 0.03893350219726562, 0.038599681854248044, 0.038431808471679686, 0.03866719818115234, 0.03852288055419922, 0.03881881713867188, 0.037525505065917966, 0.03853823852539062, 0.03861196899414063, 0.03871027374267578, 0.0388218879699707, 0.03872870254516601, 0.038642688751220705, 0.03880550384521484, 0.0386693115234375, 0.037972991943359374, 0.03834265518188477, 0.038626304626464845, 0.03872358322143555, 0.03853311920166016, 0.038572032928466796, 0.03861913681030273, 0.03886796951293945, 0.039011329650878904, 0.038809600830078124, 0.03837952041625976, 0.037607425689697264, 0.03740979385375977, 0.03724185562133789, 0.03723164749145508, 0.03744559860229492, 0.037335041046142575, 0.037253120422363284, 0.03730022430419922, 0.037302272796630856, 0.03739648056030274, 0.037308414459228514, 0.0768358383178711, 0.03727462387084961, 0.03721932983398438, 0.03744255828857422, 0.03722444915771484, 0.03723263931274414, 0.03723468780517578, 0.037351425170898435, 0.03726028823852539, 0.037222400665283206, 0.03719782257080078, 0.037326847076416016, 0.03750092697143555, 0.038975486755371096, 0.03851468658447266, 0.038324222564697266, 0.03836415863037109, 0.03865702438354492, 0.03859251022338867, 0.03796889495849609, 0.038454273223876956, 0.03858124923706055, 0.03827609634399414, 0.039785472869873044, 0.03924787139892578, 0.03859251022338867, 0.03877068710327149, 0.03864780807495117, 0.038196224212646485, 0.03844095993041992, 0.03861913681030273, 0.038670337677001954, 0.038452224731445314, 0.037917697906494144, 0.03871744155883789, 0.038529022216796875, 0.038520832061767575, 0.038778881072998046, 0.03852799987792969, 0.038389759063720705, 0.03772422409057617, 0.03858425521850586, 0.03871641540527344, 0.03861094284057617, 0.03887411117553711, 0.03873996734619141, 0.03890790557861328, 0.03827711868286133, 0.038902782440185545, 0.03904307174682617, 0.0387922248840332, 0.038702049255371095, 0.03786137771606445, 0.038809600830078124, 0.03896934509277344, 0.03885977554321289, 0.038787071228027346, 0.038757377624511716, 0.03866009521484375, 0.038470657348632815, 0.038882305145263675, 0.03882393646240234, 0.038712318420410154, 0.07939891052246094, 0.03869286346435547, 0.038795265197753906, 0.03887513732910156, 0.03879423904418945, 0.039034881591796876, 0.03891302490234375, 0.038836223602294925, 0.03861708831787109, 0.03800678253173828, 0.03891097640991211, 0.03804876708984375, 0.038882305145263675, 0.038168575286865236, 0.037981182098388674, 0.03827097702026367, 0.03884646224975586, 0.038865921020507815, 0.0377077751159668, 0.03886284637451172, 0.03846460723876953, 0.0388955192565918, 0.0387583999633789, 0.038612991333007815, 0.038899711608886715, 0.038785022735595705, 0.03849932861328125, 0.037820415496826174, 0.038653953552246094, 0.03862220764160156, 0.038809600830078124, 0.03883724975585937, 0.0387512321472168, 0.03888844680786133, 0.03887411117553711, 0.038833152770996096, 0.03888947296142578, 0.03813888168334961, 0.03830380630493164, 0.038789054870605466, 0.038876190185546874, 0.038718433380126954, 0.03883520126342774, 0.03773132705688476, 0.03853110504150391, 0.03896828842163086, 0.03829248046875, 0.037789695739746096, 0.038752254486083985, 0.03887206268310547, 0.03885158538818359, 0.038711296081542966, 0.038781951904296875, 0.03914342498779297, 0.03878400039672852, 0.038970367431640625, 0.038795265197753906, 0.03885772705078125, 0.03893145751953125, 0.03871027374267578, 0.038882305145263675, 0.03935232162475586, 0.03893657684326172, 0.0801976318359375, 0.040052734375, 0.039428096771240234, 0.038870014190673825, 0.038866943359375, 0.03884646224975586, 0.03879219055175781, 0.038886398315429685, 0.038757377624511716, 0.04078496170043945, 0.039140289306640624, 0.03867238235473633, 0.03890687942504883, 0.038779903411865234, 0.03772723388671875, 0.03863865661621094, 0.038687679290771486, 0.03894681549072266, 0.03893964767456055, 0.03887411117553711, 0.03868371200561523, 0.03804358291625977, 0.03752959823608398, 0.03767091369628906, 0.03757567977905273, 0.03858227157592774, 0.038712318420410154, 0.038975486755371096, 0.03836723327636719, 0.03893862533569336, 0.038735870361328126, 0.038797374725341796, 0.038236095428466794, 0.03860889434814453, 0.03884236907958984, 0.03882495880126953, 0.038738975524902346, 0.03886486434936524, 0.03886489486694336, 0.038742015838623044, 0.03873894500732422, 0.03862015914916992, 0.038100990295410156, 0.03866624069213867, 0.03854131317138672, 0.038796287536621094, 0.0379607048034668, 0.03861708831787109, 0.038790145874023435, 0.03870515060424805, 0.03848908615112305, 0.03781836700439453, 0.0387665901184082, 0.038441982269287106, 0.037978111267089845, 0.038556671142578124, 0.03888127899169922, 0.03868569564819336, 0.03868876647949219, 0.03892428970336914, 0.037594112396240234, 0.038866943359375, 0.038779903411865234, 0.07831346893310547, 0.03814604949951172, 0.03882700729370117, 0.038877185821533204, 0.039695358276367186, 0.03878604888916016, 0.0389826545715332, 0.038836223602294925, 0.038972415924072266, 0.03871744155883789, 0.03882393646240234, 0.03850649642944336, 0.03882495880126953, 0.0381921272277832, 0.03886796951293945, 0.038828033447265625, 0.039008255004882815, 0.0389222412109375, 0.038781951904296875, 0.03867750549316406, 0.038834175109863284, 0.03868876647949219, 0.03847577667236328, 0.03872358322143555, 0.03866726303100586, 0.03881062316894531, 0.03909939193725586, 0.038662143707275394, 0.038950912475585936, 0.038830078125, 0.038580257415771486, 0.038977504730224606, 0.03872870254516601, 0.038736961364746095, 0.03875628662109375, 0.038795265197753906, 0.038596607208251955, 0.03853414535522461, 0.03887206268310547, 0.03856076812744141, 0.03790950393676758, 0.038529022216796875, 0.03870412826538086, 0.03858432006835937, 0.037351425170898435, 0.03753881454467774, 0.038594558715820314, 0.038742015838623044, 0.038214656829833986, 0.038109184265136715, 0.03885363388061523, 0.03869696044921875, 0.038675457000732424, 0.038441982269287106, 0.03867750549316406, 0.03830476760864258, 0.0388587532043457, 0.038814720153808595, 0.038749183654785156, 0.038161407470703124, 0.03876147079467773, 0.03878297424316406, 0.03930521774291992, 0.07981977844238282, 0.03822796630859375, 0.03854131317138672, 0.03871027374267578, 0.03872153472900391, 0.03880550384521484, 0.038801406860351564, 0.0385873908996582, 0.03875635147094727, 0.03867852783203125, 0.03861606216430664, 0.038575103759765625, 0.03855052947998047, 0.03796489715576172, 0.03806198501586914, 0.03942092895507812, 0.04028108978271484, 0.038964225769042966, 0.03882393646240234, 0.03866419219970703, 0.03899084854125977, 0.03872051239013672, 0.037768192291259765, 0.03894169616699219, 0.0388935661315918, 0.03912704086303711, 0.03989913558959961, 0.038953983306884765, 0.03872972869873047, 0.03805184173583984, 0.037454849243164064, 0.03857100677490234, 0.038653953552246094, 0.03874508666992187, 0.03873689651489258, 0.038778881072998046, 0.03867340850830078, 0.03877580642700195, 0.03796480178833008, 0.03889152145385742, 0.0387665901184082, 0.03850137710571289, 0.03892633438110352, 0.03906969451904297, 0.0388403205871582, 0.03877171325683594, 0.03902873611450195, 0.03871744155883789, 0.03862015914916992, 0.03813683319091797, 0.038904830932617186, 0.03887308883666992, 0.038991870880126955, 0.03880243301391602, 0.03870207977294922, 0.03907174301147461, 0.03927449417114258, 0.038986751556396484, 0.0388218879699707, 0.03859763336181641, 0.03830579376220703, 0.03789209747314453, 0.03869081497192383, 0.07953612518310547, 0.03946086502075195, 0.03886796951293945, 0.03824127960205078, 0.037698593139648434, 0.037787647247314454, 0.03757564926147461, 0.03742924880981445, 0.037425151824951174, 0.03743027114868164, 0.037174270629882815, 0.037351425170898435, 0.0374200325012207, 0.037800960540771485, 0.037493759155273435, 0.037550079345703126, 0.03741593551635742, 0.037563392639160156, 0.037478462219238284, 0.037314495086669924, 0.037454849243164064, 0.037424129486083986, 0.03731763076782227, 0.03743129730224609, 0.03770265579223633, 0.03746201705932617, 0.03763411331176758, 0.03753977584838867, 0.037443584442138675, 0.03744255828857422, 0.0373196792602539, 0.03894784164428711, 0.03799244689941406, 0.03865497589111328, 0.03867750549316406, 0.038719486236572266, 0.03785830307006836, 0.038100990295410156, 0.03767705535888672, 0.03843379211425781, 0.038870014190673825, 0.03885055923461914, 0.0385054702758789, 0.03881881713867188, 0.03787366485595703, 0.03811635208129883, 0.03767193603515625, 0.03761459350585938, 0.03747123336791992, 0.037459968566894535, 0.03746105575561524, 0.037547969818115236, 0.037591041564941405, 0.03749990463256836, 0.03776716613769531, 0.037759998321533206, 0.03740467071533203, 0.037466110229492186, 0.03745587158203125, 0.03762176132202148, 0.03740467071533203, 0.03756851196289063, 0.037424129486083986, 0.07716556549072266, 0.03832524871826172, 0.03913113784790039, 0.03907583999633789, 0.03930316925048828, 0.04006604766845703, 0.03922431945800781, 0.03863961410522461, 0.03853311920166016, 0.03850035095214844, 0.038386688232421876, 0.03758899307250976, 0.0377077751159668, 0.03966054534912109, 0.03961654281616211, 0.039029727935791014, 0.038882305145263675, 0.03892940902709961, 0.038949951171875, 0.038799297332763674, 0.03883827209472656, 0.03873177719116211, 0.038470657348632815, 0.03772825622558594, 0.03750096130371094, 0.03739542388916016, 0.03849523162841797, 0.038932479858398435, 0.03892736053466797, 0.038724609375, 0.03894784164428711, 0.03870003128051758, 0.03879731369018555, 0.03885772705078125, 0.039008255004882815, 0.03877580642700195, 0.03863449478149414, 0.03875020980834961, 0.038730751037597655, 0.038609920501708986, 0.03889152145385742, 0.03879935836791992, 0.03799859237670898, 0.03885055923461914, 0.03883827209472656, 0.03868364715576172, 0.0378419189453125, 0.03827097702026367, 0.038882366180419924, 0.03889350509643555, 0.03880550384521484, 0.03875328063964844, 0.0384983024597168, 0.03806719970703125, 0.03889766311645508, 0.038747135162353515, 0.03880652618408203, 0.03879731369018555, 0.038317054748535154, 0.03869696044921875, 0.038836223602294925, 0.038951934814453124, 0.03869081497192383, 0.07766015625, 0.03868057632446289, 0.03878092956542969, 0.03884236907958984, 0.03863961410522461, 0.03917420959472656, 0.03880441665649414, 0.03875942230224609, 0.03831193542480469, 0.03883724975585937, 0.03871334457397461, 0.038863872528076174, 0.03870003128051758, 0.0387061767578125, 0.03883929443359375, 0.03847679901123047, 0.0389939193725586, 0.039126014709472655, 0.03876249694824219, 0.03862835311889649, 0.038225921630859375, 0.037928958892822266, 0.038201343536376955, 0.038752254486083985, 0.03788288116455078, 0.038316032409667966, 0.03892838287353516, 0.03883827209472656, 0.038575103759765625, 0.03866624069213867, 0.038742015838623044, 0.03868569564819336, 0.038523902893066404, 0.03863142395019531, 0.03812351989746094, 0.03848396682739258, 0.0384266242980957, 0.0380579833984375, 0.03916595077514649, 0.03902873611450195, 0.03880038452148438, 0.03798527908325195, 0.03868979263305664, 0.03873484802246094, 0.03874303817749023, 0.03897139358520508, 0.03858534240722656, 0.03883724975585937, 0.03884134292602539, 0.03871027374267578, 0.038441982269287106, 0.03866726303100586, 0.038435840606689455, 0.037664768218994144, 0.03777433776855469, 0.03874611282348633, 0.03870412826538086, 0.039122943878173826, 0.03954585647583008, 0.0387061767578125, 0.03868569564819336, 0.038558719635009765, 0.03892326354980469, 0.0795125732421875, 0.03869388961791992, 0.038509567260742186, 0.03903078460693359, 0.03869900894165039, 0.03888844680786133, 0.03870924758911133, 0.03882495880126953, 0.03866828918457031, 0.03871334457397461, 0.03749068832397461, 0.0382371826171875, 0.038708225250244144, 0.03870412826538086, 0.038577152252197267, 0.039311359405517575, 0.03893862533569336, 0.03851878356933594, 0.03875328063964844, 0.03947212982177734, 0.03978035354614258, 0.038626304626464845, 0.03877171325683594, 0.038569984436035154, 0.0387512321472168, 0.038697982788085936, 0.03871027374267578, 0.038152191162109376, 0.03879423904418945, 0.0386324462890625, 0.03985612869262695, 0.038866943359375, 0.038002784729003904, 0.03856582260131836, 0.03872560119628906, 0.038675457000732424, 0.038768638610839845, 0.03867750549316406, 0.03897651290893555, 0.03865599822998047, 0.03887104034423828, 0.03846963119506836, 0.038752254486083985, 0.0383109130859375, 0.039003135681152344, 0.03877273559570313, 0.03880755233764648, 0.038665214538574216, 0.0388853759765625, 0.038697982788085936, 0.03744563293457031, 0.038452224731445314, 0.03876761627197266, 0.038814720153808595, 0.03867750549316406, 0.03863961410522461, 0.03885977554321289, 0.03878604888916016, 0.03861920166015625, 0.039180225372314456, 0.038986751556396484, 0.03813785552978516, 0.037392383575439454]",tokens/s,25.56694048234806,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2029.686784,5274.861568,0.0,4628.414464,4279.64672,s,10,3.174773986816406,0.3174773986816406,0.0009221487872443522,0.3173043212890625,0.31849222106933595,0.31901463470458985,0.319432565612793,"[0.31953704833984375, 0.31614645385742185, 0.31721881103515626, 0.3171442565917969, 0.31738983154296874, 0.3167364807128906, 0.3175615539550781, 0.31837612915039065, 0.31800189208984375, 0.3166615295410156]",tokens/s,806.356613299302,kWh,3.736858639038272e-06,2.0476319108638563e-06,1.8145144724437956e-05,2.392963527434008e-05,tokens/kWh,10698031.836469762,MB,2029.686784,5274.861568,0.0,4628.414464,4463.185408,s,10,176.95450000000002,17.69545,0.0031517961497735305,17.6950830078125,17.699231054687502,17.70004228515625,17.700691269531248,"[17.69897265625, 17.69905078125, 17.700853515625, 17.69500390625, 17.695162109375, 17.69576171875, 17.693421875, 17.69222265625, 17.693814453125, 17.690236328125]",tokens/s,3.5602372361256704,kWh,0.00020892207810655237,0.00011450671390637579,0.0009965308336823604,0.0013199596256952884,tokens/kWh,47728.732586661325,,s,629,179.5270961608886,0.2854166870602364,0.03770346789831642,0.28084429931640625,0.2811535339355469,0.2813114379882813,0.5979956518554688,"[0.2810900573730469, 0.28075418090820314, 0.2808637390136719, 0.28086273193359373, 0.2813634643554688, 0.2807715759277344, 0.2808360900878906, 0.28084429931640625, 0.28089447021484376, 0.28069683837890624, 0.2808883056640625, 0.28103372192382814, 0.28076544189453123, 0.28069683837890624, 0.2809046936035156, 0.28098458862304687, 0.2806855773925781, 0.28079000854492187, 0.2813388671875, 0.2808258666992188, 0.2808616943359375, 0.28091802978515623, 0.2808821716308594, 0.28074496459960935, 0.28068453979492186, 0.28067019653320313, 0.28085964965820315, 0.28079205322265627, 0.2806651000976563, 0.28080743408203124, 0.2807859191894531, 0.2809405517578125, 0.28102041625976565, 0.28240179443359376, 0.28073675537109377, 0.2808053894042969, 0.28098355102539063, 0.28075518798828125, 0.2811678771972656, 0.2814289855957031, 0.28092312622070315, 0.2808555603027344, 0.28077566528320314, 0.2810408935546875, 0.28091595458984375, 0.2808340454101563, 0.28091189575195313, 0.2809783935546875, 0.28075726318359373, 0.28088116455078127, 0.2808842163085937, 0.28079718017578126, 0.2806927490234375, 0.28096408081054686, 0.28156112670898437, 0.28095278930664064, 0.28083203125, 0.281001953125, 0.28085964965820315, 0.28099172973632813, 0.28073165893554686, 0.28094259643554687, 0.59812353515625, 0.2807715759277344, 0.28079205322265627, 0.2807306213378906, 0.28069989013671875, 0.28109515380859373, 0.280732666015625, 0.2807408752441406, 0.28080743408203124, 0.28074188232421876, 0.2810009460449219, 0.2807080993652344, 0.28133273315429685, 0.28084326171875, 0.28075518798828125, 0.28082382202148437, 0.28070196533203123, 0.2806036376953125, 0.28080230712890625, 0.28076544189453123, 0.28069989013671875, 0.28076031494140624, 0.28054937744140623, 0.28094772338867186, 0.2809251708984375, 0.28090264892578126, 0.2812037048339844, 0.2809169921875, 0.2808821716308594, 0.28092312622070315, 0.2808821716308594, 0.2807029724121094, 0.28090673828125, 0.2810491027832031, 0.28080126953125, 0.2807142333984375, 0.28088626098632813, 0.2810163269042969, 0.28109619140625, 0.2811351013183594, 0.28115045166015623, 0.2812467346191406, 0.2808637390136719, 0.2808412170410156, 0.2813102111816406, 0.28072039794921877, 0.28143719482421875, 0.2813716430664063, 0.2811709289550781, 0.28095489501953125, 0.280875, 0.2809036865234375, 0.28120269775390627, 0.2809466857910156, 0.2809661560058594, 0.28132965087890627, 0.28103067016601563, 0.28087808227539063, 0.28130508422851563, 0.2808392028808594, 0.28099169921875, 0.28102554321289064, 0.2814218139648437, 0.5983283081054688, 0.280806396484375, 0.28079000854492187, 0.28083917236328126, 0.28077362060546873, 0.28097125244140625, 0.2809374694824219, 0.28083096313476563, 0.2809302978515625, 0.2808637390136719, 0.28075418090820314, 0.28077362060546873, 0.28114227294921873, 0.2807705688476563, 0.28075314331054685, 0.28077362060546873, 0.28073675537109377, 0.280890380859375, 0.28105523681640626, 0.28109722900390627, 0.28098458862304687, 0.28080126953125, 0.28088626098632813, 0.2811689453125, 0.2808769836425781, 0.2808493957519531, 0.28083917236328126, 0.28073370361328126, 0.280880126953125, 0.2808493957519531, 0.2811535339355469, 0.28096511840820315, 0.28092108154296874, 0.2810409240722656, 0.281107421875, 0.28079000854492187, 0.2810408935546875, 0.2809927673339844, 0.2809743347167969, 0.28115866088867186, 0.28111154174804687, 0.2810992736816406, 0.28116583251953126, 0.28090573120117185, 0.2810828857421875, 0.2810828857421875, 0.2809743347167969, 0.2810828857421875, 0.281702392578125, 0.28085964965820315, 0.2809661560058594, 0.2809671630859375, 0.2809518127441406, 0.2811023254394531, 0.281312255859375, 0.281206787109375, 0.2811064453125, 0.2809405517578125, 0.28098355102539063, 0.2809098205566406, 0.28094259643554687, 0.2807930908203125, 0.2810654602050781, 0.5982269287109375, 0.28070199584960936, 0.2806517333984375, 0.2808186950683594, 0.2806579284667969, 0.2811094970703125, 0.2807715759277344, 0.2807982177734375, 0.28059237670898435, 0.2807500915527344, 0.28073675537109377, 0.2805841979980469, 0.2806558837890625, 0.2806814575195313, 0.2806732788085938, 0.2806876220703125, 0.28078387451171877, 0.28072659301757813, 0.2806312255859375, 0.28082278442382813, 0.28084326171875, 0.28062823486328126, 0.28076544189453123, 0.280748046875, 0.28071014404296873, 0.2810214538574219, 0.2813992919921875, 0.28084530639648436, 0.28090264892578126, 0.2808401794433594, 0.2808299560546875, 0.2809927673339844, 0.28081253051757815, 0.28067636108398436, 0.2807152709960937, 0.2807992248535156, 0.2808186950683594, 0.2809610290527344, 0.2807982177734375, 0.280875, 0.280995849609375, 0.2812569580078125, 0.2808719482421875, 0.2807510986328125, 0.2810491027832031, 0.28073165893554686, 0.280985595703125, 0.28093548583984373, 0.2809241027832031, 0.280806396484375, 0.28093646240234377, 0.2808463439941406, 0.2809794616699219, 0.28087808227539063, 0.2808883056640625, 0.2814617614746094, 0.28086782836914065, 0.28073165893554686, 0.2811760559082031, 0.2809938049316406, 0.28129177856445314, 0.28095187377929687, 0.2817914123535156, 0.5984921875, 0.28099789428710936, 0.28095077514648437, 0.281122802734375, 0.2807726135253906, 0.2808883056640625, 0.28075418090820314, 0.28093646240234377, 0.28081253051757815, 0.28075213623046874, 0.2807726135253906, 0.28081253051757815, 0.28140032958984373, 0.28087808227539063, 0.28151806640625, 0.2811494445800781, 0.2811781005859375, 0.2809241638183594, 0.2809599914550781, 0.28065176391601565, 0.2806476745605469, 0.28059033203125, 0.28050125122070313, 0.2807224426269531, 0.28059039306640626, 0.2806568298339844, 0.2806446228027344, 0.2807173156738281, 0.280627197265625, 0.28063540649414065, 0.2808340454101563, 0.2807080993652344, 0.2810368041992187, 0.2810357666015625, 0.28083712768554686, 0.2806732788085938, 0.2807879638671875, 0.2809333801269531, 0.2808832092285156, 0.280869873046875, 0.2805882873535156, 0.28103067016601563, 0.2808279113769531, 0.2807982177734375, 0.2809129638671875, 0.28075820922851563, 0.28076953125, 0.2810152893066406, 0.28108184814453124, 0.2806855773925781, 0.2812528686523437, 0.2812149658203125, 0.28091494750976564, 0.28111154174804687, 0.28087908935546874, 0.2808330383300781, 0.2809251708984375, 0.28099172973632813, 0.28070401000976564, 0.2810091552734375, 0.2808401794433594, 0.2807623901367188, 0.2810029907226563, 0.5976668090820313, 0.28066201782226563, 0.280774658203125, 0.28086578369140625, 0.28072857666015627, 0.2810163879394531, 0.28058001708984376, 0.2806947937011719, 0.28056781005859377, 0.2807726135253906, 0.28080230712890625, 0.280553466796875, 0.28076544189453123, 0.280690673828125, 0.2808268737792969, 0.280875, 0.28081048583984375, 0.28086477661132814, 0.28081048583984375, 0.2810245056152344, 0.2807193603515625, 0.2807080993652344, 0.2806855773925781, 0.2810071105957031, 0.2810624084472656, 0.28088525390625, 0.28093646240234377, 0.2807060546875, 0.28088116455078127, 0.2806405029296875, 0.28068658447265626, 0.2808965148925781, 0.28075930786132813, 0.2807490539550781, 0.28122213745117186, 0.280764404296875, 0.28102655029296875, 0.28092825317382814, 0.28138290405273436, 0.28117196655273435, 0.28144537353515625, 0.2809876403808594, 0.281027587890625, 0.28075424194335935, 0.2808401184082031, 0.28082278442382813, 0.28102655029296875, 0.28107366943359374, 0.28075314331054685, 0.28084735107421877, 0.28093438720703123, 0.28076544189453123, 0.28106341552734376, 0.28107879638671873, 0.28101837158203125, 0.2814535827636719, 0.28100811767578127, 0.28086477661132814, 0.2809016418457031, 0.28101119995117185, 0.2808832092285156, 0.28088528442382815, 0.28095895385742187, 0.598540283203125, 0.2806599731445312, 0.2808401794433594, 0.2807408752441406, 0.28097842407226564, 0.2807705688476563, 0.28113101196289064, 0.28089547729492187, 0.28100607299804686, 0.28068453979492186, 0.2810439758300781, 0.28067019653320313, 0.28120269775390627, 0.2807859191894531, 0.280658935546875, 0.2809333801269531, 0.28086578369140625, 0.2808279113769531, 0.28084530639648436, 0.2808913879394531, 0.28075314331054685, 0.28059237670898435, 0.28081561279296874, 0.28066815185546873, 0.28071832275390624, 0.2806855773925781, 0.28078182983398436, 0.2810224914550781, 0.2808012390136719, 0.28073370361328126, 0.28072857666015627, 0.2806651000976563, 0.2808084411621094, 0.280742919921875, 0.2807459716796875, 0.2807408752441406, 0.28083096313476563, 0.280958984375, 0.2808985595703125, 0.2808555603027344, 0.2810992736816406, 0.2811351013183594, 0.2806947937011719, 0.28073370361328126, 0.2807859191894531, 0.2808637390136719, 0.2812098693847656, 0.2812231750488281, 0.28102041625976565, 0.2809169921875, 0.2809938049316406, 0.28092007446289063, 0.2809825134277344, 0.2808616943359375, 0.28068453979492186, 0.28086782836914065, 0.2807234497070312, 0.28062619018554685, 0.2807685241699219, 0.28067123413085937, 0.2807296142578125, 0.2808002624511719, 0.28117300415039065, 0.5992601318359375, 0.2805698547363281, 0.2806220703125, 0.28086578369140625, 0.2806927490234375, 0.28068658447265626, 0.28074188232421876, 0.28050225830078124, 0.28068658447265626, 0.2806220703125, 0.28081460571289063, 0.2807357482910156, 0.28070706176757815, 0.28100311279296875, 0.28070492553710935, 0.2808493957519531, 0.2806118469238281, 0.2809036865234375, 0.2807828369140625, 0.2809046936035156, 0.2806640625, 0.28067636108398436, 0.28069888305664065, 0.28089959716796875, 0.2806876220703125, 0.28073165893554686, 0.28076544189453123, 0.2807173156738281, 0.2807705688476563, 0.2804951171875, 0.2809938049316406, 0.2808279113769531, 0.28095693969726565, 0.28122726440429685, 0.2810163269042969, 0.28069989013671875, 0.2808842163085937, 0.28082278442382813, 0.28080126953125, 0.28084429931640625, 0.28069580078125, 0.2806753234863281, 0.28080230712890625, 0.2806804504394531, 0.2807798461914062, 0.2811206970214844, 0.28091802978515623, 0.28079718017578126, 0.28101937866210935, 0.2810828857421875, 0.2809466857910156, 0.2809190368652344, 0.2809415588378906, 0.28108389282226565, 0.28117913818359375, 0.28109515380859373, 0.2810357666015625, 0.2808545227050781, 0.280890380859375, 0.2808606872558594, 0.28100811767578127, 0.28085760498046874, 0.28095489501953125, 0.5989181518554687, 0.28075518798828125, 0.28063436889648435, 0.2807193603515625, 0.2805350341796875, 0.2809671630859375, 0.2804613037109375, 0.28068453979492186, 0.28063641357421876, 0.2805770263671875, 0.2807633972167969, 0.28062924194335936, 0.2807562255859375, 0.2807296142578125, 0.2805565490722656, 0.2806087646484375, 0.2806476745605469, 0.2805893249511719, 0.28073983764648436, 0.28069171142578125, 0.2810378112792969, 0.28060671997070313, 0.28059442138671875, 0.28090673828125, 0.28125799560546877, 0.28063436889648435, 0.28085861206054685, 0.280511474609375, 0.28059442138671875, 0.2806200256347656, 0.2808002624511719, 0.2806661071777344, 0.280890380859375, 0.28054629516601565, 0.2807828369140625, 0.28066302490234374, 0.280732666015625, 0.2807224426269531, 0.28073165893554686, 0.2807142333984375, 0.28113101196289064, 0.280958984375, 0.28107366943359374, 0.2807439270019531, 0.2809518127441406, 0.28086578369140625, 0.28080435180664065, 0.2811535339355469, 0.2808248291015625, 0.28094064331054686, 0.2808831176757812, 0.2809313354492188, 0.2811975708007812, 0.281122802734375, 0.2807705688476563, 0.2809292907714844, 0.281385986328125, 0.2810009460449219, 0.2827796630859375, 0.28104193115234377, 0.2813173828125, 0.2812630920410156, 0.2812651672363281, 0.597570556640625, 0.2809108581542969, 0.28094259643554687, 0.28096307373046875, 0.280585205078125, 0.2806241149902344, 0.28071218872070314, 0.28067840576171876, 0.2805749816894531, 0.2805186767578125, 0.28062612915039065, 0.280669189453125, 0.2806609802246094, 0.2807214050292969, 0.28059954833984374, 0.28069888305664065, 0.28063540649414065, 0.28063128662109377, 0.2805401611328125, 0.28073983764648436, 0.28051763916015626, 0.28082278442382813, 0.2806200256347656, 0.2806640625, 0.28071218872070314, 0.2807879638671875, 0.28067123413085937, 0.2807459716796875, 0.2807715759277344, 0.2807060546875, 0.28080743408203124, 0.28084326171875, 0.28085247802734375, 0.2809169921875, 0.2808770446777344, 0.28057907104492186, 0.28081048583984375, 0.28083712768554686, 0.28081765747070314, 0.28080743408203124, 0.28084326171875, 0.2808401794433594, 0.28075314331054685, 0.2806026306152344, 0.28084640502929686, 0.28075103759765624, 0.2807500915527344, 0.2811566162109375, 0.2810798034667969, 0.28084429931640625, 0.2807982177734375, 0.28074188232421876, 0.28069989013671875, 0.28078695678710935, 0.280806396484375, 0.2809016418457031, 0.28089447021484376, 0.2807982177734375, 0.28089547729492187, 0.28093438720703123, 0.2809241638183594, 0.2807132568359375, 0.28090365600585937]",tokens/s,3.50364938469401,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10516,7 +10516,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1451.905024,2418.540544,0.0,1772.09344,1621.541376,s,10,0.7316847381591797,0.07316847381591798,0.0003457241418479124,0.07326449584960938,0.07357357177734375,0.07357862548828126,0.07358266845703125,"[0.07346591949462891, 0.0728661117553711, 0.07270166778564453, 0.07324726104736329, 0.07263625335693359, 0.07288240051269532, 0.07358367919921875, 0.07328173065185548, 0.073447265625, 0.07357244873046875]",tokens/s,3498.7746313263483,kWh,8.589118671110095e-07,4.706412848463514e-07,3.7715146919278033e-06,5.101067843885164e-06,tokens/kWh,50185570.51870551,MB,1451.905024,2418.540544,0.0,1772.09344,1723.273216,s,10,38.208432373046875,3.8208432373046874,0.0009523784305681953,3.8212119140625003,3.821659497070313,3.821850988769531,3.822004182128906,"[3.8209521484375, 3.821502685546875, 3.821594482421875, 3.8205712890625, 3.82204248046875, 3.819373291015625, 3.8214716796875, 3.819123291015625, 3.82018408203125, 3.821616943359375]",tokens/s,16.488506878508232,kWh,4.511216866125423e-05,2.4723927950708465e-05,0.0001974284518234696,0.0002672645484354323,tokens/kWh,235721.49904954564,,s,629,38.80548554611203,0.06169393568539279,0.008717092816584612,0.06063616180419922,0.060711936950683595,0.060762318420410154,0.1337114239501953,"[0.060548095703125, 0.06056447982788086, 0.06056243133544922, 0.06060441589355469, 0.06070169448852539, 0.06058393478393555, 0.06068428802490235, 0.06065151977539063, 0.060698623657226565, 0.060622848510742185, 0.06061056137084961, 0.06058598327636719, 0.06065971374511719, 0.060631038665771485, 0.06055321502685547, 0.060627967834472656, 0.06062080001831055, 0.06059212875366211, 0.06057062530517578, 0.06062694549560547, 0.06059212875366211, 0.06070169448852539, 0.060641281127929686, 0.06050099182128906, 0.06061056137084961, 0.060644351959228515, 0.06061363220214844, 0.06065459060668945, 0.060680191040039064, 0.06059929656982422, 0.0606033935546875, 0.06075494384765625, 0.06057984161376953, 0.060627967834472656, 0.06067609786987305, 0.06060851287841797, 0.060625919342041014, 0.06067507171630859, 0.06063820648193359, 0.06066790390014649, 0.0607375373840332, 0.06063718414306641, 0.06057164764404297, 0.06067814254760742, 0.0606033935546875, 0.06069452667236328, 0.06069247817993164, 0.06072934341430664, 0.060672000885009764, 0.06068121719360352, 0.06060441589355469, 0.060665855407714846, 0.06066995239257812, 0.06066995239257812, 0.0605849609375, 0.06070272064208984, 0.060647422790527344, 0.06064025497436523, 0.06067302322387695, 0.060631038665771485, 0.060611583709716796, 0.06069145584106445, 0.1337518005371094, 0.060679168701171876, 0.060565502166748046, 0.060598270416259765, 0.06051942443847656, 0.06059212875366211, 0.06055731201171875, 0.06060748672485351, 0.06057062530517578, 0.06058700942993164, 0.060539905548095706, 0.06059212875366211, 0.060524543762207034, 0.06062694549560547, 0.0605296630859375, 0.06064025497436523, 0.060556289672851565, 0.0607088623046875, 0.06061772918701172, 0.0605849609375, 0.060622848510742185, 0.06062387084960937, 0.060554241180419924, 0.06061875152587891, 0.06054912185668945, 0.060633087158203126, 0.06063718414306641, 0.06061670303344727, 0.06066175842285156, 0.06064025497436523, 0.06060236740112305, 0.06065663909912109, 0.06065151977539063, 0.06146867370605469, 0.060672000885009764, 0.06063513565063477, 0.06062899017333984, 0.06063513565063477, 0.06063513565063477, 0.06066175842285156, 0.06074060821533203, 0.0606453742980957, 0.06065663909912109, 0.0606484489440918, 0.06070374298095703, 0.0606453742980957, 0.06063411331176758, 0.06067814254760742, 0.060641281127929686, 0.06077132797241211, 0.060819454193115234, 0.06069964981079102, 0.06061772918701172, 0.0607375373840332, 0.06071091079711914, 0.06056857681274414, 0.06065868759155273, 0.060695552825927736, 0.06065151977539063, 0.06061363220214844, 0.06063616180419922, 0.06097817611694336, 0.060883968353271485, 0.13362074279785155, 0.06075699234008789, 0.060611583709716796, 0.06068838500976562, 0.06058086395263672, 0.06061977767944336, 0.06060031890869141, 0.06062182235717774, 0.06057881546020508, 0.0607006721496582, 0.06058598327636719, 0.060609535217285154, 0.06054604721069336, 0.060641281127929686, 0.06057779312133789, 0.060614654541015625, 0.060682239532470705, 0.06065049743652344, 0.060854270935058595, 0.060627967834472656, 0.060611583709716796, 0.06075289535522461, 0.060668926239013675, 0.06060236740112305, 0.060581886291503906, 0.060715007781982425, 0.06063616180419922, 0.06065356826782226, 0.06065459060668945, 0.060662784576416016, 0.060657665252685546, 0.060644351959228515, 0.06066483306884766, 0.06063513565063477, 0.060652542114257815, 0.060660736083984375, 0.060652542114257815, 0.060655616760253904, 0.06063206481933594, 0.06065151977539063, 0.06070272064208984, 0.060677120208740234, 0.060706817626953125, 0.06065151977539063, 0.06060748672485351, 0.06067609786987305, 0.060674049377441405, 0.06074982452392578, 0.060652542114257815, 0.060711936950683595, 0.06064640045166016, 0.06088806533813477, 0.060649471282958986, 0.06063820648193359, 0.06074060821533203, 0.06055014419555664, 0.06067097473144531, 0.06065868759155273, 0.060693504333496094, 0.060663806915283204, 0.06061977767944336, 0.060641281127929686, 0.06065151977539063, 0.13421568298339845, 0.061047809600830075, 0.06058086395263672, 0.06061363220214844, 0.06059417724609375, 0.060660736083984375, 0.06055219268798828, 0.06056959915161133, 0.06055833435058594, 0.06059110260009765, 0.060598270416259765, 0.06062080001831055, 0.0606033935546875, 0.060644351959228515, 0.06060134506225586, 0.06059212875366211, 0.060576766967773435, 0.06063616180419922, 0.06062387084960937, 0.060624897003173826, 0.060870655059814455, 0.06065971374511719, 0.06057369613647461, 0.06063411331176758, 0.060572673797607425, 0.06064640045166016, 0.06059110260009765, 0.06067814254760742, 0.06068428802490235, 0.060662784576416016, 0.06063411331176758, 0.06064640045166016, 0.060652542114257815, 0.06065663909912109, 0.060611583709716796, 0.060641281127929686, 0.06062182235717774, 0.060668926239013675, 0.06061977767944336, 0.060657665252685546, 0.06068838500976562, 0.060642303466796874, 0.060652542114257815, 0.060644351959228515, 0.060609535217285154, 0.060624897003173826, 0.06063206481933594, 0.06063616180419922, 0.06063820648193359, 0.06064332962036133, 0.06060748672485351, 0.0606300163269043, 0.06061363220214844, 0.06065459060668945, 0.06070579147338867, 0.06060441589355469, 0.06068633651733398, 0.06062694549560547, 0.06064025497436523, 0.06063513565063477, 0.06063616180419922, 0.06065356826782226, 0.060655616760253904, 0.13350604248046874, 0.060693504333496094, 0.06060851287841797, 0.06065663909912109, 0.06060031890869141, 0.06069247817993164, 0.06064332962036133, 0.06060441589355469, 0.06056345748901367, 0.06063616180419922, 0.060576766967773435, 0.06068121719360352, 0.06066483306884766, 0.06058598327636719, 0.06065868759155273, 0.060622848510742185, 0.060655616760253904, 0.060633087158203126, 0.06061056137084961, 0.06066995239257812, 0.06062694549560547, 0.06065356826782226, 0.06086963272094727, 0.060609535217285154, 0.060611583709716796, 0.060919807434082034, 0.06071398544311524, 0.060652542114257815, 0.060660736083984375, 0.0606484489440918, 0.060668926239013675, 0.06059929656982422, 0.060655616760253904, 0.0608092155456543, 0.06080204772949219, 0.06074879837036133, 0.06071398544311524, 0.060622848510742185, 0.060660736083984375, 0.06069452667236328, 0.06069964981079102, 0.06062080001831055, 0.060665855407714846, 0.060649471282958986, 0.06063718414306641, 0.06069145584106445, 0.06063616180419922, 0.0606484489440918, 0.06067302322387695, 0.06059110260009765, 0.06071398544311524, 0.06069145584106445, 0.06063616180419922, 0.060655616760253904, 0.0607303695678711, 0.060576766967773435, 0.06069657516479492, 0.060682239532470705, 0.060644351959228515, 0.06062387084960937, 0.06065151977539063, 0.0606300163269043, 0.06076006317138672, 0.13374668884277344, 0.060639232635498044, 0.06056959915161133, 0.06061568069458008, 0.06057164764404297, 0.060706817626953125, 0.06055731201171875, 0.060556289672851565, 0.06061670303344727, 0.06055321502685547, 0.06053376007080078, 0.06062899017333984, 0.060548095703125, 0.06059929656982422, 0.060622848510742185, 0.06057369613647461, 0.06059110260009765, 0.06065868759155273, 0.06058291244506836, 0.06058905410766602, 0.06061056137084961, 0.06056755065917969, 0.06058291244506836, 0.06062694549560547, 0.06063616180419922, 0.06063513565063477, 0.06063718414306641, 0.06055526351928711, 0.060622848510742185, 0.060649471282958986, 0.06062899017333984, 0.060614654541015625, 0.06070476913452148, 0.060560382843017575, 0.06061875152587891, 0.0606453742980957, 0.06059724807739258, 0.06061670303344727, 0.060728321075439455, 0.060606464385986325, 0.060657665252685546, 0.06066175842285156, 0.06058598327636719, 0.06061260986328125, 0.06068940734863281, 0.06057984161376953, 0.06063206481933594, 0.060663806915283204, 0.06060543823242188, 0.06058905410766602, 0.060711936950683595, 0.06060031890869141, 0.06067097473144531, 0.06065151977539063, 0.060677120208740234, 0.06054604721069336, 0.06071091079711914, 0.060649471282958986, 0.06063718414306641, 0.06063820648193359, 0.06069964981079102, 0.06061875152587891, 0.06069760131835938, 0.1345075225830078, 0.06063411331176758, 0.06067097473144531, 0.060625919342041014, 0.060560382843017575, 0.06066790390014649, 0.060548095703125, 0.06058291244506836, 0.060598270416259765, 0.060606464385986325, 0.06056140899658203, 0.0605665283203125, 0.06058803176879883, 0.06055321502685547, 0.06061260986328125, 0.060576766967773435, 0.06076313781738281, 0.06071295928955078, 0.060576766967773435, 0.060572673797607425, 0.060627967834472656, 0.060576766967773435, 0.06057164764404297, 0.06062080001831055, 0.06053887939453125, 0.06092595291137695, 0.0607006721496582, 0.06059110260009765, 0.06064332962036133, 0.06065049743652344, 0.06057779312133789, 0.060633087158203126, 0.06071295928955078, 0.060647422790527344, 0.060639232635498044, 0.060668926239013675, 0.06061363220214844, 0.06061056137084961, 0.060639232635498044, 0.06059212875366211, 0.060693504333496094, 0.06067097473144531, 0.06083379364013672, 0.06059417724609375, 0.0606668815612793, 0.06058393478393555, 0.06063820648193359, 0.06070579147338867, 0.06062694549560547, 0.06063206481933594, 0.06072115325927734, 0.06065663909912109, 0.06077951812744141, 0.06069452667236328, 0.06077030563354492, 0.06057779312133789, 0.06068326568603516, 0.06058598327636719, 0.06092390441894531, 0.060906494140625, 0.060660736083984375, 0.06083686447143555, 0.060761089324951174, 0.13417062377929687, 0.060611583709716796, 0.060647422790527344, 0.06059622573852539, 0.060543998718261716, 0.06064025497436523, 0.06065868759155273, 0.06062899017333984, 0.060560382843017575, 0.06058598327636719, 0.060523521423339846, 0.06059417724609375, 0.06057369613647461, 0.06057984161376953, 0.06057164764404297, 0.06060031890869141, 0.06056857681274414, 0.060624897003173826, 0.0606453742980957, 0.06063513565063477, 0.06058700942993164, 0.060609535217285154, 0.06055833435058594, 0.060572673797607425, 0.06055116653442383, 0.060662784576416016, 0.0606453742980957, 0.06065971374511719, 0.0606300163269043, 0.06062387084960937, 0.06058700942993164, 0.060649471282958986, 0.06067097473144531, 0.060679168701171876, 0.060624897003173826, 0.060611583709716796, 0.060622848510742185, 0.060665855407714846, 0.060625919342041014, 0.06063411331176758, 0.06066790390014649, 0.060622848510742185, 0.0606300163269043, 0.060641281127929686, 0.06063820648193359, 0.06062080001831055, 0.06061670303344727, 0.06062080001831055, 0.06059929656982422, 0.0606300163269043, 0.06063206481933594, 0.06067609786987305, 0.06058803176879883, 0.06060543823242188, 0.0606484489440918, 0.060543998718261716, 0.060644351959228515, 0.06063411331176758, 0.060581886291503906, 0.06059212875366211, 0.060672000885009764, 0.06064025497436523, 0.060680191040039064, 0.1343651885986328, 0.06059212875366211, 0.06057164764404297, 0.06062080001831055, 0.06056755065917969, 0.060633087158203126, 0.06053683090209961, 0.06060134506225586, 0.06061670303344727, 0.060595199584960936, 0.06065049743652344, 0.060642303466796874, 0.06053171157836914, 0.0606033935546875, 0.06058700942993164, 0.06060441589355469, 0.060576766967773435, 0.060662784576416016, 0.06059110260009765, 0.06068940734863281, 0.06059929656982422, 0.060624897003173826, 0.06051942443847656, 0.060624897003173826, 0.06059110260009765, 0.060644351959228515, 0.06058393478393555, 0.06063820648193359, 0.06062899017333984, 0.06070169448852539, 0.060633087158203126, 0.060674049377441405, 0.060647422790527344, 0.06063411331176758, 0.06060236740112305, 0.060633087158203126, 0.060649471282958986, 0.06068428802490235, 0.060614654541015625, 0.060641281127929686, 0.06069657516479492, 0.06066175842285156, 0.06067302322387695, 0.060644351959228515, 0.06066995239257812, 0.06064025497436523, 0.060655616760253904, 0.06065049743652344, 0.060660736083984375, 0.06073241424560547, 0.060633087158203126, 0.060674049377441405, 0.06063411331176758, 0.060680191040039064, 0.06072217559814453, 0.060614654541015625, 0.060624897003173826, 0.060649471282958986, 0.06065151977539063, 0.06063616180419922, 0.0607088623046875, 0.06064025497436523, 0.06068735885620117, 0.13449215698242187, 0.060625919342041014, 0.06061260986328125, 0.060611583709716796, 0.060521472930908204, 0.06062694549560547, 0.06057881546020508, 0.06061670303344727, 0.0606033935546875, 0.06058598327636719, 0.06053887939453125, 0.06060031890869141, 0.06055526351928711, 0.06065151977539063, 0.06101196670532227, 0.06065459060668945, 0.0605849609375, 0.06064025497436523, 0.060624897003173826, 0.060652542114257815, 0.06062899017333984, 0.060672000885009764, 0.06060441589355469, 0.060598270416259765, 0.06057369613647461, 0.06062899017333984, 0.06068428802490235, 0.060649471282958986, 0.06072012710571289, 0.06067609786987305, 0.06069452667236328, 0.060657665252685546, 0.06086656188964844, 0.06068633651733398, 0.06073241424560547, 0.06067507171630859, 0.060631038665771485, 0.060680191040039064, 0.06071091079711914, 0.060662784576416016, 0.060674049377441405, 0.06072217559814453, 0.06072012710571289, 0.06066995239257812, 0.06067609786987305, 0.060609535217285154, 0.06061363220214844, 0.060674049377441405, 0.06062182235717774, 0.060662784576416016, 0.0607191047668457, 0.060662784576416016, 0.06069145584106445, 0.060709888458251954, 0.06071295928955078, 0.060631038665771485, 0.06070169448852539, 0.060624897003173826, 0.060647422790527344, 0.0607191047668457, 0.060633087158203126, 0.060622848510742185, 0.060674049377441405]",tokens/s,16.209048570016403,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10558,7 +10558,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10833,7 +10833,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2107.244544,9028.763648,0.0,8382.316544,7939.414528,s,10,6.2175233154296885,0.6217523315429688,0.0006060610905763154,0.6219114074707031,0.622343017578125,0.6224655151367188,0.6225635131835937,"[0.6211746826171874, 0.6221463012695313, 0.62206103515625, 0.6223157958984376, 0.6221713256835938, 0.6207510375976563, 0.6225880126953125, 0.620806396484375, 0.6217617797851562, 0.6217469482421875]",tokens/s,411.7395094678596,kWh,7.3365436659918894e-06,4.020108750067747e-06,3.64477579163554e-05,4.780441033241504e-05,tokens/kWh,5355154.434912305,MB,2107.244544,9028.763648,0.0,8382.316544,8233.863168,s,10,338.38613671875004,33.838613671874995,0.0013682233241316137,33.838173828125,33.839473828125,33.8409205078125,33.8420778515625,"[33.8423671875, 33.83902734375, 33.83762890625, 33.8383828125, 33.83859765625, 33.8376875, 33.837484375, 33.83784375, 33.83796484375, 33.83915234375]",tokens/s,1.8617783994018209,kWh,0.00039950132826964066,0.00021896144419255187,0.0019980115918720536,0.002616474364334246,tokens/kWh,24078.20266033073,,s,629,343.44471649169924,0.5460170373476936,0.07393127168051418,0.53711669921875,0.5374224243164063,0.5375160522460938,1.1593396875,"[0.5371217651367187, 0.5371709594726563, 0.5366651000976562, 0.537037841796875, 0.536721435546875, 0.5372631225585938, 0.5367900390625, 0.5371453247070312, 0.5368391723632813, 0.5368822021484375, 0.5365924072265625, 0.53690673828125, 0.5367193603515625, 0.5368309936523438, 0.53667431640625, 0.536953857421875, 0.53671630859375, 0.537275390625, 0.53671728515625, 0.5374269409179687, 0.5368370971679688, 0.5372139282226562, 0.5369415893554688, 0.5373849487304687, 0.5370521850585938, 0.5373521728515624, 0.53697021484375, 0.5371996459960937, 0.53709619140625, 0.5371986083984375, 0.5375160522460938, 0.53732861328125, 0.5369405517578125, 0.5373890380859375, 0.5368350830078125, 0.539198486328125, 0.5370634155273437, 0.5373604125976562, 0.5372958984375, 0.5371893920898437, 0.5370286254882812, 0.5372313842773437, 0.5371893920898437, 0.53741259765625, 0.5372507934570312, 0.5379686279296875, 0.5368719482421875, 0.5375682373046875, 0.5370552368164062, 0.5373081665039062, 0.537153564453125, 0.537343994140625, 0.5372528686523438, 0.5376102294921875, 0.53725390625, 0.537523193359375, 0.5369896850585938, 0.537259033203125, 0.5368565673828125, 0.5373132934570313, 0.5371412353515626, 0.53785498046875, 1.159841796875, 0.537038818359375, 0.537017333984375, 0.53684326171875, 0.5368760375976562, 0.537006103515625, 0.5370521850585938, 0.5366651000976562, 0.5371259155273438, 0.53661181640625, 0.5370153198242188, 0.53686474609375, 0.5367869262695313, 0.5367296142578125, 0.5369548950195312, 0.5367306518554688, 0.5370265502929688, 0.537007080078125, 0.5371494140625, 0.5369845581054687, 0.5373634643554688, 0.5367992553710937, 0.5371412353515626, 0.5370480346679688, 0.5374586791992187, 0.5371259155273438, 0.5373778076171875, 0.5375211791992187, 0.5374515380859375, 0.5371729736328125, 0.5374013671875, 0.5369978637695313, 0.53728564453125, 0.5368043823242188, 0.5373153076171875, 0.5369825439453125, 0.537270263671875, 0.5371094970703125, 0.53711669921875, 0.5369528198242187, 0.5371340942382813, 0.5368822021484375, 0.5371781005859375, 0.5371904296875, 0.5374402465820313, 0.5372866821289063, 0.5374013671875, 0.536911865234375, 0.5372446899414063, 0.53709619140625, 0.5371760864257813, 0.5372518310546875, 0.5371791381835938, 0.5371320190429687, 0.5374412841796875, 0.537218017578125, 0.5374197998046875, 0.5370767211914063, 0.53718017578125, 0.5369876708984375, 0.5377146606445312, 0.5373378295898438, 0.5373132934570313, 1.15976708984375, 0.5368463134765625, 0.537006103515625, 0.5365985107421875, 0.5367725830078125, 0.5366599731445313, 0.5368688354492187, 0.5365924072265625, 0.5370224609375, 0.53661083984375, 0.536900634765625, 0.5370091552734375, 0.53690673828125, 0.5368411865234375, 0.5370726318359375, 0.5367490844726562, 0.536974365234375, 0.5368411865234375, 0.5371555786132812, 0.5371617431640625, 0.5373317260742188, 0.5367808227539063, 0.5371094970703125, 0.5370316772460938, 0.5373429565429687, 0.537143310546875, 0.5374402465820313, 0.5370582885742188, 0.5370162963867188, 0.5369845581054687, 0.5371781005859375, 0.537079833984375, 0.53718017578125, 0.5371064453125, 0.537385986328125, 0.5369876708984375, 0.53728564453125, 0.5370982666015625, 0.5374218139648438, 0.5371422729492188, 0.5370470581054687, 0.5369978637695313, 0.5371453247070312, 0.5370398559570313, 0.5374299926757813, 0.5371023559570313, 0.5374505004882812, 0.5369210815429688, 0.5371842651367188, 0.5370501098632813, 0.5371821899414062, 0.5369190673828125, 0.5373470458984375, 0.5372252197265625, 0.53749658203125, 0.5371637573242187, 0.5375364990234375, 0.5372713012695313, 0.5375160522460938, 0.5368955078125, 0.5375631103515625, 0.5373101806640626, 0.5374607543945312, 1.1603077392578125, 0.5368985595703125, 0.53696923828125, 0.5368391723632813, 0.5368319702148437, 0.5371904296875, 0.5369384765625, 0.5367337036132812, 0.5369794311523437, 0.5366517944335938, 0.537037841796875, 0.5368688354492187, 0.5368319702148437, 0.5369343872070312, 0.5368125610351563, 0.536690673828125, 0.53686376953125, 0.5367347412109374, 0.53703271484375, 0.536848388671875, 0.5371566162109375, 0.536901611328125, 0.5371791381835938, 0.5370736694335938, 0.53732763671875, 0.5371719970703125, 0.5374044189453125, 0.5371207885742187, 0.5372057495117187, 0.5371105346679688, 0.5373675537109375, 0.5369600219726562, 0.5372620849609375, 0.5369722900390625, 0.537396240234375, 0.5369630737304687, 0.5375784912109375, 0.5370101928710938, 0.5372200927734375, 0.5372108764648438, 0.537186279296875, 0.5369098510742187, 0.5371688842773438, 0.5370357666015625, 0.5374678955078125, 0.5371658325195312, 0.5376061401367187, 0.53692724609375, 0.5371392211914062, 0.5370582885742188, 0.5371289672851562, 0.5370029907226562, 0.5373409423828125, 0.537228271484375, 0.5373900756835938, 0.5372395629882812, 0.5374474487304688, 0.5372374877929688, 0.5372262573242188, 0.5368739624023438, 0.5374719848632813, 0.5371740112304687, 0.5377188110351563, 1.159825439453125, 0.536869873046875, 0.5369948120117187, 0.5368688354492187, 0.5370460205078125, 0.536690673828125, 0.536985595703125, 0.5366886596679687, 0.5370675048828125, 0.5364070434570313, 0.53718017578125, 0.5367019653320313, 0.5370408935546875, 0.5369384765625, 0.5369620361328125, 0.5367838745117187, 0.5370101928710938, 0.536869873046875, 0.537027587890625, 0.5369927978515625, 0.5370398559570313, 0.5367500610351562, 0.5371637573242187, 0.5370224609375, 0.5372241821289062, 0.5369630737304687, 0.5373501586914062, 0.5369835815429688, 0.5371658325195312, 0.5370562744140625, 0.5373265991210937, 0.5370470581054687, 0.5372631225585938, 0.5368872680664063, 0.5373521728515624, 0.5371197509765625, 0.5373317260742188, 0.5371627807617188, 0.5372139282226562, 0.5371361083984375, 0.5373163452148437, 0.5369282836914062, 0.5371760864257813, 0.5369364624023437, 0.5373532104492188, 0.5373030395507813, 0.5374054565429688, 0.5369262084960937, 0.5374248657226562, 0.5371453247070312, 0.5373746948242187, 0.537185302734375, 0.537692138671875, 0.5372098388671875, 0.5374781494140625, 0.5371146240234375, 0.5374945068359375, 0.5370982666015625, 0.5371719970703125, 0.537037841796875, 0.5375795288085937, 0.5371279296875, 0.5373051147460938, 1.1591044921875, 0.5368514404296875, 0.5373880615234375, 0.5367183227539063, 0.5368985595703125, 0.5367828369140625, 0.5369292602539063, 0.5367296142578125, 0.5370357666015625, 0.5368729858398438, 0.5370439453125, 0.5367787475585938, 0.5368627319335938, 0.5367992553710937, 0.5369036865234375, 0.5367838745117187, 0.5368934326171875, 0.5367030029296875, 0.53711669921875, 0.5368422241210937, 0.5371361083984375, 0.5369579467773438, 0.5373798217773438, 0.5369467163085937, 0.5371658325195312, 0.537038818359375, 0.5374607543945312, 0.5370890502929687, 0.5372354736328125, 0.53711669921875, 0.5372088623046875, 0.5368576049804688, 0.5372252197265625, 0.5368668212890625, 0.5373583374023437, 0.5369968872070312, 0.5375211791992187, 0.5370316772460938, 0.5372211303710938, 0.537091064453125, 0.5372241821289062, 0.5368074340820312, 0.5371074829101562, 0.537154541015625, 0.5373409423828125, 0.5372426147460938, 0.5373746948242187, 0.53696923828125, 0.5373778076171875, 0.5370316772460938, 0.5371064453125, 0.5371883544921875, 0.5374105834960937, 0.5373143310546875, 0.5374822387695313, 0.5371658325195312, 0.537438232421875, 0.5372293090820313, 0.5373399047851563, 0.5369476928710938, 0.5372938232421876, 0.53732861328125, 0.5371525268554688, 1.1606148681640625, 0.5368709106445313, 0.537049072265625, 0.5368524780273437, 0.5368616943359374, 0.5371586303710938, 0.5369528198242187, 0.5366599731445313, 0.537049072265625, 0.5364971313476562, 0.5371094970703125, 0.536690673828125, 0.5367777099609375, 0.5367633666992188, 0.53699072265625, 0.5371033325195312, 0.5367920532226562, 0.5366876220703125, 0.5370706176757812, 0.5370224609375, 0.5373143310546875, 0.5367756958007812, 0.5372948608398438, 0.5370153198242188, 0.5372548828125, 0.536932373046875, 0.5373552856445313, 0.5369825439453125, 0.5371002807617188, 0.537217041015625, 0.537332763671875, 0.5370439453125, 0.5372569580078125, 0.5368811645507813, 0.5373009643554687, 0.5369774169921875, 0.5372927856445312, 0.53722216796875, 0.5372958984375, 0.536995849609375, 0.5370992431640625, 0.5370265502929688, 0.5373112182617188, 0.5370603637695313, 0.5374248657226562, 0.5372364501953125, 0.537511962890625, 0.5369671630859375, 0.5372569580078125, 0.5368422241210937, 0.5370869750976562, 0.5370029907226562, 0.53732861328125, 0.5371924438476563, 0.5372190551757813, 0.537006103515625, 0.5375999755859375, 0.5372948608398438, 0.5373081665039062, 0.5367767333984375, 0.53734912109375, 0.537365478515625, 0.5374453735351562, 1.1584112548828125, 0.536911865234375, 0.5371248779296875, 0.5366497192382812, 0.53728564453125, 0.5368248291015625, 0.536859619140625, 0.536838134765625, 0.5368760375976562, 0.5366333618164062, 0.5369682006835937, 0.5368125610351563, 0.5369620361328125, 0.5368678588867187, 0.5370029907226562, 0.536616943359375, 0.5369866333007812, 0.5368657836914063, 0.537059326171875, 0.536795166015625, 0.53722216796875, 0.5368893432617188, 0.5371883544921875, 0.5369292602539063, 0.5370715942382812, 0.5371924438476563, 0.537575439453125, 0.5368770751953125, 0.537238525390625, 0.5371658325195312, 0.5372815551757812, 0.537037841796875, 0.5371259155273438, 0.5369180297851562, 0.537354248046875, 0.5370726318359375, 0.537275390625, 0.5371340942382813, 0.537291748046875, 0.537017333984375, 0.5371401977539062, 0.53724365234375, 0.5372682495117187, 0.5370900268554688, 0.5373521728515624, 0.5370685424804688, 0.5372989501953125, 0.536911865234375, 0.5372057495117187, 0.5371248779296875, 0.5373235473632813, 0.5369395141601563, 0.537427978515625, 0.5376696166992188, 0.537343994140625, 0.5369886474609376, 0.537523193359375, 0.5372200927734375, 0.5373306884765625, 0.5369682006835937, 0.5373040771484375, 0.5371627807617188, 0.537533447265625, 1.15943115234375, 0.5369047241210938, 0.5374474487304688, 0.5369364624023437, 0.5368299560546875, 0.53684326171875, 0.5372139282226562, 0.5367030029296875, 0.5370921020507813, 0.5367357177734375, 0.5369508056640625, 0.536848388671875, 0.5369784545898437, 0.5365545043945312, 0.5367613525390625, 0.536711181640625, 0.5370614013671875, 0.5371105346679688, 0.5371883544921875, 0.5368330078125, 0.5372927856445312, 0.5367603149414063, 0.5372252197265625, 0.53699072265625, 0.5373624267578125, 0.5371207885742187, 0.537406494140625, 0.5369968872070312, 0.5372323608398437, 0.537059326171875, 0.5373685913085937, 0.5369948120117187, 0.5372723388671875, 0.5368565673828125, 0.5371392211914062, 0.5369763793945312, 0.5374299926757813, 0.5370664672851563, 0.5371392211914062, 0.5373184204101562, 0.5369876708984375, 0.5368934326171875, 0.5372743530273437, 0.5370818481445313, 0.537354248046875, 0.537133056640625, 0.537364501953125, 0.5369241333007813, 0.5372211303710938, 0.537133056640625, 0.537238525390625, 0.5370582885742188, 0.5374187622070312, 0.5373501586914062, 0.5373060913085937, 0.5370634155273437, 0.5375375366210937, 0.5371637573242187, 0.5374136352539063, 0.5369292602539063, 0.5373798217773438, 0.5370337524414063, 0.5373726806640625, 1.159437255859375, 0.53688525390625, 0.5370245361328125, 0.5368156127929687, 0.5369190673828125, 0.5368115234375, 0.5373972778320313, 0.5369129028320313, 0.5371268920898438, 0.5366988525390625, 0.5368770751953125, 0.5367101440429688, 0.5367337036132812, 0.536869873046875, 0.5367859497070312, 0.5366343383789063, 0.5368739624023438, 0.5368955078125, 0.5373378295898438, 0.5368309936523438, 0.5371996459960937, 0.5370316772460938, 0.53713818359375, 0.5369682006835937, 0.5373306884765625, 0.5371463623046875, 0.5374033813476562, 0.5371002807617188, 0.5371340942382813, 0.537112548828125, 0.5374371948242187, 0.5369467163085937, 0.537270263671875, 0.5368545532226563, 0.5374730224609375, 0.5369088134765625, 0.5374105834960937, 0.5371248779296875, 0.5373726806640625, 0.537154541015625, 0.53719140625, 0.5370562744140625, 0.537333740234375, 0.5370787963867187, 0.5373941650390625, 0.53711767578125, 0.5374453735351562, 0.5370869750976562, 0.537343994140625, 0.5371422729492188, 0.5372979125976562, 0.5372733154296875, 0.5372272338867188, 0.53725390625, 0.5373798217773438, 0.5372610473632813, 0.5374996337890625, 0.5372794799804688, 0.537186279296875, 0.5369476928710938, 0.5374187622070312, 0.5371514892578125, 0.53749658203125]",tokens/s,1.8314446832237186,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10862,7 +10862,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11363,7 +11363,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1852.063744,3368.5504,0.0,2722.103296,2585.357312,s,10,0.7361118698120117,0.07361118698120116,0.0003642046363895853,0.07361937713623047,0.07414464721679688,0.07420198898315429,0.07424786239624023,"[0.07425933074951171, 0.07323929595947265, 0.07386649322509765, 0.07366835021972656, 0.07318646240234375, 0.07325574493408203, 0.07413190460205078, 0.07326553344726562, 0.07365193939208985, 0.0735868148803711]",tokens/s,3477.732264599637,kWh,8.660967405115021e-07,4.745775110883439e-07,4.11042299636542e-06,5.451097247965265e-06,tokens/kWh,46963022.00360804,MB,1852.452864,3368.5504,0.0,2722.103296,2628.769792,s,10,38.912985107421875,3.8912985107421876,0.002125353517438014,3.89062548828125,3.8938447753906247,3.895329248046875,3.896516826171875,"[3.896813720703125, 3.89010009765625, 3.890788330078125, 3.890443603515625, 3.88959716796875, 3.890462646484375, 3.890949951171875, 3.890897216796875, 3.893514892578125, 3.88941748046875]",tokens/s,16.189968419560802,kWh,4.5877188203519475e-05,2.5143288266442475e-05,0.00021536239147063786,0.00028638286794059977,tokens/kWh,219985.22625685547,,s,629,39.51227389144898,0.06281760555079327,0.008801190571361054,0.061709312438964846,0.06187274169921875,0.06235340805053711,0.13552583007812502,"[0.06250291061401367, 0.0624824333190918, 0.06248860931396484, 0.06241788864135742, 0.06238412857055664, 0.062355457305908205, 0.06268928146362304, 0.06286848068237305, 0.061916160583496097, 0.061690879821777345, 0.06164377593994141, 0.06167552185058594, 0.06169702529907226, 0.061725696563720706, 0.06180147171020508, 0.06190796661376953, 0.06164275360107422, 0.06171340942382812, 0.06235033416748047, 0.06176665496826172, 0.06160793685913086, 0.06171750259399414, 0.061628414154052735, 0.06169497680664063, 0.06164275360107422, 0.061661182403564455, 0.06167244720458984, 0.0616806411743164, 0.061723648071289064, 0.061810817718505856, 0.06176448059082031, 0.06172876739501953, 0.06162124633789062, 0.06174105453491211, 0.06160793685913086, 0.06174924850463867, 0.06167859268188477, 0.0617625617980957, 0.06171340942382812, 0.062069759368896485, 0.06172172927856445, 0.06168972778320313, 0.06166220855712891, 0.06177689743041992, 0.06163558578491211, 0.06175948715209961, 0.061769729614257814, 0.06181683349609375, 0.06176665496826172, 0.061758464813232425, 0.06166425704956055, 0.061740032196044924, 0.061720577239990235, 0.06168371200561523, 0.06170640182495117, 0.06162825775146484, 0.061742080688476565, 0.06170009613037109, 0.06172774505615235, 0.061704193115234375, 0.061769729614257814, 0.06176870346069336, 0.13554074096679689, 0.06162432098388672, 0.06162432098388672, 0.06162134552001953, 0.06155356979370117, 0.06167654418945313, 0.06153932952880859, 0.06170521545410156, 0.06169804763793945, 0.06160486221313476, 0.06173798370361328, 0.061633537292480466, 0.06171964645385742, 0.06159964752197265, 0.061695999145507815, 0.061778942108154294, 0.06163148880004883, 0.061780990600585936, 0.061868030548095705, 0.06245068740844727, 0.06177689743041992, 0.06273433685302734, 0.06171750259399414, 0.061677566528320314, 0.06168678283691406, 0.061638656616210936, 0.061682689666748045, 0.06182297515869141, 0.061917182922363284, 0.06169702529907226, 0.06170624160766602, 0.06173286437988281, 0.061692928314208986, 0.061625343322753906, 0.06172774505615235, 0.06163455963134765, 0.061687808990478515, 0.061682815551757815, 0.06168051147460937, 0.061712383270263675, 0.06174822235107422, 0.061663230895996096, 0.06169001770019531, 0.061602657318115234, 0.06268313598632813, 0.06171955108642578, 0.061693984985351565, 0.061719520568847656, 0.06171955108642578, 0.061679615020751956, 0.06182809448242187, 0.06170521545410156, 0.061693950653076174, 0.06171340942382812, 0.06160179138183594, 0.06178508758544922, 0.06174822235107422, 0.061707393646240234, 0.061671295166015626, 0.06169721603393555, 0.06164358520507812, 0.06166527938842774, 0.061709312438964846, 0.13617971801757814, 0.06163046264648438, 0.06165401458740234, 0.06161407852172852, 0.06157516860961914, 0.061656063079833984, 0.06166016006469727, 0.061707263946533204, 0.061707263946533204, 0.06162944030761719, 0.06173388671875, 0.06161932754516602, 0.06166515350341797, 0.061652992248535154, 0.06172979354858398, 0.06173798370361328, 0.06173183822631836, 0.06171340942382812, 0.06171750259399414, 0.061639678955078124, 0.06162432098388672, 0.06173183822631836, 0.06170828628540039, 0.06276300811767578, 0.061687808990478515, 0.061636608123779295, 0.06173183822631836, 0.06161612701416016, 0.0616806411743164, 0.06161203384399414, 0.06170009613037109, 0.06170316696166992, 0.06180966567993164, 0.061895679473876954, 0.06174105453491211, 0.061710334777832034, 0.06173081588745117, 0.06171660614013672, 0.061677440643310544, 0.061674495697021485, 0.0617441291809082, 0.061692928314208986, 0.0617625617980957, 0.06161612701416016, 0.06178611373901367, 0.06169926452636719, 0.061711166381835936, 0.0628408317565918, 0.06211075210571289, 0.062286815643310546, 0.06192435073852539, 0.061740032196044924, 0.0617534065246582, 0.06165497589111328, 0.061690879821777345, 0.06172671890258789, 0.06164889526367188, 0.06174310302734375, 0.06171340942382812, 0.061661182403564455, 0.061671424865722656, 0.061740032196044924, 0.06165708923339844, 0.1359667205810547, 0.06163763046264648, 0.06169497680664063, 0.061628414154052735, 0.061633537292480466, 0.061764606475830076, 0.06156185531616211, 0.061704193115234375, 0.0617891845703125, 0.06193459320068359, 0.06180044937133789, 0.061663230895996096, 0.0622295036315918, 0.06165913772583008, 0.06168166351318359, 0.061690879821777345, 0.06174924850463867, 0.06168371200561523, 0.061770751953125, 0.0616519660949707, 0.06170111846923828, 0.06165913772583008, 0.061707263946533204, 0.0617441291809082, 0.06170009613037109, 0.06163148880004883, 0.06163558578491211, 0.061644798278808595, 0.06166425704956055, 0.06169190216064453, 0.061710334777832034, 0.06169497680664063, 0.061644798278808595, 0.06169497680664063, 0.061671520233154295, 0.06164572906494141, 0.06277836990356446, 0.06170009613037109, 0.06170009613037109, 0.06167859268188477, 0.06167660903930664, 0.06171231842041015, 0.06174924850463867, 0.06165094375610351, 0.06181478500366211, 0.061690879821777345, 0.06172671890258789, 0.06169190216064453, 0.06179020690917969, 0.06173388671875, 0.06178508758544922, 0.061685760498046874, 0.061876224517822265, 0.062241790771484375, 0.06170828628540039, 0.06173081588745117, 0.06157823944091797, 0.061774848937988285, 0.06169497680664063, 0.06177382278442383, 0.06168678283691406, 0.06180556869506836, 0.06197145462036133, 0.13539430236816405, 0.06167552185058594, 0.061671424865722656, 0.06164070510864258, 0.061591552734375, 0.06165401458740234, 0.06159667205810547, 0.06172876739501953, 0.06168678283691406, 0.06160281753540039, 0.061707263946533204, 0.06159667205810547, 0.0617154541015625, 0.06167244720458984, 0.061693950653076174, 0.0616519660949707, 0.06165094375610351, 0.06250291061401367, 0.06174720001220703, 0.06172159957885742, 0.06170009613037109, 0.061649921417236325, 0.061625343322753906, 0.06166732788085937, 0.06170111846923828, 0.061608959197998046, 0.06172774505615235, 0.06161305618286133, 0.06164889526367188, 0.061679615020751956, 0.06164070510864258, 0.06172265625, 0.061671390533447265, 0.061677566528320314, 0.061658111572265625, 0.061710334777832034, 0.06173183822631836, 0.061690879821777345, 0.061764606475830076, 0.06166835021972656, 0.06195308685302734, 0.06167443084716797, 0.06182809448242187, 0.06171244812011719, 0.061844417572021484, 0.06171750259399414, 0.06164582443237305, 0.06171340942382812, 0.06173593521118164, 0.06290636825561523, 0.06180249786376953, 0.06168678283691406, 0.061709312438964846, 0.06173183822631836, 0.061699073791503904, 0.061742080688476565, 0.061656063079833984, 0.061811710357666014, 0.061736961364746094, 0.061709312438964846, 0.061710334777832034, 0.061693950653076174, 0.06191308975219727, 0.13563392639160157, 0.0618526725769043, 0.061897727966308595, 0.06170316696166992, 0.06156803131103516, 0.061701087951660155, 0.06166128158569336, 0.06172867202758789, 0.0617154541015625, 0.061690879821777345, 0.06177996826171875, 0.06166527938842774, 0.06173183822631836, 0.061636608123779295, 0.0617523193359375, 0.06169702529907226, 0.0616673583984375, 0.061728736877441404, 0.061797374725341796, 0.06189875030517578, 0.0617441291809082, 0.06164275360107422, 0.061674495697021485, 0.06159360122680664, 0.06164796829223633, 0.06164368057250977, 0.061792255401611325, 0.06162739181518555, 0.061645023345947264, 0.06166812896728516, 0.06170115280151367, 0.06169699096679687, 0.061623294830322264, 0.06330368041992188, 0.06182092666625977, 0.06176051330566406, 0.06183628845214844, 0.06172774505615235, 0.06193971252441406, 0.061797374725341796, 0.0616888313293457, 0.06170329666137695, 0.06172966384887695, 0.06168166351318359, 0.06176371383666992, 0.061695873260498045, 0.061770751953125, 0.061723648071289064, 0.06177689743041992, 0.06173491287231445, 0.06175641632080078, 0.06165708923339844, 0.06165401458740234, 0.06177382278442383, 0.06164684677124024, 0.06175641632080078, 0.06164377593994141, 0.06168166351318359, 0.06175027084350586, 0.061699073791503904, 0.06170009613037109, 0.061710334777832034, 0.06166527938842774, 0.13643980407714842, 0.061716545104980466, 0.061607872009277344, 0.06166732788085937, 0.0616099853515625, 0.06168371200561523, 0.06158540725708008, 0.06174105453491211, 0.061677566528320314, 0.061644798278808595, 0.061795326232910154, 0.06170624160766602, 0.06171648025512695, 0.06165919876098633, 0.06173894500732422, 0.061732929229736326, 0.06168876647949219, 0.06168377685546875, 0.0616693115234375, 0.06164275360107422, 0.06164377593994141, 0.06178201675415039, 0.06199398422241211, 0.06298828887939453, 0.06168985748291016, 0.06167871856689453, 0.061652862548828125, 0.06170316696166992, 0.061723648071289064, 0.06167244720458984, 0.06170934295654297, 0.061823966979980466, 0.06195199966430664, 0.061687808990478515, 0.06180352020263672, 0.061764606475830076, 0.061685760498046874, 0.061674495697021485, 0.061693950653076174, 0.06170624160766602, 0.06166835021972656, 0.06171443176269531, 0.06177382278442383, 0.06163148880004883, 0.061774848937988285, 0.0616734733581543, 0.06170624160766602, 0.061707263946533204, 0.06180044937133789, 0.06171340942382812, 0.06172671890258789, 0.061739009857177736, 0.06166220855712891, 0.06182505416870117, 0.06167343902587891, 0.061732929229736326, 0.06168364715576172, 0.06287052917480469, 0.06172166442871094, 0.06172563171386719, 0.06172159957885742, 0.0617441291809082, 0.061709312438964846, 0.13548748779296876, 0.061685760498046874, 0.06184550476074219, 0.06171340942382812, 0.061636608123779295, 0.06169295883178711, 0.06171337509155273, 0.06171340942382812, 0.06171443176269531, 0.06175743865966797, 0.06177996826171875, 0.06164684677124024, 0.0617410888671875, 0.06167958450317383, 0.06168371200561523, 0.06167859268188477, 0.06231347274780273, 0.061669376373291014, 0.06177791976928711, 0.06170111846923828, 0.06169190216064453, 0.06170214462280273, 0.06165401458740234, 0.06170009613037109, 0.06169190216064453, 0.06161612701416016, 0.061682689666748045, 0.06170521545410156, 0.06170316696166992, 0.06166220855712891, 0.06175436782836914, 0.0616888313293457, 0.06170624160766602, 0.0616734733581543, 0.061725696563720706, 0.061830142974853515, 0.06189158248901367, 0.06173798370361328, 0.06172876739501953, 0.06174310302734375, 0.061740032196044924, 0.06168371200561523, 0.06173183822631836, 0.06170316696166992, 0.06178201675415039, 0.061722686767578125, 0.06168569564819336, 0.06171852874755859, 0.061712383270263675, 0.06175955200195313, 0.061817790985107424, 0.062246910095214845, 0.06236262512207031, 0.06189673614501953, 0.06172975921630859, 0.061764606475830076, 0.06163455963134765, 0.06180681610107422, 0.06167631912231445, 0.06176870346069336, 0.06178611373901367, 0.06176563262939453, 0.0617891845703125, 0.13594931030273438, 0.06170828628540039, 0.061732894897460935, 0.061699039459228514, 0.061656063079833984, 0.0616376953125, 0.06158432006835937, 0.06172467041015625, 0.06168473434448242, 0.06169497680664063, 0.06175641632080078, 0.06178620910644531, 0.061687713623046876, 0.0616806411743164, 0.06187007904052735, 0.06170214462280273, 0.06175641632080078, 0.06246201705932617, 0.06178400039672852, 0.061663230895996096, 0.06180352020263672, 0.06171750259399414, 0.06183731079101563, 0.061625343322753906, 0.061712383270263675, 0.061710334777832034, 0.06167244720458984, 0.06171443176269531, 0.06174723052978515, 0.06172361755371094, 0.061871135711669925, 0.061722591400146486, 0.06174617767333984, 0.061699073791503904, 0.061830142974853515, 0.06177996826171875, 0.061927425384521485, 0.06182195281982422, 0.061843456268310545, 0.061658206939697265, 0.06172662353515625, 0.06186086273193359, 0.06174310302734375, 0.06176153564453125, 0.06186905670166016, 0.06178815841674805, 0.06171340942382812, 0.06176563262939453, 0.061828289031982425, 0.0616824951171875, 0.061835262298583986, 0.0634439697265625, 0.06188544082641602, 0.06187519836425781, 0.06173081588745117, 0.06184243011474609, 0.06188544082641602, 0.06187212753295898, 0.06179020690917969, 0.06177587127685547, 0.0616888313293457, 0.06170521545410156, 0.0617441291809082, 0.13600973510742187, 0.062129150390625, 0.06186905670166016, 0.0616888313293457, 0.06164889526367188, 0.061658111572265625, 0.06158335876464844, 0.06171852874755859, 0.06171340942382812, 0.06164684677124024, 0.061730880737304684, 0.06165702438354492, 0.06167552185058594, 0.061641727447509766, 0.061739009857177736, 0.06163455963134765, 0.06174617767333984, 0.061677566528320314, 0.06170828628540039, 0.061725696563720706, 0.061671424865722656, 0.06169702529907226, 0.06174515151977539, 0.061661182403564455, 0.06163046264648438, 0.06156185531616211, 0.06298624038696289, 0.06177280044555664, 0.06197350311279297, 0.061710334777832034, 0.06174105453491211, 0.061628414154052735, 0.06164787292480469, 0.0615997428894043, 0.061770751953125, 0.06168678283691406, 0.06174515151977539, 0.06170111846923828, 0.06171955108642578, 0.061655040740966796, 0.06164377593994141, 0.06171852874755859, 0.061695999145507815, 0.061661182403564455, 0.06176153564453125, 0.06170316696166992, 0.061671424865722656, 0.061693023681640625, 0.061720481872558595, 0.06173593521118164, 0.06173081588745117, 0.061682689666748045, 0.061692928314208986, 0.06175539016723633, 0.061644798278808595, 0.06165708923339844, 0.06165404891967773, 0.06168368148803711, 0.06167552185058594, 0.061703231811523436, 0.061744064331054685, 0.06178303909301758, 0.06170316696166992]",tokens/s,15.919104066954871,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1490.198528,1841.823744,0.0,1195.37664,1022.155264,s,10,0.6871841278076172,0.06871841278076171,0.00033427755700674906,0.06862833786010741,0.06882823104858399,0.06926301155090332,0.06961083595275878,"[0.06969779205322266, 0.06856297302246094, 0.06858707427978515, 0.06850640106201172, 0.0686982421875, 0.06862300872802735, 0.0686336669921875, 0.06864895629882813, 0.0687316131591797, 0.06849440002441406]",tokens/s,3725.347976483958,kWh,8.097485796501648e-07,4.4370366405853753e-07,3.761924622931496e-06,5.0153768666401985e-06,tokens/kWh,51043023.64649507,MB,1490.198528,1841.823744,0.0,1195.37664,1091.597312,s,10,35.169312255859374,3.5169312255859375,0.00128683836079331,3.516599853515625,3.5173736328125,3.5190165283203125,3.5203308447265624,"[3.520659423828125, 3.51630859375, 3.515842041015625, 3.516979248046875, 3.5166650390625, 3.51653466796875, 3.516714599609375, 3.516245361328125, 3.517008544921875, 3.516354736328125]",tokens/s,17.91334432179688,kWh,4.1521653622797086e-05,2.2756081263433605e-05,0.0001967916524864682,0.0002610693873726989,tokens/kWh,241315.15622726805,,s,629,35.7307012748718,0.056805566414740584,0.008168059089268582,0.05580799865722656,0.055887052917480466,0.055967333984375,0.12452544342041016,"[0.05626367950439453, 0.05630361557006836, 0.05626367950439453, 0.056338432312011716, 0.05623910522460938, 0.05629337692260742, 0.05628313446044922, 0.056346622467041016, 0.05584896087646484, 0.05588991928100586, 0.05574758529663086, 0.055757823944091796, 0.05575884628295898, 0.055839744567871094, 0.05575680160522461, 0.055742462158203124, 0.0557209587097168, 0.05613260650634765, 0.05585203170776367, 0.0557844467163086, 0.055812095642089846, 0.05583769607543945, 0.05573529434204102, 0.05577523040771484, 0.05584588623046875, 0.05580595016479492, 0.0557946891784668, 0.055825408935546876, 0.05578137588500977, 0.05588684844970703, 0.055812095642089846, 0.055757823944091796, 0.05581414413452149, 0.05580595016479492, 0.05578137588500977, 0.05587353515625, 0.05582745742797852, 0.055779327392578126, 0.05571379089355469, 0.05586943817138672, 0.055806976318359375, 0.05578137588500977, 0.05583871841430664, 0.055760894775390625, 0.05582233428955078, 0.05581721496582031, 0.05575372695922851, 0.05586022567749024, 0.055790592193603515, 0.0557946891784668, 0.05580492782592773, 0.055793663024902344, 0.055809024810791016, 0.055818241119384764, 0.05578342437744141, 0.05583462524414062, 0.0558653450012207, 0.055798782348632815, 0.05585715103149414, 0.05593804931640625, 0.055790592193603515, 0.055820289611816405, 0.12452658843994141, 0.055755775451660154, 0.055787521362304686, 0.05579776000976563, 0.05577830505371094, 0.05574758529663086, 0.05581414413452149, 0.055787521362304686, 0.05573017501831055, 0.05575680160522461, 0.05580492782592773, 0.05572915267944336, 0.05578649520874023, 0.05576806259155274, 0.0557946891784668, 0.05575167846679688, 0.05578956985473633, 0.05579673767089844, 0.05581926345825195, 0.05581721496582031, 0.05590323257446289, 0.05579776000976563, 0.05584588623046875, 0.055739391326904295, 0.055790592193603515, 0.05578137588500977, 0.055836673736572265, 0.05580083084106445, 0.055825408935546876, 0.05577011108398437, 0.055809024810791016, 0.05585715103149414, 0.055793663024902344, 0.05574758529663086, 0.055787521362304686, 0.05576499176025391, 0.055809024810791016, 0.055766014099121096, 0.05583462524414062, 0.05594828796386719, 0.05584896087646484, 0.05576294326782227, 0.05584281539916992, 0.05577830505371094, 0.055858177185058595, 0.0560076789855957, 0.05585408020019531, 0.05578342437744141, 0.05575065612792969, 0.055859233856201174, 0.05579670333862305, 0.055782398223876956, 0.05586739349365234, 0.05582950210571289, 0.055809024810791016, 0.05578342437744141, 0.0558551025390625, 0.0561715202331543, 0.055932926177978515, 0.05579776000976563, 0.05574553680419922, 0.0558131217956543, 0.05583052825927735, 0.12455526733398438, 0.0557916145324707, 0.05582950210571289, 0.05575065612792969, 0.05580799865722656, 0.05575987243652344, 0.0558131217956543, 0.055723007202148435, 0.05578137588500977, 0.055793663024902344, 0.05577523040771484, 0.05576806259155274, 0.055795711517333986, 0.05574655914306641, 0.05582745742797852, 0.05574860763549805, 0.05578137588500977, 0.05572608184814453, 0.05576806259155274, 0.055809024810791016, 0.0557916145324707, 0.05582438278198242, 0.05580083084106445, 0.05575475311279297, 0.05576192092895508, 0.05580083084106445, 0.055844863891601565, 0.05580492782592773, 0.055823360443115234, 0.05579673767089844, 0.05576294326782227, 0.055785472869873044, 0.05581619262695312, 0.05576806259155274, 0.05587251281738281, 0.05579776000976563, 0.0557844467163086, 0.05576294326782227, 0.055839744567871094, 0.05573017501831055, 0.055815166473388675, 0.0558766098022461, 0.05590220642089844, 0.05579776000976563, 0.055818241119384764, 0.055841793060302736, 0.05584588623046875, 0.05583564758300781, 0.05581619262695312, 0.05581619262695312, 0.05587148666381836, 0.05586636734008789, 0.055820289611816405, 0.05580595016479492, 0.05584998321533203, 0.055793663024902344, 0.055785472869873044, 0.05583052825927735, 0.055785472869873044, 0.0557916145324707, 0.05583564758300781, 0.05590323257446289, 0.05586636734008789, 0.12449382019042969, 0.05575372695922851, 0.05576499176025391, 0.05578649520874023, 0.05577318572998047, 0.05608038330078125, 0.0558766098022461, 0.055744510650634765, 0.055757823944091796, 0.05576704025268555, 0.05586739349365234, 0.0557762565612793, 0.05626163101196289, 0.05574860763549805, 0.05581107330322266, 0.05573836898803711, 0.0557916145324707, 0.05576396942138672, 0.05574348831176758, 0.05576499176025391, 0.0557844467163086, 0.05575475311279297, 0.05585203170776367, 0.05574041748046875, 0.05575884628295898, 0.05578342437744141, 0.05584998321533203, 0.0558131217956543, 0.05589913558959961, 0.05588684844970703, 0.05580492782592773, 0.0557762565612793, 0.05579776000976563, 0.05576806259155274, 0.05586431884765625, 0.0557916145324707, 0.05583052825927735, 0.055790592193603515, 0.05582438278198242, 0.05580799865722656, 0.055820289611816405, 0.05578137588500977, 0.055798782348632815, 0.0558551025390625, 0.055766014099121096, 0.055825408935546876, 0.055982078552246094, 0.055766014099121096, 0.055793663024902344, 0.05594214248657227, 0.05587251281738281, 0.055806976318359375, 0.05583257675170898, 0.055965694427490234, 0.05590630340576172, 0.0557946891784668, 0.05584588623046875, 0.05582438278198242, 0.05581926345825195, 0.05581414413452149, 0.055809024810791016, 0.055823360443115234, 0.05583564758300781, 0.12459212493896485, 0.05578956985473633, 0.05584588623046875, 0.055777278900146485, 0.055769088745117185, 0.05573734283447265, 0.055818241119384764, 0.055785472869873044, 0.05583052825927735, 0.0557844467163086, 0.05582438278198242, 0.055782398223876956, 0.05585100936889648, 0.05584281539916992, 0.05582131195068359, 0.05576499176025391, 0.0557844467163086, 0.055757823944091796, 0.05580287933349609, 0.055757823944091796, 0.055841793060302736, 0.05579673767089844, 0.055818241119384764, 0.055736320495605465, 0.055801856994628904, 0.05581721496582031, 0.05588582229614258, 0.05578649520874023, 0.05585919952392578, 0.055790592193603515, 0.055836673736572265, 0.05586329650878906, 0.055874561309814455, 0.0558653450012207, 0.055793663024902344, 0.05587046432495117, 0.05577523040771484, 0.05574860763549805, 0.055839744567871094, 0.05576396942138672, 0.05587558364868164, 0.05575884628295898, 0.055861248016357425, 0.055810047149658204, 0.056005630493164066, 0.055798782348632815, 0.055818241119384764, 0.055913471221923826, 0.05584691238403321, 0.05581721496582031, 0.05577830505371094, 0.05587865447998047, 0.05581414413452149, 0.05581414413452149, 0.0557916145324707, 0.05582339096069336, 0.05581923294067383, 0.05581107330322266, 0.055856128692626954, 0.05586636734008789, 0.05587148666381836, 0.055801856994628904, 0.05587558364868164, 0.12469657897949218, 0.055790592193603515, 0.055731201171875, 0.05574860763549805, 0.055801856994628904, 0.055787521362304686, 0.0557844467163086, 0.0557496337890625, 0.055774208068847655, 0.0557844467163086, 0.05589913558959961, 0.055757823944091796, 0.055774208068847655, 0.05576806259155274, 0.05582745742797852, 0.05572710418701172, 0.055798782348632815, 0.055744510650634765, 0.055760894775390625, 0.05584281539916992, 0.05578035354614258, 0.05584384155273438, 0.05585919952392578, 0.05575680160522461, 0.055766014099121096, 0.05579673767089844, 0.055812095642089846, 0.0557946891784668, 0.05587353515625, 0.055825408935546876, 0.055828479766845705, 0.055782398223876956, 0.055831550598144535, 0.055787521362304686, 0.055877632141113284, 0.055788543701171874, 0.05577318572998047, 0.0557916145324707, 0.055841793060302736, 0.055771137237548826, 0.055823360443115234, 0.05577318572998047, 0.05582131195068359, 0.055801856994628904, 0.05582950210571289, 0.055844863891601565, 0.05596672058105469, 0.055787521362304686, 0.05583052825927735, 0.055897087097167966, 0.055853057861328125, 0.05585100936889648, 0.05584896087646484, 0.05590220642089844, 0.055806976318359375, 0.056011775970458984, 0.05590118408203125, 0.05583564758300781, 0.055809024810791016, 0.055826431274414064, 0.055774208068847655, 0.055810047149658204, 0.055856128692626954, 0.12463922882080078, 0.05575680160522461, 0.055785472869873044, 0.05579673767089844, 0.05581107330322266, 0.05583257675170898, 0.055858177185058595, 0.055795711517333986, 0.055752704620361325, 0.055806976318359375, 0.055839744567871094, 0.05576294326782227, 0.05579980850219726, 0.05575065612792969, 0.055801856994628904, 0.05574860763549805, 0.05582438278198242, 0.055736320495605465, 0.05581107330322266, 0.05579673767089844, 0.055809024810791016, 0.055801856994628904, 0.05583769607543945, 0.05578956985473633, 0.05581619262695312, 0.05584998321533203, 0.05587046432495117, 0.05583359909057617, 0.055967742919921876, 0.055815166473388675, 0.05589503860473633, 0.055825408935546876, 0.05581619262695312, 0.05591551971435547, 0.055801856994628904, 0.05583769607543945, 0.05586329650878906, 0.05574860763549805, 0.05582131195068359, 0.05574041748046875, 0.05583769607543945, 0.055839744567871094, 0.05586841583251953, 0.05581721496582031, 0.05584588623046875, 0.05578137588500977, 0.055820289611816405, 0.055841793060302736, 0.05583462524414062, 0.0557916145324707, 0.055788543701171874, 0.05590118408203125, 0.055795711517333986, 0.05579673767089844, 0.0557844467163086, 0.05580492782592773, 0.055831550598144535, 0.05579264068603516, 0.0558551025390625, 0.05587251281738281, 0.055856128692626954, 0.05582032012939453, 0.05584585571289063, 0.12463410949707031, 0.05575065612792969, 0.05573734283447265, 0.05575475311279297, 0.05578342437744141, 0.055742462158203124, 0.055790592193603515, 0.05573734283447265, 0.05578649520874023, 0.05577830505371094, 0.0558551025390625, 0.055788543701171874, 0.0557946891784668, 0.055787521362304686, 0.055820289611816405, 0.055739391326904295, 0.055820289611816405, 0.05573427200317383, 0.05605068969726563, 0.055803905487060546, 0.05579776000976563, 0.055766014099121096, 0.055853057861328125, 0.055741439819335936, 0.055752704620361325, 0.05580083084106445, 0.05584076690673828, 0.0558131217956543, 0.05589606475830078, 0.05578137588500977, 0.05597081756591797, 0.055836673736572265, 0.05578956985473633, 0.05579980850219726, 0.05587353515625, 0.05579673767089844, 0.055774208068847655, 0.05579980850219726, 0.055815166473388675, 0.05575372695922851, 0.05580492782592773, 0.05586022567749024, 0.055844863891601565, 0.055806976318359375, 0.055806976318359375, 0.055812095642089846, 0.05589606475830078, 0.055772159576416014, 0.05577318572998047, 0.05588684844970703, 0.05584793472290039, 0.05581721496582031, 0.055810047149658204, 0.05586329650878906, 0.05579673767089844, 0.055772159576416014, 0.05586227035522461, 0.05584281539916992, 0.055785472869873044, 0.05581619262695312, 0.0557946891784668, 0.05583257675170898, 0.05588787078857422, 0.1247088623046875, 0.05576806259155274, 0.05583462524414062, 0.055766014099121096, 0.055772159576416014, 0.05578956985473633, 0.05581414413452149, 0.05579673767089844, 0.055742462158203124, 0.0557762565612793, 0.05581619262695312, 0.05573529434204102, 0.05583564758300781, 0.05576704025268555, 0.055815166473388675, 0.05571583938598633, 0.05580595016479492, 0.05576704025268555, 0.055790592193603515, 0.055782398223876956, 0.05589811325073242, 0.0557946891784668, 0.055809024810791016, 0.05574758529663086, 0.055798782348632815, 0.05577830505371094, 0.05583052825927735, 0.05578137588500977, 0.05585408020019531, 0.055790592193603515, 0.05580799865722656, 0.05587251281738281, 0.0557844467163086, 0.055795711517333986, 0.05579673767089844, 0.05584588623046875, 0.055782398223876956, 0.055806976318359375, 0.05589606475830078, 0.055777278900146485, 0.055844863891601565, 0.0557762565612793, 0.056202239990234375, 0.05586431884765625, 0.05582438278198242, 0.05581414413452149, 0.055820289611816405, 0.05581926345825195, 0.055769088745117185, 0.05581619262695312, 0.05580595016479492, 0.05586943817138672, 0.05609164810180664, 0.055853057861328125, 0.0558551025390625, 0.055856128692626954, 0.056139774322509765, 0.05579980850219726, 0.05579264068603516, 0.05577318572998047, 0.05584384155273438, 0.055788543701171874, 0.05585408020019531, 0.12452249908447266, 0.055742462158203124, 0.05573734283447265, 0.055742462158203124, 0.055787521362304686, 0.05576806259155274, 0.0557946891784668, 0.05572608184814453, 0.055795711517333986, 0.055779327392578126, 0.055890945434570315, 0.055760894775390625, 0.05575475311279297, 0.05573222351074219, 0.055815166473388675, 0.055723007202148435, 0.055858177185058595, 0.05573529434204102, 0.055725055694580077, 0.055790592193603515, 0.05591961669921875, 0.05581619262695312, 0.05587865447998047, 0.05575884628295898, 0.05578956985473633, 0.05578956985473633, 0.05581414413452149, 0.055790592193603515, 0.05590937423706055, 0.05582438278198242, 0.055806976318359375, 0.05578342437744141, 0.055907329559326174, 0.055831550598144535, 0.05587558364868164, 0.055785472869873044, 0.05579264068603516, 0.055812095642089846, 0.05582950210571289, 0.05575372695922851, 0.05583871841430664, 0.0558551025390625, 0.055823360443115234, 0.05580287933349609, 0.0557946891784668, 0.05583359909057617, 0.05591756820678711, 0.055779327392578126, 0.055826431274414064, 0.05596160125732422, 0.055982078552246094, 0.055801856994628904, 0.05582233428955078, 0.05587251281738281, 0.05585100936889648, 0.0557762565612793, 0.055826431274414064, 0.055815166473388675, 0.05581926345825195, 0.05585919952392578, 0.05578137588500977, 0.05581619262695312, 0.05582950210571289]",tokens/s,17.60390861520409,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12276,7 +12276,7 @@ If this is a private repository, make sure to pass a token having permission to 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-564M,facebook/xglm-564M,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1488.359424,1579.679744,0.0,933.23264,856.69632,s,10,0.4187196159362792,0.041871961593627924,0.0014808966097482894,0.04126091194152832,0.043131636810302736,0.0445920425415039,0.04576036712646484,"[0.04605244827270508, 0.04131564712524414, 0.0411776008605957, 0.04146476745605469, 0.04118272018432617, 0.04280710220336914, 0.0412061767578125, 0.0415596809387207, 0.04101833724975586, 0.04093513488769531]",tokens/s,6113.876452326468,kWh,4.934550676908758e-07,2.7027020769764927e-07,1.3917580578494868e-06,2.155483333238012e-06,tokens/kWh,118766865.90539834,MB,1488.359424,1579.679744,0.0,933.23264,899.62496,s,10,24.019910400390625,2.4019910400390625,0.019429569356012314,2.394274169921875,2.414185791015625,2.436057373046875,2.453554638671875,"[2.457928955078125, 2.409325439453125, 2.399566650390625, 2.391402587890625, 2.39863720703125, 2.391660400390625, 2.391578369140625, 2.391282958984375, 2.391639892578125, 2.396887939453125]",tokens/s,26.228241050797365,kWh,2.945148775767949e-05,1.6140594842595416e-05,7.348580184415957e-05,0.00011907788444443448,tokens/kWh,529065.4960316985,,s,629,24.35256624984743,0.03871632154188778,0.004963175047223502,0.037943294525146484,0.03889213485717773,0.03936890869140625,0.079117353515625,"[0.040226814270019534, 0.04027084732055664, 0.040318977355957034, 0.03974041748046875, 0.040825855255126955, 0.04090367889404297, 0.03897753524780274, 0.039828479766845705, 0.04030361557006836, 0.03983871841430664, 0.04004761505126953, 0.04024524688720703, 0.03980287933349609, 0.039346176147460936, 0.038795265197753906, 0.039277568817138675, 0.03892326354980469, 0.03963289642333984, 0.039182334899902346, 0.038902782440185545, 0.039244800567626956, 0.03907788848876953, 0.0392806396484375, 0.03914035034179687, 0.03907788848876953, 0.03934003067016602, 0.03828940963745117, 0.038212608337402344, 0.03859251022338867, 0.038125568389892575, 0.03932876968383789, 0.03894169616699219, 0.03829043197631836, 0.03913216018676758, 0.03880038452148438, 0.038793216705322264, 0.039011329650878904, 0.039003135681152344, 0.039147518157958985, 0.038547454833984376, 0.03806617736816406, 0.03796889495849609, 0.03785420989990235, 0.03794432067871094, 0.03905023956298828, 0.03965235137939453, 0.03999129486083984, 0.039384063720703126, 0.039397377014160156, 0.0378419189453125, 0.03808563232421875, 0.037797889709472655, 0.037797889709472655, 0.03783987045288086, 0.037891071319580076, 0.03788083267211914, 0.03912499237060547, 0.038760448455810545, 0.03786342239379883, 0.038950912475585936, 0.0378419189453125, 0.03783270263671875, 0.0792985610961914, 0.03790643310546875, 0.03781017684936523, 0.038370304107666016, 0.03831193542480469, 0.03775078582763672, 0.03785420989990235, 0.038435840606689455, 0.038365184783935545, 0.03781017684936523, 0.037905406951904294, 0.03861094284057617, 0.03815731048583984, 0.037917697906494144, 0.037776382446289065, 0.03875532913208008, 0.03827711868286133, 0.03894784164428711, 0.0380313606262207, 0.038877185821533204, 0.037923839569091795, 0.03808870315551758, 0.03812351989746094, 0.03917004776000976, 0.0390645751953125, 0.03829862213134766, 0.03843379211425781, 0.038542335510253906, 0.03800166320800781, 0.0378869743347168, 0.038740993499755856, 0.03850239944458008, 0.03803955078125, 0.038438911437988284, 0.03892838287353516, 0.03817267227172851, 0.038012928009033206, 0.03801190567016602, 0.039065601348876954, 0.03803033447265625, 0.03791462326049805, 0.03789619064331055, 0.03887923049926758, 0.038371326446533204, 0.03790848159790039, 0.037904384613037106, 0.03786956787109375, 0.037988353729248046, 0.037943294525146484, 0.03816755294799805, 0.03861503982543945, 0.03803340911865234, 0.037966846466064456, 0.03835084915161133, 0.03848396682739258, 0.03804159927368164, 0.03789926528930664, 0.038539264678955076, 0.038340606689453126, 0.03802521514892578, 0.03793203353881836, 0.038679550170898434, 0.03826892852783203, 0.07922994995117187, 0.03890790557861328, 0.03783679962158203, 0.03792281723022461, 0.03792588806152344, 0.03827507019042969, 0.03808153533935547, 0.03808768081665039, 0.03786137771606445, 0.03799552154541016, 0.0380313606262207, 0.03805388641357422, 0.03804569625854492, 0.03840409469604492, 0.03806412887573242, 0.037905406951904294, 0.03795455932617187, 0.03794636917114258, 0.03813478469848633, 0.03800985717773438, 0.03793203353881836, 0.03811123275756836, 0.038125568389892575, 0.03792793655395508, 0.03794739151000977, 0.03824435043334961, 0.0380847053527832, 0.0379257926940918, 0.03787161636352539, 0.037819393157958986, 0.03803647994995117, 0.037956607818603515, 0.039212032318115236, 0.03805184173583984, 0.0379504623413086, 0.0379156494140625, 0.04073574447631836, 0.038391807556152346, 0.03786649703979492, 0.03888947296142578, 0.0400711669921875, 0.03781222534179687, 0.03783679962158203, 0.037981182098388674, 0.03775078582763672, 0.03779891204833984, 0.03775590515136719, 0.0378419189453125, 0.03782144165039063, 0.037781505584716796, 0.03789311981201172, 0.03784396743774414, 0.03780710220336914, 0.038095870971679685, 0.037797889709472655, 0.03782451248168945, 0.03784601593017578, 0.03794124984741211, 0.037806079864501956, 0.037781505584716796, 0.037884929656982425, 0.03780710220336914, 0.03832627105712891, 0.07955046081542969, 0.03795251083374023, 0.03794947052001953, 0.03773539352416992, 0.03782860946655273, 0.03780710220336914, 0.037874687194824216, 0.037789695739746096, 0.037867584228515626, 0.03773331069946289, 0.038191104888916014, 0.03783679962158203, 0.037872638702392575, 0.03788185501098633, 0.037850112915039064, 0.0378869743347168, 0.03800371170043945, 0.037870655059814455, 0.038172607421875, 0.037921791076660154, 0.03788390350341797, 0.03808256149291992, 0.03796377563476563, 0.037819393157958986, 0.03782553482055664, 0.03784806442260742, 0.038128639221191404, 0.038128639221191404, 0.03777228927612305, 0.03794944000244141, 0.03797196960449219, 0.03817062377929688, 0.03783168029785156, 0.037855232238769534, 0.037928958892822266, 0.03805286407470703, 0.0382371826171875, 0.03802828979492188, 0.037988353729248046, 0.037961727142333986, 0.038179840087890625, 0.03786240005493164, 0.037978111267089845, 0.0379156494140625, 0.038059009552001956, 0.038112255096435545, 0.037910526275634765, 0.03792588806152344, 0.03811942291259766, 0.03815935897827148, 0.037978111267089845, 0.037874687194824216, 0.03809996795654297, 0.03806924819946289, 0.03789311981201172, 0.037894176483154296, 0.037800926208496094, 0.038100990295410156, 0.037956607818603515, 0.037953536987304685, 0.037889022827148434, 0.03792899322509766, 0.03823203277587891, 0.07917670440673828, 0.03791462326049805, 0.03787366485595703, 0.03793305587768555, 0.03789926528930664, 0.03786240005493164, 0.038117374420166016, 0.03861913681030273, 0.03791974258422852, 0.038007808685302735, 0.03791257476806641, 0.03793817520141601, 0.03802521514892578, 0.0379422721862793, 0.03785830307006836, 0.038046718597412106, 0.037959678649902344, 0.037958656311035156, 0.03790028762817383, 0.037817344665527344, 0.03794124984741211, 0.03805184173583984, 0.03784294509887695, 0.03805388641357422, 0.03809894561767578, 0.03805286407470703, 0.03784499359130859, 0.03791974258422852, 0.03803955078125, 0.03802828979492188, 0.038161407470703124, 0.03802214431762695, 0.037894142150878905, 0.03818598556518555, 0.037917697906494144, 0.03795455932617187, 0.037975040435791016, 0.03828224182128906, 0.038002689361572264, 0.037910526275634765, 0.03793100738525391, 0.03803852844238281, 0.03797708892822266, 0.03793612670898437, 0.0378869743347168, 0.0380313606262207, 0.03825971221923828, 0.037953536987304685, 0.03790643310546875, 0.038029312133789066, 0.03808153533935547, 0.03808870315551758, 0.0379791374206543, 0.03979673767089844, 0.03898777770996094, 0.03879116821289062, 0.037969921112060545, 0.03789926528930664, 0.03830579376220703, 0.038193153381347655, 0.038182910919189454, 0.03849523162841797, 0.03826278305053711, 0.08257740783691406, 0.03822489547729492, 0.03793920135498047, 0.03778559875488281, 0.03784499359130859, 0.037872638702392575, 0.038141952514648435, 0.037921791076660154, 0.03786342239379883, 0.03808256149291992, 0.03808051300048828, 0.0378419189453125, 0.03813273620605469, 0.038711296081542966, 0.038114303588867186, 0.03786240005493164, 0.03781119918823242, 0.03850137710571289, 0.03799347305297852, 0.0378152961730957, 0.0378869743347168, 0.037989376068115234, 0.03785728073120117, 0.037850112915039064, 0.0379607048034668, 0.03787782287597656, 0.03834259033203125, 0.03791462326049805, 0.03792588806152344, 0.03778047943115234, 0.038484992980957033, 0.03798015975952149, 0.03783782577514649, 0.03778559875488281, 0.03781017684936523, 0.0379791374206543, 0.03785625457763672, 0.03782553482055664, 0.0378603515625, 0.037923839569091795, 0.0378152961730957, 0.03780505752563477, 0.03792998504638672, 0.03794124984741211, 0.03778355026245117, 0.03781119918823242, 0.03775385665893555, 0.03779891204833984, 0.03803238296508789, 0.03786956787109375, 0.03816960144042969, 0.037814273834228515, 0.03785625457763672, 0.038043647766113284, 0.03777849578857422, 0.037824447631835935, 0.037713920593261716, 0.03772108840942383, 0.03790233612060547, 0.0377784309387207, 0.037776382446289065, 0.0378152961730957, 0.037921791076660154, 0.07919411468505859, 0.03782758331298828, 0.03773747253417969, 0.03775385665893555, 0.03801804733276367, 0.037800960540771485, 0.03785728073120117, 0.03783679962158203, 0.037719039916992186, 0.038024192810058595, 0.03775590515136719, 0.037768192291259765, 0.037823486328125, 0.037910526275634765, 0.03777433776855469, 0.03780505752563477, 0.03785113525390625, 0.0377968635559082, 0.037874687194824216, 0.03781222534179687, 0.03783475112915039, 0.03781119918823242, 0.037754878997802735, 0.03797094345092773, 0.03784499359130859, 0.03780710220336914, 0.03785728073120117, 0.03789209747314453, 0.0379791374206543, 0.03779891204833984, 0.03774771118164062, 0.03775078582763672, 0.03789926528930664, 0.03783987045288086, 0.037738529205322266, 0.037819358825683595, 0.03775795364379883, 0.037889022827148434, 0.037773311614990236, 0.03777740859985351, 0.03778662490844727, 0.03790950393676758, 0.03822079849243164, 0.04011008071899414, 0.03838054275512695, 0.0379576301574707, 0.03801599884033203, 0.03980595016479492, 0.03857920074462891, 0.03786956787109375, 0.03796384048461914, 0.03790534210205078, 0.03787059020996094, 0.037891071319580076, 0.037838848114013675, 0.03804774475097656, 0.03783270263671875, 0.03785113525390625, 0.038819839477539066, 0.038046718597412106, 0.03789823913574219, 0.03790233612060547, 0.03783679962158203, 0.08027750396728515, 0.038125568389892575, 0.03789619064331055, 0.03792588806152344, 0.03793100738525391, 0.03866726303100586, 0.03827916717529297, 0.03803443145751953, 0.03789619064331055, 0.03781222534179687, 0.03788185501098633, 0.03791257476806641, 0.03783065414428711, 0.03784505462646484, 0.03773846435546875, 0.03788899230957031, 0.03809280014038086, 0.03783475112915039, 0.03798323059082031, 0.03775385665893555, 0.037966846466064456, 0.038214656829833986, 0.03794124984741211, 0.03781119918823242, 0.037868545532226565, 0.03785113525390625, 0.03793100738525391, 0.0380211181640625, 0.03793817520141601, 0.037855232238769534, 0.03802828979492188, 0.0378081283569336, 0.03788185501098633, 0.03851264190673828, 0.03827097702026367, 0.03790643310546875, 0.03789619064331055, 0.03783065414428711, 0.037907455444335936, 0.03800985717773438, 0.03793920135498047, 0.03778867340087891, 0.03791872024536133, 0.037806079864501956, 0.03793920135498047, 0.03792793655395508, 0.037855232238769534, 0.03791974258422852, 0.03787161636352539, 0.03796582412719727, 0.038024192810058595, 0.03782860946655273, 0.03795251083374023, 0.03783782577514649, 0.037956607818603515, 0.03788083267211914, 0.03803964614868164, 0.03777833557128906, 0.037766143798828124, 0.03796275329589844, 0.03784703826904297, 0.03799859237670898, 0.03801804733276367, 0.07892582702636719, 0.03798527908325195, 0.037719039916992186, 0.03776921463012695, 0.03804467010498047, 0.037907455444335936, 0.03788288116455078, 0.03789516830444336, 0.03778867340087891, 0.037966846466064456, 0.03824844741821289, 0.03791155242919922, 0.03777228927612305, 0.03783065414428711, 0.03782758331298828, 0.03794739151000977, 0.03860070419311523, 0.0380948486328125, 0.037868545532226565, 0.03780505752563477, 0.0378152961730957, 0.037822463989257815, 0.03793407821655274, 0.03778771209716797, 0.03782342529296875, 0.037855232238769534, 0.03776921463012695, 0.038046718597412106, 0.037803009033203126, 0.03780198287963867, 0.037838848114013675, 0.03789516830444336, 0.03798015975952149, 0.0378603515625, 0.03887923049926758, 0.03785625457763672, 0.03788390350341797, 0.03847884750366211, 0.03800064086914062, 0.03786342239379883, 0.03780710220336914, 0.037826591491699216, 0.03794736099243164, 0.0378787841796875, 0.03783065414428711, 0.03801702499389648, 0.03784601593017578, 0.03804569625854492, 0.03787571334838867, 0.03788288116455078, 0.03784806442260742, 0.03797094345092773, 0.038019073486328124, 0.03792281723022461, 0.037956607818603515, 0.03784601593017578, 0.03790848159790039, 0.03802316665649414, 0.037868545532226565, 0.037868545532226565, 0.03931238555908203, 0.03824435043334961, 0.038171646118164065, 0.07896473693847657, 0.03796480178833008, 0.038031391143798825, 0.038429664611816405, 0.03807231903076172, 0.03814297485351562, 0.04003839874267578, 0.038354942321777344, 0.038171646118164065, 0.03779481506347656, 0.03799347305297852, 0.03781631851196289, 0.037822463989257815, 0.03803852844238281, 0.037795841217041014, 0.03802009582519531, 0.03805184173583984, 0.037812255859375, 0.038042591094970706, 0.0378787841796875, 0.03813785552978516, 0.03794636917114258, 0.038004737854003906, 0.03808153533935547, 0.03786240005493164, 0.03785318374633789, 0.03793817520141601, 0.0379156494140625, 0.03786137771606445, 0.03794124984741211, 0.037972991943359374, 0.03801599884033203, 0.03807743835449219, 0.038757377624511716, 0.038007808685302735, 0.03791155242919922, 0.03793407821655274, 0.03837849426269531, 0.038100990295410156, 0.037975040435791016, 0.037891071319580076, 0.03786751937866211, 0.03800678253173828, 0.03780505752563477, 0.03791360092163086, 0.037907455444335936, 0.03789209747314453, 0.03788288116455078, 0.037894142150878905, 0.03790335845947266, 0.0380497932434082, 0.0378787841796875, 0.0379504623413086, 0.03782963180541992, 0.037920768737792966, 0.03815731048583984, 0.03894476699829102, 0.03841228866577148, 0.0381102066040039, 0.03795455932617187, 0.03790848159790039, 0.03784396743774414, 0.03802521514892578]",tokens/s,25.828900065262776,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1279.995904,1017.643008,0.0,371.195904,277.410816,s,10,0.4175716514587402,0.04175716514587402,0.001329291485763711,0.04133342361450196,0.042088745498657223,0.04389397258758545,0.04533815425872803,"[0.045699199676513674, 0.04128799819946289, 0.04128377532958984, 0.04097267150878906, 0.04157187271118164, 0.04107036972045899, 0.04121916961669922, 0.04137884902954102, 0.04168758392333984, 0.04140016174316406]",tokens/s,6130.684377296505,kWh,4.643787057311447e-07,2.5445859858069414e-07,1.253729761153002e-06,1.972567065464841e-06,tokens/kWh,129780124.83427167,MB,1279.995904,1017.643008,0.0,371.195904,323.048448,s,10,24.960871826171868,2.496087182617187,0.007120166670465486,2.4940638427734374,2.4991390625000003,2.5077969482421874,2.5147232568359374,"[2.516454833984375, 2.49401806640625, 2.495659423828125, 2.49662109375, 2.49185595703125, 2.493477294921875, 2.4905234375, 2.49093701171875, 2.494109619140625, 2.497215087890625]",tokens/s,25.239503026470214,kWh,2.763858284111377e-05,1.5146819142039876e-05,6.740021405084962e-05,0.00011018561603400326,tokens/kWh,571762.4701626954,,s,629,25.29028705596926,0.04020713363429131,0.0049559405951164134,0.03952540969848633,0.039936614990234376,0.040598938751220705,0.08109604888916017,"[0.04100912094116211, 0.04127743911743164, 0.04177305603027344, 0.04085964965820312, 0.041427967071533206, 0.04092620849609375, 0.04130918502807617, 0.040700927734375, 0.042106880187988284, 0.04071731185913086, 0.040460289001464846, 0.04082995223999023, 0.03992473602294922, 0.039828479766845705, 0.03968511962890625, 0.039523326873779296, 0.03983257675170898, 0.039553024291992187, 0.039913471221923826, 0.039607295989990236, 0.03934515380859375, 0.03961139297485351, 0.03963187026977539, 0.039706623077392575, 0.03935334396362305, 0.03937484741210937, 0.03951923370361328, 0.039556095123291016, 0.03966873550415039, 0.03936972808837891, 0.03945574569702148, 0.039840831756591796, 0.0411473274230957, 0.039992321014404295, 0.03950080108642578, 0.0393994255065918, 0.03961958312988281, 0.039554046630859374, 0.03956838226318359, 0.039792640686035156, 0.03939740753173828, 0.03990217590332031, 0.03950284957885742, 0.03952537536621094, 0.039384063720703126, 0.039395328521728515, 0.0393809928894043, 0.03959091186523438, 0.03945676803588867, 0.039395328521728515, 0.03952947235107422, 0.03961139297485351, 0.03948441696166992, 0.0394967041015625, 0.03940147018432617, 0.039337982177734376, 0.04001484680175781, 0.03948953628540039, 0.03933695983886719, 0.039452671051025394, 0.03969023895263672, 0.03961958312988281, 0.08116633605957031, 0.03934003067016602, 0.03947724914550781, 0.03954687881469727, 0.03953664016723633, 0.039618560791015625, 0.039354366302490236, 0.039488510131835936, 0.039809024810791016, 0.03934310531616211, 0.039357440948486325, 0.03944243240356445, 0.03957657623291016, 0.03940761566162109, 0.039419902801513675, 0.03942399978637695, 0.0395489273071289, 0.039482368469238284, 0.03993907165527344, 0.03930419158935547, 0.0394967041015625, 0.03933491134643555, 0.039373825073242184, 0.03941785430908203, 0.039444480895996094, 0.039467041015625, 0.039427040100097656, 0.03946086502075195, 0.039478271484375, 0.039577598571777346, 0.03945471954345703, 0.039298046112060545, 0.039298046112060545, 0.03948339080810547, 0.039370750427246096, 0.03942502212524414, 0.03943833541870117, 0.039602176666259765, 0.03958988952636719, 0.03945881652832031, 0.039532543182373044, 0.04032716751098633, 0.04127542495727539, 0.040283103942871094, 0.03961139297485351, 0.03957657623291016, 0.03948953628540039, 0.03951615905761719, 0.039367679595947266, 0.03977833557128906, 0.039686111450195315, 0.04003123092651367, 0.04087910461425781, 0.040207359313964845, 0.03972403335571289, 0.03946393585205078, 0.039387134552001955, 0.039554046630859374, 0.03936972808837891, 0.03947724914550781, 0.0394005126953125, 0.03951507186889648, 0.03940662384033203, 0.08121955108642578, 0.039504894256591795, 0.039790592193603515, 0.03975270462036133, 0.03982131195068359, 0.039556129455566406, 0.039650272369384766, 0.03966259384155273, 0.03948339080810547, 0.03957145690917969, 0.03962265777587891, 0.039591934204101564, 0.03957452774047852, 0.03968716812133789, 0.03951103973388672, 0.039772159576416014, 0.03950592041015625, 0.039828479766845705, 0.03968819046020508, 0.039564289093017575, 0.039532543182373044, 0.03951615905761719, 0.03973324966430664, 0.03973222351074219, 0.03958476638793945, 0.03959807968139648, 0.03997183990478516, 0.03969740676879883, 0.03954380798339844, 0.0401080322265625, 0.03958784103393555, 0.03946188735961914, 0.03951718521118164, 0.03965542221069336, 0.03964416122436523, 0.039585792541503906, 0.039599105834960936, 0.039634944915771485, 0.0393891830444336, 0.03955712127685547, 0.03965542221069336, 0.03983161544799805, 0.03946899032592773, 0.039577598571777346, 0.039572479248046875, 0.03957555389404297, 0.039698463439941406, 0.03968406295776367, 0.039599105834960936, 0.03945779037475586, 0.03973529434204102, 0.038989822387695314, 0.03940147018432617, 0.03963904190063477, 0.03952435302734375, 0.03980287933349609, 0.039634944915771485, 0.03947724914550781, 0.03947417449951172, 0.0395786247253418, 0.039591934204101564, 0.03949465560913086, 0.03950899124145508, 0.08075161743164062, 0.03947520065307617, 0.03970457458496094, 0.03956329727172852, 0.03949052810668945, 0.03964108657836914, 0.03945369720458984, 0.03959091186523438, 0.03958476638793945, 0.03949465560913086, 0.03940147018432617, 0.03942502212524414, 0.039354366302490236, 0.039539710998535156, 0.03950387191772461, 0.03942297744750976, 0.039923713684082034, 0.03959603118896484, 0.03958169555664062, 0.03936870574951172, 0.039532543182373044, 0.039547904968261716, 0.039572479248046875, 0.03954687881469727, 0.03948748779296875, 0.03956531143188476, 0.04043366241455078, 0.040976383209228515, 0.039962623596191404, 0.039689216613769535, 0.039373825073242184, 0.03946188735961914, 0.03932364654541016, 0.03958476638793945, 0.039539710998535156, 0.03945471954345703, 0.039446529388427735, 0.039629825592041014, 0.039375873565673826, 0.03932467269897461, 0.039597057342529295, 0.040233985900878906, 0.04049612808227539, 0.03984896087646484, 0.039357440948486325, 0.03942195129394531, 0.03954483032226563, 0.039456832885742185, 0.03950073623657226, 0.03953049468994141, 0.03962060928344727, 0.039444480895996094, 0.039754753112792966, 0.03976806259155274, 0.04006911849975586, 0.039629825592041014, 0.039419902801513675, 0.03955199813842773, 0.03977830505371094, 0.039602176666259765, 0.040025089263916014, 0.03968819046020508, 0.03958169555664062, 0.08158719635009766, 0.03952844619750977, 0.03982643127441406, 0.03968000030517578, 0.03958476638793945, 0.03935334396362305, 0.039572479248046875, 0.040210430145263674, 0.03953561782836914, 0.039654399871826174, 0.03932057571411133, 0.03947417449951172, 0.03926323318481445, 0.039636993408203126, 0.039689216613769535, 0.03954278564453125, 0.039504894256591795, 0.0395601921081543, 0.03949055862426758, 0.0393963508605957, 0.03964518356323242, 0.039634944915771485, 0.039430145263671876, 0.03942297744750976, 0.03927347183227539, 0.039599105834960936, 0.039367679595947266, 0.039485439300537106, 0.039806976318359374, 0.040528896331787106, 0.039532543182373044, 0.039362560272216796, 0.03942502212524414, 0.039651329040527344, 0.03954483032226563, 0.03994521713256836, 0.040018943786621096, 0.03948748779296875, 0.039367679595947266, 0.039653377532958986, 0.03947110366821289, 0.03956633758544922, 0.03923660659790039, 0.039523326873779296, 0.03932876968383789, 0.039406593322753904, 0.039462913513183595, 0.03944140625, 0.03933184051513672, 0.039863296508789066, 0.03953664016723633, 0.039504894256591795, 0.039191551208496093, 0.040030208587646485, 0.03945881652832031, 0.03943324661254883, 0.03956835174560547, 0.0393891830444336, 0.03934822463989258, 0.039390209197998044, 0.03933292770385742, 0.03960108947753906, 0.03945881652832031, 0.08104959869384766, 0.03950592041015625, 0.03941891098022461, 0.039413726806640625, 0.03958988952636719, 0.03938102340698242, 0.03919769668579102, 0.03935843276977539, 0.03926323318481445, 0.03933184051513672, 0.03955507278442383, 0.039395328521728515, 0.0393963508605957, 0.039449600219726565, 0.03967795181274414, 0.039828479766845705, 0.039708671569824217, 0.0393994255065918, 0.03942195129394531, 0.03932160186767578, 0.039772159576416014, 0.0416184310913086, 0.03949363327026367, 0.03941273498535156, 0.0393114242553711, 0.039466945648193356, 0.03957555389404297, 0.039329792022705076, 0.039397377014160156, 0.03927859115600586, 0.039390239715576175, 0.039481311798095706, 0.03924889755249023, 0.039397377014160156, 0.03929600143432617, 0.039414783477783204, 0.039723007202148435, 0.039465984344482424, 0.04114329528808594, 0.04011625671386719, 0.039448543548583986, 0.03947315216064453, 0.03952537536621094, 0.039370750427246096, 0.03939123153686523, 0.03942604827880859, 0.039684097290039064, 0.03952742385864258, 0.040645633697509766, 0.039919647216796875, 0.03952329635620117, 0.03967078399658203, 0.03952540969848633, 0.03946083068847656, 0.039446529388427735, 0.03928268814086914, 0.03962879943847656, 0.039470142364501956, 0.03950073623657226, 0.03953868865966797, 0.03960422515869141, 0.03969945526123047, 0.03978854370117187, 0.08111411285400391, 0.040289279937744144, 0.0393072624206543, 0.03955199813842773, 0.039501823425292966, 0.0393891830444336, 0.0393359375, 0.03960934448242188, 0.03942604827880859, 0.039554046630859374, 0.03937279891967774, 0.03945369720458984, 0.03937177658081055, 0.039611423492431644, 0.03978031921386719, 0.0395335693359375, 0.039449600219726565, 0.03945574569702148, 0.03949158477783203, 0.039651329040527344, 0.039408641815185545, 0.039446529388427735, 0.039370750427246096, 0.039618560791015625, 0.03969126510620117, 0.03963391876220703, 0.039793663024902344, 0.03998822402954102, 0.03951513671875, 0.03942502212524414, 0.03936460876464844, 0.03952844619750977, 0.03944550323486328, 0.039468032836914066, 0.03973836898803711, 0.039488510131835936, 0.039398399353027344, 0.03951103973388672, 0.039433216094970705, 0.03953049468994141, 0.039384063720703126, 0.039610366821289066, 0.03942195129394531, 0.03937484741210937, 0.039444480895996094, 0.039605247497558595, 0.0394700813293457, 0.0394598388671875, 0.03948134231567383, 0.03951103973388672, 0.039702529907226565, 0.03990528106689453, 0.03952844619750977, 0.03959807968139648, 0.039534591674804685, 0.03949772644042969, 0.03940966415405273, 0.039390209197998044, 0.03936153411865234, 0.03949260711669922, 0.03947724914550781, 0.039478271484375, 0.03961964797973633, 0.08132498931884766, 0.039577598571777346, 0.039392257690429686, 0.039580673217773435, 0.039521278381347655, 0.03939328002929687, 0.039190528869628906, 0.03920588684082031, 0.039411712646484375, 0.0393891830444336, 0.03947110366821289, 0.03920588684082031, 0.03942911911010742, 0.03978137588500977, 0.03956531143188476, 0.03957452774047852, 0.039624702453613284, 0.03947110366821289, 0.03937996673583984, 0.03949676895141602, 0.03960723114013672, 0.03957350540161133, 0.03959299087524414, 0.04049200057983399, 0.040079360961914064, 0.03965235137939453, 0.03947315216064453, 0.039523326873779296, 0.03956633758544922, 0.03941785430908203, 0.03943219375610352, 0.03935539245605469, 0.03939328002929687, 0.03963596725463867, 0.03949772644042969, 0.039539710998535156, 0.03936460876464844, 0.039382015228271484, 0.03985203170776367, 0.04112588882446289, 0.03990835189819336, 0.039303199768066406, 0.03935740661621094, 0.03930828857421875, 0.039629825592041014, 0.039370750427246096, 0.03955507278442383, 0.039414783477783204, 0.039392257690429686, 0.03963391876220703, 0.039408641815185545, 0.03947727966308594, 0.03938198471069336, 0.03943116760253906, 0.039287841796875, 0.03957244873046875, 0.03941686248779297, 0.03938608169555664, 0.03933695983886719, 0.03938816070556641, 0.03966566467285156, 0.03951001739501953, 0.03967795181274414, 0.0816015396118164, 0.03950899124145508, 0.03951308822631836, 0.039567359924316405, 0.03965951919555664, 0.03955820846557617, 0.03948230361938477, 0.03946086502075195, 0.03948339080810547, 0.03967385482788086, 0.039572479248046875, 0.03956326293945313, 0.03950387191772461, 0.03951103973388672, 0.03965446472167969, 0.03958572769165039, 0.03975372695922851, 0.039676959991455076, 0.03963081741333008, 0.03968307113647461, 0.039874561309814455, 0.03963596725463867, 0.039359489440917966, 0.03943936157226562, 0.03984998321533203, 0.03992166519165039, 0.039825408935546876, 0.03975987243652344, 0.03933695983886719, 0.03944755172729492, 0.03970355224609375, 0.03947520065307617, 0.03924070358276367, 0.039408641815185545, 0.03939328002929687, 0.0395786247253418, 0.039427135467529295, 0.039470016479492186, 0.03935641479492188, 0.039585792541503906, 0.03967488098144531, 0.039550975799560545, 0.039375873565673826, 0.039498752593994144, 0.03950387191772461, 0.03968307113647461, 0.0395601921081543, 0.03951513671875, 0.039567359924316405, 0.03954278564453125, 0.040025089263916014, 0.03950694274902344, 0.03945369720458984, 0.03981414413452149, 0.03958169555664062, 0.03971583938598633, 0.040190975189208986, 0.039646209716796874, 0.03933184051513672, 0.03954380798339844, 0.03963289642333984, 0.03962777709960937, 0.039547904968261716, 0.081259521484375, 0.03940454483032227, 0.03937996673583984, 0.03957452774047852, 0.039414783477783204, 0.039419902801513675, 0.03935644912719727, 0.03931337738037109, 0.03946700668334961, 0.0394117431640625, 0.03942396926879883, 0.039762943267822266, 0.039406593322753904, 0.039550975799560545, 0.039481407165527345, 0.03942700958251953, 0.0395489273071289, 0.03993600082397461, 0.04035379028320313, 0.04098361587524414, 0.03966969680786133, 0.03950080108642578, 0.039616512298583983, 0.03977830505371094, 0.039462913513183595, 0.03949260711669922, 0.039669761657714846, 0.03949977493286133, 0.03933388900756836, 0.039332862854003905, 0.039398399353027344, 0.03946700668334961, 0.039672832489013675, 0.039769088745117184, 0.040182785034179686, 0.04107059097290039, 0.04016742324829101, 0.03948441696166992, 0.03967078399658203, 0.03942604827880859, 0.03934207916259766, 0.03936460876464844, 0.03937279891967774, 0.03931545639038086, 0.04087807846069336, 0.04114636611938476, 0.039907329559326174, 0.03947110366821289, 0.03962879943847656, 0.03942707061767578, 0.03959609603881836, 0.03944236755371094, 0.039609375, 0.03951305770874024, 0.03944857788085938, 0.03932160186767578, 0.03956531143188476, 0.03942399978637695, 0.03932160186767578, 0.03932057571411133, 0.039365631103515625, 0.03976499176025391, 0.03956326293945313]",tokens/s,24.871208405344607,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-6.7b,facebook/opt-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2019.69664,5274.861568,0.0,4628.414464,4279.64672,s,10,3.1774854736328124,0.31774854736328123,0.0007817722146641278,0.3174602966308594,0.3187982666015625,0.3189695983886719,0.3191066638183594,"[0.31848922729492185, 0.3174812927246094, 0.31692730712890627, 0.31720437622070313, 0.3174393005371094, 0.31661932373046875, 0.3174072570800781, 0.31876019287109375, 0.31914093017578127, 0.3180162658691406]",tokens/s,805.6685140634671,kWh,3.737864001757569e-06,2.0481879523572387e-06,1.80164467048112e-05,2.3802498658926007e-05,tokens/kWh,10755173.38193397,MB,2019.69664,5274.861568,0.0,4628.414464,4463.185408,s,10,176.9617734375,17.69617734375,0.0029728792469339506,17.696136718749997,17.6995603515625,17.699568261718753,17.699574589843753,"[17.69631640625, 17.695939453125, 17.698544921875, 17.693357421875, 17.690775390625, 17.69253125, 17.699216796875, 17.69955859375, 17.699576171875, 17.69595703125]",tokens/s,3.5600909041664055,kWh,0.00020889662316689894,0.00011449271197744792,0.0009934635569229903,0.0013168528920673372,tokens/kWh,47841.33473033258,,s,629,179.53831649780273,0.2854345254337086,0.0377383923728263,0.28085964965820315,0.28117113647460934,0.2812770324707031,0.5980694702148438,"[0.2809251708984375, 0.280669189453125, 0.2807091064453125, 0.2809108581542969, 0.28111666870117186, 0.28064971923828125, 0.2810654602050781, 0.28062106323242186, 0.2806558837890625, 0.2807910461425781, 0.28068453979492186, 0.2808770446777344, 0.2806640625, 0.2807613525390625, 0.28071832275390624, 0.28078387451171877, 0.2807152709960937, 0.28067431640625, 0.28070501708984374, 0.2807490539550781, 0.28121600341796876, 0.2806056823730469, 0.28095077514648437, 0.28073471069335937, 0.28082278442382813, 0.2807633972167969, 0.2808299560546875, 0.2809405517578125, 0.28083712768554686, 0.28074700927734375, 0.28070401000976564, 0.28072857666015627, 0.2808084411621094, 0.2823638916015625, 0.2806558837890625, 0.2807715759277344, 0.2807767028808594, 0.2810624084472656, 0.28082278442382813, 0.2810224609375, 0.28100811767578127, 0.2810163269042969, 0.2808545227050781, 0.28095077514648437, 0.2809190368652344, 0.28101324462890626, 0.2807828369140625, 0.2812323913574219, 0.2809671630859375, 0.28106137084960936, 0.2808330383300781, 0.2810439758300781, 0.28064971923828125, 0.28111770629882815, 0.2812231750488281, 0.2810408935546875, 0.28118118286132815, 0.28099072265625, 0.2808832092285156, 0.2812590026855469, 0.2809385070800781, 0.2808832092285156, 0.5983795166015625, 0.28064871215820314, 0.2806937561035156, 0.2806876220703125, 0.2805964660644531, 0.28077362060546873, 0.28084735107421877, 0.2808135681152344, 0.28077978515625, 0.2806302795410156, 0.2807633972167969, 0.2808637390136719, 0.281501708984375, 0.2808135681152344, 0.28127743530273436, 0.2808412170410156, 0.2809886779785156, 0.2807439270019531, 0.2807940979003906, 0.28082382202148437, 0.28080126953125, 0.28092620849609373, 0.2807767028808594, 0.28085964965820315, 0.2809610290527344, 0.2808668212890625, 0.28077978515625, 0.2808545227050781, 0.2811832275390625, 0.28089959716796875, 0.28109515380859373, 0.28077566528320314, 0.2809292907714844, 0.28073983764648436, 0.28083712768554686, 0.2807214050292969, 0.28081561279296874, 0.28083096313476563, 0.28084735107421877, 0.2808258666992188, 0.28108901977539064, 0.28077566528320314, 0.28080230712890625, 0.2808279113769531, 0.28091494750976564, 0.280700927734375, 0.28091494750976564, 0.28119143676757813, 0.28127642822265625, 0.2812569580078125, 0.2810091552734375, 0.2808279113769531, 0.28108697509765623, 0.2808934326171875, 0.28086273193359373, 0.28114431762695313, 0.28093438720703123, 0.28095489501953125, 0.2810040283203125, 0.2808350830078125, 0.28096307373046875, 0.2807848815917969, 0.2808197021484375, 0.5980743408203125, 0.28070401000976564, 0.2806609802246094, 0.28080743408203124, 0.280690673828125, 0.2807357482910156, 0.28078900146484376, 0.28069888305664065, 0.28073883056640625, 0.28077566528320314, 0.280637451171875, 0.2809395141601562, 0.28084225463867185, 0.280669189453125, 0.28070196533203123, 0.28089242553710936, 0.2807613525390625, 0.28102554321289064, 0.280958984375, 0.28090777587890625, 0.28084530639648436, 0.2811146240234375, 0.2807060546875, 0.2808821716308594, 0.28075314331054685, 0.2806824951171875, 0.2807562255859375, 0.2807132263183594, 0.2805975036621094, 0.280627197265625, 0.2807439270019531, 0.280848388671875, 0.2807879638671875, 0.2806599731445312, 0.2806077575683594, 0.28076031494140624, 0.28118426513671874, 0.28088525390625, 0.2811043701171875, 0.281143310546875, 0.2812733459472656, 0.2809036865234375, 0.28103372192382814, 0.28068453979492186, 0.28123956298828123, 0.28111566162109375, 0.28099072265625, 0.28118118286132815, 0.2812323913574219, 0.28117300415039065, 0.2811197509765625, 0.2810224609375, 0.28101937866210935, 0.28089447021484376, 0.28143719482421875, 0.2813931579589844, 0.28145767211914063, 0.28113409423828123, 0.2810245056152344, 0.2813102111816406, 0.2813634643554688, 0.2808913879394531, 0.2811832275390625, 0.5980569458007813, 0.280806396484375, 0.28072549438476563, 0.28071218872070314, 0.2806876220703125, 0.2807807922363281, 0.28076235961914064, 0.2810705871582031, 0.2806241149902344, 0.2806855773925781, 0.28086578369140625, 0.2805698547363281, 0.2807582702636719, 0.28055859375, 0.28054937744140623, 0.28078387451171877, 0.2807562255859375, 0.2807173156738281, 0.28059033203125, 0.2807162780761719, 0.2807500915527344, 0.28070196533203123, 0.2806824951171875, 0.2808340454101563, 0.28079718017578126, 0.28069171142578125, 0.2811125793457031, 0.2808883056640625, 0.28106341552734376, 0.28073471069335937, 0.28064254760742186, 0.2807705688476563, 0.2809989013671875, 0.28072857666015627, 0.28084225463867185, 0.28058624267578125, 0.28073370361328126, 0.28079513549804686, 0.28098458862304687, 0.28078900146484376, 0.2809722900390625, 0.2808965148925781, 0.2811125793457031, 0.28095693969726565, 0.281196533203125, 0.28067636108398436, 0.28092312622070315, 0.2809385070800781, 0.28081561279296874, 0.28086477661132814, 0.2808913879394531, 0.28067431640625, 0.28132455444335935, 0.280922119140625, 0.28108697509765623, 0.2814136352539062, 0.2810091552734375, 0.28069989013671875, 0.280838134765625, 0.28130712890625, 0.2808555603027344, 0.2809016418457031, 0.2813931579589844, 0.5983590698242187, 0.2807306213378906, 0.28087808227539063, 0.2805196838378906, 0.2805125122070313, 0.2804551696777344, 0.2805780334472656, 0.2806200256347656, 0.2805002136230469, 0.28068453979492186, 0.28058010864257815, 0.2806302795410156, 0.2806098022460938, 0.2805196838378906, 0.2809016418457031, 0.28069989013671875, 0.28064971923828125, 0.28055755615234373, 0.28064666748046874, 0.28067019653320313, 0.2806671447753906, 0.28063641357421876, 0.28048590087890624, 0.2806886291503906, 0.2806722412109375, 0.2805166015625, 0.2806087646484375, 0.2805964660644531, 0.28062106323242186, 0.2810009460449219, 0.2807500915527344, 0.2806753234863281, 0.28076235961914064, 0.28126412963867187, 0.28134194946289065, 0.281164794921875, 0.28111666870117186, 0.2810347595214844, 0.28092620849609373, 0.2809241638183594, 0.28083096313476563, 0.28105523681640626, 0.2809354248046875, 0.2808258666992188, 0.281017333984375, 0.2810296325683594, 0.2807910461425781, 0.2810521545410156, 0.28102859497070315, 0.2807940979003906, 0.2811197509765625, 0.28095590209960936, 0.2808842163085937, 0.2810460205078125, 0.2808913879394531, 0.2809292907714844, 0.2807777404785156, 0.28074188232421876, 0.28088525390625, 0.280995849609375, 0.28099172973632813, 0.28075213623046874, 0.28117196655273435, 0.5974200439453125, 0.28051559448242186, 0.28070501708984374, 0.28070501708984374, 0.280669189453125, 0.28124365234375, 0.2808965148925781, 0.2807500915527344, 0.2808545227050781, 0.2804951171875, 0.280748046875, 0.28090264892578126, 0.28080435180664065, 0.2806118469238281, 0.28069580078125, 0.2808616943359375, 0.28078387451171877, 0.2806773681640625, 0.28072039794921877, 0.2808463439941406, 0.28071014404296873, 0.2810408935546875, 0.28070501708984374, 0.28084530639648436, 0.2805821533203125, 0.2807767028808594, 0.2807459716796875, 0.2806599731445312, 0.28074188232421876, 0.28069580078125, 0.2806732788085938, 0.2806507568359375, 0.2808084411621094, 0.2809466857910156, 0.2810572814941406, 0.2805780334472656, 0.28073883056640625, 0.28082891845703123, 0.2810071105957031, 0.28080331420898436, 0.28094976806640626, 0.28130303955078123, 0.28097842407226564, 0.2809968566894531, 0.2808493957519531, 0.2807193603515625, 0.2809456787109375, 0.2811535339355469, 0.2808330383300781, 0.28075418090820314, 0.280853515625, 0.2809169921875, 0.28076953125, 0.280890380859375, 0.2809374694824219, 0.281280517578125, 0.28133172607421875, 0.28101324462890626, 0.28093438720703123, 0.28075314331054685, 0.2808053894042969, 0.28079205322265627, 0.2809333801269531, 0.5982566528320312, 0.2806384582519531, 0.28069989013671875, 0.28062823486328126, 0.2805821533203125, 0.28079205322265627, 0.2808053894042969, 0.28074496459960935, 0.28080743408203124, 0.2807490539550781, 0.2807940979003906, 0.2807173156738281, 0.28145458984375, 0.281248779296875, 0.28092825317382814, 0.2808186950683594, 0.280995849609375, 0.2806609802246094, 0.2807244873046875, 0.2807807922363281, 0.28073779296875, 0.28084326171875, 0.2809169921875, 0.2808883056640625, 0.2809671630859375, 0.2806528015136719, 0.28088525390625, 0.2807490539550781, 0.28113714599609374, 0.2809948120117188, 0.28080331420898436, 0.28099993896484377, 0.28106137084960936, 0.2807500915527344, 0.28100302124023435, 0.280774658203125, 0.28079718017578126, 0.28105319213867186, 0.2808555603027344, 0.2810224609375, 0.2812200927734375, 0.28123544311523435, 0.2811412353515625, 0.28102554321289064, 0.28113201904296875, 0.2809661560058594, 0.280995849609375, 0.2812538757324219, 0.2814095458984375, 0.2813102111816406, 0.28100506591796875, 0.28089447021484376, 0.28107672119140625, 0.28094464111328127, 0.28103884887695313, 0.28113613891601563, 0.2810777587890625, 0.28092312622070315, 0.2810982360839844, 0.280875, 0.2811453552246094, 0.28104193115234377, 0.28097842407226564, 0.6000148315429688, 0.28073370361328126, 0.2808883056640625, 0.28085964965820315, 0.2808688659667969, 0.28098355102539063, 0.280806396484375, 0.28103988647460937, 0.2807173156738281, 0.2808053894042969, 0.28107879638671873, 0.2807080993652344, 0.2809938049316406, 0.28080947875976564, 0.28094259643554687, 0.28093646240234377, 0.2807214050292969, 0.2808412170410156, 0.28081048583984375, 0.2815068054199219, 0.28095590209960936, 0.28094976806640626, 0.28077978515625, 0.2809466857910156, 0.2811985778808594, 0.2808299560546875, 0.28084530639648436, 0.28066201782226563, 0.2807275390625, 0.280890380859375, 0.28104702758789063, 0.28076748657226563, 0.2809395141601562, 0.2808934326171875, 0.2809876403808594, 0.28116683959960936, 0.28092007446289063, 0.2807060546875, 0.2810091552734375, 0.28097024536132814, 0.280890380859375, 0.28086782836914065, 0.28100811767578127, 0.28064358520507815, 0.2809088134765625, 0.2808115234375, 0.2807173156738281, 0.28098150634765623, 0.28097537231445313, 0.2808688659667969, 0.2809374694824219, 0.28094976806640626, 0.2810654602050781, 0.28078387451171877, 0.2811709289550781, 0.2815887451171875, 0.2812333984375, 0.28107879638671873, 0.2810439758300781, 0.281069580078125, 0.28113818359375, 0.28109515380859373, 0.2812569580078125, 0.6002237548828125, 0.280869873046875, 0.2811043701171875, 0.28084225463867185, 0.2808985595703125, 0.28092007446289063, 0.28068658447265626, 0.28079513549804686, 0.28076953125, 0.28095281982421877, 0.2808002624511719, 0.28057089233398436, 0.28077978515625, 0.28081460571289063, 0.2807982177734375, 0.2810357666015625, 0.28085760498046874, 0.28081561279296874, 0.28080947875976564, 0.2812375183105469, 0.28106341552734376, 0.2807726135253906, 0.2809108581542969, 0.2809692077636719, 0.2808115234375, 0.28083712768554686, 0.2811781005859375, 0.28099172973632813, 0.2809948120117188, 0.2808340454101563, 0.28090264892578126, 0.28101324462890626, 0.2809876403808594, 0.2817751159667969, 0.2810347595214844, 0.2808299560546875, 0.281091064453125, 0.28080435180664065, 0.28114227294921873, 0.2809620361328125, 0.2809395141601562, 0.28100811767578127, 0.28102041625976565, 0.2808565673828125, 0.28097024536132814, 0.2810491027832031, 0.281064453125, 0.281091064453125, 0.2807859191894531, 0.2808688659667969, 0.28093646240234377, 0.28071832275390624, 0.2809395141601562, 0.2809599914550781, 0.2808821716308594, 0.28085247802734375, 0.28103372192382814, 0.2807777404785156, 0.2810357666015625, 0.28092108154296874, 0.28099789428710936, 0.28096820068359374, 0.28156927490234374, 0.5991085815429688, 0.280742919921875, 0.2808258666992188, 0.2806661071777344, 0.28075418090820314, 0.2806241149902344, 0.2807306213378906, 0.2808197021484375, 0.2806886291503906, 0.2807244873046875, 0.28067019653320313, 0.2805350341796875, 0.2807224426269531, 0.280869873046875, 0.28091390991210935, 0.2807828369140625, 0.2808197021484375, 0.28051046752929687, 0.2806558837890625, 0.280774658203125, 0.28073165893554686, 0.2806384582519531, 0.280711181640625, 0.28101223754882815, 0.28087091064453124, 0.2806732788085938, 0.28076748657226563, 0.28072653198242187, 0.28075418090820314, 0.28080230712890625, 0.2811269226074219, 0.28098046875, 0.28079513549804686, 0.281059326171875, 0.2811566162109375, 0.2811617431640625, 0.2810163269042969, 0.28065689086914064, 0.2807767028808594, 0.2808002624511719, 0.28075726318359373, 0.28086782836914065, 0.28077566528320314, 0.2806456298828125, 0.28085964965820315, 0.28072039794921877, 0.2807767028808594, 0.2809374694824219, 0.28083096313476563, 0.2811514892578125, 0.2812262268066406, 0.2808821716308594, 0.2809948120117188, 0.2811064453125, 0.28104193115234377, 0.2811760559082031, 0.28112997436523435, 0.28082891845703123, 0.2809241638183594, 0.28104806518554687, 0.2811535339355469, 0.2811269226074219, 0.28101223754882815]",tokens/s,3.5034304223728086,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12706,7 +12706,7 @@ OSError: 8 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,EleutherAI/pythia-1.4b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12834,7 +12834,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: CodeGenForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12893,7 +12893,7 @@ OSError: / does not appear to have a file named config.json. Checkout 'https://h ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-2-1_6b,stabilityai/stablelm-2-1_6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1443.536896,2426.929152,0.0,1780.482048,1621.410304,s,10,0.7068105316162109,0.07068105316162109,0.00033069583366084575,0.07070081710815429,0.07103355484008789,0.07120412101745605,0.07134057395935059,"[0.07069222259521485, 0.0702221450805664, 0.07099565124511718, 0.07054032135009766, 0.07070941162109375, 0.07019967651367187, 0.07137468719482422, 0.07050953674316406, 0.07078886413574219, 0.07077801513671875]",tokens/s,3621.904153219448,kWh,8.280014542525558e-07,4.5352803657666435e-07,3.715746950839271e-06,4.997276441668491e-06,tokens/kWh,51227904.43718713,MB,1443.536896,2426.929152,0.0,1780.482048,1723.272192,s,10,38.065419677734376,3.8065419677734376,0.002289567625927924,3.8060361328125003,3.8076869140625003,3.8103506591796874,3.8124816552734373,"[3.813014404296875, 3.806056640625, 3.804952880859375, 3.806771484375, 3.8046533203125, 3.805589599609375, 3.80634765625, 3.806015625, 3.807094970703125, 3.804923095703125]",tokens/s,16.55045459458066,kWh,4.493906886566492e-05,2.4628168189011863e-05,0.00019878599838376527,0.000268353235438442,tokens/kWh,234765.19631697034,,s,629,38.640658420562765,0.061431889380862865,0.008435353672886576,0.06039961624145508,0.06047948837280273,0.06056448059082031,0.13117497192382813,"[0.06107852935791016, 0.06100889587402344, 0.06097612762451172, 0.06086963272094727, 0.061018112182617185, 0.06080614471435547, 0.06082867050170898, 0.06069145584106445, 0.06044672012329102, 0.060388351440429686, 0.06043033599853516, 0.06039449691772461, 0.06051430511474609, 0.060368896484375, 0.06042828750610352, 0.06034124755859375, 0.06040063858032227, 0.0605747184753418, 0.06040063858032227, 0.06043340682983398, 0.06040063858032227, 0.06036070251464844, 0.06043648147583008, 0.06034329605102539, 0.060440574645996094, 0.060382209777832034, 0.060477439880371096, 0.060386302947998044, 0.060402687072753904, 0.06038937759399414, 0.06038425445556641, 0.060371967315673826, 0.06056140899658203, 0.060434432983398435, 0.06046003341674805, 0.0605665283203125, 0.0605030403137207, 0.06035968017578125, 0.06042828750610352, 0.060386302947998044, 0.06044672012329102, 0.060416000366210934, 0.060462078094482424, 0.060398593902587894, 0.06042316818237305, 0.060455936431884766, 0.06044876861572265, 0.06036070251464844, 0.06043135833740235, 0.06038323211669922, 0.060483585357666014, 0.06042316818237305, 0.060598270416259765, 0.06038118362426758, 0.06044569778442383, 0.06175641632080078, 0.060464126586914066, 0.060442623138427735, 0.06043648147583008, 0.06043545532226562, 0.060418048858642576, 0.060424190521240234, 0.13181951904296876, 0.060440574645996094, 0.060472320556640625, 0.060391422271728515, 0.06032588958740234, 0.0603955192565918, 0.060331008911132813, 0.060371967315673826, 0.06035148620605469, 0.06034431838989258, 0.06032179260253906, 0.060355583190917966, 0.060352512359619144, 0.060385280609130856, 0.0603422737121582, 0.06037913513183594, 0.06034124755859375, 0.06037606430053711, 0.06037401580810547, 0.06042214584350586, 0.060349441528320315, 0.06038118362426758, 0.06040063858032227, 0.06038118362426758, 0.06034534454345703, 0.06042009735107422, 0.06041190338134766, 0.06039961624145508, 0.0603770866394043, 0.06039244842529297, 0.06035660934448242, 0.060388351440429686, 0.06036172866821289, 0.06043135833740235, 0.06046515274047851, 0.06044364929199219, 0.06042316818237305, 0.060483585357666014, 0.06034841537475586, 0.06043648147583008, 0.06038016128540039, 0.060409854888916016, 0.06041497421264649, 0.06037606430053711, 0.06034739303588867, 0.06041190338134766, 0.060437503814697265, 0.06042009735107422, 0.060368896484375, 0.060365825653076174, 0.06037094497680664, 0.06043545532226562, 0.06042316818237305, 0.06043852615356445, 0.060368896484375, 0.060437503814697265, 0.06047436904907227, 0.060375038146972655, 0.060437503814697265, 0.06037606430053711, 0.06049280166625977, 0.060416000366210934, 0.060424190521240234, 0.13166899108886718, 0.06033919906616211, 0.06037094497680664, 0.0603873291015625, 0.060402687072753904, 0.06045798492431641, 0.060355583190917966, 0.06040678405761719, 0.0603504638671875, 0.06037401580810547, 0.06034636688232422, 0.06035865783691406, 0.060352512359619144, 0.06037299346923828, 0.06032588958740234, 0.060369918823242184, 0.06034841537475586, 0.060391422271728515, 0.06036787033081055, 0.060404735565185545, 0.060355583190917966, 0.0603873291015625, 0.06036070251464844, 0.06039756774902344, 0.06034636688232422, 0.06039244842529297, 0.06037094497680664, 0.060412960052490236, 0.060467166900634764, 0.060371967315673826, 0.06035456085205078, 0.0603873291015625, 0.06039449691772461, 0.06044672012329102, 0.06039654541015625, 0.06040063858032227, 0.060439552307128906, 0.06043135833740235, 0.06035148620605469, 0.060416000366210934, 0.06039039993286133, 0.06038425445556641, 0.060385280609130856, 0.06039654541015625, 0.06032896041870117, 0.06044364929199219, 0.06042316818237305, 0.06042931365966797, 0.060483585357666014, 0.06036377716064453, 0.060401664733886716, 0.06041907119750976, 0.06040576171875, 0.060402687072753904, 0.06036787033081055, 0.06038425445556641, 0.06044467163085938, 0.060409854888916016, 0.060437503814697265, 0.060369918823242184, 0.06047641754150391, 0.060407806396484375, 0.06041702270507812, 0.13118156433105468, 0.06036684799194336, 0.06035763168334961, 0.0603770866394043, 0.06031155014038086, 0.06035968017578125, 0.06035148620605469, 0.06037606430053711, 0.06033817672729492, 0.060862464904785155, 0.06047129440307617, 0.060401664733886716, 0.0603504638671875, 0.06036684799194336, 0.060369918823242184, 0.060418048858642576, 0.06037606430053711, 0.06039039993286133, 0.06034534454345703, 0.06039244842529297, 0.06033203125, 0.06042316818237305, 0.06031872177124024, 0.06051737594604492, 0.06032896041870117, 0.06043852615356445, 0.06038937759399414, 0.0605296630859375, 0.060375038146972655, 0.06068735885620117, 0.06043033599853516, 0.06040678405761719, 0.0603504638671875, 0.060437503814697265, 0.06040883255004883, 0.06041497421264649, 0.060470272064208984, 0.060434432983398435, 0.060523521423339846, 0.06047334289550781, 0.06040576171875, 0.06041190338134766, 0.060385280609130856, 0.06039756774902344, 0.0603770866394043, 0.06044876861572265, 0.06042726516723633, 0.06041907119750976, 0.06034124755859375, 0.060368896484375, 0.06036479949951172, 0.06041292953491211, 0.060382209777832034, 0.06074265670776367, 0.06057984161376953, 0.06045183944702148, 0.060477439880371096, 0.06036787033081055, 0.0604139518737793, 0.06040883255004883, 0.06046515274047851, 0.060416000366210934, 0.06042828750610352, 0.13133517456054689, 0.06037299346923828, 0.060365825653076174, 0.06038937759399414, 0.060333057403564455, 0.060365825653076174, 0.06033919906616211, 0.060352512359619144, 0.06033407974243164, 0.06035763168334961, 0.06035456085205078, 0.06034124755859375, 0.060336128234863284, 0.06036684799194336, 0.06036479949951172, 0.06038425445556641, 0.06032691192626953, 0.060369918823242184, 0.06036070251464844, 0.06038118362426758, 0.06033203125, 0.06039039993286133, 0.06033407974243164, 0.06036275100708008, 0.060355583190917966, 0.060418048858642576, 0.060385280609130856, 0.06042009735107422, 0.060331008911132813, 0.0603955192565918, 0.060355583190917966, 0.060385280609130856, 0.06039244842529297, 0.06041702270507812, 0.060426239013671876, 0.060424190521240234, 0.06043033599853516, 0.06041497421264649, 0.060355583190917966, 0.06041292953491211, 0.06036275100708008, 0.060407806396484375, 0.06040678405761719, 0.06038323211669922, 0.060437503814697265, 0.06044364929199219, 0.06044569778442383, 0.06041907119750976, 0.060353534698486325, 0.060418048858642576, 0.060385280609130856, 0.060416000366210934, 0.06038937759399414, 0.0604139518737793, 0.060385280609130856, 0.06043648147583008, 0.06046822357177734, 0.0603504638671875, 0.06047129440307617, 0.06038425445556641, 0.06048051071166992, 0.06044876861572265, 0.0604139518737793, 0.130845703125, 0.06035456085205078, 0.06035968017578125, 0.0603955192565918, 0.0603238410949707, 0.06036479949951172, 0.060352512359619144, 0.06036787033081055, 0.060355583190917966, 0.06032896041870117, 0.0603873291015625, 0.06034534454345703, 0.06035865783691406, 0.06034739303588867, 0.06038118362426758, 0.0603504638671875, 0.06037094497680664, 0.060385280609130856, 0.06036377716064453, 0.060401664733886716, 0.06036377716064453, 0.060391422271728515, 0.06037811279296875, 0.06038937759399414, 0.06030950546264648, 0.060393470764160156, 0.060416000366210934, 0.060475391387939455, 0.06039756774902344, 0.06038425445556641, 0.06041497421264649, 0.06039654541015625, 0.06035865783691406, 0.06042316818237305, 0.06040883255004883, 0.06039449691772461, 0.06043545532226562, 0.06064025497436523, 0.060382209777832034, 0.060442623138427735, 0.06040063858032227, 0.060409854888916016, 0.06037606430053711, 0.06038425445556641, 0.06032896041870117, 0.06041702270507812, 0.060424190521240234, 0.06051839828491211, 0.060369918823242184, 0.06036787033081055, 0.06040576171875, 0.06050099182128906, 0.060424190521240234, 0.060402687072753904, 0.060368896484375, 0.060486656188964844, 0.06061670303344727, 0.06039654541015625, 0.06042214584350586, 0.060404735565185545, 0.060560382843017575, 0.06044467163085938, 0.06042009735107422, 0.13205914306640626, 0.06037299346923828, 0.06082252883911133, 0.06047436904907227, 0.06030950546264648, 0.06037811279296875, 0.060336128234863284, 0.060368896484375, 0.06032281494140625, 0.06037811279296875, 0.06031564712524414, 0.06044876861572265, 0.0604661750793457, 0.06039039993286133, 0.060353534698486325, 0.060386302947998044, 0.06032896041870117, 0.060382209777832034, 0.06035865783691406, 0.06039756774902344, 0.06032998275756836, 0.06042931365966797, 0.06038425445556641, 0.060442623138427735, 0.06044876861572265, 0.060391422271728515, 0.0603422737121582, 0.06039654541015625, 0.0604139518737793, 0.06036479949951172, 0.060353534698486325, 0.06038937759399414, 0.060440574645996094, 0.06048051071166992, 0.06043135833740235, 0.06040063858032227, 0.06046822357177734, 0.06043545532226562, 0.060393470764160156, 0.060421119689941405, 0.06042521667480469, 0.060450817108154295, 0.060391422271728515, 0.06040063858032227, 0.060365825653076174, 0.060416000366210934, 0.06047641754150391, 0.06045183944702148, 0.06034124755859375, 0.06060441589355469, 0.060424190521240234, 0.060447742462158206, 0.06042828750610352, 0.06047948837280273, 0.06037299346923828, 0.060418048858642576, 0.06048972702026367, 0.060391422271728515, 0.06043135833740235, 0.060369918823242184, 0.06054195022583008, 0.06044876861572265, 0.06041088104248047, 0.13115802001953125, 0.06042828750610352, 0.06035763168334961, 0.06038323211669922, 0.06030847930908203, 0.06037913513183594, 0.06053887939453125, 0.06036684799194336, 0.06036070251464844, 0.06045695877075195, 0.060365825653076174, 0.060375038146972655, 0.06034431838989258, 0.060369918823242184, 0.06037299346923828, 0.060369918823242184, 0.06036172866821289, 0.060494846343994144, 0.060402687072753904, 0.060462078094482424, 0.06046105575561524, 0.06041907119750976, 0.060421119689941405, 0.06037913513183594, 0.06034636688232422, 0.06039449691772461, 0.06035456085205078, 0.06039247894287109, 0.06034223937988281, 0.060368896484375, 0.060440574645996094, 0.06048153686523437, 0.06046105575561524, 0.060442623138427735, 0.060488704681396485, 0.06050201416015625, 0.060416000366210934, 0.06039449691772461, 0.060382209777832034, 0.06044569778442383, 0.06039654541015625, 0.06040371322631836, 0.06044569778442383, 0.060369918823242184, 0.060401664733886716, 0.0604139518737793, 0.06041497421264649, 0.06039654541015625, 0.06044364929199219, 0.0603873291015625, 0.06037606430053711, 0.060434432983398435, 0.06038425445556641, 0.060424190521240234, 0.06037913513183594, 0.060407806396484375, 0.060470272064208984, 0.06036070251464844, 0.06043648147583008, 0.060527614593505856, 0.060483585357666014, 0.06039961624145508, 0.06042009735107422, 0.1314693145751953, 0.060409854888916016, 0.06036377716064453, 0.060426273345947266, 0.06032380676269531, 0.06034636688232422, 0.06032998275756836, 0.06037811279296875, 0.06034841537475586, 0.06038118362426758, 0.060472320556640625, 0.060371967315673826, 0.06034534454345703, 0.060349441528320315, 0.06040063858032227, 0.06034329605102539, 0.06041907119750976, 0.06053376007080078, 0.06042931365966797, 0.060382209777832034, 0.06047948837280273, 0.060418048858642576, 0.06037094497680664, 0.06039756774902344, 0.06042214584350586, 0.06038118362426758, 0.06034329605102539, 0.06052556610107422, 0.06041190338134766, 0.06060851287841797, 0.06038937759399414, 0.06039961624145508, 0.06040371322631836, 0.06044160079956055, 0.06049792098999023, 0.0604139518737793, 0.060505088806152345, 0.06044160079956055, 0.06043135833740235, 0.06070169448852539, 0.060440574645996094, 0.06040576171875, 0.06042828750610352, 0.060391422271728515, 0.060369918823242184, 0.060447742462158206, 0.06047641754150391, 0.06041907119750976, 0.060447742462158206, 0.06072217559814453, 0.060470272064208984, 0.060469249725341796, 0.06040063858032227, 0.060439552307128906, 0.060385280609130856, 0.06046310424804688, 0.06047436904907227, 0.06038937759399414, 0.060418048858642576, 0.06040576171875, 0.060464126586914066, 0.06039039993286133, 0.06039244842529297, 0.13141401672363281, 0.06035456085205078, 0.060371967315673826, 0.06040883255004883, 0.0603422737121582, 0.060402687072753904, 0.06033715057373047, 0.06034022521972656, 0.06035660934448242, 0.06035763168334961, 0.06033817672729492, 0.060368896484375, 0.060333057403564455, 0.060388351440429686, 0.060368896484375, 0.060393470764160156, 0.06032896041870117, 0.06041292953491211, 0.06034739303588867, 0.06040063858032227, 0.06032793426513672, 0.06037913513183594, 0.06035456085205078, 0.0603770866394043, 0.06034124755859375, 0.06042828750610352, 0.06037401580810547, 0.060442623138427735, 0.06034739303588867, 0.060388351440429686, 0.06035968017578125, 0.060409854888916016, 0.060355583190917966, 0.060442623138427735, 0.06044364929199219, 0.060437503814697265, 0.06044876861572265, 0.06044467163085938, 0.0603770866394043, 0.06041292953491211, 0.060386302947998044, 0.06039961624145508, 0.06038118362426758, 0.06043340682983398, 0.060369918823242184, 0.06043033599853516, 0.06046515274047851, 0.06043033599853516, 0.0603351058959961, 0.060391422271728515, 0.06037401580810547, 0.06042521667480469, 0.0603955192565918, 0.06042009735107422, 0.060371967315673826, 0.06050406265258789, 0.0604661750793457, 0.06038425445556641, 0.06045183944702148, 0.060386302947998044, 0.06047129440307617, 0.060393470764160156, 0.06040678405761719]",tokens/s,16.278190530658136,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-110B,Qwen/Qwen1.5-110B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12935,7 +12935,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciCoder-1b,Deci/DeciCoder-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13354,7 +13354,7 @@ Checkout your internet connection or see how to run the library in offline mode ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-13b,huggyllama/llama-13b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2418.307072,9028.763648,0.0,8382.316544,7939.283456,s,10,6.167790039062499,0.61677900390625,0.0015724819730622269,0.6163188171386719,0.6186426635742188,0.6196823547363282,0.6205141076660157,"[0.6154488525390625, 0.6207220458984375, 0.6163983764648437, 0.616369384765625, 0.6161586303710938, 0.6162682495117188, 0.6171541137695312, 0.61841162109375, 0.6153932495117187, 0.6154655151367188]",tokens/s,415.059524365573,kWh,7.267181906435225e-06,3.982036919842357e-06,3.6222087801172345e-05,4.747130662744993e-05,tokens/kWh,5392731.276791313,MB,2418.307072,9028.763648,0.0,8382.316544,8233.862144,s,10,338.13460546874995,33.81346054687499,0.001160851078231068,33.813755859375,33.814566015625,33.8146169921875,33.8146577734375,"[33.81088671875, 33.81180859375, 33.81329296875, 33.8145546875, 33.81344140625, 33.814203125, 33.81466796875, 33.81398828125, 33.81423828125, 33.8135234375]",tokens/s,1.8631633373538987,kWh,0.0003992094929516316,0.00021880157859535757,0.0019938841196158203,0.0026118951911628095,tokens/kWh,24120.4165516123,,s,629,343.1500421142575,0.5455485566204417,0.0733377889745091,0.5367131958007813,0.5369653442382812,0.5370488891601563,1.1532696044921875,"[0.5365380859375, 0.5367070922851562, 0.5365258178710938, 0.5365678100585938, 0.5363978271484375, 0.5368893432617188, 0.5364459228515625, 0.5367337036132812, 0.53631591796875, 0.5366466674804687, 0.5364623413085937, 0.5367838745117187, 0.5364408569335938, 0.5364674682617188, 0.536364013671875, 0.5366456298828125, 0.5363200073242187, 0.5367244873046875, 0.5366456298828125, 0.5368350830078125, 0.5363937377929687, 0.53671728515625, 0.5367060546875, 0.5366753540039062, 0.536754150390625, 0.5368975219726563, 0.5367050170898438, 0.5365545043945312, 0.5365575561523438, 0.5367234497070312, 0.5365892944335937, 0.5366651000976562, 0.536605712890625, 0.536995849609375, 0.5365831909179688, 0.5367439575195313, 0.5365278930664062, 0.5368299560546875, 0.5366886596679687, 0.536685546875, 0.536537109375, 0.5367859497070312, 0.5366128540039062, 0.5368207397460938, 0.5366558837890625, 0.5369497680664063, 0.5365565185546874, 0.5367982177734375, 0.5366558837890625, 0.5367808227539063, 0.5366763305664063, 0.5368923950195312, 0.5366241455078125, 0.536774658203125, 0.53675927734375, 0.5367725830078125, 0.5367039794921875, 0.536869873046875, 0.5366456298828125, 0.537069580078125, 0.5366548461914062, 0.5369712524414062, 1.158197265625, 0.5365196533203125, 0.5368145751953125, 0.536332275390625, 0.536406005859375, 0.5366067504882812, 0.5366845703125, 0.5364172973632813, 0.5368176879882812, 0.5368688354492187, 0.536642578125, 0.536406005859375, 0.5366333618164062, 0.5364541625976562, 0.5366098022460938, 0.5361817626953125, 0.53676953125, 0.5361551513671875, 0.536685546875, 0.5366466674804687, 0.5368422241210937, 0.5364224243164063, 0.5367869262695313, 0.5369600219726562, 0.5367777099609375, 0.5367060546875, 0.5369262084960937, 0.536658935546875, 0.5365780639648438, 0.5367879638671875, 0.5365667724609375, 0.5368043823242188, 0.53661181640625, 0.5367131958007813, 0.5368822021484375, 0.5366190185546875, 0.53648486328125, 0.5366988525390625, 0.5368545532226563, 0.5366681518554688, 0.5364121704101562, 0.5366712036132812, 0.5368248291015625, 0.5367808227539063, 0.5365872802734375, 0.5368678588867187, 0.536943603515625, 0.5366456298828125, 0.5367275390625, 0.5368524780273437, 0.5368688354492187, 0.5366353759765625, 0.536722412109375, 0.5367449340820313, 0.5368883056640625, 0.5366231079101562, 0.5369866333007812, 0.5368473510742188, 0.5367838745117187, 0.5364449462890625, 0.5368104858398437, 0.5368524780273437, 0.5368975219726563, 1.1535626220703126, 0.5364807739257812, 0.5365350341796875, 0.5364326171875, 0.5363937377929687, 0.5365145874023437, 0.5367337036132812, 0.5366046752929687, 0.5366548461914062, 0.5365155639648438, 0.5366333618164062, 0.5364786987304687, 0.5366466674804687, 0.5365094604492188, 0.5365811157226562, 0.5364182739257812, 0.5364715576171875, 0.536364013671875, 0.53667431640625, 0.536796142578125, 0.5367265014648438, 0.5366466674804687, 0.536521728515625, 0.5366763305664063, 0.5366128540039062, 0.5367654418945312, 0.536900634765625, 0.5364971313476562, 0.5365094604492188, 0.5370931396484375, 0.536764404296875, 0.5365831909179688, 0.5366036376953125, 0.5365330200195313, 0.5368893432617188, 0.5366026000976563, 0.5368606567382812, 0.536816650390625, 0.5369476928710938, 0.5367490844726562, 0.536848388671875, 0.5366753540039062, 0.53680126953125, 0.5367080688476562, 0.5367828369140625, 0.5370787963867187, 0.5374157104492188, 0.536700927734375, 0.5368831787109375, 0.5369343872070312, 0.5367613525390625, 0.536642578125, 0.537175048828125, 0.5367551879882813, 0.5368934326171875, 0.5366619873046875, 0.5368678588867187, 0.536922119140625, 0.5367337036132812, 0.536595458984375, 0.5369989013671875, 0.5368309936523438, 0.5367244873046875, 1.1537080078125, 0.5367060546875, 0.5366784057617188, 0.5363886108398438, 0.5365841674804688, 0.5365995483398438, 0.536474609375, 0.5364500732421875, 0.5364602661132812, 0.5364244384765625, 0.53661083984375, 0.5364818115234375, 0.5368258666992187, 0.5365626831054687, 0.5365032958984375, 0.5363486938476563, 0.5364080810546875, 0.536258544921875, 0.5372548828125, 0.5367050170898438, 0.5369456787109375, 0.5366630249023437, 0.5366046752929687, 0.5366466674804687, 0.53688525390625, 0.53661083984375, 0.5367572631835937, 0.5366036376953125, 0.5366927490234376, 0.5367080688476562, 0.5365616455078125, 0.5367736206054687, 0.5366415405273437, 0.5366763305664063, 0.5369446411132812, 0.53675927734375, 0.536859619140625, 0.53688525390625, 0.5369733276367188, 0.5365831909179688, 0.5367142333984375, 0.5367131958007813, 0.536679443359375, 0.5372713012695313, 0.5369129028320313, 0.5368627319335938, 0.5370460205078125, 0.53677978515625, 0.5368248291015625, 0.5368831787109375, 0.5366865844726563, 0.5365606689453125, 0.5368023071289062, 0.5370009765625, 0.536958984375, 0.536795166015625, 0.5366497192382812, 0.5369476928710938, 0.5372333984375, 0.5365616455078125, 0.5370224609375, 0.5368002319335937, 0.53704296875, 1.153555419921875, 0.536501220703125, 0.5367725830078125, 0.5364859008789062, 0.5365391235351562, 0.53650634765625, 0.536647705078125, 0.5364469604492188, 0.5366046752929687, 0.536479736328125, 0.5366087646484375, 0.5368074340820312, 0.5366292724609375, 0.5366722412109375, 0.53648486328125, 0.5362606201171874, 0.5367900390625, 0.53659033203125, 0.5366394653320312, 0.536732666015625, 0.5366343383789063, 0.5365166015625, 0.536553466796875, 0.536722412109375, 0.5367357177734375, 0.5368555297851563, 0.536791015625, 0.5365360717773437, 0.5365933837890625, 0.5367091064453124, 0.5368463134765625, 0.53677978515625, 0.5366712036132812, 0.5366599731445313, 0.5368442993164062, 0.5365278930664062, 0.536806396484375, 0.53661181640625, 0.5368319702148437, 0.5365330200195313, 0.5365309448242187, 0.5367255249023437, 0.5366333618164062, 0.5368831787109375, 0.5366343383789063, 0.5369200439453125, 0.5369671630859375, 0.5370470581054687, 0.5367920532226562, 0.5366865844726563, 0.536700927734375, 0.5368156127929687, 0.536958984375, 0.5368831787109375, 0.5370194091796875, 0.5367992553710937, 0.5370132446289062, 0.5367459716796875, 0.536838134765625, 0.536479736328125, 0.5374484252929688, 0.53677978515625, 0.5368411865234375, 1.153471435546875, 0.5364951171875, 0.5366415405273437, 0.5368473510742188, 0.5363834838867187, 0.5366210327148437, 0.536531982421875, 0.5365493774414063, 0.536679443359375, 0.536647705078125, 0.5366435546875, 0.5364019165039062, 0.5366824951171875, 0.5364306030273438, 0.5364357299804687, 0.536369140625, 0.536427490234375, 0.5365278930664062, 0.5366179809570313, 0.5367347412109374, 0.5368822021484375, 0.5366978759765625, 0.5366753540039062, 0.536853515625, 0.53673779296875, 0.536754150390625, 0.5367623901367188, 0.5366763305664063, 0.5367183227539063, 0.5367817993164062, 0.5367337036132812, 0.5367459716796875, 0.5368053588867188, 0.5364777221679687, 0.5369343872070312, 0.5366507568359375, 0.5367244873046875, 0.536754150390625, 0.536985595703125, 0.5369190673828125, 0.5368668212890625, 0.5366292724609375, 0.536880126953125, 0.5368176879882812, 0.5366988525390625, 0.5369866333007812, 0.5370194091796875, 0.5367459716796875, 0.5368370971679688, 0.5365135498046875, 0.53667431640625, 0.53671630859375, 0.5368494262695312, 0.53669580078125, 0.53698046875, 0.5367992553710937, 0.536774658203125, 0.5369190673828125, 0.5369292602539063, 0.5370726318359375, 0.5369497680664063, 0.5366865844726563, 0.5370203857421875, 1.154618408203125, 0.5365596313476563, 0.5364193115234375, 0.5363773193359375, 0.5362728881835938, 0.5365473022460937, 0.5363855590820312, 0.5364039916992187, 0.5364664306640625, 0.5363722534179688, 0.536680419921875, 0.5364766845703125, 0.5371268920898438, 0.5366917114257812, 0.5365166015625, 0.5365975341796875, 0.5365575561523438, 0.5364418334960938, 0.5368084716796875, 0.5367070922851562, 0.5367449340820313, 0.5367603149414063, 0.5367551879882813, 0.536754150390625, 0.5367510986328125, 0.5367684936523438, 0.5367675170898437, 0.5366599731445313, 0.5364961547851562, 0.5366578979492187, 0.536806396484375, 0.5367562255859375, 0.5367408447265625, 0.536627197265625, 0.5369661254882813, 0.5367470092773438, 0.5369343872070312, 0.5368074340820312, 0.5368576049804688, 0.536795166015625, 0.5367285766601563, 0.5365309448242187, 0.5366722412109375, 0.5369343872070312, 0.5367562255859375, 0.5368893432617188, 0.5369651489257813, 0.5368514404296875, 0.5369467163085937, 0.5367101440429688, 0.5367296142578125, 0.5370715942382812, 0.53705419921875, 0.5369313354492188, 0.5368729858398438, 0.5368361206054687, 0.5369497680664063, 0.5369282836914062, 0.5368883056640625, 0.5366251220703125, 0.5370183715820313, 0.5368709106445313, 0.5369528198242187, 1.1556219482421874, 0.5364859008789062, 0.5367736206054687, 0.5364080810546875, 0.536553466796875, 0.5366824951171875, 0.5366098022460938, 0.5365892944335937, 0.53646337890625, 0.536279052734375, 0.5366405029296875, 0.5365493774414063, 0.53646337890625, 0.5365104370117187, 0.5364838256835938, 0.5365084228515625, 0.5366610107421875, 0.5364357299804687, 0.536647705078125, 0.5364889526367187, 0.536748046875, 0.5367930908203125, 0.5366671142578125, 0.5367449340820313, 0.5367019653320313, 0.5370091552734375, 0.536827880859375, 0.536573974609375, 0.5367787475585938, 0.5366098022460938, 0.536838134765625, 0.5366077270507813, 0.5368043823242188, 0.5365616455078125, 0.5371207885742187, 0.5365278930664062, 0.5368494262695312, 0.5366507568359375, 0.5369343872070312, 0.5367572631835937, 0.5367551879882813, 0.5369886474609376, 0.5369292602539063, 0.5371954956054688, 0.536795166015625, 0.5366947631835938, 0.536932373046875, 0.5366435546875, 0.5369630737304687, 0.5365084228515625, 0.537143310546875, 0.5367654418945312, 0.536848388671875, 0.5367890014648438, 0.5368955078125, 0.5366302490234375, 0.5367152709960937, 0.5368729858398438, 0.5369559326171875, 0.5366732788085937, 0.5370153198242188, 0.536826904296875, 0.536880126953125, 1.1527025146484375, 0.536489990234375, 0.5366835327148437, 0.5366138916015625, 0.5365176391601563, 0.5364889526367187, 0.536453125, 0.5365125122070312, 0.5366312866210937, 0.5364080810546875, 0.5364920043945313, 0.5364111328125, 0.5365258178710938, 0.5365238037109376, 0.5367664794921875, 0.5364930419921875, 0.5364602661132812, 0.5364971313476562, 0.536774658203125, 0.5366835327148437, 0.5369774169921875, 0.5367152709960937, 0.53650634765625, 0.53692724609375, 0.5366179809570313, 0.5366988525390625, 0.5368176879882812, 0.5365545043945312, 0.5367337036132812, 0.536700927734375, 0.5366466674804687, 0.536774658203125, 0.5367050170898438, 0.53663232421875, 0.5369866333007812, 0.5367091064453124, 0.5367388305664063, 0.5366220703125, 0.5368125610351563, 0.5369733276367188, 0.5369313354492188, 0.53682177734375, 0.53686474609375, 0.5370634155273437, 0.5370572509765625, 0.536827880859375, 0.536806396484375, 0.536573974609375, 0.5369467163085937, 0.5368350830078125, 0.5367930908203125, 0.536605712890625, 0.5369088134765625, 0.536859619140625, 0.5371924438476563, 0.5367551879882813, 0.5368657836914063, 0.5368893432617188, 0.5367039794921875, 0.5366220703125, 0.5370398559570313, 0.5369682006835937, 0.5367817993164062, 1.1527506103515626, 0.5365135498046875, 0.5366722412109375, 0.536648681640625, 0.5365913696289063, 0.536416259765625, 0.5366036376953125, 0.5365339965820313, 0.5371463623046875, 0.53639990234375, 0.536574951171875, 0.5365350341796875, 0.5370501098632813, 0.5364541625976562, 0.5363189697265625, 0.5364039916992187, 0.5364520874023437, 0.5365821533203124, 0.5366282348632813, 0.5366784057617188, 0.5367418823242187, 0.5366732788085937, 0.536711181640625, 0.5367203979492188, 0.5366394653320312, 0.5366917114257812, 0.5369784545898437, 0.53665380859375, 0.5367470092773438, 0.5368893432617188, 0.536816650390625, 0.5366619873046875, 0.5365606689453125, 0.5364520874023437, 0.536932373046875, 0.5365616455078125, 0.5367787475585938, 0.5365401611328126, 0.5368002319335937, 0.5364520874023437, 0.5367603149414063, 0.5366005859375, 0.536791015625, 0.5371688842773438, 0.53677978515625, 0.5367255249023437, 0.5369937744140625, 0.536795166015625, 0.5367992553710937, 0.5370582885742188, 0.5368258666992187, 0.536658935546875, 0.5370296020507812, 0.5368555297851563, 0.5368770751953125, 0.5367736206054687, 0.5367725830078125, 0.5368186645507812, 0.536853515625, 0.5365596313476563, 0.5368340454101562, 0.5368145751953125, 0.5369313354492188]",tokens/s,1.8330174058103814,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-7B,Qwen/Qwen-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13383,7 +13383,7 @@ ChildProcessError: Traceback (most recent call last): ImportError: This modeling file requires the following packages that were not found in your environment: transformers_stream_generator. Run `pip install transformers_stream_generator` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-72B,Qwen/Qwen1.5-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13425,7 +13425,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 96.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Deci/DeciLM-7B,Deci/DeciLM-7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13496,7 +13496,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-j-6b,EleutherAI/gpt-j-6b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13568,7 +13568,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-160m,EleutherAI/pythia-160m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13683,7 +13683,7 @@ ChildProcessError: Traceback (most recent call last): RuntimeError: The size of tensor a (60) must match the size of tensor b (32) at non-singleton dimension 2 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.3b,EleutherAI/pythia-1.3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13712,7 +13712,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-2.7b,EleutherAI/pythia-2.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13774,7 +13774,7 @@ ValueError: OPTForCausalLM does not support an attention implementation through ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,mistralai/Mistral-7B-v0.1,mistralai/Mistral-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5086.445568,5956.435968,0.0,5309.988864,5047.4112,s,10,3.3951274108886715,0.33951274108886714,0.00038503905864011734,0.33945068359375,0.33990344848632814,0.34012059326171873,0.34029430908203123,"[0.33982321166992185, 0.3391507568359375, 0.33938748168945315, 0.33922314453125, 0.34033773803710937, 0.3395138854980469, 0.33896197509765624, 0.3398551940917969, 0.3395834045410156, 0.3392906188964844]",tokens/s,754.0217759692035,kWh,4.00638022908458e-06,2.1953130347537806e-06,1.9920034454532867e-05,2.6121727718371226e-05,tokens/kWh,9800270.593126083,MB,5086.445568,5956.435968,0.0,5309.988864,5098.333184,s,10,188.221390625,18.822139062500003,0.0011505271698466547,18.8222666015625,18.8231517578125,18.82387861328125,18.82446009765625,"[18.821546875, 18.82124609375, 18.822708984375, 18.82460546875, 18.822560546875, 18.82055859375, 18.8222265625, 18.822306640625, 18.820640625, 18.822990234375]",tokens/s,3.3471222261616944,kWh,0.00022222505159952025,0.00012179794381854662,0.0010774225934188715,0.0014214455888369383,tokens/kWh,44321.077426219425,,s,629,190.97790853881838,0.3036214762143376,0.0403576770019219,0.29874688720703124,0.2989168701171875,0.2990350280761719,0.6382961645507812,"[0.2991718444824219, 0.29857177734375, 0.2985912170410156, 0.2988011474609375, 0.29874789428710935, 0.2987694091796875, 0.29878680419921877, 0.29858917236328125, 0.29871923828125, 0.29875506591796874, 0.29862911987304686, 0.29869976806640625, 0.29868133544921877, 0.29870693969726564, 0.2988308410644531, 0.2986854248046875, 0.2986752014160156, 0.2989803466796875, 0.29859634399414064, 0.2987222900390625, 0.2987008056640625, 0.2986854248046875, 0.2986875, 0.2986332092285156, 0.2987663269042969, 0.29869259643554685, 0.2985953369140625, 0.29869976806640625, 0.298566650390625, 0.29867929077148436, 0.29865780639648437, 0.29856668090820315, 0.29851031494140623, 0.2987509765625, 0.2991912841796875, 0.2989486083984375, 0.29886465454101563, 0.2987008056640625, 0.29902334594726565, 0.2987386779785156, 0.29886257934570315, 0.29884622192382815, 0.29862298583984376, 0.2987376708984375, 0.29878475952148437, 0.2987100219726562, 0.2987274169921875, 0.29880218505859374, 0.2987960205078125, 0.29913088989257813, 0.2988216247558594, 0.29875405883789063, 0.2986885070800781, 0.29871206665039063, 0.2987673645019531, 0.29861785888671877, 0.29866598510742187, 0.2988287963867188, 0.2988287963867188, 0.29879705810546875, 0.2987857971191406, 0.2987950134277344, 0.6382120971679688, 0.29854412841796873, 0.2985953369140625, 0.2985646057128906, 0.29854104614257815, 0.298724365234375, 0.29859634399414064, 0.29881240844726564, 0.29854617309570314, 0.29873358154296875, 0.2988380126953125, 0.29890969848632815, 0.2988410949707031, 0.29858612060546874, 0.2986250305175781, 0.2987253723144531, 0.29870693969726564, 0.29874893188476564, 0.2987386779785156, 0.2988083190917969, 0.29867315673828126, 0.2986280822753906, 0.29873358154296875, 0.2987376708984375, 0.2986905517578125, 0.298777587890625, 0.29880523681640625, 0.29867828369140625, 0.2988226623535156, 0.2988072814941406, 0.2987427978515625, 0.29883905029296876, 0.29878680419921877, 0.2985922546386719, 0.29876837158203123, 0.29872332763671877, 0.29868646240234376, 0.2986690673828125, 0.2988482666015625, 0.29870181274414065, 0.29881857299804687, 0.29884518432617185, 0.2989219970703125, 0.2987427978515625, 0.2989219970703125, 0.2988748779296875, 0.2987110290527344, 0.29886566162109374, 0.2988779602050781, 0.29880523681640625, 0.29869464111328126, 0.298745849609375, 0.2987427978515625, 0.29859942626953123, 0.2986700744628906, 0.29881857299804687, 0.2987734985351562, 0.29879806518554686, 0.298703857421875, 0.2987110290527344, 0.298756103515625, 0.2991769714355469, 0.2987376708984375, 0.6383523559570312, 0.2986219482421875, 0.29856051635742187, 0.29854617309570314, 0.2987694091796875, 0.29871923828125, 0.298534912109375, 0.2986680297851563, 0.29889434814453125, 0.2986250305175781, 0.29877044677734377, 0.29878375244140626, 0.2986680297851563, 0.2991124572753906, 0.2987796630859375, 0.2987530212402344, 0.2987222900390625, 0.2987888488769531, 0.2988287963867188, 0.2987950134277344, 0.2988482666015625, 0.29875405883789063, 0.29862603759765627, 0.2985953369140625, 0.2988308410644531, 0.2987632751464844, 0.2986342468261719, 0.299082763671875, 0.2986588134765625, 0.2988001403808594, 0.29873358154296875, 0.29896295166015624, 0.29876123046875, 0.2986895446777344, 0.298682373046875, 0.2987694091796875, 0.29866494750976563, 0.298819580078125, 0.2986956787109375, 0.2985513000488281, 0.29881240844726564, 0.2988011474609375, 0.298814453125, 0.29863833618164065, 0.29881753540039063, 0.2988830871582031, 0.29885440063476565, 0.29881036376953124, 0.29866290283203123, 0.298682373046875, 0.29867315673828126, 0.29896395874023435, 0.29883700561523435, 0.2987694091796875, 0.2988328857421875, 0.29885748291015624, 0.2988564453125, 0.29895269775390626, 0.2987806701660156, 0.29891278076171873, 0.2988830871582031, 0.29890765380859374, 0.2989168701171875, 0.6389586181640625, 0.29872128295898437, 0.2986536865234375, 0.29879910278320315, 0.2986854248046875, 0.29872128295898437, 0.29872845458984376, 0.2987356262207031, 0.29869464111328126, 0.29878375244140626, 0.2991523742675781, 0.2986168212890625, 0.29872640991210936, 0.2986588134765625, 0.29912472534179685, 0.29872845458984376, 0.2986956787109375, 0.29874688720703124, 0.2986332092285156, 0.29874789428710935, 0.29879806518554686, 0.29864346313476564, 0.29869876098632814, 0.29862399291992187, 0.29865676879882813, 0.298787841796875, 0.2986106872558594, 0.29867724609375, 0.2988134460449219, 0.298756103515625, 0.29874072265625, 0.2988482666015625, 0.2986065979003906, 0.29880422973632814, 0.298756103515625, 0.2998691711425781, 0.2986905517578125, 0.298714111328125, 0.2987028503417969, 0.2987509765625, 0.2988083190917969, 0.2988206176757813, 0.2987182006835937, 0.298703857421875, 0.29875201416015623, 0.29893121337890627, 0.2988267517089844, 0.2989168701171875, 0.2986741638183594, 0.2989014892578125, 0.29890866088867185, 0.2989270935058594, 0.29876837158203123, 0.29879910278320315, 0.29877862548828127, 0.29884622192382815, 0.29892095947265623, 0.2989322204589844, 0.298756103515625, 0.2989404296875, 0.29872845458984376, 0.2988206176757813, 0.29878375244140626, 0.63944189453125, 0.29866494750976563, 0.29873458862304686, 0.29941351318359377, 0.2987222900390625, 0.2986700744628906, 0.298650634765625, 0.29863015747070315, 0.29849395751953123, 0.2986536865234375, 0.29863729858398436, 0.29870489501953124, 0.2986311950683594, 0.2986546936035156, 0.298714111328125, 0.298745849609375, 0.29884210205078127, 0.29883187866210936, 0.2987581481933594, 0.2987724914550781, 0.2988584899902344, 0.298640380859375, 0.2987530212402344, 0.29866598510742187, 0.2987734985351562, 0.2988441467285156, 0.29874993896484375, 0.2986393737792969, 0.2987663269042969, 0.2986833801269531, 0.29886053466796875, 0.29871206665039063, 0.29884622192382815, 0.2987632751464844, 0.2985902099609375, 0.29863116455078126, 0.29862911987304686, 0.29875506591796874, 0.2986752014160156, 0.2986854248046875, 0.29865472412109373, 0.298703857421875, 0.2987304992675781, 0.29888204956054687, 0.2987386779785156, 0.2986557312011719, 0.2988277893066406, 0.29886465454101563, 0.2989363098144531, 0.298903564453125, 0.2989363098144531, 0.2989404296875, 0.29923941040039065, 0.29895782470703125, 0.298714111328125, 0.2987028503417969, 0.2989168701171875, 0.2988216247558594, 0.29880935668945313, 0.29886465454101563, 0.29877658081054687, 0.2988164978027344, 0.29866598510742187, 0.6385172729492188, 0.2987161560058594, 0.29870181274414065, 0.29873458862304686, 0.2985277404785156, 0.29870181274414065, 0.29862399291992187, 0.2987222900390625, 0.29852978515625, 0.2984530029296875, 0.29863833618164065, 0.2985779113769531, 0.2986393737792969, 0.2987673645019531, 0.2987304992675781, 0.2987028503417969, 0.2987581481933594, 0.298576904296875, 0.2986188659667969, 0.29863116455078126, 0.29871307373046874, 0.2985830383300781, 0.29869259643554685, 0.29868441772460935, 0.29859634399414064, 0.29861785888671877, 0.2987222900390625, 0.2986536865234375, 0.2988011474609375, 0.2985912170410156, 0.2987427978515625, 0.29869259643554685, 0.2987960205078125, 0.2987673645019531, 0.2987694091796875, 0.29872946166992187, 0.29888204956054687, 0.29878475952148437, 0.2987581481933594, 0.29880422973632814, 0.2987530212402344, 0.29877044677734377, 0.298819580078125, 0.29879910278320315, 0.29877862548828127, 0.29886465454101563, 0.29878271484375, 0.29871923828125, 0.298576904296875, 0.2992762756347656, 0.29873663330078126, 0.2988359680175781, 0.29874688720703124, 0.29913394165039064, 0.2987376708984375, 0.29887387084960937, 0.29878988647460936, 0.29888101196289063, 0.29867724609375, 0.29872128295898437, 0.2988380126953125, 0.2989219970703125, 0.29882366943359373, 0.6379100341796875, 0.29849395751953123, 0.2987642822265625, 0.2986956787109375, 0.2986536865234375, 0.2986342468261719, 0.29868646240234376, 0.29869259643554685, 0.2986905517578125, 0.298967041015625, 0.2986536865234375, 0.2986956787109375, 0.29867111206054686, 0.2986670227050781, 0.29867724609375, 0.29880218505859374, 0.2986833801269531, 0.2987028503417969, 0.29863116455078126, 0.29869363403320315, 0.2986977233886719, 0.29881753540039063, 0.2988114013671875, 0.2987325439453125, 0.2990878601074219, 0.2988226623535156, 0.2988031921386719, 0.29867623901367185, 0.2988410949707031, 0.29869158935546875, 0.2987182006835937, 0.29871514892578127, 0.29878988647460936, 0.2987304992675781, 0.29869363403320315, 0.29874789428710935, 0.298756103515625, 0.2987591552734375, 0.2987161560058594, 0.29865676879882813, 0.29869876098632814, 0.29875711059570315, 0.2986875, 0.29886566162109374, 0.29875711059570315, 0.2986752014160156, 0.2987008056640625, 0.29882571411132813, 0.2986741638183594, 0.29876019287109373, 0.29889434814453125, 0.2987161560058594, 0.29932748413085936, 0.29880422973632814, 0.2988687438964844, 0.29856051635742187, 0.29914111328125, 0.29886566162109374, 0.29886566162109374, 0.298956787109375, 0.29882366943359373, 0.29876123046875, 0.29887387084960937, 0.6389237670898438, 0.2986998291015625, 0.29864031982421874, 0.2987253723144531, 0.2986250305175781, 0.2985779113769531, 0.2987386779785156, 0.2985850830078125, 0.29863629150390625, 0.29863729858398436, 0.29869158935546875, 0.29864141845703124, 0.2990602111816406, 0.2987724914550781, 0.2986608581542969, 0.2989906005859375, 0.2988503112792969, 0.29879705810546875, 0.29870489501953124, 0.298724365234375, 0.29857073974609377, 0.29866494750976563, 0.29862405395507813, 0.29871405029296877, 0.2985922546386719, 0.29864346313476564, 0.29879296875, 0.299109375, 0.2994646911621094, 0.2987386779785156, 0.29884722900390626, 0.29866598510742187, 0.29875711059570315, 0.2986332092285156, 0.29882571411132813, 0.2988114013671875, 0.29872332763671877, 0.2986680297851563, 0.2987653198242187, 0.29900082397460936, 0.2987509765625, 0.29862399291992187, 0.2986475524902344, 0.29878170776367186, 0.2986905517578125, 0.29885336303710935, 0.298756103515625, 0.29877044677734377, 0.2987724914550781, 0.29862911987304686, 0.2986526794433594, 0.298903564453125, 0.298745849609375, 0.29878988647460936, 0.2988226623535156, 0.2988164978027344, 0.2988072814941406, 0.29877862548828127, 0.29900390625, 0.29874072265625, 0.29883392333984377, 0.2988072814941406, 0.298956787109375, 0.6386237182617187, 0.29877658081054687, 0.2986854248046875, 0.29864651489257815, 0.2986045532226563, 0.2988031921386719, 0.2985267333984375, 0.2986618957519531, 0.29867315673828126, 0.29864651489257815, 0.29855230712890624, 0.29865472412109373, 0.2986639404296875, 0.2986680297851563, 0.2987386779785156, 0.2985472106933594, 0.2987100219726562, 0.2987182006835937, 0.2986219482421875, 0.29865164184570314, 0.29867724609375, 0.2986342468261719, 0.2985973815917969, 0.2986598510742188, 0.29906228637695315, 0.2986956787109375, 0.29896600341796875, 0.29861785888671877, 0.2990868530273437, 0.298724365234375, 0.2988769226074219, 0.2987591552734375, 0.29865164184570314, 0.298787841796875, 0.29884005737304686, 0.298682373046875, 0.29856768798828126, 0.2986956787109375, 0.29869158935546875, 0.2987182006835937, 0.2987448425292969, 0.2987745361328125, 0.2987950134277344, 0.29871923828125, 0.29873663330078126, 0.29879193115234376, 0.29870489501953124, 0.2987960205078125, 0.2988062744140625, 0.29881549072265623, 0.29877658081054687, 0.2988441467285156, 0.2987806701660156, 0.2988707885742187, 0.29862399291992187, 0.2986117248535156, 0.2987694091796875, 0.29873663330078126, 0.29883905029296876, 0.29911654663085935, 0.29874176025390625, 0.29878375244140626, 0.29891787719726565, 0.638328857421875, 0.2986250305175781, 0.2985748596191406, 0.29873663330078126, 0.29864346313476564, 0.29865472412109373, 0.29872845458984376, 0.2986977233886719, 0.29866598510742187, 0.29880935668945313, 0.29860147094726563, 0.29870693969726564, 0.29879193115234376, 0.2987386779785156, 0.29876019287109373, 0.2987950134277344, 0.29864141845703124, 0.29870693969726564, 0.29863629150390625, 0.29877044677734377, 0.29883905029296876, 0.29858917236328125, 0.2987734985351562, 0.29867623901367185, 0.29884518432617185, 0.2988001403808594, 0.298745849609375, 0.29894451904296876, 0.29879910278320315, 0.298703857421875, 0.29870489501953124, 0.29902334594726565, 0.2988717956542969, 0.298745849609375, 0.29904281616210937, 0.29879193115234376, 0.2988031921386719, 0.2987653198242187, 0.2987325439453125, 0.2988206176757813, 0.29881036376953124, 0.29879092407226565, 0.298650634765625, 0.29869158935546875, 0.29877862548828127, 0.29889739990234376, 0.29879705810546875, 0.29879193115234376, 0.29879296875, 0.29889022827148437, 0.29869876098632814, 0.2989056091308594, 0.2989189147949219, 0.2988441467285156, 0.29881857299804687, 0.29889022827148437, 0.29874789428710935, 0.29886257934570315, 0.2987694091796875, 0.298851318359375, 0.2988492736816406, 0.29875711059570315, 0.2988687438964844]",tokens/s,3.293574659040466,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-1.3B,EleutherAI/gpt-neo-1.3B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13949,7 +13949,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-7b,huggyllama/llama-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,4110.757888,5287.44448,0.0,4640.997376,4330.668544,s,10,3.1969297485351564,0.31969297485351567,0.001348702205900169,0.31914218139648437,0.32145095520019534,0.32196441497802736,0.322375182800293,"[0.3186290283203125, 0.3224778747558594, 0.31885498046875, 0.31913238525390625, 0.3191519775390625, 0.31846505737304687, 0.3200827941894531, 0.3207144470214844, 0.32133685302734377, 0.3180843505859375]",tokens/s,800.7683000144123,kWh,3.758179541263316e-06,2.0593209790774835e-06,1.8989615886123856e-05,2.4807116406464656e-05,tokens/kWh,10319619.411036715,MB,4110.757888,5287.44448,0.0,4640.997376,4515.667456,s,10,175.15052539062503,17.515052539062502,0.0018447663069760805,17.515354492187498,17.517081640624998,17.517545703125,17.517916953125,"[17.516314453125, 17.513849609375, 17.516134765625, 17.511609375, 17.515056640625, 17.513658203125, 17.516978515625, 17.51565234375, 17.518009765625, 17.51326171875]",tokens/s,3.596906138848048,kWh,0.00020681544852339562,0.0001133520352390417,0.0010167293237994686,0.001336896807561906,tokens/kWh,47124.055980725156,,s,629,177.750416656494,0.2825920773553166,0.03806230539059325,0.278002685546875,0.2781634521484375,0.27826237182617186,0.5973247680664062,"[0.27860992431640624, 0.27796890258789064, 0.2778388366699219, 0.2779924621582031, 0.278066162109375, 0.27786444091796875, 0.2777835388183594, 0.27789825439453125, 0.2779504699707031, 0.27782144165039063, 0.2779309997558594, 0.27789825439453125, 0.27791461181640625, 0.27783782958984377, 0.27793408203125, 0.27778662109375, 0.2780508117675781, 0.27777227783203123, 0.2779862976074219, 0.27807537841796875, 0.2779053955078125, 0.27794329833984377, 0.2780487670898438, 0.2779258728027344, 0.2780487670898438, 0.27790643310546875, 0.27796786499023435, 0.2779258728027344, 0.2781552734375, 0.2780569458007813, 0.2780641174316406, 0.27803851318359374, 0.2781112365722656, 0.2779862976074219, 0.27808154296875, 0.27809381103515624, 0.27819314575195314, 0.2780559387207031, 0.2779658203125, 0.27812966918945314, 0.2781440124511719, 0.27802215576171874, 0.2781583251953125, 0.2780712890625, 0.2780262451171875, 0.27805184936523436, 0.2781255798339844, 0.27800167846679685, 0.278181884765625, 0.2779955139160156, 0.2780979309082031, 0.2779791259765625, 0.2781634521484375, 0.27801702880859375, 0.27805902099609375, 0.27803033447265624, 0.2780702819824219, 0.2781051025390625, 0.2779955139160156, 0.27840716552734374, 0.27825457763671874, 0.27805902099609375, 0.6008422241210938, 0.278034423828125, 0.27797402954101563, 0.2779924621582031, 0.27792486572265623, 0.27854949951171876, 0.27780300903320315, 0.27780300903320315, 0.27782962036132813, 0.27787673950195313, 0.27778866577148437, 0.2779053955078125, 0.27800677490234377, 0.27789926147460936, 0.2778347473144531, 0.2777978820800781, 0.2779156494140625, 0.27812353515625, 0.27781939697265623, 0.27812249755859375, 0.2779033508300781, 0.27796173095703125, 0.2778306579589844, 0.2780006408691406, 0.2777835388183594, 0.2780692443847656, 0.277818359375, 0.2778603515625, 0.27797503662109374, 0.27809588623046877, 0.27791360473632815, 0.27798016357421873, 0.27808050537109374, 0.2780610656738281, 0.27797708129882814, 0.2779371643066406, 0.27804672241210937, 0.27804672241210937, 0.2779084777832031, 0.2779504699707031, 0.2779473876953125, 0.27842559814453127, 0.27789825439453125, 0.27808050537109374, 0.27795660400390626, 0.2780354614257812, 0.27798016357421873, 0.2781337585449219, 0.278034423828125, 0.2782637939453125, 0.27796685791015624, 0.27807745361328123, 0.2779402160644531, 0.27826995849609376, 0.27799859619140627, 0.27812863159179685, 0.2779176940917969, 0.2780999755859375, 0.2779873352050781, 0.2780979309082031, 0.27799755859375, 0.27807949829101564, 0.2780159912109375, 0.5972192993164063, 0.2780426330566406, 0.27779583740234376, 0.2778480529785156, 0.2779473876953125, 0.2780057678222656, 0.2779320373535156, 0.2778316650390625, 0.2777558898925781, 0.27795559692382815, 0.2778880004882813, 0.27797607421875, 0.2779033508300781, 0.2783354797363281, 0.2778040466308594, 0.27798220825195313, 0.2778470458984375, 0.27807537841796875, 0.2782801818847656, 0.2779371643066406, 0.2779781188964844, 0.2780334167480469, 0.2780231628417969, 0.2780211181640625, 0.27808154296875, 0.2785587158203125, 0.2779914245605469, 0.27797607421875, 0.2779873352050781, 0.27805902099609375, 0.2780835876464844, 0.2781051025390625, 0.27797708129882814, 0.2781767578125, 0.27785830688476565, 0.277865478515625, 0.27793304443359373, 0.2779924621582031, 0.27794638061523436, 0.2780702819824219, 0.2779484558105469, 0.27794427490234375, 0.2778828735351562, 0.27815115356445314, 0.27796173095703125, 0.2780702819824219, 0.2779371643066406, 0.2781368408203125, 0.2780641174316406, 0.2786897888183594, 0.277907470703125, 0.27796685791015624, 0.2779893798828125, 0.2781440124511719, 0.2782853088378906, 0.27813479614257813, 0.2779914245605469, 0.27810302734375, 0.27808255004882815, 0.2780641174316406, 0.2779453430175781, 0.2784460754394531, 0.27808563232421873, 0.5973657836914062, 0.27796890258789064, 0.27786138916015624, 0.27774566650390625, 0.27787982177734377, 0.2779381713867187, 0.2778122253417969, 0.2779596862792969, 0.2777999267578125, 0.27791668701171873, 0.2776084594726563, 0.2777763977050781, 0.27786444091796875, 0.277960693359375, 0.2777794494628906, 0.2778050537109375, 0.27780914306640625, 0.27791873168945314, 0.2777989196777344, 0.278118408203125, 0.2779361267089844, 0.27799655151367186, 0.2777415771484375, 0.27795559692382815, 0.277928955078125, 0.27811328125, 0.2781317138671875, 0.27800473022460936, 0.278029296875, 0.2780559387207031, 0.27793408203125, 0.2779279479980469, 0.2780426330566406, 0.2780579833984375, 0.2779381713867187, 0.27791668701171873, 0.27802215576171874, 0.2780968933105469, 0.2780088195800781, 0.2779258728027344, 0.2781665344238281, 0.2780159912109375, 0.2779873352050781, 0.2779945068359375, 0.2779811706542969, 0.27802215576171874, 0.27794125366210937, 0.27793304443359373, 0.27801190185546876, 0.27817166137695315, 0.2779422607421875, 0.2779924621582031, 0.2779320373535156, 0.27813784790039064, 0.2779095153808594, 0.2779996032714844, 0.27801089477539065, 0.2781603698730469, 0.27780609130859374, 0.278012939453125, 0.27797402954101563, 0.2780262451171875, 0.27819622802734373, 0.5975726318359375, 0.2780139465332031, 0.2778470458984375, 0.27806719970703125, 0.27787161254882814, 0.27796890258789064, 0.2779207763671875, 0.2778746948242187, 0.27775079345703124, 0.27796273803710936, 0.27774053955078126, 0.27780096435546875, 0.27782861328125, 0.27800372314453126, 0.2778880004882813, 0.2778480529785156, 0.27787982177734377, 0.27799038696289063, 0.2778757019042969, 0.278012939453125, 0.27793304443359373, 0.2780631103515625, 0.27790234375, 0.2780334167480469, 0.27802520751953125, 0.2781419372558594, 0.27788595581054687, 0.27802725219726565, 0.27800677490234377, 0.2781614074707031, 0.2779781188964844, 0.2780211181640625, 0.27803237915039064, 0.2782033996582031, 0.27809381103515624, 0.2779648132324219, 0.27801702880859375, 0.27804568481445313, 0.27800372314453126, 0.2782637939453125, 0.27806002807617186, 0.27815936279296877, 0.27799655151367186, 0.2780784606933594, 0.2779934692382813, 0.2781634521484375, 0.27805389404296876, 0.2781481018066406, 0.2780784606933594, 0.27816754150390627, 0.278044677734375, 0.27812762451171874, 0.27803033447265624, 0.27824127197265625, 0.27803851318359374, 0.2779924621582031, 0.2779473876953125, 0.2782156677246094, 0.27799859619140627, 0.277923828125, 0.2781163635253906, 0.27806207275390626, 0.2780477294921875, 0.596780029296875, 0.27795660400390626, 0.27787673950195313, 0.27799655151367186, 0.27799655151367186, 0.2779637756347656, 0.27791256713867185, 0.2778070983886719, 0.2779002990722656, 0.2779228210449219, 0.27778561401367186, 0.2780426330566406, 0.27786956787109374, 0.2779729919433594, 0.2781552734375, 0.277907470703125, 0.27780096435546875, 0.2779596862792969, 0.2777630615234375, 0.27805184936523436, 0.27790643310546875, 0.2779781188964844, 0.27776614379882814, 0.2779596862792969, 0.27791256713867185, 0.2780692443847656, 0.2778828735351562, 0.2778746948242187, 0.27790234375, 0.27818801879882815, 0.2778746948242187, 0.2779228210449219, 0.2780006408691406, 0.27844403076171875, 0.27789108276367186, 0.27794638061523436, 0.2780211181640625, 0.27810714721679686, 0.27806515502929685, 0.2781481018066406, 0.2779986267089844, 0.27808969116210935, 0.2780999755859375, 0.27813479614257813, 0.2778173522949219, 0.2779811706542969, 0.277970947265625, 0.2781951904296875, 0.2780405883789063, 0.27821978759765625, 0.2779084777832031, 0.2779719543457031, 0.27793408203125, 0.27815936279296877, 0.2778623962402344, 0.27813479614257813, 0.2779402160644531, 0.27821261596679686, 0.27808154296875, 0.2780405883789063, 0.2779207763671875, 0.2780631103515625, 0.2781051025390625, 0.5984942016601562, 0.2780979309082031, 0.277855224609375, 0.27779483032226565, 0.27799038696289063, 0.278066162109375, 0.27786956787109374, 0.2779422607421875, 0.27826174926757813, 0.2779832458496094, 0.277796875, 0.27789413452148437, 0.27803955078125, 0.27804672241210937, 0.2779361267089844, 0.2780231628417969, 0.27792486572265623, 0.27805184936523436, 0.2779658203125, 0.2780712890625, 0.2781317138671875, 0.2780928039550781, 0.2779443054199219, 0.2780149841308594, 0.2779504699707031, 0.2780702819824219, 0.2779525146484375, 0.27796990966796875, 0.27806719970703125, 0.278034423828125, 0.2780405883789063, 0.278002685546875, 0.27802215576171874, 0.27820645141601563, 0.2779361267089844, 0.2779596862792969, 0.2780201110839844, 0.2780692443847656, 0.277960693359375, 0.2780354614257812, 0.2779791259765625, 0.2780498046875, 0.2780415954589844, 0.278066162109375, 0.27822695922851565, 0.2780313720703125, 0.27807232666015624, 0.27818701171875, 0.2780415954589844, 0.2782801818847656, 0.27808767700195314, 0.2781061096191406, 0.2780712890625, 0.278044677734375, 0.27798834228515623, 0.27814093017578123, 0.27845428466796873, 0.27822592163085935, 0.2780159912109375, 0.2780559387207031, 0.2779934692382813, 0.2782310485839844, 0.2781583251953125, 0.5990051879882813, 0.27789312744140626, 0.2778961791992188, 0.2781265869140625, 0.27796786499023435, 0.2780559387207031, 0.27769650268554685, 0.27773849487304686, 0.2780559387207031, 0.27800372314453126, 0.2776801147460938, 0.27787060546875, 0.27782144165039063, 0.27800473022460936, 0.27786444091796875, 0.2779299926757812, 0.2782156677246094, 0.27800167846679685, 0.2778316650390625, 0.27791668701171873, 0.2779422607421875, 0.2783897705078125, 0.2778204040527344, 0.2779105224609375, 0.27800372314453126, 0.2780231628417969, 0.277970947265625, 0.277939208984375, 0.27832012939453127, 0.2780641174316406, 0.27799755859375, 0.27813067626953125, 0.27803851318359374, 0.278150146484375, 0.27794638061523436, 0.2779176940917969, 0.27802215576171874, 0.2780631103515625, 0.27795455932617186, 0.27806207275390626, 0.2780682373046875, 0.27805490112304687, 0.2779576416015625, 0.2780088195800781, 0.277907470703125, 0.278139892578125, 0.27790130615234376, 0.27797607421875, 0.2780682373046875, 0.27826278686523437, 0.277939208984375, 0.27811431884765625, 0.2783999938964844, 0.27817166137695315, 0.2779648132324219, 0.27796685791015624, 0.27807537841796875, 0.27856692504882813, 0.27819827270507813, 0.2779945068359375, 0.278118408203125, 0.2781051025390625, 0.2781685791015625, 0.5997485961914063, 0.27802215576171874, 0.2779996032714844, 0.27782962036132813, 0.27801089477539065, 0.2780426330566406, 0.2778408813476563, 0.2778839111328125, 0.277833740234375, 0.27784600830078127, 0.27785726928710935, 0.27782962036132813, 0.2780631103515625, 0.27799038696289063, 0.2778828735351562, 0.27795147705078127, 0.2778327026367188, 0.27798016357421873, 0.2779279479980469, 0.2780078125, 0.2779719543457031, 0.2780078125, 0.2779309997558594, 0.278034423828125, 0.27794943237304687, 0.278086669921875, 0.277823486328125, 0.2779084777832031, 0.2780201110839844, 0.27817062377929686, 0.27802520751953125, 0.27811328125, 0.2780989379882812, 0.27811532592773436, 0.27788595581054687, 0.27801803588867186, 0.27810202026367187, 0.27812249755859375, 0.27801702880859375, 0.27796173095703125, 0.27799038696289063, 0.27811737060546876, 0.27811019897460937, 0.27808154296875, 0.2779719543457031, 0.2780282897949219, 0.2780211181640625, 0.278066162109375, 0.27813479614257813, 0.2783088684082031, 0.27800677490234377, 0.2780928039550781, 0.2781051025390625, 0.27819314575195314, 0.27791668701171873, 0.278002685546875, 0.27808255004882815, 0.27803955078125, 0.2780610656738281, 0.2780631103515625, 0.2780364685058594, 0.2781470642089844, 0.2780149841308594, 0.5994915771484375, 0.2779402160644531, 0.27785113525390626, 0.2778450012207031, 0.2779842529296875, 0.27790130615234376, 0.27784396362304686, 0.2778245239257813, 0.277923828125, 0.2779576416015625, 0.2778316650390625, 0.27778253173828127, 0.2778204040527344, 0.2779279479980469, 0.27787982177734377, 0.2778787841796875, 0.277796875, 0.277907470703125, 0.27791973876953124, 0.27798220825195313, 0.27790438842773435, 0.27808050537109374, 0.27791156005859374, 0.27804364013671873, 0.27786138916015624, 0.27808050537109374, 0.27792178344726565, 0.2780641174316406, 0.27806719970703125, 0.2780426330566406, 0.2778623962402344, 0.278029296875, 0.278066162109375, 0.2781890563964844, 0.27787982177734377, 0.27800677490234377, 0.2780415954589844, 0.2781695861816406, 0.2779719543457031, 0.2780477294921875, 0.27800985717773435, 0.2780211181640625, 0.2778736572265625, 0.27810714721679686, 0.27792486572265623, 0.27814093017578123, 0.2779535217285156, 0.2780579833984375, 0.2779596862792969, 0.2782637939453125, 0.27796786499023435, 0.27802215576171874, 0.2780498046875, 0.2780979309082031, 0.27786138916015624, 0.2779873352050781, 0.27800473022460936, 0.27812045288085935, 0.2780364685058594, 0.27805490112304687, 0.2780139465332031, 0.2782054443359375, 0.27814910888671873]",tokens/s,3.5386696235742376,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neox-20b,EleutherAI/gpt-neox-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14164,7 +14164,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: FalconForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14363,7 +14363,7 @@ OSError: M is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-12b,EleutherAI/pythia-12b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14783,7 +14783,7 @@ Traceback (most recent call last): OSError: Incorrect path_or_model_id: '-'. Please provide either the path to a local folder or the repo_id of a model on the Hub. ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/polyglot-ko-12.8b,EleutherAI/polyglot-ko-12.8b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14912,7 +14912,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,huggyllama/llama-30b,huggyllama/llama-30b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,6436.08576,20389.036032,0.0,19742.588928,18821.508608,s,10,15.157101196289062,1.5157101196289062,0.003170114166599588,1.5143783569335938,1.5197610107421875,1.5208548095703125,1.5217298486328126,"[1.5126845703125, 1.512470458984375, 1.5143804931640625, 1.5141666259765625, 1.514376220703125, 1.512878173828125, 1.515317626953125, 1.5193604736328126, 1.5195179443359375, 1.5219486083984375]",tokens/s,168.89773096103423,kWh,1.786541634135776e-05,9.79020476837144e-06,8.850873747360267e-05,0.00011616435858333187,tokens/kWh,2203774.058773418,MB,6438.002688,20389.036032,0.0,19742.588928,19408.2304,s,10,843.2092499999999,84.320925,0.002661197598122423,84.32021093750001,84.32392734375,84.325639453125,84.327009140625,"[84.323546875, 84.3200234375, 84.3196015625, 84.321078125, 84.3183359375, 84.3178828125, 84.3218359375, 84.3203984375, 84.3191953125, 84.3273515625]",tokens/s,0.7471455039185113,kWh,0.000995496139741606,0.0005456199585125352,0.004966448584266792,0.006507564682520933,tokens/kWh,9681.040922914766,,s,629,855.5100844726549,1.3601114220550974,0.18017856913180233,1.33838134765625,1.33879111328125,1.3389119140624999,2.8532972265625003,"[1.33838330078125, 1.338135498046875, 1.338103759765625, 1.337954345703125, 1.338292236328125, 1.338123291015625, 1.3379686279296874, 1.337951171875, 1.33804443359375, 1.3383045654296875, 1.3380516357421874, 1.337988037109375, 1.3382154541015625, 1.3380885009765624, 1.3382850341796875, 1.3384017333984375, 1.338282958984375, 1.33829833984375, 1.338177490234375, 1.3382943115234376, 1.338187744140625, 1.33867822265625, 1.338323974609375, 1.338177490234375, 1.3380833740234375, 1.33817236328125, 1.3384478759765626, 1.3384119873046876, 1.3385123291015626, 1.338522705078125, 1.338946533203125, 1.3383382568359374, 1.3386168212890626, 1.33869775390625, 1.3384273681640626, 1.3386065673828125, 1.3384437255859376, 1.33874072265625, 1.33851953125, 1.3385645751953126, 1.33836083984375, 1.3386669921875, 1.3389434814453125, 1.338682373046875, 1.3389168701171874, 1.3387059326171875, 1.338524658203125, 1.338112060546875, 1.3382738037109374, 1.338470458984375, 1.33846630859375, 1.338481689453125, 1.338217529296875, 1.3385523681640625, 1.3385206298828125, 1.3382369384765624, 1.337903076171875, 1.338323974609375, 1.338281982421875, 1.338577880859375, 1.33874072265625, 1.3389075927734375, 2.85612548828125, 1.3380362548828124, 1.3384171142578125, 1.338639404296875, 1.337964599609375, 1.33838134765625, 1.3382850341796875, 1.3384775390625, 1.3385257568359374, 1.3377781982421875, 1.338303466796875, 1.337987060546875, 1.3382584228515626, 1.338038330078125, 1.33821240234375, 1.3383648681640625, 1.338181640625, 1.33842431640625, 1.3387275390625, 1.3381058349609376, 1.3383424072265624, 1.338270751953125, 1.3384775390625, 1.3381263427734376, 1.338206298828125, 1.33825732421875, 1.3382103271484376, 1.3385584716796874, 1.338059814453125, 1.3389271240234375, 1.33803515625, 1.3390428466796875, 1.3384058837890624, 1.3389639892578125, 1.3388267822265625, 1.338429443359375, 1.33853076171875, 1.338356689453125, 1.338428466796875, 1.338724365234375, 1.3384376220703125, 1.338541015625, 1.3387283935546874, 1.3388216552734375, 1.3385850830078125, 1.3389854736328124, 1.3382000732421875, 1.3383629150390626, 1.3383331298828125, 1.3383916015625, 1.3385103759765624, 1.3383638916015625, 1.3384805908203126, 1.338123291015625, 1.3385267333984374, 1.3388809814453124, 1.33817041015625, 1.3378948974609375, 1.33848779296875, 1.33816015625, 1.33851953125, 1.33850927734375, 1.3387950439453125, 2.853568603515625, 1.338101806640625, 1.33815087890625, 1.3382379150390624, 1.33825439453125, 1.3381744384765626, 1.3384129638671876, 1.3381632080078125, 1.3381375732421874, 1.3378150634765624, 1.3382593994140626, 1.33808740234375, 1.337988037109375, 1.3381693115234374, 1.3381324462890625, 1.3380936279296876, 1.338429443359375, 1.338197021484375, 1.33809765625, 1.3381221923828126, 1.338328125, 1.3383106689453126, 1.3385123291015626, 1.338229736328125, 1.338282958984375, 1.3386322021484376, 1.338462158203125, 1.338744873046875, 1.338181640625, 1.3385010986328125, 1.3385123291015626, 1.3384263916015624, 1.3385533447265625, 1.3387633056640624, 1.3384591064453124, 1.338155029296875, 1.338788818359375, 1.338418212890625, 1.3386004638671876, 1.338249267578125, 1.3388902587890625, 1.3383546142578124, 1.3384744873046874, 1.3390633544921875, 1.338366943359375, 1.3388748779296875, 1.3383424072265624, 1.338250244140625, 1.3385625, 1.3382236328125, 1.338692626953125, 1.338902587890625, 1.33848779296875, 1.338251220703125, 1.33867626953125, 1.3384058837890624, 1.33861376953125, 1.3383055419921874, 1.338281982421875, 1.338312744140625, 1.338599365234375, 1.33861376953125, 1.3388585205078125, 2.85312109375, 1.3381170654296874, 1.338102783203125, 1.3384007568359375, 1.3381734619140624, 1.3384765625, 1.3382646484375, 1.337976806640625, 1.3380946044921875, 1.3380577392578126, 1.338217529296875, 1.3380413818359376, 1.3382645263671875, 1.3380198974609374, 1.338167236328125, 1.3382799072265625, 1.3384786376953124, 1.3382943115234376, 1.3386229248046875, 1.3382103271484376, 1.337924560546875, 1.3382840576171875, 1.33852978515625, 1.3382205810546874, 1.3383424072265624, 1.3382379150390624, 1.3384171142578125, 1.33878173828125, 1.3382154541015625, 1.3386751708984375, 1.3386956787109374, 1.338599365234375, 1.3385687255859375, 1.3383782958984376, 1.338692626953125, 1.3380567626953126, 1.338788818359375, 1.3382574462890624, 1.3384222412109374, 1.3383055419921874, 1.338651611328125, 1.3384990234375, 1.3384417724609374, 1.3387919921875, 1.338693603515625, 1.339273193359375, 1.3382420654296876, 1.3387283935546874, 1.3383782958984376, 1.3382728271484374, 1.3388123779296874, 1.3383935546875, 1.338482666015625, 1.3383321533203125, 1.33850732421875, 1.338587158203125, 1.3388779296875, 1.33827685546875, 1.3387684326171876, 1.3382061767578124, 1.3386629638671874, 1.33879296875, 1.3388123779296874, 2.85336572265625, 1.3381539306640624, 1.3380556640625, 1.338523681640625, 1.33804443359375, 1.3385030517578125, 1.338166259765625, 1.3381856689453124, 1.337996337890625, 1.3377669677734374, 1.3383956298828126, 1.33794921875, 1.3381846923828125, 1.337997314453125, 1.3380955810546875, 1.3382861328125, 1.3384765625, 1.3382677001953125, 1.3384693603515625, 1.338175537109375, 1.3383792724609376, 1.3383802490234376, 1.3384591064453124, 1.3381478271484375, 1.33817041015625, 1.33816015625, 1.33810888671875, 1.3385738525390625, 1.3384058837890624, 1.338608642578125, 1.338224609375, 1.338683349609375, 1.33831982421875, 1.338618896484375, 1.3387069091796875, 1.3382994384765625, 1.33880419921875, 1.338345458984375, 1.3385123291015626, 1.33853076171875, 1.338577880859375, 1.338240966796875, 1.33871826171875, 1.3390069580078126, 1.33830859375, 1.3391697998046874, 1.3380157470703125, 1.33829736328125, 1.338114013671875, 1.338156005859375, 1.3383414306640624, 1.3386527099609375, 1.3385123291015626, 1.3385001220703125, 1.338565673828125, 1.338566650390625, 1.338281005859375, 1.33831884765625, 1.3384908447265624, 1.3383905029296874, 1.3385390625, 1.3388421630859375, 1.3386884765625, 2.851640380859375, 1.338324951171875, 1.3379962158203125, 1.3380853271484374, 1.3381488037109375, 1.3381253662109376, 1.3382379150390624, 1.338017822265625, 1.338619873046875, 1.337976806640625, 1.33848681640625, 1.337903076171875, 1.338588134765625, 1.3379849853515624, 1.3383536376953125, 1.3382103271484376, 1.33810888671875, 1.3383853759765625, 1.3383260498046874, 1.3380946044921875, 1.338181640625, 1.3383946533203126, 1.3384130859375, 1.3385594482421874, 1.3382113037109375, 1.338124267578125, 1.338328125, 1.3386270751953124, 1.3382451171875, 1.338799072265625, 1.3383802490234376, 1.3385257568359374, 1.338417236328125, 1.33832080078125, 1.3384765625, 1.3382215576171874, 1.3384263916015624, 1.338396728515625, 1.33853076171875, 1.3386475830078124, 1.3387509765625, 1.3384683837890625, 1.3387899169921875, 1.3387919921875, 1.3385533447265625, 1.3391268310546875, 1.3381468505859375, 1.338397705078125, 1.3383485107421875, 1.33824609375, 1.33825634765625, 1.338388427734375, 1.3384888916015625, 1.33812841796875, 1.3388359375, 1.338244140625, 1.3383956298828126, 1.338101806640625, 1.3383690185546875, 1.3383741455078124, 1.3385155029296876, 1.3386322021484376, 1.33859228515625, 2.854215576171875, 1.338356689453125, 1.3385379638671875, 1.3381529541015624, 1.33810791015625, 1.3381365966796874, 1.3386229248046875, 1.3378641357421874, 1.338134521484375, 1.3378743896484375, 1.3383321533203125, 1.3382052001953124, 1.3383260498046874, 1.338207275390625, 1.3384837646484375, 1.338671142578125, 1.338228759765625, 1.3382061767578124, 1.338640380859375, 1.338203125, 1.3383311767578125, 1.3386884765625, 1.338472412109375, 1.33840380859375, 1.3382052001953124, 1.33817138671875, 1.3381181640625, 1.338577880859375, 1.3382061767578124, 1.338861572265625, 1.3389066162109375, 1.338660888671875, 1.3386341552734375, 1.3385523681640625, 1.3385430908203124, 1.338102783203125, 1.3386177978515625, 1.3383690185546875, 1.338588134765625, 1.3384847412109375, 1.3383792724609376, 1.3389332275390624, 1.3387847900390626, 1.3388912353515625, 1.338240966796875, 1.3393377685546874, 1.3380147705078125, 1.338346435546875, 1.3382327880859375, 1.33857177734375, 1.338472412109375, 1.33872021484375, 1.338575927734375, 1.33836181640625, 1.3385318603515626, 1.338082275390625, 1.3387274169921874, 1.3383782958984376, 1.3383505859375, 1.338428466796875, 1.3386341552734375, 1.3385738525390625, 1.3386802978515624, 2.85828515625, 1.33825439453125, 1.338598388671875, 1.338102783203125, 1.3382010498046875, 1.3382891845703124, 1.338208251953125, 1.337956298828125, 1.3384530029296875, 1.3379942626953125, 1.338576904296875, 1.3380638427734375, 1.338123291015625, 1.3382757568359376, 1.33821240234375, 1.338217529296875, 1.33817041015625, 1.3384171142578125, 1.338460205078125, 1.338281005859375, 1.33812841796875, 1.3382943115234376, 1.3385491943359376, 1.338522705078125, 1.338440673828125, 1.338386474609375, 1.338228759765625, 1.3383526611328125, 1.3381888427734374, 1.338619873046875, 1.3384171142578125, 1.3388001708984374, 1.33842431640625, 1.3388319091796874, 1.3386240234375, 1.3381939697265626, 1.33859228515625, 1.3382901611328124, 1.338334228515625, 1.3385902099609375, 1.33840380859375, 1.3385789794921874, 1.338714111328125, 1.3387325439453126, 1.338462158203125, 1.3388338623046876, 1.3383720703125, 1.33831884765625, 1.338462158203125, 1.33825537109375, 1.3387796630859374, 1.3383209228515625, 1.3384263916015624, 1.3384837646484375, 1.338556396484375, 1.338260498046875, 1.338534912109375, 1.3380771484375, 1.338418212890625, 1.338197021484375, 1.3385902099609375, 1.3388994140625, 1.33888818359375, 2.8586865234375, 1.338609619140625, 1.338398681640625, 1.338145751953125, 1.338203125, 1.338523681640625, 1.3383199462890625, 1.33819482421875, 1.33796142578125, 1.3378519287109376, 1.338119140625, 1.338113037109375, 1.338135498046875, 1.3380516357421874, 1.33864453125, 1.33827587890625, 1.338181640625, 1.3382677001953125, 1.3383045654296875, 1.337975830078125, 1.338377197265625, 1.3383536376953125, 1.3386004638671876, 1.33821435546875, 1.338138671875, 1.3382799072265625, 1.3388739013671875, 1.33859228515625, 1.33829833984375, 1.3384765625, 1.338504150390625, 1.3385277099609374, 1.338335205078125, 1.33824609375, 1.3385943603515624, 1.33834033203125, 1.338545166015625, 1.3384049072265625, 1.3381488037109375, 1.338460205078125, 1.3383106689453126, 1.3387652587890626, 1.3386905517578125, 1.338962890625, 1.338398681640625, 1.338982421875, 1.3382379150390624, 1.3384539794921875, 1.3381795654296875, 1.3384007568359375, 1.33836083984375, 1.338472412109375, 1.338291259765625, 1.338218505859375, 1.338640380859375, 1.338345458984375, 1.3385645751953126, 1.3381519775390625, 1.3383629150390626, 1.3388759765625, 1.3386044921875, 1.33867724609375, 1.3387591552734375, 2.86118408203125, 1.3389434814453125, 1.33874072265625, 1.338156005859375, 1.338240966796875, 1.3385543212890625, 1.3379993896484375, 1.3379625244140625, 1.33812841796875, 1.33773828125, 1.3383751220703124, 1.3379461669921875, 1.3382154541015625, 1.338399658203125, 1.338345458984375, 1.3383638916015625, 1.338197998046875, 1.3385594482421874, 1.3383792724609376, 1.33813037109375, 1.3387049560546875, 1.338354736328125, 1.3387908935546875, 1.338270751953125, 1.338091552734375, 1.33819287109375, 1.3384530029296875, 1.338455078125, 1.338661865234375, 1.338914794921875, 1.33852978515625, 1.33894140625, 1.3384273681640626, 1.33825634765625, 1.3386331787109376, 1.3382471923828125, 1.338392578125, 1.338302490234375, 1.3384324951171875, 1.33825537109375, 1.3385830078125, 1.3388687744140626, 1.3387059326171875, 1.338957763671875, 1.33882470703125, 1.33918408203125, 1.3382420654296876, 1.3383424072265624, 1.3381488037109375, 1.3381571044921876, 1.338545166015625, 1.338265625, 1.33840283203125, 1.338265625, 1.3387376708984375, 1.338366943359375, 1.3383372802734375, 1.338070068359375, 1.3385430908203124, 1.3384478759765626, 1.3390029296875, 1.3386383056640625, 1.3390069580078126]",tokens/s,0.7352338814190846,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-125m,EleutherAI/gpt-neo-125m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15040,7 +15040,7 @@ OSError: google/recurrentgemma-7b is not a local folder and is not a valid model If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,EleutherAI/pythia-410m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15385,7 +15385,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-72B,Qwen/Qwen-72B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15488,7 +15488,7 @@ ChildProcessError: Traceback (most recent call last): torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 96.00 MiB. GPU ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/gpt-neo-2.7B,EleutherAI/gpt-neo-2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15790,7 +15790,7 @@ OSError: 0 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.30.1,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,EleutherAI/pythia-6.7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch