diff --git "a/perf-df-gptq-1xA10.csv" "b/perf-df-gptq-1xA10.csv" --- "a/perf-df-gptq-1xA10.csv" +++ "b/perf-df-gptq-1xA10.csv" @@ -1,5 +1,5 @@ config.name,config.backend.name,config.backend.version,config.backend._target_,config.backend.task,config.backend.library,config.backend.model,config.backend.processor,config.backend.device,config.backend.device_ids,config.backend.seed,config.backend.inter_op_num_threads,config.backend.intra_op_num_threads,config.backend.model_kwargs.trust_remote_code,config.backend.processor_kwargs.trust_remote_code,config.backend.hub_kwargs.trust_remote_code,config.backend.no_weights,config.backend.device_map,config.backend.torch_dtype,config.backend.eval_mode,config.backend.to_bettertransformer,config.backend.low_cpu_mem_usage,config.backend.attn_implementation,config.backend.cache_implementation,config.backend.autocast_enabled,config.backend.autocast_dtype,config.backend.torch_compile,config.backend.torch_compile_target,config.backend.quantization_scheme,config.backend.quantization_config.bits,config.backend.quantization_config.use_exllama ,config.backend.quantization_config.version,config.backend.quantization_config.model_seqlen,config.backend.deepspeed_inference,config.backend.peft_type,config.scenario.name,config.scenario._target_,config.scenario.iterations,config.scenario.duration,config.scenario.warmup_runs,config.scenario.input_shapes.batch_size,config.scenario.input_shapes.num_choices,config.scenario.input_shapes.sequence_length,config.scenario.new_tokens,config.scenario.latency,config.scenario.memory,config.scenario.energy,config.scenario.generate_kwargs.max_new_tokens,config.scenario.generate_kwargs.min_new_tokens,config.launcher.name,config.launcher._target_,config.launcher.device_isolation,config.launcher.device_isolation_action,config.launcher.numactl,config.launcher.start_method,config.environment.cpu,config.environment.cpu_count,config.environment.cpu_ram_mb,config.environment.system,config.environment.machine,config.environment.platform,config.environment.processor,config.environment.python_version,config.environment.gpu,config.environment.gpu_count,config.environment.gpu_vram_mb,config.environment.optimum_benchmark_version,config.environment.optimum_benchmark_commit,config.environment.transformers_version,config.environment.transformers_commit,config.environment.accelerate_version,config.environment.accelerate_commit,config.environment.diffusers_version,config.environment.diffusers_commit,config.environment.optimum_version,config.environment.optimum_commit,config.environment.timm_version,config.environment.timm_commit,config.environment.peft_version,config.environment.peft_commit,report.traceback,config.backend.hub_kwargs.revision,config.backend.hub_kwargs.force_download,config.backend.hub_kwargs.local_files_only,report.prefill.memory.unit,report.prefill.memory.max_ram,report.prefill.memory.max_global_vram,report.prefill.memory.max_process_vram,report.prefill.memory.max_reserved,report.prefill.memory.max_allocated,report.prefill.latency.unit,report.prefill.latency.count,report.prefill.latency.total,report.prefill.latency.mean,report.prefill.latency.stdev,report.prefill.latency.p50,report.prefill.latency.p90,report.prefill.latency.p95,report.prefill.latency.p99,report.prefill.latency.values,report.prefill.throughput.unit,report.prefill.throughput.value,report.prefill.energy.unit,report.prefill.energy.cpu,report.prefill.energy.ram,report.prefill.energy.gpu,report.prefill.energy.total,report.prefill.efficiency.unit,report.prefill.efficiency.value,report.decode.memory.unit,report.decode.memory.max_ram,report.decode.memory.max_global_vram,report.decode.memory.max_process_vram,report.decode.memory.max_reserved,report.decode.memory.max_allocated,report.decode.latency.unit,report.decode.latency.count,report.decode.latency.total,report.decode.latency.mean,report.decode.latency.stdev,report.decode.latency.p50,report.decode.latency.p90,report.decode.latency.p95,report.decode.latency.p99,report.decode.latency.values,report.decode.throughput.unit,report.decode.throughput.value,report.decode.energy.unit,report.decode.energy.cpu,report.decode.energy.ram,report.decode.energy.gpu,report.decode.energy.total,report.decode.efficiency.unit,report.decode.efficiency.value,report.per_token.memory,report.per_token.latency.unit,report.per_token.latency.count,report.per_token.latency.total,report.per_token.latency.mean,report.per_token.latency.stdev,report.per_token.latency.p50,report.per_token.latency.p90,report.per_token.latency.p95,report.per_token.latency.p99,report.per_token.latency.values,report.per_token.throughput.unit,report.per_token.throughput.value,report.per_token.energy,report.per_token.efficiency -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -67,7 +67,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e96e0-031d72e41a863ba549f483a8;7453247c-4b5a-453a-bcd2-6782e969a678) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe8b5-763db71f294b5ab2360c3ea7;3470eca0-a7a9-41ba-8e0b-3c536cbff4d1) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -99,7 +99,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1674.15808,2173.17376,0.0,1526.726656,1394.863104,s,10,0.6633986434936523,0.06633986434936523,0.0002834213305971893,0.06622074890136719,0.06651487426757811,0.06681279602050781,0.06705113342285156,"[0.0671107177734375, 0.06620326232910156, 0.06610300445556641, 0.06609532928466796, 0.06620972442626953, 0.06618291473388672, 0.06644866943359375, 0.06623177337646484, 0.06639087677001954, 0.06642237091064453]",tokens/s,3858.916543028016,kWh,7.835787422872801e-07,4.2936535715154407e-07,3.713461396079291e-06,4.926405495518115e-06,tokens/kWh,51964865.70845631,MB,1674.15808,2173.17376,0.0,1526.726656,1491.575808,s,10,36.005891601562496,3.6005891601562494,0.002651652094642963,3.5998477783203127,3.6018425781250003,3.6050617919921875,3.6076371630859376,"[3.608281005859375, 3.5999453125, 3.598923828125, 3.599460205078125, 3.601127197265625, 3.59905322265625, 3.60047607421875, 3.6000146484375, 3.599750244140625, 3.59885986328125]",tokens/s,17.497136495646753,kWh,4.243640493478207e-05,2.3257412183964045e-05,0.0001990910619585201,0.0002647848790772662,tokens/kWh,237928.99435777872,,s,629,36.54402151870729,0.058098603368374045,0.007920234388562994,0.05711872100830078,0.05725793228149414,0.05743477630615234,0.12371951416015625,"[0.057945152282714844, 0.05798495864868164, 0.0580055046081543, 0.05788979339599609, 0.057850879669189455, 0.057829376220703124, 0.057862144470214844, 0.057864192962646485, 0.057673728942871094, 0.05731020736694336, 0.05710335922241211, 0.05719347381591797, 0.05710233688354492, 0.057106433868408205, 0.0570860481262207, 0.0570530891418457, 0.05704908752441406, 0.057076736450195314, 0.05712300872802734, 0.057064254760742186, 0.05702041625976562, 0.05703577423095703, 0.057115646362304685, 0.05706547164916992, 0.05710745620727539, 0.05711974334716797, 0.0571495361328125, 0.05713209533691406, 0.05724041748046875, 0.05711155319213867, 0.05710233688354492, 0.05710847854614258, 0.05713817596435547, 0.057109504699707034, 0.057128959655761716, 0.05708492660522461, 0.05728460693359375, 0.05741158294677735, 0.05719039916992188, 0.057150463104248046, 0.05721913528442383, 0.05724051284790039, 0.057250816345214846, 0.05715660858154297, 0.05710131072998047, 0.05713715362548828, 0.057148414611816405, 0.05711769485473633, 0.05705625534057617, 0.05710233688354492, 0.05716889572143555, 0.0571146240234375, 0.05715660858154297, 0.05712384033203125, 0.05728665542602539, 0.057501697540283205, 0.05726105499267578, 0.05715865707397461, 0.0571228141784668, 0.057123870849609376, 0.057141216278076175, 0.05713612747192383, 0.12372582244873047, 0.057003009796142576, 0.0569804801940918, 0.057055233001708984, 0.05701836776733398, 0.0569989128112793, 0.05700723266601562, 0.05702028656005859, 0.0569989128112793, 0.056995838165283204, 0.057055233001708984, 0.05706752014160156, 0.05704819107055664, 0.057043838500976565, 0.05712793731689453, 0.05704601669311524, 0.05708390426635742, 0.057033729553222653, 0.05710438537597656, 0.05716582489013672, 0.05708697509765625, 0.05704806518554688, 0.05710233688354492, 0.05702143859863281, 0.05708185577392578, 0.05744537734985351, 0.05713100814819336, 0.057090049743652345, 0.057082878112792966, 0.057150463104248046, 0.057082878112792966, 0.0571228141784668, 0.05711155319213867, 0.057132030487060545, 0.057106433868408205, 0.057106433868408205, 0.05711872100830078, 0.05711155319213867, 0.05713407897949219, 0.05718220901489258, 0.05731532669067383, 0.05795532989501953, 0.057283584594726565, 0.05720371246337891, 0.05717708969116211, 0.05717299270629883, 0.05711974334716797, 0.05710847854614258, 0.05734502410888672, 0.05723852920532226, 0.05714022445678711, 0.05715251159667969, 0.05715577697753906, 0.05716153717041016, 0.05712691116333008, 0.05717520141601563, 0.05716463851928711, 0.05722326278686524, 0.057159584045410154, 0.05724774551391602, 0.0571514892578125, 0.057145343780517575, 0.057210880279541014, 0.12386406707763672, 0.05759385681152344, 0.0571412467956543, 0.05713919830322266, 0.057073665618896485, 0.05700505447387695, 0.05704601669311524, 0.057079807281494144, 0.05709414291381836, 0.057032703399658206, 0.057040897369384766, 0.057082878112792966, 0.057025535583496094, 0.05706854248046875, 0.05706854248046875, 0.057082878112792966, 0.05710335922241211, 0.05707059097290039, 0.057062496185302736, 0.05714323043823242, 0.05711868667602539, 0.05705215835571289, 0.057106433868408205, 0.05709107208251953, 0.0570880012512207, 0.05703987121582031, 0.057116703033447264, 0.05707772827148438, 0.0571146240234375, 0.05711360168457031, 0.05710847854614258, 0.05707468795776367, 0.05710847854614258, 0.05711667251586914, 0.05715353775024414, 0.057100318908691404, 0.05713916778564453, 0.05713407897949219, 0.05711052703857422, 0.05710335922241211, 0.05710131072998047, 0.057215999603271485, 0.0570748176574707, 0.05712371063232422, 0.0571228141784668, 0.05708390426635742, 0.057079807281494144, 0.057355262756347655, 0.057128959655761716, 0.05716787338256836, 0.057128959655761716, 0.05713817596435547, 0.05713926315307617, 0.057100223541259765, 0.05719670486450195, 0.057151329040527346, 0.05712998580932617, 0.057078784942626956, 0.05714329528808594, 0.05711360168457031, 0.0571228141784668, 0.057112575531005856, 0.057183231353759766, 0.12367574310302734, 0.056995742797851565, 0.05700198364257812, 0.05703475189208984, 0.05727849578857422, 0.057269214630126956, 0.057047073364257815, 0.05706441497802734, 0.057073665618896485, 0.05708083343505859, 0.057063423156738284, 0.05713100814819336, 0.05715967941284179, 0.05700505447387695, 0.057012222290039063, 0.057041919708251954, 0.05705625534057617, 0.05701836776733398, 0.05703987121582031, 0.057034816741943356, 0.05715244674682617, 0.05731737518310547, 0.0571228141784668, 0.0570511360168457, 0.05713715362548828, 0.0571956787109375, 0.0572127685546875, 0.05714739227294922, 0.05711769485473633, 0.05712998580932617, 0.05721299362182617, 0.05710124969482422, 0.0571146240234375, 0.05709209442138672, 0.05710335922241211, 0.05717913436889648, 0.0571228141784668, 0.05706649780273437, 0.05721916961669922, 0.05742480087280273, 0.057145343780517575, 0.05712080001831055, 0.057209823608398436, 0.05713715362548828, 0.057196544647216796, 0.0571412467956543, 0.057204734802246096, 0.05710131072998047, 0.0570880012512207, 0.05712384033203125, 0.05712486267089844, 0.05714944076538086, 0.057112575531005856, 0.05715865707397461, 0.05710969543457031, 0.05712876892089844, 0.05718527984619141, 0.05717196655273438, 0.0571514892578125, 0.05713407897949219, 0.05715967941284179, 0.05710028839111328, 0.05712899017333985, 0.12372886657714843, 0.05697945785522461, 0.057043968200683595, 0.057011199951171876, 0.05699379348754883, 0.05702963256835938, 0.05703987121582031, 0.05699686431884766, 0.056997886657714845, 0.057017345428466794, 0.05705526351928711, 0.057045982360839846, 0.057024513244628906, 0.05703168106079102, 0.05710438537597656, 0.05736959838867187, 0.057125022888183594, 0.05710943984985352, 0.05705001449584961, 0.0570511360168457, 0.05708083343505859, 0.057106433868408205, 0.0570777587890625, 0.0571412467956543, 0.05782236862182617, 0.057167713165283206, 0.057115646362304685, 0.05717708969116211, 0.05708492660522461, 0.057109504699707034, 0.05706547164916992, 0.05717094421386719, 0.05709107208251953, 0.057148414611816405, 0.05710745620727539, 0.057166847229003906, 0.05711360168457031, 0.05718425750732422, 0.05723648071289063, 0.057148414611816405, 0.05711872100830078, 0.05720787048339844, 0.05713708877563477, 0.05714432144165039, 0.05713411331176758, 0.0571893424987793, 0.05716787338256836, 0.05747609710693359, 0.05769830322265625, 0.05727849578857422, 0.05716886520385742, 0.05756313705444336, 0.05713715362548828, 0.057181182861328124, 0.05716585540771484, 0.057141216278076175, 0.057130016326904294, 0.05735625457763672, 0.057161727905273435, 0.05717401504516602, 0.057186302185058595, 0.05720576095581055, 0.057133056640625, 0.12370329284667969, 0.057022464752197265, 0.05708390426635742, 0.057109504699707034, 0.05707571029663086, 0.05704601669311524, 0.05707263946533203, 0.05702860641479492, 0.05706649780273437, 0.05703680038452148, 0.05714031982421875, 0.05703158569335937, 0.0570296630859375, 0.057093086242675783, 0.0570860481262207, 0.0570428466796875, 0.057093120574951174, 0.05715456008911133, 0.057109504699707034, 0.057038848876953124, 0.057106433868408205, 0.05713100814819336, 0.0570849609375, 0.05713507080078125, 0.05710847854614258, 0.05710131072998047, 0.05709619140625, 0.05708803176879883, 0.05723542404174805, 0.05714227294921875, 0.05707171249389648, 0.05706332778930664, 0.057166847229003906, 0.05710745620727539, 0.057085983276367186, 0.0571176643371582, 0.05715353775024414, 0.057128959655761716, 0.057161727905273435, 0.05716889572143555, 0.05708595275878906, 0.05713407897949219, 0.057151519775390625, 0.05714838409423828, 0.05710131072998047, 0.05712076950073242, 0.05712588882446289, 0.05708390426635742, 0.05710956954956055, 0.057257919311523436, 0.05713817596435547, 0.057111648559570315, 0.057140129089355465, 0.05717708969116211, 0.05717606353759765, 0.057207809448242185, 0.057199615478515625, 0.05715353775024414, 0.05714739227294922, 0.0571514892578125, 0.057159839630126955, 0.05725167846679687, 0.057390079498291016, 0.12398899078369141, 0.057022464752197265, 0.05720678329467774, 0.05741056060791016, 0.05723244857788086, 0.05702649688720703, 0.05707059097290039, 0.05711155319213867, 0.057188350677490236, 0.05710444641113281, 0.0570398063659668, 0.057169921875, 0.05716889572143555, 0.05702668762207031, 0.057043838500976565, 0.057090049743652345, 0.057041919708251954, 0.057085983276367186, 0.05702652740478516, 0.05705830383300781, 0.057245697021484375, 0.057621505737304686, 0.05722828674316406, 0.057294849395751954, 0.05706444931030273, 0.05710847854614258, 0.05715763092041016, 0.0570880012512207, 0.057071617126464844, 0.05710438537597656, 0.057063423156738284, 0.05707263946533203, 0.05708083343505859, 0.05711667251586914, 0.05706854248046875, 0.057051200866699216, 0.057162689208984374, 0.05719244766235351, 0.057093120574951174, 0.05715558242797852, 0.057145343780517575, 0.05710649490356445, 0.05712480163574219, 0.05714438247680664, 0.05721097564697265, 0.05712368011474609, 0.0571228141784668, 0.05718220901489258, 0.05710847854614258, 0.05710540771484375, 0.05713100814819336, 0.05725798416137695, 0.057396224975585934, 0.057283584594726565, 0.057161727905273435, 0.05711478424072266, 0.057111392974853514, 0.057166847229003906, 0.05711052703857422, 0.05714329528808594, 0.057194496154785154, 0.057161727905273435, 0.057212928771972656, 0.12372697448730469, 0.05699059295654297, 0.05705215835571289, 0.057063423156738284, 0.05701545715332031, 0.05705302429199219, 0.05720371246337891, 0.05743513488769531, 0.05712588882446289, 0.057027584075927736, 0.0570777587890625, 0.05704908752441406, 0.05708083343505859, 0.057078784942626956, 0.05707468795776367, 0.05732659149169922, 0.05708697509765625, 0.05708083343505859, 0.05706444931030273, 0.05707263946533203, 0.05718425750732422, 0.057059326171875, 0.05710745620727539, 0.057071617126464844, 0.05712179183959961, 0.057032703399658206, 0.05708697509765625, 0.05711155319213867, 0.057133056640625, 0.05724364852905273, 0.05734105682373047, 0.05738790512084961, 0.05716793441772461, 0.057132991790771484, 0.05713612747192383, 0.05710233688354492, 0.05710745620727539, 0.05714128112792969, 0.05710742568969727, 0.057118751525878905, 0.05711151885986328, 0.05716582489013672, 0.05710540771484375, 0.05714432144165039, 0.05711772918701172, 0.05720470428466797, 0.05712076950073242, 0.05711974334716797, 0.05743820953369141, 0.05712899017333985, 0.05714019012451172, 0.05718732833862305, 0.05710438537597656, 0.05714438247680664, 0.057135040283203126, 0.057219070434570314, 0.05716070556640625, 0.057175041198730465, 0.05717401504516602, 0.05718527984619141, 0.05712998580932617, 0.05715763092041016, 0.05714022445678711, 0.12395622253417969, 0.05701836776733398, 0.057207809448242185, 0.05751193618774414, 0.05706547164916992, 0.0569989128112793, 0.05712588882446289, 0.057063423156738284, 0.05708492660522461, 0.057040897369384766, 0.05708697509765625, 0.05706854248046875, 0.057059326171875, 0.057076736450195314, 0.05708697509765625, 0.057095169067382816, 0.05706854248046875, 0.0571514892578125, 0.05708083343505859, 0.057102401733398436, 0.05703878402709961, 0.0571412467956543, 0.0570695686340332, 0.057043968200683595, 0.057038848876953124, 0.057037887573242185, 0.05709817504882812, 0.05703475189208984, 0.057188350677490236, 0.057099262237548826, 0.05706547164916992, 0.05705830383300781, 0.057078784942626956, 0.05714739227294922, 0.05712998580932617, 0.05710550308227539, 0.05721488189697266, 0.0571514892578125, 0.05715865707397461, 0.05713510513305664, 0.057166847229003906, 0.05712899017333985, 0.05711459350585937, 0.057202686309814454, 0.05711667251586914, 0.05711872100830078, 0.05710028839111328, 0.05714022445678711, 0.057175041198730465, 0.05734297561645508, 0.05758771133422851, 0.05724364852905273, 0.05717401504516602, 0.05717401504516602, 0.05718732833862305, 0.0571228141784668, 0.057178112030029295, 0.05714944076538086, 0.05714944076538086, 0.05716070556640625, 0.05714636611938476, 0.05710335922241211, 0.057178112030029295, 0.12406784057617187, 0.057033729553222653, 0.057090049743652345, 0.057212928771972656, 0.057355262756347655, 0.0573040657043457, 0.05709619140625, 0.05710847854614258, 0.05710540771484375, 0.05705318450927734, 0.057087039947509764, 0.057307071685791015, 0.057095169067382816, 0.057003009796142576, 0.057043968200683595, 0.057073665618896485, 0.057063423156738284, 0.05705830383300781, 0.05702041625976562, 0.05706444931030273, 0.05709619140625, 0.05713100814819336, 0.05702361679077148, 0.057046913146972654, 0.057022464752197265, 0.057073665618896485, 0.05743423843383789, 0.05708582305908203, 0.05706649780273437, 0.05704499053955078, 0.05712691116333008, 0.05705420684814453, 0.057057281494140626, 0.05709107208251953, 0.05710438537597656, 0.05707263946533203, 0.057076736450195314, 0.057436161041259766, 0.05707689666748047, 0.057113441467285156, 0.05710233688354492, 0.0571228141784668, 0.057112575531005856, 0.057071617126464844, 0.05707571029663086, 0.05714432144165039, 0.05711052703857422, 0.05705830383300781, 0.05711360168457031, 0.05709107208251953, 0.05712489700317383, 0.05710128021240234, 0.057132030487060545, 0.05710438537597656, 0.05719039916992188, 0.057338878631591796, 0.05713715362548828, 0.057150463104248046, 0.05713612747192383, 0.05713919830322266, 0.05708492660522461, 0.0570880012512207, 0.05712588882446289]",tokens/s,17.212117710635876,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -138,7 +138,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d460b-46efe85357291070747b3d2b;56f72314-bc17-4100-b3e6-4351676e2cd2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe92e-17475c804f76b0ac19270810;5d98dde6-361d-47bb-8027-cc801427a92a) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -195,7 +195,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpxwiirte3/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpbfif100k/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1243.045888,849.870848,0.0,203.423744,184.496128,s,17,0.16700380897521971,0.009823753469130572,0.00032406108228421964,0.009798175811767578,0.009877977752685547,0.010117497825622558,0.010848511009216308,"[0.011031264305114747, 0.009838335990905762, 0.009837311744689941, 0.009870592117309571, 0.009787679672241211, 0.00969315242767334, 0.009745439529418946, 0.009798175811767578, 0.00945907211303711, 0.009712384223937989, 0.009529888153076172, 0.009813471794128418, 0.009889056205749511, 0.009819840431213379, 0.009827679634094238, 0.009773152351379395, 0.009577312469482423]",tokens/s,26059.285873208773,kWh,1.1371295855728884e-07,6.228751583782483e-08,2.1097974316073842e-07,3.8698021755585207e-07,tokens/kWh,661532523.8506593,MB,1243.045888,849.870848,0.0,203.423744,193.844224,s,17,9.859052612304687,0.5799442713120404,0.009384098949348406,0.5792615356445312,0.582433935546875,0.589327099609375,0.609600576171875,"[0.6146689453125, 0.5820621337890625, 0.5795172729492187, 0.57863232421875, 0.5712827758789063, 0.5796158447265625, 0.5781971435546875, 0.5771149291992187, 0.5810087280273437, 0.5711693725585938, 0.5720592651367188, 0.5829916381835938, 0.5792615356445312, 0.5767430419921875, 0.573796875, 0.58035546875, 0.5805753173828125]",tokens/s,108.6311273624129,kWh,6.535953210618106e-06,3.5814351743626475e-06,1.157450808028629e-05,2.1691896465267043e-05,tokens/kWh,2904310.3769592154,,s,1070,10.00479437446596,0.00935027511632331,0.0012516390282944286,0.009197567939758301,0.009368793487548828,0.009731482028961182,0.019336427612304686,"[0.010523648262023925, 0.010109951972961426, 0.009951231956481933, 0.009835552215576172, 0.00968393611907959, 0.01023692798614502, 0.009938943862915038, 0.009607168197631836, 0.009637887954711915, 0.00972697639465332, 0.01003929615020752, 0.009821184158325195, 0.009737215995788574, 0.009621503829956055, 0.010085375785827636, 0.009829376220703125, 0.009803775787353516, 0.009574399948120118, 0.009894911766052245, 0.009881600379943848, 0.009778176307678223, 0.009654272079467773, 0.009691136360168457, 0.009679871559143067, 0.009843711853027343, 0.00973516845703125, 0.009602047920227052, 0.009592831611633301, 0.009702400207519531, 0.009350144386291503, 0.009743359565734864, 0.009999456405639649, 0.00969206428527832, 0.009797632217407226, 0.009861120223999023, 0.00974233627319336, 0.009634816169738769, 0.009597951889038087, 0.00927948760986328, 0.009200639724731445, 0.009762816429138184, 0.00970956802368164, 0.009640959739685059, 0.009963520050048828, 0.010007552146911621, 0.009977919578552246, 0.0097074556350708, 0.009759743690490723, 0.009867263793945312, 0.009690112113952636, 0.009538559913635255, 0.009515007972717286, 0.00961945629119873, 0.00930406379699707, 0.009401344299316406, 0.009560064315795898, 0.009651200294494629, 0.009743359565734864, 0.009855999946594238, 0.00983142375946045, 0.009650176048278808, 0.009390080451965332, 0.019545087814331053, 0.009326592445373535, 0.009299967765808105, 0.009189375877380371, 0.00922316837310791, 0.00914739227294922, 0.009115648269653321, 0.009276479721069335, 0.009284543991088866, 0.009153535842895508, 0.009074687957763672, 0.009202688217163087, 0.009183232307434081, 0.009252863883972168, 0.009224224090576171, 0.009331680297851563, 0.009191424369812011, 0.009270272254943847, 0.009176159858703613, 0.009543583869934083, 0.009376768112182618, 0.00923750400543213, 0.009227264404296874, 0.009970687866210937, 0.009368703842163087, 0.009185152053833008, 0.009369600296020507, 0.009630720138549804, 0.009368576049804688, 0.009257984161376954, 0.009250847816467286, 0.009198559761047363, 0.00923136043548584, 0.009145343780517578, 0.009137311935424804, 0.009149439811706543, 0.009200480461120606, 0.009233407974243164, 0.009220095634460449, 0.009228287696838379, 0.009299967765808105, 0.009211903572082519, 0.009084927558898925, 0.00913920021057129, 0.009142271995544434, 0.009117695808410644, 0.00911571216583252, 0.009382847785949708, 0.00935321617126465, 0.009257984161376954, 0.009208831787109375, 0.009076767921447754, 0.009084896087646484, 0.009037823677062988, 0.00918015956878662, 0.009111552238464356, 0.00919961643218994, 0.009134079933166504, 0.009264127731323242, 0.00928767967224121, 0.00912281608581543, 0.009064448356628419, 0.009074687957763672, 0.019499008178710937, 0.009242624282836913, 0.00923033618927002, 0.009169919967651367, 0.00912384033203125, 0.009058303833007812, 0.009085951805114746, 0.009171968460083007, 0.009233407974243164, 0.009173024177551269, 0.009084896087646484, 0.009076736450195312, 0.009096192359924317, 0.009189375877380371, 0.009201663970947266, 0.009220095634460449, 0.009248767852783203, 0.00922316837310791, 0.00923033618927002, 0.009216064453125, 0.00903878402709961, 0.009159680366516113, 0.00916380786895752, 0.009100255966186523, 0.009077759742736816, 0.00923033618927002, 0.009047039985656738, 0.009165823936462402, 0.00921497631072998, 0.00922111988067627, 0.00922214412689209, 0.00921292781829834, 0.009175040245056153, 0.009157631874084473, 0.009191424369812011, 0.0091146240234375, 0.009119744300842286, 0.009169919967651367, 0.009189375877380371, 0.0092293119430542, 0.009218048095703125, 0.009198592185974122, 0.00918835163116455, 0.009192447662353515, 0.00918835163116455, 0.009200639724731445, 0.009191424369812011, 0.009219072341918945, 0.009126912117004395, 0.009227264404296874, 0.009175040245056153, 0.009267200469970703, 0.009219072341918945, 0.009129983901977539, 0.009157631874084473, 0.009280511856079102, 0.009280511856079102, 0.009463808059692384, 0.009468928337097168, 0.009313280105590821, 0.009226240158081055, 0.009244671821594238, 0.009248767852783203, 0.019546112060546874, 0.009198592185974122, 0.00923033618927002, 0.009241600036621094, 0.009197567939758301, 0.009257984161376954, 0.00921395206451416, 0.00919961643218994, 0.009201663970947266, 0.009218048095703125, 0.009255935668945312, 0.009029631614685058, 0.009183232307434081, 0.009269280433654785, 0.009225184440612794, 0.009245696067810059, 0.009326720237731933, 0.009116543769836425, 0.009233407974243164, 0.009166848182678223, 0.00922111988067627, 0.00918015956878662, 0.009134079933166504, 0.009088000297546387, 0.009176063537597656, 0.009102335929870605, 0.009143296241760255, 0.009162752151489258, 0.009256959915161133, 0.00919654369354248, 0.009159680366516113, 0.009198592185974122, 0.009145343780517578, 0.009176063537597656, 0.009148415565490722, 0.00923136043548584, 0.00923750400543213, 0.009146368026733399, 0.009104384422302245, 0.009142271995544434, 0.00910540771484375, 0.00912384033203125, 0.00918015956878662, 0.009149439811706543, 0.009192511558532715, 0.009216959953308106, 0.009167872428894042, 0.009246720314025878, 0.009243647575378417, 0.009190400123596192, 0.00920576000213623, 0.009155584335327148, 0.009167872428894042, 0.009138175964355469, 0.009133055686950683, 0.009053183555603026, 0.00913920021057129, 0.009001055717468261, 0.009059231758117675, 0.009116671562194823, 0.00921292781829834, 0.009240575790405273, 0.009183232307434081, 0.01943142318725586, 0.009120767593383788, 0.009217023849487305, 0.009161727905273438, 0.009099264144897461, 0.009162752151489258, 0.00921497631072998, 0.009093119621276855, 0.009167872428894042, 0.009184255599975585, 0.009181183815002441, 0.009235456466674804, 0.009172991752624511, 0.009210880279541016, 0.00916592025756836, 0.009098143577575683, 0.009030655860900879, 0.009049087524414063, 0.009181183815002441, 0.009111552238464356, 0.009085951805114746, 0.009141247749328613, 0.009209856033325196, 0.009234432220458985, 0.009207807540893554, 0.009176063537597656, 0.009184255599975585, 0.009239551544189453, 0.009192447662353515, 0.009158656120300293, 0.009186304092407227, 0.009191424369812011, 0.009202688217163087, 0.009116671562194823, 0.009224191665649414, 0.009161727905273438, 0.009167872428894042, 0.009200639724731445, 0.009211903572082519, 0.0091627836227417, 0.009301983833312988, 0.00931942367553711, 0.009126912117004395, 0.008987648010253906, 0.008691712379455567, 0.008762368202209473, 0.008802304267883301, 0.00879923152923584, 0.008743935585021973, 0.00882585620880127, 0.008767487525939942, 0.008882176399230958, 0.008848383903503418, 0.008884223937988281, 0.008896512031555176, 0.0088340482711792, 0.008876031875610351, 0.008733695983886718, 0.008738816261291504, 0.008808575630187988, 0.008762240409851074, 0.008844287872314453, 0.008932352066040039, 0.019290111541748048, 0.009226240158081055, 0.009192543983459473, 0.009072735786437988, 0.009117504119873047, 0.009209856033325196, 0.009234432220458985, 0.009071616172790528, 0.009153535842895508, 0.009234432220458985, 0.009253888130187989, 0.009182208061218262, 0.009101311683654785, 0.009162752151489258, 0.009194496154785157, 0.009153535842895508, 0.009136128425598144, 0.009181183815002441, 0.00923750400543213, 0.009248767852783203, 0.009215999603271484, 0.009622528076171874, 0.00941977596282959, 0.009262080192565919, 0.009164799690246582, 0.009107456207275391, 0.009246720314025878, 0.009148415565490722, 0.009133055686950683, 0.009167872428894042, 0.009257984161376954, 0.009179136276245118, 0.009218048095703125, 0.009142271995544434, 0.009201663970947266, 0.009152511596679687, 0.009197567939758301, 0.009241600036621094, 0.009190400123596192, 0.00921497631072998, 0.009203712463378906, 0.009235456466674804, 0.009233407974243164, 0.009073792457580566, 0.009089920043945312, 0.009043968200683594, 0.009189375877380371, 0.00921497631072998, 0.009186304092407227, 0.009152511596679687, 0.009183232307434081, 0.00921395206451416, 0.009274368286132812, 0.009203712463378906, 0.009209856033325196, 0.009239583969116212, 0.009121760368347168, 0.009207807540893554, 0.009187328338623046, 0.00923033618927002, 0.009146368026733399, 0.00918015956878662, 0.00918835163116455, 0.019396608352661132, 0.009326592445373535, 0.009152511596679687, 0.009268223762512207, 0.009243647575378417, 0.009166848182678223, 0.008897536277770996, 0.008979455947875976, 0.009034751892089844, 0.009119744300842286, 0.009065471649169921, 0.009208831787109375, 0.009220095634460449, 0.009226240158081055, 0.00922111988067627, 0.009209856033325196, 0.00921292781829834, 0.009148447990417481, 0.009192416191101074, 0.009165823936462402, 0.009200639724731445, 0.009162752151489258, 0.009201663970947266, 0.009051136016845703, 0.009189375877380371, 0.009262080192565919, 0.009207903861999512, 0.00918825626373291, 0.009133055686950683, 0.00921497631072998, 0.009157631874084473, 0.009128959655761718, 0.009217087745666503, 0.009224127769470215, 0.009115839958190918, 0.009102144241333009, 0.009166848182678223, 0.00921395206451416, 0.009158656120300293, 0.00919974422454834, 0.009130047798156739, 0.009087807655334472, 0.009056256294250489, 0.009099264144897461, 0.009063424110412598, 0.009078783988952637, 0.009101311683654785, 0.009162752151489258, 0.009463808059692384, 0.009272319793701172, 0.009198592185974122, 0.009144319534301757, 0.009219072341918945, 0.009159680366516113, 0.009142271995544434, 0.009260031700134277, 0.009185279846191406, 0.00923136043548584, 0.009182208061218262, 0.009290752410888671, 0.009267200469970703, 0.009079808235168458, 0.009135104179382325, 0.019449855804443358, 0.009274368286132812, 0.009172991752624511, 0.009284607887268067, 0.009233407974243164, 0.009042943954467773, 0.009087072372436524, 0.00907356834411621, 0.0091146240234375, 0.009109503746032714, 0.009274368286132812, 0.009159680366516113, 0.00913920021057129, 0.009234432220458985, 0.009345184326171875, 0.009156448364257812, 0.009245696067810059, 0.009245696067810059, 0.0092293119430542, 0.009239551544189453, 0.00911673641204834, 0.009136063575744629, 0.009192447662353515, 0.009241600036621094, 0.009242624282836913, 0.009210880279541016, 0.009164799690246582, 0.009126912117004395, 0.009116671562194823, 0.00922111988067627, 0.009138175964355469, 0.009243647575378417, 0.009153535842895508, 0.009244671821594238, 0.009288703918457031, 0.009179136276245118, 0.009073663711547851, 0.009178208351135255, 0.009131936073303223, 0.009789440155029297, 0.010276864051818848, 0.009390080451965332, 0.009272319793701172, 0.00929792022705078, 0.009158656120300293, 0.009267264366149903, 0.009231295585632324, 0.009220095634460449, 0.009408512115478516, 0.00936355209350586, 0.00907049560546875, 0.008773632049560547, 0.008762368202209473, 0.00891596794128418, 0.008843263626098634, 0.008750080108642578, 0.008733695983886718, 0.008853504180908203, 0.008744959831237792, 0.008860671997070312, 0.00880947208404541, 0.008872960090637207, 0.008772607803344726, 0.019050495147705078, 0.00914739227294922, 0.009171968460083007, 0.00913920021057129, 0.009129983901977539, 0.009084927558898925, 0.009102399826049806, 0.009157567977905274, 0.009215999603271484, 0.009225215911865235, 0.009127936363220214, 0.009496576309204101, 0.00930303955078125, 0.009183232307434081, 0.00932249641418457, 0.0092293119430542, 0.00920787239074707, 0.009202688217163087, 0.009164735794067383, 0.00921497631072998, 0.009233407974243164, 0.00934502410888672, 0.009215999603271484, 0.009211903572082519, 0.009240575790405273, 0.009356287956237793, 0.009263104438781738, 0.009226240158081055, 0.009272319793701172, 0.009249792098999024, 0.009227264404296874, 0.009277440071105958, 0.009200672149658203, 0.009244640350341797, 0.009247743606567382, 0.00940339183807373, 0.00921395206451416, 0.009118720054626465, 0.009260031700134277, 0.009226240158081055, 0.00923033618927002, 0.009190400123596192, 0.009373696327209472, 0.009288703918457031, 0.009221152305603027, 0.009135071754455566, 0.009248767852783203, 0.009151679992675781, 0.00914412784576416, 0.00918015956878662, 0.009190400123596192, 0.009176063537597656, 0.009245696067810059, 0.00921292781829834, 0.009134079933166504, 0.009086976051330567, 0.009086112022399902, 0.009161567687988282, 0.009284607887268067, 0.00914739227294922, 0.009242624282836913, 0.009239551544189453, 0.009207807540893554, 0.019359743118286133, 0.009250816345214843, 0.009277440071105958, 0.009189375877380371, 0.009217023849487305, 0.009102335929870605, 0.009195520401000976, 0.009217023849487305, 0.009215999603271484, 0.009209856033325196, 0.009276415824890137, 0.009144319534301757, 0.00900710391998291, 0.008902655601501466, 0.00880025577545166, 0.008871935844421386, 0.008897536277770996, 0.008839167594909669, 0.008781824111938476, 0.008749055862426757, 0.00880947208404541, 0.008901632308959961, 0.008910847663879394, 0.00892518424987793, 0.008903679847717285, 0.008893440246582032, 0.008972288131713867, 0.00890777587890625, 0.00890675163269043, 0.008847519874572754, 0.008818528175354005, 0.008945664405822755, 0.008953856468200684, 0.008910847663879394, 0.008886272430419923, 0.008861760139465331, 0.008858559608459472, 0.00888319969177246, 0.009025535583496093, 0.00923033618927002, 0.009121888160705567, 0.009133983612060546, 0.009175040245056153, 0.00913920021057129, 0.00922316837310791, 0.009239551544189453, 0.009241600036621094, 0.009255935668945312, 0.009226240158081055, 0.009177087783813476, 0.009233407974243164, 0.009143296241760255, 0.00920473575592041, 0.00920473575592041, 0.009117695808410644, 0.009062399864196777, 0.009056415557861328, 0.009175999641418457, 0.009165727615356445, 0.009145343780517578, 0.009138175964355469, 0.009235456466674804, 0.009246784210205078, 0.01884774398803711, 0.00892204761505127, 0.008887295722961425, 0.00892416000366211, 0.008919039726257324, 0.008806400299072266, 0.008905728340148926, 0.00890880012512207, 0.008914943695068359, 0.008866815567016602, 0.008857600212097168, 0.00890880012512207, 0.008857600212097168, 0.008770560264587402, 0.008869888305664063, 0.008758272171020508, 0.008865792274475098, 0.008824831962585449, 0.009273344039916993, 0.009147520065307618, 0.009206656455993653, 0.009214015960693359, 0.00923539161682129, 0.0092293119430542, 0.009240575790405273, 0.009120767593383788, 0.009254912376403808, 0.009246720314025878, 0.009211903572082519, 0.009234432220458985, 0.00912281608581543, 0.009153535842895508, 0.009148415565490722, 0.009217023849487305, 0.009240575790405273, 0.009051136016845703, 0.009158656120300293, 0.009244671821594238, 0.009165823936462402, 0.009259103775024414, 0.00910428810119629, 0.009156607627868652, 0.009071616172790528, 0.009088000297546387, 0.009150464057922364, 0.009191424369812011, 0.009203712463378906, 0.009197567939758301, 0.009211903572082519, 0.009085951805114746, 0.009252863883972168, 0.009261055946350098, 0.009144319534301757, 0.0089169921875, 0.00882380771636963, 0.00890675163269043, 0.008846336364746094, 0.00879923152923584, 0.00942080020904541, 0.009156607627868652, 0.009150464057922364, 0.00923033618927002, 0.009256959915161133, 0.019325952529907226, 0.009145343780517578, 0.00921292781829834, 0.00920473575592041, 0.009232383728027344, 0.009175040245056153, 0.009247743606567382, 0.009246720314025878, 0.009155584335327148, 0.009269248008728028, 0.00930406379699707, 0.009315327644348144, 0.009176063537597656, 0.00922111988067627, 0.009227264404296874, 0.00912281608581543, 0.00921395206451416, 0.009234432220458985, 0.009240575790405273, 0.009179136276245118, 0.009125887870788574, 0.00921292781829834, 0.00922537612915039, 0.00909398365020752, 0.009247743606567382, 0.009356287956237793, 0.009209856033325196, 0.009286656379699706, 0.009254912376403808, 0.009262080192565919, 0.009192447662353515, 0.00913100814819336, 0.009253888130187989, 0.009264127731323242, 0.009240575790405273, 0.009261055946350098, 0.009332736015319825, 0.009190400123596192, 0.009409536361694336, 0.009316351890563965, 0.009192447662353515, 0.009083904266357423, 0.00921513557434082, 0.009099103927612305, 0.00910848045349121, 0.009153535842895508, 0.009072640419006347, 0.00912281608581543, 0.009566207885742188, 0.009355263710021973, 0.00941158390045166, 0.009475071907043458, 0.010164223670959472, 0.009293824195861817, 0.009284671783447265, 0.009354175567626953, 0.009158656120300293, 0.009213088035583496, 0.009136992454528809, 0.009243647575378417, 0.009215999603271484, 0.009174015998840332, 0.00939417552947998, 0.019589120864868165, 0.009201663970947266, 0.009107456207275391, 0.00919654369354248, 0.009144319534301757, 0.00919654369354248, 0.009233407974243164, 0.009263104438781738, 0.009289728164672852, 0.009229375839233398, 0.009210816383361816, 0.009252863883972168, 0.009246720314025878, 0.00920473575592041, 0.00913920021057129, 0.009211903572082519, 0.009214048385620116, 0.009109408378601073, 0.009092096328735352, 0.009267200469970703, 0.009323519706726074, 0.009115648269653321, 0.009198592185974122, 0.009293824195861817, 0.0091843843460083, 0.009113471984863281, 0.009152511596679687, 0.009136128425598144, 0.009265151977539063, 0.009224320411682129, 0.009140095710754394, 0.009064448356628419, 0.009190400123596192, 0.009260031700134277, 0.00921395206451416, 0.009155584335327148, 0.009053407669067383, 0.009170720100402832, 0.0092293119430542, 0.009143296241760255, 0.009264127731323242, 0.009234432220458985, 0.00914739227294922, 0.0092293119430542, 0.009291775703430176, 0.009253888130187989, 0.009151488304138184, 0.009143296241760255, 0.009112575531005859, 0.009107456207275391, 0.009168895721435547, 0.009296895980834961, 0.00920473575592041, 0.00914739227294922, 0.009156607627868652, 0.009156607627868652, 0.009170944213867188, 0.009110527992248535, 0.009191424369812011, 0.009135104179382325, 0.009087008476257324, 0.009227231979370117, 0.009193471908569336, 0.019509248733520508, 0.00922214412689209, 0.009254912376403808, 0.009165823936462402, 0.00908902359008789, 0.009220095634460449, 0.009126976013183593, 0.009373632431030274, 0.009495552062988282, 0.009178112030029297, 0.008806400299072266, 0.008794112205505371, 0.008841216087341308, 0.00890067195892334, 0.008871871948242188, 0.008905728340148926, 0.008902655601501466, 0.00890675163269043, 0.008870911598205567, 0.009075712203979493, 0.009273344039916993, 0.009257984161376954, 0.009076736450195312, 0.009226304054260254, 0.009189311981201171, 0.009143296241760255, 0.00908083152770996, 0.00911359977722168, 0.00919961643218994, 0.009136128425598144, 0.009226240158081055, 0.009202688217163087, 0.00930303955078125, 0.009268223762512207, 0.009245696067810059, 0.009154560089111329, 0.009047039985656738, 0.00915875244140625, 0.009263008117675782, 0.009268223762512207, 0.009255935668945312, 0.00932863998413086, 0.00923750400543213, 0.00943513584136963, 0.00921395206451416, 0.009183232307434081, 0.009143296241760255, 0.00923033618927002, 0.009234432220458985, 0.009183232307434081, 0.00921395206451416, 0.009129983901977539, 0.009178112030029297, 0.00922316837310791, 0.009220095634460449, 0.009103360176086426, 0.009099264144897461, 0.009206784248352052, 0.009102335929870605, 0.008966143608093263, 0.008952832221984864, 0.009235456466674804, 0.009166879653930664, 0.019551200866699217, 0.009203712463378906, 0.009207807540893554, 0.009195520401000976, 0.00920473575592041, 0.009081855773925781, 0.009138208389282226, 0.009206751823425292, 0.009069727897644044, 0.00909500789642334, 0.009075712203979493, 0.009251839637756347, 0.009143296241760255, 0.009044992446899413, 0.009095168113708496, 0.00909721565246582, 0.009273344039916993, 0.009202688217163087, 0.009220095634460449, 0.009200639724731445, 0.00919654369354248, 0.009187328338623046, 0.009194496154785157, 0.009220128059387206, 0.00885756778717041, 0.008859647750854491, 0.008879103660583497, 0.008896608352661133, 0.008827808380126954, 0.008890368461608887, 0.008872960090637207, 0.009155584335327148, 0.009268223762512207, 0.00923852825164795, 0.009135104179382325, 0.009187328338623046, 0.00908083152770996, 0.009074687957763672, 0.009110527992248535, 0.009248767852783203, 0.00918015956878662, 0.009179136276245118, 0.009178112030029297, 0.009066495895385742, 0.008925215721130371, 0.008839136123657226, 0.008696831703186036, 0.008950783729553222, 0.008897536277770996, 0.008856575965881347, 0.008797183990478515, 0.00890880012512207, 0.009142271995544434, 0.009152576446533204, 0.009179072380065918, 0.009263104438781738, 0.009179136276245118, 0.009220095634460449, 0.009257984161376954, 0.009208864212036132, 0.009212896347045899, 0.009203712463378906, 0.00923852825164795, 0.01962188720703125, 0.009197567939758301, 0.009276415824890137, 0.009210880279541016, 0.009266176223754884, 0.00922316837310791, 0.00920576000213623, 0.009263168334960938, 0.009200575828552247, 0.009222208023071289, 0.00924665641784668, 0.009208831787109375, 0.009190400123596192, 0.009347071647644043, 0.009125887870788574, 0.009094143867492676, 0.009088031768798828, 0.009106528282165528, 0.00907968044281006, 0.009120767593383788, 0.009215999603271484, 0.008967167854309082, 0.009009152412414552, 0.008977408409118653, 0.009110527992248535, 0.009375743865966797, 0.00921292781829834, 0.009167872428894042, 0.009269248008728028, 0.009095168113708496, 0.009136128425598144, 0.009124863624572753, 0.00910028839111328, 0.009101311683654785, 0.009169024467468262, 0.009172863960266113, 0.009210880279541016, 0.009251839637756347, 0.009948160171508789, 0.00923852825164795, 0.009289728164672852, 0.00926425552368164, 0.009149312019348144, 0.0092293119430542, 0.009250816345214843, 0.009380864143371583, 0.009245696067810059, 0.00917199993133545, 0.00909001636505127, 0.00918835163116455, 0.0091658878326416, 0.009192383766174317, 0.009256959915161133, 0.009270272254943847, 0.009243647575378417, 0.009161727905273438, 0.009264127731323242, 0.009207807540893554, 0.009226240158081055, 0.009208831787109375, 0.00914739227294922, 0.009169919967651367, 0.00922111988067627, 0.018906112670898437, 0.008923135757446288, 0.009217023849487305, 0.009284607887268067, 0.009157631874084473, 0.009200639724731445, 0.00923852825164795, 0.009200639724731445, 0.009148415565490722, 0.009266176223754884, 0.009198592185974122, 0.009218048095703125, 0.00910540771484375, 0.009266176223754884, 0.009264127731323242, 0.009206815719604492, 0.009223135948181152, 0.009200639724731445, 0.009989119529724122, 0.009961471557617188, 0.009261055946350098, 0.009271295547485351, 0.009243647575378417, 0.009207807540893554, 0.009073663711547851, 0.00886787223815918, 0.008962016105651855, 0.009142271995544434, 0.009162752151489258, 0.009256959915161133, 0.009211903572082519, 0.00918950366973877, 0.009131903648376466, 0.009225215911865235, 0.00922316837310791, 0.009207807540893554, 0.009172991752624511, 0.009079808235168458, 0.009226240158081055, 0.00923033618927002, 0.009140224456787109, 0.009226240158081055, 0.009263104438781738, 0.009155584335327148, 0.00922111988067627, 0.009246848106384277, 0.00925068759918213, 0.009235456466674804, 0.009288703918457031, 0.00912179183959961, 0.00919961643218994, 0.009259008407592773, 0.009135104179382325, 0.009215999603271484, 0.009168895721435547, 0.009155584335327148, 0.009294848442077636, 0.009154560089111329, 0.009138175964355469, 0.009104448318481445, 0.009225152015686034, 0.00921497631072998, 0.009239551544189453]",tokens/s,106.94872477647667,, @@ -350,7 +350,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e9102-722a1aff7a9b811b384b3d5a;506ea068-ad20-4bb2-ab00-bdc6ab8df4f9) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe2b3-719ba0561b78b68727c9c69f;ac78b6c1-0230-48e6-93b3-d7c6affca947) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -439,7 +439,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95ec-2455f19d47a0895441583118;eca943ab-445d-4bb4-9cc5-88276dbf66bf) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe8de-3d422ac9308d6d2b4146251a;42ae17d6-bbc2-435c-8a72-a378c525f46f) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -499,7 +499,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90d9-5b9a6074020ff5b8333fb3ff;d0a1da00-037e-4adf-b50e-51ccd92bfb39) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe286-3d9c0040514896443ccdf0f8;87073e3e-d30f-491b-8f27-043430182ed9) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -731,7 +731,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmplocd80u0/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpk8pb0kn3/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -828,7 +828,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e8f49-700530d145a395d35e1431af;6133d759-4244-4593-a1f3-5fc2a6cd3398) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe0ea-246e979c01f2029801c52138;67a81354-cb78-4097-928e-abc28f7f610d) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -913,7 +913,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9691-4c83ece21fecdff46d68fcd6;0d66b73e-f87d-452f-a0a1-2b63446bb361) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe85f-740a64bf709e4e4a4c5e1eef;ab77d85c-065c-4744-942c-7180f944cf65) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1029,7 +1029,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95c4-30d550e51efdbaee71ec7ffc;e136d0fa-072e-4951-8576-9e8795c2b640) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe792-370fd5271dd284fd6dac44b6;17ab6342-9be2-46d9-93a9-88883decb5e1) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1103,7 +1103,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1847.382016,3393.716224,0.0,2749.366272,2585.357312,s,10,0.823750862121582,0.0823750862121582,0.0014776796762060414,0.08214677047729493,0.0831559356689453,0.0847691680908203,0.08605975402832032,"[0.08638240051269531, 0.0818265609741211, 0.08206022644042969, 0.08117469024658203, 0.08168768310546876, 0.08057820892333985, 0.08223331451416016, 0.08279743957519531, 0.0825384979248047, 0.08247183990478515]",tokens/s,3107.7357459835416,kWh,9.516499386893379e-07,5.214546148738008e-07,4.1613655513126345e-06,5.634470104875773e-06,tokens/kWh,45434618.55951124,MB,1847.382016,3395.813376,0.0,2749.366272,2628.768768,s,10,45.41338720703124,4.541338720703125,0.018165998348059254,4.543606689453124,4.563778857421875,4.565016137695313,4.566005961914063,"[4.51978564453125, 4.5209658203125, 4.52827197265625, 4.56350390625, 4.56018798828125, 4.5468212890625, 4.56625341796875, 4.551375, 4.54039208984375, 4.515830078125]",tokens/s,13.872561346898575,kWh,5.326320814291636e-05,2.919147736644736e-05,0.00022564251273608083,0.0003080971982454445,tokens/kWh,204480.92471717735,,s,629,46.07871485900881,0.07325709834500603,0.009850363871964385,0.07204454040527344,0.07273390045166016,0.07303168182373046,0.1544689245605469,"[0.07243673706054687, 0.0722165756225586, 0.07173836517333984, 0.07148953247070312, 0.07192166137695312, 0.07207936096191406, 0.07273779296875, 0.0723599395751953, 0.07181619262695313, 0.07186022186279296, 0.07170764923095703, 0.07144550323486328, 0.07165235137939453, 0.07186022186279296, 0.07191142272949219, 0.07151513671875, 0.07160524749755859, 0.07166259002685547, 0.07287091064453124, 0.07176089477539062, 0.0714567642211914, 0.07153663635253907, 0.07149772644042969, 0.07140966033935547, 0.07148134613037109, 0.07159705352783204, 0.07194419097900391, 0.07158477020263672, 0.07170559692382812, 0.07132978820800781, 0.07170662689208984, 0.07147110748291016, 0.071552001953125, 0.07141478729248046, 0.07180697631835938, 0.07170150756835937, 0.07141990661621093, 0.07142912292480469, 0.07150080108642579, 0.07151513671875, 0.07165542602539063, 0.0717096939086914, 0.07222067260742188, 0.07158477020263672, 0.07184486389160157, 0.07146701049804688, 0.07157350158691406, 0.07190630340576172, 0.07302861022949218, 0.07173939514160156, 0.07169023895263672, 0.07154898834228515, 0.07212742614746094, 0.0715120620727539, 0.07172710418701173, 0.07148748779296875, 0.07151001739501953, 0.07134003448486329, 0.07155404663085937, 0.07149158477783203, 0.07145369720458984, 0.07155404663085937, 0.15449702453613282, 0.07144448089599609, 0.07186227416992187, 0.07158477020263672, 0.07216435241699219, 0.07135027313232421, 0.07165235137939453, 0.07149158477783203, 0.07154994964599609, 0.07159705352783204, 0.0714241943359375, 0.07140435028076172, 0.07174861145019532, 0.0725555191040039, 0.07297945404052734, 0.07174553680419922, 0.07139737701416016, 0.07143023681640626, 0.07148944091796874, 0.07140966033935547, 0.0728647689819336, 0.07160934448242187, 0.07147622680664062, 0.07147212982177735, 0.07145471954345703, 0.0713359375, 0.07237222290039062, 0.07177113342285156, 0.07253708648681641, 0.07189708709716797, 0.0723056640625, 0.07202098846435546, 0.07208038330078125, 0.07169229125976563, 0.07199641418457031, 0.07196774291992188, 0.07131443023681641, 0.0715335693359375, 0.07154994964599609, 0.0714803237915039, 0.07152639770507813, 0.07153459167480469, 0.07189094543457031, 0.07155404663085937, 0.07141171264648437, 0.07154176330566406, 0.07424921417236328, 0.07197798156738282, 0.07158377838134766, 0.07196975708007812, 0.07168409729003906, 0.07160524749755859, 0.07153049468994141, 0.07144755554199218, 0.07139532470703125, 0.0712806396484375, 0.07142400360107422, 0.07159398651123047, 0.07154688262939453, 0.07156735992431641, 0.07152333068847656, 0.07149263763427735, 0.07195951843261719, 0.15509400939941406, 0.07232307434082032, 0.07224524688720703, 0.07204454040527344, 0.07226675415039062, 0.07196057891845703, 0.07289667510986328, 0.0720239028930664, 0.07299593353271484, 0.07253084564208985, 0.07200870513916016, 0.07150694274902344, 0.07152639770507813, 0.07155404663085937, 0.07140863800048829, 0.07150182342529297, 0.0714229736328125, 0.07174569702148438, 0.0724744644165039, 0.07129190063476562, 0.07246540832519531, 0.0721612777709961, 0.07162163543701172, 0.07225856018066407, 0.07228518676757813, 0.0722165756225586, 0.07262515258789062, 0.07154278564453125, 0.07134719848632813, 0.07153254699707032, 0.07247666931152344, 0.07277875518798828, 0.07165337371826172, 0.07139942169189453, 0.07149874877929688, 0.0718071060180664, 0.0725533447265625, 0.07195648193359375, 0.07154790496826172, 0.07206604766845703, 0.07182854461669921, 0.071468994140625, 0.07151821136474609, 0.07143424224853516, 0.07143014526367188, 0.07146099090576172, 0.07209868621826172, 0.07252684783935547, 0.07136255645751953, 0.07155404663085937, 0.07135948944091797, 0.07153971099853515, 0.07153254699707032, 0.07149568176269532, 0.07168428802490234, 0.07149037170410157, 0.07146086120605469, 0.07137894439697266, 0.07127654266357422, 0.07141375732421874, 0.07227699279785156, 0.07252582550048828, 0.07163085174560548, 0.15379866027832031, 0.07165235137939453, 0.07263651275634765, 0.07239673614501953, 0.07263549041748046, 0.07231270599365235, 0.07247277069091797, 0.07208735656738281, 0.07200460815429688, 0.07240704345703125, 0.07304908752441407, 0.07273369598388672, 0.07322624206542969, 0.07284633636474609, 0.07248076629638672, 0.07288422393798828, 0.07182540893554687, 0.07239679718017578, 0.07256473541259766, 0.07270195007324219, 0.07253094482421875, 0.07234559631347656, 0.07266508483886719, 0.07261901092529296, 0.0730072021484375, 0.07291177368164062, 0.07205683135986328, 0.07246233367919921, 0.07198822021484375, 0.07254937744140624, 0.072521728515625, 0.07264256286621094, 0.07221965026855469, 0.07235788726806641, 0.07210495758056641, 0.07256781005859375, 0.07240499114990234, 0.073059326171875, 0.0725524444580078, 0.072195068359375, 0.0725196762084961, 0.07145471954345703, 0.07244390106201172, 0.07228006744384766, 0.07420928192138672, 0.072700927734375, 0.0726456298828125, 0.07191961669921874, 0.07188992309570312, 0.0720148468017578, 0.07125299072265626, 0.0723394546508789, 0.07194419097900391, 0.07213772583007813, 0.07230156707763671, 0.0723947525024414, 0.0721295394897461, 0.07237222290039062, 0.07256371307373047, 0.07262515258789062, 0.07254937744140624, 0.07255859375, 0.07262413024902344, 0.1548278045654297, 0.072353759765625, 0.0725002212524414, 0.07245209503173829, 0.07242854309082031, 0.07192063903808593, 0.07217766571044922, 0.07246953582763672, 0.07247049713134765, 0.0727531509399414, 0.0725770263671875, 0.07248998260498046, 0.07175382232666015, 0.07248995208740235, 0.07272748565673828, 0.07259442901611328, 0.07262310028076172, 0.07351808166503906, 0.07339826965332032, 0.07235794830322266, 0.07229228973388672, 0.07209574127197266, 0.07278079986572265, 0.0727224349975586, 0.07274598693847656, 0.07179264068603515, 0.07222579193115235, 0.07242342376708985, 0.07207017517089843, 0.0719615707397461, 0.07255654144287109, 0.07277670288085937, 0.07267021179199219, 0.07186227416992187, 0.07147929382324218, 0.07170457458496093, 0.07202098846435546, 0.07212032318115234, 0.07165548706054688, 0.07158470153808594, 0.07155404663085937, 0.07349555206298829, 0.07213772583007813, 0.07260262298583985, 0.07264870452880859, 0.07200153350830078, 0.07254528045654297, 0.071657470703125, 0.07168000030517578, 0.07270604705810547, 0.07262003326416015, 0.07252070617675781, 0.07246652984619141, 0.07231581115722656, 0.0721295394897461, 0.07235379028320313, 0.07264669036865234, 0.0744959716796875, 0.07266918182373047, 0.07270710754394531, 0.0722820816040039, 0.07173632049560547, 0.07150592041015626, 0.15439666748046876, 0.07259750366210938, 0.0715857925415039, 0.07145779418945312, 0.07145267486572265, 0.07161650848388672, 0.072521728515625, 0.07253606414794922, 0.0724295654296875, 0.07247872161865235, 0.07240396881103515, 0.07220531463623046, 0.07265996551513672, 0.07269376373291016, 0.07269171142578125, 0.07250125122070313, 0.07236608123779296, 0.07237222290039062, 0.07195340728759765, 0.07279923248291016, 0.07360511779785156, 0.07294566345214844, 0.07247769927978516, 0.07262515258789062, 0.07273471832275391, 0.07258316802978515, 0.07195452880859375, 0.0722000961303711, 0.07212854766845703, 0.07230051422119141, 0.07157759857177734, 0.07145779418945312, 0.0715448989868164, 0.071901123046875, 0.07262310028076172, 0.07254425811767579, 0.07270809936523437, 0.07179366302490234, 0.0733564453125, 0.07144534301757813, 0.07159603118896485, 0.0715857925415039, 0.0716390380859375, 0.07154994964599609, 0.07138201904296874, 0.07153868865966796, 0.07152639770507813, 0.07149362945556641, 0.07145577239990235, 0.0716349105834961, 0.07142707061767578, 0.07146189117431641, 0.07142707061767578, 0.07152230072021484, 0.07156735992431641, 0.07205580902099609, 0.0724869155883789, 0.07231078338623047, 0.07246233367919921, 0.07301737976074218, 0.07265891265869141, 0.07283404541015626, 0.07256678771972656, 0.15526809692382812, 0.07210291290283204, 0.07220326232910156, 0.07217664337158203, 0.07258521270751953, 0.07277362823486327, 0.07267737579345702, 0.07268045043945312, 0.07258627319335938, 0.07265583801269532, 0.07280332946777343, 0.07242342376708985, 0.0724245147705078, 0.07229945373535156, 0.07221759796142578, 0.07265382385253906, 0.07256678771972656, 0.0719617919921875, 0.07145760345458985, 0.07243059539794922, 0.07212134552001953, 0.07256781005859375, 0.07259961700439453, 0.07362655639648437, 0.07254118347167969, 0.07281970977783203, 0.07263948822021485, 0.07236198425292968, 0.0738150405883789, 0.07245311737060547, 0.07157247924804687, 0.07246131134033203, 0.0724726104736328, 0.07284937286376954, 0.07257804870605469, 0.07242444610595702, 0.07258726501464843, 0.07241436767578124, 0.07216009521484375, 0.07275212860107422, 0.07261695861816406, 0.07253094482421875, 0.07270502471923829, 0.07222271728515625, 0.07260467529296875, 0.07262310028076172, 0.07250534057617188, 0.0722001953125, 0.07243775939941406, 0.07218994903564453, 0.07228006744384766, 0.07340134429931641, 0.07345664215087891, 0.072163330078125, 0.0721233901977539, 0.07228928375244141, 0.07184178924560547, 0.07190630340576172, 0.07204761505126953, 0.07256269073486328, 0.07192269134521484, 0.07259750366210938, 0.07253606414794922, 0.155936767578125, 0.07303372955322265, 0.07266304016113281, 0.07238143920898438, 0.07211436462402344, 0.07191020965576173, 0.0720711669921875, 0.07248486328125, 0.07192988586425782, 0.07246336364746093, 0.07251964569091797, 0.07230873870849609, 0.0725791015625, 0.07243260955810547, 0.07255763244628906, 0.07227590179443359, 0.07335628509521484, 0.07273471832275391, 0.07248281860351563, 0.07255142211914062, 0.07281868743896484, 0.07529267120361328, 0.07326207733154297, 0.07283302307128907, 0.0724869155883789, 0.07157862091064453, 0.07152639770507813, 0.07212850952148438, 0.07244290924072265, 0.07218582153320313, 0.07301427459716797, 0.07195954895019531, 0.07162982177734376, 0.0713338851928711, 0.0723927001953125, 0.0724510726928711, 0.07208857727050781, 0.07149158477783203, 0.07150694274902344, 0.07147007751464844, 0.07175475311279297, 0.07147929382324218, 0.07141273498535156, 0.07240191650390625, 0.07315865325927734, 0.07248281860351563, 0.07143218994140625, 0.07166361236572266, 0.07237344360351562, 0.07268537902832031, 0.07203228759765624, 0.07159190368652343, 0.07179468536376953, 0.07157453155517578, 0.07146701049804688, 0.07154994964599609, 0.07143833923339844, 0.0720005111694336, 0.0722422103881836, 0.07253823852539062, 0.07229526519775391, 0.07265484619140625, 0.07151923370361328, 0.15542373657226563, 0.07168102264404297, 0.07164620971679687, 0.07139328002929687, 0.07188992309570312, 0.07269683074951172, 0.07254835510253907, 0.0723773422241211, 0.07155609893798828, 0.07221043395996093, 0.071583740234375, 0.07193702697753906, 0.07257292938232422, 0.07212134552001953, 0.07240396881103515, 0.07164518737792969, 0.07148851013183594, 0.07235382080078125, 0.07236399841308594, 0.07167692565917969, 0.07185100555419922, 0.0719288330078125, 0.07146803283691407, 0.07169843292236328, 0.07143730926513672, 0.07199641418457031, 0.07272351837158203, 0.07263839721679688, 0.07242447662353516, 0.07246947479248046, 0.0725002212524414, 0.0726087646484375, 0.07243673706054687, 0.0725382080078125, 0.07221135711669922, 0.07237840270996093, 0.07187760162353515, 0.07260671997070313, 0.07360921478271484, 0.07258624267578125, 0.07270604705810547, 0.07254425811767579, 0.07218892669677734, 0.07190220642089844, 0.07209779357910157, 0.07169741058349609, 0.07152333068847656, 0.07145894622802734, 0.07228096008300781, 0.07201996612548828, 0.07152435302734375, 0.07199436950683594, 0.07146086120605469, 0.07159606170654297, 0.07153250885009765, 0.07242854309082031, 0.07224524688720703, 0.07124889373779297, 0.07148457336425781, 0.07174845123291015, 0.07160323333740234, 0.07141065979003906, 0.07256371307373047, 0.15466802978515626, 0.07162790679931641, 0.07138700866699219, 0.07146701049804688, 0.0714229736328125, 0.0716605453491211, 0.07148851013183594, 0.0715888671875, 0.07149772644042969, 0.07135142517089844, 0.07130406188964844, 0.07140351867675782, 0.07143955230712891, 0.07148115539550781, 0.07158477020263672, 0.07155302429199219, 0.07171071624755859, 0.07169945526123046, 0.07139737701416016, 0.07129917144775391, 0.07149353790283203, 0.07145062255859375, 0.07140966033935547, 0.07134925079345703, 0.07143526458740235, 0.07143321228027344, 0.07270912170410156, 0.07148236846923828, 0.07134003448486329, 0.07151513671875, 0.0714803237915039, 0.07152333068847656, 0.07151923370361328, 0.07155097961425781, 0.07148339080810547, 0.07168000030517578, 0.07140659332275391, 0.07138508605957031, 0.07145267486572265, 0.07127040100097656, 0.07239167785644532, 0.07333273315429688, 0.0716072998046875, 0.07177728271484375, 0.07187251281738281, 0.07153254699707032, 0.07148544311523437, 0.07157247924804687, 0.07153667449951172, 0.07155606079101562, 0.0714772491455078, 0.07148544311523437, 0.0713534698486328, 0.07227993774414063, 0.0716072998046875, 0.07318630218505859, 0.0728053741455078, 0.07240191650390625, 0.07223500823974609, 0.07155404663085937, 0.07140351867675782, 0.07203433227539062, 0.07255753326416016]",tokens/s,13.650554316122056,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1129,7 +1129,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpw763jynu/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmprpakw4k4/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -1162,7 +1162,7 @@ ImportError: This modeling file requires the following packages that were not fo ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1389.85472,1596.45696,0.0,950.009856,857.516544,s,10,0.5760096969604492,0.057600969696044925,0.00029658166146349625,0.057502817153930665,0.057649400329589845,0.05806812362670898,0.058403102264404295,"[0.05848684692382813, 0.05749667358398437, 0.05749142456054687, 0.05744214248657226, 0.05752342224121094, 0.05748796844482422, 0.05755635070800781, 0.05751923370361328, 0.057503841400146485, 0.057501792907714844]",tokens/s,4444.3696234783665,kWh,6.794432075848831e-07,3.7230317196598e-07,2.898270198562978e-06,3.950016578113841e-06,tokens/kWh,64809854.5759121,MB,1389.85472,1596.45696,0.0,950.009856,888.373248,s,10,30.99354663085937,3.099354663085937,0.004435069727024205,3.0984003906249997,3.1061409912109372,3.1076104125976562,3.108785949707031,"[3.109079833984375, 3.09672705078125, 3.099322021484375, 3.0984990234375, 3.0983017578125, 3.09989990234375, 3.094158447265625, 3.095660400390625, 3.105814453125, 3.096083740234375]",tokens/s,20.326812142651896,kWh,3.656450825214082e-05,2.003794690709175e-05,0.0001537495773418402,0.00021035203250107276,tokens/kWh,299497.9380561902,,s,629,31.461226520538325,0.05001784820435347,0.006851285101616403,0.049097728729248044,0.0493834228515625,0.04992880630493164,0.10677215362548828,"[0.05049651336669922, 0.050405376434326174, 0.050948097229003904, 0.05020876693725586, 0.05066547012329101, 0.05147238540649414, 0.050577407836914064, 0.049240062713623044, 0.05071974563598633, 0.049844223022460936, 0.04920832061767578, 0.04908544158935547, 0.04913663864135742, 0.04943360137939453, 0.04917657470703125, 0.049121280670166016, 0.04936294555664063, 0.049102848052978515, 0.049140735626220705, 0.049107967376708986, 0.0490967025756836, 0.04909568023681641, 0.0490618896484375, 0.04907827377319336, 0.04910079956054687, 0.04911206436157227, 0.04908544158935547, 0.04910489654541016, 0.049099777221679686, 0.04909363174438477, 0.049083393096923826, 0.049111038208007815, 0.04908249664306641, 0.04901465606689453, 0.049122303009033204, 0.04907724761962891, 0.04908031845092774, 0.04905779266357422, 0.049084415435791014, 0.049073150634765625, 0.04904345703125, 0.04912025451660156, 0.04911308670043945, 0.04913151931762695, 0.04909056091308594, 0.04909465789794922, 0.0490967025756836, 0.04912332916259766, 0.04904959869384766, 0.04907212829589844, 0.0491673583984375, 0.04907827377319336, 0.04908236694335937, 0.049081344604492184, 0.04986470413208008, 0.04910079956054687, 0.04906393432617188, 0.04904652786254883, 0.04906598281860351, 0.049119232177734375, 0.04907622528076172, 0.04908236694335937, 0.10679808044433593, 0.04906598281860351, 0.04905881500244141, 0.04911001586914063, 0.04905472183227539, 0.049189888000488284, 0.04913049697875976, 0.04910489654541016, 0.04904140853881836, 0.049073150634765625, 0.04907622528076172, 0.04949094390869141, 0.04909875106811523, 0.049068031311035154, 0.04907212829589844, 0.049102848052978515, 0.049102848052978515, 0.04915507125854492, 0.049083393096923826, 0.049084415435791014, 0.049081344604492184, 0.04906598281860351, 0.04904755020141602, 0.049116161346435545, 0.049070079803466796, 0.05003366470336914, 0.04913356781005859, 0.049099777221679686, 0.049075199127197267, 0.049083393096923826, 0.049012737274169924, 0.04907417678833008, 0.04910182571411133, 0.04908236694335937, 0.04904243087768555, 0.04910489654541016, 0.04907827377319336, 0.049075199127197267, 0.04907622528076172, 0.04906393432617188, 0.049091583251953126, 0.04907827377319336, 0.04909875106811523, 0.04911001586914063, 0.04906393432617188, 0.049895423889160156, 0.04930559921264648, 0.0490967025756836, 0.049137664794921876, 0.049105918884277344, 0.04907212829589844, 0.049067008972167966, 0.049138687133789063, 0.04910489654541016, 0.04961587142944336, 0.04918476867675781, 0.049168384552001954, 0.04962815856933594, 0.04916940689086914, 0.04909568023681641, 0.04903936004638672, 0.049099777221679686, 0.04940902328491211, 0.10675814056396485, 0.04907929611206055, 0.049081344604492184, 0.0491038703918457, 0.049137664794921876, 0.049672191619873046, 0.04973158264160156, 0.049089534759521485, 0.04917657470703125, 0.04911206436157227, 0.04907110214233398, 0.049073150634765625, 0.04910079956054687, 0.04909260940551758, 0.04909465789794922, 0.04919500732421875, 0.050680831909179686, 0.049188865661621096, 0.049064960479736325, 0.049084415435791014, 0.04910899353027344, 0.049067008972167966, 0.051506175994873046, 0.04994559860229492, 0.049116161346435545, 0.04910185623168945, 0.049081310272216794, 0.049089534759521485, 0.04911001586914063, 0.04913459014892578, 0.04911820983886719, 0.04904857635498047, 0.049114112854003904, 0.04907110214233398, 0.04906291198730469, 0.049135616302490234, 0.04920832061767578, 0.049073150634765625, 0.04910079956054687, 0.04906905746459961, 0.0491069450378418, 0.049051647186279294, 0.049073150634765625, 0.04907827377319336, 0.049081344604492184, 0.04908544158935547, 0.0491069450378418, 0.04909875106811523, 0.04913971328735352, 0.049073150634765625, 0.04913356781005859, 0.04909465789794922, 0.04909465789794922, 0.04904550552368164, 0.04907622528076172, 0.049064960479736325, 0.04907929611206055, 0.04911308670043945, 0.049122303009033204, 0.04905472183227539, 0.049053695678710936, 0.049102848052978515, 0.04907212829589844, 0.10670285034179687, 0.04906598281860351, 0.04905267333984375, 0.04907417678833008, 0.04902297592163086, 0.04910182571411133, 0.04911718368530273, 0.049089534759521485, 0.049053695678710936, 0.04913151931762695, 0.04907724761962891, 0.049097728729248044, 0.049083393096923826, 0.04906598281860351, 0.04904857635498047, 0.049084415435791014, 0.04911206436157227, 0.04911308670043945, 0.0490618896484375, 0.049121280670166016, 0.049089534759521485, 0.04902912139892578, 0.0491038703918457, 0.049138687133789063, 0.0490618896484375, 0.049259521484375, 0.04914688110351562, 0.049067008972167966, 0.05005414581298828, 0.04943155288696289, 0.049070079803466796, 0.049051647186279294, 0.04918272018432617, 0.04906086349487305, 0.049073150634765625, 0.04916428756713867, 0.049070079803466796, 0.04912639999389649, 0.049083393096923826, 0.04919500732421875, 0.04906598281860351, 0.04906086349487305, 0.04960255813598633, 0.049073150634765625, 0.04903219223022461, 0.049242111206054685, 0.04934143829345703, 0.049083393096923826, 0.04913356781005859, 0.05064089584350586, 0.04961382293701172, 0.04916326522827148, 0.049143806457519534, 0.04909465789794922, 0.04909363174438477, 0.04930867385864258, 0.04911820983886719, 0.049242111206054685, 0.04964659118652344, 0.049168384552001954, 0.04912947082519531, 0.04917248153686524, 0.049056766510009765, 0.10683699035644531, 0.04909056091308594, 0.049157119750976565, 0.049089534759521485, 0.049242111206054685, 0.04905779266357422, 0.049152000427246094, 0.04907827377319336, 0.04910079956054687, 0.049091583251953126, 0.049616897583007816, 0.049105918884277344, 0.04906905746459961, 0.04906086349487305, 0.04916326522827148, 0.049084415435791014, 0.049051647186279294, 0.0493199348449707, 0.049083393096923826, 0.049067008972167966, 0.04910899353027344, 0.04909056091308594, 0.04917555236816406, 0.04982070541381836, 0.0490731201171875, 0.049263614654541016, 0.04909056091308594, 0.04911718368530273, 0.04912025451660156, 0.04908748626708984, 0.04907929611206055, 0.049154048919677736, 0.049355777740478515, 0.049138687133789063, 0.04907622528076172, 0.049102848052978515, 0.049408000946044923, 0.04905881500244141, 0.04905062484741211, 0.04926464080810547, 0.05005516815185547, 0.04912332916259766, 0.049070079803466796, 0.049261600494384765, 0.049047519683837894, 0.049135616302490234, 0.049086463928222655, 0.04916326522827148, 0.04908031845092774, 0.04905779266357422, 0.04910489654541016, 0.049116161346435545, 0.049102848052978515, 0.04904755020141602, 0.049068031311035154, 0.049107967376708986, 0.05044326400756836, 0.04923187255859375, 0.04912332916259766, 0.049053695678710936, 0.04904550552368164, 0.049099777221679686, 0.049075199127197267, 0.10677760314941406, 0.04906086349487305, 0.04904447937011719, 0.04905779266357422, 0.049353729248046874, 0.04914790344238281, 0.0491069450378418, 0.04914585494995117, 0.04906291198730469, 0.049064960479736325, 0.04909056091308594, 0.04905062484741211, 0.049157119750976565, 0.049089534759521485, 0.04904447937011719, 0.04907929611206055, 0.051057662963867184, 0.0493568000793457, 0.04907929611206055, 0.04907417678833008, 0.04909465789794922, 0.049068031311035154, 0.04905984115600586, 0.0490967025756836, 0.04909465789794922, 0.049114112854003904, 0.0501923828125, 0.050111488342285154, 0.049091583251953126, 0.04909465789794922, 0.049122303009033204, 0.04905984115600586, 0.04907417678833008, 0.05060403060913086, 0.05001113510131836, 0.04911206436157227, 0.049064960479736325, 0.04910182571411133, 0.04907724761962891, 0.049067008972167966, 0.04907417678833008, 0.049114112854003904, 0.0491673583984375, 0.04914585494995117, 0.04900556945800781, 0.049070079803466796, 0.04913151931762695, 0.049068031311035154, 0.04913663864135742, 0.049143806457519534, 0.04906393432617188, 0.049097728729248044, 0.04910182571411133, 0.04909568023681641, 0.04909465789794922, 0.04905574417114258, 0.04909875106811523, 0.0490885124206543, 0.049091583251953126, 0.04907827377319336, 0.04912947082519531, 0.04908031845092774, 0.049073150634765625, 0.10681855773925782, 0.04911206436157227, 0.049053695678710936, 0.04912639999389649, 0.0492564468383789, 0.04907827377319336, 0.049084415435791014, 0.049067008972167966, 0.04919705581665039, 0.04909260940551758, 0.04912844848632812, 0.049122303009033204, 0.049089534759521485, 0.049097728729248044, 0.04908031845092774, 0.04907110214233398, 0.049056766510009765, 0.049173503875732424, 0.04913971328735352, 0.04907622528076172, 0.049102848052978515, 0.04907827377319336, 0.04908236694335937, 0.04906086349487305, 0.04930047988891602, 0.049223678588867184, 0.04916326522827148, 0.04913151931762695, 0.04917248153686524, 0.049084415435791014, 0.04916633605957031, 0.04911206436157227, 0.04910182571411133, 0.04905779266357422, 0.049170433044433595, 0.04912639999389649, 0.049086463928222655, 0.04912844848632812, 0.0491069450378418, 0.049064960479736325, 0.049111038208007815, 0.04907827377319336, 0.04906393432617188, 0.04908544158935547, 0.04906291198730469, 0.04913459014892578, 0.04911513519287109, 0.049084415435791014, 0.04907724761962891, 0.04908031845092774, 0.049124351501464845, 0.04911001586914063, 0.049124351501464845, 0.049086463928222655, 0.049084415435791014, 0.04907827377319336, 0.049075199127197267, 0.04916326522827148, 0.04912844848632812, 0.04907110214233398, 0.04905779266357422, 0.04910079956054687, 0.04921139144897461, 0.10679500579833984, 0.04903219223022461, 0.04903014373779297, 0.049089534759521485, 0.04902707290649414, 0.04915609741210938, 0.0491253776550293, 0.04910182571411133, 0.049514495849609375, 0.049067008972167966, 0.04910489654541016, 0.04907212829589844, 0.04908544158935547, 0.049111038208007815, 0.04907724761962891, 0.04912025451660156, 0.049102848052978515, 0.04916428756713867, 0.04926259231567383, 0.04909260940551758, 0.04908544158935547, 0.04905984115600586, 0.049075199127197267, 0.049119232177734375, 0.04907622528076172, 0.0492492790222168, 0.04908544158935547, 0.04903219223022461, 0.04915097427368164, 0.049075199127197267, 0.04907417678833008, 0.049084415435791014, 0.04909363174438477, 0.04907929611206055, 0.049160190582275394, 0.049124351501464845, 0.049084415435791014, 0.049089534759521485, 0.04903628921508789, 0.04908748626708984, 0.04905574417114258, 0.04930047988891602, 0.0490885124206543, 0.049876991271972655, 0.049073150634765625, 0.049081344604492184, 0.04921855926513672, 0.04910079956054687, 0.0491038703918457, 0.049119232177734375, 0.04906905746459961, 0.049056766510009765, 0.049086463928222655, 0.049291263580322264, 0.04934656143188477, 0.049522689819335934, 0.049116161346435545, 0.04921139144897461, 0.04907417678833008, 0.04907724761962891, 0.04917555236816406, 0.04904243087768555, 0.049070079803466796, 0.10677760314941406, 0.04971212768554688, 0.04909568023681641, 0.049225727081298826, 0.04921036911010742, 0.049083393096923826, 0.049067008972167966, 0.049160190582275394, 0.04909465789794922, 0.04918374252319336, 0.049148929595947265, 0.049081344604492184, 0.04911820983886719, 0.049075199127197267, 0.049111038208007815, 0.04916223907470703, 0.04908544158935547, 0.049083393096923826, 0.04907622528076172, 0.04908236694335937, 0.049084415435791014, 0.04906905746459961, 0.049170433044433595, 0.04947251129150391, 0.04919705581665039, 0.04907827377319336, 0.04929740905761719, 0.0501319694519043, 0.04920115280151367, 0.04922777557373047, 0.04910899353027344, 0.049091583251953126, 0.04911513519287109, 0.049091583251953126, 0.04905881500244141, 0.04925132751464844, 0.04913356781005859, 0.04906905746459961, 0.04914585494995117, 0.04909568023681641, 0.049111038208007815, 0.04907827377319336, 0.04944384002685547, 0.049111038208007815, 0.04939263916015625, 0.055959552764892576, 0.04918067169189453, 0.049107967376708986, 0.04907929611206055, 0.04907212829589844, 0.04911308670043945, 0.04913356781005859, 0.04953497695922852, 0.049137664794921876, 0.0493834228515625, 0.04908544158935547, 0.04901683044433594, 0.049097728729248044, 0.04940697479248047, 0.049903617858886716, 0.049137664794921876, 0.049086463928222655, 0.04905779266357422, 0.10681753540039063, 0.04907417678833008, 0.04907622528076172, 0.049124351501464845, 0.04902604675292969, 0.04909465789794922, 0.049188865661621096, 0.04913049697875976, 0.04907417678833008, 0.04910489654541016, 0.049105918884277344, 0.0490885124206543, 0.04907724761962891, 0.049091583251953126, 0.04906291198730469, 0.049152000427246094, 0.04910182571411133, 0.04914688110351562, 0.04909465789794922, 0.04907724761962891, 0.04907724761962891, 0.04907724761962891, 0.04907110214233398, 0.0491253776550293, 0.04909568023681641, 0.049116161346435545, 0.04909465789794922, 0.04915097427368164, 0.049132545471191405, 0.04907212829589844, 0.049081344604492184, 0.049111038208007815, 0.049114112854003904, 0.04907417678833008, 0.04904755020141602, 0.04917657470703125, 0.04907212829589844, 0.04914995193481445, 0.04906598281860351, 0.04911308670043945, 0.049225727081298826, 0.04926054382324219, 0.049086463928222655, 0.04920835113525391, 0.04906800079345703, 0.049084415435791014, 0.049124351501464845, 0.049091583251953126, 0.04918476867675781, 0.04911718368530273, 0.049805313110351565, 0.04936908721923828, 0.0493834228515625, 0.04906598281860351, 0.04907212829589844, 0.049070079803466796, 0.04913459014892578, 0.04940492630004883, 0.04909568023681641, 0.04909260940551758, 0.04974694442749023, 0.049086463928222655, 0.04910079956054687]",tokens/s,19.992863265816414,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1201,7 +1201,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4680-65267c5367b98e643d8c8494;206a362f-168f-40b5-92ef-51ee8675e54a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe9aa-6a2e8bef425c434820a299ca;32993da6-387a-4637-b6c5-cf17d34ba54a) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1258,7 +1258,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp8_yh_pe3/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpsmlr85vg/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -1300,7 +1300,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e959c-237996b20310ed6e5b08a88c;774e0bb5-df11-46a6-beec-622e73521546) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe76b-77ff8ae96edb0cdb5a218871;4885aeee-9c6f-4224-98da-4c55d1c9c617) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1371,7 +1371,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9640-76bd665a142df2135469a6c4;4a5cae97-d0cb-436f-b3f3-4cc3fb9af705) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe80e-3a7d207b247029fa519e7093;b6f27841-342e-4fb3-ae7a-bad8629ef801) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1403,7 +1403,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5020.81536,8173.125632,0.0,7526.678528,6833.613824,s,10,3.326392669677735,0.3326392669677734,0.0030324363165483534,0.33195497131347657,0.3341614532470703,0.3376804428100586,0.3404956344604492,"[0.3411994323730469, 0.33337945556640625, 0.33058718872070314, 0.3303564453125, 0.33089251708984374, 0.33060626220703127, 0.3314384460449219, 0.332591064453125, 0.332870361328125, 0.33247149658203123]",tokens/s,769.6024655585886,kWh,3.897490269607968e-06,2.1353897317655327e-06,2.0114881683290285e-05,2.6147761684663786e-05,tokens/kWh,9790512.973435482,MB,5020.81536,8173.125632,0.0,7526.678528,7110.583808,s,10,183.47133203125,18.347133203125,0.004664256892410071,18.349021484375,18.351409570312498,18.35145087890625,18.35148392578125,"[18.348591796875, 18.3397421875, 18.34241796875, 18.338681640625, 18.3502734375, 18.348642578125, 18.349400390625, 18.350689453125, 18.3514921875, 18.351400390625]",tokens/s,3.4337789616782985,kWh,0.00021652181481321653,0.0001186722658167949,0.0010773519962895067,0.0014125460769195183,tokens/kWh,44600.31501230066,,s,629,186.15873925781258,0.29595983983753976,0.03945461951219622,0.29114471435546874,0.29161677246093753,0.29189117431640627,0.6223596313476563,"[0.29115188598632813, 0.2917990417480469, 0.2910863342285156, 0.29130035400390625, 0.2909051818847656, 0.2912214050292969, 0.2909573059082031, 0.29103411865234374, 0.2910054321289062, 0.2911651916503906, 0.2917939147949219, 0.2913239135742188, 0.2910597839355469, 0.2918880615234375, 0.29148263549804687, 0.29144268798828127, 0.2912911376953125, 0.29127783203125, 0.2912788391113281, 0.29161370849609375, 0.291304443359375, 0.29117236328125, 0.2911733703613281, 0.29094400024414063, 0.29102081298828125, 0.2911559753417969, 0.2910597229003906, 0.2914805908203125, 0.2913699951171875, 0.29076275634765625, 0.29096661376953126, 0.29093161010742186, 0.2912389221191406, 0.29129931640625, 0.2909480895996094, 0.2912174072265625, 0.2912204895019531, 0.29107199096679687, 0.2910310363769531, 0.29094091796875, 0.29076480102539065, 0.29079766845703126, 0.29086407470703124, 0.29078524780273435, 0.2911088562011719, 0.29134234619140625, 0.2910627746582031, 0.29169970703125, 0.29144064331054687, 0.2908927917480469, 0.2910648193359375, 0.2910535583496094, 0.2909266052246094, 0.29100030517578124, 0.2910525512695312, 0.2911098937988281, 0.2910382080078125, 0.29060504150390626, 0.2909347839355469, 0.2907238464355469, 0.29095013427734373, 0.29088467407226565, 0.6249952392578125, 0.29085287475585936, 0.2911539306640625, 0.290904052734375, 0.2911037292480469, 0.2912143249511719, 0.2912419738769531, 0.291493896484375, 0.2911508483886719, 0.291056640625, 0.2912942504882812, 0.2909091491699219, 0.2910955505371094, 0.2910320739746094, 0.2910310363769531, 0.29128805541992187, 0.2911457214355469, 0.29099114990234376, 0.291136474609375, 0.291462158203125, 0.29125222778320314, 0.2911016845703125, 0.290882568359375, 0.290977783203125, 0.2910105590820313, 0.2909910888671875, 0.29104946899414064, 0.2909276123046875, 0.2908713073730469, 0.29091635131835936, 0.2908846130371094, 0.29086822509765625, 0.29094400024414063, 0.2912911376953125, 0.291083251953125, 0.2909296569824219, 0.2908917846679687, 0.2909429626464844, 0.29090509033203127, 0.29101669311523437, 0.2909276123046875, 0.2912174072265625, 0.29108837890625, 0.29093682861328124, 0.29119793701171875, 0.29107916259765626, 0.2934343566894531, 0.29095834350585936, 0.2907484130859375, 0.29136895751953124, 0.29149798583984377, 0.29114471435546874, 0.29111294555664063, 0.29156658935546875, 0.29093792724609374, 0.2908681640625, 0.2908518371582031, 0.2906265563964844, 0.2906163330078125, 0.2909542541503906, 0.29111505126953124, 0.2914887390136719, 0.29113345336914065, 0.6221834106445312, 0.291056640625, 0.2913434143066406, 0.29181332397460935, 0.2915881042480469, 0.2911866760253906, 0.2914805908203125, 0.29122662353515627, 0.29128805541992187, 0.29116824340820313, 0.2914232177734375, 0.2919321594238281, 0.29145498657226565, 0.2910115966796875, 0.29119488525390624, 0.2909060974121094, 0.29112115478515627, 0.29129421997070315, 0.29111294555664063, 0.29093887329101564, 0.2918696899414063, 0.2910146484375, 0.2910586853027344, 0.2929541015625, 0.29112115478515627, 0.29093887329101564, 0.29098599243164064, 0.2910157775878906, 0.2910462951660156, 0.291080322265625, 0.29077490234375, 0.2908006286621094, 0.2908897399902344, 0.29091943359375, 0.290830322265625, 0.29089688110351564, 0.2908344421386719, 0.29086004638671875, 0.29178369140625, 0.29098904418945315, 0.29101568603515626, 0.2907648315429687, 0.29071978759765627, 0.29085177612304686, 0.2908037109375, 0.2908999633789062, 0.2911928405761719, 0.29247796630859374, 0.29141607666015623, 0.2910228576660156, 0.2908846130371094, 0.2909020080566406, 0.2911375427246094, 0.290845703125, 0.2914642028808594, 0.29115493774414064, 0.2909491271972656, 0.290914306640625, 0.2911805419921875, 0.29089483642578123, 0.2907873229980469, 0.2908323974609375, 0.290597900390625, 0.6217267456054687, 0.29077197265625, 0.2910330810546875, 0.2909470825195313, 0.29107510375976564, 0.29102178955078123, 0.29123788452148436, 0.29106893920898436, 0.29189324951171874, 0.2911856689453125, 0.2912204895019531, 0.29157272338867185, 0.29196185302734373, 0.2910382080078125, 0.2912010498046875, 0.29094088745117186, 0.29163418579101563, 0.290946044921875, 0.29115493774414064, 0.2909962158203125, 0.2909952087402344, 0.2908518371582031, 0.29090310668945313, 0.29097158813476565, 0.29139556884765627, 0.2911918029785156, 0.29115493774414064, 0.2910177307128906, 0.291470458984375, 0.2910995178222656, 0.2909153137207031, 0.29085287475585936, 0.29129421997070315, 0.2910177307128906, 0.29091738891601565, 0.290850830078125, 0.2908016662597656, 0.2909204406738281, 0.290703369140625, 0.2909480895996094, 0.290808837890625, 0.2907821960449219, 0.2912767944335938, 0.29103521728515624, 0.2930513000488281, 0.2909010009765625, 0.2909573059082031, 0.2909429626464844, 0.29101364135742186, 0.2907362060546875, 0.29087738037109373, 0.29123077392578123, 0.2909122009277344, 0.2908518371582031, 0.29091021728515626, 0.29078021240234375, 0.291042236328125, 0.29075250244140627, 0.2906828918457031, 0.29115399169921874, 0.29102072143554686, 0.291009521484375, 0.29132696533203123, 0.6224281616210937, 0.29084161376953127, 0.29141094970703124, 0.29116824340820313, 0.29133004760742187, 0.2911754150390625, 0.2917089233398438, 0.2911651916503906, 0.29145394897460936, 0.29128704833984376, 0.29129010009765627, 0.29106585693359377, 0.2920130615234375, 0.2914622497558594, 0.2913115234375, 0.291346435546875, 0.291209228515625, 0.292105224609375, 0.29109759521484374, 0.2908907470703125, 0.29111090087890623, 0.2913320922851563, 0.291009521484375, 0.2909900817871094, 0.2909204406738281, 0.29106893920898436, 0.29144268798828127, 0.2911283264160156, 0.2910525512695312, 0.2916495361328125, 0.2910535583496094, 0.2909675598144531, 0.291346435546875, 0.2914672546386719, 0.2912665710449219, 0.29141607666015623, 0.29099212646484374, 0.29119384765625, 0.2911212463378906, 0.29125518798828126, 0.29113345336914065, 0.2911068115234375, 0.2909552612304688, 0.2910054321289062, 0.29141812133789063, 0.29120306396484374, 0.29097164916992185, 0.29154507446289063, 0.29104742431640623, 0.29088870239257814, 0.2910064697265625, 0.29087435913085935, 0.2910402526855469, 0.2909849548339844, 0.29111199951171873, 0.29143136596679686, 0.291462158203125, 0.2940467224121094, 0.29119488525390624, 0.2914396057128906, 0.2912563171386719, 0.2915348510742187, 0.291083251953125, 0.622940185546875, 0.2911180725097656, 0.291894287109375, 0.2918707885742188, 0.29123678588867186, 0.291156982421875, 0.2911457214355469, 0.2911201171875, 0.29126556396484377, 0.2914580078125, 0.2914068603515625, 0.2910320739746094, 0.29122354125976563, 0.2909378662109375, 0.29136383056640625, 0.29118463134765626, 0.2914672546386719, 0.2917652587890625, 0.2911856689453125, 0.29107098388671876, 0.29113446044921876, 0.29076788330078124, 0.29069720458984377, 0.29090304565429687, 0.29082623291015625, 0.29081088256835935, 0.291056640625, 0.29096347045898435, 0.29087435913085935, 0.291009521484375, 0.2907371520996094, 0.2910607299804687, 0.29096038818359377, 0.2913167419433594, 0.2914723815917969, 0.2914847717285156, 0.2912306213378906, 0.291462158203125, 0.2936002502441406, 0.2915155029296875, 0.2913965148925781, 0.29102386474609376, 0.2910607299804687, 0.2909644775390625, 0.29098904418945315, 0.29122970581054686, 0.29106997680664065, 0.29096649169921873, 0.2909337463378906, 0.2908917846679687, 0.2909982604980469, 0.2911559753417969, 0.29184307861328124, 0.2915625, 0.29157272338867185, 0.29153689575195313, 0.29130859375, 0.29116513061523436, 0.2910146484375, 0.2910955505371094, 0.2916024169921875, 0.291262451171875, 0.2909020080566406, 0.6232821655273437, 0.2913167419433594, 0.29129421997070315, 0.2912972717285156, 0.2918635559082031, 0.29147442626953124, 0.2911283264160156, 0.2916812744140625, 0.2915973205566406, 0.2912419738769531, 0.2914723815917969, 0.291281982421875, 0.29155218505859376, 0.29142630004882814, 0.29158706665039064, 0.2915564270019531, 0.2916771240234375, 0.29130136108398436, 0.2912850646972656, 0.2912767333984375, 0.2914969482421875, 0.2918369140625, 0.2913320922851563, 0.29271755981445313, 0.29146624755859374, 0.2911907958984375, 0.2909296569824219, 0.2910955505371094, 0.291083251953125, 0.2912593994140625, 0.2912706604003906, 0.29119589233398435, 0.29104638671875, 0.29097268676757815, 0.29096038818359377, 0.29111294555664063, 0.2913320922851563, 0.291304443359375, 0.29161575317382815, 0.2912389221191406, 0.29126040649414064, 0.2912554626464844, 0.2910042419433594, 0.2909931640625, 0.2908856201171875, 0.29083340454101564, 0.2908917846679687, 0.290946044921875, 0.2909757385253906, 0.29078323364257813, 0.29109762573242187, 0.29080374145507815, 0.29088555908203123, 0.29080780029296877, 0.2909491271972656, 0.2907904052734375, 0.2916864013671875, 0.2912890930175781, 0.2907801513671875, 0.29119692993164065, 0.29122457885742187, 0.29137411499023436, 0.291031005859375, 0.6241597290039063, 0.29123788452148436, 0.29148980712890626, 0.29125018310546874, 0.29131057739257815, 0.29118463134765626, 0.29127166748046873, 0.291272705078125, 0.2918543395996094, 0.2915051574707031, 0.291631103515625, 0.29105560302734373, 0.2923868103027344, 0.2912337951660156, 0.2911866760253906, 0.291061767578125, 0.292490234375, 0.2912214965820312, 0.291267578125, 0.2910013427734375, 0.29129421997070315, 0.2910822448730469, 0.29093582153320313, 0.29120001220703123, 0.29091943359375, 0.2908856201171875, 0.29145703125, 0.2914949035644531, 0.29112933349609377, 0.29123788452148436, 0.2912767944335938, 0.2912133178710937, 0.29131365966796874, 0.29124609375, 0.2912194519042969, 0.29173348999023435, 0.2911201171875, 0.2912788391113281, 0.2910044250488281, 0.2910013427734375, 0.29118771362304685, 0.2910330810546875, 0.2915604553222656, 0.29146316528320315, 0.2914334716796875, 0.29083953857421874, 0.29135769653320315, 0.2911631469726563, 0.29084979248046877, 0.2908221435546875, 0.29091122436523437, 0.29104229736328124, 0.2908825988769531, 0.29070947265625, 0.29156146240234376, 0.292242431640625, 0.29155123901367186, 0.29115493774414064, 0.29105670166015624, 0.2912132568359375, 0.29169766235351563, 0.2913536071777344, 0.291162109375, 0.6245776977539063, 0.2910738830566406, 0.29133721923828126, 0.29104638671875, 0.29137100219726564, 0.2918984375, 0.291302490234375, 0.29108514404296876, 0.2911313781738281, 0.29162203979492185, 0.2911015319824219, 0.29115188598632813, 0.29129421997070315, 0.2920244140625, 0.29145590209960937, 0.29112423706054685, 0.291620849609375, 0.29239004516601563, 0.29183676147460935, 0.29133517456054686, 0.29116928100585937, 0.2914068603515625, 0.29107916259765626, 0.2920724487304687, 0.2911928405761719, 0.29108837890625, 0.2911365051269531, 0.29100747680664063, 0.2909665832519531, 0.29107705688476565, 0.2915195007324219, 0.29198641967773437, 0.2912174072265625, 0.2914068603515625, 0.29100747680664063, 0.29123480224609377, 0.29113037109375, 0.29115188598632813, 0.2909010009765625, 0.2909767761230469, 0.29136282348632814, 0.2908538818359375, 0.29098599243164064, 0.29118463134765626, 0.2911078491210938, 0.29112628173828126, 0.29127783203125, 0.2916915283203125, 0.2914672546386719, 0.29112628173828126, 0.2914969482421875, 0.29150311279296875, 0.29141299438476564, 0.2911457214355469, 0.2911907958984375, 0.2912061462402344, 0.29095834350585936, 0.29087335205078124, 0.2910064697265625, 0.2916515808105469, 0.29124505615234375, 0.2910699462890625, 0.2909389038085938, 0.6244362182617188, 0.2915553283691406, 0.29160562133789064, 0.29136474609375, 0.2912962646484375, 0.2911641540527344, 0.29157272338867185, 0.2914303894042969, 0.29142630004882814, 0.291240966796875, 0.29158303833007815, 0.291295166015625, 0.2915502014160156, 0.291373046875, 0.2917918701171875, 0.29154507446289063, 0.2914375610351562, 0.2918666687011719, 0.291716064453125, 0.29142630004882814, 0.2913177490234375, 0.29200588989257814, 0.29093582153320313, 0.2909337463378906, 0.29117440795898436, 0.29120620727539065, 0.29325714111328127, 0.2910044250488281, 0.29144064331054687, 0.291175537109375, 0.29100634765625, 0.29097576904296873, 0.29103927612304686, 0.29099923706054687, 0.29096038818359377, 0.2911754150390625, 0.2910146484375, 0.29110272216796873, 0.2911170654296875, 0.2914334716796875, 0.29108428955078125, 0.29090304565429687, 0.2909296569824219, 0.290904052734375, 0.2912860107421875, 0.2911866760253906, 0.29128500366210935, 0.2911088562011719, 0.29134951782226565, 0.29139764404296875, 0.29150411987304686, 0.291240966796875, 0.29109146118164064, 0.29109658813476563, 0.2911150207519531, 0.2918696899414063, 0.29107916259765626, 0.29097369384765626, 0.29076480102539065, 0.2908671875, 0.2908917846679687, 0.29114776611328125, 0.2908713073730469]",tokens/s,3.3788368061995393,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1429,11 +1429,11 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpr265qkfc/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpri8jsnva/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4706.717696,21498.42944,0.0,20851.982336,20236.60032,s,10,15.640249145507811,1.5640249145507812,0.00168203754809238,1.563786376953125,1.564931494140625,1.5666584350585937,1.5680399877929687,"[1.5630980224609374, 1.5644747314453125, 1.5625794677734375, 1.56453466796875, 1.5645272216796875, 1.5645477294921875, 1.5624769287109375, 1.5628692626953125, 1.5683853759765625, 1.5627557373046874]",tokens/s,163.68025702041214,kWh,1.8462728808323543e-05,1.0117648402811028e-05,9.129585081439906e-05,0.00011987622802553362,tokens/kWh,2135535.9958896274,MB,4706.717696,21498.42944,0.0,20851.982336,20339.706368,s,10,879.7075468749999,87.97075468749999,0.002247932535373388,87.97129296874999,87.973384375,87.973496875,87.973586875,"[87.9672265625, 87.968984375, 87.9679609375, 87.9719375, 87.972140625, 87.973609375, 87.973359375, 87.968734375, 87.9729453125, 87.9706484375]",tokens/s,0.7161470902892214,kWh,0.0010385013203819595,0.0005691906291757731,0.005147801034904198,0.0067554929844619305,tokens/kWh,9325.744271351337,,s,629,892.387736694337,1.4187404398956058,0.1859067280964402,1.39627001953125,1.3965248535156252,1.3967189697265625,2.9604817578125,"[1.39608984375, 1.3963448486328125, 1.3963612060546875, 1.3961072998046875, 1.3964923095703126, 1.3963653564453125, 1.396220947265625, 1.3965025634765624, 1.3960928955078125, 1.3962034912109376, 1.3961072998046875, 1.396042724609375, 1.396294677734375, 1.3963448486328125, 1.3963970947265625, 1.3963560791015626, 1.39631201171875, 1.396305908203125, 1.3963643798828125, 1.3958726806640624, 1.3961134033203124, 1.3959710693359375, 1.3963284912109375, 1.396166748046875, 1.3961839599609376, 1.3961236572265625, 1.3963602294921875, 1.3963294677734375, 1.3962445068359375, 1.396177001953125, 1.3961810302734374, 1.3960733642578125, 1.3959208984375, 1.396359130859375, 1.3958072509765624, 1.39637451171875, 1.396179931640625, 1.396137939453125, 1.39983154296875, 1.396505615234375, 1.3963970947265625, 1.3963028564453126, 1.396326416015625, 1.3964442138671875, 1.3961175537109376, 1.396173828125, 1.396177978515625, 1.3965987548828125, 1.3961728515625, 1.3964912109375, 1.3962935791015625, 1.39636328125, 1.39614208984375, 1.3962025146484376, 1.396126708984375, 1.3964083251953125, 1.3963018798828124, 1.396030517578125, 1.3959925537109374, 1.39601513671875, 1.395998779296875, 1.39626904296875, 2.96171923828125, 1.396209716796875, 1.3962506103515624, 1.3962557373046875, 1.3960858154296876, 1.3961513671875, 1.39597412109375, 1.39620654296875, 1.396126708984375, 1.396136962890625, 1.3962086181640625, 1.396041748046875, 1.39642578125, 1.396341796875, 1.3963878173828126, 1.396432861328125, 1.3963028564453126, 1.395948486328125, 1.3960714111328125, 1.3977159423828125, 1.395883056640625, 1.3964984130859375, 1.3963028564453126, 1.3964410400390626, 1.3959066162109375, 1.3964769287109375, 1.396209716796875, 1.3961768798828125, 1.39627001953125, 1.396104248046875, 1.396179931640625, 1.396157470703125, 1.39639404296875, 1.3961943359375, 1.39627001953125, 1.3962998046875, 1.39653125, 1.3978797607421876, 1.39601513671875, 1.39614111328125, 1.3961656494140624, 1.3962823486328124, 1.396073486328125, 1.3962823486328124, 1.3959659423828126, 1.3963233642578126, 1.3965946044921875, 1.3960704345703125, 1.396041748046875, 1.3960120849609374, 1.396105224609375, 1.3962178955078124, 1.39605810546875, 1.396042724609375, 1.3964912109375, 1.3962711181640626, 1.396250732421875, 1.396602783203125, 1.3962762451171875, 1.396516845703125, 1.396326416015625, 1.398640625, 1.3963304443359374, 2.96060302734375, 1.39647900390625, 1.39653125, 1.3961614990234374, 1.3960181884765626, 1.3962342529296874, 1.39601611328125, 1.3958338623046875, 1.39621484375, 1.3961512451171876, 1.3961553955078125, 1.3965394287109374, 1.39652294921875, 1.3962301025390624, 1.396216796875, 1.39635400390625, 1.3961707763671876, 1.396279296875, 1.39607861328125, 1.3962249755859375, 1.395962890625, 1.3963018798828124, 1.396126708984375, 1.3988648681640625, 1.3960888671875, 1.3964031982421874, 1.3962301025390624, 1.3964400634765626, 1.3959556884765625, 1.3960714111328125, 1.3961871337890626, 1.396095947265625, 1.3961861572265626, 1.3961195068359376, 1.3960130615234374, 1.395988525390625, 1.3964349365234374, 1.396515869140625, 1.3959833984375, 1.396220947265625, 1.3962230224609375, 1.396178955078125, 1.396315185546875, 1.3963509521484374, 1.3963929443359375, 1.3960755615234375, 1.3961195068359376, 1.3990062255859375, 1.396305908203125, 1.3959075927734375, 1.3962926025390625, 1.3961728515625, 1.3962230224609375, 1.39626806640625, 1.3965035400390624, 1.3963131103515625, 1.3960294189453124, 1.3960611572265624, 1.396111328125, 1.396221923828125, 1.39627734375, 1.3961727294921875, 1.3960447998046874, 2.961617919921875, 1.3961595458984375, 1.39620458984375, 1.39593212890625, 1.3961533203125, 1.3960263671875, 1.3961451416015624, 1.3962352294921876, 1.3962076416015625, 1.3960560302734375, 1.3962998046875, 1.3964493408203125, 1.39588720703125, 1.396454345703125, 1.3963284912109375, 1.39639501953125, 1.3963642578125, 1.3962127685546875, 1.396188232421875, 1.396221923828125, 1.396116455078125, 1.3964298095703125, 1.3964974365234375, 1.39631201171875, 1.3961666259765626, 1.396533203125, 1.3962506103515624, 1.396453369140625, 1.3960457763671874, 1.396447265625, 1.396515869140625, 1.39631005859375, 1.396305908203125, 1.3964073486328126, 1.396190185546875, 1.3960980224609374, 1.39657421875, 1.396528076171875, 1.3965926513671876, 1.4005042724609376, 1.396791259765625, 1.3964676513671874, 1.3963724365234376, 1.3963826904296874, 1.3961553955078125, 1.396410400390625, 1.39633251953125, 1.3959403076171875, 1.396326416015625, 1.3962608642578125, 1.3963551025390626, 1.3964451904296875, 1.396463623046875, 1.39631103515625, 1.39641552734375, 1.3962301025390624, 1.396401123046875, 1.3960284423828124, 1.3962137451171874, 1.3964359130859374, 1.3964512939453124, 1.3963233642578126, 1.39616357421875, 2.96169482421875, 1.39649951171875, 1.39685986328125, 1.396106201171875, 1.3962987060546874, 1.396274169921875, 1.3959752197265625, 1.3962496337890624, 1.3964339599609374, 1.39608984375, 1.3961011962890626, 1.3962977294921874, 1.3963929443359375, 1.39652197265625, 1.3963203125, 1.3962506103515624, 1.396452392578125, 1.3964031982421874, 1.3964390869140626, 1.396611083984375, 1.3960335693359376, 1.3959803466796874, 1.395914794921875, 1.3961964111328125, 1.3962803955078125, 1.3961553955078125, 1.3965179443359375, 1.39654248046875, 1.3962127685546875, 1.3964461669921875, 1.396337646484375, 1.3962884521484376, 1.3965230712890624, 1.3961533203125, 1.39621484375, 1.3961461181640624, 1.3964666748046874, 1.396431884765625, 1.3962578125, 1.396168701171875, 1.3964727783203126, 1.4002913818359375, 1.396400146484375, 1.3963284912109375, 1.3961728515625, 1.396716552734375, 1.396369384765625, 1.3964410400390626, 1.3962874755859376, 1.3961717529296875, 1.39602734375, 1.396189208984375, 1.39620458984375, 1.3962147216796874, 1.3963602294921875, 1.396046875, 1.39650048828125, 1.3963643798828125, 1.396154296875, 1.396696044921875, 1.396347900390625, 1.3963560791015626, 1.3960509033203126, 2.96171630859375, 1.3963919677734375, 1.3963355712890626, 1.39618505859375, 1.3962926025390625, 1.396115478515625, 1.3960120849609374, 1.396125732421875, 1.396279296875, 1.3963253173828125, 1.3961246337890625, 1.3961092529296875, 1.3965772705078126, 1.396653076171875, 1.3962025146484376, 1.395925048828125, 1.39633154296875, 1.396427734375, 1.3964052734375, 1.3962967529296875, 1.39624755859375, 1.396263916015625, 1.3963182373046874, 1.396494384765625, 1.3964830322265624, 1.3963775634765625, 1.396209716796875, 1.396411376953125, 1.3961912841796875, 1.39610107421875, 1.39620556640625, 1.39635205078125, 1.3962015380859376, 1.396042724609375, 1.3961400146484375, 1.396209716796875, 1.3963079833984375, 1.3961953125, 1.40067431640625, 1.3964073486328126, 1.396336669921875, 1.3963826904296874, 1.396295654296875, 1.39614208984375, 1.3964093017578125, 1.396348876953125, 1.39626806640625, 1.39626806640625, 1.3963868408203124, 1.3961544189453126, 1.3960079345703125, 1.39607861328125, 1.396130859375, 1.3962720947265626, 1.396485107421875, 1.39627001953125, 1.3966878662109374, 1.396177978515625, 1.3960714111328125, 1.39645751953125, 1.3967821044921875, 1.3963929443359375, 1.396221923828125, 2.9624052734375, 1.39640625, 1.396473876953125, 1.396378662109375, 1.3961103515625, 1.3961666259765626, 1.3961380615234376, 1.3961973876953124, 1.3964195556640624, 1.3963243408203125, 1.39621484375, 1.39620556640625, 1.3961348876953126, 1.3965609130859375, 1.3961318359375, 1.396334716796875, 1.396403076171875, 1.3967677001953125, 1.3963458251953125, 1.39630078125, 1.396005859375, 1.396190185546875, 1.396221923828125, 1.3991322021484376, 1.396274169921875, 1.3965567626953126, 1.3959332275390626, 1.39675439453125, 1.3960386962890625, 1.396147216796875, 1.3964830322265624, 1.3962813720703124, 1.3961563720703125, 1.3962669677734374, 1.3960653076171874, 1.3961922607421875, 1.3961605224609375, 1.3963775634765625, 1.396552734375, 1.39624853515625, 1.396236328125, 1.396547607421875, 1.396465576171875, 1.3962352294921876, 1.3962802734375, 1.3960560302734375, 1.3962137451171874, 1.39620556640625, 1.396074462890625, 1.396125732421875, 1.396263916015625, 1.3962332763671874, 1.396306884765625, 1.396263916015625, 1.39645849609375, 1.396369384765625, 1.3965506591796875, 1.396136962890625, 1.395937255859375, 1.3961185302734376, 1.3962281494140625, 1.3962301025390624, 1.400385498046875, 2.960169921875, 1.3960386962890625, 1.3961124267578124, 1.3962587890625, 1.3961328125, 1.3968507080078125, 1.396115478515625, 1.39623828125, 1.3961759033203125, 1.396401123046875, 1.39621484375, 1.3962935791015625, 1.39642578125, 1.3964779052734375, 1.3963294677734375, 1.3960775146484374, 1.397832763671875, 1.396600830078125, 1.39633154296875, 1.3963253173828125, 1.3960980224609374, 1.3963857421875, 1.3962301025390624, 1.39618408203125, 1.396242431640625, 1.3963233642578126, 1.3963560791015626, 1.3964493408203125, 1.39633154296875, 1.396285400390625, 1.3962772216796875, 1.3962454833984375, 1.3963140869140624, 1.396157470703125, 1.39630078125, 1.3960755615234375, 1.3961595458984375, 1.39652392578125, 1.396315185546875, 1.3964451904296875, 1.3963427734375, 1.3962332763671874, 1.39616357421875, 1.3960242919921875, 1.3964410400390626, 1.396295654296875, 1.396368408203125, 1.396148193359375, 1.3961932373046875, 1.396262939453125, 1.3962987060546874, 1.3963458251953125, 1.396189208984375, 1.3962659912109374, 1.3962874755859376, 1.3962987060546874, 1.3963837890625, 1.3963499755859374, 1.3961451416015624, 1.396453369140625, 1.3965670166015625, 1.3963551025390626, 1.3961656494140624, 2.9656298828125, 1.3966131591796875, 1.396486083984375, 1.396496337890625, 1.3961390380859375, 1.3962127685546875, 1.396168701171875, 1.39631005859375, 1.3962650146484374, 1.39608984375, 1.396369384765625, 1.396193359375, 1.3961533203125, 1.396337646484375, 1.396303955078125, 1.3962352294921876, 1.3963612060546875, 1.3983529052734376, 1.39636328125, 1.396357177734375, 1.396263916015625, 1.3963140869140624, 1.3959822998046876, 1.396474853515625, 1.396347900390625, 1.396595703125, 1.396111328125, 1.3961328125, 1.3962701416015626, 1.3961102294921874, 1.3961707763671876, 1.39612158203125, 1.3963203125, 1.3962547607421876, 1.396042724609375, 1.396252685546875, 1.3960447998046874, 1.396010986328125, 1.3960858154296876, 1.3964871826171874, 1.396463623046875, 1.396033447265625, 1.3962823486328124, 1.3962874755859376, 1.3961461181640624, 1.396464599609375, 1.3964410400390626, 1.396326416015625, 1.396304931640625, 1.396130859375, 1.3960826416015626, 1.396231201171875, 1.39595166015625, 1.3962086181640625, 1.39643505859375, 1.3962188720703126, 1.39656396484375, 1.396241455078125, 1.396109375, 1.396360107421875, 1.3962454833984375, 1.396515869140625, 1.4013460693359374, 2.9601669921875, 1.396242431640625, 1.3961563720703125, 1.3967093505859376, 1.3961451416015624, 1.3960601806640625, 1.3962076416015625, 1.396126708984375, 1.396243408203125, 1.3964246826171876, 1.3964666748046874, 1.39620654296875, 1.39607958984375, 1.39641552734375, 1.396093994140625, 1.3960858154296876, 1.3963212890625, 1.396093994140625, 1.3965445556640625, 1.39627001953125, 1.3963775634765625, 1.396485107421875, 1.3963131103515625, 1.3964759521484376, 1.3963275146484375, 1.3965045166015626, 1.39883935546875, 1.3962230224609375, 1.3962762451171875, 1.396242431640625, 1.3961400146484375, 1.3962178955078124, 1.3964820556640625, 1.396148193359375, 1.3961688232421876, 1.3961829833984376, 1.396322265625, 1.396190185546875, 1.396115478515625, 1.39614208984375, 1.3964031982421874, 1.39636328125, 1.3966151123046875, 1.3964134521484375, 1.3963355712890626, 1.396177978515625, 1.396125732421875, 1.396169677734375, 1.396275146484375, 1.3965240478515626, 1.3966817626953125, 1.3962578125, 1.396337646484375, 1.3967205810546874, 1.3967247314453124, 1.396284423828125, 1.396473876953125, 1.3963509521484374, 1.3962127685546875, 1.3961492919921874, 1.3961031494140625, 1.3963387451171876, 1.39633251953125]",tokens/s,0.7048505645427167,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1462,7 +1462,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4abe-1f54ed505cd0f6a56cec5e76;fcf4516b-68aa-4605-8dde-0907dc1144ee) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664feb34-65d9b4837b5682c8799dbe81;83a711b4-fb3c-4e38-8070-c202c16e13f0) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -1547,7 +1547,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9669-704d343a3e559a4146f0cf02;2cff9d5e-d586-4e09-b451-92ddfc277f41) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe837-51569e814b03f25f20a19a90;c528d2cd-1131-417d-8718-3264c50b50ff) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1609,7 +1609,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90b1-5a517a6e3dc96a1f1bf6605d;f47cdfc3-c109-4c9c-85d4-9fdb4b07968b) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe25d-22e570ad24f6166434159d33;cb1e02a0-bc28-4df2-9c42-97c96653357c) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -1736,7 +1736,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9614-04336de0684ae89e579d2595;d247b014-ddd7-4ad2-b541-f0f06d59c992) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe7e3-6e090bba374634442728bdd9;f37345d4-5e3d-4bee-b7d5-67d913ff2f6e) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1809,7 +1809,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e912a-2f9867c81bdb86316e10c233;5aeeba98-52c6-4365-a9cc-f4164aa149a6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe2de-116b068527e3277c5de789b2;e2ac5f4b-c20b-4d2c-afcb-2d70e68ac2a1) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1868,10 +1868,10 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpygas2d7b/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpv2mmpztp/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1910,7 +1910,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4633-44e7eda25cd46d70467e35e0;c4416f25-ed48-4a56-9f83-569e67477d5c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe956-647ae9e467a8e02949da1b99;b0c7b1aa-f073-4e1f-933a-8e3f80529fc5) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1941,7 +1941,7 @@ OSError: B is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -1967,7 +1967,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpnr7aenwu/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpc95vbk_9/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5946.503168,7053.246464,0.0,6406.79936,6098.60864,s,10,4.370441192626953,0.4370441192626952,0.0009281353036040856,0.43696878051757815,0.4381058013916016,0.4383167251586914,0.4384854641723633,"[0.4385276489257813, 0.4359717712402344, 0.43771954345703123, 0.43744598388671874, 0.4380589294433594, 0.4362575378417969, 0.43786093139648435, 0.4364915771484375, 0.436230712890625, 0.43587655639648437]",tokens/s,585.7532196792366,kWh,5.148596204997261e-06,2.8211948946868096e-06,2.4996191494520076e-05,3.296598259420415e-05,tokens/kWh,7765580.754902423,MB,5946.503168,7055.343616,0.0,6408.896512,6157.952,s,10,244.94424023437497,24.494424023437496,0.005272769015065994,24.493744140625,24.502037109375,24.502107421875,24.502163671875,"[24.487064453125, 24.502177734375, 24.49926953125, 24.4968515625, 24.502021484375, 24.493955078125, 24.493533203125, 24.4922109375, 24.487486328125, 24.489669921875]",tokens/s,2.5720139383444343,kWh,0.00028916327015789236,0.00015848618370445323,0.00135038218658508,0.0017980316404474258,tokens/kWh,35038.315557296286,,s,629,248.48397949218753,0.3950460723246224,0.05195367916433446,0.388748291015625,0.38915277099609374,0.38942371215820315,0.8254186791992187,"[0.38863052368164064, 0.3885834350585938, 0.38854656982421876, 0.38854452514648435, 0.3885762634277344, 0.3887677307128906, 0.3885189208984375, 0.3887728576660156, 0.3885434875488281, 0.38884658813476564, 0.38842572021484373, 0.3884328918457031, 0.3885711364746094, 0.3883417663574219, 0.3885240173339844, 0.38941387939453126, 0.3884031982421875, 0.3887196044921875, 0.38854452514648435, 0.3887032470703125, 0.3886264343261719, 0.38861721801757815, 0.3888977966308594, 0.3890206604003906, 0.388917236328125, 0.3884503173828125, 0.388843505859375, 0.3885404052734375, 0.38881793212890625, 0.38851788330078124, 0.3882977294921875, 0.3884666748046875, 0.388701171875, 0.38850970458984374, 0.3885455322265625, 0.3883397216796875, 0.3884103698730469, 0.388495361328125, 0.3894988708496094, 0.3882915954589844, 0.3888158569335938, 0.3886602783203125, 0.3886519775390625, 0.3887790222167969, 0.3884134521484375, 0.38843392944335936, 0.3885281372070313, 0.3888619384765625, 0.38980096435546874, 0.38914764404296875, 0.3887196044921875, 0.38860186767578125, 0.38848306274414063, 0.38838375854492185, 0.3885035400390625, 0.3884605407714844, 0.38845645141601565, 0.38830081176757814, 0.3884513244628906, 0.3889899597167969, 0.3887943725585937, 0.3889407958984375, 0.8262952880859376, 0.38884658813476564, 0.38921319580078123, 0.388748291015625, 0.3887923278808594, 0.38922137451171873, 0.3888977966308594, 0.3885906066894531, 0.3892326354980469, 0.38878106689453124, 0.38923776245117186, 0.38898483276367185, 0.3888721923828125, 0.38877694702148435, 0.38888754272460935, 0.3888670654296875, 0.38904830932617185, 0.3887943725585937, 0.38909030151367185, 0.38900018310546874, 0.3890636901855469, 0.3890831298828125, 0.3892695007324219, 0.38899404907226565, 0.39072256469726563, 0.3887984619140625, 0.3887677307128906, 0.3885977478027344, 0.38868377685546873, 0.38868069458007815, 0.38871551513671876, 0.3887615966796875, 0.38870834350585937, 0.38858139038085937, 0.388885498046875, 0.3888230285644531, 0.3896135559082031, 0.39009588623046876, 0.3886540832519531, 0.38893875122070315, 0.3886039123535156, 0.3890165710449219, 0.3886899108886719, 0.38870529174804686, 0.38863665771484374, 0.3886868591308594, 0.38889166259765623, 0.3890595703125, 0.38893771362304685, 0.3889479675292969, 0.388917236328125, 0.38900222778320315, 0.38909951782226565, 0.38838067626953127, 0.388495361328125, 0.3884666748046875, 0.389074951171875, 0.38862539672851565, 0.38855474853515626, 0.3885578308105469, 0.388843505859375, 0.389080078125, 0.3892039794921875, 0.8267151489257812, 0.3888455810546875, 0.389074951171875, 0.38913229370117186, 0.38876263427734375, 0.3887595520019531, 0.3888875732421875, 0.38918243408203124, 0.3892049865722656, 0.38914047241210936, 0.3887319030761719, 0.38899713134765623, 0.38865509033203127, 0.38901556396484377, 0.3886766052246094, 0.38860800170898435, 0.388864013671875, 0.38877593994140625, 0.3888302001953125, 0.388790283203125, 0.3885578308105469, 0.38857318115234374, 0.3885926513671875, 0.3890616455078125, 0.3892623291015625, 0.3890933837890625, 0.38865817260742186, 0.3889070129394531, 0.38874929809570313, 0.38892340087890626, 0.3887185974121094, 0.3888066711425781, 0.38876568603515627, 0.3885875244140625, 0.38916607666015623, 0.38873394775390624, 0.3885383605957031, 0.38869403076171877, 0.38857217407226563, 0.38951425170898435, 0.38901556396484377, 0.38918759155273436, 0.3889551391601562, 0.3887442016601563, 0.3890032653808594, 0.38889675903320314, 0.3889725341796875, 0.3889837951660156, 0.38876568603515627, 0.38967501831054685, 0.3893606262207031, 0.38946200561523436, 0.3886407775878906, 0.38881381225585937, 0.38867251586914064, 0.38852505493164063, 0.38857217407226563, 0.38876568603515627, 0.38846771240234373, 0.38870529174804686, 0.3890585632324219, 0.38904730224609374, 0.38865716552734375, 0.8268502807617187, 0.3887595520019531, 0.3887820739746094, 0.38875750732421877, 0.3886417846679687, 0.38880767822265627, 0.38890087890625, 0.38849432373046877, 0.38883224487304685, 0.3887718505859375, 0.38920294189453125, 0.3890390930175781, 0.3896995849609375, 0.38909747314453125, 0.3887442016601563, 0.388569091796875, 0.3890595703125, 0.3884718017578125, 0.38869195556640623, 0.38880563354492187, 0.38886605834960936, 0.38868377685546873, 0.3896105041503906, 0.38871450805664065, 0.3889029235839844, 0.38878311157226564, 0.3886878662109375, 0.38854757690429687, 0.38863873291015627, 0.3887370300292969, 0.38861004638671875, 0.38876263427734375, 0.3888015441894531, 0.3884984436035156, 0.38875546264648436, 0.38851174926757814, 0.388959228515625, 0.38882406616210935, 0.3891138610839844, 0.38870733642578126, 0.38867251586914064, 0.38911282348632814, 0.38889166259765623, 0.3884820556640625, 0.38871450805664065, 0.3886540832519531, 0.388600830078125, 0.389032958984375, 0.389180419921875, 0.38953677368164064, 0.38902886962890626, 0.3890667419433594, 0.389001220703125, 0.3887820739746094, 0.38870016479492187, 0.3886346130371094, 0.3887544250488281, 0.38863873291015627, 0.38883328247070315, 0.3885629577636719, 0.3887093811035156, 0.38889675903320314, 0.38913946533203125, 0.8280023193359375, 0.38870834350585937, 0.38864794921875, 0.3887124328613281, 0.388748291015625, 0.3888793640136719, 0.3885383605957031, 0.38862744140625, 0.38867352294921875, 0.3893258361816406, 0.3888793640136719, 0.38871551513671876, 0.38890802001953123, 0.38877490234375, 0.38909747314453125, 0.3890729064941406, 0.3887790222167969, 0.38876876831054685, 0.3889479675292969, 0.38905447387695313, 0.388927490234375, 0.38944769287109376, 0.3890114440917969, 0.3895459899902344, 0.3894302673339844, 0.3890831298828125, 0.38988287353515627, 0.388738037109375, 0.3887667236328125, 0.38885888671875, 0.3889950866699219, 0.3887370300292969, 0.3887564697265625, 0.38875546264648436, 0.3886991271972656, 0.389185546875, 0.38876058959960935, 0.38907086181640627, 0.38890188598632813, 0.3888619384765625, 0.38933401489257813, 0.388853759765625, 0.38896026611328127, 0.38860595703125, 0.38862744140625, 0.38865817260742186, 0.388463623046875, 0.3885875244140625, 0.3883212890625, 0.3891998596191406, 0.3895500793457031, 0.3895326843261719, 0.389032958984375, 0.3888066711425781, 0.388822021484375, 0.388811767578125, 0.38866329956054685, 0.38884146118164065, 0.38888958740234375, 0.38867147827148435, 0.38877490234375, 0.3891199951171875, 0.38873599243164064, 0.8254505004882813, 0.38886605834960936, 0.3887943725585937, 0.3888025512695312, 0.38903604125976565, 0.38877694702148435, 0.38867864990234374, 0.3889029235839844, 0.3888517150878906, 0.38892645263671877, 0.38902374267578127, 0.38975079345703123, 0.389823486328125, 0.389465087890625, 0.3888455810546875, 0.3890862121582031, 0.38899200439453124, 0.388611083984375, 0.3885342712402344, 0.3887175598144531, 0.38892340087890626, 0.38893875122070315, 0.3891138610839844, 0.38856704711914064, 0.3884707946777344, 0.3899740295410156, 0.388389892578125, 0.388780029296875, 0.3886213073730469, 0.3885793151855469, 0.3890093994140625, 0.38884454345703123, 0.3886991271972656, 0.388748291015625, 0.3889151916503906, 0.38854757690429687, 0.3891988525390625, 0.3890667419433594, 0.388822021484375, 0.38893670654296875, 0.3886960754394531, 0.38903399658203125, 0.38867251586914064, 0.38844415283203126, 0.3887595520019531, 0.3886796875, 0.3885424499511719, 0.3885875244140625, 0.3885660095214844, 0.388421630859375, 0.38844927978515625, 0.3885557861328125, 0.3888015441894531, 0.3884236755371094, 0.38832846069335936, 0.388347900390625, 0.388389892578125, 0.38839910888671875, 0.38838580322265626, 0.38835302734375, 0.3885363159179688, 0.38859982299804685, 0.38915277099609374, 0.8270069580078125, 0.388790283203125, 0.38852301025390623, 0.38860800170898435, 0.38870529174804686, 0.3886397399902344, 0.3888097229003906, 0.38888140869140625, 0.38900531005859373, 0.3891435546875, 0.38884454345703123, 0.38879129028320314, 0.38904934692382814, 0.38878924560546874, 0.38913946533203125, 0.38866329956054685, 0.3886970825195312, 0.3890196533203125, 0.38881076049804686, 0.38890087890625, 0.3889612731933594, 0.38866021728515626, 0.3888097229003906, 0.38903604125976565, 0.38920907592773435, 0.389138427734375, 0.38885888671875, 0.38871142578125, 0.38858648681640623, 0.3885793151855469, 0.38836428833007813, 0.38872164916992186, 0.3887953796386719, 0.3888721923828125, 0.389064697265625, 0.3893247985839844, 0.3885936584472656, 0.38854757690429687, 0.3885936584472656, 0.3884789733886719, 0.38885479736328127, 0.38876058959960935, 0.38854656982421876, 0.3885404052734375, 0.38906982421875, 0.38860287475585936, 0.38849945068359376, 0.38863665771484374, 0.38878106689453124, 0.3889336242675781, 0.38914559936523435, 0.38887322998046875, 0.3893893127441406, 0.3887462463378906, 0.38847796630859377, 0.38849127197265626, 0.38849740600585936, 0.3886684265136719, 0.3886510009765625, 0.38863565063476563, 0.38843597412109376, 0.3889039306640625, 0.38858853149414063, 0.8253368530273437, 0.38870220947265627, 0.38867864990234374, 0.3885823974609375, 0.3886929931640625, 0.3884881896972656, 0.38854656982421876, 0.38842059326171874, 0.38854757690429687, 0.38858648681640623, 0.388569091796875, 0.3887452087402344, 0.3891435546875, 0.38859982299804685, 0.3888302001953125, 0.38854144287109377, 0.3885936584472656, 0.38863052368164064, 0.38876980590820315, 0.3885404052734375, 0.38905447387695313, 0.38915277099609374, 0.38901556396484377, 0.38856704711914064, 0.3885404052734375, 0.38846157836914064, 0.38859469604492186, 0.38851071166992185, 0.38895001220703124, 0.38856805419921875, 0.38871450805664065, 0.38837246704101563, 0.3885987854003906, 0.38849127197265626, 0.38861618041992185, 0.38849639892578125, 0.38856192016601565, 0.38879641723632813, 0.38922341918945313, 0.38905752563476564, 0.38980096435546874, 0.3889090576171875, 0.3889070129394531, 0.3885916137695313, 0.38894183349609374, 0.3886796875, 0.3886407775878906, 0.3885383605957031, 0.38921829223632814, 0.38887322998046875, 0.38906265258789063, 0.3887585144042969, 0.38897869873046875, 0.38905548095703124, 0.38911077880859374, 0.3886458740234375, 0.3885455322265625, 0.3886878662109375, 0.3890585632324219, 0.38870220947265627, 0.38884658813476564, 0.38864486694335937, 0.3889561462402344, 0.82583349609375, 0.3893094482421875, 0.38874725341796873, 0.38864697265625, 0.3890738525390625, 0.3888230285644531, 0.3885772705078125, 0.3886264343261719, 0.38845541381835935, 0.38850457763671875, 0.38880767822265627, 0.3886039123535156, 0.3885486145019531, 0.388485107421875, 0.3883673706054688, 0.3887646789550781, 0.3885353088378906, 0.388632568359375, 0.38862539672851565, 0.3885936584472656, 0.38859674072265626, 0.3886069641113281, 0.388485107421875, 0.38892440795898436, 0.38891622924804686, 0.38909133911132815, 0.388453369140625, 0.38876058959960935, 0.3886592102050781, 0.3885567932128906, 0.3885189208984375, 0.3885557861328125, 0.388716552734375, 0.3882700805664063, 0.38850457763671875, 0.3887790222167969, 0.38870733642578126, 0.3885281372070313, 0.38864382934570313, 0.3885977478027344, 0.38863565063476563, 0.3885660095214844, 0.38846463012695315, 0.38862234497070314, 0.3884236755371094, 0.3884984436035156, 0.388347900390625, 0.38830181884765624, 0.38828134155273436, 0.3885066223144531, 0.38897152709960936, 0.38908721923828127, 0.3887431640625, 0.3887175598144531, 0.3884800109863281, 0.38863665771484374, 0.39151205444335935, 0.3889510498046875, 0.38859674072265626, 0.38852301025390623, 0.38847796630859377, 0.3888824462890625, 0.38877593994140625, 0.824754150390625, 0.38834994506835935, 0.38857217407226563, 0.38853323364257814, 0.3888844909667969, 0.3888005065917969, 0.3891640319824219, 0.3885987854003906, 0.3888486328125, 0.38872781372070314, 0.38896435546875, 0.3892623291015625, 0.3890882568359375, 0.3891937255859375, 0.38884658813476564, 0.3889827880859375, 0.38853323364257814, 0.3885260925292969, 0.38840933227539065, 0.38836428833007813, 0.388558837890625, 0.38827621459960937, 0.3885209655761719, 0.3886376953125, 0.38875341796875, 0.3887790222167969, 0.38885888671875, 0.3887298583984375, 0.38882406616210935, 0.38858035278320313, 0.388927490234375, 0.3888988037109375, 0.38889166259765623, 0.38841650390625, 0.3884298095703125, 0.3886039123535156, 0.3884247131347656, 0.3889837951660156, 0.38898175048828126, 0.3884728393554687, 0.3891671142578125, 0.3885987854003906, 0.3886120910644531, 0.388411376953125, 0.38858139038085937, 0.3884390258789063, 0.3885076599121094, 0.388558837890625, 0.388738037109375, 0.38887322998046875, 0.38831716918945314, 0.3885014953613281, 0.3886868591308594, 0.38870220947265627, 0.38861721801757815, 0.38846258544921874, 0.3886673889160156, 0.3884789733886719, 0.38866943359375, 0.388358154296875, 0.3887667236328125, 0.38856500244140624, 0.388558837890625]",tokens/s,2.5313503159658475,, @@ -2010,7 +2010,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9454-448cb808587209f658c62e33;e3f2f902-a61f-4619-b833-916cd5900149) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe624-5e02ccc75e795647381d98aa;3c98b460-d173-489c-acbf-af620b6d912f) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2068,7 +2068,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp8rfm226j/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpvl5wtkub/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1279.791104,1019.74016,0.0,373.293056,277.410816,s,10,0.3806483840942383,0.03806483840942383,0.0010374051807092283,0.03839825630187988,0.038631001281738284,0.03909522857666015,0.03946661041259766,"[0.03955945587158203, 0.03852783966064453, 0.03837411117553711, 0.038422401428222654, 0.03847126388549805, 0.038292991638183595, 0.03834467315673828, 0.03843852615356445, 0.036084991455078125, 0.03613212966918945]",tokens/s,6725.366787229584,kWh,4.3617786259962035e-07,2.3900503699099743e-07,1.1732389459702955e-06,1.8484218455609134e-06,tokens/kWh,138496523.73174343,MB,1280.086016,1019.74016,0.0,373.293056,323.047424,s,10,23.712255126953128,2.371225512695313,0.003882883813537052,2.3718930664062503,2.376249291992188,2.3762570922851562,2.3762633325195313,"[2.362728515625, 2.37624755859375, 2.373291748046875, 2.367230224609375, 2.371970947265625, 2.371815185546875, 2.376264892578125, 2.37161865234375, 2.37223681640625, 2.3688505859375]",tokens/s,26.56854004931377,kWh,2.7777078536237154e-05,1.5222735153430559e-05,6.781520513363374e-05,0.00011081501882330147,tokens/kWh,568514.9961527847,,s,629,24.01549103546141,0.03818043089898478,0.004509392475924156,0.03761459350585938,0.03786444854736328,0.03806208114624023,0.07505829895019543,"[0.03638784027099609, 0.03659775924682617, 0.036880382537841795, 0.03663974380493164, 0.03670937728881836, 0.03717529678344727, 0.036736000061035154, 0.03655987167358398, 0.03629260635375976, 0.03701657485961914, 0.03663974380493164, 0.0363612174987793, 0.03665100860595703, 0.038844417572021485, 0.03839078521728516, 0.038013950347900394, 0.03768627166748047, 0.03772825622558594, 0.037766143798828124, 0.03749990463256836, 0.037585918426513674, 0.03749683380126953, 0.03755929565429687, 0.03760537719726562, 0.037743614196777346, 0.03752140808105469, 0.0375551986694336, 0.03762995147705078, 0.037495807647705076, 0.03751935958862305, 0.03787571334838867, 0.03770880126953125, 0.0376360969543457, 0.03753472137451172, 0.037599231719970705, 0.03759718322753906, 0.03856595230102539, 0.03801900863647461, 0.037601280212402347, 0.03762073516845703, 0.0376545295715332, 0.03752243041992188, 0.03753881454467774, 0.03760435104370117, 0.03775795364379883, 0.037550079345703126, 0.03786649703979492, 0.03831398391723633, 0.03769036865234375, 0.037601280212402347, 0.03755929565429687, 0.037498878479003905, 0.03746303939819336, 0.03752345657348633, 0.03794432067871094, 0.037713920593261716, 0.03760435104370117, 0.037495807647705076, 0.037591041564941405, 0.03748454284667969, 0.03763916778564453, 0.03788083267211914, 0.07681126403808594, 0.03763711929321289, 0.03831808090209961, 0.03748966217041016, 0.03763507080078125, 0.03758489608764649, 0.03759308624267578, 0.037579776763916016, 0.03772419357299805, 0.03755926513671875, 0.03763097763061524, 0.03747020721435547, 0.03756748962402344, 0.03761663818359375, 0.037901313781738284, 0.03767398452758789, 0.03793817520141601, 0.03772415924072266, 0.037601280212402347, 0.03751321411132812, 0.03760332870483398, 0.03770470428466797, 0.03760435104370117, 0.03767705535888672, 0.03810611343383789, 0.0376995849609375, 0.03758694458007812, 0.03785728073120117, 0.03757878494262695, 0.03765244674682617, 0.03786956787109375, 0.03756032180786133, 0.037838848114013675, 0.037773311614990236, 0.037525505065917966, 0.037884929656982425, 0.03775692749023438, 0.03761151885986328, 0.0378869743347168, 0.03772518539428711, 0.037797889709472655, 0.03774566268920899, 0.037759998321533206, 0.03785113525390625, 0.038040576934814455, 0.03770470428466797, 0.0377077751159668, 0.03755929565429687, 0.037599231719970705, 0.037661697387695314, 0.037553150177001955, 0.037823486328125, 0.03759820938110352, 0.03769651031494141, 0.037648384094238284, 0.03772313690185547, 0.03791667175292969, 0.03774259185791016, 0.03760435104370117, 0.03763711929321289, 0.03762483215332031, 0.03786444854736328, 0.03789516830444336, 0.07634636688232421, 0.03765350341796875, 0.03746713638305664, 0.03770470428466797, 0.03758796691894531, 0.03751116943359375, 0.03785318374633789, 0.037889022827148434, 0.03752345657348633, 0.0374774398803711, 0.037649345397949216, 0.03765760040283203, 0.037743614196777346, 0.0377149429321289, 0.03764329528808594, 0.03746198272705078, 0.037561344146728515, 0.037763072967529294, 0.03768832015991211, 0.03759823989868164, 0.037550048828125, 0.03745382308959961, 0.03772825622558594, 0.03768012619018555, 0.03773952102661133, 0.03763097763061524, 0.03759820938110352, 0.03765657424926758, 0.0381921272277832, 0.037905406951904294, 0.03770675277709961, 0.03752345657348633, 0.03779072189331055, 0.0376627197265625, 0.03760844802856445, 0.03748761749267578, 0.03757670211791992, 0.0375654411315918, 0.03750713729858399, 0.03755820846557617, 0.0375551986694336, 0.03763302230834961, 0.037645313262939455, 0.03746099090576172, 0.03763507080078125, 0.0376258544921875, 0.03769651031494141, 0.0375203857421875, 0.037610496520996094, 0.037612545013427735, 0.03768627166748047, 0.03968307113647461, 0.03810201644897461, 0.03744870376586914, 0.03748761749267578, 0.03760844802856445, 0.037705726623535156, 0.0375654411315918, 0.037533695220947266, 0.037482494354248046, 0.037713920593261716, 0.037579776763916016, 0.037495807647705076, 0.07634534454345702, 0.03747840118408203, 0.037424129486083986, 0.03769753646850586, 0.03747020721435547, 0.03760639953613281, 0.03746815872192383, 0.03759308624267578, 0.03759001541137695, 0.037582847595214845, 0.037572608947753904, 0.037525505065917966, 0.03752959823608398, 0.03775590515136719, 0.03758899307250976, 0.037459968566894535, 0.0377446403503418, 0.037495807647705076, 0.03755929565429687, 0.037438465118408204, 0.03756851196289063, 0.037602302551269534, 0.037515262603759765, 0.03751321411132812, 0.03760639953613281, 0.037541889190673826, 0.03759206390380859, 0.03758694458007812, 0.03766681671142578, 0.037548030853271484, 0.037512191772460936, 0.037582847595214845, 0.037493759155273435, 0.037449726104736326, 0.03746815872192383, 0.03757056045532227, 0.037823486328125, 0.037482494354248046, 0.037579776763916016, 0.03767091369628906, 0.03766886520385742, 0.03751424026489258, 0.03748147201538086, 0.037495807647705076, 0.03765555191040039, 0.03773132705688476, 0.03749273681640625, 0.03760639953613281, 0.03768217468261719, 0.037533695220947266, 0.037394432067871096, 0.03765350341796875, 0.03762278366088867, 0.03748966217041016, 0.03768729782104492, 0.0376514892578125, 0.03766985702514648, 0.03745280075073242, 0.037541889190673826, 0.03762073516845703, 0.03751731109619141, 0.037572608947753904, 0.03764223861694336, 0.07663206481933593, 0.037482494354248046, 0.03758694458007812, 0.03754598236083984, 0.03749785614013672, 0.03743641662597656, 0.03762790298461914, 0.03742105484008789, 0.03755929565429687, 0.037628929138183595, 0.03765350341796875, 0.03744768142700195, 0.03746918487548828, 0.03767500686645508, 0.03757670211791992, 0.03751222229003906, 0.03749884796142578, 0.03762995147705078, 0.03771596908569336, 0.03759718322753906, 0.03778252792358398, 0.03771596908569336, 0.03758182525634766, 0.03766886520385742, 0.0377262077331543, 0.03749683380126953, 0.0374917106628418, 0.037571582794189456, 0.03759513473510742, 0.03750912094116211, 0.037544960021972655, 0.0376627197265625, 0.03753574371337891, 0.037466110229492186, 0.03755212783813477, 0.03748044967651367, 0.03741798400878906, 0.03748761749267578, 0.03759414291381836, 0.03739030456542969, 0.03777740859985351, 0.03743129730224609, 0.037466110229492186, 0.037466110229492186, 0.03757056045532227, 0.03745177459716797, 0.03782451248168945, 0.03930624008178711, 0.03780198287963867, 0.037553150177001955, 0.037572608947753904, 0.037628929138183595, 0.03765350341796875, 0.03766579055786133, 0.03758796691894531, 0.03759718322753906, 0.03784396743774414, 0.038711296081542966, 0.03820236968994141, 0.037833728790283204, 0.03772931289672852, 0.03775075149536133, 0.03758796691894531, 0.07656038665771485, 0.03744153594970703, 0.037580799102783204, 0.03758489608764649, 0.03763507080078125, 0.03759308624267578, 0.03766681671142578, 0.0375470085144043, 0.0374835205078125, 0.03772313690185547, 0.037703678131103514, 0.037541889190673826, 0.0375551986694336, 0.037652511596679684, 0.037497825622558593, 0.03801190567016602, 0.037644287109375, 0.03753881454467774, 0.03759718322753906, 0.03752345657348633, 0.037541889190673826, 0.03757363128662109, 0.0374835205078125, 0.03762176132202148, 0.03808870315551758, 0.03783782577514649, 0.03767705535888672, 0.03754086303710937, 0.03700121688842774, 0.03764223861694336, 0.03752959823608398, 0.037765121459960936, 0.03765862274169922, 0.03768012619018555, 0.03785116958618164, 0.03752854537963867, 0.03746201705932617, 0.037591041564941405, 0.03764022445678711, 0.03770057678222656, 0.03760435104370117, 0.037569534301757815, 0.03757875061035156, 0.037550079345703126, 0.03768320083618164, 0.03784806442260742, 0.037806079864501956, 0.037579776763916016, 0.03758796691894531, 0.03765248107910156, 0.03765862274169922, 0.03774566268920899, 0.03764121627807617, 0.03772313690185547, 0.03784294509887695, 0.037817344665527344, 0.03783987045288086, 0.03761971282958984, 0.03763814544677734, 0.03765862274169922, 0.0377077751159668, 0.03765862274169922, 0.03762995147705078, 0.07627571105957032, 0.03767603302001953, 0.03746713638305664, 0.03762278366088867, 0.03759001541137695, 0.037689342498779296, 0.03778559875488281, 0.03764019012451172, 0.03788390350341797, 0.03751424026489258, 0.03749273681640625, 0.03770880126953125, 0.03754291152954101, 0.037703678131103514, 0.03769036865234375, 0.037574657440185545, 0.03785420989990235, 0.03765657424926758, 0.03755929565429687, 0.037599231719970705, 0.03741593551635742, 0.03765555191040039, 0.0375838737487793, 0.038059009552001956, 0.03752243041992188, 0.03781324768066406, 0.03762790298461914, 0.03782761764526367, 0.03764118576049805, 0.03759513473510742, 0.03786444854736328, 0.037617664337158206, 0.03751731109619141, 0.037566463470458986, 0.03778252792358398, 0.037596160888671876, 0.03751321411132812, 0.03806412887573242, 0.03936972808837891, 0.03786342239379883, 0.03775897598266602, 0.037601280212402347, 0.03762483215332031, 0.03776921463012695, 0.03782860946655273, 0.037449726104736326, 0.037705726623535156, 0.0375654411315918, 0.03755929565429687, 0.03753267288208008, 0.03763302230834961, 0.03753881454467774, 0.037580799102783204, 0.03768832015991211, 0.037574657440185545, 0.037531646728515625, 0.03764940643310547, 0.03754393768310547, 0.03761151885986328, 0.037515262603759765, 0.03761459350585938, 0.03788288116455078, 0.03977830505371094, 0.07646412658691407, 0.03738828659057617, 0.03792588806152344, 0.037748737335205076, 0.03746713638305664, 0.037897216796875, 0.037553150177001955, 0.03762176132202148, 0.037343231201171875, 0.037651454925537106, 0.037405696868896485, 0.0374835205078125, 0.03782758331298828, 0.03766681671142578, 0.03759001541137695, 0.0375838737487793, 0.0378081283569336, 0.03773235321044922, 0.03765760040283203, 0.03766886520385742, 0.03765964889526367, 0.03771187210083008, 0.03750707244873047, 0.03753472137451172, 0.037531646728515625, 0.03755110549926758, 0.03769343948364258, 0.03748659133911133, 0.037596160888671876, 0.03750707244873047, 0.03753881454467774, 0.037820415496826174, 0.03772415924072266, 0.03755929565429687, 0.037553150177001955, 0.03769036865234375, 0.03775897598266602, 0.037594112396240234, 0.03757875061035156, 0.03759513473510742, 0.03752755355834961, 0.03780505752563477, 0.03753267288208008, 0.03759820938110352, 0.03762790298461914, 0.03850447845458985, 0.03777225494384766, 0.03763916778564453, 0.03750297546386719, 0.037561344146728515, 0.037645313262939455, 0.03790848159790039, 0.037582847595214845, 0.037553150177001955, 0.037664768218994144, 0.03753574371337891, 0.03770675277709961, 0.03749075317382813, 0.03756332778930664, 0.03781324768066406, 0.03759718322753906, 0.037773311614990236, 0.037571582794189456, 0.07186124420166015, 0.037190654754638675, 0.03747225570678711, 0.03754908752441406, 0.037579742431640625, 0.03760639953613281, 0.03770982360839844, 0.03739136123657227, 0.03773440170288086, 0.0376360969543457, 0.03769241714477539, 0.03769241714477539, 0.03752345657348633, 0.03768627166748047, 0.03762688064575195, 0.03752140808105469, 0.037733375549316404, 0.03770675277709961, 0.037754878997802735, 0.03753472137451172, 0.03745792007446289, 0.03804876708984375, 0.03773747253417969, 0.037572608947753904, 0.03754291152954101, 0.03757056045532227, 0.03814297485351562, 0.03765350341796875, 0.03766988754272461, 0.03759001541137695, 0.03777536010742188, 0.03753472137451172, 0.037574657440185545, 0.03759718322753906, 0.03750707244873047, 0.03766886520385742, 0.03754393768310547, 0.03745280075073242, 0.03773952102661133, 0.03783168029785156, 0.03754086303710937, 0.03762688064575195, 0.037483551025390624, 0.03754902267456055, 0.03769036865234375, 0.03759718322753906, 0.037456897735595705, 0.03748761749267578, 0.038013950347900394, 0.037768192291259765, 0.03762073516845703, 0.03767705535888672, 0.037928958892822266, 0.03753779220581055, 0.03802828979492188, 0.037585918426513674, 0.039975936889648435, 0.03809791946411133, 0.03769343948364258, 0.037564414978027344, 0.03762278366088867, 0.03766988754272461, 0.03770470428466797, 0.07192781066894531, 0.03523993682861328, 0.03559731292724609, 0.038373374938964845, 0.03784908676147461, 0.03781017684936523, 0.03761663818359375, 0.037765121459960936, 0.037564414978027344, 0.037615615844726565, 0.03755417633056641, 0.038040576934814455, 0.03775692749023438, 0.03775692749023438, 0.037582847595214845, 0.037716991424560545, 0.03759308624267578, 0.03775283050537109, 0.03753062438964844, 0.03760435104370117, 0.03752447891235351, 0.037569534301757815, 0.03765964889526367, 0.03779891204833984, 0.03758182525634766, 0.037698558807373043, 0.037585918426513674, 0.03753062438964844, 0.03767398452758789, 0.037664768218994144, 0.0376627197265625, 0.037787647247314454, 0.03765657424926758, 0.037833728790283204, 0.037781505584716796, 0.037684223175048825, 0.037719039916992186, 0.03794944000244141, 0.037571582794189456, 0.03768627166748047, 0.037579776763916016, 0.037574657440185545, 0.03738521575927734, 0.03818188858032227, 0.0376627197265625, 0.03749990463256836, 0.037795841217041014, 0.03761151885986328, 0.037803009033203126, 0.0377077751159668, 0.03760435104370117, 0.03778867340087891, 0.037615615844726565, 0.03751731109619141, 0.03779379272460937, 0.03823820877075195, 0.03774054336547852, 0.03762688064575195, 0.03750400161743164, 0.037553150177001955, 0.037956607818603515, 0.03771289443969727, 0.037722110748291016]",tokens/s,26.191427819286087,, @@ -2232,7 +2232,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9404-3587ba101c78e553145b7941;9ee5f8d3-7fd1-4bc0-930a-9e6fe1f28919) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe5c8-54665ba708d8e01455b008b0;fd4e5e09-e52f-4879-abf6-c2c9128750a1) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2263,7 +2263,7 @@ OSError: m is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2302,7 +2302,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46f7-232df3e435d4184b7657b9f4;0b0c7fc6-a9ba-4f99-95d6-64a0e4bb93b2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fea2f-3d6b0c036d33d37444aa4ea4;56d5a2e2-16c4-4742-9f19-0c8baa25497e) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2333,7 +2333,7 @@ OSError: 1 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2372,7 +2372,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46a9-0052fe5c40b74b847994d5ee;41ac9e3e-84d3-40ec-8c5e-5f8b89239aa0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe9d1-0a51856b63b1aebc3e218cfb;2865687d-bc70-452a-af9c-7b78b38a63b6) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2404,7 +2404,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1984.876544,5989.9904,0.0,5343.543296,5028.441088,s,10,3.1335410461425783,0.31335410461425783,0.00016607950218654725,0.3133529968261719,0.31348611145019534,0.31359719696044924,0.31368606536865234,"[0.31370828247070315, 0.31336306762695315, 0.31337255859375, 0.3133429260253906, 0.31346142578125, 0.3132883911132813, 0.3129918212890625, 0.31331317138671877, 0.3133724670410156, 0.3133269348144531]",tokens/s,816.9671187653299,kWh,3.7004872897846836e-06,2.027674437186988e-06,1.9071647201746034e-05,2.4799808928717706e-05,tokens/kWh,10322660.175964376,MB,1984.876544,5989.9904,0.0,5343.543296,5239.77216,s,10,173.81857031250001,17.38185703125,0.0016428170344984726,17.382506835937498,17.383453710937502,17.38379912109375,17.384075449218752,"[17.38330078125, 17.380638671875, 17.379466796875, 17.379962890625, 17.38414453125, 17.379732421875, 17.3823359375, 17.38293359375, 17.383376953125, 17.382677734375]",tokens/s,3.624468886536999,kWh,0.0002051448724760364,0.00011243641709052894,0.00104923167966246,0.0013668129692290254,tokens/kWh,46092.62672970995,,s,629,176.36203536987298,0.2803847939107679,0.037244301398409944,0.2758819580078125,0.2761283325195313,0.2762205078125,0.589428603515625,"[0.2761820068359375, 0.2757201843261719, 0.27577255249023436, 0.27555621337890623, 0.2757478332519531, 0.2757447814941406, 0.27579086303710937, 0.27595059204101563, 0.27581134033203125, 0.27607244873046877, 0.2759741516113281, 0.275857421875, 0.2758164367675781, 0.2757877807617187, 0.2759137268066406, 0.2758492126464844, 0.27594342041015624, 0.27582977294921873, 0.27569561767578127, 0.2758164367675781, 0.27580416870117186, 0.2757386169433594, 0.27586868286132815, 0.27584613037109373, 0.2762127380371094, 0.27622296142578123, 0.275810302734375, 0.27572940063476564, 0.2759485473632812, 0.2759763488769531, 0.275898193359375, 0.27574169921875, 0.27577139282226565, 0.27591680908203126, 0.27613592529296876, 0.27610623168945314, 0.2760693664550781, 0.2760191955566406, 0.27601715087890627, 0.2756700134277344, 0.27597006225585935, 0.27576025390625, 0.27574667358398436, 0.27606219482421873, 0.27612057495117187, 0.2758451232910156, 0.2759208984375, 0.2757877807617187, 0.27588711547851563, 0.2758451232910156, 0.2758226013183594, 0.27603045654296876, 0.27588507080078123, 0.2760099792480469, 0.275977294921875, 0.27598635864257814, 0.276126708984375, 0.2760325012207031, 0.27589016723632814, 0.2759362487792969, 0.27633560180664063, 0.2762076110839844, 0.589686767578125, 0.27579296875, 0.2758912048339844, 0.2755696105957031, 0.27566488647460935, 0.27578164672851563, 0.27568658447265626, 0.27567391967773436, 0.27574270629882813, 0.27591476440429685, 0.2757673034667969, 0.27571923828125, 0.27575289916992185, 0.27581951904296875, 0.27591476440429685, 0.27598641967773435, 0.27602740478515625, 0.275999755859375, 0.2758871765136719, 0.2757139892578125, 0.2757969970703125, 0.2758831176757813, 0.27586141967773437, 0.27572940063476564, 0.2759465026855469, 0.2758922119140625, 0.2758031921386719, 0.2759197998046875, 0.2758338623046875, 0.2759362487792969, 0.2759137268066406, 0.27588299560546875, 0.27584109497070314, 0.2757446899414063, 0.27593112182617185, 0.2759884948730469, 0.2759075927734375, 0.27627212524414063, 0.27612265014648435, 0.2758819580078125, 0.27578469848632814, 0.2760181884765625, 0.2756884460449219, 0.2756751403808594, 0.27570791625976565, 0.2758963317871094, 0.275962890625, 0.27599258422851564, 0.27579290771484377, 0.27598028564453125, 0.2757253112792969, 0.27567828369140623, 0.2760723876953125, 0.27621682739257813, 0.27621682739257813, 0.2760048522949219, 0.27609600830078124, 0.2760130615234375, 0.2758963317871094, 0.2758092651367188, 0.27579290771484377, 0.27617587280273437, 0.27585638427734377, 0.5895167846679688, 0.2755778503417969, 0.2755225524902344, 0.2756474914550781, 0.27567718505859373, 0.2757580871582031, 0.2758758544921875, 0.275894287109375, 0.27570791625976565, 0.2759065551757813, 0.27611953735351563, 0.27581234741210936, 0.2757314453125, 0.2757550048828125, 0.27572634887695313, 0.27565890502929685, 0.2756616516113281, 0.27558297729492187, 0.2756833190917969, 0.2757171325683594, 0.2758512573242187, 0.27572940063476564, 0.27575091552734377, 0.2757437438964844, 0.2757099609375, 0.27583181762695314, 0.2759813232421875, 0.27583282470703124, 0.275989501953125, 0.2758287353515625, 0.27593728637695314, 0.27578179931640623, 0.2758705749511719, 0.2757949523925781, 0.27586358642578124, 0.27575906372070313, 0.27582977294921873, 0.2758369140625, 0.2759813232421875, 0.2758338623046875, 0.27582769775390625, 0.27598855590820315, 0.27586758422851565, 0.275989501953125, 0.27594241333007813, 0.2760263671875, 0.275962890625, 0.2760714111328125, 0.27587994384765624, 0.27599871826171873, 0.27599871826171873, 0.27596697998046876, 0.27598028564453125, 0.2760345458984375, 0.27599563598632815, 0.27605401611328123, 0.27608575439453126, 0.2759219055175781, 0.27582464599609374, 0.2757857360839844, 0.2760765380859375, 0.2763509826660156, 0.275989501953125, 0.5894297485351563, 0.27584515380859376, 0.2756576843261719, 0.2755870666503906, 0.27570791625976565, 0.2757959594726562, 0.2756833190917969, 0.27583486938476565, 0.27573556518554687, 0.2757662658691406, 0.27571405029296875, 0.2756496887207031, 0.27601290893554686, 0.2758656005859375, 0.275873779296875, 0.27624551391601565, 0.27586456298828127, 0.2756526184082031, 0.2758502502441406, 0.27597311401367186, 0.27586050415039065, 0.27589117431640625, 0.2757529602050781, 0.27563827514648437, 0.27569561767578127, 0.27572122192382814, 0.2756976623535156, 0.2758430786132812, 0.2757335510253906, 0.27578875732421876, 0.27579290771484377, 0.275736572265625, 0.275926025390625, 0.2760550537109375, 0.27593112182617185, 0.27577035522460935, 0.2759977111816406, 0.27599359130859374, 0.27578469848632814, 0.27566900634765623, 0.2758645935058594, 0.27597943115234375, 0.2758408203125, 0.2757447814941406, 0.2761717834472656, 0.27593215942382815, 0.27588815307617187, 0.27595159912109374, 0.27595162963867187, 0.2760038452148437, 0.2761451416015625, 0.276168701171875, 0.27605810546875, 0.275915771484375, 0.2760796203613281, 0.27607553100585935, 0.27600076293945314, 0.27597311401367186, 0.2757519226074219, 0.27574169921875, 0.2759772033691406, 0.27622503662109377, 0.2758954162597656, 0.5895504760742187, 0.2757806396484375, 0.27574270629882813, 0.2755809020996094, 0.2756556701660156, 0.2758553466796875, 0.27582977294921873, 0.2757611389160156, 0.2757662658691406, 0.27569271850585936, 0.27588079833984375, 0.2757529602050781, 0.27588812255859374, 0.2761379699707031, 0.27587481689453125, 0.275842041015625, 0.275810302734375, 0.276291748046875, 0.2757876281738281, 0.2760099792480469, 0.27596798706054687, 0.27568960571289064, 0.2761757507324219, 0.27643392944335937, 0.276031494140625, 0.27558212280273436, 0.27611221313476564, 0.27607858276367186, 0.2758093566894531, 0.27569964599609376, 0.2758922119140625, 0.27593011474609375, 0.2758440856933594, 0.27588507080078123, 0.27579290771484377, 0.27576934814453125, 0.2757867431640625, 0.2758184814453125, 0.2761134033203125, 0.27586868286132815, 0.27572341918945314, 0.2761408996582031, 0.27610418701171874, 0.27574685668945315, 0.27607040405273436, 0.2760775451660156, 0.2758922119140625, 0.2760028076171875, 0.2758963317871094, 0.276030517578125, 0.27591058349609376, 0.2758502502441406, 0.27591064453125, 0.2759444580078125, 0.2758912048339844, 0.27591476440429685, 0.2760570983886719, 0.276738037109375, 0.27606533813476564, 0.27590652465820314, 0.2762925720214844, 0.276173828125, 0.2762342529296875, 0.5897278442382813, 0.27554098510742187, 0.27550811767578126, 0.27563723754882813, 0.2758369140625, 0.2758635559082031, 0.2757877807617187, 0.2758359069824219, 0.275704833984375, 0.27570175170898437, 0.27571917724609374, 0.27570687866210936, 0.27575704956054686, 0.2761584777832031, 0.2761574401855469, 0.2758307800292969, 0.2756147155761719, 0.27577651977539064, 0.2757969970703125, 0.27572735595703124, 0.2758666687011719, 0.2757724609375, 0.2758092041015625, 0.27582977294921873, 0.2756864013671875, 0.2757314453125, 0.2757181396484375, 0.27575924682617187, 0.27573336791992187, 0.275800048828125, 0.27577752685546875, 0.27585433959960937, 0.2759710693359375, 0.27594033813476565, 0.275999755859375, 0.2761985778808594, 0.27582138061523437, 0.275778564453125, 0.2763182067871094, 0.27580517578125, 0.27590347290039063, 0.27603662109375, 0.2758963317871094, 0.2759710693359375, 0.27588507080078123, 0.2759229736328125, 0.2758747863769531, 0.2758604736328125, 0.2758092651367188, 0.27608984375, 0.2761431884765625, 0.27591671752929686, 0.2758430786132812, 0.27598745727539065, 0.2759690246582031, 0.2759546813964844, 0.2761134033203125, 0.276063232421875, 0.27581439208984376, 0.275926025390625, 0.27579800415039063, 0.2760110168457031, 0.27595162963867187, 0.589216796875, 0.27562698364257815, 0.2756802673339844, 0.27580416870117186, 0.27592498779296876, 0.27604788208007813, 0.27590963745117186, 0.276082763671875, 0.2759833679199219, 0.27581637573242185, 0.275736572265625, 0.27580722045898437, 0.27586151123046876, 0.2759065551757813, 0.2757027893066406, 0.2757949523925781, 0.2757376098632813, 0.2758860778808594, 0.27602841186523436, 0.2759024658203125, 0.27579800415039063, 0.27583181762695314, 0.27587686157226565, 0.27673715209960936, 0.2757733154296875, 0.27583181762695314, 0.27580108642578127, 0.275800048828125, 0.2759126892089844, 0.2759957580566406, 0.27601397705078123, 0.2760673217773438, 0.275962890625, 0.27576422119140626, 0.2758727722167969, 0.2758164367675781, 0.2757949523925781, 0.27575704956054686, 0.27581747436523435, 0.27594342041015624, 0.27643902587890623, 0.27627212524414063, 0.27580526733398436, 0.2757590026855469, 0.27588507080078123, 0.2759137268066406, 0.275884033203125, 0.27603353881835935, 0.27581234741210936, 0.2759168701171875, 0.27608978271484375, 0.27585433959960937, 0.27584310913085935, 0.27578057861328126, 0.27595672607421873, 0.27586663818359375, 0.27584820556640627, 0.275800048828125, 0.2760386657714844, 0.27609906005859375, 0.2759393310546875, 0.276068359375, 0.2758584289550781, 0.5894256591796875, 0.2757519226074219, 0.27562905883789063, 0.27567922973632814, 0.275926025390625, 0.2760058898925781, 0.27571923828125, 0.275737548828125, 0.27581951904296875, 0.2757867431640625, 0.27557989501953123, 0.27591064453125, 0.27603558349609375, 0.27598236083984373, 0.2759444274902344, 0.27592498779296876, 0.2765977478027344, 0.2759024658203125, 0.27591476440429685, 0.2757949523925781, 0.2757334899902344, 0.2758748779296875, 0.2756525573730469, 0.27586764526367186, 0.27580621337890626, 0.2758922119140625, 0.275768310546875, 0.2757969970703125, 0.2758154296875, 0.27576217651367185, 0.2761134033203125, 0.27618304443359376, 0.27594342041015624, 0.2758114013671875, 0.2758677673339844, 0.27579885864257814, 0.2757120056152344, 0.27593728637695314, 0.27627316284179687, 0.27602740478515625, 0.27586773681640625, 0.27601507568359374, 0.27580307006835936, 0.27578982543945313, 0.27589324951171873, 0.27602841186523436, 0.275984375, 0.2762659912109375, 0.2760888366699219, 0.27614004516601565, 0.2759772033691406, 0.2761553955078125, 0.276178955078125, 0.2760140686035156, 0.2759884948730469, 0.27596188354492185, 0.2760160827636719, 0.27593331909179686, 0.2758255310058594, 0.27580416870117186, 0.27598028564453125, 0.27591680908203126, 0.2759393310546875, 0.5894686889648437, 0.27560858154296874, 0.2758042297363281, 0.2758890991210938, 0.27572122192382814, 0.27565158081054686, 0.27570687866210936, 0.2757386169433594, 0.27570892333984376, 0.27567718505859373, 0.2758983764648438, 0.27579290771484377, 0.27593829345703125, 0.2757734375, 0.27579904174804687, 0.27577035522460935, 0.2757867431640625, 0.27646975708007815, 0.27583999633789064, 0.27571505737304686, 0.2758973388671875, 0.27630694580078125, 0.27613592529296876, 0.2760110168457031, 0.2761154479980469, 0.27583181762695314, 0.2757969970703125, 0.27583282470703124, 0.27583181762695314, 0.27591168212890627, 0.27576934814453125, 0.2758462219238281, 0.27575399780273435, 0.2756771240234375, 0.2758359069824219, 0.2758389892578125, 0.27591168212890627, 0.27591064453125, 0.27612774658203126, 0.27606536865234377, 0.27584912109375, 0.2759342041015625, 0.27576217651367185, 0.27572735595703124, 0.27581134033203125, 0.27600897216796877, 0.2760663757324219, 0.27599148559570313, 0.2761246643066406, 0.275998779296875, 0.2759300537109375, 0.27614105224609375, 0.2763601989746094, 0.2759393310546875, 0.27599359130859374, 0.2758359069824219, 0.27598028564453125, 0.2767350769042969, 0.2761839599609375, 0.27600076293945314, 0.2762076110839844, 0.27613592529296876, 0.27586868286132815, 0.5895372924804687, 0.2757632446289062, 0.2756290283203125, 0.2755625, 0.2757171325683594, 0.27591680908203126, 0.27584613037109373, 0.2757356872558594, 0.27580813598632814, 0.2757386169433594, 0.2757887878417969, 0.2759915466308594, 0.2760028991699219, 0.2759228515625, 0.2758154296875, 0.27574169921875, 0.2757253112792969, 0.2758440856933594, 0.275962890625, 0.27582565307617185, 0.2758287353515625, 0.27589529418945313, 0.27577651977539064, 0.2758656005859375, 0.27606646728515627, 0.2761624145507813, 0.27685784912109374, 0.2756823120117188, 0.2758031311035156, 0.27605099487304685, 0.27588818359375, 0.27613067626953125, 0.27619744873046875, 0.2759935302734375, 0.275873779296875, 0.2757857360839844, 0.27572735595703124, 0.276126708984375, 0.2760447998046875, 0.2759004211425781, 0.2759741516113281, 0.27599462890625, 0.2757498779296875, 0.2757662658691406, 0.2758410339355469, 0.2759075927734375, 0.275857421875, 0.2758912048339844, 0.27594033813476565, 0.2759230041503906, 0.27591571044921875, 0.27589324951171873, 0.27589529418945313, 0.27597210693359375, 0.27605606079101563, 0.27605401611328123, 0.276168701171875, 0.2760570983886719, 0.27591168212890627, 0.2757959594726562, 0.27595672607421873, 0.27602944946289065, 0.2759638977050781]",tokens/s,3.566527221580528,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2472,7 +2472,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e96d2-43a574b123c5f91e42555463;fc2cf641-1f4f-4631-89b7-3825df3b5160) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe8a6-1d34900d627f1ba24224a6da;1fb8940e-b4f8-4a3c-a96a-3c00f670eb7f) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2532,7 +2532,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2571,7 +2571,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45fe-4cd77c8f4f4f9b135da7ca55;4b8c223f-cbef-4e99-8e64-3950ed38701b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe921-075ce5db770a44425edd7a74;83cad86c-2e87-419e-aa92-79092baf4278) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2781,7 +2781,7 @@ ChildProcessError: Traceback (most recent call last): TypeError: DeciCoderAttention.forward() got an unexpected keyword argument 'cache_position' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -2839,7 +2839,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90f5-5378bd5e3d9dfe2e6ebb8df9;e5e319ea-d0a8-418c-aba0-3b4b7092dc2e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe2a6-3b328ea24aa0bded58a41cb2;c114ad18-d410-41a3-a8b6-96dc782216d0) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -2947,7 +2947,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3044,7 +3044,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e96fa-389fdaa604b593d521cd76c6;131ef9a4-3297-4a04-b6ad-ef219fdf2456) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe8d0-683a3e377d7dfa117effd447;4c678107-8b15-4418-bb19-cd5325f773ae) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3104,7 +3104,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90cb-3211d2e53e13d9677563d606;ac80c622-f989-46ae-9240-002fa0dc64ec) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe279-138cee6f3009ff3457d023f7;a13cdf8e-855c-4087-bd54-feb44a489f4b) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -3629,7 +3629,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e8f37-2e73dcf538a9fa1241c36d80;d7a306d0-3b4f-4f66-bb73-310e3759da9f) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe0d9-6c2e9c7d5eadcbee200bb3cb;3ad66679-2e3a-4012-abde-94cb642f5eb4) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -3714,7 +3714,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9684-1f3565e31b0df31666fb8b27;71f42a52-5a4f-4247-a4cc-b330be432361) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe852-4deaae8e749e00be28026a11;a9562898-d78f-4de6-b8f0-af4a2e70df6b) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3858,7 +3858,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95b6-2a67ec8273249ef61dcbddc8;a634ef19-a2db-431d-bd9c-9ffab024000b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe785-5940803836739116398189ed;98a8c4c0-7157-4a86-b5a6-79d4d49eecd2) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3932,7 +3932,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1861.566464,3395.813376,0.0,2749.366272,2585.357312,s,10,0.732101089477539,0.0732101089477539,0.0009905213877066726,0.07305867004394531,0.07430353927612304,0.07468780326843262,0.07499521446228027,"[0.07394802856445312, 0.07507206726074218, 0.07251001739501953, 0.07249107360839843, 0.07266223907470704, 0.07207814025878906, 0.07345510101318359, 0.07379513549804688, 0.07421814727783203, 0.07187113952636719]",tokens/s,3496.7848522489344,kWh,8.519276714534592e-07,4.66812359389341e-07,4.015554771193711e-06,5.334294802036511e-06,tokens/kWh,47991348.34135247,MB,1861.566464,3395.813376,0.0,2749.366272,2628.768768,s,10,38.62193530273437,3.862193530273438,0.003503694519058953,3.8613565673828125,3.8657285400390626,3.8678885620117187,3.8696165795898434,"[3.870048583984375, 3.86524853515625, 3.86500830078125, 3.861310546875, 3.8626630859375, 3.861402587890625, 3.860045654296875, 3.859800048828125, 3.858201904296875, 3.8582060546875]",tokens/s,16.311973883799574,kWh,4.5592295868243344e-05,2.4984092023898362e-05,0.00021485331476820774,0.00028542970266034936,tokens/kWh,220719.84594738425,,s,629,39.21789340972901,0.062349592066341816,0.008754170010469397,0.06125347137451172,0.06140214309692383,0.06161048431396485,0.13420894287109375,"[0.06200627136230469, 0.062050304412841796, 0.062051326751708984, 0.062050304412841796, 0.062045185089111325, 0.06205952072143555, 0.062058494567871096, 0.06193766403198242, 0.061431808471679686, 0.06132940673828125, 0.061346817016601565, 0.06124851226806641, 0.06131001663208008, 0.06127609634399414, 0.06120550537109375, 0.06132633590698242, 0.06146662521362305, 0.061208576202392576, 0.06135193634033203, 0.06129971313476563, 0.06123110580444336, 0.06118707275390625, 0.061292545318603515, 0.06125568008422851, 0.06133350372314453, 0.061290496826171874, 0.06124652862548828, 0.061225921630859374, 0.061346912384033205, 0.061209503173828124, 0.061295616149902345, 0.06128947067260742, 0.061267967224121096, 0.06126182556152344, 0.06134374237060547, 0.06130483245849609, 0.06131302261352539, 0.061306880950927733, 0.06135603332519531, 0.06126182556152344, 0.06134783935546875, 0.06131097412109375, 0.06137241744995117, 0.06140518569946289, 0.06134783935546875, 0.0612853775024414, 0.06133657455444336, 0.06125875091552734, 0.061308097839355466, 0.06119721603393555, 0.061290401458740235, 0.061284351348876956, 0.06131100845336914, 0.061276126861572265, 0.0613171501159668, 0.061335521697998045, 0.06157120132446289, 0.06169484710693359, 0.06147891235351562, 0.06137855911254883, 0.06133555221557617, 0.06140723037719727, 0.13694259643554688, 0.06134783935546875, 0.06128639984130859, 0.061176830291748044, 0.06126182556152344, 0.06133760070800781, 0.06124851226806641, 0.06124748611450195, 0.061262847900390625, 0.06136524963378906, 0.061467647552490234, 0.0615208969116211, 0.06147686386108398, 0.06129663848876953, 0.06119014358520508, 0.06120755386352539, 0.06123212814331055, 0.06132633590698242, 0.061279232025146485, 0.06128742218017578, 0.0612935676574707, 0.06128025436401367, 0.061241344451904295, 0.06126489639282227, 0.06128332901000977, 0.061233150482177735, 0.06123519897460938, 0.061358081817626954, 0.061282302856445314, 0.06133046340942383, 0.061308895111083984, 0.06129663848876953, 0.06127206420898437, 0.06125568008422851, 0.06132428741455078, 0.06128639984130859, 0.06130483245849609, 0.06126694488525391, 0.06124031829833984, 0.06144204711914063, 0.06149631881713867, 0.06135500717163086, 0.06128646469116211, 0.06134777450561523, 0.06132428741455078, 0.06129663848876953, 0.06375430297851563, 0.06132217788696289, 0.06123519897460938, 0.06129971313476563, 0.061276161193847656, 0.061272224426269534, 0.06125347137451172, 0.06133964920043945, 0.06128339385986328, 0.06130579376220703, 0.06134483337402344, 0.06128326416015625, 0.061246463775634766, 0.061332576751708986, 0.06133852767944336, 0.061446144104003904, 0.061350910186767575, 0.13430271911621094, 0.06127001571655273, 0.061233150482177735, 0.06136012649536133, 0.06128947067260742, 0.06129459381103516, 0.06125056076049805, 0.06130793762207031, 0.06123107147216797, 0.061254657745361325, 0.0612935676574707, 0.06134067153930664, 0.06121779251098633, 0.06132940673828125, 0.06141030502319336, 0.06151679992675781, 0.06136217498779297, 0.061445121765136716, 0.06127523040771484, 0.061303775787353514, 0.06122387313842773, 0.061254657745361325, 0.06121062469482422, 0.062117889404296876, 0.06129971313476563, 0.061297664642333986, 0.061246463775634766, 0.06128742218017578, 0.06134080123901367, 0.06140198516845703, 0.061385726928710936, 0.06134988784790039, 0.06122598266601562, 0.06134067153930664, 0.06146662521362305, 0.06136422348022461, 0.06127308654785156, 0.06131097412109375, 0.06125260925292969, 0.06135398483276367, 0.061334560394287106, 0.06134473419189453, 0.06128639984130859, 0.06140927886962891, 0.06132940673828125, 0.061383678436279294, 0.06133760070800781, 0.062399486541748046, 0.06130995178222656, 0.061292545318603515, 0.061216766357421876, 0.061344768524169924, 0.061290496826171874, 0.061297664642333986, 0.061300735473632816, 0.06129971313476563, 0.06125158309936524, 0.06135193634033203, 0.06126489639282227, 0.06127001571655273, 0.061341697692871094, 0.06126489639282227, 0.06131609725952149, 0.1343057861328125, 0.06122393417358398, 0.06128332901000977, 0.06137446212768555, 0.061221889495849606, 0.061254657745361325, 0.06116352081298828, 0.06125875091552734, 0.0611512336730957, 0.06126492691040039, 0.06120751953125, 0.06121779251098633, 0.0611962890625, 0.06132633590698242, 0.061233150482177735, 0.061298686981201174, 0.06121779251098633, 0.061222911834716794, 0.06113391876220703, 0.06125968170166016, 0.06115020751953125, 0.0611778564453125, 0.061143039703369144, 0.06125875091552734, 0.06121779251098633, 0.06122905731201172, 0.06124649429321289, 0.06131824111938477, 0.061295486450195315, 0.061472766876220705, 0.061192192077636716, 0.06132326507568359, 0.06136832046508789, 0.061267967224121096, 0.06120550537109375, 0.06134067153930664, 0.06122905731201172, 0.06128742218017578, 0.061238273620605466, 0.06129971313476563, 0.06122905731201172, 0.061338623046875, 0.06120243072509766, 0.06127718353271484, 0.06223360061645508, 0.06153113555908203, 0.06150451278686524, 0.061418495178222655, 0.06119935989379883, 0.06135295867919922, 0.06118195343017578, 0.06127206420898437, 0.06126182556152344, 0.06143078231811523, 0.06123212814331055, 0.06126694488525391, 0.06118707275390625, 0.06128639984130859, 0.061200382232666016, 0.061346817016601565, 0.061312000274658204, 0.0612147216796875, 0.061295616149902345, 0.13448602294921874, 0.06163667297363281, 0.0612740478515625, 0.06121574401855469, 0.06132121658325195, 0.06135603332519531, 0.06117068862915039, 0.06136422348022461, 0.0611962890625, 0.06124544143676758, 0.061176864624023435, 0.061287391662597654, 0.06121779251098633, 0.06123110580444336, 0.06122598266601562, 0.06128947067260742, 0.06124236679077148, 0.06216089630126953, 0.06116659164428711, 0.06125875091552734, 0.06119833755493164, 0.06118297576904297, 0.06118809509277344, 0.06124860763549805, 0.06115318298339844, 0.06120243072509766, 0.061222911834716794, 0.06121881484985352, 0.06118707275390625, 0.061338623046875, 0.06123110580444336, 0.061228031158447264, 0.06126899337768555, 0.0612567024230957, 0.06119014358520508, 0.06126182556152344, 0.06132633590698242, 0.061192192077636716, 0.0611860466003418, 0.061295616149902345, 0.06120652770996094, 0.0612393913269043, 0.061224864959716796, 0.06126800155639649, 0.06119113540649414, 0.06131916809082031, 0.061198463439941404, 0.06121868896484375, 0.06128639984130859, 0.06151065444946289, 0.061228031158447264, 0.061271167755126955, 0.06120230484008789, 0.06123417663574219, 0.06121267318725586, 0.061297664642333986, 0.061246463775634766, 0.06344601440429687, 0.06122086334228516, 0.06132428741455078, 0.0612751350402832, 0.06124031829833984, 0.061357120513916015, 0.13396780395507812, 0.06137753677368164, 0.06133760070800781, 0.06117068862915039, 0.061176830291748044, 0.06124240112304687, 0.061154430389404296, 0.06114390563964844, 0.061158401489257816, 0.06123929595947265, 0.061139968872070315, 0.06118707275390625, 0.06120755386352539, 0.06121574401855469, 0.06114713668823242, 0.061330432891845706, 0.06128947067260742, 0.061262847900390625, 0.06122598266601562, 0.061257728576660155, 0.06117068862915039, 0.0612044792175293, 0.06126182556152344, 0.06124544143676758, 0.061168670654296875, 0.06120240020751953, 0.0611778564453125, 0.06120550537109375, 0.06122598266601562, 0.06124748611450195, 0.0611778564453125, 0.06125568008422851, 0.06140227127075195, 0.0613282241821289, 0.0612782096862793, 0.06129663848876953, 0.06118195343017578, 0.06134067153930664, 0.06349004745483398, 0.061344768524169924, 0.06118809509277344, 0.06125363159179688, 0.06121267318725586, 0.0612782096862793, 0.06121574401855469, 0.061282302856445314, 0.06121574401855469, 0.0612567024230957, 0.061241344451904295, 0.061473823547363284, 0.06134579086303711, 0.06122390365600586, 0.06124236679077148, 0.06125056076049805, 0.061193214416503904, 0.061243392944335937, 0.061246463775634766, 0.06125056076049805, 0.061238273620605466, 0.061230079650878906, 0.0612147216796875, 0.0612782096862793, 0.06139289474487305, 0.13551513671875, 0.061222911834716794, 0.061300865173339845, 0.06122995376586914, 0.06116864013671875, 0.061241344451904295, 0.06116147232055664, 0.06121574401855469, 0.061233150482177735, 0.06125056076049805, 0.061224960327148435, 0.061298686981201174, 0.0611860466003418, 0.061274112701416014, 0.0612782096862793, 0.06128742218017578, 0.06179743957519531, 0.061714366912841795, 0.06137036895751953, 0.06115532684326172, 0.06111948776245117, 0.06116659164428711, 0.061200382232666016, 0.06237392044067383, 0.06122288131713867, 0.06122412872314453, 0.061155136108398435, 0.061274112701416014, 0.0611778564453125, 0.061254657745361325, 0.061224960327148435, 0.06123110580444336, 0.06135500717163086, 0.06133452987670898, 0.061189121246337894, 0.061244415283203124, 0.06118611145019531, 0.06124025726318359, 0.06119935989379883, 0.061228031158447264, 0.061192192077636716, 0.061260799407958984, 0.06117068862915039, 0.06125888061523437, 0.061265792846679684, 0.061200382232666016, 0.06116150283813476, 0.06126895904541016, 0.061183998107910156, 0.06129459381103516, 0.061162494659423826, 0.061208576202392576, 0.0611962890625, 0.061230079650878906, 0.06120451354980469, 0.06117782211303711, 0.06115327835083008, 0.06122905731201172, 0.06116659164428711, 0.061230079650878906, 0.061257728576660155, 0.061224960327148435, 0.06124748611450195, 0.13553868103027344, 0.06118195343017578, 0.061238273620605466, 0.06116864013671875, 0.06116761779785156, 0.06121881484985352, 0.06114918518066406, 0.061135936737060546, 0.06112966537475586, 0.06120243072509766, 0.0611409912109375, 0.06120550537109375, 0.06121062469482422, 0.0611962890625, 0.06115327835083008, 0.061298686981201174, 0.061943809509277345, 0.06136524963378906, 0.06138470458984375, 0.06130588912963867, 0.061340641021728516, 0.06137241744995117, 0.06116966247558594, 0.06127206420898437, 0.061456382751464846, 0.06138777542114258, 0.061489280700683595, 0.061464447021484375, 0.06123110580444336, 0.061284351348876956, 0.06120652770996094, 0.061193214416503904, 0.061192192077636716, 0.06128025436401367, 0.06137036895751953, 0.06136627197265625, 0.06142771148681641, 0.061208576202392576, 0.06112768173217773, 0.06126185607910156, 0.06128329467773438, 0.06123110580444336, 0.06121574401855469, 0.06126182556152344, 0.06118707275390625, 0.061271041870117185, 0.061193214416503904, 0.06123417663574219, 0.061163551330566404, 0.06126399993896484, 0.06111932754516602, 0.061269153594970704, 0.061199199676513674, 0.06125980758666992, 0.061153247833251954, 0.061228031158447264, 0.061243392944335937, 0.06127308654785156, 0.06119833755493164, 0.061274112701416014, 0.06121062469482422, 0.06120550537109375, 0.06126182556152344, 0.13602406311035156, 0.06120140838623047, 0.061265918731689455, 0.06120550537109375, 0.06120755386352539, 0.06119731140136719, 0.06116761779785156, 0.06120345687866211, 0.0611256332397461, 0.061224960327148435, 0.06121881484985352, 0.061208576202392576, 0.06116556930541992, 0.061209632873535154, 0.06117071914672852, 0.06118393707275391, 0.061159423828125, 0.06195302581787109, 0.06117478561401367, 0.06131014251708984, 0.06121555328369141, 0.061200382232666016, 0.061189121246337894, 0.061211647033691405, 0.06113382339477539, 0.06121267318725586, 0.06122393417358398, 0.061211647033691405, 0.06117478561401367, 0.06120652770996094, 0.061213695526123046, 0.06119116973876953, 0.06121062469482422, 0.06128844833374023, 0.061402111053466796, 0.06131315231323242, 0.06120947265625, 0.061208576202392576, 0.06116556930541992, 0.061249534606933595, 0.06116147232055664, 0.06124236679077148, 0.061209632873535154, 0.06124848175048828, 0.06123734283447266, 0.061222816467285154, 0.06129059219360351, 0.06121871948242188, 0.06120243072509766, 0.06126387023925781, 0.061208576202392576, 0.06143283081054687, 0.06129459381103516, 0.061241344451904295, 0.061193214416503904, 0.06121984100341797, 0.06117171096801758, 0.061183998107910156, 0.06115737533569336, 0.061206687927246095, 0.06120636749267578, 0.06126694488525391, 0.06128947067260742, 0.1337159729003906, 0.06119833755493164, 0.061267967224121096, 0.061118465423583984, 0.06119731140136719, 0.06125363159179688, 0.06119116973876953, 0.06113689422607422, 0.06114918518066406, 0.06121779251098633, 0.061126655578613284, 0.06121068954467773, 0.0611736946105957, 0.06115020751953125, 0.061172737121582034, 0.0612147216796875, 0.06117375946044922, 0.061211647033691405, 0.06116352081298828, 0.061233150482177735, 0.06114508819580078, 0.06125875091552734, 0.06129971313476563, 0.061185054779052735, 0.061112289428710935, 0.061224960327148435, 0.06247423934936523, 0.06127414321899414, 0.06117987060546875, 0.061175807952880856, 0.06119935989379883, 0.06124448013305664, 0.061127616882324216, 0.06123929595947265, 0.06120550537109375, 0.061259777069091796, 0.06121779251098633, 0.06124236679077148, 0.061239326477050784, 0.06118502426147461, 0.06120547103881836, 0.06126489639282227, 0.061208576202392576, 0.06122905731201172, 0.061249534606933595, 0.06124236679077148, 0.061159423828125, 0.06130278396606445, 0.06115225601196289, 0.06123929595947265, 0.06117068862915039, 0.06123929595947265, 0.061241344451904295, 0.061357120513916015, 0.061324222564697266, 0.06140313720703125, 0.061230079650878906, 0.06125875091552734, 0.06115737533569336, 0.06119935989379883, 0.06123622512817383, 0.06115430450439453, 0.061200382232666016]",tokens/s,16.038597316498404,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -3991,7 +3991,7 @@ ImportError: This modeling file requires the following packages that were not fo ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1495.560192,1596.45696,0.0,950.009856,857.516544,s,10,0.5835707511901856,0.05835707511901857,0.00031171236882634643,0.0583786563873291,0.05868130149841309,0.05881489009857178,0.05892176097869873,"[0.05856409454345703, 0.058012256622314455, 0.0584984016418457, 0.05846012878417969, 0.058651615142822265, 0.05804627227783203, 0.05894847869873047, 0.05788921737670898, 0.0582031021118164, 0.058297183990478514]",tokens/s,4386.785997719917,kWh,6.829061891761589e-07,3.7418846269126377e-07,2.9153571813525047e-06,3.972451833219927e-06,tokens/kWh,64443827.3257792,MB,1495.560192,1596.45696,0.0,950.009856,888.373248,s,10,31.284889648437503,3.12848896484375,0.0114589981267836,3.1283529052734376,3.1444966552734375,3.1450495971679686,3.145491950683594,"[3.13634814453125, 3.115281494140625, 3.119911376953125, 3.136393310546875, 3.11410107421875, 3.132646728515625, 3.144373779296875, 3.12405908203125, 3.116172119140625, 3.1456025390625]",tokens/s,20.137517091464787,kWh,3.707178993826504e-05,2.0317029893987028e-05,0.00015490235240604643,0.00021229117223829852,tokens/kWh,296762.2220733795,,s,629,31.75902714920044,0.05049129912432502,0.0069426659511907,0.04940083312988281,0.050250752258300786,0.05062901916503906,0.10768740234375,"[0.05231513595581055, 0.05160345458984375, 0.05017190551757812, 0.05024256134033203, 0.050050048828125, 0.05151641464233398, 0.04965580749511719, 0.05403955078125, 0.05475328063964844, 0.05024563217163086, 0.04980428695678711, 0.04989235305786133, 0.049364990234375, 0.0493383674621582, 0.04931379318237305, 0.04937318420410156, 0.04939263916015625, 0.049358848571777345, 0.04939980697631836, 0.04938240051269531, 0.049672191619873046, 0.049539070129394534, 0.049344512939453126, 0.04935475158691406, 0.04934963226318359, 0.04931891250610351, 0.049522689819335934, 0.05018521499633789, 0.049339393615722656, 0.04933222579956055, 0.04935270309448242, 0.04942131042480469, 0.04936908721923828, 0.04948787307739258, 0.04943667221069336, 0.049882110595703126, 0.04938751983642578, 0.04944076919555664, 0.0493383674621582, 0.049337345123291014, 0.049358848571777345, 0.049345535278320314, 0.04937011337280273, 0.04935987091064453, 0.049391616821289064, 0.04938751983642578, 0.0493834228515625, 0.04934143829345703, 0.049342464447021485, 0.04940083312988281, 0.049360897064208986, 0.04937420654296875, 0.04940288162231445, 0.04937830352783203, 0.04984729766845703, 0.04938547134399414, 0.04934860610961914, 0.04936601638793945, 0.049353729248046874, 0.049393665313720705, 0.0493383674621582, 0.04930867385864258, 0.10750975799560547, 0.04935065460205078, 0.04933324813842774, 0.04929843139648438, 0.049293312072753906, 0.04932198333740234, 0.049410049438476565, 0.04968550491333008, 0.04969472122192383, 0.049876991271972655, 0.04935987091064453, 0.04935168075561523, 0.049329151153564454, 0.04934041595458984, 0.04936601638793945, 0.04931686401367188, 0.04940288162231445, 0.04935168075561523, 0.049334270477294925, 0.04972339248657227, 0.049358848571777345, 0.049342464447021485, 0.049320960998535154, 0.049320960998535154, 0.049329151153564454, 0.04931071853637695, 0.0493383674621582, 0.04930047988891602, 0.049375232696533204, 0.04935065460205078, 0.049334270477294925, 0.04937830352783203, 0.04933222579956055, 0.049345535278320314, 0.04934143829345703, 0.04934656143188477, 0.04940902328491211, 0.04937625503540039, 0.04938649749755859, 0.04937011337280273, 0.04934143829345703, 0.04936806488037109, 0.049360897064208986, 0.04933017730712891, 0.04935475158691406, 0.049334270477294925, 0.049358848571777345, 0.04938956832885742, 0.049334270477294925, 0.04938751983642578, 0.04947763061523437, 0.0494284782409668, 0.04937830352783203, 0.05029785537719727, 0.04992716979980469, 0.05004185485839844, 0.04987289428710937, 0.05022515106201172, 0.04950732803344727, 0.04943769454956055, 0.049324031829833984, 0.049342464447021485, 0.049882110595703126, 0.10801459503173828, 0.04929843139648438, 0.04931379318237305, 0.049314815521240236, 0.04931584167480469, 0.04934963226318359, 0.049326080322265625, 0.049360897064208986, 0.04932505416870117, 0.04939263916015625, 0.04934041595458984, 0.049364990234375, 0.04934963226318359, 0.04933324813842774, 0.049353729248046874, 0.049334270477294925, 0.049363967895507815, 0.049363967895507815, 0.049393665313720705, 0.04936703872680664, 0.04932505416870117, 0.049377281188964846, 0.050105342864990236, 0.05029785537719727, 0.050336769104003906, 0.04940492630004883, 0.04940083312988281, 0.04941721725463867, 0.049361919403076174, 0.04937932968139649, 0.04937625503540039, 0.04940185546875, 0.04936601638793945, 0.04934656143188477, 0.049364990234375, 0.04937830352783203, 0.049391616821289064, 0.04933324813842774, 0.04941721725463867, 0.04939980697631836, 0.04940492630004883, 0.049397758483886715, 0.04936908721923828, 0.04934656143188477, 0.04988518524169922, 0.0497151985168457, 0.04937420654296875, 0.04940902328491211, 0.05056409454345703, 0.049788928985595705, 0.04936703872680664, 0.04937625503540039, 0.04938751983642578, 0.049364990234375, 0.04935065460205078, 0.04957081604003906, 0.05004492950439453, 0.050049022674560545, 0.04997119903564453, 0.05030400085449219, 0.050100223541259765, 0.050097152709960936, 0.04945100784301758, 0.10798079681396484, 0.049688575744628906, 0.04928102493286133, 0.05053849411010742, 0.049363967895507815, 0.04932812881469727, 0.04932198333740234, 0.04979916763305664, 0.05087641525268555, 0.04971724700927734, 0.04935168075561523, 0.05140582275390625, 0.049631233215332034, 0.049377281188964846, 0.049331199645996096, 0.04936806488037109, 0.04941721725463867, 0.049363967895507815, 0.05114777755737305, 0.049432575225830076, 0.04938547134399414, 0.049549312591552735, 0.04955033493041992, 0.04931584167480469, 0.049858558654785154, 0.04995993423461914, 0.04933529663085937, 0.049391616821289064, 0.049535999298095705, 0.05039411163330078, 0.0502927360534668, 0.049617919921875, 0.05028966522216797, 0.05063679885864258, 0.050097152709960936, 0.04943155288696289, 0.04942233657836914, 0.051917823791503906, 0.050206718444824217, 0.05040639877319336, 0.049342464447021485, 0.0502210578918457, 0.050051071166992187, 0.050310142517089845, 0.05020467376708984, 0.05010227203369141, 0.04941312026977539, 0.04939878463745117, 0.04935168075561523, 0.049355777740478515, 0.04938444900512695, 0.04937932968139649, 0.049442817687988284, 0.04939980697631836, 0.04936908721923828, 0.04941516876220703, 0.049364990234375, 0.04942950439453125, 0.04940185546875, 0.04997836685180664, 0.049767425537109375, 0.050016254425048826, 0.05046783828735352, 0.10819379425048828, 0.04933017730712891, 0.049972225189208984, 0.049909759521484375, 0.05004288101196289, 0.04941823959350586, 0.049337345123291014, 0.04937830352783203, 0.049337345123291014, 0.04939878463745117, 0.049339393615722656, 0.04941516876220703, 0.049324031829833984, 0.049360897064208986, 0.049347583770751956, 0.04935168075561523, 0.04934963226318359, 0.049397758483886715, 0.049391616821289064, 0.04940902328491211, 0.04942131042480469, 0.049364990234375, 0.049345535278320314, 0.04933631896972656, 0.04935475158691406, 0.04940390396118164, 0.04935270309448242, 0.04930047988891602, 0.0493568000793457, 0.04939980697631836, 0.04935782241821289, 0.04941721725463867, 0.049342464447021485, 0.04935475158691406, 0.04936908721923828, 0.04933017730712891, 0.04935987091064453, 0.04935987091064453, 0.049344512939453126, 0.04958924865722656, 0.04940185546875, 0.0493834228515625, 0.04938854217529297, 0.049342464447021485, 0.04937113571166992, 0.04938240051269531, 0.04934860610961914, 0.04935270309448242, 0.04939468765258789, 0.04938444900512695, 0.04934963226318359, 0.049430526733398435, 0.049396736145019535, 0.049396736145019535, 0.04944179153442383, 0.04937830352783203, 0.049377281188964846, 0.04944384002685547, 0.04940492630004883, 0.049448959350585936, 0.049462272644042966, 0.04976947021484375, 0.0502743034362793, 0.10758860778808593, 0.0498524169921875, 0.04991078567504883, 0.04995276641845703, 0.049888256072998044, 0.04983705520629883, 0.049756160736083986, 0.05019340896606445, 0.04939263916015625, 0.049460224151611325, 0.049739776611328126, 0.04960358428955078, 0.04971724700927734, 0.04931379318237305, 0.04972032165527344, 0.049842174530029294, 0.05082726287841797, 0.04992921447753906, 0.04943769454956055, 0.05039513778686523, 0.049898494720458986, 0.0500469741821289, 0.05020774459838867, 0.04951347351074219, 0.049334270477294925, 0.04935475158691406, 0.04934860610961914, 0.04938649749755859, 0.0493199348449707, 0.0493199348449707, 0.050383872985839843, 0.04939059066772461, 0.04940083312988281, 0.04937011337280273, 0.04941619110107422, 0.049977344512939455, 0.05010739135742188, 0.050111488342285154, 0.050871295928955076, 0.0503900146484375, 0.04935168075561523, 0.05034598541259765, 0.049408000946044923, 0.04936703872680664, 0.04938751983642578, 0.04934656143188477, 0.049344512939453126, 0.049339393615722656, 0.0493383674621582, 0.049380352020263675, 0.04935270309448242, 0.049342464447021485, 0.04933631896972656, 0.049501182556152344, 0.050086910247802735, 0.04981043243408203, 0.049620990753173826, 0.05106687927246094, 0.05060812759399414, 0.04976639938354492, 0.04935782241821289, 0.04937011337280273, 0.04936908721923828, 0.1085450210571289, 0.049393665313720705, 0.04937113571166992, 0.04931891250610351, 0.05010739135742188, 0.049708030700683595, 0.04968447875976562, 0.049334270477294925, 0.04932505416870117, 0.04934860610961914, 0.049320960998535154, 0.049719295501708984, 0.049906688690185545, 0.04990873718261719, 0.05004083251953125, 0.0499681282043457, 0.05612748718261719, 0.05035007858276367, 0.04996710586547851, 0.049882110595703126, 0.05125120162963867, 0.049593345642089844, 0.04933529663085937, 0.04941516876220703, 0.049931262969970705, 0.04936703872680664, 0.04978585433959961, 0.05063270568847656, 0.04977459335327149, 0.04984012985229492, 0.04968447875976562, 0.04980326461791992, 0.050301952362060545, 0.04990771102905273, 0.05057843017578125, 0.05055897521972656, 0.05073100662231445, 0.05051903915405274, 0.050508800506591796, 0.04954111862182617, 0.04937625503540039, 0.04937420654296875, 0.05008588790893555, 0.049739776611328126, 0.049358848571777345, 0.049355777740478515, 0.050086910247802735, 0.04985343933105469, 0.04955955123901367, 0.04940083312988281, 0.04934041595458984, 0.04935782241821289, 0.049979393005371096, 0.0498790397644043, 0.04941516876220703, 0.049742847442626956, 0.049702911376953124, 0.05007360076904297, 0.05108224105834961, 0.049438720703125, 0.04939980697631836, 0.04979814529418945, 0.04958720016479492, 0.10789478302001954, 0.04976435089111328, 0.04958924865722656, 0.05006131362915039, 0.04940083312988281, 0.049323009490966796, 0.04935065460205078, 0.05004492950439453, 0.049926143646240234, 0.049912830352783204, 0.04959641647338867, 0.04933324813842774, 0.04933017730712891, 0.04934656143188477, 0.04933324813842774, 0.04937830352783203, 0.04995276641845703, 0.04990156936645508, 0.04940697479248047, 0.049498111724853515, 0.049345535278320314, 0.04936294555664063, 0.049301502227783206, 0.04935065460205078, 0.04938137435913086, 0.04965580749511719, 0.04968960189819336, 0.04995993423461914, 0.05003366470336914, 0.050113536834716796, 0.04981145477294922, 0.05002342224121094, 0.050050048828125, 0.05027123260498047, 0.05000703811645508, 0.050563072204589846, 0.04972339248657227, 0.04941107177734375, 0.04931891250610351, 0.04941312026977539, 0.04932812881469727, 0.04937932968139649, 0.049360897064208986, 0.04934041595458984, 0.04936703872680664, 0.049329151153564454, 0.04932710266113281, 0.04939980697631836, 0.049364990234375, 0.04937932968139649, 0.049358848571777345, 0.0506163215637207, 0.0494284782409668, 0.04937830352783203, 0.049710079193115236, 0.04936703872680664, 0.04936908721923828, 0.04934860610961914, 0.04936908721923828, 0.04931584167480469, 0.04931584167480469, 0.049396736145019535, 0.04930252838134765, 0.10772582244873047, 0.04937625503540039, 0.049344512939453126, 0.04930867385864258, 0.04931891250610351, 0.049342464447021485, 0.04934143829345703, 0.049377281188964846, 0.049334270477294925, 0.04939468765258789, 0.05003776168823242, 0.05007564926147461, 0.049783809661865235, 0.049353729248046874, 0.04938649749755859, 0.04937420654296875, 0.04935475158691406, 0.04942438507080078, 0.04941107177734375, 0.04940288162231445, 0.04943667221069336, 0.04938854217529297, 0.04933529663085937, 0.04938547134399414, 0.04940083312988281, 0.049353729248046874, 0.04933324813842774, 0.04938240051269531, 0.04935270309448242, 0.04931379318237305, 0.04938137435913086, 0.049977344512939455, 0.05000908660888672, 0.04952678298950195, 0.04935065460205078, 0.04938854217529297, 0.04945305633544922, 0.04934348678588867, 0.049355777740478515, 0.04933529663085937, 0.04937113571166992, 0.04939980697631836, 0.04935065460205078, 0.049364990234375, 0.04974694442749023, 0.04986880111694336, 0.049620990753173826, 0.049397758483886715, 0.0493568000793457, 0.049377281188964846, 0.049345535278320314, 0.050026496887207034, 0.04940288162231445, 0.04940697479248047, 0.04942335891723633, 0.04935068893432617, 0.0493792953491211, 0.049476608276367184, 0.04939263916015625, 0.04941823959350586, 0.04937318420410156, 0.04934143829345703, 0.049941505432128906, 0.10781798553466797, 0.05007564926147461, 0.050239486694335936, 0.049704959869384766, 0.049331199645996096, 0.049331199645996096, 0.04935270309448242, 0.049724414825439454, 0.050086910247802735, 0.05024153518676758, 0.04992409515380859, 0.04992512130737305, 0.05034188842773438, 0.05062348937988281, 0.0499159049987793, 0.04936908721923828, 0.04935987091064453, 0.049410049438476565, 0.049393665313720705, 0.049408000946044923, 0.0497151985168457, 0.04990771102905273, 0.049549312591552735, 0.05006335830688476, 0.050083839416503906, 0.05160140609741211, 0.050644992828369144, 0.04997529602050781, 0.049903617858886716, 0.05058560180664062, 0.051154945373535154, 0.050195457458496094, 0.049993728637695314, 0.04950527954101563, 0.04934143829345703, 0.05148979187011719, 0.05039616012573242, 0.04994047927856445, 0.0500684814453125, 0.05000191879272461, 0.049860607147216796, 0.050184192657470705, 0.05008281707763672, 0.05017702484130859, 0.04991692733764649, 0.04950425720214844, 0.04989952087402344, 0.04970905685424805, 0.049391616821289064, 0.050065406799316405, 0.04981248092651367, 0.04980121612548828, 0.04990566253662109, 0.049358848571777345, 0.04936294555664063, 0.04978585433959961, 0.0497520637512207, 0.049873920440673826, 0.04991897583007812, 0.049999870300292966, 0.04938956832885742, 0.051095550537109374, 0.04938956832885742]",tokens/s,19.805392559571388,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4030,7 +4030,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4673-7105734d35b6d0eb662aaac0;49f80694-dd7f-4175-a694-4a26c3dbe03a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe99d-6681c5997875c2181c3bbbbb;4984ac69-a5e1-4ddb-bfd8-84206f603c37) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4129,7 +4129,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e958e-67f1024a741c71cc0ab49342;205bd8c6-7540-45de-9f1c-ebe8e9cb4b08) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe75d-018dc94c2aaa6d20598db50b;53954103-f9f3-4528-b77a-dbdda497718b) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4228,7 +4228,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9631-7210f3ae2490bb805e16b1a7;02dd6005-a978-4ac9-a7e7-4700aba3baab) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe801-7418fd22278af5943f44ca86;d74961cf-c086-4aea-bbe0-eedd328686cd) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4260,7 +4260,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5006.11072,8173.125632,0.0,7526.678528,6833.613824,s,10,3.234631896972656,0.3234631896972656,0.001290890363598202,0.32324838256835936,0.32507350158691406,0.3250783767700195,0.3250822769165039,"[0.3225212097167969, 0.32506610107421874, 0.32166519165039065, 0.3220767822265625, 0.32293771362304685, 0.3222037658691406, 0.32355905151367187, 0.3244464111328125, 0.3250724182128906, 0.325083251953125]",tokens/s,791.4347231893512,kWh,3.805942098713583e-06,2.0851093678459165e-06,2.0185346009376354e-05,2.6076397475935855e-05,tokens/kWh,9817307.020122128,MB,5006.11072,8173.125632,0.0,7526.678528,7110.583808,s,10,176.69872265624997,17.669872265624996,0.0012983881421126473,17.6700712890625,17.6713546875,17.67136875,17.67138,"[17.667875, 17.670087890625, 17.67133984375, 17.67002734375, 17.6713515625, 17.6700546875, 17.66839453125, 17.66792578125, 17.6713828125, 17.670283203125]",tokens/s,3.5653907992623313,kWh,0.00020855451250034905,0.00011430324463261854,0.0010754413360468291,0.0013982990931797968,tokens/kWh,45054.7385085798,,s,629,179.32957586669917,0.2851026643349749,0.03850587607379773,0.2804192810058594,0.28063436889648435,0.28075439453125,0.6034208325195313,"[0.2807080993652344, 0.2805329895019531, 0.28025140380859376, 0.28053607177734374, 0.28031283569335935, 0.2804653930664063, 0.2804398193359375, 0.2805975036621094, 0.2809886779785156, 0.2807132263183594, 0.2803640441894531, 0.28059136962890624, 0.28041012573242186, 0.28036813354492185, 0.2803312683105469, 0.2803927001953125, 0.2804172668457031, 0.28025445556640627, 0.2803117980957031, 0.28033331298828124, 0.2803240966796875, 0.2801163635253906, 0.280345458984375, 0.28016845703125, 0.280369140625, 0.28039883422851564, 0.2807910461425781, 0.280511474609375, 0.280563720703125, 0.28041317749023437, 0.2805340270996094, 0.2804440002441406, 0.28055950927734374, 0.2804561767578125, 0.28033535766601564, 0.28021453857421874, 0.2805248107910156, 0.2804111328125, 0.28032101440429685, 0.2803333740234375, 0.2802308349609375, 0.2801756286621094, 0.28041830444335936, 0.28040704345703127, 0.2803896179199219, 0.2802872314453125, 0.2804316101074219, 0.2803599243164063, 0.2805401611328125, 0.28073471069335937, 0.28038763427734376, 0.28052883911132814, 0.28028314208984373, 0.2804653930664063, 0.2805145568847656, 0.28037631225585935, 0.28035687255859376, 0.28009164428710936, 0.280563720703125, 0.2805452880859375, 0.2803619689941406, 0.2802472839355469, 0.606171142578125, 0.28023602294921873, 0.2805667724609375, 0.28041830444335936, 0.28034356689453127, 0.2801756286621094, 0.2803046264648438, 0.2802882690429688, 0.2804449157714844, 0.2802883605957031, 0.2804110412597656, 0.280369140625, 0.2804029541015625, 0.2803896179199219, 0.28032205200195315, 0.28044595336914063, 0.2805125122070313, 0.28050225830078124, 0.28046234130859377, 0.2804019165039062, 0.2804981689453125, 0.2805166015625, 0.28044082641601564, 0.2806159362792969, 0.28032818603515625, 0.28030157470703126, 0.2801919860839844, 0.28039065551757814, 0.2802534484863281, 0.2802708740234375, 0.28019607543945313, 0.28049099731445315, 0.2804510803222656, 0.2804951171875, 0.2803937377929687, 0.28054119873046873, 0.28021759033203125, 0.28069683837890624, 0.28063232421875, 0.280690673828125, 0.28052377319335936, 0.28034970092773437, 0.2803292236328125, 0.2804285583496094, 0.28027392578125, 0.2803363952636719, 0.2830735473632813, 0.280342529296875, 0.2805002136230469, 0.28026470947265625, 0.280332275390625, 0.28067636108398436, 0.28094464111328127, 0.28055142211914064, 0.2803056640625, 0.28043777465820313, 0.280479736328125, 0.2805698547363281, 0.28024627685546877, 0.2804009094238281, 0.2803466186523437, 0.2804971618652344, 0.2802001953125, 0.6034718627929687, 0.2802196350097656, 0.2803599243164063, 0.28019610595703126, 0.2802452392578125, 0.28016439819335937, 0.28058108520507813, 0.28036813354492185, 0.280489990234375, 0.28041217041015626, 0.2804766845703125, 0.280406005859375, 0.28043264770507814, 0.28037017822265625, 0.2804029541015625, 0.2804766845703125, 0.2804409484863281, 0.28033114624023436, 0.280384521484375, 0.28046234130859377, 0.2805831604003906, 0.2805032958984375, 0.2803189697265625, 0.28047564697265626, 0.2804776916503906, 0.2805032958984375, 0.2803240966796875, 0.2804725646972656, 0.280263671875, 0.2804572143554688, 0.2804981689453125, 0.28048895263671875, 0.28037939453125, 0.28040704345703127, 0.28037939453125, 0.28038247680664063, 0.28024627685546877, 0.28041522216796877, 0.280531982421875, 0.28053912353515625, 0.2805320739746094, 0.28040591430664064, 0.280352783203125, 0.28057293701171876, 0.2807500915527344, 0.2806947937011719, 0.2803097839355469, 0.28375958251953126, 0.2804418640136719, 0.28035174560546877, 0.2806302795410156, 0.2805452880859375, 0.2805770263671875, 0.28036505126953126, 0.2803804016113281, 0.2804449157714844, 0.28060671997070313, 0.28033535766601564, 0.2802309265136719, 0.2807715759277344, 0.2806528015136719, 0.28054833984375, 0.2802769775390625, 0.6030131225585937, 0.280258544921875, 0.28030975341796877, 0.28019097900390627, 0.28033331298828124, 0.2803947448730469, 0.2806056823730469, 0.2804992065429687, 0.280352783203125, 0.28040499877929687, 0.28040704345703127, 0.2803046264648438, 0.2813982849121094, 0.2808186950683594, 0.28054425048828124, 0.2802554931640625, 0.280458251953125, 0.28053811645507815, 0.280384521484375, 0.2803640441894531, 0.28048382568359376, 0.2803180236816406, 0.2802380065917969, 0.280384521484375, 0.280332275390625, 0.2803558349609375, 0.280263671875, 0.28042752075195315, 0.2803292236328125, 0.28048382568359376, 0.28054022216796876, 0.28073977661132815, 0.28061285400390623, 0.28051358032226564, 0.28070404052734377, 0.28075820922851563, 0.2804500427246094, 0.2803436279296875, 0.2801786193847656, 0.28026571655273436, 0.28032000732421875, 0.28032000732421875, 0.28045208740234373, 0.28043777465820313, 0.28250009155273437, 0.2803670959472656, 0.28039065551757814, 0.28047259521484375, 0.2804735717773438, 0.28055657958984376, 0.2805083923339844, 0.28037017822265625, 0.28052276611328125, 0.2804500427246094, 0.28042648315429686, 0.2804561767578125, 0.28051864624023437, 0.28045208740234373, 0.280205322265625, 0.28048281860351565, 0.280342529296875, 0.28044287109375, 0.28014181518554687, 0.6037985229492188, 0.28016024780273435, 0.2806005859375, 0.2805433349609375, 0.2804161682128906, 0.2803446044921875, 0.2804192810058594, 0.28039065551757814, 0.28035482788085936, 0.2803814392089844, 0.28049099731445315, 0.2804490356445313, 0.280406005859375, 0.28048590087890624, 0.2803670959472656, 0.28037939453125, 0.2804244384765625, 0.2816696472167969, 0.2804664306640625, 0.2806661071777344, 0.2807828369140625, 0.2805749816894531, 0.28032205200195315, 0.2804725646972656, 0.2803814392089844, 0.2803445739746094, 0.2803363952636719, 0.28047564697265626, 0.28027496337890623, 0.2807398071289062, 0.28061285400390623, 0.2807162780761719, 0.2804971618652344, 0.28034970092773437, 0.28027493286132815, 0.28037734985351564, 0.2802923583984375, 0.28038861083984373, 0.280300537109375, 0.28038861083984373, 0.2803998718261719, 0.2803978271484375, 0.2803036193847656, 0.28044390869140623, 0.2803814392089844, 0.2804592590332031, 0.28030877685546873, 0.2804520568847656, 0.2805401611328125, 0.28040093994140625, 0.28047869873046877, 0.2802821044921875, 0.28053607177734374, 0.280384521484375, 0.28043060302734374, 0.28053607177734374, 0.2805882873535156, 0.2829209594726563, 0.28022271728515624, 0.28046746826171876, 0.28047564697265626, 0.28037734985351564, 0.2803466186523437, 0.6032896118164063, 0.28028729248046874, 0.280363037109375, 0.2801775817871094, 0.28032717895507814, 0.28021353149414063, 0.28027389526367186, 0.2802452392578125, 0.2804510803222656, 0.2805073852539062, 0.28084429931640625, 0.2804039611816406, 0.28040093994140625, 0.28040298461914065, 0.2804899291992187, 0.2803292236328125, 0.28033535766601564, 0.2803394470214844, 0.280374267578125, 0.2803261413574219, 0.28048794555664064, 0.28052581787109376, 0.28022271728515624, 0.2803998718261719, 0.28032818603515625, 0.2804019165039062, 0.2803507080078125, 0.28050125122070313, 0.28041217041015626, 0.28047579956054686, 0.2804621887207031, 0.28067019653320313, 0.2804643859863281, 0.2803978271484375, 0.28047360229492185, 0.2806077575683594, 0.28034356689453127, 0.2804471740722656, 0.2830160522460938, 0.28041522216796877, 0.2804878845214844, 0.2803507080078125, 0.28064254760742186, 0.28075726318359373, 0.28046847534179686, 0.2805565490722656, 0.2804930419921875, 0.280406005859375, 0.28038247680664063, 0.28038861083984373, 0.28051763916015626, 0.28036300659179686, 0.2802964477539063, 0.2803466186523437, 0.28047052001953127, 0.280489990234375, 0.280458251953125, 0.2804019165039062, 0.28022988891601563, 0.2804500427246094, 0.280489990234375, 0.28060671997070313, 0.28033535766601564, 0.6047180786132812, 0.2801838073730469, 0.2804490356445313, 0.28032305908203126, 0.2802616271972656, 0.28019915771484377, 0.2802657775878906, 0.2804325866699219, 0.28046847534179686, 0.2803947448730469, 0.28039065551757814, 0.28033642578125, 0.28051556396484373, 0.2802339782714844, 0.28041522216796877, 0.2803292236328125, 0.28062619018554685, 0.28064254760742186, 0.280531982421875, 0.28036300659179686, 0.2804572143554688, 0.2804643859863281, 0.2804633483886719, 0.28211508178710937, 0.28030157470703126, 0.2803814392089844, 0.2802503662109375, 0.28055142211914064, 0.2803292236328125, 0.28034765625, 0.28033023071289065, 0.28057089233398436, 0.28050125122070313, 0.2804154052734375, 0.2804631958007813, 0.28067019653320313, 0.2802503662109375, 0.28074188232421876, 0.28085147094726565, 0.2804930419921875, 0.28036813354492185, 0.280416259765625, 0.2802923583984375, 0.2804019165039062, 0.2803046264648438, 0.2803619689941406, 0.2803456115722656, 0.2802769775390625, 0.280453125, 0.2803978271484375, 0.28041317749023437, 0.28027392578125, 0.28071218872070314, 0.28065689086914064, 0.2804090881347656, 0.28041217041015626, 0.28040704345703127, 0.28029541015625, 0.2801346435546875, 0.28052581787109376, 0.2804592590332031, 0.28048590087890624, 0.2802903137207031, 0.6054840087890625, 0.28021453857421874, 0.280310791015625, 0.28032205200195315, 0.28034765625, 0.28025753784179686, 0.2805084533691406, 0.2806282043457031, 0.28064358520507815, 0.28045111083984375, 0.28058212280273437, 0.2803957824707031, 0.28048794555664064, 0.28039886474609377, 0.2804111022949219, 0.2803732604980469, 0.28168295288085937, 0.28037734985351564, 0.280384521484375, 0.2804725646972656, 0.280384521484375, 0.28036505126953126, 0.28022476196289064, 0.2803456115722656, 0.2803947448730469, 0.28043365478515625, 0.2802872314453125, 0.2803343505859375, 0.28028005981445314, 0.2804500427246094, 0.28037118530273436, 0.280627197265625, 0.28052581787109376, 0.28042138671875, 0.28047052001953127, 0.2803825378417969, 0.28021856689453123, 0.2804111328125, 0.280310791015625, 0.28049612426757814, 0.28038348388671874, 0.280332275390625, 0.2804029541015625, 0.2804561767578125, 0.2804613037109375, 0.28042138671875, 0.2804080505371094, 0.280585205078125, 0.2807091064453125, 0.28042138671875, 0.28039678955078123, 0.2802759704589844, 0.28042648315429686, 0.2803343505859375, 0.2803670959472656, 0.28048590087890624, 0.28051864624023437, 0.28044512939453126, 0.28021951293945313, 0.2805595397949219, 0.28056781005859377, 0.28054425048828124, 0.28022271728515624, 0.6061547241210937, 0.2803138427734375, 0.280384521484375, 0.2803660888671875, 0.28042657470703125, 0.280196044921875, 0.2803947448730469, 0.28044699096679687, 0.28030670166015625, 0.2804285583496094, 0.28048895263671875, 0.2804090881347656, 0.2807029724121094, 0.280764404296875, 0.2804920349121094, 0.280521728515625, 0.2805084228515625, 0.28190618896484376, 0.2804491271972656, 0.28042230224609377, 0.2804757080078125, 0.28043154907226564, 0.28037118530273436, 0.28051864624023437, 0.28060467529296873, 0.2805248107910156, 0.2803466186523437, 0.28050942993164063, 0.2803037109375, 0.2804366455078125, 0.2803835144042969, 0.28081353759765626, 0.28074496459960935, 0.28037835693359375, 0.28037631225585935, 0.2805555114746094, 0.28037939453125, 0.2805688171386719, 0.2804551696777344, 0.280510498046875, 0.2805585632324219, 0.2803660888671875, 0.280416259765625, 0.28055859375, 0.28047052001953127, 0.2804244384765625, 0.2805206909179688, 0.28048593139648437, 0.2806968078613281, 0.28052993774414064, 0.2805851745605469, 0.28038247680664063, 0.28042138671875, 0.2804561767578125, 0.28033740234375, 0.2804776916503906, 0.2805770263671875, 0.2803240966796875, 0.28041522216796877, 0.280511474609375, 0.28045413208007813, 0.2804766845703125, 0.28029034423828125, 0.6060789184570312, 0.28018893432617187, 0.28060467529296873, 0.28061285400390623, 0.28044287109375, 0.2802237548828125, 0.2804111328125, 0.28039678955078123, 0.2804664306640625, 0.2803947448730469, 0.2804111328125, 0.2804019165039062, 0.28042752075195315, 0.2803138427734375, 0.28037631225585935, 0.2802904052734375, 0.28029736328125, 0.28034970092773437, 0.28041934204101565, 0.2805073852539062, 0.2807215576171875, 0.28053692626953125, 0.28042752075195315, 0.2806824951171875, 0.2805852661132813, 0.280438720703125, 0.28213861083984376, 0.280531982421875, 0.28032000732421875, 0.28051559448242186, 0.28042648315429686, 0.280711181640625, 0.2804653930664063, 0.28054425048828124, 0.2804316101074219, 0.2805401611328125, 0.2803804016113281, 0.2804695129394531, 0.28054629516601565, 0.2805770263671875, 0.28039486694335936, 0.280461181640625, 0.28036300659179686, 0.28038247680664063, 0.2803312683105469, 0.2803261413574219, 0.28028005981445314, 0.28033023071289065, 0.28046234130859377, 0.2805422058105469, 0.28042755126953123, 0.2803332824707031, 0.2804695129394531, 0.28044082641601564, 0.2805206909179688, 0.28069171142578125, 0.28068865966796874, 0.2804121398925781, 0.28018893432617187, 0.2804930419921875, 0.28053094482421875, 0.2804398193359375, 0.2802923583984375]",tokens/s,3.5075084350143877,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4290,7 +4290,7 @@ ValueError: FalconForCausalLM does not support an attention implementation throu ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4778.385408,21498.42944,0.0,20851.982336,20236.60032,s,10,15.667565917968751,1.566756591796875,0.0018490956286824453,1.5662525024414062,1.5679145385742188,1.569856866455078,1.5714107287597656,"[1.565250732421875, 1.566271728515625, 1.5669622802734375, 1.56748291015625, 1.5658883056640625, 1.565689697265625, 1.56708056640625, 1.5662332763671876, 1.5717991943359375, 1.5649072265625]",tokens/s,163.39487661347562,kWh,1.848894233504931e-05,1.0131964364536543e-05,9.316793564540516e-05,0.000121788842344991,tokens/kWh,2101998.796201949,MB,4784.119808,21498.42944,0.0,20851.982336,20339.706368,s,10,881.343109375,88.1343109375,0.002956912666877524,88.1353125,88.13739453125,88.137447265625,88.137489453125,"[88.129875, 88.130765625, 88.1368203125, 88.137265625, 88.1375, 88.1343046875, 88.1363203125, 88.13146875, 88.1373828125, 88.13140625]",tokens/s,0.7148180921806506,kWh,0.0010404256961577469,0.0005702452815513971,0.005204286163425598,0.006814957141134741,tokens/kWh,9244.37215015413,,s,629,894.0462337646492,1.4213771601981693,0.18617884218031844,1.398877197265625,1.3992523681640625,1.3995548828125002,2.96542931640625,"[1.39861083984375, 1.398836181640625, 1.39867138671875, 1.3985986328125, 1.3985372314453124, 1.3985423583984375, 1.3984840087890624, 1.3988331298828125, 1.3987451171875, 1.3984296875, 1.398603759765625, 1.39862939453125, 1.398592529296875, 1.39898779296875, 1.399804931640625, 1.3989600830078126, 1.3987215576171874, 1.3987052001953124, 1.3989468994140626, 1.39868359375, 1.398623291015625, 1.39879931640625, 1.3986170654296874, 1.398488037109375, 1.399183349609375, 1.3986939697265626, 1.39886279296875, 1.3986058349609376, 1.3987020263671874, 1.3987952880859376, 1.398541259765625, 1.3985986328125, 1.3989161376953125, 1.398920166015625, 1.3990419921875, 1.39896826171875, 1.3991865234375, 1.398741943359375, 1.398930419921875, 1.3989847412109375, 1.3989520263671875, 1.3987255859375, 1.398935546875, 1.39886083984375, 1.3990206298828125, 1.3991453857421876, 1.39919775390625, 1.3996278076171875, 1.3989007568359375, 1.3989898681640625, 1.3989866943359375, 1.3988802490234375, 1.3991854248046875, 1.399235595703125, 1.398929443359375, 1.3990225830078125, 1.3991126708984376, 1.398993896484375, 1.39955615234375, 1.3990440673828124, 1.3991434326171874, 1.39884033203125, 2.965783447265625, 1.3988966064453126, 1.398802490234375, 1.3985321044921875, 1.39858740234375, 1.3986273193359375, 1.3983150634765624, 1.3987532958984374, 1.3986478271484375, 1.3984962158203125, 1.3987645263671875, 1.3986058349609376, 1.3986898193359374, 1.39869189453125, 1.3983323974609374, 1.398645751953125, 1.3987481689453125, 1.39890478515625, 1.3990318603515626, 1.4008525390625, 1.3985054931640626, 1.3984737548828126, 1.3986109619140625, 1.3987440185546876, 1.3986488037109375, 1.3987255859375, 1.398550537109375, 1.3988331298828125, 1.3985792236328125, 1.398634521484375, 1.3985382080078126, 1.3987164306640625, 1.3986856689453124, 1.39848291015625, 1.3987113037109375, 1.3989376220703125, 1.398762451171875, 1.3987420654296876, 1.3987840576171875, 1.3990584716796874, 1.399045166015625, 1.401311279296875, 1.39856591796875, 1.3989765625, 1.398443115234375, 1.398836181640625, 1.39945166015625, 1.3988555908203124, 1.3991495361328126, 1.3991956787109374, 1.39893359375, 1.399204833984375, 1.3988291015625, 1.399236572265625, 1.399004150390625, 1.399005126953125, 1.3992908935546875, 1.3988536376953125, 1.399035888671875, 1.3992274169921874, 1.3993287353515624, 1.399700439453125, 1.3991219482421875, 2.966549560546875, 1.39862939453125, 1.39866015625, 1.399478271484375, 1.399192626953125, 1.3989632568359376, 1.398771728515625, 1.3986805419921875, 1.3987911376953126, 1.3989549560546874, 1.3987236328125, 1.3990697021484375, 1.3988494873046875, 1.3987430419921876, 1.3989058837890624, 1.39859765625, 1.3985792236328125, 1.398561767578125, 1.3986058349609376, 1.3985843505859374, 1.3987799072265625, 1.3989171142578125, 1.39850244140625, 1.40042041015625, 1.3986375732421874, 1.398804443359375, 1.398970458984375, 1.3991331787109376, 1.3986907958984376, 1.3987430419921876, 1.3988720703125, 1.399077880859375, 1.398720458984375, 1.3989119873046876, 1.398973388671875, 1.39915576171875, 1.39900927734375, 1.3989549560546874, 1.398836181640625, 1.3991864013671875, 1.39887109375, 1.39875537109375, 1.4007685546875, 1.39913525390625, 1.39900927734375, 1.3990001220703125, 1.39928369140625, 1.3991044921875, 1.3991658935546876, 1.3989150390625, 1.39884130859375, 1.3990972900390626, 1.3989857177734375, 1.3992447509765624, 1.399131103515625, 1.399405517578125, 1.3992642822265624, 1.3992550048828125, 1.3990594482421874, 1.3992591552734375, 1.3989908447265624, 1.39915771484375, 1.3989212646484375, 2.96693359375, 1.3985638427734375, 1.398752197265625, 1.398823974609375, 1.398572021484375, 1.3988055419921874, 1.398634521484375, 1.3986221923828126, 1.3986856689453124, 1.39827197265625, 1.398539306640625, 1.3988157958984375, 1.4004111328125, 1.39886279296875, 1.3987327880859375, 1.39856689453125, 1.3985545654296876, 1.39869287109375, 1.3988526611328125, 1.39871533203125, 1.39894580078125, 1.3987052001953124, 1.398551513671875, 1.3985792236328125, 1.3989222412109374, 1.399103515625, 1.398645751953125, 1.3989754638671874, 1.3987388916015624, 1.3983846435546874, 1.398825927734375, 1.3987491455078125, 1.3987706298828124, 1.398856689453125, 1.3985126953125, 1.3989344482421875, 1.3986201171875, 1.3990635986328126, 1.3985853271484374, 1.3990738525390625, 1.399287841796875, 1.39908203125, 1.3989232177734374, 1.3990615234375, 1.403336669921875, 1.399041015625, 1.398888427734375, 1.3992376708984375, 1.399426025390625, 1.3991126708984376, 1.39894677734375, 1.39915673828125, 1.3997025146484374, 1.3994208984375, 1.39913623046875, 1.398823974609375, 1.3998172607421875, 1.399552978515625, 1.3991270751953124, 1.3991854248046875, 1.399057373046875, 1.3992447509765624, 1.3991864013671875, 2.965474365234375, 1.3988095703125, 1.3988209228515625, 1.3986990966796875, 1.39848193359375, 1.3987869873046874, 1.398802490234375, 1.3984625244140625, 1.3988065185546874, 1.3984798583984375, 1.3986632080078125, 1.39890380859375, 1.3988720703125, 1.398951904296875, 1.3987686767578125, 1.398593505859375, 1.3987645263671875, 1.401100341796875, 1.398867919921875, 1.3988475341796875, 1.398614990234375, 1.398646728515625, 1.3986273193359375, 1.39869287109375, 1.3989990234375, 1.3990072021484374, 1.398619140625, 1.3986314697265625, 1.398825927734375, 1.3988597412109376, 1.3990645751953126, 1.398625244140625, 1.3988720703125, 1.3991177978515625, 1.398873046875, 1.3989488525390625, 1.3991280517578124, 1.398972412109375, 1.398724609375, 1.3990164794921875, 1.39892431640625, 1.398846435546875, 1.399014404296875, 1.398878173828125, 1.3987020263671874, 1.3988331298828125, 1.3991075439453124, 1.3988905029296874, 1.3992171630859376, 1.39901025390625, 1.3989990234375, 1.3992232666015625, 1.4028953857421875, 1.3991341552734375, 1.3992559814453125, 1.39980908203125, 1.399646240234375, 1.399251953125, 1.3990379638671875, 1.3988916015625, 1.39907373046875, 1.399131103515625, 1.398908935546875, 2.96527978515625, 1.3986693115234374, 1.3986282958984375, 1.3985587158203125, 1.39848193359375, 1.3986611328125, 1.3984716796875, 1.3986427001953126, 1.3985526123046874, 1.3987379150390624, 1.3991270751953124, 1.3989129638671876, 1.3986580810546876, 1.3988095703125, 1.39888232421875, 1.398624267578125, 1.3986683349609375, 1.3986641845703125, 1.3986754150390626, 1.398550537109375, 1.398556640625, 1.3986058349609376, 1.398283203125, 1.3985740966796876, 1.398614990234375, 1.398750244140625, 1.3989530029296875, 1.3992540283203125, 1.3987236328125, 1.39867138671875, 1.3988074951171876, 1.398635498046875, 1.398919189453125, 1.3987952880859376, 1.3987451171875, 1.3988116455078126, 1.3987747802734376, 1.398972412109375, 1.404094482421875, 1.3990953369140624, 1.39894580078125, 1.3994281005859375, 1.3991240234375, 1.399256103515625, 1.398873046875, 1.39911572265625, 1.3989263916015624, 1.399004150390625, 1.39951513671875, 1.3991126708984376, 1.3989796142578126, 1.39894677734375, 1.3989365234375, 1.399294921875, 1.3989478759765626, 1.398964111328125, 1.3990972900390626, 1.3990020751953125, 1.3989478759765626, 1.399329833984375, 1.39915673828125, 1.3990286865234376, 1.3991751708984375, 2.966761474609375, 1.3985423583984375, 1.39884130859375, 1.3985556640625, 1.3985556640625, 1.3988095703125, 1.39877783203125, 1.3987860107421874, 1.3986641845703125, 1.39853515625, 1.3985361328125, 1.398656982421875, 1.3986968994140625, 1.39901953125, 1.3986212158203124, 1.3989027099609375, 1.3986119384765625, 1.398583251953125, 1.398730712890625, 1.3989017333984375, 1.398697998046875, 1.398677490234375, 1.3984931640625, 1.401703369140625, 1.3986641845703125, 1.39890185546875, 1.3988289794921875, 1.3988167724609375, 1.398576171875, 1.3986744384765626, 1.398751220703125, 1.3987318115234375, 1.398581298828125, 1.39869384765625, 1.39877783203125, 1.3993779296875, 1.399005126953125, 1.3989232177734374, 1.3988587646484374, 1.39894677734375, 1.39911474609375, 1.399098388671875, 1.398730712890625, 1.39896826171875, 1.398935546875, 1.39915576171875, 1.39913623046875, 1.3991424560546875, 1.3990655517578126, 1.3991065673828125, 1.3990860595703125, 1.399183349609375, 1.3989744873046874, 1.39915576171875, 1.3991240234375, 1.399055419921875, 1.402797119140625, 1.3991197509765625, 1.398992919921875, 1.3993922119140625, 1.3990738525390625, 1.399083984375, 1.3990860595703125, 2.965937255859375, 1.3989212646484375, 1.3988382568359374, 1.398813720703125, 1.3986324462890625, 1.398729736328125, 1.3987880859375, 1.3985526123046874, 1.398720458984375, 1.39859765625, 1.39854443359375, 1.3988331298828125, 1.3987318115234375, 1.3988648681640625, 1.3985648193359375, 1.3985484619140625, 1.400943603515625, 1.3987593994140626, 1.398794189453125, 1.3986282958984375, 1.3985535888671874, 1.398667236328125, 1.3985853271484374, 1.3986427001953126, 1.398782958984375, 1.3989068603515624, 1.3985126953125, 1.39886083984375, 1.398930419921875, 1.39879931640625, 1.3986181640625, 1.3988238525390626, 1.3986898193359374, 1.3985914306640626, 1.3986754150390626, 1.3989530029296875, 1.3988914794921874, 1.3989222412109374, 1.3990400390625, 1.399320556640625, 1.398939697265625, 1.3989180908203125, 1.3988443603515626, 1.3991177978515625, 1.3987430419921876, 1.3990850830078125, 1.3988494873046875, 1.3990185546875, 1.399118896484375, 1.398983642578125, 1.399162841796875, 1.3992745361328125, 1.398998046875, 1.398919189453125, 1.3992181396484376, 1.399088134765625, 1.39907177734375, 1.3991177978515625, 1.399078857421875, 1.399394287109375, 1.399125, 1.3990706787109375, 1.3991875, 2.971739013671875, 1.399003173828125, 1.3988106689453126, 1.3985955810546875, 1.3986651611328125, 1.3988382568359374, 1.3985577392578126, 1.39854541015625, 1.3985228271484376, 1.398729736328125, 1.3986324462890625, 1.3985679931640624, 1.3984625244140625, 1.399056396484375, 1.398541259765625, 1.398877197265625, 1.39889453125, 1.4004193115234376, 1.3987532958984374, 1.3991875, 1.39894482421875, 1.39869384765625, 1.3984102783203125, 1.3985894775390626, 1.3989273681640626, 1.3987164306640625, 1.398676513671875, 1.3989805908203126, 1.398752197265625, 1.3987369384765624, 1.3986201171875, 1.3987164306640625, 1.3989530029296875, 1.3989693603515625, 1.3985843505859374, 1.3986744384765626, 1.3989273681640626, 1.3991177978515625, 1.398738037109375, 1.39888427734375, 1.399014404296875, 1.3989857177734375, 1.3989949951171874, 1.3990389404296875, 1.3988638916015625, 1.3990963134765626, 1.39890380859375, 1.3990758056640624, 1.3991700439453125, 1.399078857421875, 1.399034912109375, 1.3990799560546876, 1.398781982421875, 1.3989775390625, 1.3993604736328125, 1.3990277099609374, 1.3989385986328124, 1.3992847900390626, 1.4042296142578126, 1.3992314453125, 1.3990748291015624, 1.399125, 1.39915673828125, 2.9653134765625, 1.3987911376953126, 1.3988863525390625, 1.3984849853515624, 1.398365234375, 1.3986478271484375, 1.3986856689453124, 1.3987255859375, 1.398328369140625, 1.3986314697265625, 1.3986007080078124, 1.398823974609375, 1.398498291015625, 1.39887109375, 1.3987860107421874, 1.3986385498046876, 1.3992396240234375, 1.3990430908203124, 1.3988321533203125, 1.3987052001953124, 1.3982330322265626, 1.3985146484375, 1.3988536376953125, 1.3986785888671875, 1.398445068359375, 1.39894677734375, 1.4012242431640625, 1.3989908447265624, 1.3988177490234375, 1.3987337646484375, 1.39881982421875, 1.3986683349609375, 1.3987255859375, 1.3986907958984376, 1.3990216064453125, 1.3991240234375, 1.398856689453125, 1.399024658203125, 1.398556640625, 1.3989744873046874, 1.3988157958984375, 1.3987318115234375, 1.3989478759765626, 1.3992376708984375, 1.398750244140625, 1.3990389404296875, 1.39900927734375, 1.399330810546875, 1.3993922119140625, 1.39889453125, 1.399046142578125, 1.3991455078125, 1.39875537109375, 1.3991259765625, 1.39901025390625, 1.399046142578125, 1.399300048828125, 1.39921923828125, 1.399109619140625, 1.3992078857421875, 1.39896630859375, 1.399267333984375, 1.3988720703125]",tokens/s,0.7035430341800197,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4319,7 +4319,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4aad-04710e0424df0f9646b94647;b7c59044-3a5a-4a4c-82da-e7be2ec0c625) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664feb26-2c881bda6ae9c23668b2dd37;9e31c27b-645b-4d74-ae47-dc74570736e7) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -4404,7 +4404,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e965c-049692ca67c1615c40646dce;a3bef62f-cda1-471e-999d-49ff959fec24) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe829-5a0bba330ed7c3cf2b291ee1;ee139925-daee-4b19-8d12-21ccca2a6390) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4436,7 +4436,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2101.428224,2846.359552,0.0,2199.912448,2075.149824,s,10,1.3471448974609377,0.13471448974609376,0.0003936742891810882,0.13468195343017578,0.13517516784667968,0.13523564758300782,0.13528403137207032,"[0.13474790954589844, 0.13450845336914063, 0.1341354217529297, 0.13502685546875, 0.13516172790527345, 0.13461599731445312, 0.1343356475830078, 0.1350916748046875, 0.13422508239746095, 0.13529612731933593]",tokens/s,1900.3152554896055,kWh,1.5858272093313712e-06,8.689580453501548e-07,7.364979966052824e-06,9.81976522073435e-06,tokens/kWh,26069869.721473403,MB,2101.428224,2846.359552,0.0,2199.912448,2180.296192,s,10,73.0066201171875,7.3006620117187495,0.0009503587647667944,7.30020068359375,7.302343115234375,7.302360034179688,7.302373569335938,"[7.30233935546875, 7.300029296875, 7.29990771484375, 7.30075146484375, 7.3012353515625, 7.30013525390625, 7.30026611328125, 7.302376953125, 7.2997060546875, 7.29987255859375]",tokens/s,8.629354420034067,kWh,8.619521513139761e-05,4.723979546581904e-05,0.00039969104012294997,0.0005331260507201667,tokens/kWh,118170.92771005514,,s,629,74.10269908905025,0.11781033241502431,0.016038562819056714,0.11586867523193359,0.11600281524658203,0.11604725799560547,0.25048523010253904,"[0.116384765625, 0.11645645141601563, 0.1162772445678711, 0.11587481689453125, 0.1159024658203125, 0.11586969757080078, 0.11578265380859375, 0.11597516632080078, 0.1156864013671875, 0.11581644439697265, 0.11589324951171875, 0.11574578857421874, 0.11588921356201172, 0.11588909149169922, 0.11581951904296875, 0.11583385467529297, 0.1159362564086914, 0.11588198089599609, 0.11589017486572266, 0.11591168212890625, 0.1158635482788086, 0.11604070281982422, 0.11596390533447265, 0.1157959976196289, 0.11587888336181641, 0.11584614562988281, 0.11581337738037109, 0.11577651214599609, 0.11592499542236329, 0.11576525115966797, 0.11587276458740234, 0.11586150360107422, 0.11592499542236329, 0.11580006408691407, 0.11592396545410157, 0.11582361602783203, 0.11592396545410157, 0.11594649505615234, 0.11588301086425781, 0.11581542205810547, 0.1159208984375, 0.11584716796875, 0.11586563110351562, 0.11596489715576172, 0.11587071990966796, 0.11584307098388671, 0.11591577911376953, 0.11582669067382813, 0.11593830108642578, 0.11596393585205078, 0.11587375640869141, 0.115810302734375, 0.11594649505615234, 0.11581852722167969, 0.1158983383178711, 0.11596390533447265, 0.11589939117431641, 0.11581132507324218, 0.11588607788085938, 0.11584207916259766, 0.11590652465820313, 0.11594035339355468, 0.25061375427246096, 0.1157396469116211, 0.11593421173095703, 0.11573350524902344, 0.11583283233642579, 0.11590656280517578, 0.1157949447631836, 0.11581951904296875, 0.11594751739501953, 0.11577753448486328, 0.1157325439453125, 0.11582969665527344, 0.1158440933227539, 0.11578982543945313, 0.11597516632080078, 0.11578982543945313, 0.11583590698242187, 0.11588915252685547, 0.11580518341064452, 0.11581439971923828, 0.11588607788085938, 0.11584207916259766, 0.11584819030761719, 0.1159147186279297, 0.1158328628540039, 0.11581846618652344, 0.11587891387939453, 0.115736572265625, 0.11584716796875, 0.115884033203125, 0.1159004135131836, 0.11583385467529297, 0.11596288299560546, 0.11589734649658204, 0.11583795166015624, 0.11584819030761719, 0.11587379455566406, 0.11582771301269532, 0.11601203155517578, 0.11582669067382813, 0.11606221008300781, 0.11587174224853515, 0.1159188461303711, 0.11580006408691407, 0.11604991912841797, 0.1158440933227539, 0.11589631652832032, 0.1158809585571289, 0.11587071990966796, 0.1158656005859375, 0.11594342041015625, 0.11577651214599609, 0.11590553283691406, 0.1160079345703125, 0.11591782379150391, 0.11579596710205078, 0.115989501953125, 0.11589631652832032, 0.11592294311523438, 0.11589324951171875, 0.11592806243896485, 0.11582975769042969, 0.11595673370361329, 0.2503679962158203, 0.11580723571777343, 0.11585945892333985, 0.11577651214599609, 0.11584204864501953, 0.11591065979003906, 0.11576531219482422, 0.11575801849365235, 0.11578572845458984, 0.11581337738037109, 0.11575910186767578, 0.11594547271728516, 0.11584102630615234, 0.11585638427734375, 0.11582157135009766, 0.11580620574951171, 0.1157734375, 0.11602944183349609, 0.11576422119140625, 0.11584102630615234, 0.1157949447631836, 0.11588102722167969, 0.1157825927734375, 0.1159731216430664, 0.11577958679199218, 0.11586867523193359, 0.11582669067382813, 0.11583590698242187, 0.11576217651367188, 0.11594239807128906, 0.11580006408691407, 0.11592704010009766, 0.11579084777832031, 0.1158635482788086, 0.11574272155761718, 0.11594649505615234, 0.11582259368896484, 0.11602022552490235, 0.11587481689453125, 0.11585842895507813, 0.11580723571777343, 0.11598233795166016, 0.11587891387939453, 0.11591065979003906, 0.11588915252685547, 0.11585638427734375, 0.11586867523193359, 0.11598336029052735, 0.11583999633789062, 0.11589734649658204, 0.11589324951171875, 0.1158656005859375, 0.11593421173095703, 0.1160263671875, 0.11586048126220704, 0.11587379455566406, 0.11587583923339843, 0.11592294311523438, 0.11588822174072265, 0.11591670227050781, 0.1158318099975586, 0.1159208984375, 0.11615641784667968, 0.25116671752929687, 0.11583999633789062, 0.1161871337890625, 0.11583078765869141, 0.11588710021972656, 0.11593328094482422, 0.11585628509521484, 0.11573350524902344, 0.11615744018554687, 0.11579084777832031, 0.11575296020507812, 0.11597209930419922, 0.11578880310058594, 0.11577037048339844, 0.11582975769042969, 0.11580928039550781, 0.11577139282226563, 0.11601510620117188, 0.11582566070556641, 0.11580723571777343, 0.1158440933227539, 0.11590962982177734, 0.11577446746826171, 0.1160447998046875, 0.11583692932128906, 0.11578982543945313, 0.1157734375, 0.11585330963134766, 0.1157918701171875, 0.11602432250976563, 0.11584921264648437, 0.11583999633789062, 0.11580825805664062, 0.11589222717285157, 0.11580313873291016, 0.11603865814208984, 0.11587891387939453, 0.11582669067382813, 0.1158656005859375, 0.11589529418945313, 0.1158123550415039, 0.11602432250976563, 0.11588915252685547, 0.11576934051513672, 0.115884033203125, 0.1159004135131836, 0.11582669067382813, 0.11619020843505859, 0.11592499542236329, 0.11584614562988281, 0.11585638427734375, 0.11584716796875, 0.11582259368896484, 0.11603046417236328, 0.11590860748291015, 0.11579289245605469, 0.11586867523193359, 0.11590348815917968, 0.1158656005859375, 0.11615641784667968, 0.11596595001220703, 0.11582566070556641, 0.11584307098388671, 0.25133978271484375, 0.1157918701171875, 0.1157918701171875, 0.11581340789794922, 0.11580105590820312, 0.1159208984375, 0.11578470611572265, 0.11577446746826171, 0.11591474914550781, 0.11574169921875, 0.11582157135009766, 0.11597824096679688, 0.11573248291015625, 0.11590144348144531, 0.11589631652832032, 0.11577958679199218, 0.11584614562988281, 0.11606937408447265, 0.11580518341064452, 0.11582259368896484, 0.1160816650390625, 0.11583897399902343, 0.11587993621826172, 0.11600691223144531, 0.1157570571899414, 0.11595673370361329, 0.11589734649658204, 0.11579596710205078, 0.11585945892333985, 0.11595468902587891, 0.11575296020507812, 0.11594547271728516, 0.11589529418945313, 0.1158656005859375, 0.11584204864501953, 0.1160099868774414, 0.11586969757080078, 0.11585433959960938, 0.1160263671875, 0.11582771301269532, 0.1159024658203125, 0.1159731216430664, 0.11586252593994141, 0.11589939117431641, 0.11600077056884765, 0.11582771301269532, 0.11592294311523438, 0.11600691223144531, 0.11583999633789062, 0.11584819030761719, 0.11597618865966797, 0.11576012420654297, 0.11593421173095703, 0.11615846252441406, 0.11584819030761719, 0.11588915252685547, 0.11598233795166016, 0.1158635482788086, 0.11595263671875, 0.11600179290771484, 0.11590348815917968, 0.11580928039550781, 0.11602432250976563, 0.2507939910888672, 0.11579084777832031, 0.1159208984375, 0.1157754898071289, 0.11588505554199219, 0.11589529418945313, 0.11586867523193359, 0.11585228729248047, 0.11579392242431641, 0.11576115417480469, 0.11573554992675782, 0.11589119720458985, 0.11571405029296875, 0.11585126495361328, 0.11578777313232422, 0.11587276458740234, 0.1157754898071289, 0.1159393310546875, 0.1158287353515625, 0.11589734649658204, 0.11582463836669922, 0.11588710021972656, 0.11579801940917969, 0.1159731216430664, 0.11576115417480469, 0.11590348815917968, 0.11575603485107422, 0.1158440933227539, 0.1157734375, 0.11595673370361329, 0.1157918701171875, 0.11586252593994141, 0.11587481689453125, 0.11588301086425781, 0.1157918701171875, 0.11593421173095703, 0.11580108642578125, 0.11593830108642578, 0.11585330963134766, 0.11588813018798828, 0.11581443023681641, 0.1159474868774414, 0.11639091491699219, 0.11594445037841797, 0.11590348815917968, 0.11587071990966796, 0.11583078765869141, 0.11599052429199219, 0.11584204864501953, 0.1159208984375, 0.11589222717285157, 0.11586870574951172, 0.11583074951171875, 0.11595673370361329, 0.11585433959960938, 0.11585536193847656, 0.11588301086425781, 0.11593116760253906, 0.11590550231933594, 0.11593421173095703, 0.11583385467529297, 0.11593830108642578, 0.11590656280517578, 0.2504949798583984, 0.11576934051513672, 0.11589529418945313, 0.11576012420654297, 0.11585228729248047, 0.11593727874755859, 0.1158635482788086, 0.11573760223388672, 0.11580518341064452, 0.11576934051513672, 0.11574066925048829, 0.11602124786376954, 0.11579698944091797, 0.11577651214599609, 0.11578368377685547, 0.11584307098388671, 0.11574578857421874, 0.11600691223144531, 0.11581132507324218, 0.1158123550415039, 0.11582669067382813, 0.11593011474609374, 0.11580210876464844, 0.1160683822631836, 0.11585635375976562, 0.11582566070556641, 0.11578982543945313, 0.11580825805664062, 0.11583283233642579, 0.11603456115722656, 0.11585638427734375, 0.11580620574951171, 0.11578060913085937, 0.11590656280517578, 0.11582771301269532, 0.11603865814208984, 0.11590758514404297, 0.11584716796875, 0.11587481689453125, 0.11587071990966796, 0.11587071990966796, 0.11605197143554688, 0.11588813018798828, 0.11577855682373046, 0.11590144348144531, 0.11589939117431641, 0.11595059204101563, 0.11598643493652344, 0.11587071990966796, 0.11583999633789062, 0.1158656005859375, 0.11585228729248047, 0.11581542205810547, 0.11605404663085937, 0.11617174530029296, 0.11582157135009766, 0.11587379455566406, 0.11591782379150391, 0.11587174224853515, 0.11600281524658203, 0.11595161437988281, 0.11582259368896484, 0.1158748779296875, 0.2516458892822266, 0.11581337738037109, 0.11582259368896484, 0.11585842895507813, 0.11578675079345703, 0.11597516632080078, 0.11578368377685547, 0.11580723571777343, 0.1159208984375, 0.11576422119140625, 0.11574476623535156, 0.11594239807128906, 0.11575296020507812, 0.11589324951171875, 0.11591372680664062, 0.11579698944091797, 0.115884033203125, 0.1160284194946289, 0.11581337738037109, 0.11583999633789062, 0.11594956970214844, 0.11583385467529297, 0.1159188461303711, 0.11604889678955078, 0.11577037048339844, 0.11590656280517578, 0.1158666229248047, 0.11572428894042969, 0.11584512329101562, 0.11597926330566406, 0.11575910186767578, 0.11589939117431641, 0.11591474914550781, 0.11590758514404297, 0.11587481689453125, 0.11601715087890625, 0.1158635482788086, 0.11585842895507813, 0.11596492767333984, 0.11588813018798828, 0.11595366668701172, 0.11598028564453125, 0.11586764526367188, 0.1159393310546875, 0.1159731216430664, 0.11584307098388671, 0.11590860748291015, 0.11597618865966797, 0.11586048126220704, 0.11581439971923828, 0.11600281524658203, 0.11580108642578125, 0.11589836883544923, 0.11604070281982422, 0.11583795166015624, 0.11589836883544923, 0.11596902465820312, 0.1158809585571289, 0.11591474914550781, 0.11605094146728516, 0.1159331817626953, 0.11579392242431641, 0.11675033569335938, 0.2504601593017578, 0.11579801940917969, 0.11592499542236329, 0.11581439971923828, 0.11587379455566406, 0.11589222717285157, 0.11578060913085937, 0.11579596710205078, 0.11579904174804688, 0.1157754898071289, 0.11574578857421874, 0.11589836883544923, 0.11598540496826172, 0.11587071990966796, 0.11578572845458984, 0.11582566070556641, 0.11574784088134765, 0.11597926330566406, 0.11573248291015625, 0.11590144348144531, 0.11580518341064452, 0.1159331817626953, 0.11605811309814452, 0.11600691223144531, 0.11582463836669922, 0.11598336029052735, 0.11579596710205078, 0.11595980834960938, 0.11576934051513672, 0.11596800231933593, 0.11575910186767578, 0.115884033203125, 0.11577651214599609, 0.11587789154052734, 0.1157754898071289, 0.1159362564086914, 0.11577651214599609, 0.11590656280517578, 0.11589119720458985, 0.11586867523193359, 0.11583798217773437, 0.11596489715576172, 0.11581340789794922, 0.11586966705322266, 0.11587789154052734, 0.11586665344238281, 0.11582768249511718, 0.11593011474609374, 0.11577446746826171, 0.11590451049804687, 0.11585228729248047, 0.11581849670410156, 0.11582361602783203, 0.11599359893798829, 0.11584102630615234, 0.11587891387939453, 0.11591680145263672, 0.11588813018798828, 0.11583087921142578, 0.11594844818115234, 0.11587379455566406, 0.11590451049804687, 0.11583283233642579, 0.25145138549804685, 0.11581747436523437, 0.1159188461303711, 0.11576627349853516, 0.1158656005859375, 0.11593421173095703, 0.11584716796875, 0.11570893096923827, 0.1157734375, 0.11578368377685547, 0.11572838592529297, 0.11592396545410157, 0.11577958679199218, 0.1157580795288086, 0.11577855682373046, 0.1158318099975586, 0.11572838592529297, 0.11599667358398437, 0.11578675079345703, 0.11578777313232422, 0.11580518341064452, 0.1159024658203125, 0.11576422119140625, 0.11599565124511718, 0.11581542205810547, 0.11582361602783203, 0.11586252593994141, 0.11583385467529297, 0.11588607788085938, 0.11603046417236328, 0.11584921264648437, 0.1158318099975586, 0.1157918701171875, 0.11593113708496093, 0.11624447631835938, 0.11602022552490235, 0.11585126495361328, 0.11579392242431641, 0.11585638427734375, 0.11593727874755859, 0.11582259368896484, 0.11600895690917969, 0.11589427185058594, 0.11582463836669922, 0.11588813018798828, 0.11592704010009766, 0.11582361602783203, 0.11600895690917969, 0.11590758514404297, 0.11593727874755859, 0.11590348815917968, 0.11586457824707032, 0.11585228729248047, 0.11603968048095703, 0.11589222717285157, 0.11578572845458984, 0.11583692932128906, 0.11589324951171875, 0.115810302734375, 0.11596902465820312, 0.11585842895507813, 0.11576319885253906, 0.11587686157226562]",tokens/s,8.488219831832597,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4494,7 +4494,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90a3-0c2bd8e43d2539372c3ce383;aae54c60-3ade-4e50-8322-39225b6869ff) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe250-6ce0a9154f0639e524d7a7aa;b35a9428-d4ed-4000-a422-75f2f99cef93) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -4609,7 +4609,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4677,7 +4677,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9607-41ba9ef3719a5ef44333e173;6e1d0866-525d-4d72-902f-a9b7db75ba93) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe7d5-5b8937da6851ffc85bf7583f;750a4985-ef91-4e9c-b71e-5daca2cb745e) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4806,7 +4806,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e911c-594492965643646a33cc1b0d;e3dbef5f-d415-429b-89c5-aaa0b518e2ba) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe2ce-1f5357a91fee9d2d53d1dded;549eabdc-0fa0-4f1a-8c00-5042777a04a9) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4896,7 +4896,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: XGLMForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -4935,7 +4935,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4626-4f1e8baa0bb78ed54ac396e6;c554b746-f9a4-417e-878c-5a1bddfe1ad5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe949-247396407db822ea611721e3;d38fe93d-c04f-4286-a2cf-736db3fbc714) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4966,7 +4966,7 @@ OSError: B is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5035,7 +5035,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9447-34eecfeb16a07c3870777e31;720e3a55-57d7-4718-b546-e1b850f52ed2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe60a-0bb32ce63deaf98b35ac8b8f;2cb2cb9e-b8ae-4fa9-863e-ae1a5dd888a5) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5066,7 +5066,7 @@ OSError: s is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5213,7 +5213,7 @@ ImportError: This modeling file requires the following packages that were not fo ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-13b-hf,meta-llama/Llama-2-13b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1511.15776,9028.763648,0.0,8382.316544,7939.283456,s,10,6.163353820800782,0.6163353820800781,0.0007470918632091674,0.6161679382324219,0.6170639099121094,0.6176385223388672,0.6180982122802734,"[0.6155059814453125, 0.618213134765625, 0.6159963989257813, 0.6163756713867188, 0.6165255126953125, 0.6156544799804687, 0.6169362182617187, 0.615810791015625, 0.6163394775390625, 0.6159961547851562]",tokens/s,415.3582731791616,kWh,7.268342706892226e-06,3.9827401672268555e-06,3.547922772975501e-05,4.673031060387409e-05,tokens/kWh,5478243.065193254,MB,1511.15776,9028.763648,0.0,8382.316544,8233.862144,s,10,338.06373828125,33.806373828125004,0.0007984609104005154,33.806359375,33.807255078125,33.8074712890625,33.8076442578125,"[33.80585546875, 33.80540234375, 33.80515234375, 33.8057265625, 33.80694140625, 33.80619140625, 33.807046875, 33.80720703125, 33.80652734375, 33.8076875]",tokens/s,1.8635539061449868,kWh,0.0003991127639015516,0.00021874844576166437,0.0019414585204858576,0.002559319730149073,tokens/kWh,24615.91619751645,,s,629,343.074377746582,0.5454282635080796,0.07326490605602001,0.5365985107421875,0.5368582153320313,0.53691865234375,1.152982744140625,"[0.53646435546875, 0.5365278930664062, 0.5363026123046875, 0.53646337890625, 0.5364152221679688, 0.5364141845703125, 0.5362595825195312, 0.5367633666992188, 0.5365442504882812, 0.536468505859375, 0.53633740234375, 0.5372354736328125, 0.5362872314453125, 0.5365032958984375, 0.5362749633789062, 0.5365442504882812, 0.53628515625, 0.5367285766601563, 0.53627392578125, 0.536791015625, 0.5363978271484375, 0.53667431640625, 0.536563720703125, 0.5366732788085937, 0.536838134765625, 0.5366947631835938, 0.536542236328125, 0.5366845703125, 0.5367408447265625, 0.5366077270507813, 0.5364725952148437, 0.536553466796875, 0.5364869384765625, 0.5367019653320313, 0.53639990234375, 0.5365125122070312, 0.5365892944335937, 0.5366753540039062, 0.5366527709960938, 0.5365985107421875, 0.5364367065429687, 0.5364971313476562, 0.5365022583007812, 0.536669189453125, 0.5366814575195312, 0.536880126953125, 0.5365841674804688, 0.5367070922851562, 0.5366159057617188, 0.5367337036132812, 0.536542236328125, 0.5366917114257812, 0.5367449340820313, 0.5366814575195312, 0.5366302490234375, 0.5369927978515625, 0.5365278930664062, 0.5368565673828125, 0.5363169555664062, 0.536754150390625, 0.5366220703125, 0.536722412109375, 1.1552481689453125, 0.5366220703125, 0.53665380859375, 0.5363517456054687, 0.5367039794921875, 0.5362749633789062, 0.5365791015625, 0.5363875732421876, 0.5367131958007813, 0.5366507568359375, 0.5364859008789062, 0.5363333129882812, 0.536616943359375, 0.5362718505859375, 0.5364049682617188, 0.5362831420898437, 0.5364920043945313, 0.5363026123046875, 0.5364510498046875, 0.5363446044921875, 0.5369446411132812, 0.536501220703125, 0.5365360717773437, 0.5365380859375, 0.5368473510742188, 0.5363783569335937, 0.5365238037109376, 0.536322021484375, 0.5364418334960938, 0.5365258178710938, 0.536616943359375, 0.5366353759765625, 0.5365247802734375, 0.5364019165039062, 0.5366640625, 0.5364889526367187, 0.5364951171875, 0.5365380859375, 0.5368422241210937, 0.5365944213867188, 0.5364357299804687, 0.5369364624023437, 0.5368053588867188, 0.5364827880859375, 0.5365667724609375, 0.536827880859375, 0.5369098510742187, 0.5367900390625, 0.5367030029296875, 0.536511474609375, 0.5367756958007812, 0.5364295654296874, 0.5365196533203125, 0.5367234497070312, 0.5368340454101562, 0.53665380859375, 0.53675830078125, 0.5367869262695313, 0.5368842163085937, 0.5364653930664063, 0.5368156127929687, 0.5366251220703125, 0.5368719482421875, 1.1530455322265625, 0.5364807739257812, 0.5364070434570313, 0.5360916748046874, 0.5364459228515625, 0.5364664306640625, 0.5362728881835938, 0.5364561767578125, 0.5363179321289062, 0.5363015747070312, 0.5362913208007812, 0.5364818115234375, 0.5364510498046875, 0.5365811157226562, 0.5363619995117187, 0.536431640625, 0.5364766845703125, 0.5363486938476563, 0.5366661376953125, 0.5364039916992187, 0.5365667724609375, 0.5368156127929687, 0.5365278930664062, 0.536563720703125, 0.5365821533203124, 0.53661083984375, 0.5367490844726562, 0.5365084228515625, 0.5365073852539063, 0.5364838256835938, 0.5369405517578125, 0.5366835327148437, 0.5364766845703125, 0.5364111328125, 0.536685546875, 0.5365514526367188, 0.5366098022460938, 0.5366026000976563, 0.536764404296875, 0.5367050170898438, 0.5365821533203124, 0.5366384887695312, 0.5366681518554688, 0.5365350341796875, 0.5368494262695312, 0.5367725830078125, 0.5366947631835938, 0.536479736328125, 0.5366835327148437, 0.5368370971679688, 0.5367193603515625, 0.5365616455078125, 0.5368872680664063, 0.5366405029296875, 0.53665380859375, 0.536580078125, 0.5366814575195312, 0.536880126953125, 0.5368361206054687, 0.536501220703125, 0.5368955078125, 0.5366251220703125, 0.536826904296875, 1.153522705078125, 0.53634765625, 0.5365821533203124, 0.5365667724609375, 0.5363824462890625, 0.536453125, 0.5365995483398438, 0.5364766845703125, 0.5363558349609375, 0.5361654052734375, 0.5363988647460938, 0.5362063598632812, 0.5364572143554688, 0.5363435668945312, 0.5363967895507813, 0.5363067016601563, 0.5365360717773437, 0.5364100952148437, 0.536468505859375, 0.5365360717773437, 0.5365391235351562, 0.5364019165039062, 0.5366036376953125, 0.5365155639648438, 0.5365933837890625, 0.5365678100585938, 0.5367091064453124, 0.536384521484375, 0.5364132080078124, 0.536500244140625, 0.5366210327148437, 0.5367265014648438, 0.5366651000976562, 0.536332275390625, 0.536911865234375, 0.5366210327148437, 0.5366445922851563, 0.5366497192382812, 0.5367306518554688, 0.5367572631835937, 0.5367490844726562, 0.5369937744140625, 0.5365831909179688, 0.5366712036132812, 0.5368893432617188, 0.5367122192382813, 0.5367859497070312, 0.5366405029296875, 0.5365596313476563, 0.5366138916015625, 0.5370316772460938, 0.5365647583007812, 0.5366998901367187, 0.53686474609375, 0.5368668212890625, 0.536426513671875, 0.5366978759765625, 0.5366220703125, 0.5366784057617188, 0.5365032958984375, 0.5368002319335937, 0.5369036865234375, 0.536853515625, 1.1535677490234375, 0.536469482421875, 0.536595458984375, 0.53646435546875, 0.5364049682617188, 0.5364408569335938, 0.536322021484375, 0.5363179321289062, 0.5363466186523438, 0.5362728881835938, 0.536573974609375, 0.5363169555664062, 0.53655859375, 0.536431640625, 0.5364777221679687, 0.5363599243164062, 0.5363363647460937, 0.53625244140625, 0.5365616455078125, 0.53661083984375, 0.5366302490234375, 0.5364889526367187, 0.53648486328125, 0.5369354248046875, 0.536395751953125, 0.5367234497070312, 0.5367879638671875, 0.5365196533203125, 0.5365933837890625, 0.5366507568359375, 0.5365043334960937, 0.5367255249023437, 0.53703271484375, 0.5365176391601563, 0.5368576049804688, 0.5369425659179687, 0.5368248291015625, 0.5367470092773438, 0.5365944213867188, 0.5366569213867187, 0.5365504150390625, 0.5365616455078125, 0.5368729858398438, 0.5366384887695312, 0.5368923950195312, 0.5366179809570313, 0.5368237915039062, 0.53652685546875, 0.536647705078125, 0.5364500732421875, 0.5366619873046875, 0.5367900390625, 0.53671728515625, 0.5366435546875, 0.536826904296875, 0.5366722412109375, 0.5369088134765625, 0.5367930908203125, 0.536685546875, 0.53656884765625, 0.5368576049804688, 0.5367675170898437, 0.5367654418945312, 1.152680908203125, 0.536322021484375, 0.5365278930664062, 0.5362452392578125, 0.53627392578125, 0.5367684936523438, 0.5364306030273438, 0.5364336547851563, 0.5366886596679687, 0.5364623413085937, 0.5366046752929687, 0.5362985229492188, 0.5363906860351563, 0.5365616455078125, 0.5363148803710938, 0.5364786987304687, 0.5363906860351563, 0.5362974853515625, 0.5365278930664062, 0.5366231079101562, 0.53686474609375, 0.5363804321289063, 0.5363988647460938, 0.5364992065429688, 0.5364520874023437, 0.5365924072265625, 0.5365821533203124, 0.5365094604492188, 0.5365350341796875, 0.5365524291992188, 0.5366722412109375, 0.5365995483398438, 0.5367675170898437, 0.5365350341796875, 0.5367684936523438, 0.5365657348632813, 0.5366814575195312, 0.5366005859375, 0.5367756958007812, 0.5364551391601563, 0.5366712036132812, 0.53652685546875, 0.5366896362304687, 0.5367971801757813, 0.5366548461914062, 0.5366865844726563, 0.5367285766601563, 0.536542236328125, 0.5367767333984375, 0.5366026000976563, 0.53661083984375, 0.536585205078125, 0.53707470703125, 0.5368156127929687, 0.5367285766601563, 0.5366968383789062, 0.536796142578125, 0.5366876220703125, 0.5368606567382812, 0.5364725952148437, 0.53682177734375, 0.5371248779296875, 0.5368145751953125, 1.154049072265625, 0.536754150390625, 0.536500244140625, 0.5364705200195312, 0.536279052734375, 0.5364592895507813, 0.536521728515625, 0.5364541625976562, 0.536553466796875, 0.536447998046875, 0.5365278930664062, 0.5361920166015625, 0.5366558837890625, 0.5363538208007812, 0.536290283203125, 0.5365933837890625, 0.5363732299804688, 0.5363538208007812, 0.5366149291992187, 0.5364777221679687, 0.5366548461914062, 0.5363466186523438, 0.5363824462890625, 0.5367030029296875, 0.53656884765625, 0.5366384887695312, 0.536764404296875, 0.5365238037109376, 0.5364807739257812, 0.5365186767578125, 0.5366179809570313, 0.536680419921875, 0.536521728515625, 0.536573974609375, 0.5373849487304687, 0.53646337890625, 0.5365442504882812, 0.5368811645507813, 0.5369241333007813, 0.5366845703125, 0.536647705078125, 0.53655859375, 0.5366302490234375, 0.5366517944335938, 0.5368319702148437, 0.5368565673828125, 0.5366558837890625, 0.5367070922851562, 0.5367879638671875, 0.5363773193359375, 0.5366231079101562, 0.5366876220703125, 0.5366527709960938, 0.536853515625, 0.5369262084960937, 0.53659033203125, 0.53675830078125, 0.5368442993164062, 0.5366282348632813, 0.536585205078125, 0.536901611328125, 0.5365821533203124, 0.5368975219726563, 1.1528212890625, 0.5364930419921875, 0.5364346923828125, 0.5364602661132812, 0.5364786987304687, 0.5365841674804688, 0.536700927734375, 0.5364203491210937, 0.5365206909179687, 0.5362360229492188, 0.5365258178710938, 0.536453125, 0.5364039916992187, 0.5362821044921875, 0.536416259765625, 0.536489990234375, 0.5365043334960937, 0.5367490844726562, 0.5366927490234376, 0.536537109375, 0.536605712890625, 0.5364295654296874, 0.5365186767578125, 0.5366405029296875, 0.5365330200195313, 0.5366036376953125, 0.5368033447265625, 0.5365196533203125, 0.536685546875, 0.5364827880859375, 0.536585205078125, 0.5366415405273437, 0.5366220703125, 0.5365125122070312, 0.5367930908203125, 0.5364500732421875, 0.536690673828125, 0.5365841674804688, 0.5366343383789063, 0.5364725952148437, 0.5364572143554688, 0.5365831909179688, 0.5368606567382812, 0.536900634765625, 0.536642578125, 0.5369200439453125, 0.536875, 0.536627197265625, 0.5367439575195313, 0.5369313354492188, 0.5367183227539063, 0.5367459716796875, 0.5367490844726562, 0.5367183227539063, 0.5367930908203125, 0.5367992553710937, 0.5367244873046875, 0.5365708618164062, 0.5366456298828125, 0.536489990234375, 0.5367398681640625, 0.5367736206054687, 0.536911865234375, 1.153685546875, 0.5364326171875, 0.5368453369140626, 0.5364326171875, 0.5363701782226562, 0.5364049682617188, 0.5363363647460937, 0.5364193115234375, 0.53652685546875, 0.5361500244140625, 0.5366282348632813, 0.53642138671875, 0.5365309448242187, 0.5364653930664063, 0.5363671264648437, 0.5362626342773438, 0.5362943725585938, 0.5363496704101562, 0.5363967895507813, 0.5364561767578125, 0.536637451171875, 0.536521728515625, 0.5364653930664063, 0.5365862426757813, 0.5365125122070312, 0.5366251220703125, 0.5366927490234376, 0.5372610473632813, 0.536563720703125, 0.5365821533203124, 0.5365442504882812, 0.5365206909179687, 0.536384521484375, 0.5364254760742188, 0.5367828369140625, 0.5363783569335937, 0.5367131958007813, 0.536658935546875, 0.5366210327148437, 0.5367654418945312, 0.5365166015625, 0.5369682006835937, 0.53671728515625, 0.5366220703125, 0.53667431640625, 0.536658935546875, 0.5370972290039062, 0.536573974609375, 0.5367091064453124, 0.5366876220703125, 0.5367388305664063, 0.5366159057617188, 0.5368125610351563, 0.5367398681640625, 0.5367869262695313, 0.5366179809570313, 0.5367490844726562, 0.5368197021484375, 0.5369180297851562, 0.5365380859375, 0.5368033447265625, 0.5367613525390625, 0.5369190673828125, 1.1530567626953125, 0.5364326171875, 0.536469482421875, 0.5364152221679688, 0.5364408569335938, 0.5365975341796875, 0.53661083984375, 0.536395751953125, 0.5363558349609375, 0.5362308959960937, 0.536479736328125, 0.5363179321289062, 0.5364981689453125, 0.536501220703125, 0.5363660888671875, 0.5362943725585938, 0.536474609375, 0.5364049682617188, 0.5366415405273437, 0.5366968383789062, 0.536627197265625, 0.536537109375, 0.5364306030273438, 0.5364510498046875, 0.536933349609375, 0.5366517944335938, 0.536753173828125, 0.5365309448242187, 0.5369077758789063, 0.5366138916015625, 0.5365667724609375, 0.5371023559570313, 0.5366179809570313, 0.5367234497070312, 0.5368606567382812, 0.5366128540039062, 0.5367398681640625, 0.5365196533203125, 0.5367510986328125, 0.5366845703125, 0.536595458984375, 0.5366077270507813, 0.5368237915039062, 0.5367767333984375, 0.5368995971679688, 0.5368330078125, 0.5367787475585938, 0.536690673828125, 0.5368862915039062, 0.5365944213867188, 0.5366364135742188, 0.5365504150390625, 0.53686474609375, 0.5368176879882812, 0.536826904296875, 0.5364766845703125, 0.536711181640625, 0.5368995971679688, 0.5367941284179687, 0.5364572143554688, 0.53665380859375, 0.536563720703125, 0.5366558837890625]",tokens/s,1.8334216741321963,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,4113.342464,5287.44448,0.0,4640.997376,4330.668544,s,10,3.2019090576171876,0.3201909057617187,0.0017134996701674667,0.31955131530761716,0.3223881805419922,0.3223929733276367,0.32239680755615235,"[0.3185635986328125, 0.3218643493652344, 0.318020263671875, 0.3190321350097656, 0.3193435974121094, 0.3183670349121094, 0.319759033203125, 0.32239776611328125, 0.3221741638183594, 0.3223871154785156]",tokens/s,799.5230201525817,kWh,3.7596130847103066e-06,2.060095877009083e-06,1.8658895135439252e-05,2.447860409715864e-05,tokens/kWh,10458112.6841998,MB,4113.342464,5287.44448,0.0,4640.997376,4515.667456,s,10,175.162884765625,17.5162884765625,0.0013239684065771513,17.516396484375,17.5170955078125,17.51822841796875,17.51913474609375,"[17.5166953125, 17.51646484375, 17.51684375, 17.515330078125, 17.516587890625, 17.51393359375, 17.519361328125, 17.51530859375, 17.516328125, 17.51603125]",tokens/s,3.596652343577039,kWh,0.00020683056154391835,0.0001133576330349002,0.0010022863660645621,0.0013224745606433806,tokens/kWh,47637.96739451128,,s,629,177.76777615356434,0.2826196759198163,0.038106541032305166,0.27801907348632815,0.2782052307128906,0.2782926879882812,0.5975155737304687,"[0.27855975341796874, 0.27801702880859375, 0.2778101806640625, 0.2780354614257812, 0.27816244506835935, 0.2778398742675781, 0.2780078125, 0.27787161254882814, 0.2779852905273438, 0.27780914306640625, 0.27788491821289063, 0.2780006408691406, 0.2780405883789063, 0.27786956787109374, 0.27781939697265623, 0.27792691040039064, 0.27795147705078127, 0.27795660400390626, 0.2780201110839844, 0.27802520751953125, 0.2780088195800781, 0.2779525146484375, 0.2779207763671875, 0.2779402160644531, 0.2781061096191406, 0.27823513793945315, 0.2779893798828125, 0.27797503662109374, 0.27814501953125, 0.2779105224609375, 0.2781542358398438, 0.2779914245605469, 0.2781327209472656, 0.27813067626953125, 0.2779105224609375, 0.2780354614257812, 0.27810202026367187, 0.27796890258789064, 0.2780477294921875, 0.27807232666015624, 0.2782822265625, 0.2780149841308594, 0.27799859619140627, 0.27797402954101563, 0.2781419372558594, 0.2780426330566406, 0.2779207763671875, 0.278086669921875, 0.2782423095703125, 0.27799755859375, 0.2780201110839844, 0.27810202026367187, 0.2780508117675781, 0.27797607421875, 0.2779842529296875, 0.27804568481445313, 0.27818701171875, 0.2779914245605469, 0.27808255004882815, 0.2781388854980469, 0.2782474365234375, 0.2780313720703125, 0.6001817626953125, 0.27799859619140627, 0.27790234375, 0.2779320373535156, 0.27799655151367186, 0.27801907348632815, 0.2778521728515625, 0.27775384521484375, 0.27788595581054687, 0.2779473876953125, 0.27779583740234376, 0.2782218322753906, 0.27791156005859374, 0.27793304443359373, 0.277886962890625, 0.2779207763671875, 0.27789208984375, 0.27796890258789064, 0.27791461181640625, 0.278029296875, 0.27789312744140626, 0.2780139465332031, 0.2779402160644531, 0.2780508117675781, 0.2779996032714844, 0.2780682373046875, 0.27781939697265623, 0.2778675231933594, 0.2782105712890625, 0.27816244506835935, 0.2780088195800781, 0.27801089477539065, 0.27811737060546876, 0.27817166137695315, 0.27833856201171875, 0.27794842529296876, 0.27794329833984377, 0.27803955078125, 0.2779781188964844, 0.2778316650390625, 0.2780487670898438, 0.2780835876464844, 0.2780579833984375, 0.2780610656738281, 0.2779648132324219, 0.2781368408203125, 0.278066162109375, 0.27846142578125, 0.2780498046875, 0.2782576599121094, 0.2779914245605469, 0.27807333374023435, 0.27804568481445313, 0.27823309326171874, 0.27794842529296876, 0.2780641174316406, 0.27811431884765625, 0.2781614074707031, 0.27800985717773435, 0.2780682373046875, 0.278029296875, 0.2781337585449219, 0.27806719970703125, 0.5975459594726562, 0.2780262451171875, 0.2778736572265625, 0.2778173522949219, 0.277960693359375, 0.2780528564453125, 0.27834878540039065, 0.2778900451660156, 0.27787673950195313, 0.2779422607421875, 0.27794638061523436, 0.2780139465332031, 0.27793408203125, 0.27799655151367186, 0.27793304443359373, 0.2779371643066406, 0.2779084777832031, 0.2779934692382813, 0.27786343383789064, 0.27802520751953125, 0.2779924621582031, 0.2782013549804688, 0.27794329833984377, 0.2780282897949219, 0.2780559387207031, 0.27812966918945314, 0.27786648559570315, 0.2779525146484375, 0.27802725219726565, 0.27815731811523436, 0.27795455932617186, 0.27812249755859375, 0.2780569458007813, 0.27815731811523436, 0.277923828125, 0.2779637756347656, 0.27835391235351564, 0.27807437133789065, 0.27791256713867185, 0.2779658203125, 0.2780928039550781, 0.27812966918945314, 0.2783221740722656, 0.2781583251953125, 0.2779832458496094, 0.27802520751953125, 0.27819622802734373, 0.2780426330566406, 0.2780415954589844, 0.27818292236328124, 0.27791156005859374, 0.2779729919433594, 0.2780845947265625, 0.27821261596679686, 0.2780334167480469, 0.27809588623046877, 0.27812863159179685, 0.2781747131347656, 0.2780631103515625, 0.2779945068359375, 0.27803033447265624, 0.2781747131347656, 0.2780282897949219, 0.5974374389648438, 0.27794943237304687, 0.27812762451171874, 0.27768115234375, 0.2779719543457031, 0.27803955078125, 0.27782247924804687, 0.2778367919921875, 0.27792691040039064, 0.27802725219726565, 0.2777835388183594, 0.2779033508300781, 0.27785626220703125, 0.27808563232421873, 0.2782443542480469, 0.2781737060546875, 0.2779832458496094, 0.2779924621582031, 0.2779576416015625, 0.27801907348632815, 0.27789413452148437, 0.2780528564453125, 0.27785626220703125, 0.27797503662109374, 0.27792691040039064, 0.27814501953125, 0.2779156494140625, 0.27798016357421873, 0.2780313720703125, 0.2781163635253906, 0.27802420043945314, 0.2779832458496094, 0.27799755859375, 0.27810714721679686, 0.2778880004882813, 0.2779371643066406, 0.27798834228515623, 0.27807949829101564, 0.27803033447265624, 0.2779576416015625, 0.2780262451171875, 0.2780784606933594, 0.27793408203125, 0.27805902099609375, 0.27813784790039064, 0.2781317138671875, 0.27809381103515624, 0.2779576416015625, 0.27808154296875, 0.2782586975097656, 0.2782494812011719, 0.2780610656738281, 0.27814501953125, 0.2781214599609375, 0.2779207763671875, 0.277923828125, 0.2779832458496094, 0.27808154296875, 0.27801907348632815, 0.27799755859375, 0.2781317138671875, 0.27814910888671873, 0.2781051025390625, 0.5979115600585938, 0.2780415954589844, 0.2779361267089844, 0.27780813598632814, 0.27787774658203124, 0.2779586486816406, 0.2779361267089844, 0.27787774658203124, 0.278002685546875, 0.278002685546875, 0.2778736572265625, 0.27790234375, 0.2779648132324219, 0.2780006408691406, 0.27782861328125, 0.277970947265625, 0.2783703002929688, 0.2780405883789063, 0.27791360473632815, 0.27801702880859375, 0.2779422607421875, 0.27796786499023435, 0.2783999938964844, 0.27802520751953125, 0.278012939453125, 0.2780917663574219, 0.277907470703125, 0.2779371643066406, 0.27806207275390626, 0.2780498046875, 0.27799038696289063, 0.2780682373046875, 0.27811328125, 0.2783170471191406, 0.27801190185546876, 0.27803237915039064, 0.2780487670898438, 0.27803237915039064, 0.2779955139160156, 0.27786138916015624, 0.2780262451171875, 0.2781327209472656, 0.27797503662109374, 0.2780948486328125, 0.2779361267089844, 0.27810714721679686, 0.27803851318359374, 0.27794943237304687, 0.2780702819824219, 0.27821774291992185, 0.2779535217285156, 0.2780139465332031, 0.27819314575195314, 0.2782003173828125, 0.2779729919433594, 0.27806515502929685, 0.278012939453125, 0.2781890563964844, 0.27835595703125, 0.27808767700195314, 0.27808563232421873, 0.2781542358398438, 0.278086669921875, 0.596664306640625, 0.27794638061523436, 0.27794125366210937, 0.2780057678222656, 0.27801803588867186, 0.27811328125, 0.27796173095703125, 0.277855224609375, 0.2778839111328125, 0.278002685546875, 0.2778757019042969, 0.27769650268554685, 0.2778746948242187, 0.2780262451171875, 0.27781427001953124, 0.2780006408691406, 0.27796685791015624, 0.27805389404296876, 0.27779379272460936, 0.2778880004882813, 0.27790438842773435, 0.27795559692382815, 0.2778480529785156, 0.27800167846679685, 0.27791360473632815, 0.27811019897460937, 0.2778890380859375, 0.27791973876953124, 0.2779453430175781, 0.2782668762207031, 0.27817062377929686, 0.27803750610351563, 0.2780006408691406, 0.2780569458007813, 0.2779402160644531, 0.2778787841796875, 0.2779443054199219, 0.2780559387207031, 0.27797708129882814, 0.2779002990722656, 0.2781419372558594, 0.2780415954589844, 0.277960693359375, 0.27822592163085935, 0.27799755859375, 0.2780968933105469, 0.27796890258789064, 0.27799655151367186, 0.278034423828125, 0.2782085266113281, 0.2779279479980469, 0.27796890258789064, 0.27808050537109374, 0.27819110107421874, 0.2779781188964844, 0.2779361267089844, 0.2780231628417969, 0.2781419372558594, 0.2780088195800781, 0.27789926147460936, 0.27808563232421873, 0.2782248840332031, 0.2780631103515625, 0.5981163330078125, 0.2783621215820313, 0.2783006591796875, 0.2778358154296875, 0.27795965576171877, 0.2779729919433594, 0.2779207763671875, 0.2778347473144531, 0.27787774658203124, 0.27798016357421873, 0.277865478515625, 0.2778245239257813, 0.2778685302734375, 0.27808563232421873, 0.27831195068359377, 0.27789208984375, 0.2779105224609375, 0.27801907348632815, 0.2779402160644531, 0.27796890258789064, 0.27797503662109374, 0.27805184936523436, 0.277907470703125, 0.2780078125, 0.27791873168945314, 0.2780928039550781, 0.2779361267089844, 0.27801702880859375, 0.2780907592773437, 0.27824127197265625, 0.27799038696289063, 0.2781542358398438, 0.2781747131347656, 0.27819827270507813, 0.27798019409179686, 0.27835592651367186, 0.2779781188964844, 0.2782853088378906, 0.27827609252929686, 0.2779586486816406, 0.27801190185546876, 0.27814093017578123, 0.27794329833984377, 0.27867544555664064, 0.27796786499023435, 0.27813479614257813, 0.27802520751953125, 0.27811532592773436, 0.27813580322265624, 0.2783088684082031, 0.2785464172363281, 0.278076416015625, 0.2782156677246094, 0.27824331665039065, 0.27822900390625, 0.27794329833984377, 0.27808767700195314, 0.27831195068359377, 0.27817266845703126, 0.27806719970703125, 0.27806002807617186, 0.2782515258789062, 0.2780887145996094, 0.60076953125, 0.2780528564453125, 0.27787774658203124, 0.2780835876464844, 0.2779893798828125, 0.2780364685058594, 0.27775180053710935, 0.2778787841796875, 0.2780692443847656, 0.2780835876464844, 0.2782095336914063, 0.27780813598632814, 0.2778828735351562, 0.27790234375, 0.27783782958984377, 0.2779309997558594, 0.27794125366210937, 0.2780784606933594, 0.277897216796875, 0.2780211181640625, 0.2779945068359375, 0.27791973876953124, 0.2779525146484375, 0.278066162109375, 0.27791256713867185, 0.2780282897949219, 0.27787673950195313, 0.2778757019042969, 0.27794842529296876, 0.27817779541015625, 0.2779361267089844, 0.2780845947265625, 0.27806207275390626, 0.27812249755859375, 0.27798834228515623, 0.27791668701171873, 0.27791360473632815, 0.277960693359375, 0.2780917663574219, 0.2781583251953125, 0.27809588623046877, 0.27802725219726565, 0.27801190185546876, 0.2780579833984375, 0.27796273803710936, 0.27820440673828123, 0.2784163818359375, 0.2779811706542969, 0.278076416015625, 0.2782392272949219, 0.27798220825195313, 0.2780231628417969, 0.27797708129882814, 0.2781951904296875, 0.2779914245605469, 0.2780313720703125, 0.277970947265625, 0.2780968933105469, 0.2780917663574219, 0.278002685546875, 0.2780364685058594, 0.27814501953125, 0.278044677734375, 0.6006200561523437, 0.27797503662109374, 0.2778890380859375, 0.2780282897949219, 0.2779402160644531, 0.27803851318359374, 0.277928955078125, 0.27785931396484376, 0.2780006408691406, 0.2779156494140625, 0.2777651062011719, 0.2779658203125, 0.27800677490234377, 0.2779637756347656, 0.2778880004882813, 0.27796173095703125, 0.2779176940917969, 0.2780159912109375, 0.27834878540039065, 0.27800985717773435, 0.27794842529296876, 0.2780631103515625, 0.2779105224609375, 0.27805490112304687, 0.27783578491210936, 0.2781808776855469, 0.2778818664550781, 0.27799755859375, 0.2780211181640625, 0.27812045288085935, 0.27800473022460936, 0.27808154296875, 0.2780897216796875, 0.2780979309082031, 0.277960693359375, 0.27794329833984377, 0.2780313720703125, 0.27806719970703125, 0.2780088195800781, 0.2779719543457031, 0.27808563232421873, 0.27815731811523436, 0.2779842529296875, 0.2779084777832031, 0.2779156494140625, 0.2780907592773437, 0.27803750610351563, 0.2781327209472656, 0.27808563232421873, 0.27822283935546877, 0.27808050537109374, 0.27801803588867186, 0.2780968933105469, 0.278297607421875, 0.2783078308105469, 0.2780487670898438, 0.27810302734375, 0.2782310485839844, 0.2781327209472656, 0.27803851318359374, 0.2779381713867187, 0.2782033996582031, 0.27808050537109374, 0.6008156127929688, 0.2780989379882812, 0.27795147705078127, 0.2778542175292969, 0.2780559387207031, 0.2779955139160156, 0.2778808288574219, 0.2778542175292969, 0.2780201110839844, 0.27807745361328123, 0.27784295654296876, 0.27786956787109374, 0.27791156005859374, 0.2778890380859375, 0.27790234375, 0.27802520751953125, 0.27789926147460936, 0.27797402954101563, 0.2777374572753906, 0.27824026489257814, 0.2779862976074219, 0.27802420043945314, 0.2778818664550781, 0.27795455932617186, 0.27798220825195313, 0.2781051025390625, 0.278150146484375, 0.277886962890625, 0.2780579833984375, 0.27816754150390627, 0.2780149841308594, 0.2779729919433594, 0.27802420043945314, 0.27813784790039064, 0.27803237915039064, 0.27795147705078127, 0.27801089477539065, 0.278044677734375, 0.27811224365234377, 0.27802215576171874, 0.2780487670898438, 0.27812863159179685, 0.27805389404296876, 0.27803955078125, 0.2779996032714844, 0.27808767700195314, 0.2780692443847656, 0.27806207275390626, 0.27810302734375, 0.27824639892578124, 0.27794125366210937, 0.27805389404296876, 0.2780415954589844, 0.2782115783691406, 0.2779934692382813, 0.27796786499023435, 0.2781685791015625, 0.2781685791015625, 0.27801089477539065, 0.2780498046875, 0.2780282897949219, 0.27819314575195314, 0.27838156127929686]",tokens/s,3.538324063055383,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5416,7 +5416,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e93f6-696d33247adc22470daa6867;fa0ec27b-b5d0-4552-826a-0c85cd4f63b7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe5ba-594b4f67776af4a93ef87e21;181a08e0-34eb-4fb9-b6c5-a01b759f9412) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5447,7 +5447,7 @@ OSError: m is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5486,7 +5486,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46ea-61d76feb58130df55d641f84;7f9e84b9-71a3-44d7-85f1-712870cd7708) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fea21-1a7ad53f3b02c10807d625b2;7e6ab414-c109-47d3-9e9f-5434bcaaf964) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5517,7 +5517,7 @@ OSError: 1 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5556,7 +5556,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d469b-66f77aa61a43508b5c626b31;358c1803-3172-4738-9732-66dd1e6b88db) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe9c4-55b64e7d01c3a9104d0851f2;6e1a7e56-f631-4747-8fa5-343dc2ea6083) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5616,7 +5616,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5684,7 +5684,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e96e7-0b91509115cf686f56c321a4;de7ff37d-21b5-4702-88dc-010cad13155b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe8bc-421465a40cc0148d2b4045df;fe0bd07f-fbfc-4915-84ba-aa659493f832) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5716,7 +5716,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1595.359232,2173.17376,0.0,1526.726656,1394.863104,s,10,0.6623263702392578,0.06623263702392578,0.00023606018108147416,0.06618220901489258,0.06632227020263672,0.06662019195556641,0.06685852935791016,"[0.0669181137084961, 0.06611609649658202, 0.06618179321289062, 0.06608083343505859, 0.06618262481689453, 0.0661866226196289, 0.06623158264160156, 0.06609548950195313, 0.0660771484375, 0.06625606536865235]",tokens/s,3865.163935833069,kWh,7.826929992836482e-07,4.287589798161134e-07,3.6917340056288315e-06,4.903185984728593e-06,tokens/kWh,52210950.34888227,MB,1595.359232,2173.17376,0.0,1526.726656,1491.575808,s,10,35.95646752929687,3.5956467529296865,0.0008135899821663264,3.595584716796875,3.596604443359375,3.5969305175781248,3.5971913769531247,"[3.597256591796875, 3.595404052734375, 3.595923583984375, 3.595748046875, 3.59598876953125, 3.59542138671875, 3.594245849609375, 3.596531982421875, 3.5951494140625, 3.5947978515625]",tokens/s,17.521187238059028,kWh,4.238840278257247e-05,2.323122906644022e-05,0.00019878792797797785,0.0002644075598269905,tokens/kWh,238268.52772750793,,s,629,36.49370518112184,0.058018609190972695,0.007909574934194006,0.0570511360168457,0.05715558242797852,0.05729751129150391,0.12360420166015626,"[0.05711052703857422, 0.05699379348754883, 0.05696409606933594, 0.05699071884155273, 0.05715558242797852, 0.0573040657043457, 0.057024513244628906, 0.056997886657714845, 0.057022464752197265, 0.057194496154785154, 0.05726105499267578, 0.0569989128112793, 0.056989696502685545, 0.05699686431884766, 0.05699379348754883, 0.05700412750244141, 0.05706742477416992, 0.057040897369384766, 0.05701017761230469, 0.05697536087036133, 0.05695904159545898, 0.05706028747558594, 0.05730614471435547, 0.05744534301757812, 0.05733171081542969, 0.05712076950073242, 0.057059326171875, 0.05705318450927734, 0.05705011367797851, 0.05711769485473633, 0.05714742279052734, 0.056989662170410155, 0.05705420684814453, 0.05701836776733398, 0.05701939010620117, 0.0570030403137207, 0.05712380981445313, 0.05712486267089844, 0.057032703399658206, 0.05707369613647461, 0.05707772827148438, 0.057095169067382816, 0.057022464752197265, 0.0570695686340332, 0.05705011367797851, 0.057033729553222653, 0.05706752014160156, 0.05705011367797851, 0.05706547164916992, 0.057128959655761716, 0.05711769485473633, 0.057082878112792966, 0.057059326171875, 0.057078784942626956, 0.0570777587890625, 0.05708697509765625, 0.05709827041625976, 0.05704188919067383, 0.05709619140625, 0.05708083343505859, 0.05706752014160156, 0.05705011367797851, 0.12362137603759765, 0.056991840362548826, 0.05700495910644531, 0.05700812911987305, 0.05699379348754883, 0.05698457717895508, 0.05697945785522461, 0.056925182342529294, 0.05695897674560547, 0.056953857421875, 0.0570327377319336, 0.05699580764770508, 0.056987648010253904, 0.05701324844360352, 0.05697945785522461, 0.05703987121582031, 0.05702041625976562, 0.057047039031982424, 0.057025535583496094, 0.056981502532958986, 0.05700198364257812, 0.05697439956665039, 0.05703571319580078, 0.0569796142578125, 0.05699465560913086, 0.057000961303710934, 0.05707059097290039, 0.057035839080810544, 0.05702860641479492, 0.057064384460449216, 0.05706752014160156, 0.05705625534057617, 0.05702656173706055, 0.05705215835571289, 0.05708499145507812, 0.057032638549804685, 0.0570695686340332, 0.05703785705566406, 0.057029598236083986, 0.05702963256835938, 0.057057281494140626, 0.05709823989868164, 0.05710847854614258, 0.05726105499267578, 0.05748339080810547, 0.05710015869140625, 0.057079807281494144, 0.05710438537597656, 0.05715251159667969, 0.05705420684814453, 0.057055233001708984, 0.05708902359008789, 0.057047039031982424, 0.05703475189208984, 0.0572221450805664, 0.05709632110595703, 0.057057151794433596, 0.05707574462890625, 0.05714326477050781, 0.057073665618896485, 0.05711872100830078, 0.05728361511230469, 0.05749756622314453, 0.12373709106445313, 0.05693644714355469, 0.05699993515014649, 0.05710847854614258, 0.05697843170166016, 0.056965118408203126, 0.05702041625976562, 0.05701324844360352, 0.05697228622436523, 0.05692940902709961, 0.05695987319946289, 0.05700812911987305, 0.056997886657714845, 0.057006080627441405, 0.05700198364257812, 0.057027584075927736, 0.057040897369384766, 0.05698252868652344, 0.05696716690063477, 0.05702143859863281, 0.05704297637939453, 0.05697635269165039, 0.057030654907226565, 0.05721193695068359, 0.057038814544677734, 0.057003009796142576, 0.057043968200683595, 0.057062400817871096, 0.05704499053955078, 0.057024513244628906, 0.057032703399658206, 0.05708390426635742, 0.05703680038452148, 0.05703475189208984, 0.057022464752197265, 0.05706854248046875, 0.05707468795776367, 0.05700812911987305, 0.05709823989868164, 0.05709823989868164, 0.057055233001708984, 0.05706038284301758, 0.05729481506347656, 0.05743308639526367, 0.05715251159667969, 0.05734604644775391, 0.05720985412597656, 0.05734604644775391, 0.05710745620727539, 0.05708595275878906, 0.057115646362304685, 0.05706649780273437, 0.057112575531005856, 0.05718425750732422, 0.057188350677490236, 0.05706444931030273, 0.05712384033203125, 0.057057342529296874, 0.05707155227661133, 0.05709107208251953, 0.057095169067382816, 0.057075809478759766, 0.05712067031860352, 0.12360908508300782, 0.05700505447387695, 0.05700719833374023, 0.056930206298828126, 0.05698457717895508, 0.057017345428466794, 0.05741567993164062, 0.057556991577148435, 0.05706649780273437, 0.05702963256835938, 0.056997886657714845, 0.0569804801940918, 0.05706854248046875, 0.057014270782470705, 0.056998046875, 0.05701820755004883, 0.05705830383300781, 0.05699379348754883, 0.05703782272338867, 0.056992767333984375, 0.057009151458740234, 0.05701529693603516, 0.057009151458740234, 0.05697552108764648, 0.057010017395019534, 0.057027584075927736, 0.05704499053955078, 0.05703987121582031, 0.05708492660522461, 0.057025535583496094, 0.057040897369384766, 0.05707468795776367, 0.05709209442138672, 0.05703680038452148, 0.057041919708251954, 0.05706041717529297, 0.057095104217529294, 0.05701849746704102, 0.05703667068481445, 0.05710028839111328, 0.05714022445678711, 0.05703680038452148, 0.05714022445678711, 0.05715865707397461, 0.057128959655761716, 0.05706444931030273, 0.057076736450195314, 0.057082878112792966, 0.05706649780273437, 0.05704816055297852, 0.05709507369995117, 0.05713407897949219, 0.057079807281494144, 0.05707263946533203, 0.05713612747192383, 0.05708185577392578, 0.05710028839111328, 0.05710745620727539, 0.05711974334716797, 0.05707468795776367, 0.05711155319213867, 0.0570695686340332, 0.05712691116333008, 0.12370432281494141, 0.056957950592041014, 0.05696409606933594, 0.056973312377929686, 0.05696921539306641, 0.056987648010253904, 0.056957950592041014, 0.05700710296630859, 0.057003009796142576, 0.05695187377929688, 0.057024513244628906, 0.05704697418212891, 0.05699993515014649, 0.056989696502685545, 0.05700931167602539, 0.05735203170776367, 0.057017345428466794, 0.05703168106079102, 0.056995838165283204, 0.05699686431884766, 0.05700198364257812, 0.057017345428466794, 0.057009151458740234, 0.05702876663208008, 0.05699260711669922, 0.05698867034912109, 0.057145343780517575, 0.057256961822509764, 0.057063552856445314, 0.05710015869140625, 0.057556991577148435, 0.057073665618896485, 0.05703782272338867, 0.057024513244628906, 0.05715251159667969, 0.0570511360168457, 0.05703987121582031, 0.057071617126464844, 0.05708902359008789, 0.0570777587890625, 0.05708185577392578, 0.05713919830322266, 0.0570777587890625, 0.05705625534057617, 0.05719859313964844, 0.05732044982910156, 0.05711769485473633, 0.05706547164916992, 0.05709120178222656, 0.05709708786010742, 0.05709721755981445, 0.0570511360168457, 0.05716275024414062, 0.057096225738525394, 0.057081825256347654, 0.05710745620727539, 0.057106433868408205, 0.05708390426635742, 0.05707468795776367, 0.05706752014160156, 0.05710131072998047, 0.0570880012512207, 0.057113662719726566, 0.12363871765136719, 0.05696819305419922, 0.056992767333984375, 0.05697536087036133, 0.05693644714355469, 0.057076736450195314, 0.05710662460327148, 0.0570263671875, 0.05702860641479492, 0.05704601669311524, 0.05699071884155273, 0.05697433471679687, 0.056997886657714845, 0.05704499053955078, 0.057030784606933595, 0.056956798553466796, 0.05703168106079102, 0.05705830383300781, 0.05708390426635742, 0.05699174499511719, 0.05696819305419922, 0.05703782272338867, 0.05703683090209961, 0.05704806518554688, 0.05700707244873047, 0.05702860641479492, 0.05705625534057617, 0.057014270782470705, 0.05703577423095703, 0.05702143859863281, 0.056995838165283204, 0.05702143859863281, 0.05706854248046875, 0.057047039031982424, 0.05701324844360352, 0.05704601669311524, 0.05715763092041016, 0.057232383728027345, 0.05703987121582031, 0.05702963256835938, 0.05712998580932617, 0.05705011367797851, 0.05703081512451172, 0.05702742385864258, 0.05706547164916992, 0.057059326171875, 0.05707379150390625, 0.05710835266113281, 0.05705625534057617, 0.057079936981201174, 0.057076606750488285, 0.057164798736572264, 0.057196544647216796, 0.05732966232299805, 0.057592830657958984, 0.057112575531005856, 0.05710335922241211, 0.057071617126464844, 0.05714432144165039, 0.05708902359008789, 0.057055233001708984, 0.057073665618896485, 0.05708595275878906, 0.12372172546386719, 0.05698867034912109, 0.056992767333984375, 0.05697536087036133, 0.057003009796142576, 0.0569620475769043, 0.05696921539306641, 0.05695388793945313, 0.05703164672851563, 0.05694668960571289, 0.056956928253173826, 0.05697228622436523, 0.05703577423095703, 0.05697228622436523, 0.05700812911987305, 0.05702143859863281, 0.05709721755981445, 0.05698867034912109, 0.05698355102539063, 0.056957950592041014, 0.05701939010620117, 0.056981502532958986, 0.057025535583496094, 0.057180160522460936, 0.057044063568115234, 0.056990623474121094, 0.057009151458740234, 0.057024513244628906, 0.05704806518554688, 0.05703782272338867, 0.05708185577392578, 0.057025535583496094, 0.057060352325439455, 0.05702348709106445, 0.0570777587890625, 0.057062400817871096, 0.057059326171875, 0.05714739227294922, 0.05711155319213867, 0.05710233688354492, 0.057063423156738284, 0.05705318450927734, 0.05711155319213867, 0.05704601669311524, 0.05703577423095703, 0.05702143859863281, 0.0570830078125, 0.057073535919189455, 0.057073665618896485, 0.05704294586181641, 0.057128959655761716, 0.05704294586181641, 0.057057281494140626, 0.05704499053955078, 0.05711155319213867, 0.05706649780273437, 0.05709107208251953, 0.05711667251586914, 0.05715251159667969, 0.05710438537597656, 0.05712076950073242, 0.05711155319213867, 0.05711872100830078, 0.12358975982666015, 0.05694758224487305, 0.05695078277587891, 0.056957950592041014, 0.057017345428466794, 0.05695283126831055, 0.05697228622436523, 0.057017345428466794, 0.05719859313964844, 0.05720064163208008, 0.0571412467956543, 0.05701017761230469, 0.057033729553222653, 0.05706444931030273, 0.057047039031982424, 0.05733478546142578, 0.05700812911987305, 0.0570511360168457, 0.05716275024414062, 0.056997886657714845, 0.05701017761230469, 0.056998943328857424, 0.05703164672851563, 0.057011199951171876, 0.05700713729858398, 0.05699375915527344, 0.05715456008911133, 0.05720476913452149, 0.057122783660888674, 0.05708390426635742, 0.05703987121582031, 0.05700505447387695, 0.05701631927490235, 0.05708902359008789, 0.05744844818115234, 0.057059326171875, 0.05708806228637695, 0.057064384460449216, 0.05704806518554688, 0.0570511360168457, 0.05708902359008789, 0.0570511360168457, 0.05705625534057617, 0.05706137466430664, 0.05708185577392578, 0.057106433868408205, 0.057062400817871096, 0.05704499053955078, 0.057115646362304685, 0.05706854248046875, 0.057071617126464844, 0.05741567993164062, 0.057178112030029295, 0.057296897888183596, 0.057106433868408205, 0.05709619140625, 0.057076736450195314, 0.057033729553222653, 0.05713612747192383, 0.057079807281494144, 0.057062400817871096, 0.05707059097290039, 0.05710851287841797, 0.12359164428710938, 0.05696102523803711, 0.05696307373046875, 0.056927230834960936, 0.056970241546630856, 0.05700719833374023, 0.057005985260009766, 0.05705318450927734, 0.057025535583496094, 0.057006080627441405, 0.05697740936279297, 0.056970241546630856, 0.057027584075927736, 0.05699174499511719, 0.056976383209228515, 0.056973312377929686, 0.05703782272338867, 0.05705625534057617, 0.057024513244628906, 0.0571146240234375, 0.05732966232299805, 0.057014270782470705, 0.05700198364257812, 0.05699686431884766, 0.05703577423095703, 0.057027584075927736, 0.05703184127807617, 0.05703561782836914, 0.057078784942626956, 0.057043968200683595, 0.057055233001708984, 0.05703680038452148, 0.05706649780273437, 0.05700198364257812, 0.05705011367797851, 0.057024513244628906, 0.05729792022705078, 0.05719049453735352, 0.05705104064941406, 0.05710540771484375, 0.057043968200683595, 0.057115646362304685, 0.05704806518554688, 0.05711667251586914, 0.05705420684814453, 0.057041919708251954, 0.05709209442138672, 0.05705318450927734, 0.057078784942626956, 0.05709318542480469, 0.057083873748779296, 0.05706953430175781, 0.057076736450195314, 0.05718527984619141, 0.05712076950073242, 0.057071617126464844, 0.057079807281494144, 0.05706547164916992, 0.057112575531005856, 0.05709414291381836, 0.05710438537597656, 0.05715558242797852, 0.0571514892578125, 0.12375552368164063, 0.05699071884155273, 0.057012222290039063, 0.05696307373046875, 0.05695590209960937, 0.05699071884155273, 0.057009151458740234, 0.057006080627441405, 0.05699689483642578, 0.05694166564941406, 0.05700902557373047, 0.0569620475769043, 0.057000961303710934, 0.05699385452270508, 0.05700396728515625, 0.056959999084472655, 0.057025535583496094, 0.056990753173828124, 0.056986591339111325, 0.05701939010620117, 0.057032798767089846, 0.05698652648925781, 0.05702143859863281, 0.05705318450927734, 0.05699686431884766, 0.05698252868652344, 0.05719039916992188, 0.057040897369384766, 0.05702963256835938, 0.05699993515014649, 0.057030654907226565, 0.057079807281494144, 0.05699071884155273, 0.05704294586181641, 0.05701529693603516, 0.05702143859863281, 0.05702348709106445, 0.0570030403137207, 0.057078750610351565, 0.05704499053955078, 0.05704908752441406, 0.057003009796142576, 0.05709619140625, 0.05712179183959961, 0.057060352325439455, 0.057012222290039063, 0.05711052703857422, 0.05706752014160156, 0.057079807281494144, 0.057040897369384766, 0.05708601760864258, 0.05706335830688476, 0.057082878112792966, 0.05705625534057617, 0.05712179183959961, 0.05722521591186523, 0.05745971298217773, 0.057296897888183596, 0.057099262237548826, 0.05711052703857422, 0.05738598251342773, 0.05713715362548828, 0.05708902359008789]",tokens/s,17.23584922052205,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -5755,7 +5755,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4612-6c529ab8327c4f0f237f558e;776a7514-0aef-4b79-a89e-a3094ef92247) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe935-70d00e6820da0315007320d8;ff6be5eb-eec2-4277-af08-f5c263221d17) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5812,7 +5812,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpbkle7igv/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp2te35u5k/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1235.00544,849.870848,0.0,203.423744,184.496128,s,18,0.17785792160034175,0.009880995644463434,0.00028849658241711547,0.009792351722717285,0.010203827285766602,0.010547976160049438,0.010567451219558715,"[0.01054368019104004, 0.010040608406066895, 0.00977449607849121, 0.009805983543395996, 0.010572319984436035, 0.009761152267456054, 0.009838656425476074, 0.009730496406555177, 0.009720447540283204, 0.009830143928527832, 0.009709216117858887, 0.009674847602844238, 0.0093340482711792, 0.009970623970031739, 0.010058176040649415, 0.00970019245147705, 0.01001411247253418, 0.009778719902038574]",tokens/s,25908.32029598587,kWh,1.152578154553566e-07,6.315625308503158e-08,2.135530163707643e-07,3.919670849111525e-07,tokens/kWh,653116064.7278017,MB,1235.33312,849.870848,0.0,203.423744,193.844224,s,18,10.402695068359375,0.577927503797743,0.008339520037857094,0.5779442443847655,0.5829411499023438,0.5867088104248047,0.5985698382568359,"[0.6015350952148437, 0.5610368041992188, 0.577316650390625, 0.5771664428710938, 0.5780480346679687, 0.5784367065429687, 0.58244775390625, 0.5793241577148438, 0.5705869140625, 0.5801618041992187, 0.5777095947265625, 0.5614797973632812, 0.5778404541015625, 0.5821588134765625, 0.5775224609375, 0.577691650390625, 0.5781395263671875, 0.5840924072265625]",tokens/s,109.01021250244577,kWh,6.8063746160841315e-06,3.729610197448205e-06,1.1949573998074713e-05,2.248555881160705e-05,tokens/kWh,2801798.2798577095,,s,1133,10.560176177978532,0.009320543846406458,0.0012759106372352618,0.009179136276245118,0.00931942367553711,0.009467494201660154,0.01931186187744141,"[0.010136575698852538, 0.010117119789123535, 0.010518527984619141, 0.011207679748535156, 0.010182656288146973, 0.010301440238952637, 0.01010585594177246, 0.009370623588562011, 0.009408512115478516, 0.009662464141845703, 0.010227711677551269, 0.009968640327453614, 0.009417728424072265, 0.009302016258239745, 0.009689087867736817, 0.00939417552947998, 0.009448448181152343, 0.009433088302612304, 0.009310208320617675, 0.009481216430664062, 0.00941055965423584, 0.009150464057922364, 0.009300992012023926, 0.009226240158081055, 0.009285632133483887, 0.009370623588562011, 0.009431039810180664, 0.009575424194335937, 0.009665535926818849, 0.009380864143371583, 0.0092293119430542, 0.00941875171661377, 0.009263104438781738, 0.009482239723205567, 0.009341952323913574, 0.00931123161315918, 0.009253888130187989, 0.009623552322387695, 0.00926416015625, 0.009291744232177735, 0.009316351890563965, 0.009242624282836913, 0.00931942367553711, 0.009393152236938476, 0.009688128471374511, 0.009570240020751953, 0.009524224281311035, 0.009476096153259277, 0.009392127990722657, 0.009502719879150391, 0.00921497631072998, 0.009432064056396485, 0.00919961643218994, 0.009102335929870605, 0.00894159984588623, 0.00897225570678711, 0.009049087524414063, 0.009372672080993653, 0.00982323169708252, 0.009829376220703125, 0.010189824104309082, 0.009583616256713867, 0.019946495056152345, 0.009302016258239745, 0.00921395206451416, 0.009126912117004395, 0.00923136043548584, 0.009269248008728028, 0.00922316837310791, 0.00919961643218994, 0.009209856033325196, 0.008895487785339355, 0.008706048011779785, 0.008743935585021973, 0.008759296417236329, 0.008770560264587402, 0.00881766414642334, 0.008759296417236329, 0.00869375991821289, 0.00881868839263916, 0.00879308795928955, 0.008857600212097168, 0.008679424285888672, 0.008689663887023925, 0.008700927734375, 0.008813568115234375, 0.008691712379455567, 0.008765439987182617, 0.008687616348266602, 0.008707072257995606, 0.008688639640808106, 0.008725503921508788, 0.00882688045501709, 0.008848383903503418, 0.008791040420532227, 0.00880947208404541, 0.008836095809936523, 0.00879308795928955, 0.008792063713073731, 0.008837120056152344, 0.008756223678588868, 0.00872755241394043, 0.008868864059448242, 0.008816639900207519, 0.008864768028259277, 0.008865792274475098, 0.008829952239990235, 0.00869478416442871, 0.008721407890319824, 0.008704000473022461, 0.008741888046264648, 0.008798208236694336, 0.008839167594909669, 0.008837151527404784, 0.008818655967712403, 0.009099360466003417, 0.009208736419677734, 0.009070591926574707, 0.009049087524414063, 0.009135104179382325, 0.009229472160339355, 0.00910319995880127, 0.009110527992248535, 0.009154560089111329, 0.009095168113708496, 0.019355648040771483, 0.009136128425598144, 0.009048064231872559, 0.009202688217163087, 0.009135104179382325, 0.009116671562194823, 0.009179136276245118, 0.00918015956878662, 0.009082880020141602, 0.009126912117004395, 0.009077759742736816, 0.009078783988952637, 0.00909823989868164, 0.0090316801071167, 0.00920576000213623, 0.009142271995544434, 0.009226240158081055, 0.009197567939758301, 0.009185279846191406, 0.009160703659057617, 0.009063424110412598, 0.009071616172790528, 0.009044992446899413, 0.009179136276245118, 0.009272319793701172, 0.009160799980163574, 0.009067423820495605, 0.009197567939758301, 0.009176063537597656, 0.00908083152770996, 0.009227264404296874, 0.009239551544189453, 0.009094143867492676, 0.009169919967651367, 0.009203712463378906, 0.009226240158081055, 0.009219072341918945, 0.00922214412689209, 0.009215999603271484, 0.00923136043548584, 0.009049087524414063, 0.009220095634460449, 0.009171968460083007, 0.009270272254943847, 0.009129983901977539, 0.009188575744628906, 0.009053983688354493, 0.009191424369812011, 0.009149439811706543, 0.009174015998840332, 0.00919654369354248, 0.009157631874084473, 0.009161727905273438, 0.009043968200683594, 0.009170975685119629, 0.009255904197692871, 0.009160863876342773, 0.009115488052368163, 0.009215999603271484, 0.00920576000213623, 0.009184255599975585, 0.00912179183959961, 0.009195520401000976, 0.019381248474121093, 0.009170944213867188, 0.009167872428894042, 0.009063424110412598, 0.009252863883972168, 0.009083904266357423, 0.009169919967651367, 0.009103360176086426, 0.00918835163116455, 0.00912179183959961, 0.009159680366516113, 0.009136159896850586, 0.009019359588623047, 0.00902143955230713, 0.009095168113708496, 0.009164799690246582, 0.00909823989868164, 0.009163776397705077, 0.009117695808410644, 0.009158656120300293, 0.009170944213867188, 0.009185312271118164, 0.009213919639587403, 0.009047039985656738, 0.009116671562194823, 0.009211903572082519, 0.009175040245056153, 0.009174015998840332, 0.00913920021057129, 0.00919654369354248, 0.009142271995544434, 0.009171968460083007, 0.009154560089111329, 0.009160703659057617, 0.009229375839233398, 0.009232319831848145, 0.009142271995544434, 0.009174015998840332, 0.009191424369812011, 0.009164799690246582, 0.00908902359008789, 0.00909823989868164, 0.009107456207275391, 0.009192447662353515, 0.009194496154785157, 0.009175040245056153, 0.009224191665649414, 0.009218048095703125, 0.00922316837310791, 0.00923136043548584, 0.009201663970947266, 0.009085951805114746, 0.009075712203979493, 0.009074687957763672, 0.009172991752624511, 0.009033727645874023, 0.009083904266357423, 0.009138175964355469, 0.009157631874084473, 0.00914739227294922, 0.00923750400543213, 0.009316351890563965, 0.009243647575378417, 0.02026393508911133, 0.009143296241760255, 0.009061375617980956, 0.00904207992553711, 0.009099103927612305, 0.00919961643218994, 0.009197567939758301, 0.009195520401000976, 0.009160703659057617, 0.009228287696838379, 0.009159680366516113, 0.009232383728027344, 0.009218048095703125, 0.009169919967651367, 0.009163776397705077, 0.00920473575592041, 0.009119744300842286, 0.009210880279541016, 0.00923852825164795, 0.009668607711791993, 0.009360383987426758, 0.009329664230346679, 0.009226240158081055, 0.009163871765136718, 0.009235360145568848, 0.009151488304138184, 0.009088000297546387, 0.00922316837310791, 0.00910848045349121, 0.00919654369354248, 0.009192447662353515, 0.009226240158081055, 0.009247743606567382, 0.009218048095703125, 0.009179136276245118, 0.009210880279541016, 0.00918835163116455, 0.009234432220458985, 0.009203712463378906, 0.009241600036621094, 0.009250816345214843, 0.009194496154785157, 0.009051136016845703, 0.009227264404296874, 0.009227264404296874, 0.009191424369812011, 0.009276415824890137, 0.00912281608581543, 0.009060352325439454, 0.009172991752624511, 0.009186304092407227, 0.009084032058715821, 0.0090316801071167, 0.00905510425567627, 0.009248767852783203, 0.009243647575378417, 0.009118720054626465, 0.008986623764038085, 0.00892416000366211, 0.00885043239593506, 0.008862719535827637, 0.00911359977722168, 0.009030655860900879, 0.019396608352661132, 0.009150464057922364, 0.00912281608581543, 0.009197567939758301, 0.009101311683654785, 0.009109503746032714, 0.009166848182678223, 0.00908902359008789, 0.009137151718139648, 0.009185279846191406, 0.009181183815002441, 0.009253952026367187, 0.009087008476257324, 0.00908995246887207, 0.009132032394409179, 0.009161727905273438, 0.00913920021057129, 0.009211903572082519, 0.009278464317321777, 0.009169919967651367, 0.009077759742736816, 0.009086976051330567, 0.009208831787109375, 0.009174015998840332, 0.009181183815002441, 0.009260031700134277, 0.00909823989868164, 0.009146368026733399, 0.009189375877380371, 0.009160703659057617, 0.009184255599975585, 0.009249792098999024, 0.00914739227294922, 0.00920473575592041, 0.009245696067810059, 0.009291775703430176, 0.009069567680358886, 0.009195520401000976, 0.009206784248352052, 0.009132032394409179, 0.009063424110412598, 0.009133055686950683, 0.009143296241760255, 0.009078783988952637, 0.009349311828613281, 0.009133888244628905, 0.00906668758392334, 0.009192255973815918, 0.009144319534301757, 0.009219072341918945, 0.009079839706420898, 0.009255904197692871, 0.009261055946350098, 0.009198592185974122, 0.009141247749328613, 0.009293824195861817, 0.009278464317321777, 0.009257984161376954, 0.009269248008728028, 0.009266176223754884, 0.009197567939758301, 0.009148415565490722, 0.009181183815002441, 0.019737600326538086, 0.009267200469970703, 0.009135104179382325, 0.00921497631072998, 0.009219072341918945, 0.009197567939758301, 0.009191424369812011, 0.00922316837310791, 0.009259008407592773, 0.009307135581970214, 0.009315327644348144, 0.009284607887268067, 0.00931942367553711, 0.009195520401000976, 0.009155584335327148, 0.009136128425598144, 0.00922214412689209, 0.009190400123596192, 0.009202688217163087, 0.009332736015319825, 0.009200639724731445, 0.009227264404296874, 0.009165823936462402, 0.009000960350036622, 0.008986623764038085, 0.009043968200683594, 0.009018367767333984, 0.009227392196655273, 0.009224063873291015, 0.009268223762512207, 0.00930303955078125, 0.009286720275878907, 0.009290687561035156, 0.009325568199157714, 0.009293888092041016, 0.009355199813842774, 0.009337856292724609, 0.009239551544189453, 0.009166848182678223, 0.009272319793701172, 0.009359359741210938, 0.009284607887268067, 0.009266176223754884, 0.009273344039916993, 0.009251839637756347, 0.009235456466674804, 0.009286656379699706, 0.009333760261535644, 0.009286656379699706, 0.009310208320617675, 0.009302016258239745, 0.009225215911865235, 0.009176063537597656, 0.009224191665649414, 0.009250816345214843, 0.00930406379699707, 0.009120767593383788, 0.009150464057922364, 0.009233471870422363, 0.009338815689086915, 0.009276415824890137, 0.009270272254943847, 0.009217023849487305, 0.019334144592285156, 0.009140224456787109, 0.009145343780517578, 0.009179167747497559, 0.009268192291259766, 0.009254912376403808, 0.009278464317321777, 0.009160703659057617, 0.009176063537597656, 0.009175040245056153, 0.00941875171661377, 0.009135104179382325, 0.009235456466674804, 0.009289728164672852, 0.009273344039916993, 0.009245696067810059, 0.009188447952270508, 0.009137120246887207, 0.009135040283203125, 0.00913100814819336, 0.009266176223754884, 0.009107456207275391, 0.009163904190063476, 0.009179007530212403, 0.009244671821594238, 0.00909823989868164, 0.009096287727355956, 0.009116576194763184, 0.009085951805114746, 0.009220159530639649, 0.009242560386657715, 0.009169919967651367, 0.009153535842895508, 0.00923136043548584, 0.009198592185974122, 0.009219072341918945, 0.00921292781829834, 0.009152671813964845, 0.009095135688781739, 0.009108351707458497, 0.00913100814819336, 0.009149439811706543, 0.009228287696838379, 0.00913100814819336, 0.009441280364990234, 0.009235456466674804, 0.00921292781829834, 0.00928767967224121, 0.009194496154785157, 0.009140224456787109, 0.009195520401000976, 0.009189375877380371, 0.009197567939758301, 0.00923136043548584, 0.009284607887268067, 0.00919654369354248, 0.009151488304138184, 0.009153535842895508, 0.009083935737609863, 0.009085920333862305, 0.009227264404296874, 0.009158656120300293, 0.009119744300842286, 0.01919385528564453, 0.008897536277770996, 0.008768511772155761, 0.00910848045349121, 0.009161727905273438, 0.009086976051330567, 0.009133055686950683, 0.009117695808410644, 0.009067520141601563, 0.00921497631072998, 0.009129983901977539, 0.00918015956878662, 0.009044992446899413, 0.008796159744262694, 0.008827903747558593, 0.008866815567016602, 0.008691712379455567, 0.008760319709777833, 0.008839167594909669, 0.00882380771636963, 0.0088340482711792, 0.008731648445129395, 0.008802304267883301, 0.008795136451721192, 0.008844287872314453, 0.00890675163269043, 0.00881772804260254, 0.008837056159973145, 0.00872447967529297, 0.008742912292480469, 0.008830975532531739, 0.008797344207763671, 0.008948575973510743, 0.00951296043395996, 0.009158656120300293, 0.009202688217163087, 0.009095168113708496, 0.009115648269653321, 0.009158656120300293, 0.009143296241760255, 0.009332736015319825, 0.009127936363220214, 0.009206784248352052, 0.009300992012023926, 0.009198592185974122, 0.009124863624572753, 0.009181183815002441, 0.009224191665649414, 0.009182208061218262, 0.009182208061218262, 0.009177087783813476, 0.009157631874084473, 0.009219072341918945, 0.009235456466674804, 0.00921292781829834, 0.009206784248352052, 0.00922111988067627, 0.009272319793701172, 0.009179136276245118, 0.009088000297546387, 0.009075712203979493, 0.009242624282836913, 0.009273344039916993, 0.019775487899780272, 0.009277440071105958, 0.00930406379699707, 0.009160896301269532, 0.009081664085388184, 0.009125887870788574, 0.009217087745666503, 0.009294783592224121, 0.009259008407592773, 0.009174015998840332, 0.009237536430358887, 0.009176032066345215, 0.009359359741210938, 0.009254912376403808, 0.009415679931640625, 0.009275391578674316, 0.00912281608581543, 0.00922214412689209, 0.00928876781463623, 0.009192383766174317, 0.009159680366516113, 0.009207807540893554, 0.009309184074401856, 0.009120767593383788, 0.009166848182678223, 0.008999936103820801, 0.009118720054626465, 0.00923033618927002, 0.00922323226928711, 0.009146304130554199, 0.009227264404296874, 0.00930303955078125, 0.009255935668945312, 0.009110527992248535, 0.009076736450195312, 0.00912281608581543, 0.009102335929870605, 0.00913920021057129, 0.009243647575378417, 0.009208831787109375, 0.00909721565246582, 0.009232383728027344, 0.009280511856079102, 0.009257984161376954, 0.009200639724731445, 0.009266176223754884, 0.009227264404296874, 0.009320480346679687, 0.009240544319152832, 0.009115648269653321, 0.009124896049499513, 0.009084896087646484, 0.009176063537597656, 0.009248767852783203, 0.009246720314025878, 0.009260031700134277, 0.009182208061218262, 0.009068544387817384, 0.009130047798156739, 0.00921183967590332, 0.009153535842895508, 0.009088000297546387, 0.009128959655761718, 0.019221536636352538, 0.00914735984802246, 0.00881766414642334, 0.008739839553833008, 0.009056256294250489, 0.009115648269653321, 0.009055232048034668, 0.009151488304138184, 0.009211903572082519, 0.009256959915161133, 0.009225215911865235, 0.009248767852783203, 0.009208831787109375, 0.00923033618927002, 0.009128959655761718, 0.009155584335327148, 0.0091146240234375, 0.009055264472961426, 0.009093088150024414, 0.009335807800292969, 0.009111552238464356, 0.009262080192565919, 0.00919961643218994, 0.009070591926574707, 0.009192447662353515, 0.009271295547485351, 0.009233407974243164, 0.009169919967651367, 0.009148415565490722, 0.009164799690246582, 0.009200639724731445, 0.009159680366516113, 0.009225215911865235, 0.009076736450195312, 0.009156607627868652, 0.00912384033203125, 0.009118720054626465, 0.009143296241760255, 0.009150464057922364, 0.009190431594848633, 0.009420767784118652, 0.009259008407592773, 0.009239551544189453, 0.009190400123596192, 0.009144415855407715, 0.009088992118835449, 0.009233344078063964, 0.00922316837310791, 0.00910028839111328, 0.009146431922912597, 0.00923846435546875, 0.00919654369354248, 0.009192447662353515, 0.009241600036621094, 0.00913920021057129, 0.009236479759216308, 0.009218048095703125, 0.009034912109375, 0.009146207809448243, 0.00923852825164795, 0.00923033618927002, 0.009185279846191406, 0.00914739227294922, 0.019207231521606444, 0.009263039588928223, 0.009186304092407227, 0.009096320152282716, 0.009118592262268067, 0.009195520401000976, 0.009234432220458985, 0.009151488304138184, 0.009243647575378417, 0.00921395206451416, 0.009193471908569336, 0.009243647575378417, 0.00923750400543213, 0.009211903572082519, 0.009115648269653321, 0.009201663970947266, 0.009084927558898925, 0.009203712463378906, 0.009178112030029297, 0.009164799690246582, 0.009107456207275391, 0.00879923152923584, 0.0087326717376709, 0.008692735671997071, 0.008788991928100585, 0.008741888046264648, 0.008717311859130859, 0.008870911598205567, 0.008846336364746094, 0.008821760177612305, 0.008831999778747558, 0.008837120056152344, 0.008755200386047364, 0.008786944389343262, 0.008811552047729493, 0.008701919555664062, 0.008704000473022461, 0.008717311859130859, 0.008651776313781738, 0.008627200126647949, 0.008673279762268067, 0.008647744178771972, 0.008747967720031738, 0.008684543609619141, 0.008839167594909669, 0.008827903747558593, 0.008866815567016602, 0.008782848358154297, 0.00884227180480957, 0.008824799537658691, 0.008820735931396484, 0.008838144302368164, 0.008849408149719238, 0.00881766414642334, 0.008777728080749512, 0.00868556785583496, 0.008679424285888672, 0.008851455688476563, 0.008749055862426757, 0.008771583557128907, 0.008869888305664063, 0.008725503921508788, 0.008758272171020508, 0.018522111892700196, 0.008760479927062989, 0.008744799613952636, 0.008678400039672851, 0.008663040161132812, 0.008689663887023925, 0.008782848358154297, 0.008729599952697753, 0.009275391578674316, 0.010090496063232422, 0.009278464317321777, 0.009252863883972168, 0.009244671821594238, 0.00923852825164795, 0.009389056205749511, 0.009120767593383788, 0.009095168113708496, 0.00923852825164795, 0.009381888389587402, 0.009172991752624511, 0.009146368026733399, 0.009207807540893554, 0.009226240158081055, 0.009163776397705077, 0.009165823936462402, 0.009182208061218262, 0.009246720314025878, 0.009202688217163087, 0.009207807540893554, 0.009197567939758301, 0.009206784248352052, 0.009366527557373047, 0.009355263710021973, 0.00913100814819336, 0.00922214412689209, 0.009150464057922364, 0.009242624282836913, 0.009143296241760255, 0.009187328338623046, 0.00920473575592041, 0.009209856033325196, 0.009149439811706543, 0.009115648269653321, 0.009124863624572753, 0.009211903572082519, 0.009187328338623046, 0.009228287696838379, 0.009172991752624511, 0.009270272254943847, 0.009181183815002441, 0.009215999603271484, 0.009220095634460449, 0.009201663970947266, 0.009146400451660156, 0.009382880210876465, 0.009256959915161133, 0.009209856033325196, 0.009274368286132812, 0.009193471908569336, 0.009179136276245118, 0.009100319862365723, 0.009211872100830078, 0.009215999603271484, 0.019696672439575194, 0.009102304458618165, 0.009134079933166504, 0.00912384033203125, 0.009239551544189453, 0.009202688217163087, 0.009096192359924317, 0.00909004783630371, 0.00912281608581543, 0.009125887870788574, 0.009203712463378906, 0.009104384422302245, 0.00882585620880127, 0.008884223937988281, 0.008749055862426757, 0.008752127647399903, 0.008795136451721192, 0.008794112205505371, 0.008836095809936523, 0.008820735931396484, 0.008721407890319824, 0.009194496154785157, 0.009233407974243164, 0.009241600036621094, 0.009246720314025878, 0.009147583961486816, 0.009139007568359375, 0.009153535842895508, 0.009278464317321777, 0.009261055946350098, 0.009248767852783203, 0.009250816345214843, 0.009294848442077636, 0.009308159828186035, 0.009265151977539063, 0.009347071647644043, 0.009260031700134277, 0.009329664230346679, 0.009267200469970703, 0.0092293119430542, 0.009225215911865235, 0.009184255599975585, 0.0092293119430542, 0.00918835163116455, 0.009201663970947266, 0.009164799690246582, 0.009148415565490722, 0.009138208389282226, 0.009210847854614257, 0.009318400382995605, 0.009245696067810059, 0.009461759567260742, 0.010915840148925781, 0.010038271903991699, 0.010377216339111327, 0.009389056205749511, 0.009617568016052246, 0.009375583648681641, 0.009366527557373047, 0.009326592445373535, 0.009265151977539063, 0.009374719619750976, 0.009284607887268067, 0.01962700843811035, 0.009049087524414063, 0.009162752151489258, 0.009194496154785157, 0.00922316837310791, 0.009119744300842286, 0.00921497631072998, 0.009175040245056153, 0.009163776397705077, 0.00919961643218994, 0.009193471908569336, 0.009232383728027344, 0.009225215911865235, 0.009160703659057617, 0.009127936363220214, 0.00922214412689209, 0.009052255630493163, 0.009175968170166016, 0.009187328338623046, 0.009208831787109375, 0.009189408302307129, 0.009068511962890624, 0.009040896415710448, 0.009053183555603026, 0.009175040245056153, 0.009203712463378906, 0.009224191665649414, 0.009067520141601563, 0.009135104179382325, 0.009217023849487305, 0.009209856033325196, 0.009096192359924317, 0.009144319534301757, 0.009155584335327148, 0.009078783988952637, 0.009629695892333985, 0.009241600036621094, 0.009183232307434081, 0.009161727905273438, 0.009234432220458985, 0.009186304092407227, 0.009131168365478516, 0.009117535591125488, 0.009124863624572753, 0.00909721565246582, 0.009162816047668457, 0.009056192398071289, 0.009151488304138184, 0.009083904266357423, 0.009171968460083007, 0.00918835163116455, 0.00910028839111328, 0.009140224456787109, 0.009065471649169921, 0.00922316837310791, 0.009200672149658203, 0.0090982084274292, 0.00923033618927002, 0.009254912376403808, 0.009189375877380371, 0.009062399864196777, 0.009104384422302245, 0.009064448356628419, 0.019264511108398438, 0.009159744262695313, 0.00922822380065918, 0.009208831787109375, 0.00918835163116455, 0.009182208061218262, 0.00918835163116455, 0.009112575531005859, 0.009053183555603026, 0.009227295875549317, 0.009236448287963868, 0.009200639724731445, 0.009247743606567382, 0.009157631874084473, 0.009063424110412598, 0.009496576309204101, 0.009324543952941895, 0.00922316837310791, 0.009200639724731445, 0.009242624282836913, 0.00919961643218994, 0.009125887870788574, 0.00922214412689209, 0.009150464057922364, 0.009102335929870605, 0.0090316801071167, 0.009200639724731445, 0.009150464057922364, 0.009085951805114746, 0.009141247749328613, 0.009163776397705077, 0.009049087524414063, 0.009164799690246582, 0.009156607627868652, 0.00902143955230713, 0.009132032394409179, 0.009043968200683594, 0.009084927558898925, 0.009071616172790528, 0.009217023849487305, 0.009185279846191406, 0.009183232307434081, 0.009183232307434081, 0.009169919967651367, 0.00922214412689209, 0.009191424369812011, 0.009251839637756347, 0.00914739227294922, 0.009112640380859375, 0.009141183853149415, 0.009069567680358886, 0.009070591926574707, 0.009040063858032226, 0.00905401611328125, 0.009086976051330567, 0.009063424110412598, 0.009263104438781738, 0.009186304092407227, 0.009119903564453125, 0.009044832229614257, 0.009127936363220214, 0.009224191665649414, 0.009486495971679687, 0.01975075149536133, 0.009183232307434081, 0.0092293119430542, 0.009153535842895508, 0.009065471649169921, 0.009033760070800782, 0.009042911529541016, 0.009149439811706543, 0.009255935668945312, 0.009170944213867188, 0.009193471908569336, 0.00920576000213623, 0.009220095634460449, 0.009232383728027344, 0.009138175964355469, 0.009169919967651367, 0.009307135581970214, 0.009273344039916993, 0.009115648269653321, 0.009132032394409179, 0.009265151977539063, 0.009175040245056153, 0.009054207801818847, 0.009225215911865235, 0.009192447662353515, 0.009267200469970703, 0.009179136276245118, 0.009255071640014649, 0.00908784008026123, 0.009150464057922364, 0.009152640342712402, 0.009082752227783203, 0.009157631874084473, 0.00933683204650879, 0.009025535583496093, 0.008992768287658692, 0.008881343841552734, 0.008942399978637695, 0.008988672256469727, 0.00909004783630371, 0.009201663970947266, 0.009261055946350098, 0.009135104179382325, 0.009195648193359374, 0.009181056022644043, 0.00919961643218994, 0.009078783988952637, 0.00912281608581543, 0.00922214412689209, 0.009189375877380371, 0.009096192359924317, 0.009136128425598144, 0.009146368026733399, 0.009236479759216308, 0.009179136276245118, 0.009235456466674804, 0.009173088073730469, 0.00908073616027832, 0.009496576309204101, 0.009350144386291503, 0.009266176223754884, 0.00928767967224121, 0.009235456466674804, 0.01946224021911621, 0.008913824081420899, 0.009961471557617188, 0.009366527557373047, 0.009101311683654785, 0.00914739227294922, 0.009067520141601563, 0.009072768211364745, 0.012043135643005371, 0.009969663619995118, 0.009380864143371583, 0.00935321617126465, 0.00932249641418457, 0.00920473575592041, 0.01004851245880127, 0.009363455772399902, 0.009171968460083007, 0.00919961643218994, 0.009158656120300293, 0.009293824195861817, 0.009206784248352052, 0.009227264404296874, 0.009201727867126464, 0.00920569610595703, 0.009234432220458985, 0.009376768112182618, 0.009207807540893554, 0.009278464317321777, 0.009200639724731445, 0.009198592185974122, 0.009110655784606934, 0.009169792175292969, 0.009219072341918945, 0.009275391578674316, 0.009152511596679687, 0.009171968460083007, 0.009192447662353515, 0.009001983642578124, 0.009166848182678223, 0.009206784248352052, 0.009124863624572753, 0.00912179183959961, 0.009092096328735352, 0.009110527992248535, 0.009146368026733399, 0.009148415565490722, 0.00910028839111328, 0.009198592185974122, 0.009185279846191406, 0.009109503746032714, 0.009128959655761718, 0.009093119621276855, 0.00921497631072998, 0.009187328338623046, 0.00921292781829834, 0.009129983901977539, 0.009110527992248535, 0.009033727645874023, 0.009046015739440917, 0.009069567680358886, 0.009292799949645996, 0.009153535842895508, 0.009305120468139649]",tokens/s,107.28987669378871,, @@ -5967,7 +5967,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e9109-2080809168b17ac15c8f64bf;415dcc35-daaa-470d-8d27-b3b819a90127) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe2ba-4d910175168f97f217d94ac8;adcbdf56-3429-4047-870c-c746e5647ddc) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -6056,7 +6056,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95f3-2f4d4e275cc1eee75285153f;bc345723-dc59-457f-b8e4-8ec12a5dc059) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe8e5-09b91a36228231da118ddf67;6ea180c1-bed1-45f4-b6e9-139d21e4c2cd) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6116,7 +6116,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90df-3258b65940d59eca0bad301f;451b501e-7be9-4011-8eea-f4c1550615c4) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe28c-44ce44d9161768241345fea7;23a4a98e-04a3-420a-b10d-083ee8aec7ff) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -6348,7 +6348,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp4reqj10r/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpa458x0s4/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -6445,7 +6445,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e8f50-16e5cda37f8eb65052a7121e;6e8e1370-2e52-4448-b92b-f426364af56e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe0f1-5be03c5172083b27366087e4;6893c147-732d-44ea-ac35-dda617b2ff0e) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -6530,7 +6530,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9697-221c696916e11a014aaef2e9;6a5b7981-4663-4f24-8c3f-241b4d26c9e2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe866-24b97d5a5d742d4c32ccbe50;727a3c8f-2bab-491a-94db-05c99ab9faba) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6646,7 +6646,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95cb-6fa7931a69eddedd30d830ec;aa86f0f3-0638-4c9e-aca6-ebb149dc96b4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe799-339ace757c7c577339e7a3a4;2db37b4a-c00b-4032-8807-029413af323b) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6720,7 +6720,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1844.498432,3393.716224,0.0,2749.366272,2585.357312,s,10,0.817713020324707,0.08177130203247071,0.0017450673609261166,0.08117209243774415,0.08364522857666015,0.0849723907470703,0.08603412048339844,"[0.08629955291748047, 0.08335030364990234, 0.08125138854980468, 0.0808326416015625, 0.0810927963256836, 0.08075574493408202, 0.08150611114501953, 0.0821209945678711, 0.08023673248291016, 0.08026675415039063]",tokens/s,3130.6826923013223,kWh,9.541744039156959e-07,5.228433619278451e-07,4.351788875694268e-06,5.828806641537809e-06,tokens/kWh,43919796.236791916,MB,1844.498432,3395.813376,0.0,2749.366272,2628.768768,s,10,45.24670263671875,4.524670263671875,0.011893284876975272,4.52194580078125,4.53378251953125,4.544506494140625,4.553085673828124,"[4.55523046875, 4.5133427734375, 4.5271787109375, 4.52349462890625, 4.52039697265625, 4.527849609375, 4.5313994140625, 4.51648095703125, 4.51110009765625, 4.52022900390625]",tokens/s,13.923666549985024,kWh,5.321501913543884e-05,2.9164818669173503e-05,0.00023333654016370403,0.0003157163779683163,tokens/kWh,199546.18890985235,,s,629,45.90427850341802,0.0729797750451796,0.009729252653977099,0.07159503936767578,0.07251025695800782,0.07281602630615235,0.15288290954589845,"[0.07412633514404297, 0.07440076446533203, 0.0745185317993164, 0.07484521484375, 0.07359587097167969, 0.07270297241210938, 0.07295487976074219, 0.0725032958984375, 0.0718704605102539, 0.0717998046875, 0.07187251281738281, 0.07140557098388672, 0.07241011047363281, 0.0728279037475586, 0.07255449676513671, 0.07251353454589844, 0.07184178924560547, 0.07262105560302734, 0.07237939453125, 0.072342529296875, 0.07164313507080078, 0.07216537475585938, 0.07139225769042969, 0.07148646545410156, 0.07167488098144531, 0.07242137908935548, 0.07154073333740234, 0.07204863739013671, 0.07214796447753906, 0.07182643127441406, 0.07173836517333984, 0.07221862030029297, 0.0719452133178711, 0.07195136260986328, 0.07149260711669922, 0.07235382080078125, 0.07215001678466797, 0.07205680084228516, 0.07161958312988281, 0.07250943756103516, 0.07279821014404297, 0.07184793853759766, 0.07290367889404296, 0.0722012176513672, 0.07165235137939453, 0.072595458984375, 0.07175475311279297, 0.07177327728271485, 0.07174237060546874, 0.07140249633789063, 0.07150489807128907, 0.07124172973632813, 0.07160115051269532, 0.07147634887695313, 0.07142079925537109, 0.07171481323242188, 0.07252889251708984, 0.07654195404052734, 0.07267225646972657, 0.07200358581542969, 0.07152947235107422, 0.07169741058349609, 0.15563571166992188, 0.07144140625, 0.07148544311523437, 0.0714229736328125, 0.07137792205810548, 0.07154073333740234, 0.07146803283691407, 0.07168921661376954, 0.07148748779296875, 0.07125196838378907, 0.072015869140625, 0.0714434585571289, 0.07144652557373046, 0.07144755554199218, 0.07143526458740235, 0.07141580963134765, 0.07135846710205078, 0.07186329650878906, 0.07159503936767578, 0.07138301086425781, 0.07177523040771484, 0.07191756439208985, 0.07138511657714844, 0.07143011474609375, 0.0714691162109375, 0.07133177947998047, 0.07206400299072266, 0.07146598052978516, 0.07148339080810547, 0.07149465942382813, 0.07135244750976563, 0.07228915405273438, 0.07199129486083984, 0.07153971099853515, 0.07214195251464844, 0.07162457275390625, 0.07148236846923828, 0.07145574188232422, 0.07147622680664062, 0.07152435302734375, 0.07147622680664062, 0.07173017883300781, 0.07242240142822266, 0.07143628692626953, 0.07144550323486328, 0.07149062347412109, 0.07370745849609375, 0.071552001953125, 0.07143014526367188, 0.07157561492919921, 0.07137686157226562, 0.07150895690917969, 0.0715448989868164, 0.071289794921875, 0.07157247924804687, 0.07189810943603515, 0.07256575775146484, 0.0720179214477539, 0.07132978820800781, 0.07143628692626953, 0.07151615905761718, 0.071478271484375, 0.07195954895019531, 0.15412019348144532, 0.07261593627929687, 0.07204249572753907, 0.07153766632080077, 0.07150694274902344, 0.07190835571289063, 0.07203942108154297, 0.07188070678710938, 0.0718387222290039, 0.07253708648681641, 0.07149874877929688, 0.07151312255859375, 0.07189910125732422, 0.07219609832763672, 0.07241522979736328, 0.07219609832763672, 0.07200665283203125, 0.07194111633300782, 0.07144857788085937, 0.07177216339111328, 0.07156428527832032, 0.07152435302734375, 0.071404541015625, 0.07229337310791016, 0.0715982437133789, 0.07232189178466797, 0.07304499053955078, 0.07218585968017578, 0.07157247924804687, 0.07147212982177735, 0.07146905517578125, 0.0714250259399414, 0.07145369720458984, 0.0720711669921875, 0.07135027313232421, 0.07165337371826172, 0.07138304138183593, 0.07151001739501953, 0.07149158477783203, 0.0714434585571289, 0.07209881591796875, 0.07256575775146484, 0.07190835571289063, 0.07150592041015626, 0.07148441314697265, 0.07140665435791016, 0.07138195037841796, 0.07330815887451173, 0.07221965026855469, 0.07159101104736328, 0.07144233703613281, 0.0722165756225586, 0.07188787078857421, 0.07187558746337891, 0.07241216278076172, 0.07215513610839844, 0.0714250259399414, 0.07158169555664062, 0.07205580902099609, 0.07226470184326172, 0.07152249908447265, 0.07164704132080078, 0.07141785430908203, 0.15294053649902345, 0.07213772583007813, 0.07148236846923828, 0.07144550323486328, 0.07141990661621093, 0.07155097961425781, 0.07134003448486329, 0.0716882553100586, 0.07207417297363282, 0.07252582550048828, 0.07232307434082032, 0.07238553619384766, 0.07231795501708985, 0.07151103973388671, 0.07148134613037109, 0.07183052825927734, 0.07228518676757813, 0.07243673706054687, 0.07223193359375, 0.07178342437744141, 0.07154994964599609, 0.07213056182861328, 0.07134528350830079, 0.07198809814453125, 0.07157759857177734, 0.07243981170654297, 0.07139225769042969, 0.07143321228027344, 0.0725002212524414, 0.07220428466796874, 0.07141069030761718, 0.07173426818847656, 0.0720742416381836, 0.07153049468994141, 0.07139839935302734, 0.07158169555664062, 0.07130623626708985, 0.07154605102539062, 0.07140230560302735, 0.07144652557373046, 0.07150796508789062, 0.07149465942382813, 0.07262413024902344, 0.07216844940185548, 0.073059326171875, 0.07163391876220702, 0.07140966033935547, 0.07162879943847657, 0.07215309143066406, 0.07194834899902344, 0.07152428436279297, 0.07163085174560548, 0.07161036682128906, 0.07260262298583985, 0.07228108978271484, 0.07144550323486328, 0.0714434585571289, 0.07155712127685547, 0.07133695983886719, 0.07153254699707032, 0.0716072998046875, 0.07149568176269532, 0.07149158477783203, 0.15309414672851562, 0.07136255645751953, 0.07175580596923828, 0.07164310455322266, 0.07139532470703125, 0.07184896087646485, 0.07149158477783203, 0.07176089477539062, 0.0714567642211914, 0.07160524749755859, 0.07143116760253906, 0.07155097961425781, 0.07128268432617188, 0.07143628692626953, 0.0721080322265625, 0.07152742767333985, 0.07195340728759765, 0.0717998046875, 0.07170150756835937, 0.07143628692626953, 0.07166976165771484, 0.07139328002929687, 0.07143334197998047, 0.07141567993164062, 0.07132784271240235, 0.07156521606445312, 0.07134419250488282, 0.07141779327392578, 0.07136870574951172, 0.07146701049804688, 0.07180595397949219, 0.07261695861816406, 0.07249817657470703, 0.07219513702392578, 0.07214688110351562, 0.0726087646484375, 0.07276953887939454, 0.07257190704345703, 0.07155097961425781, 0.07176396942138671, 0.07192063903808593, 0.07148041534423828, 0.07149251556396484, 0.07153049468994141, 0.0720097885131836, 0.07163180541992188, 0.07188992309570312, 0.07222169494628906, 0.07132569885253906, 0.07175081634521484, 0.07155388641357421, 0.07143321228027344, 0.07141580963134765, 0.07153868865966796, 0.07144963073730469, 0.07194624328613282, 0.07234966278076171, 0.0735129623413086, 0.07201398468017578, 0.07184162902832031, 0.07150899505615234, 0.07152639770507813, 0.07161958312988281, 0.15273472595214843, 0.0721817626953125, 0.07285453033447266, 0.0722841567993164, 0.07220633697509765, 0.07234150695800781, 0.0714803237915039, 0.07147212982177735, 0.07221247863769531, 0.07245523071289063, 0.0723117446899414, 0.07214387512207031, 0.07150694274902344, 0.07175373077392579, 0.07198617553710937, 0.07157350158691406, 0.0715489273071289, 0.07148851013183594, 0.07142707061767578, 0.07142092895507812, 0.0714629135131836, 0.07176806640625, 0.07141785430908203, 0.07138406372070312, 0.07139647674560547, 0.07157030487060546, 0.07146803283691407, 0.07146495819091797, 0.07141375732421874, 0.07154790496826172, 0.07151107025146484, 0.07144342041015625, 0.07185100555419922, 0.07181517028808594, 0.07148646545410156, 0.0713328628540039, 0.07203743743896485, 0.07147821044921875, 0.07322930908203125, 0.07284429168701172, 0.07214694213867187, 0.07164927673339844, 0.07231283569335938, 0.07260262298583985, 0.07171686553955078, 0.07191961669921874, 0.07252889251708984, 0.07248281860351563, 0.07143628692626953, 0.07152742767333985, 0.07151001739501953, 0.0715335693359375, 0.07144652557373046, 0.0727040023803711, 0.07292620849609376, 0.07252684783935547, 0.07237529754638672, 0.07170252990722656, 0.07182028961181641, 0.0714229736328125, 0.07204557037353515, 0.07154176330566406, 0.07145881652832031, 0.15435366821289062, 0.07202508544921875, 0.07218994903564453, 0.07186739349365234, 0.07218073272705078, 0.07134719848632813, 0.07238559722900391, 0.07221548461914062, 0.07241216278076172, 0.07184178924560547, 0.07194214630126954, 0.07146803283691407, 0.07222271728515625, 0.07166976165771484, 0.07240601348876953, 0.0716042251586914, 0.07220451354980469, 0.07180572509765625, 0.0720343017578125, 0.07153049468994141, 0.07171891021728516, 0.07154073333740234, 0.07239577484130859, 0.07312076568603515, 0.07245823669433593, 0.07169741058349609, 0.07252790069580078, 0.07206703948974609, 0.07212345886230469, 0.07148537445068359, 0.07223910522460937, 0.0724295654296875, 0.0715489273071289, 0.07143526458740235, 0.07208550262451172, 0.0715868148803711, 0.07253298950195312, 0.07192985534667969, 0.0719288330078125, 0.0714424285888672, 0.07221759796142578, 0.07254220581054688, 0.07175885009765624, 0.07147622680664062, 0.07144960021972656, 0.07139942169189453, 0.07147622680664062, 0.07175270080566407, 0.07210086059570313, 0.07172198486328125, 0.07231078338623047, 0.07225138854980469, 0.07237836456298828, 0.07164521789550782, 0.07197487640380859, 0.07171071624755859, 0.07147110748291016, 0.07145785522460937, 0.07145772552490234, 0.07148953247070312, 0.07153049468994141, 0.07147929382324218, 0.07183257293701172, 0.15419290161132812, 0.0719319076538086, 0.07157453155517578, 0.07133900451660156, 0.07158086395263671, 0.07154566192626953, 0.07155609893798828, 0.07177932739257813, 0.07223910522460937, 0.07164825439453125, 0.07151103973388671, 0.07149874877929688, 0.07160832214355468, 0.07286483001708985, 0.07185404968261719, 0.07166358184814453, 0.07239065551757813, 0.07162777709960938, 0.07142400360107422, 0.07145779418945312, 0.07137689971923829, 0.07147929382324218, 0.07150899505615234, 0.07147929382324218, 0.07143014526367188, 0.07144550323486328, 0.07150489807128907, 0.07147929382324218, 0.07139942169189453, 0.07138111877441407, 0.07176595306396484, 0.07276025390625, 0.07230156707763671, 0.07154994964599609, 0.07149874877929688, 0.071984130859375, 0.07214704132080078, 0.07197277069091797, 0.07147622680664062, 0.07143628692626953, 0.07141990661621093, 0.07155097961425781, 0.07132876586914062, 0.07161753845214844, 0.07156735992431641, 0.07142707061767578, 0.07145574188232422, 0.07148236846923828, 0.07152947235107422, 0.07150387573242188, 0.07297023773193359, 0.07153459167480469, 0.07137391662597656, 0.07174543762207031, 0.07197593688964844, 0.07133900451660156, 0.07137586975097657, 0.07162675476074219, 0.07164524841308594, 0.07149254608154297, 0.07288114929199219, 0.07158067321777344, 0.07160832214355468, 0.1534167022705078, 0.07255654144287109, 0.07152025604248047, 0.07147110748291016, 0.07152127838134766, 0.07155718231201172, 0.07147615814208984, 0.07213878631591797, 0.0719688949584961, 0.07150473785400391, 0.07153254699707032, 0.0713677749633789, 0.07146691131591797, 0.07148236846923828, 0.07166361236572266, 0.07142195129394531, 0.07141273498535156, 0.07221247863769531, 0.07152947235107422, 0.07151411437988281, 0.07147727966308594, 0.07141168212890625, 0.07130111694335937, 0.07153561401367188, 0.07139328002929687, 0.07157868957519531, 0.07202604675292969, 0.07173222351074218, 0.07142809295654297, 0.07216230773925782, 0.07140249633789063, 0.07125708770751953, 0.07135030364990234, 0.07157039642333984, 0.07152537536621094, 0.07153663635253907, 0.07136067199707032, 0.07143817901611328, 0.07143218994140625, 0.07128883361816406, 0.07141171264648437, 0.07147212982177735, 0.07205785369873047, 0.0722841567993164, 0.07163801574707031, 0.07146086120605469, 0.07142400360107422, 0.07148851013183594, 0.0714567642211914, 0.07157453155517578, 0.07137894439697266, 0.07154994964599609, 0.07136358642578125, 0.07146806335449218, 0.07140144348144531, 0.07132978820800781, 0.07141990661621093, 0.0713338851928711, 0.07141683197021484, 0.07153254699707032, 0.072774658203125, 0.07168000030517578, 0.07149465942382813, 0.15201997375488283, 0.07148339080810547, 0.07143628692626953, 0.07144038391113282, 0.07139942169189453, 0.07197293090820313, 0.07162464141845704, 0.07138406372070312, 0.07135743713378906, 0.07148544311523437, 0.07146803283691407, 0.07145881652832031, 0.07143218994140625, 0.07152742767333985, 0.07162982177734376, 0.07121612548828125, 0.07146495819091797, 0.0720404510498047, 0.0724695053100586, 0.07181721496582032, 0.07156018829345703, 0.07254528045654297, 0.07181414031982422, 0.07141072082519531, 0.0720823974609375, 0.0720711669921875, 0.07299686431884765, 0.07165644836425782, 0.07156121826171875, 0.071299072265625, 0.07145369720458984, 0.07160329437255859, 0.07154678344726563, 0.07152537536621094, 0.07154688262939453, 0.0715868148803711, 0.0715704345703125, 0.0713881607055664, 0.07157350158691406, 0.0714260482788086, 0.07144038391113282, 0.07158988952636719, 0.07161138916015625, 0.07159295654296875, 0.07171071624755859, 0.07247052764892578, 0.07238553619384766, 0.07274291229248046, 0.07178034973144531, 0.0724695053100586, 0.07223808288574218, 0.07157257843017578, 0.07217967987060547, 0.07244691467285157, 0.07151718139648437, 0.07182745361328124, 0.07148748779296875, 0.07153971099853515, 0.07172608184814454, 0.07232614135742188, 0.07174861145019532, 0.07182540893554687, 0.07194419097900391]",tokens/s,13.702426451450826,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6746,7 +6746,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp4xqzchjj/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpiwd48qth/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -6779,7 +6779,7 @@ ImportError: This modeling file requires the following packages that were not fo ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1530.867712,1596.45696,0.0,950.009856,857.516544,s,10,0.5920607986450195,0.05920607986450195,0.0026561766441009003,0.0581931037902832,0.06018668594360351,0.06362006111145019,0.06636676124572753,"[0.06705343627929687, 0.05793779373168945, 0.05804457473754883, 0.0580964469909668, 0.058221630096435543, 0.05848179244995117, 0.05885744094848633, 0.05942371368408203, 0.05777939224243164, 0.05816457748413086]",tokens/s,4323.880260032033,kWh,6.794170991518794e-07,3.722834273722928e-07,2.8646399034599733e-06,3.916340429984146e-06,tokens/kWh,65367146.849651255,MB,1530.867712,1596.45696,0.0,950.009856,888.373248,s,10,31.443047607421878,3.144304760742187,0.028395623245064967,3.1506854248046876,3.17336376953125,3.1794501953125,3.1843193359375,"[3.18553662109375, 3.156751708984375, 3.09842138671875, 3.14605859375, 3.1401103515625, 3.124360107421875, 3.17201123046875, 3.16763037109375, 3.09685498046875, 3.155312255859375]",tokens/s,20.036225745855937,kWh,3.7348117898973627e-05,2.046855207133909e-05,0.0001546499027747399,0.0002124665727450526,tokens/kWh,296517.2317981346,,s,629,31.91495677947996,0.050739199967376754,0.006915167421256304,0.050097152709960936,0.05089812393188477,0.05135626220703125,0.10733584594726563,"[0.053369857788085937, 0.05259980773925781, 0.054150142669677735, 0.052741119384765625, 0.05313024139404297, 0.05274009704589844, 0.051294208526611325, 0.05223526382446289, 0.05079244613647461, 0.05039411163330078, 0.05090918350219727, 0.05163212966918945, 0.05048524856567383, 0.05044838333129883, 0.05030297470092773, 0.05091430282592774, 0.05080780792236328, 0.05040332794189453, 0.050255870819091795, 0.050356224060058595, 0.05037363052368164, 0.05037875366210937, 0.050547710418701174, 0.04912332916259766, 0.04912025451660156, 0.04912332916259766, 0.0490885124206543, 0.04913971328735352, 0.04907724761962891, 0.04911718368530273, 0.04908748626708984, 0.049127422332763675, 0.04910182571411133, 0.0491253776550293, 0.04909568023681641, 0.050716670989990234, 0.05085184097290039, 0.050508800506591796, 0.0506798095703125, 0.050367488861083984, 0.05057228851318359, 0.050566143035888675, 0.05034086227416992, 0.050678783416748044, 0.049223678588867184, 0.05280051040649414, 0.05099929428100586, 0.050407424926757816, 0.05042892837524414, 0.05095219039916992, 0.04916223907470703, 0.04919705581665039, 0.04914790344238281, 0.049111038208007815, 0.04911718368530273, 0.04910182571411133, 0.05056409454345703, 0.05402521514892578, 0.05089177703857422, 0.05039513778686523, 0.05034393692016602, 0.05038489532470703, 0.10725990295410157, 0.050229248046875, 0.050310142517089845, 0.05069926452636719, 0.05130342483520508, 0.050307071685791016, 0.0505354232788086, 0.05059481430053711, 0.05064396667480469, 0.05037158584594727, 0.05051801681518555, 0.04907724761962891, 0.04911308670043945, 0.050462718963623046, 0.05050777435302734, 0.05088358306884765, 0.050391040802001956, 0.05048320007324219, 0.05041459274291992, 0.05044736099243164, 0.05044736099243164, 0.050460670471191404, 0.050427902221679685, 0.05039308929443359, 0.050344959259033206, 0.05039513778686523, 0.049148929595947265, 0.0491253776550293, 0.04916223907470703, 0.04913459014892578, 0.04912639999389649, 0.04981350326538086, 0.04912639999389649, 0.04919705581665039, 0.049159168243408206, 0.049157119750976565, 0.04910899353027344, 0.04915507125854492, 0.04911513519287109, 0.04909260940551758, 0.049114112854003904, 0.04910182571411133, 0.0498073616027832, 0.05100646209716797, 0.050508800506591796, 0.05055897521972656, 0.05046783828735352, 0.050941951751708986, 0.05068902587890625, 0.05045043182373047, 0.05036544036865234, 0.05045145416259766, 0.050590721130371094, 0.04966912078857422, 0.050441215515136716, 0.05103308868408203, 0.05064704132080078, 0.050852863311767575, 0.05050982284545898, 0.05041049575805664, 0.050492416381835936, 0.05037363052368164, 0.05058150482177735, 0.10736537933349609, 0.04913049697875976, 0.04905472183227539, 0.04906086349487305, 0.049152000427246094, 0.0491038703918457, 0.049081344604492184, 0.049067008972167966, 0.04907417678833008, 0.04910489654541016, 0.04908748626708984, 0.04914995193481445, 0.049124351501464845, 0.049091583251953126, 0.04906291198730469, 0.049084415435791014, 0.04908236694335937, 0.04909260940551758, 0.05048422241210938, 0.05063782501220703, 0.050508800506591796, 0.04910489654541016, 0.0491038703918457, 0.04911001586914063, 0.04911308670043945, 0.049111038208007815, 0.04913356781005859, 0.049105918884277344, 0.049084415435791014, 0.04912639999389649, 0.04911206436157227, 0.04907929611206055, 0.049326080322265625, 0.049114112854003904, 0.04914585494995117, 0.049188865661621096, 0.04914176177978516, 0.04909568023681641, 0.04910899353027344, 0.04911718368530273, 0.04911206436157227, 0.049105918884277344, 0.049097728729248044, 0.04914483261108398, 0.049091583251953126, 0.04911513519287109, 0.04906393432617188, 0.049127422332763675, 0.04910182571411133, 0.0490967025756836, 0.04913971328735352, 0.04911513519287109, 0.04911513519287109, 0.04909465789794922, 0.04910182571411133, 0.04917145538330078, 0.0491069450378418, 0.04909465789794922, 0.049122303009033204, 0.049099777221679686, 0.04906598281860351, 0.049099777221679686, 0.049105918884277344, 0.10742886352539062, 0.049107967376708986, 0.04905984115600586, 0.04907724761962891, 0.04909875106811523, 0.0491253776550293, 0.049312767028808595, 0.05077299118041992, 0.050925567626953126, 0.05039923095703125, 0.050457599639892575, 0.05040435028076172, 0.04911820983886719, 0.049099777221679686, 0.04909875106811523, 0.049114112854003904, 0.04911206436157227, 0.04909465789794922, 0.04914790344238281, 0.04913971328735352, 0.050928638458251956, 0.05084774398803711, 0.050667518615722655, 0.05039820861816406, 0.050269184112548826, 0.04910182571411133, 0.04914790344238281, 0.049119232177734375, 0.049185791015625, 0.049132545471191405, 0.049127422332763675, 0.04911820983886719, 0.04907929611206055, 0.049152000427246094, 0.04907417678833008, 0.0491069450378418, 0.049138687133789063, 0.04914688110351562, 0.04910489654541016, 0.04907929611206055, 0.049121280670166016, 0.049132545471191405, 0.049097728729248044, 0.049122303009033204, 0.05336064147949219, 0.05167923355102539, 0.050607105255126954, 0.05047500610351562, 0.05101567840576172, 0.050618366241455076, 0.0505425910949707, 0.05124095916748047, 0.0514150390625, 0.050776065826416014, 0.05151129531860352, 0.05127884674072265, 0.05094297790527344, 0.05072281646728516, 0.05063577651977539, 0.050683902740478515, 0.05096243286132812, 0.05053849411010742, 0.05067161560058594, 0.10759474945068359, 0.04911001586914063, 0.05073715209960938, 0.0518389778137207, 0.050948097229003904, 0.050482177734375, 0.05047091293334961, 0.050267135620117184, 0.05040435028076172, 0.050427902221679685, 0.05044326400756836, 0.05060812759399414, 0.050902015686035154, 0.050563072204589846, 0.05043814468383789, 0.050444286346435545, 0.049122303009033204, 0.04912332916259766, 0.04911513519287109, 0.04915507125854492, 0.04906905746459961, 0.04910697555541992, 0.04911510467529297, 0.04911718368530273, 0.0493383674621582, 0.04910079956054687, 0.04908031845092774, 0.049170433044433595, 0.04909465789794922, 0.04909875106811523, 0.049127422332763675, 0.04950630569458008, 0.05036441421508789, 0.05189529418945313, 0.05050470352172851, 0.05053235244750977, 0.05045657730102539, 0.05043814468383789, 0.0505425910949707, 0.05018521499633789, 0.050299903869628904, 0.05061734390258789, 0.05047296142578125, 0.050907135009765625, 0.050531326293945314, 0.05062041473388672, 0.05048320007324219, 0.05062041473388672, 0.04913459014892578, 0.04912332916259766, 0.04910899353027344, 0.04914176177978516, 0.04912947082519531, 0.049152000427246094, 0.049121280670166016, 0.04916223907470703, 0.0491776008605957, 0.04912025451660156, 0.049102848052978515, 0.049083393096923826, 0.04906291198730469, 0.04909260940551758, 0.049073150634765625, 0.10775552368164062, 0.04906393432617188, 0.049977344512939455, 0.04905984115600586, 0.04905881500244141, 0.04907724761962891, 0.049067008972167966, 0.04910079956054687, 0.050547710418701174, 0.05044838333129883, 0.05040435028076172, 0.05038489532470703, 0.050411518096923826, 0.050530303955078126, 0.049105918884277344, 0.04919500732421875, 0.049165313720703124, 0.0491673583984375, 0.04914278411865235, 0.04911820983886719, 0.04912332916259766, 0.049091583251953126, 0.049083393096923826, 0.049116161346435545, 0.04909056091308594, 0.049122303009033204, 0.04914278411865235, 0.04921855926513672, 0.05071769714355469, 0.05073100662231445, 0.05098495864868164, 0.04917555236816406, 0.04910079956054687, 0.050476032257080077, 0.049168384552001954, 0.049132545471191405, 0.0491253776550293, 0.049173503875732424, 0.05061119842529297, 0.04907622528076172, 0.04909363174438477, 0.049138687133789063, 0.04909260940551758, 0.04914176177978516, 0.04910489654541016, 0.04911206436157227, 0.04911718368530273, 0.04912947082519531, 0.05040332794189453, 0.05049446487426758, 0.050476032257080077, 0.05055590438842773, 0.0491253776550293, 0.04912947082519531, 0.04910489654541016, 0.04912025451660156, 0.049135616302490234, 0.04914691162109375, 0.05030294418334961, 0.05048422241210938, 0.05081087875366211, 0.050493438720703124, 0.050885631561279294, 0.10820505523681641, 0.049056766510009765, 0.04908236694335937, 0.04904959869384766, 0.050695167541503904, 0.050836479187011716, 0.05061529541015625, 0.05053747177124023, 0.05067571258544922, 0.050444286346435545, 0.050441215515136716, 0.05077299118041992, 0.05043199920654297, 0.05038694381713867, 0.05033267211914062, 0.04909875106811523, 0.049107967376708986, 0.04911513519287109, 0.049089534759521485, 0.04910489654541016, 0.04909875106811523, 0.04909056091308594, 0.049099777221679686, 0.049154048919677736, 0.050462718963623046, 0.050477054595947264, 0.050326526641845705, 0.050402305603027345, 0.05139148712158203, 0.05121331024169922, 0.05090508651733398, 0.05076377487182617, 0.05069926452636719, 0.05075763320922851, 0.05053849411010742, 0.05055078506469726, 0.05094707107543945, 0.05069209671020508, 0.050945022583007815, 0.05057843017578125, 0.050895870208740236, 0.050700286865234374, 0.050388992309570314, 0.05045862579345703, 0.05046681594848633, 0.05048320007324219, 0.05045452880859375, 0.05044224166870117, 0.05035724639892578, 0.050424831390380856, 0.05047091293334961, 0.05087846374511719, 0.05183488082885742, 0.051074047088623044, 0.0504637451171875, 0.05063372802734375, 0.05080883026123047, 0.05051903915405274, 0.05108736038208008, 0.05058768081665039, 0.05035209655761719, 0.05030912017822266, 0.05164134216308594, 0.10880716705322266, 0.05060403060913086, 0.050336769104003906, 0.05051084899902344, 0.05053440093994141, 0.05089791870117188, 0.05114777755737305, 0.05058355331420898, 0.05031321716308594, 0.050487297058105465, 0.050840576171875, 0.05264691162109375, 0.05059993743896484, 0.050508800506591796, 0.050541568756103515, 0.05067161560058594, 0.05086003112792969, 0.05064191818237305, 0.05047500610351562, 0.05081497573852539, 0.05063167953491211, 0.05087641525268555, 0.050523136138916014, 0.04912639999389649, 0.04913049697875976, 0.049111038208007815, 0.04917248153686524, 0.049119232177734375, 0.04918272018432617, 0.05031219100952149, 0.050563072204589846, 0.05087846374511719, 0.05064601516723633, 0.05058969497680664, 0.05079142379760742, 0.05066342544555664, 0.04978483200073242, 0.050385921478271485, 0.050577407836914064, 0.05047808074951172, 0.050473983764648435, 0.05041561508178711, 0.0503900146484375, 0.050563072204589846, 0.05063372802734375, 0.05047296142578125, 0.05037670516967773, 0.050348033905029295, 0.05040127944946289, 0.050367488861083984, 0.05035520172119141, 0.050770942687988284, 0.04911820983886719, 0.049121280670166016, 0.0491069450378418, 0.049099777221679686, 0.04949094390869141, 0.04917145538330078, 0.04914483261108398, 0.05048524856567383, 0.05045145416259766, 0.05037670516967773, 0.050609153747558595, 0.10711244964599609, 0.04911001586914063, 0.04906291198730469, 0.04905881500244141, 0.0491038703918457, 0.049086463928222655, 0.04910079956054687, 0.04908544158935547, 0.049081344604492184, 0.050293758392333986, 0.04912639999389649, 0.04912844848632812, 0.049102848052978515, 0.04908544158935547, 0.04906905746459961, 0.049099777221679686, 0.04908748626708984, 0.049091583251953126, 0.04913663864135742, 0.049102848052978515, 0.04908031845092774, 0.049091583251953126, 0.0491069450378418, 0.049116161346435545, 0.04908236694335937, 0.049122303009033204, 0.0491069450378418, 0.049105918884277344, 0.0490967025756836, 0.049122303009033204, 0.0491253776550293, 0.04909875106811523, 0.049143806457519534, 0.04910182571411133, 0.049097728729248044, 0.04914278411865235, 0.04913971328735352, 0.049089534759521485, 0.049084415435791014, 0.04914585494995117, 0.04909056091308594, 0.04913459014892578, 0.04911513519287109, 0.04911718368530273, 0.049097728729248044, 0.049099777221679686, 0.050816001892089846, 0.04915302276611328, 0.04910390472412109, 0.049142753601074216, 0.04914585494995117, 0.04913356781005859, 0.04913049697875976, 0.049107967376708986, 0.049091583251953126, 0.049116161346435545, 0.04912332916259766, 0.049097728729248044, 0.04913459014892578, 0.0490885124206543, 0.0491038703918457, 0.04909465789794922, 0.049111038208007815, 0.10746163177490234, 0.04907724761962891, 0.05045452880859375, 0.05030297470092773, 0.05086003112792969, 0.050482177734375, 0.050685951232910156, 0.050541568756103515, 0.05048115158081055, 0.05044326400756836, 0.050525184631347655, 0.049083393096923826, 0.0491253776550293, 0.049102848052978515, 0.0491069450378418, 0.04909260940551758, 0.04929945755004883, 0.04914790344238281, 0.04919500732421875, 0.049135616302490234, 0.04911513519287109, 0.049121280670166016, 0.04907110214233398, 0.04911001586914063, 0.04907622528076172, 0.04908031845092774, 0.04915507125854492, 0.05031628799438476, 0.05039718246459961, 0.05027942276000977, 0.050367488861083984, 0.05046988677978516, 0.05051289749145508, 0.05048934555053711, 0.050482177734375, 0.050097152709960936, 0.05110067367553711, 0.050667518615722655, 0.05048422241210938, 0.04909875106811523, 0.04913971328735352, 0.049114112854003904, 0.050479103088378906, 0.050695167541503904, 0.05079859161376953, 0.05037670516967773, 0.05090508651733398, 0.05190041732788086, 0.051672065734863284, 0.04914483261108398, 0.04911820983886719, 0.04909363174438477, 0.05062144088745117, 0.05076172637939453, 0.05055692672729492, 0.050702335357666016, 0.05098495864868164, 0.05084774398803711, 0.050667518615722655, 0.050898944854736325, 0.05171916961669922, 0.05083545684814453, 0.050374656677246096]",tokens/s,19.708627661511404,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -6818,7 +6818,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4686-76337dd644ea2cd31e133602;6b892cba-92b7-418f-8039-dfcbbb891db1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe9b0-02abc949063abdeb0b2cb06d;17a44f74-4f89-4ac0-9e73-b254fb5a091e) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6875,7 +6875,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmprnainiq2/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp87rs9a03/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): @@ -6917,7 +6917,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95a3-0cd292c37548d0d72a32a751;b8b5e108-8b14-4ccc-9579-6701b798ec1c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe772-10c0a013016c4f11722672e5;f11c1985-b947-4d35-baad-8cad3d9f6871) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6988,7 +6988,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9647-19e8ea46308c0c9f39958dd3;75375c8d-b5af-4d7f-bb75-fbd18b04576e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe815-66d59d92565a71c15244f194;676a111c-0386-45d6-9fa6-a99a48b36411) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7020,7 +7020,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5016.45312,8173.125632,0.0,7526.678528,6833.613824,s,10,3.326199401855469,0.33261994018554686,0.003080048658705347,0.3318737487792969,0.3344139831542969,0.33784259643554687,0.34058548706054687,"[0.34127120971679686, 0.3336520690917969, 0.33123345947265626, 0.33050531005859374, 0.330927734375, 0.32976422119140625, 0.33141766357421876, 0.332329833984375, 0.33259820556640624, 0.33249969482421876]",tokens/s,769.647183079867,kWh,3.900687001298405e-06,2.137033586950565e-06,2.0206181039483795e-05,2.6243901627732764e-05,tokens/kWh,9754647.141699262,MB,5016.45312,8173.125632,0.0,7526.678528,7110.583808,s,10,183.519623046875,18.3519623046875,0.004288348522690653,18.3511787109375,18.358202734375,18.358737109375,18.359164609375,"[18.359271484375, 18.348994140625, 18.349845703125, 18.3483828125, 18.3546015625, 18.358083984375, 18.350859375, 18.351498046875, 18.353595703125, 18.344490234375]",tokens/s,3.4328754034062285,kWh,0.00021665113780019963,0.00011874283365397788,0.001081981506265517,0.0014173754777196945,tokens/kWh,44448.34907215681,,s,629,186.2062080688478,0.2960353069457038,0.03944703464753294,0.2912143249511719,0.2916943786621094,0.2919354309082031,0.62257517578125,"[0.29116619873046873, 0.2915358581542969, 0.2909962158203125, 0.29132696533203123, 0.29103924560546873, 0.29121844482421877, 0.29109759521484374, 0.29120819091796873, 0.2913249206542969, 0.2912593994140625, 0.2912757873535156, 0.29111911010742186, 0.2910730285644531, 0.2911375427246094, 0.29116928100585937, 0.29137100219726564, 0.2911713256835938, 0.29115289306640624, 0.291462158203125, 0.2916761474609375, 0.29171713256835935, 0.2913617858886719, 0.29141400146484375, 0.29209906005859376, 0.29184716796875, 0.29161471557617186, 0.29145599365234376, 0.2914078674316406, 0.29128500366210935, 0.29134030151367185, 0.2919741516113281, 0.29177548217773436, 0.29137921142578127, 0.29178778076171874, 0.2918133850097656, 0.2917447814941406, 0.29174578857421873, 0.2921359252929687, 0.29115188598632813, 0.290977783203125, 0.2910382080078125, 0.2909767761230469, 0.2910525512695312, 0.2908293151855469, 0.29078936767578123, 0.2910259094238281, 0.2909624328613281, 0.29080474853515625, 0.29093582153320313, 0.2909808959960937, 0.29161981201171877, 0.29114471435546874, 0.290914306640625, 0.29103717041015625, 0.2916168518066406, 0.29117019653320314, 0.29111090087890623, 0.29126046752929685, 0.2911712646484375, 0.2911160278320313, 0.29122662353515627, 0.2909982604980469, 0.6255001831054687, 0.29120819091796873, 0.2914027404785156, 0.2912962646484375, 0.29111090087890623, 0.29109246826171875, 0.291209228515625, 0.2909593505859375, 0.29141094970703124, 0.291378173828125, 0.2910279541015625, 0.29116824340820313, 0.29119793701171875, 0.29114163208007815, 0.2911231994628906, 0.29109146118164064, 0.2910648193359375, 0.29119488525390624, 0.29128704833984376, 0.29122457885742187, 0.2912767944335938, 0.29190347290039065, 0.29122354125976563, 0.2913259582519531, 0.2910484619140625, 0.29144677734375, 0.29113037109375, 0.2916597900390625, 0.2917017517089844, 0.2913945617675781, 0.291080322265625, 0.2912357177734375, 0.29142630004882814, 0.29292755126953124, 0.2914487609863281, 0.29130239868164065, 0.2910894775390625, 0.2912685546875, 0.2911784973144531, 0.29153689575195313, 0.2912389221191406, 0.29098809814453125, 0.2912736511230469, 0.2915430297851562, 0.29123175048828126, 0.2910330810546875, 0.29101364135742186, 0.29123992919921876, 0.29120306396484374, 0.29105767822265627, 0.29113345336914065, 0.29084671020507813, 0.2909900817871094, 0.29090304565429687, 0.290951171875, 0.29124609375, 0.2915348510742187, 0.2909532470703125, 0.2907913818359375, 0.2914089050292969, 0.2912010192871094, 0.2910945129394531, 0.29110693359375, 0.6228212280273437, 0.2909962158203125, 0.29121636962890624, 0.2909593505859375, 0.291162109375, 0.2909972534179687, 0.29106689453125, 0.2911058044433594, 0.29145703125, 0.291662841796875, 0.2912471008300781, 0.29108941650390624, 0.2911866760253906, 0.2912204895019531, 0.2912194519042969, 0.29111090087890623, 0.2910064697265625, 0.29104129028320314, 0.2909378662109375, 0.2911313781738281, 0.29149594116210936, 0.29147647094726564, 0.2911918029785156, 0.2929029541015625, 0.2915798645019531, 0.29134539794921877, 0.29141195678710935, 0.29131787109375, 0.2910135498046875, 0.291483642578125, 0.2914928588867188, 0.29125323486328125, 0.2910013427734375, 0.29123992919921876, 0.29137716674804687, 0.29116619873046873, 0.29122354125976563, 0.2912204895019531, 0.29115805053710936, 0.29145083618164064, 0.29133416748046875, 0.29131875610351565, 0.2913157348632813, 0.29126959228515625, 0.2911016845703125, 0.2911048583984375, 0.29092343139648436, 0.2929295349121094, 0.29161471557617186, 0.29102386474609376, 0.291267578125, 0.2909962158203125, 0.2914375610351562, 0.2910167236328125, 0.2913781433105469, 0.2911907958984375, 0.2911426696777344, 0.2909788513183594, 0.2909173583984375, 0.2910310363769531, 0.2912215881347656, 0.29127874755859373, 0.2908927917480469, 0.6225029296875, 0.291240966796875, 0.29127987670898436, 0.2913617858886719, 0.2913382263183594, 0.2910033874511719, 0.2911457214355469, 0.2911641540527344, 0.29143243408203123, 0.2911856689453125, 0.2912788391113281, 0.2914222106933594, 0.2920099792480469, 0.2911703186035156, 0.2913064880371094, 0.2911098937988281, 0.2911754150390625, 0.2914334716796875, 0.29155123901367186, 0.29108941650390624, 0.29109963989257814, 0.291051513671875, 0.2908641357421875, 0.2909010009765625, 0.29108941650390624, 0.29113037109375, 0.29089797973632814, 0.29110470581054687, 0.290946044921875, 0.2912204895019531, 0.2911457214355469, 0.29170281982421875, 0.2912634582519531, 0.2912041015625, 0.2915431213378906, 0.29133712768554687, 0.2911508483886719, 0.29135565185546874, 0.29162701416015624, 0.29137615966796876, 0.2912163391113281, 0.29111294555664063, 0.2913996887207031, 0.2912686157226563, 0.2932623291015625, 0.2910310363769531, 0.2911180725097656, 0.2912808837890625, 0.29141299438476564, 0.29114471435546874, 0.290977783203125, 0.29087554931640625, 0.2910574951171875, 0.29096038818359377, 0.29089483642578123, 0.2909276123046875, 0.2909982604980469, 0.2909296569824219, 0.2908979187011719, 0.29132696533203123, 0.29107199096679687, 0.2911231994628906, 0.2913863830566406, 0.622603271484375, 0.2917027893066406, 0.29134951782226565, 0.29125018310546874, 0.2917386169433594, 0.2914908142089844, 0.29173348999023435, 0.2911733703613281, 0.29134335327148436, 0.2910812072753906, 0.2911098937988281, 0.29106381225585937, 0.29109658813476563, 0.29120205688476564, 0.2910320739746094, 0.2910947265625, 0.29100115966796875, 0.292279296875, 0.2911508483886719, 0.2910453796386719, 0.29116928100585937, 0.29125735473632813, 0.29149594116210936, 0.29107611083984375, 0.2910126037597656, 0.29110272216796873, 0.291125244140625, 0.29108941650390624, 0.2909861145019531, 0.29115072631835937, 0.2912153625488281, 0.29102386474609376, 0.2911457214355469, 0.29115289306640624, 0.29117236328125, 0.29140069580078126, 0.29126553344726563, 0.29102386474609376, 0.29119589233398435, 0.2918922119140625, 0.2919342041015625, 0.29135769653320315, 0.29173965454101564, 0.29154611206054687, 0.29135565185546874, 0.2912491455078125, 0.2913812561035156, 0.29117745971679687, 0.29169253540039064, 0.29132186889648437, 0.2913034362792969, 0.29131365966796874, 0.29162701416015624, 0.2916198425292969, 0.29157684326171873, 0.29161575317382815, 0.29180108642578123, 0.2936002502441406, 0.29074740600585935, 0.2910423583984375, 0.29112109375, 0.29105889892578124, 0.290897705078125, 0.6212341918945312, 0.29106689453125, 0.2912808837890625, 0.291009521484375, 0.291135498046875, 0.2910699462890625, 0.2918184814453125, 0.2910679016113281, 0.2911058044433594, 0.2919096374511719, 0.29177752685546876, 0.29155328369140626, 0.29200897216796873, 0.29156658935546875, 0.29147137451171873, 0.29141812133789063, 0.29161895751953126, 0.29161663818359373, 0.292021240234375, 0.29205810546875, 0.2920222778320313, 0.2917530212402344, 0.2912767333984375, 0.29116110229492187, 0.29100747680664063, 0.29129010009765627, 0.2913894348144531, 0.29166693115234377, 0.29114886474609375, 0.2912081298828125, 0.29109759521484374, 0.2914437255859375, 0.2911088562011719, 0.2910955505371094, 0.29103512573242185, 0.2910525512695312, 0.2909982604980469, 0.29115191650390626, 0.293800048828125, 0.29129302978515625, 0.29165670776367186, 0.29106381225585937, 0.290946044921875, 0.29162905883789064, 0.2911651916503906, 0.29107098388671876, 0.2913679504394531, 0.29179495239257813, 0.29130035400390625, 0.2910279541015625, 0.29130853271484375, 0.2909982604980469, 0.29144580078125, 0.29087945556640626, 0.29143243408203123, 0.29117953491210935, 0.2911651916503906, 0.29131265258789063, 0.2911754150390625, 0.2913474426269531, 0.2915758056640625, 0.2920316162109375, 0.29120806884765627, 0.6234173583984375, 0.291557373046875, 0.2918380126953125, 0.29149484252929686, 0.29161676025390626, 0.29160858154296876, 0.2914078674316406, 0.29153176879882814, 0.2913587341308594, 0.2913239135742188, 0.29197210693359377, 0.29134848022460935, 0.2913802185058594, 0.29154815673828127, 0.291346435546875, 0.2913414306640625, 0.2914947814941406, 0.2916659240722656, 0.2913699951171875, 0.2909399108886719, 0.2911231994628906, 0.29095013427734373, 0.290977783203125, 0.2927216491699219, 0.29113345336914065, 0.29114981079101565, 0.29107403564453127, 0.2911918029785156, 0.2909122619628906, 0.291061767578125, 0.29106893920898436, 0.2917857360839844, 0.29131570434570314, 0.2910535583496094, 0.2911559753417969, 0.29109146118164064, 0.2910341796875, 0.29110470581054687, 0.29129931640625, 0.2910627746582031, 0.2909686279296875, 0.29092755126953124, 0.2909470825195313, 0.2911006774902344, 0.2909060974121094, 0.29136587524414065, 0.2909696044921875, 0.2912542724609375, 0.2915430297851562, 0.2912122802734375, 0.2912204895019531, 0.29156558227539064, 0.29109658813476563, 0.2909214782714844, 0.29097164916992185, 0.29119692993164065, 0.291061767578125, 0.29099417114257814, 0.29109042358398435, 0.29147442626953124, 0.29125531005859373, 0.291282958984375, 0.2911488037109375, 0.6243696899414063, 0.2917304382324219, 0.2916188049316406, 0.2910771484375, 0.29127786254882815, 0.2910637512207031, 0.2911365051269531, 0.29100747680664063, 0.2916546630859375, 0.2911180725097656, 0.2911170654296875, 0.29096652221679686, 0.2911590270996094, 0.2909399108886719, 0.29097268676757815, 0.29142630004882814, 0.29239706420898437, 0.29116110229492187, 0.2918369140625, 0.29131570434570314, 0.29151852416992186, 0.2918962707519531, 0.2915921936035156, 0.2912767944335938, 0.29122775268554685, 0.291442626953125, 0.2913709716796875, 0.2911436767578125, 0.29121023559570314, 0.2912071533203125, 0.29118975830078125, 0.2910627746582031, 0.29127783203125, 0.2911385498046875, 0.2912276611328125, 0.29141299438476564, 0.2915604553222656, 0.29117645263671876, 0.2910699462890625, 0.2913536071777344, 0.29105767822265627, 0.291409912109375, 0.29124301147460935, 0.29146929931640625, 0.2910986328125, 0.2909337463378906, 0.2911016845703125, 0.2913239135742188, 0.29193624877929686, 0.2917181396484375, 0.2913301696777344, 0.2909848327636719, 0.2914642028808594, 0.2911160278320313, 0.291262451171875, 0.2912665710449219, 0.29157888793945314, 0.2910177307128906, 0.290872314453125, 0.2911313781738281, 0.291019775390625, 0.290977783203125, 0.2908876647949219, 0.6242959594726563, 0.29109146118164064, 0.29125836181640624, 0.29093069458007814, 0.2913310852050781, 0.290766845703125, 0.29151742553710935, 0.2912389221191406, 0.29116928100585937, 0.29127996826171876, 0.2913391723632813, 0.2912143249511719, 0.29121844482421877, 0.2917212219238281, 0.29165771484375, 0.2913197937011719, 0.2915440673828125, 0.292537353515625, 0.29179800415039064, 0.29155841064453125, 0.2919966735839844, 0.291230712890625, 0.2913167419433594, 0.29126553344726563, 0.2913648681640625, 0.2911160278320313, 0.2914089050292969, 0.29123077392578123, 0.2909859313964844, 0.29102694702148435, 0.29122457885742187, 0.2911651916503906, 0.2914170837402344, 0.2914068603515625, 0.29143243408203123, 0.29179800415039064, 0.2914170837402344, 0.29126553344726563, 0.2910986328125, 0.2912225341796875, 0.2914498596191406, 0.29127783203125, 0.2909962158203125, 0.2915625, 0.2915471496582031, 0.29119692993164065, 0.29103616333007815, 0.2915000305175781, 0.29133517456054686, 0.29093582153320313, 0.29118771362304685, 0.2910597229003906, 0.2912747497558594, 0.2911539306640625, 0.29163418579101563, 0.2911805419921875, 0.29106893920898436, 0.29166387939453126, 0.2909224853515625, 0.2912563171386719, 0.29143243408203123, 0.2914846801757813, 0.2908927917480469, 0.6241136474609374, 0.29096652221679686, 0.2912112731933594, 0.29113446044921876, 0.2913536071777344, 0.2909900817871094, 0.29119488525390624, 0.2911098937988281, 0.2912286682128906, 0.29107199096679687, 0.2915706787109375, 0.2912286682128906, 0.2910863342285156, 0.2910955505371094, 0.2912276611328125, 0.2910618286132812, 0.29106475830078127, 0.29117645263671876, 0.29103411865234374, 0.29098599243164064, 0.29118881225585935, 0.29100946044921877, 0.2909153137207031, 0.29101568603515626, 0.2915973205566406, 0.2913239135742188, 0.2927995300292969, 0.2914518737792969, 0.2909532165527344, 0.29111294555664063, 0.29151028442382815, 0.29141094970703124, 0.2911375427246094, 0.29108944702148437, 0.290989013671875, 0.29102694702148435, 0.29118875122070315, 0.2913966064453125, 0.2912143249511719, 0.2912030944824219, 0.2909511413574219, 0.29100851440429687, 0.290956298828125, 0.2911313781738281, 0.29113958740234375, 0.29146829223632814, 0.29121636962890624, 0.291093505859375, 0.29149798583984377, 0.2911119689941406, 0.291536865234375, 0.29114675903320314, 0.2909624328613281, 0.2908794860839844, 0.2909542541503906, 0.29093582153320313, 0.29088870239257814, 0.29089181518554685, 0.29084463500976565, 0.29150106811523435, 0.29094400024414063, 0.29157684326171873, 0.2909276123046875]",tokens/s,3.3779754527165613,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7046,11 +7046,11 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpjula_18j/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmps0v9meys/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4755.23072,21498.42944,0.0,20851.982336,20236.60032,s,10,15.643266601562498,1.5643266601562498,0.001716851893423332,1.563902587890625,1.5660734985351563,1.567352362060547,1.5683754528808596,"[1.5631807861328124, 1.5647735595703125, 1.5629490966796875, 1.56462890625, 1.563903564453125, 1.5627757568359375, 1.565789306640625, 1.563901611328125, 1.5686312255859376, 1.5627327880859374]",tokens/s,163.6486844598237,kWh,1.846329511867629e-05,1.0117909425098331e-05,9.183712902519626e-05,0.00012041833356897088,tokens/kWh,2125922.1284055826,MB,4755.23072,21498.42944,0.0,20851.982336,20339.706368,s,10,879.77465625,87.97746562500001,0.0022909340685085054,87.97753515625,87.97981953125,87.980624609375,87.981268671875,"[87.974140625, 87.978578125, 87.9789609375, 87.9772265625, 87.979640625, 87.97715625, 87.97784375, 87.97353125, 87.9814296875, 87.9761484375]",tokens/s,0.716092462455496,kWh,0.0010385792660216491,0.0005692333667929961,0.005151564232359207,0.006759376865173853,tokens/kWh,9320.38577765846,,s,629,892.4578986816405,1.4188519851854382,0.185906583948783,1.396368408203125,1.3966589599609376,1.3968528564453124,2.960364462890625,"[1.39618505859375, 1.39667041015625, 1.3964708251953124, 1.396378662109375, 1.3963243408203125, 1.3962271728515625, 1.3963038330078126, 1.3964288330078125, 1.3964349365234374, 1.396369384765625, 1.3964697265625, 1.3965823974609375, 1.3962998046875, 1.3964708251953124, 1.3964349365234374, 1.3963857421875, 1.396304931640625, 1.3962802734375, 1.3960714111328125, 1.396305908203125, 1.3961553955078125, 1.3963427734375, 1.396221923828125, 1.3963756103515625, 1.39614111328125, 1.396317138671875, 1.3963294677734375, 1.3963673095703124, 1.396279296875, 1.3963018798828124, 1.3959813232421876, 1.3964195556640624, 1.396304931640625, 1.396304931640625, 1.396275146484375, 1.3961236572265625, 1.3962794189453125, 1.396401123046875, 1.39871435546875, 1.396484130859375, 1.396558837890625, 1.396505615234375, 1.3966868896484375, 1.3962772216796875, 1.3962833251953124, 1.3960928955078125, 1.396380615234375, 1.3968824462890626, 1.3964442138671875, 1.396285400390625, 1.3964237060546876, 1.396358154296875, 1.396305908203125, 1.3966787109375, 1.3965537109375, 1.3964185791015624, 1.3962916259765625, 1.3968353271484375, 1.396358154296875, 1.396221923828125, 1.39633251953125, 1.396252685546875, 2.962231201171875, 1.396673583984375, 1.3970401611328125, 1.396295654296875, 1.3965557861328124, 1.396348876953125, 1.3965035400390624, 1.3961922607421875, 1.3968536376953125, 1.39629052734375, 1.396274169921875, 1.396199462890625, 1.3962281494140625, 1.3965833740234375, 1.396538330078125, 1.3962445068359375, 1.3963079833984375, 1.3963243408203125, 1.396515869140625, 1.3962762451171875, 1.396073486328125, 1.3978162841796875, 1.396114501953125, 1.396137939453125, 1.3962198486328126, 1.3962987060546874, 1.3961595458984375, 1.3963714599609376, 1.396452392578125, 1.3961502685546876, 1.3962720947265626, 1.3963868408203124, 1.3964185791015624, 1.3962239990234375, 1.396464599609375, 1.3964482421875, 1.396427734375, 1.396380615234375, 1.396588623046875, 1.3964810791015625, 1.3967718505859374, 1.3981634521484374, 1.3962802734375, 1.396275146484375, 1.3962301025390624, 1.396253662109375, 1.396327392578125, 1.3961728515625, 1.3966571044921876, 1.396474853515625, 1.3965128173828125, 1.3963858642578124, 1.396339599609375, 1.3961318359375, 1.3965147705078125, 1.3963387451171876, 1.396506591796875, 1.3962393798828125, 1.3964605712890625, 1.3965179443359375, 1.39639599609375, 1.3985382080078126, 1.3965301513671875, 2.9603779296875, 1.396199462890625, 1.3967340087890625, 1.3962291259765625, 1.3964554443359376, 1.396252685546875, 1.3964892578125, 1.3960601806640625, 1.39671240234375, 1.3963714599609376, 1.3963509521484374, 1.3961092529296875, 1.39653125, 1.3965814208984375, 1.3964400634765626, 1.3963172607421874, 1.39658349609375, 1.3963939208984375, 1.396515869140625, 1.3962587890625, 1.3965823974609375, 1.3966663818359375, 1.396378662109375, 1.3987901611328124, 1.3961759033203125, 1.3963653564453125, 1.3963018798828124, 1.3964093017578125, 1.39652197265625, 1.3961544189453126, 1.3962332763671874, 1.39641650390625, 1.39623828125, 1.3962393798828125, 1.3961964111328125, 1.396336669921875, 1.3964564208984376, 1.3965137939453125, 1.3963756103515625, 1.3962701416015626, 1.396137939453125, 1.396569091796875, 1.39635205078125, 1.3963131103515625, 1.396252685546875, 1.3963427734375, 1.3963970947265625, 1.3995396728515626, 1.39681591796875, 1.396443115234375, 1.396485107421875, 1.3962608642578125, 1.3962506103515624, 1.3963192138671876, 1.3964769287109375, 1.396210693359375, 1.3968311767578125, 1.396547607421875, 1.39646875, 1.39649951171875, 1.396304931640625, 1.3964041748046876, 1.3961563720703125, 2.961944580078125, 1.39614208984375, 1.396427734375, 1.39610009765625, 1.3963775634765625, 1.397116943359375, 1.3965689697265624, 1.396379638671875, 1.396430908203125, 1.3963284912109375, 1.3966080322265626, 1.3961544189453126, 1.3964501953125, 1.396547607421875, 1.396279296875, 1.3963203125, 1.3963448486328125, 1.396454345703125, 1.3965128173828125, 1.396253662109375, 1.396316162109375, 1.396046875, 1.396062255859375, 1.3964974365234375, 1.396443115234375, 1.3962864990234376, 1.3960980224609374, 1.3964810791015625, 1.3962884521484376, 1.396199462890625, 1.396241455078125, 1.3960233154296875, 1.3961175537109376, 1.396675537109375, 1.3962762451171875, 1.3964237060546876, 1.3964083251953125, 1.3962884521484376, 1.3965762939453126, 1.3964708251953124, 1.396631591796875, 1.3964022216796874, 1.396336669921875, 1.396317138671875, 1.400958984375, 1.396559814453125, 1.3964820556640625, 1.39627001953125, 1.3966888427734374, 1.3963427734375, 1.396474853515625, 1.3962342529296874, 1.3963448486328125, 1.39620654296875, 1.396552734375, 1.3967205810546874, 1.3966029052734374, 1.3963560791015626, 1.3962332763671874, 1.3963212890625, 1.3964134521484375, 1.396279296875, 1.39620654296875, 2.96119921875, 1.3962393798828125, 1.39626806640625, 1.3964124755859375, 1.3966131591796875, 1.396216796875, 1.3965548095703124, 1.3961031494140625, 1.396580322265625, 1.396380615234375, 1.3969398193359375, 1.396273193359375, 1.39626904296875, 1.39641650390625, 1.3962547607421876, 1.39647900390625, 1.396178955078125, 1.3985556640625, 1.3966253662109376, 1.3968516845703125, 1.3966131591796875, 1.3960867919921875, 1.39594140625, 1.3963447265625, 1.396115478515625, 1.396316162109375, 1.3962301025390624, 1.3962618408203125, 1.3963406982421875, 1.3959669189453126, 1.396400146484375, 1.3961195068359376, 1.3963663330078124, 1.396368408203125, 1.3962496337890624, 1.39649951171875, 1.3963284912109375, 1.3964493408203125, 1.3962188720703126, 1.396115478515625, 1.3963602294921875, 1.3963663330078124, 1.3966182861328125, 1.3964237060546876, 1.3963038330078126, 1.3963653564453125, 1.39626806640625, 1.396369384765625, 1.3965137939453125, 1.3962486572265624, 1.3964317626953124, 1.3962711181640626, 1.39637353515625, 1.3964647216796875, 1.39634375, 1.3963253173828125, 1.396611083984375, 1.4011822509765626, 1.3966837158203125, 1.3966038818359374, 1.39641650390625, 1.3965537109375, 1.3965322265625, 2.960329833984375, 1.3966304931640625, 1.396526123046875, 1.3960919189453125, 1.396494384765625, 1.39626904296875, 1.3971036376953125, 1.39640625, 1.3964461669921875, 1.3961175537109376, 1.3965537109375, 1.3965025634765624, 1.3961400146484375, 1.396515869140625, 1.396490234375, 1.396189208984375, 1.3961563720703125, 1.3964564208984376, 1.3965404052734376, 1.396285400390625, 1.3961553955078125, 1.396074462890625, 1.395947509765625, 1.3964676513671874, 1.3965732421875, 1.3963448486328125, 1.3962977294921874, 1.3964267578125, 1.3964461669921875, 1.39627001953125, 1.3965732421875, 1.396262939453125, 1.39629052734375, 1.39640625, 1.3961185302734376, 1.3963243408203125, 1.3963826904296874, 1.3964442138671875, 1.400627197265625, 1.396809814453125, 1.396582275390625, 1.396357177734375, 1.39643798828125, 1.3963756103515625, 1.3962301025390624, 1.3963294677734375, 1.39649951171875, 1.3963336181640624, 1.3967288818359376, 1.396634521484375, 1.3963294677734375, 1.3961072998046875, 1.3963233642578126, 1.3963182373046874, 1.39650146484375, 1.396305908203125, 1.3965128173828125, 1.396368408203125, 1.3962659912109374, 1.39628955078125, 1.3963355712890626, 1.39620654296875, 1.39628955078125, 2.96319189453125, 1.39616259765625, 1.3965670166015625, 1.39634375, 1.39635205078125, 1.3960509033203126, 1.39610107421875, 1.3959833984375, 1.3964052734375, 1.3961861572265626, 1.39652197265625, 1.39647998046875, 1.3962496337890624, 1.3968414306640624, 1.3966519775390625, 1.3960438232421875, 1.3962802734375, 1.3961964111328125, 1.3965936279296876, 1.3964493408203125, 1.3962025146484376, 1.396211669921875, 1.3961871337890626, 1.3990369873046875, 1.39656396484375, 1.39658447265625, 1.3963192138671876, 1.3963756103515625, 1.396262939453125, 1.3963499755859374, 1.3965291748046875, 1.3964564208984376, 1.396116455078125, 1.3963765869140625, 1.396496337890625, 1.396341796875, 1.3961666259765626, 1.396453369140625, 1.396306884765625, 1.3964974365234375, 1.396357177734375, 1.396156494140625, 1.396138916015625, 1.39642578125, 1.3961502685546876, 1.396420654296875, 1.3963642578125, 1.39631201171875, 1.396749267578125, 1.396473876953125, 1.396526123046875, 1.396474853515625, 1.3964871826171874, 1.396432861328125, 1.396547607421875, 1.3964410400390626, 1.4002022705078125, 1.3966510009765625, 1.39624658203125, 1.396135986328125, 1.3961492919921874, 1.3963140869140624, 1.39603759765625, 2.96123388671875, 1.3961964111328125, 1.3963387451171876, 1.3962198486328126, 1.39618408203125, 1.3961973876953124, 1.3963714599609376, 1.396177978515625, 1.3965966796875, 1.396106201171875, 1.396715576171875, 1.3962823486328124, 1.3963294677734375, 1.3963028564453126, 1.396380615234375, 1.3961964111328125, 1.398497314453125, 1.3963970947265625, 1.3965679931640624, 1.396317138671875, 1.3965782470703125, 1.3963919677734375, 1.3962659912109374, 1.3962720947265626, 1.396201416015625, 1.396317138671875, 1.3962393798828125, 1.396137939453125, 1.3964708251953124, 1.3964288330078125, 1.3963857421875, 1.3960755615234375, 1.3963480224609375, 1.396232177734375, 1.3961451416015624, 1.3963182373046874, 1.396641845703125, 1.3966080322265626, 1.396378662109375, 1.396273193359375, 1.3963345947265624, 1.3962322998046874, 1.3966744384765626, 1.3964073486328126, 1.3963243408203125, 1.3962373046875, 1.3963499755859374, 1.3963345947265624, 1.3965465087890625, 1.396432861328125, 1.39626904296875, 1.39639404296875, 1.3961748046875, 1.3964892578125, 1.3966878662109374, 1.396232177734375, 1.396611083984375, 1.3963765869140625, 1.39654248046875, 1.39650048828125, 1.39634375, 1.3964093017578125, 1.3962454833984375, 2.96591357421875, 1.3961614990234374, 1.396864013671875, 1.3961820068359374, 1.396379638671875, 1.3964666748046874, 1.3964503173828124, 1.3963642578125, 1.39664892578125, 1.39633154296875, 1.396452392578125, 1.3962650146484374, 1.396241455078125, 1.3963756103515625, 1.3963775634765625, 1.396111328125, 1.3964503173828124, 1.3980078125, 1.3967994384765625, 1.396642822265625, 1.3961103515625, 1.3965987548828125, 1.396348876953125, 1.3962291259765625, 1.396262939453125, 1.39631201171875, 1.3963212890625, 1.3961759033203125, 1.3965260009765625, 1.3965496826171875, 1.396384765625, 1.3963182373046874, 1.396358154296875, 1.3962158203125, 1.3962720947265626, 1.396420654296875, 1.3962281494140625, 1.3967728271484374, 1.3964512939453124, 1.39666845703125, 1.3965936279296876, 1.3961728515625, 1.39624853515625, 1.3961861572265626, 1.39656396484375, 1.3964697265625, 1.396485107421875, 1.396537353515625, 1.3965670166015625, 1.3964073486328126, 1.3962578125, 1.3964390869140626, 1.3966990966796875, 1.39630078125, 1.3968045654296875, 1.3967421875, 1.3966285400390626, 1.3964482421875, 1.396533203125, 1.396506591796875, 1.396284423828125, 1.3962977294921874, 1.400701904296875, 2.960291748046875, 1.3964564208984376, 1.396536376953125, 1.396294677734375, 1.396358154296875, 1.39620458984375, 1.3963621826171875, 1.3960714111328125, 1.396537353515625, 1.396304931640625, 1.3964451904296875, 1.396430908203125, 1.3963642578125, 1.3963837890625, 1.396828125, 1.3963817138671875, 1.39660498046875, 1.396589599609375, 1.39631201171875, 1.396283447265625, 1.396177978515625, 1.39618505859375, 1.396168701171875, 1.396116455078125, 1.39642578125, 1.396654052734375, 1.399869384765625, 1.3961328125, 1.3965076904296876, 1.3962025146484376, 1.3964339599609374, 1.3961707763671876, 1.3962506103515624, 1.396093994140625, 1.3964124755859375, 1.396494384765625, 1.3964676513671874, 1.396410400390625, 1.396167724609375, 1.3962772216796875, 1.396236328125, 1.3962117919921875, 1.39633251953125, 1.396305908203125, 1.396552734375, 1.3963140869140624, 1.39652197265625, 1.396462646484375, 1.3968414306640624, 1.39652099609375, 1.396305908203125, 1.396357177734375, 1.3965721435546874, 1.39648193359375, 1.396537353515625, 1.396272216796875, 1.396579345703125, 1.3963602294921875, 1.3965322265625, 1.396306884765625, 1.3964041748046876, 1.396275146484375, 1.3961829833984376]",tokens/s,0.7047951516023033,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7079,7 +7079,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4ac5-3c9fa7a5650a59c842e84cf5;eff27cf1-71d9-4a38-a3d4-70b70d9a9fc8) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664feb3b-37d0916c26cd3b5108c1725d;503ff323-a7de-45dc-b8ec-3117cb67a2b3) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -7164,7 +7164,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9670-029fa01e630f16054e255509;64c5c163-98ba-46ba-a745-7edb4bea35f1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe83d-6aef8c200ec5ab9f7dc88bd4;bc5814c5-1714-49bc-b507-be17d053e781) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7226,7 +7226,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90b7-0bce57740c98bc325a04a391;82d56c37-4219-4eb1-9307-0b8337f9b813) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe263-3115b9dd73b6217d1f68d917;21b6cb9d-8d9c-4ed4-ac63-d514f463bfbd) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -7353,7 +7353,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e961b-526e746554ea78e347de5ec0;ec21f860-b615-4d97-a404-39ed466fb26a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe7e9-66bcac851ef198ed5e316754;d190559c-e2b9-44fe-a38f-2fb38c8c68a6) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7426,7 +7426,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9130-30299aef003754d473efd042;82e6d984-8c35-43a8-82b7-54aa2bd090e8) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe2e4-719e06fa6475de9a43a91868;b28e91b8-8280-4f3e-b1ab-c665c5c4b7e9) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7485,10 +7485,10 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpxyn267e1/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpdvfzq_xb/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7527,7 +7527,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4639-26c9c8cc29cf5fbf6742e3c0;22642277-9e05-4615-93bb-511fda798606) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe95d-3788d64974871b3e177d46f5;fe219b41-b80b-4896-a5e7-87b9da4baacf) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7558,7 +7558,7 @@ OSError: B is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7584,7 +7584,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpxlwhpvmm/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp7xljhxlx/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5937.815552,7053.246464,0.0,6406.79936,6098.60864,s,10,4.369107391357422,0.4369107391357422,0.0007208936334109947,0.43653796386718746,0.43786113586425784,0.4381150161743164,0.4383181204223633,"[0.438368896484375, 0.43641900634765624, 0.43653814697265625, 0.43647592163085935, 0.43767733764648437, 0.43614486694335936, 0.43780471801757814, 0.43653778076171873, 0.4368502807617187, 0.4362904357910156]",tokens/s,585.9320384442743,kWh,5.155067728913349e-06,2.8244480818023514e-06,2.4793715487132108e-05,3.2773231297847804e-05,tokens/kWh,7811252.960485814,MB,5937.815552,7055.343616,0.0,6408.896512,6157.952,s,10,245.01072265624998,24.501072265624998,0.0034318484925660374,24.5004931640625,24.5048515625,24.5070751953125,24.508854101562502,"[24.50105078125, 24.49864453125, 24.50170703125, 24.49968359375, 24.496791015625, 24.509298828125, 24.50153515625, 24.504357421875, 24.499935546875, 24.49771875]",tokens/s,2.571316035355277,kWh,0.0002892534986599056,0.000158535909402625,0.0013504422735896686,0.0017982316816521993,tokens/kWh,35034.417779869254,,s,629,248.55180075073253,0.3951538962650754,0.05191103267325261,0.38883840942382814,0.3892619018554688,0.38950277099609376,0.8256093530273438,"[0.38878106689453124, 0.38896536254882813, 0.388632568359375, 0.38873394775390624, 0.38894488525390625, 0.3886663818359375, 0.3890759582519531, 0.3891517333984375, 0.3889776611328125, 0.38876263427734375, 0.38862646484375, 0.38881890869140623, 0.3890196533203125, 0.3885875244140625, 0.38878616333007815, 0.3895060424804688, 0.38884146118164065, 0.38903399658203125, 0.389644287109375, 0.38915277099609374, 0.3891998596191406, 0.3889284973144531, 0.38931558227539065, 0.3889561462402344, 0.3889131469726563, 0.38835916137695314, 0.38861004638671875, 0.38877081298828126, 0.388674560546875, 0.3886622619628906, 0.3889428405761719, 0.38882098388671876, 0.3894353942871094, 0.3886909484863281, 0.38898175048828126, 0.38895718383789063, 0.38856500244140624, 0.3886120910644531, 0.389538818359375, 0.38871450805664065, 0.38864486694335937, 0.3886510009765625, 0.3886561279296875, 0.38851071166992185, 0.3891835021972656, 0.38876364135742186, 0.3887032470703125, 0.3888701477050781, 0.38866943359375, 0.3892899780273438, 0.3888373718261719, 0.388600830078125, 0.3887615966796875, 0.3901071472167969, 0.3886766052246094, 0.3887728576660156, 0.3887442016601563, 0.3886929931640625, 0.3890124816894531, 0.3890810852050781, 0.38887628173828126, 0.388568115234375, 0.8254780883789062, 0.3889141845703125, 0.38875750732421877, 0.38877593994140625, 0.3889254455566406, 0.3885977478027344, 0.3893872680664062, 0.38873907470703123, 0.38861721801757815, 0.3886346130371094, 0.38904217529296875, 0.3888578491210937, 0.38904013061523435, 0.38923876953125, 0.38863873291015627, 0.38872164916992186, 0.38861822509765626, 0.3884800109863281, 0.3887462463378906, 0.3886069641113281, 0.3893196716308594, 0.38932070922851564, 0.3890186157226562, 0.3887646789550781, 0.38994842529296875, 0.38882098388671876, 0.3890513916015625, 0.3891650695800781, 0.38888037109375, 0.3886991882324219, 0.3889468688964844, 0.3885434875488281, 0.3888199768066406, 0.3885844421386719, 0.388600830078125, 0.3886417846679687, 0.3889479675292969, 0.3886417846679687, 0.3886336059570312, 0.3887790222167969, 0.38892236328125, 0.38850970458984374, 0.38881793212890625, 0.3887523803710938, 0.3886868591308594, 0.38856500244140624, 0.3889899597167969, 0.38897457885742187, 0.38872164916992186, 0.3891282043457031, 0.38919577026367186, 0.3890595703125, 0.3892817993164063, 0.38935140991210937, 0.3886397399902344, 0.38855474853515626, 0.3887595520019531, 0.3887523803710938, 0.3889899597167969, 0.38894900512695313, 0.38868582153320314, 0.38889370727539063, 0.3889704895019531, 0.8256952514648438, 0.38874215698242187, 0.38912716674804687, 0.3890257873535156, 0.38883328247070315, 0.3886970825195312, 0.3889541015625, 0.38887527465820315, 0.389212158203125, 0.389137451171875, 0.38895819091796874, 0.3886970825195312, 0.3888670654296875, 0.38879641723632813, 0.38865716552734375, 0.3888598937988281, 0.3887564697265625, 0.3886991271972656, 0.38894900512695313, 0.3889837951660156, 0.38897357177734376, 0.3888793640136719, 0.38878106689453124, 0.38885479736328127, 0.3889776611328125, 0.38883840942382814, 0.38856396484375, 0.3887400817871094, 0.38860800170898435, 0.3888199768066406, 0.38868582153320314, 0.3887298583984375, 0.38889166259765623, 0.38933505249023437, 0.3892725830078125, 0.3889756164550781, 0.38903604125976565, 0.38886810302734376, 0.38863665771484374, 0.3886458740234375, 0.3889704895019531, 0.38879129028320314, 0.3888148498535156, 0.3887943725585937, 0.3887175598144531, 0.38925106811523436, 0.3889131469726563, 0.38866534423828125, 0.3887677307128906, 0.38883840942382814, 0.388853759765625, 0.388853759765625, 0.3888025512695312, 0.38896536254882813, 0.38888958740234375, 0.3887513732910156, 0.38878823852539063, 0.38919064331054687, 0.38868582153320314, 0.39011123657226565, 0.3893196716308594, 0.38929306030273436, 0.38961663818359377, 0.8260413208007813, 0.38885992431640626, 0.3889643249511719, 0.3887523803710938, 0.3887247314453125, 0.3887185974121094, 0.38895718383789063, 0.38912408447265623, 0.3887984619140625, 0.3889039306640625, 0.388822021484375, 0.3887298583984375, 0.388811767578125, 0.3887718505859375, 0.38874215698242187, 0.3888404541015625, 0.38900531005859373, 0.38877694702148435, 0.3888035888671875, 0.3886673889160156, 0.38864999389648436, 0.38919064331054687, 0.38883941650390624, 0.38895718383789063, 0.38876263427734375, 0.38873394775390624, 0.3887442016601563, 0.3888988037109375, 0.38905035400390625, 0.38884658813476564, 0.38873294067382813, 0.3889889221191406, 0.3889121398925781, 0.3887595520019531, 0.3887544250488281, 0.38894488525390625, 0.38860800170898435, 0.38863873291015627, 0.3886981201171875, 0.38873294067382813, 0.3888630065917969, 0.3888568115234375, 0.3888783264160156, 0.38862539672851565, 0.3889613037109375, 0.389233642578125, 0.38876263427734375, 0.38913946533203125, 0.3891650695800781, 0.3887237243652344, 0.38868582153320314, 0.38899713134765623, 0.3889704895019531, 0.3898193969726563, 0.3890257873535156, 0.3886673889160156, 0.38894692993164065, 0.3890390930175781, 0.3892991943359375, 0.38878515625, 0.3889059753417969, 0.3887462463378906, 0.3888424987792969, 0.826693603515625, 0.3888025512695312, 0.3887523803710938, 0.3885926513671875, 0.3887923278808594, 0.3886120910644531, 0.38872576904296874, 0.38866021728515626, 0.3889807434082031, 0.38920294189453125, 0.38865304565429687, 0.38867559814453123, 0.3889039306640625, 0.388706298828125, 0.3886264343261719, 0.3890257873535156, 0.3886796875, 0.38881893920898436, 0.388748291015625, 0.3887544250488281, 0.3898101806640625, 0.38882098388671876, 0.3889725341796875, 0.38886605834960936, 0.3887820739746094, 0.38895001220703124, 0.3888015441894531, 0.38883840942382814, 0.3887093811035156, 0.38877593994140625, 0.3885823974609375, 0.3885844421386719, 0.3888087158203125, 0.388632568359375, 0.388895751953125, 0.3889407958984375, 0.3887400817871094, 0.38869403076171877, 0.38845440673828124, 0.388706298828125, 0.3888496704101562, 0.38918142700195313, 0.3888035888671875, 0.38866534423828125, 0.3889039306640625, 0.38894488525390625, 0.3888486328125, 0.38889984130859373, 0.3886970825195312, 0.3889612731933594, 0.3889326171875, 0.38907904052734377, 0.38912612915039063, 0.388790283203125, 0.3889459228515625, 0.38889675903320314, 0.3888025512695312, 0.38870016479492187, 0.38883428955078125, 0.3888517150878906, 0.38923776245117186, 0.38900222778320315, 0.38878616333007815, 0.8254668579101563, 0.389001220703125, 0.3892408447265625, 0.38915377807617185, 0.38880767822265627, 0.38871142578125, 0.38886297607421877, 0.3891148681640625, 0.38876364135742186, 0.38882720947265625, 0.3886867980957031, 0.3885404052734375, 0.38891827392578127, 0.38905035400390625, 0.38921829223632814, 0.3889141845703125, 0.3891558532714844, 0.38889471435546874, 0.3890810852050781, 0.38909234619140626, 0.3887319030761719, 0.38957876586914064, 0.3891712036132812, 0.3887923583984375, 0.3895316162109375, 0.38924493408203126, 0.38886605834960936, 0.38888653564453124, 0.38975283813476563, 0.38893875122070315, 0.38879641723632813, 0.38891009521484377, 0.38910565185546875, 0.3892593078613281, 0.39087921142578125, 0.38871551513671876, 0.38870834350585937, 0.38860186767578125, 0.3886960754394531, 0.38911077880859374, 0.38857318115234374, 0.3887657470703125, 0.3889714660644531, 0.38878106689453124, 0.3889807434082031, 0.3888148498535156, 0.3887872009277344, 0.3890739135742188, 0.38941082763671875, 0.3891589050292969, 0.38924798583984377, 0.3888977966308594, 0.38902792358398436, 0.3890799560546875, 0.3894364013671875, 0.3890677795410156, 0.388959228515625, 0.38916094970703125, 0.3891568603515625, 0.3888988037109375, 0.38902987670898437, 0.38889370727539063, 0.38949786376953127, 0.8270551147460937, 0.38860186767578125, 0.38860595703125, 0.38868377685546873, 0.38870834350585937, 0.38866021728515626, 0.3888844909667969, 0.38888858032226564, 0.3890831298828125, 0.388947998046875, 0.38878512573242185, 0.389180419921875, 0.38898175048828126, 0.38901556396484377, 0.38881381225585937, 0.38938214111328123, 0.38876568603515627, 0.38882098388671876, 0.3885455322265625, 0.3887093811035156, 0.3890534362792969, 0.38878515625, 0.3887400817871094, 0.3889725341796875, 0.3888455810546875, 0.38887628173828126, 0.3890677795410156, 0.38898175048828126, 0.38893771362304685, 0.3888015441894531, 0.38900735473632814, 0.3888015441894531, 0.38892440795898436, 0.3890124816894531, 0.3895828552246094, 0.3893197021484375, 0.3891650390625, 0.3890083923339844, 0.38891827392578127, 0.38878311157226564, 0.3886766052246094, 0.38947021484375, 0.3891251220703125, 0.38870834350585937, 0.3887933349609375, 0.3888517150878906, 0.38911181640625, 0.38871142578125, 0.38868582153320314, 0.3887319030761719, 0.388611083984375, 0.38874215698242187, 0.388890625, 0.3886889038085938, 0.38894692993164065, 0.38870220947265627, 0.3885906066894531, 0.38906878662109373, 0.3886192626953125, 0.38860186767578125, 0.3894384765625, 0.38954290771484373, 0.3893800964355469, 0.825660400390625, 0.38914047241210936, 0.3887677307128906, 0.38940365600585936, 0.38903704833984376, 0.3888568420410156, 0.38899917602539064, 0.3896258544921875, 0.3887247314453125, 0.3890565185546875, 0.3885342712402344, 0.38850048828125, 0.38885888671875, 0.38866021728515626, 0.38881793212890625, 0.3884103698730469, 0.38903604125976565, 0.38858648681640623, 0.3889070129394531, 0.3887185974121094, 0.38888858032226564, 0.3892193298339844, 0.3894497375488281, 0.38906060791015623, 0.3886120910644531, 0.388600830078125, 0.38877490234375, 0.3888005065917969, 0.38874725341796873, 0.38922955322265623, 0.3888926696777344, 0.3885957336425781, 0.38874212646484374, 0.3889346618652344, 0.3888650207519531, 0.38876876831054685, 0.388664306640625, 0.38930841064453126, 0.3888230285644531, 0.38908517456054686, 0.38917633056640627, 0.38882611083984375, 0.3889510498046875, 0.38876058959960935, 0.3885772705078125, 0.3890186157226562, 0.3886346130371094, 0.3894353942871094, 0.389170166015625, 0.3888036193847656, 0.38960125732421874, 0.3888005065917969, 0.3895419006347656, 0.3890739135742188, 0.3897927551269531, 0.38913229370117186, 0.38885888671875, 0.3889837951660156, 0.388969482421875, 0.38935653686523436, 0.38904013061523435, 0.3888988037109375, 0.38913433837890626, 0.8261171264648437, 0.38934527587890627, 0.3890943908691406, 0.3892561950683594, 0.38904730224609374, 0.3889776611328125, 0.38878823852539063, 0.38880459594726563, 0.38874215698242187, 0.38939239501953127, 0.3890657348632813, 0.3885916137695313, 0.38864999389648436, 0.3887585144042969, 0.388790283203125, 0.3887032470703125, 0.3896012878417969, 0.3887442016601563, 0.3886336059570312, 0.38861004638671875, 0.389037109375, 0.3888260498046875, 0.388701171875, 0.38865304565429687, 0.38868069458007815, 0.38857318115234374, 0.3886213073730469, 0.3887431640625, 0.38874725341796873, 0.38863873291015627, 0.388595703125, 0.38875341796875, 0.38857522583007814, 0.3887923278808594, 0.38920806884765624, 0.3891712036132812, 0.38883123779296874, 0.3889407958984375, 0.388748291015625, 0.3887503356933594, 0.38887527465820315, 0.38860287475585936, 0.38913742065429685, 0.3887790222167969, 0.3885629577636719, 0.38873394775390624, 0.38880563354492187, 0.38863565063476563, 0.38857318115234374, 0.3885987854003906, 0.38900531005859373, 0.3888097229003906, 0.38852197265625, 0.3885834350585938, 0.38872576904296874, 0.38862234497070314, 0.3915376586914063, 0.38877694702148435, 0.3889428405761719, 0.38899917602539064, 0.38912716674804687, 0.38925726318359377, 0.3891875610351562, 0.8258375854492187, 0.388811767578125, 0.38908212280273435, 0.3887953796386719, 0.38926849365234373, 0.38856192016601565, 0.38865304565429687, 0.38886297607421877, 0.38894692993164065, 0.38873086547851565, 0.3887933349609375, 0.3884656677246094, 0.38854656982421876, 0.38860287475585936, 0.3886929931640625, 0.3886868591308594, 0.38878411865234375, 0.3888066711425781, 0.3886141357421875, 0.3886141357421875, 0.38892340087890626, 0.38862335205078125, 0.3893319702148437, 0.3891773376464844, 0.38893875122070315, 0.38863565063476563, 0.3889203186035156, 0.3887298583984375, 0.38881689453125, 0.3891302490234375, 0.3888824462890625, 0.3886489562988281, 0.3886243896484375, 0.3889407958984375, 0.38874111938476563, 0.38874111938476563, 0.38891009521484377, 0.3886141357421875, 0.38878823852539063, 0.38871142578125, 0.3888291931152344, 0.38855270385742186, 0.38883328247070315, 0.388885498046875, 0.38914764404296875, 0.38934017944335936, 0.38926025390625, 0.38955938720703126, 0.3897701416015625, 0.3890442199707031, 0.38865716552734375, 0.3886612548828125, 0.3888455810546875, 0.38867559814453123, 0.38898483276367185, 0.3890862121582031, 0.38879129028320314, 0.38848614501953127, 0.3886970825195312, 0.3888097229003906, 0.3888128356933594, 0.3887206115722656, 0.38876980590820315]",tokens/s,2.530659597315939,, @@ -7627,7 +7627,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e945a-2cfe57cd2e2cf77e51e4eb5a;cd371f52-321a-4863-bf6b-f2f0f3f39933) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe62a-49f3e76d29bf38ad15fc50bb;17950160-1099-4f69-8cf4-660952bc2ace) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7685,7 +7685,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpht08bq6c/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpqmc8m0_b/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1269.682176,1019.74016,0.0,373.293056,277.410816,s,10,0.37923129653930665,0.037923129653930666,0.0012152886645715089,0.038048974990844725,0.038519168853759764,0.0396308479309082,0.04052019119262695,"[0.04074252700805664, 0.03821686553955078, 0.03797974395751953, 0.03823471832275391, 0.03626841735839844, 0.03608185577392578, 0.03733708953857422, 0.03811382293701172, 0.03827212905883789, 0.03798412704467773]",tokens/s,6750.497713035296,kWh,4.301884120160883e-07,2.3572264180256752e-07,1.164744366137984e-06,1.83065541995664e-06,tokens/kWh,139840626.0453229,MB,1269.682176,1019.74016,0.0,373.293056,323.047424,s,10,23.493592773437502,2.3493592773437504,0.014632907695307939,2.3561588134765623,2.362567529296875,2.3635712402343754,2.3643742089843753,"[2.34352734375, 2.36090478515625, 2.362344482421875, 2.364574951171875, 2.338309326171875, 2.331607421875, 2.357506103515625, 2.3608046875, 2.3548115234375, 2.3192021484375]",tokens/s,26.81582191687153,kWh,2.642717914328431e-05,1.4482863030778391e-05,6.568572578086276e-05,0.00010659576795492543,tokens/kWh,591017.8350292469,,s,629,23.794291763305676,0.037828762739754634,0.004511848848474209,0.03739033508300781,0.03767849044799805,0.03791810684204102,0.07427948760986337,"[0.03534745788574219, 0.03654348754882813, 0.03639910507202149, 0.03678003311157227, 0.03678003311157227, 0.036736000061035154, 0.036125694274902344, 0.03654348754882813, 0.036519935607910156, 0.0366110725402832, 0.03650764846801758, 0.03670425415039062, 0.03646771240234375, 0.037282817840576174, 0.03782860946655273, 0.03862527847290039, 0.037749759674072264, 0.03757056045532227, 0.037465087890625, 0.0372408332824707, 0.03734630584716797, 0.03750092697143555, 0.03727872085571289, 0.03741388702392578, 0.03740262222290039, 0.03726131057739258, 0.037303295135498044, 0.03731353759765625, 0.037188640594482424, 0.037268447875976565, 0.03742617416381836, 0.03730124664306641, 0.03739136123657227, 0.037250049591064455, 0.03735039901733399, 0.03733913421630859, 0.03732787322998047, 0.03716505432128906, 0.03726233673095703, 0.0372490234375, 0.03732070541381836, 0.03729305648803711, 0.03743129730224609, 0.03723980712890625, 0.037269504547119144, 0.03735039901733399, 0.03727769470214844, 0.03724800109863281, 0.03734124755859375, 0.037339073181152344, 0.037269504547119144, 0.03736678314208984, 0.0371701774597168, 0.037266433715820314, 0.03737190246582031, 0.037318656921386716, 0.03726847839355469, 0.03745894241333008, 0.03732787322998047, 0.03729715347290039, 0.037501953125, 0.037501953125, 0.07589580535888672, 0.03728793716430664, 0.037407745361328126, 0.037207038879394534, 0.03731353759765625, 0.03784089660644531, 0.037353473663330077, 0.03739340972900391, 0.0374015998840332, 0.03751731109619141, 0.03856486511230469, 0.037754878997802735, 0.037369857788085936, 0.03734220886230469, 0.03731353759765625, 0.0373125114440918, 0.03733606338500976, 0.037425151824951174, 0.037303295135498044, 0.037362686157226564, 0.038662143707275394, 0.037748737335205076, 0.03751116943359375, 0.03752243041992188, 0.03749785614013672, 0.03739340972900391, 0.03728793716430664, 0.03726540756225586, 0.03750707244873047, 0.03739340972900391, 0.03738828659057617, 0.03735859298706055, 0.037395454406738284, 0.03731558227539063, 0.037868545532226565, 0.037456897735595705, 0.03736883163452148, 0.03723161697387695, 0.03758489608764649, 0.03761151885986328, 0.03742822265625, 0.037384193420410154, 0.03741900634765625, 0.03744976043701172, 0.03728688049316406, 0.03746815872192383, 0.03732275390625, 0.037525505065917966, 0.0375203857421875, 0.03739340972900391, 0.03733913421630859, 0.03781631851196289, 0.03749785614013672, 0.03746815872192383, 0.03744870376586914, 0.037324798583984374, 0.03752345657348633, 0.03728691101074219, 0.0374015998840332, 0.03738828659057617, 0.037454849243164064, 0.037454849243164064, 0.03749273681640625, 0.0757176284790039, 0.03736166381835938, 0.03737702560424805, 0.03760332870483398, 0.03744768142700195, 0.037354496002197264, 0.03724595260620117, 0.037343231201171875, 0.03800678253173828, 0.03777740859985351, 0.03753472137451172, 0.03747430419921875, 0.0374835205078125, 0.037585918426513674, 0.037338111877441404, 0.0374466552734375, 0.03750912094116211, 0.037379070281982424, 0.03765964889526367, 0.03735551834106445, 0.03742310333251953, 0.03738726425170898, 0.03740364837646484, 0.037495807647705076, 0.03738214492797851, 0.037310462951660156, 0.03733401489257813, 0.03747430419921875, 0.03743436813354492, 0.0373831672668457, 0.03754393768310547, 0.037370880126953124, 0.03760025787353516, 0.037601280212402347, 0.037510143280029294, 0.0373043212890625, 0.0373125114440918, 0.03747225570678711, 0.03762688064575195, 0.037282817840576174, 0.037294078826904296, 0.03753267288208008, 0.037607425689697264, 0.037400577545166014, 0.037495807647705076, 0.03745075225830078, 0.038043647766113284, 0.037773311614990236, 0.03741798400878906, 0.03741183853149414, 0.03771289443969727, 0.03767193603515625, 0.03745280075073242, 0.0372674560546875, 0.03750092697143555, 0.038141952514648435, 0.037969921112060545, 0.037384193420410154, 0.03739852905273437, 0.03727977752685547, 0.03753366470336914, 0.037572608947753904, 0.03738726425170898, 0.07624192047119141, 0.03899084854125977, 0.037817344665527344, 0.03795356750488281, 0.03747836685180664, 0.03777536010742188, 0.03754905700683594, 0.03737497711181641, 0.03731763076782227, 0.03754598236083984, 0.0381317138671875, 0.037395454406738284, 0.03748556900024414, 0.03738726425170898, 0.037493759155273435, 0.037318656921386716, 0.03763916778564453, 0.03733299255371094, 0.037407745361328126, 0.03729817581176758, 0.03746918487548828, 0.03743129730224609, 0.03740364837646484, 0.03732275390625, 0.03753267288208008, 0.03765350341796875, 0.03748863983154297, 0.03742617416381836, 0.03767708969116211, 0.03730940628051758, 0.03756748962402344, 0.03845017623901367, 0.03810201644897461, 0.03751321411132812, 0.03726847839355469, 0.03728998565673828, 0.03749273681640625, 0.03731148910522461, 0.03733606338500976, 0.03725107192993164, 0.03753881454467774, 0.03738726425170898, 0.037282817840576174, 0.03728179168701172, 0.0374917106628418, 0.037351425170898435, 0.03746819305419922, 0.037434337615966796, 0.037440513610839846, 0.03745587158203125, 0.03747532653808594, 0.03769651031494141, 0.03751935958862305, 0.037397502899169925, 0.03741491317749023, 0.037466110229492186, 0.03749478530883789, 0.037566463470458986, 0.03733401489257813, 0.03742822265625, 0.03756851196289063, 0.037501953125, 0.03759820938110352, 0.07184793853759766, 0.03510374450683594, 0.03521535873413086, 0.03516928100585937, 0.03511603164672852, 0.03513241577148438, 0.03520000076293945, 0.035148799896240236, 0.03514470291137695, 0.03511500930786133, 0.0367718391418457, 0.03750400161743164, 0.03728998565673828, 0.037482494354248046, 0.0374835205078125, 0.03756032180786133, 0.037303295135498044, 0.037424129486083986, 0.03768729782104492, 0.03739648056030274, 0.03758492660522461, 0.03751830291748047, 0.03748659133911133, 0.03748863983154297, 0.037410816192626956, 0.03736883163452148, 0.03754396820068359, 0.03754390335083008, 0.037591041564941405, 0.03752140808105469, 0.03739136123657227, 0.03755212783813477, 0.03739340972900391, 0.03741491317749023, 0.03769139099121094, 0.037408767700195314, 0.038091777801513675, 0.03745382308959961, 0.0376360969543457, 0.03739136123657227, 0.03732070541381836, 0.03746303939819336, 0.03743334579467773, 0.037364734649658206, 0.037526527404785154, 0.0374015998840332, 0.037416961669921874, 0.037416961669921874, 0.03750604629516602, 0.037427200317382815, 0.037367809295654295, 0.037498878479003905, 0.037397502899169925, 0.03749683380126953, 0.037622814178466794, 0.03773948669433594, 0.037416961669921874, 0.03744464111328125, 0.03751728057861328, 0.0374835205078125, 0.03757056045532227, 0.037477375030517575, 0.037720062255859374, 0.0716236801147461, 0.035062782287597655, 0.03518668746948242, 0.035019775390625, 0.03531980895996094, 0.035253246307373046, 0.03521843338012695, 0.034840576171875, 0.03494911956787109, 0.03526863861083984, 0.03520406341552734, 0.035253246307373046, 0.03508428955078125, 0.0372490234375, 0.037318656921386716, 0.03726335906982422, 0.03732787322998047, 0.03786342239379883, 0.037288959503173826, 0.0374835205078125, 0.03732992172241211, 0.03781631851196289, 0.038765567779541016, 0.03768012619018555, 0.03770982360839844, 0.037367809295654295, 0.03745894241333008, 0.03721318435668945, 0.03721420669555664, 0.03736678314208984, 0.03769343948364258, 0.03735756683349609, 0.03937484741210937, 0.03793203353881836, 0.037556224822998044, 0.03736883163452148, 0.03729510498046875, 0.03745177459716797, 0.03745792007446289, 0.037561344146728515, 0.037422080993652344, 0.037353473663330077, 0.037594112396240234, 0.03732070541381836, 0.03734732818603516, 0.037340160369873046, 0.037397502899169925, 0.037362686157226564, 0.03757875061035156, 0.03736064147949219, 0.037321727752685545, 0.03742822265625, 0.03740364837646484, 0.037318656921386716, 0.03724492645263672, 0.03730022430419922, 0.03753472137451172, 0.0374200325012207, 0.03728486251831055, 0.03722751998901367, 0.03739136123657227, 0.037362686157226564, 0.037375999450683595, 0.07522509002685547, 0.03725209426879883, 0.03725823974609375, 0.0374466552734375, 0.03733916854858398, 0.03731145477294922, 0.03730124664306641, 0.03746406555175781, 0.037375999450683595, 0.03750707244873047, 0.037292030334472655, 0.037459968566894535, 0.03742924880981445, 0.037405696868896485, 0.03738009643554688, 0.0372592658996582, 0.03754905700683594, 0.03736067199707031, 0.03739644622802735, 0.037394432067871096, 0.037394432067871096, 0.037392383575439454, 0.037389312744140625, 0.0372674560546875, 0.03746713638305664, 0.037416961669921874, 0.03733606338500976, 0.03751321411132812, 0.03757056045532227, 0.037351425170898435, 0.03736883163452148, 0.037335041046142575, 0.03735756683349609, 0.03738521575927734, 0.03734630584716797, 0.03721932983398438, 0.037397502899169925, 0.03740364837646484, 0.037465087890625, 0.03727974319458008, 0.03751116943359375, 0.03736883163452148, 0.03720294570922852, 0.03728179168701172, 0.03764019012451172, 0.037395454406738284, 0.037294078826904296, 0.037379070281982424, 0.03743231964111328, 0.03737705612182617, 0.03722953414916992, 0.03727360153198242, 0.03775692749023438, 0.03735039901733399, 0.03786652755737305, 0.03738211059570312, 0.03764019012451172, 0.03744255828857422, 0.03744153594970703, 0.03740671920776367, 0.038042625427246096, 0.03756032180786133, 0.03753267288208008, 0.07615692901611328, 0.03728486251831055, 0.03751628875732422, 0.03750092697143555, 0.03723161697387695, 0.03742105484008789, 0.03728384017944336, 0.03728793716430664, 0.03723369598388672, 0.037351390838623044, 0.0372592658996582, 0.03703603363037109, 0.03736576080322265, 0.03826483154296875, 0.03825356674194336, 0.037498878479003905, 0.03745075225830078, 0.03750604629516602, 0.03733606338500976, 0.0373493766784668, 0.037405696868896485, 0.037548030853271484, 0.037897216796875, 0.0374015998840332, 0.03749273681640625, 0.03787980651855469, 0.03769548797607422, 0.03733401489257813, 0.037424129486083986, 0.03724288177490234, 0.03732275390625, 0.03731353759765625, 0.037348350524902346, 0.0373043212890625, 0.03746099090576172, 0.03751731109619141, 0.037424129486083986, 0.037477375030517575, 0.03738623809814453, 0.037395454406738284, 0.0372828483581543, 0.0373001937866211, 0.03731455993652344, 0.03749683380126953, 0.037678081512451174, 0.03893964767456055, 0.037498878479003905, 0.03726847839355469, 0.03724595260620117, 0.037381118774414065, 0.03728998565673828, 0.03745177459716797, 0.03747635269165039, 0.037425151824951174, 0.03755724716186523, 0.03739033508300781, 0.03731763076782227, 0.03747532653808594, 0.03732070541381836, 0.03763302230834961, 0.0374917106628418, 0.03746303939819336, 0.037392383575439454, 0.07604838562011719, 0.03738726425170898, 0.03732275390625, 0.037763072967529294, 0.03760537719726562, 0.037493759155273435, 0.03744153594970703, 0.037367809295654295, 0.037353473663330077, 0.03748966217041016, 0.037321727752685545, 0.03739136123657227, 0.03747020721435547, 0.03746201705932617, 0.03758899307250976, 0.037288959503173826, 0.03742617416381836, 0.037359615325927735, 0.03741900634765625, 0.03733708953857422, 0.0372592658996582, 0.03737497711181641, 0.03758182525634766, 0.03737395095825195, 0.03727052688598633, 0.03736678314208984, 0.037477375030517575, 0.03746201705932617, 0.03737702560424805, 0.03690291213989258, 0.03730022430419922, 0.037541889190673826, 0.03752140808105469, 0.03749683380126953, 0.03740979385375977, 0.037354496002197264, 0.037292030334472655, 0.03724288177490234, 0.03722956848144531, 0.03722956848144531, 0.0375203857421875, 0.03734630584716797, 0.03717836761474609, 0.0371671028137207, 0.037372928619384765, 0.037341182708740234, 0.03730636978149414, 0.03713945770263672, 0.03719987106323242, 0.03745177459716797, 0.03732889556884766, 0.03736166381835938, 0.0373493766784668, 0.03739648056030274, 0.03730739212036133, 0.03730739212036133, 0.03730944061279297, 0.03724390411376953, 0.037353473663330077, 0.03740467071533203, 0.0375203857421875, 0.03734630584716797, 0.03746406555175781, 0.07565721893310547, 0.0372674560546875, 0.0373125114440918, 0.037479423522949216, 0.03725721740722656, 0.03723468780517578, 0.0374200325012207, 0.03733196640014649, 0.03729817581176758, 0.03751321411132812, 0.03730739212036133, 0.03743334579467773, 0.03721932983398438, 0.03751833724975586, 0.03735244750976562, 0.037408767700195314, 0.03565260696411133, 0.035312641143798826, 0.03527782440185547, 0.0352542724609375, 0.03772415924072266, 0.0377077751159668, 0.03799244689941406, 0.03948953628540039, 0.037106689453125, 0.03671449661254883, 0.03655372619628906, 0.03666534423828125, 0.03620249557495117, 0.03514470291137695, 0.03519692611694336, 0.035335166931152344, 0.03523788833618164, 0.03600076675415039, 0.03778662490844727, 0.03708927917480469, 0.03679846572875976, 0.036724769592285156, 0.03671651077270508, 0.0366192626953125, 0.036624416351318356, 0.03651375961303711, 0.03685990524291992, 0.03674524688720703, 0.03836924743652344, 0.03679436874389649, 0.03653734588623047, 0.03676671981811523, 0.03650559997558594, 0.03653529739379883, 0.036706302642822264, 0.03682099151611328, 0.036528129577636716, 0.03685887908935547, 0.03667865753173828, 0.036674560546875, 0.036618240356445314, 0.03663052749633789, 0.03661721420288086, 0.03660595321655274, 0.0365926399230957, 0.036544513702392575, 0.036792320251464845]",tokens/s,26.434911627418618,, @@ -7849,7 +7849,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e940b-67209a1f111b85cf3d6eb775;9d195ddc-44d2-4504-b63b-7929ae6f233f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe5ce-553b65097a4f664d76754a49;a3292b5d-4ac9-443d-8e2c-9e8b20883e8e) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7880,7 +7880,7 @@ OSError: m is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7919,7 +7919,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46fe-327bfa9f17e3d2523c8de62b;57f2ca23-f104-4797-84ff-cc62037d71cb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fea35-5d3828444f286ae61c03dac4;72e64028-0b43-4d6f-8bd0-80ef15954f89) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7950,7 +7950,7 @@ OSError: 1 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -7989,7 +7989,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46af-77e75bfc7e5cf56323b1370d;9cb15cc2-3c31-4120-99ed-38196ca97c03) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe9d9-494001e0380713a5162d8805;239eb66a-1610-423c-8be0-b646f7125557) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8021,7 +8021,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1983.488,5989.9904,0.0,5343.543296,5028.441088,s,10,3.1337609558105464,0.31337609558105467,0.00027686268578395025,0.3133651885986328,0.31370623779296875,0.3137779235839844,0.31383527221679686,"[0.313849609375, 0.3131971130371094, 0.313050048828125, 0.3129657897949219, 0.3134375, 0.3131690673828125, 0.3132928771972656, 0.3134548034667969, 0.3136538391113281, 0.3136903076171875]",tokens/s,816.9097886210202,kWh,3.701830148282979e-06,2.028091685247091e-06,1.8870336276812644e-05,2.4600258110342712e-05,tokens/kWh,10406394.878124049,MB,1983.488,5989.9904,0.0,5343.543296,5239.77216,s,10,173.840755859375,17.384075585937502,0.0013451275670699147,17.3843994140625,17.385715820312498,17.38575537109375,17.38578701171875,"[17.382298828125, 17.382341796875, 17.383392578125, 17.385794921875, 17.38570703125, 17.38494921875, 17.382314453125, 17.385146484375, 17.383849609375, 17.3849609375]",tokens/s,3.6240063320342775,kWh,0.0002051686882082787,0.00011244981813404137,0.0010459510989129874,0.0013635696052553075,tokens/kWh,46202.26188468334,,s,629,176.38418560791018,0.2804200089155965,0.0372555521516828,0.2759075927734375,0.27617626953125,0.2762659851074219,0.5894606005859375,"[0.27620758056640626, 0.27565890502929685, 0.27574066162109373, 0.27581832885742186, 0.2757529602050781, 0.27581134033203125, 0.2757877807617187, 0.2757376098632813, 0.27581439208984376, 0.27559832763671877, 0.27564544677734376, 0.2756812744140625, 0.2757611389160156, 0.2756966247558594, 0.27579290771484377, 0.2757939147949219, 0.27595166015625, 0.2758276672363281, 0.275894287109375, 0.2764021911621094, 0.27620352172851564, 0.2758912048339844, 0.2757949523925781, 0.27583999633789064, 0.2758635559082031, 0.27592398071289065, 0.27586663818359375, 0.2758133850097656, 0.27602023315429686, 0.27596798706054687, 0.2758226013183594, 0.27598037719726565, 0.2760109252929687, 0.2758604736328125, 0.27593215942382815, 0.2759710693359375, 0.2758410339355469, 0.27593829345703125, 0.2758512573242187, 0.2756894836425781, 0.27609088134765625, 0.275820556640625, 0.27575704956054686, 0.27588095092773435, 0.27588711547851563, 0.27587686157226565, 0.27591680908203126, 0.2757427978515625, 0.2760017395019531, 0.27596697998046876, 0.2759280700683594, 0.2759126892089844, 0.27610317993164063, 0.27607244873046877, 0.27608062744140627, 0.2761144409179688, 0.2762260437011719, 0.27591680908203126, 0.2758604736328125, 0.27589529418945313, 0.27602227783203126, 0.27591680908203126, 0.5894880981445313, 0.2758164367675781, 0.27574072265625, 0.2757887268066406, 0.2757550048828125, 0.27588812255859374, 0.27581747436523435, 0.27586968994140626, 0.2757969970703125, 0.2757457885742188, 0.2757058410644531, 0.2757478332519531, 0.2758533020019531, 0.2759290771484375, 0.27576217651367185, 0.2758482360839844, 0.27584304809570315, 0.2757908935546875, 0.2758450927734375, 0.27569357299804687, 0.2757437438964844, 0.2757253112792969, 0.2757058410644531, 0.2757754821777344, 0.2762158203125, 0.27628851318359376, 0.2757662658691406, 0.27578369140625, 0.2758215637207031, 0.27607858276367186, 0.2759485473632812, 0.2758533020019531, 0.2758533020019531, 0.27594033813476565, 0.2760499267578125, 0.2760499267578125, 0.27593011474609375, 0.27593215942382815, 0.27595672607421873, 0.2758092651367188, 0.2756904907226562, 0.2759833679199219, 0.27587686157226565, 0.2761185302734375, 0.2759977111816406, 0.27580108642578127, 0.2761430969238281, 0.2759444580078125, 0.2758430786132812, 0.2761820068359375, 0.2760028076171875, 0.2757887878417969, 0.27588507080078123, 0.2760099792480469, 0.2759075927734375, 0.2759270324707031, 0.2759741516113281, 0.27580517578125, 0.27592498779296876, 0.2758184814453125, 0.2762926025390625, 0.27641445922851565, 0.27615130615234373, 0.5894983520507813, 0.27609600830078124, 0.2757376098632813, 0.2758164367675781, 0.2758512573242187, 0.2758912048339844, 0.27580517578125, 0.27581134033203125, 0.2758553466796875, 0.27580621337890626, 0.27591781616210936, 0.2756812744140625, 0.2757213439941406, 0.275948486328125, 0.2758133239746094, 0.2760325012207031, 0.27610931396484373, 0.27586968994140626, 0.2757376098632813, 0.27578469848632814, 0.27573248291015623, 0.27593829345703125, 0.2760468444824219, 0.2759813537597656, 0.2757641906738281, 0.27599462890625, 0.2759126892089844, 0.275926025390625, 0.27571615600585936, 0.27582763671875, 0.2758359985351562, 0.2757620849609375, 0.2760693664550781, 0.27618917846679686, 0.2759270324707031, 0.275736572265625, 0.2762188720703125, 0.2762076110839844, 0.2761646118164062, 0.2758727722167969, 0.2757396545410156, 0.2762577819824219, 0.2761164855957031, 0.2759045104980469, 0.2758502502441406, 0.2759444580078125, 0.2759137268066406, 0.2758912048339844, 0.2758287353515625, 0.2760447998046875, 0.2758389892578125, 0.27603045654296876, 0.2760561218261719, 0.2759925231933594, 0.27622308349609376, 0.2759412841796875, 0.27583282470703124, 0.27588711547851563, 0.27578164672851563, 0.2758533020019531, 0.27593524169921874, 0.276252685546875, 0.275904541015625, 0.589043701171875, 0.27563827514648437, 0.2755625, 0.27563623046875, 0.2757171325683594, 0.27581747436523435, 0.2759045104980469, 0.2758502502441406, 0.2761154479980469, 0.2760601501464844, 0.2758635559082031, 0.2757611389160156, 0.27609805297851564, 0.2761154479980469, 0.27605093383789064, 0.2759741516113281, 0.27581951904296875, 0.2759577941894531, 0.2758522644042969, 0.27575704956054686, 0.27579290771484377, 0.27595672607421873, 0.275873779296875, 0.27607040405273436, 0.27620248413085935, 0.2759495544433594, 0.2758533020019531, 0.2758236083984375, 0.27593112182617185, 0.2758287353515625, 0.27597210693359375, 0.27595162963867187, 0.2758983764648438, 0.27593011474609375, 0.2758912048339844, 0.2757918701171875, 0.27586251831054687, 0.27610931396484373, 0.2759608459472656, 0.2759813232421875, 0.27586764526367186, 0.2759833679199219, 0.2760919189453125, 0.27588812255859374, 0.27643902587890623, 0.27614822387695315, 0.2762680358886719, 0.276347900390625, 0.27591680908203126, 0.2761134033203125, 0.2760673217773438, 0.27598745727539065, 0.27604583740234373, 0.27612057495117187, 0.2760058898925781, 0.2759280700683594, 0.27597927856445315, 0.27601629638671876, 0.27583575439453123, 0.2760191955566406, 0.27625982666015625, 0.2762239990234375, 0.27597515869140626, 0.5897318115234375, 0.27561370849609373, 0.27581951904296875, 0.275766357421875, 0.2758869934082031, 0.2758778991699219, 0.2757601318359375, 0.27578265380859374, 0.27584716796875, 0.2758042297363281, 0.2758931884765625, 0.27586456298828127, 0.2757571105957031, 0.27601300048828126, 0.27611135864257813, 0.27604788208007813, 0.2758584289550781, 0.27632229614257814, 0.27613388061523436, 0.27607757568359376, 0.2758492126464844, 0.2757734375, 0.27580416870117186, 0.27582769775390625, 0.2757601318359375, 0.27602841186523436, 0.27597006225585935, 0.27583282470703124, 0.2757949523925781, 0.27580722045898437, 0.27597311401367186, 0.2761154479980469, 0.27593011474609375, 0.27602227783203126, 0.27585638427734377, 0.27580020141601563, 0.27591873168945313, 0.275884033203125, 0.2760130615234375, 0.27586968994140626, 0.2759444580078125, 0.27615435791015625, 0.27587890625, 0.27576849365234374, 0.2758582763671875, 0.27598028564453125, 0.27599462890625, 0.2760612487792969, 0.2757805480957031, 0.27640011596679687, 0.27632846069335937, 0.27581951904296875, 0.275924072265625, 0.27632733154296873, 0.2762403869628906, 0.2761011352539062, 0.27605810546875, 0.2767032470703125, 0.27598745727539065, 0.27586151123046876, 0.27600076293945314, 0.2762557373046875, 0.275926025390625, 0.589389892578125, 0.27571295166015625, 0.27556558227539063, 0.2757244567871094, 0.2759483947753906, 0.27599359130859374, 0.2757550659179687, 0.2757950134277344, 0.27580197143554686, 0.27577957153320315, 0.27570791625976565, 0.27570687866210936, 0.27567718505859373, 0.27588711547851563, 0.2757447814941406, 0.27579086303710937, 0.2757283935546875, 0.27566387939453124, 0.27588198852539064, 0.27588507080078123, 0.2758993835449219, 0.2761841735839844, 0.2760744018554688, 0.2758851623535156, 0.2758378601074219, 0.27610418701171874, 0.27607763671875, 0.27606112670898436, 0.27606439208984374, 0.2759279479980469, 0.2759290771484375, 0.27578469848632814, 0.2759977111816406, 0.2759977111816406, 0.27600076293945314, 0.2758512573242187, 0.27591781616210936, 0.27596493530273436, 0.2764861450195312, 0.27593215942382815, 0.27619635009765625, 0.27639193725585937, 0.27593215942382815, 0.27573452758789063, 0.27578469848632814, 0.2759638977050781, 0.275947509765625, 0.2759741516113281, 0.2757969970703125, 0.27604071044921874, 0.27595672607421873, 0.2758133850097656, 0.2759813232421875, 0.27590963745117186, 0.27603662109375, 0.2761707458496094, 0.27597515869140626, 0.27635302734375, 0.2761717834472656, 0.2761226196289063, 0.2761871337890625, 0.2763735046386719, 0.2762178649902344, 0.5899192504882812, 0.27583486938476565, 0.2756280212402344, 0.2756782531738281, 0.27569354248046873, 0.27574887084960936, 0.2757326354980469, 0.27568521118164063, 0.27571505737304686, 0.2757939147949219, 0.2757519226074219, 0.27565057373046875, 0.2759813232421875, 0.2763233337402344, 0.27594033813476565, 0.27570892333984376, 0.27588095092773435, 0.27579913330078126, 0.27584194946289065, 0.2757959594726562, 0.2757601318359375, 0.275915771484375, 0.2758717346191406, 0.2766981201171875, 0.27584820556640627, 0.27586663818359375, 0.2758164367675781, 0.2758860778808594, 0.27581951904296875, 0.2759198913574219, 0.2758471374511719, 0.27602227783203126, 0.2758389892578125, 0.27607757568359376, 0.27607858276367186, 0.2759741516113281, 0.27615130615234373, 0.2758215637207031, 0.2759270324707031, 0.2756812744140625, 0.2758584289550781, 0.2761246643066406, 0.2758021240234375, 0.2757969970703125, 0.27563006591796874, 0.27589016723632814, 0.275768310546875, 0.27595980834960937, 0.27612570190429686, 0.2761656188964844, 0.2760376281738281, 0.2757305297851563, 0.2759229125976562, 0.27594232177734374, 0.27598641967773435, 0.2760079345703125, 0.2759444580078125, 0.27622503662109377, 0.2760816650390625, 0.2758338623046875, 0.27582977294921873, 0.2762280883789062, 0.27591476440429685, 0.5897728271484375, 0.27567822265625, 0.27565875244140625, 0.27598745727539065, 0.2760601501464844, 0.2760601501464844, 0.2759466552734375, 0.27583779907226563, 0.27587890625, 0.2757652893066406, 0.2757242431640625, 0.2757949523925781, 0.2759208984375, 0.27594546508789064, 0.275768310546875, 0.2758553466796875, 0.27644723510742186, 0.2758533020019531, 0.27591064453125, 0.2758492431640625, 0.2760283813476562, 0.2760479431152344, 0.27597714233398435, 0.2759331970214844, 0.27583999633789064, 0.2758778991699219, 0.27581134033203125, 0.27582769775390625, 0.27572634887695313, 0.27586251831054687, 0.2757887878417969, 0.2759343566894531, 0.275971923828125, 0.27577752685546875, 0.2759813232421875, 0.2760325012207031, 0.2758430786132812, 0.27591476440429685, 0.2760652770996094, 0.2760570983886719, 0.2758092651367188, 0.2759495544433594, 0.27585946655273436, 0.27593112182617185, 0.27596697998046876, 0.2762127380371094, 0.27619430541992185, 0.27613287353515625, 0.2758840637207031, 0.27600994873046875, 0.27602328491210937, 0.2760018310546875, 0.27603350830078127, 0.2758922119140625, 0.2761584777832031, 0.27600076293945314, 0.2760468444824219, 0.27607553100585935, 0.276173828125, 0.27598541259765624, 0.2760816650390625, 0.2761902160644531, 0.27597927856445315, 0.5897615356445313, 0.27575604248046875, 0.27561575317382814, 0.27574887084960936, 0.27571505737304686, 0.2759024658203125, 0.275662841796875, 0.27566796875, 0.275810302734375, 0.2758042907714844, 0.27597607421875, 0.27585638427734377, 0.27588507080078123, 0.2757857360839844, 0.2758389892578125, 0.27602841186523436, 0.2758359069824219, 0.27657421875, 0.2758512573242187, 0.2756966247558594, 0.2758164367675781, 0.2758133850097656, 0.2758604736328125, 0.27580416870117186, 0.27576321411132815, 0.2758656005859375, 0.27587890625, 0.27585638427734377, 0.2758307800292969, 0.2758800048828125, 0.2761092529296875, 0.27588507080078123, 0.2758287353515625, 0.2758860778808594, 0.27581439208984376, 0.27581134033203125, 0.2760130615234375, 0.27615640258789065, 0.27614718627929685, 0.27591781616210936, 0.2758912048339844, 0.27607040405273436, 0.2759321899414062, 0.2759085693359375, 0.2759393310546875, 0.2760263671875, 0.27615545654296875, 0.2762147216796875, 0.27608062744140627, 0.2761134033203125, 0.2759915466308594, 0.2759741516113281, 0.27621478271484373, 0.27639398193359377, 0.2759977111816406, 0.27599359130859374, 0.2760386962890625, 0.27610015869140625, 0.27581329345703126, 0.2759362487792969, 0.2759834289550781, 0.27614407348632813, 0.27588198852539064, 0.5901117553710937, 0.2758656005859375, 0.27585638427734377, 0.2758584289550781, 0.2758635559082031, 0.2759761962890625, 0.2757877807617187, 0.2757386169433594, 0.27582464599609374, 0.27579904174804687, 0.27584613037109373, 0.2758553466796875, 0.2758135070800781, 0.2758398742675781, 0.2757652587890625, 0.27591064453125, 0.27581951904296875, 0.2758533020019531, 0.2759761962890625, 0.27597006225585935, 0.275915771484375, 0.27576422119140626, 0.27607244873046877, 0.2762127380371094, 0.27575091552734377, 0.27577752685546875, 0.27675442504882813, 0.27577957153320315, 0.2757877807617187, 0.2758184814453125, 0.2759004211425781, 0.27587994384765624, 0.2758584289550781, 0.2757611389160156, 0.2758389892578125, 0.27578369140625, 0.2759444580078125, 0.2761922607421875, 0.27635617065429685, 0.2758768005371094, 0.27578369140625, 0.27603353881835935, 0.2757949523925781, 0.27571609497070315, 0.27609906005859375, 0.2762629089355469, 0.276031494140625, 0.2760970153808594, 0.275957763671875, 0.2759761962890625, 0.2758922424316406, 0.2758799133300781, 0.276068359375, 0.27603155517578126, 0.27607647705078125, 0.2759610290527344, 0.2761746520996094, 0.27612570190429686, 0.27617483520507813, 0.2762178649902344, 0.2758922424316406, 0.27611236572265624, 0.27603558349609375]",tokens/s,3.5660793388712486,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8089,7 +8089,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e96c5-2e2140c84e678cba1d190585;581d82bf-67ba-445d-a568-1218d8584711) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe898-0a7f98764c8574815d88dc8d;722e36ff-2261-4d32-815a-6c7d2f4291e3) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8121,7 +8121,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1565.773824,2168.979456,0.0,1522.532352,1394.863616,s,10,0.7274297943115234,0.07274297943115235,0.0013986768503876123,0.07226521682739258,0.07347540283203124,0.07515016555786132,0.0764899757385254,"[0.0768249282836914, 0.07210610961914063, 0.07184393310546874, 0.07232908630371093, 0.07202118682861328, 0.07247756958007813, 0.0723966064453125, 0.07220134735107422, 0.07212579345703125, 0.07310323333740235]",tokens/s,3519.2399596759906,kWh,8.529549022372679e-07,4.6738028940643143e-07,3.933354465626244e-06,5.253689657269945e-06,tokens/kWh,48727659.359504156,MB,1566.101504,2168.979456,0.0,1522.532352,1491.576832,s,10,39.543352050781245,3.9543352050781246,0.008196721752012559,3.9515069580078124,3.9657108154296874,3.966758361816406,3.967596398925781,"[3.967805908203125, 3.948567138671875, 3.95064404296875, 3.945499267578125, 3.96494189453125, 3.96547802734375, 3.952369873046875, 3.9549228515625, 3.945201171875, 3.947921875]",tokens/s,15.931881525647071,kWh,4.658322652049964e-05,2.5530277081401715e-05,0.0002073834256987738,0.0002794969293006751,tokens/kWh,225404.98086197698,,s,629,40.12977773284911,0.0637993286690765,0.008641086270954917,0.06266368103027344,0.06314926147460938,0.06341548919677734,0.1351290490722656,"[0.06447001647949219, 0.06381260681152344, 0.06451712036132813, 0.06395187377929687, 0.06415462493896484, 0.06395904159545898, 0.06358937454223633, 0.06393247985839844, 0.06341020965576172, 0.06255913543701172, 0.06289203262329102, 0.06282035064697265, 0.06282547378540039, 0.062635009765625, 0.06270771026611328, 0.06300057601928712, 0.06288188934326172, 0.06253660964965821, 0.06260019302368164, 0.06271699142456055, 0.0629185905456543, 0.06360678482055664, 0.06323507308959961, 0.06260121536254883, 0.06290636825561523, 0.06293503952026368, 0.06285619354248047, 0.06276403045654297, 0.06258179092407226, 0.06296060943603515, 0.06260432052612305, 0.06311523056030273, 0.06329446411132812, 0.06290329742431641, 0.06268313598632813, 0.062635009765625, 0.06278656005859375, 0.06255001449584961, 0.06302822494506836, 0.06287564849853515, 0.06296575927734376, 0.06306713485717773, 0.06284902572631836, 0.06260432052612305, 0.06263395309448243, 0.06258073425292969, 0.06253977584838867, 0.06282854461669922, 0.06293913650512695, 0.0629483528137207, 0.062635009765625, 0.06271702575683594, 0.06260835266113281, 0.06260115051269531, 0.062458881378173826, 0.06255820846557616, 0.06286761474609374, 0.06274646377563477, 0.06261043167114258, 0.0626566390991211, 0.06248025512695313, 0.06259711837768554, 0.13514854431152343, 0.06287974548339843, 0.06256435012817382, 0.06253260803222656, 0.06251007843017578, 0.06260326385498047, 0.06248755264282226, 0.06260224151611328, 0.0625080337524414, 0.06253670501708984, 0.06258687973022461, 0.06260224151611328, 0.06252544021606446, 0.0625602569580078, 0.062499839782714846, 0.0625726089477539, 0.06289503860473633, 0.06260224151611328, 0.062561279296875, 0.06254694366455078, 0.06257561492919922, 0.06253158569335937, 0.06252134323120118, 0.06264012908935547, 0.06252441787719727, 0.06250188827514648, 0.06263296127319336, 0.06261248016357422, 0.06253260803222656, 0.06250092697143554, 0.062495681762695314, 0.062494720458984375, 0.06260326385498047, 0.06264831924438477, 0.06355766296386718, 0.0630865592956543, 0.0629678077697754, 0.06314086532592773, 0.06258278274536133, 0.06254796981811524, 0.06267903900146485, 0.06278860855102539, 0.0626237449645996, 0.06257459259033203, 0.06258687973022461, 0.06263091278076172, 0.0631982078552246, 0.06278153610229492, 0.06252124786376953, 0.062424064636230465, 0.06357913589477539, 0.06316953659057617, 0.06260224151611328, 0.06253875350952148, 0.06283571243286133, 0.06263091278076172, 0.06268825531005859, 0.062740478515625, 0.06261555099487305, 0.06264831924438477, 0.06265155029296875, 0.0625590705871582, 0.06258902359008789, 0.1349620819091797, 0.06290022277832032, 0.06273023986816406, 0.06258278274536133, 0.06256537628173828, 0.06241484832763672, 0.06243644714355469, 0.06253769683837891, 0.06264825439453126, 0.06266265487670898, 0.06263910293579102, 0.0626872329711914, 0.06254899215698242, 0.06250291061401367, 0.06315827178955079, 0.0630835189819336, 0.06262579345703125, 0.0625541114807129, 0.06296575927734376, 0.06356377410888672, 0.06273535919189453, 0.06261248016357422, 0.06268825531005859, 0.0627446403503418, 0.0627281265258789, 0.0624793586730957, 0.06259609603881836, 0.0625162239074707, 0.0626165771484375, 0.06249065780639648, 0.06313776016235352, 0.06331596755981446, 0.06296166229248047, 0.06251827239990235, 0.06264012908935547, 0.062561279296875, 0.0625602569580078, 0.06269747161865234, 0.06257459259033203, 0.06271078491210938, 0.06256038284301758, 0.06276697540283203, 0.06258380889892579, 0.06290636825561523, 0.06270566558837891, 0.06256332778930664, 0.062481407165527345, 0.06269132614135742, 0.0627322883605957, 0.06273023986816406, 0.06274867248535156, 0.06260326385498047, 0.06264934539794922, 0.06341632080078125, 0.06303948974609375, 0.06266265487670898, 0.06273331069946289, 0.06255615997314454, 0.06252851104736327, 0.06263091278076172, 0.06242508697509765, 0.06251929473876953, 0.06264524841308594, 0.13569229125976562, 0.06256332778930664, 0.06257161712646485, 0.06252329635620117, 0.06241689682006836, 0.06251827239990235, 0.0628070411682129, 0.06277529525756836, 0.06263296127319336, 0.06275276947021484, 0.06252441787719727, 0.0628039665222168, 0.06260121536254883, 0.0625541114807129, 0.06245785522460937, 0.0629299201965332, 0.06257561492919922, 0.06257766342163086, 0.06255615997314454, 0.06261145782470703, 0.06267084884643555, 0.062483455657958986, 0.06249574279785156, 0.06257049560546875, 0.0625797119140625, 0.06259097671508788, 0.06254899215698242, 0.06255513763427735, 0.06255923080444335, 0.06248448181152344, 0.06252236938476563, 0.06307942581176758, 0.06321664047241211, 0.06248038482666016, 0.06251110458374023, 0.062478336334228515, 0.06253158569335937, 0.06253260803222656, 0.06252748870849609, 0.06256332778930664, 0.06271283340454102, 0.06266368103027344, 0.06257356643676758, 0.06252032089233399, 0.06264217758178711, 0.0626165771484375, 0.06250291061401367, 0.06261151885986328, 0.06268204879760743, 0.06258483123779297, 0.06261558532714843, 0.06265241622924805, 0.06281110382080078, 0.06251638412475585, 0.06255599975585938, 0.06269132614135742, 0.06260224151611328, 0.06267187118530274, 0.06256332778930664, 0.06258483123779297, 0.06260224151611328, 0.06268211364746094, 0.0627701759338379, 0.13546188354492186, 0.06358937454223633, 0.06271692657470704, 0.0625797119140625, 0.0627845115661621, 0.06304460906982422, 0.06299238586425782, 0.06287257766723633, 0.06339686584472656, 0.06300262451171874, 0.0626688003540039, 0.06257459259033203, 0.06258687973022461, 0.06256435012817382, 0.06278144073486328, 0.06308249664306641, 0.06297600173950195, 0.06380134582519531, 0.06340934371948242, 0.06293689727783203, 0.06308863830566407, 0.06269644927978515, 0.06269747161865234, 0.0626104965209961, 0.06275372695922851, 0.06255308914184571, 0.06283878326416016, 0.06302105712890625, 0.06272204971313476, 0.06288281631469726, 0.06287155151367188, 0.06313369750976562, 0.06276710510253906, 0.06288998413085937, 0.0630978546142578, 0.06291558456420898, 0.0626237449645996, 0.062740478515625, 0.06290636825561523, 0.06285414505004883, 0.06296678543090821, 0.06321459197998047, 0.06271283340454102, 0.06288281631469726, 0.06283161544799805, 0.06350028610229493, 0.06359552001953125, 0.06299955368041993, 0.06314700698852539, 0.06274662399291993, 0.06401229095458984, 0.06346259307861328, 0.06288467025756836, 0.06280294418334961, 0.0630374412536621, 0.0629749755859375, 0.06288383865356445, 0.06269440078735351, 0.06262988662719726, 0.06266777420043945, 0.06251110458374023, 0.06259097671508788, 0.06276505661010742, 0.1356810302734375, 0.0628848648071289, 0.06297600173950195, 0.0627957763671875, 0.06275993728637695, 0.06265651321411132, 0.06285110473632813, 0.06305072021484374, 0.06309273529052735, 0.06269030380249023, 0.06300467300415039, 0.06298419189453125, 0.06283878326416016, 0.06258380889892579, 0.06298009490966797, 0.06349414443969727, 0.06279065704345703, 0.06271692657470704, 0.06281727981567382, 0.06321049499511719, 0.06390067291259766, 0.06274662399291993, 0.0630302734375, 0.0632360954284668, 0.06287462234497071, 0.06285107040405273, 0.06283366394042969, 0.06292787170410157, 0.06309171295166016, 0.06284697723388671, 0.0633908805847168, 0.06323696136474609, 0.06265651321411132, 0.06304665756225586, 0.06274252700805664, 0.06292582321166992, 0.06268108749389649, 0.06265359878540039, 0.06276182556152343, 0.06291046524047851, 0.062814208984375, 0.0628848648071289, 0.06260736083984375, 0.06319718551635742, 0.06301593780517578, 0.06291046524047851, 0.06317772674560547, 0.06293196868896485, 0.06287871932983398, 0.06329446411132812, 0.063172607421875, 0.06317670440673828, 0.06291251373291015, 0.06272000122070312, 0.0629749755859375, 0.06316953659057617, 0.06267801666259766, 0.06295257568359375, 0.06293081665039063, 0.06283468627929688, 0.06287475204467774, 0.06263283157348633, 0.06303846359252929, 0.13538819885253905, 0.06253769683837891, 0.06298009490966797, 0.06262271881103515, 0.06244147109985351, 0.06257459259033203, 0.06248448181152344, 0.062561279296875, 0.06243840026855469, 0.06266060638427734, 0.06255615997314454, 0.0625602569580078, 0.06251827239990235, 0.06253670501708984, 0.06262579345703125, 0.06250905609130859, 0.0625162239074707, 0.06275174331665039, 0.062491649627685546, 0.06252544021606446, 0.06253772735595703, 0.06270156860351563, 0.06256752014160157, 0.062464927673339846, 0.062461952209472656, 0.06248246383666992, 0.06255408096313476, 0.06254387283325195, 0.06258790588378907, 0.06255104064941407, 0.06267903900146485, 0.0632279052734375, 0.06326886367797852, 0.06286649703979492, 0.0625304946899414, 0.0626063346862793, 0.06283059310913086, 0.0628674545288086, 0.06294118499755859, 0.06314803314208985, 0.06268928146362304, 0.06283776092529297, 0.06278348922729492, 0.06264831924438477, 0.06287667083740234, 0.0631541748046875, 0.06334259033203125, 0.06290240097045899, 0.06362099075317383, 0.06321971130371094, 0.06279987335205078, 0.06298316955566406, 0.06268211364746094, 0.06257356643676758, 0.06284492874145507, 0.06280089569091797, 0.06280806350708008, 0.06302105712890625, 0.06263919830322266, 0.06264822387695312, 0.06268415832519532, 0.06266777420043945, 0.06285414505004883, 0.13541477966308593, 0.0627701759338379, 0.06268825531005859, 0.06313779067993164, 0.06299552154541016, 0.06270678329467773, 0.06255599975585938, 0.06287667083740234, 0.06264422225952149, 0.06314086532592773, 0.06269235229492187, 0.06259814453125, 0.06276095962524414, 0.0626063346862793, 0.06283878326416016, 0.06278553771972656, 0.06273334503173827, 0.06341424179077149, 0.06318694305419922, 0.06259199905395507, 0.06277734375, 0.0626984977722168, 0.06281625747680664, 0.06286131286621094, 0.06263097763061523, 0.06298515319824219, 0.06283161544799805, 0.06271283340454102, 0.06281216049194335, 0.06291353607177734, 0.06285619354248047, 0.06308147048950195, 0.06286028671264648, 0.06315520095825196, 0.06290636825561523, 0.0625715217590332, 0.0626319351196289, 0.06263808059692383, 0.06276403045654297, 0.06299955368041993, 0.06254387283325195, 0.06315929412841798, 0.0626698226928711, 0.06250086212158203, 0.06255417633056641, 0.06266566467285156, 0.06308454513549805, 0.06267801666259766, 0.06262271881103515, 0.06260224151611328, 0.0628930549621582, 0.06298419189453125, 0.06266675186157226, 0.06257561492919922, 0.06261043167114258, 0.0626954231262207, 0.06256947326660156, 0.06256742477416992, 0.06259814453125, 0.062483455657958986, 0.06245785522460937, 0.06260739135742187, 0.06276399993896484, 0.13507891845703124, 0.06250188827514648, 0.06251520156860352, 0.06254899215698242, 0.0627691535949707, 0.0625172462463379, 0.06258483123779297, 0.06251110458374023, 0.0627589111328125, 0.06260428619384766, 0.0625797119140625, 0.06265548706054687, 0.06242611312866211, 0.06266265487670898, 0.06277119827270508, 0.06264422225952149, 0.062453758239746096, 0.06272614288330078, 0.06254899215698242, 0.06250700759887695, 0.0625541114807129, 0.06259199905395507, 0.06264422225952149, 0.06261862564086915, 0.06245068740844727, 0.06270054244995117, 0.0625797119140625, 0.06258073425292969, 0.06257868957519531, 0.06262783813476562, 0.06285107040405273, 0.06257664108276367, 0.06262992095947266, 0.06257558441162109, 0.06262886428833007, 0.0626319351196289, 0.0626616325378418, 0.06258892822265626, 0.06262787246704102, 0.06260732650756835, 0.06287871932983398, 0.06262886428833007, 0.06252953720092773, 0.06264422225952149, 0.06258995056152344, 0.06258483123779297, 0.06291046524047851, 0.06279372787475586, 0.06252236938476563, 0.06269241714477539, 0.06262675094604492, 0.06250188827514648, 0.06257356643676758, 0.0625797119140625, 0.06252646255493165, 0.06253158569335937, 0.06251520156860352, 0.06253488159179688, 0.06273411178588867, 0.06257868957519531, 0.0625428466796875, 0.06268108749389649, 0.06285007858276367, 0.1364827880859375, 0.06258995056152344, 0.06280400085449218, 0.06276399993896484, 0.06260736083984375, 0.06255513763427735, 0.06256332778930664, 0.06250495910644531, 0.06252134323120118, 0.06254899215698242, 0.06257766342163086, 0.06255206298828125, 0.06252134323120118, 0.06252851104736327, 0.0626063346862793, 0.06365081787109375, 0.06285721588134766, 0.06258380889892579, 0.06250291061401367, 0.06255308914184571, 0.06253260803222656, 0.06257459259033203, 0.06257049560546875, 0.0626165771484375, 0.06249881744384766, 0.0625459213256836, 0.06259609603881836, 0.0625459213256836, 0.06249676895141602, 0.06287360000610352, 0.06255718231201172, 0.06255923080444335, 0.06253363037109375, 0.06260326385498047, 0.06264934539794922, 0.06255615997314454, 0.06262476730346679, 0.06248857498168945, 0.06258278274536133, 0.06264230346679688, 0.06293900680541992, 0.06301798248291016, 0.06250188827514648, 0.06278656005859375, 0.06277939224243165, 0.06267596817016602, 0.06271078491210938, 0.06272409439086914, 0.06257766342163086, 0.06339788818359375, 0.06294425582885742, 0.06258483123779297, 0.06268415832519532, 0.06261964797973633, 0.06261145782470703, 0.06269747161865234, 0.06256435012817382, 0.06262886428833007, 0.06264217758178711, 0.0625758399963379, 0.06268700790405274, 0.06271078491210938, 0.06260736083984375]",tokens/s,15.674146121300794,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -8160,7 +8160,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45f1-7e81833649f084f75ae3e8bb;6c56cf7b-6d1e-47da-b09d-8ff6fd214d36) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe914-526203057a923f92549cf8c0;56000c78-f918-4482-ba40-d0454eab1202) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8344,7 +8344,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90e6-0818f0421682b34f4074d1a7;169abcd9-97eb-4023-b1ec-42b19018c046) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe293-52eac2fb6b9beabc4f49c479;5aeee17d-e984-40b6-a0cf-508f24fdb1f6) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -8433,7 +8433,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e96ed-2733547050d3a40c0fb4d2d5;8cf77899-f7b8-4aca-a6d0-83f1f01097fa) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe8c3-04ba71447647b0aa72185c5d;b6f37f1d-159e-4ad8-9245-e302230d505d) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8493,7 +8493,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90be-0ea1728f19417b6d105d9589;41da7afb-2257-4bd2-89c2-10883533913f) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe26a-4e63059f3fc176d076f9a76f;a7d94172-b238-4271-9fd4-1455ae00743a) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -8794,7 +8794,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e8f29-34384bf7687492b9384aacaf;604b0e94-beb6-42f7-a8a8-279e02dea638) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe0cc-0ce50cf451be51397a5bd1e3;6a5510fa-d4ae-4d40-a590-09947f84d30f) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -8879,7 +8879,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9677-5967da7d5dc5b2bc5912abf8;49fd7622-b2e3-4601-a71b-c42d4cd99cce) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe844-66c732da738ffed153251327;eed5f341-1fb8-4fef-b962-572e5a654002) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8995,7 +8995,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95a9-6bbac167140322f206cb28e3;0a7f50a9-73e0-4a78-a32b-b9bf10733378) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe778-29a5a73c496e814f6cc78bae;7687c3fb-30fb-420c-ba8c-f731676bf8d8) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9100,7 +9100,7 @@ ImportError: This modeling file requires the following packages that were not fo ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1415.028736,1613.234176,0.0,966.787072,857.647616,s,10,0.6064277153015137,0.06064277153015137,0.0005360740203387311,0.06047083282470703,0.0607635929107666,0.061500197029113766,0.06208948032379151,"[0.06223680114746094, 0.06054751968383789, 0.060445377349853516, 0.06039302444458008, 0.06052438354492187, 0.06041523361206055, 0.06038729476928711, 0.06038188934326172, 0.06059990310668945, 0.060496288299560545]",tokens/s,4221.44294432054,kWh,7.132483155053942e-07,3.908248181578091e-07,3.0210064328674308e-06,4.125079566530634e-06,tokens/kWh,62059409.005607806,MB,1415.323648,1613.234176,0.0,966.787072,888.374272,s,10,32.58626342773437,3.258626342773437,0.022443080457647806,3.2567353515625,3.28173505859375,3.29157529296875,3.29944748046875,"[3.23838671875, 3.249800048828125, 3.2428310546875, 3.27954833984375, 3.278740478515625, 3.263670654296875, 3.23083203125, 3.30141552734375, 3.269190673828125, 3.231847900390625]",tokens/s,19.333299793550527,kWh,3.788478694387389e-05,2.0760127218265596e-05,0.00015591334293673197,0.00021455825709887146,tokens/kWh,293626.5462436559,,s,629,33.078812648773194,0.05258952726355039,0.0071681337276366264,0.05231718444824219,0.05284106216430664,0.0530991096496582,0.11080524047851563,"[0.05068697738647461, 0.051525630950927735, 0.05200588989257812, 0.052103168487548826, 0.052073471069335936, 0.05199052810668945, 0.05201715087890625, 0.05012582397460937, 0.052483070373535154, 0.05346918487548828, 0.052569087982177735, 0.05195673751831055, 0.05133107376098633, 0.051763198852539063, 0.05237145614624023, 0.052596736907958984, 0.051773441314697265, 0.05259366226196289, 0.05216972732543945, 0.052529151916503904, 0.052765697479248044, 0.0515491828918457, 0.04983910369873047, 0.05230284881591797, 0.0522977294921875, 0.05245542526245117, 0.05235302352905274, 0.05288140869140625, 0.05037875366210937, 0.050211841583251954, 0.05019955062866211, 0.050116607666015625, 0.04986470413208008, 0.051664894104003906, 0.052329471588134766, 0.05205196762084961, 0.05258649444580078, 0.052342784881591796, 0.05034086227416992, 0.050369537353515625, 0.05033062362670898, 0.04984115219116211, 0.05033062362670898, 0.05038694381713867, 0.05068288040161133, 0.05038796615600586, 0.05106585693359375, 0.05185126495361328, 0.051507198333740234, 0.051432449340820315, 0.051535873413085936, 0.04977151870727539, 0.0497786865234375, 0.05021593475341797, 0.05030400085449219, 0.04981248092651367, 0.05093273544311523, 0.05149388885498047, 0.051681278228759765, 0.05255782318115235, 0.05168230438232422, 0.04980633544921875, 0.11058585357666016, 0.04979097747802735, 0.049904640197753904, 0.05267865753173828, 0.05265817642211914, 0.050579456329345705, 0.05008486557006836, 0.04976230239868164, 0.04998553466796875, 0.04978483200073242, 0.049890304565429686, 0.049855487823486325, 0.04997119903564453, 0.052187137603759766, 0.05042892837524414, 0.0507852783203125, 0.05044940948486328, 0.050223102569580076, 0.05026406478881836, 0.05023436737060547, 0.050184192657470705, 0.05034905624389648, 0.05034393692016602, 0.05225164794921875, 0.052517887115478515, 0.05013708877563477, 0.050223102569580076, 0.05145907211303711, 0.05279743957519531, 0.05248819351196289, 0.051158016204833984, 0.05059993743896484, 0.05127782440185547, 0.05255782318115235, 0.052775936126708986, 0.052316192626953126, 0.05245539093017578, 0.050318336486816405, 0.05116108703613281, 0.05267660903930664, 0.054093822479248044, 0.05301964950561523, 0.052506622314453126, 0.052410369873046876, 0.05212876892089844, 0.051972095489501956, 0.05219839859008789, 0.05250252914428711, 0.05223628616333008, 0.05237760162353516, 0.053710849761962894, 0.05283942413330078, 0.05223526382446289, 0.0523581428527832, 0.05287628936767578, 0.05238272094726563, 0.05229875183105469, 0.05228646469116211, 0.05243494415283203, 0.05242777633666992, 0.05363916778564453, 0.05273190307617188, 0.052566017150878906, 0.11094220733642578, 0.05019033432006836, 0.050956287384033204, 0.05254553604125976, 0.05244825744628906, 0.05204684829711914, 0.05245439910888672, 0.05242367935180664, 0.05231206512451172, 0.05263564682006836, 0.05255782318115235, 0.0524318733215332, 0.052716545104980465, 0.052664321899414064, 0.05250252914428711, 0.05235507202148437, 0.05257011032104492, 0.05230899047851562, 0.050320384979248046, 0.05015961456298828, 0.050748416900634766, 0.052381694793701174, 0.052342784881591796, 0.05246771240234375, 0.05037055969238281, 0.05060300827026367, 0.05248716735839844, 0.05240217590332031, 0.05265203094482422, 0.050683902740478515, 0.0501288948059082, 0.050293758392333986, 0.050490367889404295, 0.05036441421508789, 0.050032638549804685, 0.05148364639282227, 0.0502845458984375, 0.05041561508178711, 0.050303009033203124, 0.05023433685302734, 0.05023231887817383, 0.050206718444824217, 0.0507586555480957, 0.05252710342407227, 0.052517921447753906, 0.05146723175048828, 0.05027635192871094, 0.050748416900634766, 0.05018009567260742, 0.050216960906982425, 0.051141632080078124, 0.05246361541748047, 0.05256499099731445, 0.05231718444824219, 0.05246566390991211, 0.05212160110473633, 0.05244927978515625, 0.05036646270751953, 0.05015552139282226, 0.05051801681518555, 0.052411392211914064, 0.05209395217895508, 0.05238886260986328, 0.11298611450195313, 0.0523765754699707, 0.05231513595581055, 0.05312307357788086, 0.05240217590332031, 0.052104190826416014, 0.053207038879394535, 0.05267865753173828, 0.05237760162353516, 0.052574207305908206, 0.05244416046142578, 0.052373504638671874, 0.05249331283569336, 0.05246464157104492, 0.050260990142822266, 0.05041971206665039, 0.0502927360534668, 0.05025689697265625, 0.05030604934692383, 0.05012684631347656, 0.05252608108520508, 0.052446208953857425, 0.052754432678222656, 0.05061939239501953, 0.0504453125, 0.050482177734375, 0.05023027038574219, 0.052479999542236325, 0.05250559997558594, 0.052446208953857425, 0.052877311706542966, 0.05242675018310547, 0.052441089630126954, 0.050579456329345705, 0.052395008087158204, 0.052908031463623044, 0.05234790420532227, 0.052558849334716794, 0.05271142578125, 0.05265203094482422, 0.052590591430664066, 0.053103614807128906, 0.05225676727294922, 0.05123072052001953, 0.05249126434326172, 0.05229568099975586, 0.051487743377685545, 0.05287936019897461, 0.05262745666503906, 0.05251583862304687, 0.05263872146606445, 0.05307699203491211, 0.05070438385009766, 0.05263052749633789, 0.052421630859375, 0.052670463562011716, 0.051538944244384766, 0.05237247848510742, 0.052279296875, 0.05085184097290039, 0.05279641723632812, 0.051361793518066405, 0.052721664428710936, 0.1127936019897461, 0.05269606399536133, 0.05232844924926758, 0.05220454406738281, 0.05260697555541992, 0.05229875183105469, 0.052217857360839844, 0.05261414337158203, 0.053215232849121094, 0.052416511535644535, 0.05246771240234375, 0.052311038970947264, 0.05242675018310547, 0.05270016098022461, 0.052348926544189454, 0.05273395156860351, 0.050767871856689455, 0.051264511108398435, 0.05222707366943359, 0.05251379013061523, 0.05266636657714844, 0.05070336151123047, 0.050182144165039064, 0.05242675018310547, 0.052324352264404295, 0.05240217590332031, 0.052327423095703124, 0.05314252853393555, 0.05281792068481445, 0.05065523147583008, 0.050270206451416014, 0.05028966522216797, 0.05028966522216797, 0.05036236953735351, 0.050283519744873044, 0.0518205451965332, 0.05236326217651367, 0.05251686477661133, 0.05257113647460938, 0.052819969177246094, 0.05247795104980469, 0.05360537719726562, 0.05272063827514648, 0.052255744934082034, 0.052201473236083984, 0.052383743286132815, 0.05026508712768555, 0.05025996780395508, 0.05021491241455078, 0.05033267211914062, 0.05140070343017578, 0.053119998931884765, 0.05229568099975586, 0.05256294250488281, 0.05291929626464844, 0.05241856002807617, 0.05234175872802734, 0.050427902221679685, 0.05264179229736328, 0.05267251205444336, 0.053569534301757815, 0.053510143280029294, 0.05232230377197265, 0.11092787170410157, 0.05015961456298828, 0.05025689697265625, 0.05018624114990235, 0.05173452758789063, 0.05220454406738281, 0.052337696075439456, 0.05243593597412109, 0.05226291275024414, 0.052590591430664066, 0.052490238189697266, 0.05240524673461914, 0.05285171127319336, 0.05051801681518555, 0.05058150482177735, 0.051422206878662106, 0.05263052749633789, 0.052741119384765625, 0.052395008087158204, 0.05244518280029297, 0.05238988876342773, 0.05300121688842773, 0.05050572967529297, 0.05062246322631836, 0.05045145416259766, 0.05044224166870117, 0.05186150360107422, 0.05226496124267578, 0.05252505493164063, 0.05279129409790039, 0.05262745666503906, 0.050383872985839843, 0.05048422241210938, 0.050427902221679685, 0.05287321472167969, 0.052603904724121096, 0.052410369873046876, 0.05239807891845703, 0.05242777633666992, 0.05248819351196289, 0.05239603042602539, 0.052393985748291017, 0.050290687561035156, 0.05049753570556641, 0.05035007858276367, 0.05136076736450195, 0.05246156692504883, 0.052350975036621096, 0.053114879608154295, 0.05045555114746094, 0.05034700775146484, 0.05047500610351562, 0.05049753570556641, 0.051848190307617184, 0.0525035514831543, 0.05285273742675781, 0.05243084716796875, 0.05331148910522461, 0.05238784027099609, 0.05285068893432617, 0.05247180938720703, 0.05247078323364258, 0.052446208953857425, 0.11184844970703126, 0.052255744934082034, 0.052190208435058595, 0.05280767822265625, 0.05235200119018555, 0.05249433517456055, 0.05231001663208008, 0.052621311187744144, 0.05276364898681641, 0.05246156692504883, 0.05265919876098633, 0.050348033905029295, 0.05070131301879883, 0.050411518096923826, 0.05027840042114258, 0.05043814468383789, 0.05031628799438476, 0.050299903869628904, 0.050255870819091795, 0.050270206451416014, 0.05240422439575195, 0.051282943725585936, 0.050713600158691405, 0.05041049575805664, 0.050339839935302735, 0.050323455810546876, 0.050283519744873044, 0.05053747177124023, 0.05309132766723633, 0.052590591430664066, 0.05257932662963867, 0.05262335968017578, 0.05248921585083008, 0.052585472106933595, 0.052499454498291014, 0.050531326293945314, 0.05124607849121094, 0.050500606536865236, 0.05043711853027344, 0.05168332672119141, 0.052348926544189454, 0.052261886596679685, 0.05278003311157226, 0.050457599639892575, 0.05321932983398438, 0.05246054458618164, 0.05041254425048828, 0.0507770881652832, 0.050407424926757816, 0.05213798522949219, 0.05028761672973633, 0.050288639068603515, 0.050329601287841794, 0.050293758392333986, 0.050539520263671874, 0.050344959259033206, 0.05049856185913086, 0.050283519744873044, 0.05032755279541016, 0.05024460983276367, 0.05046988677978516, 0.05064089584350586, 0.05021286392211914, 0.11070054626464844, 0.051236862182617186, 0.05164031982421875, 0.0504535026550293, 0.05021388626098633, 0.049906688690185545, 0.04987596893310547, 0.050718719482421876, 0.051402751922607424, 0.05298691177368164, 0.052868064880371095, 0.05224857711791992, 0.05240729522705078, 0.052896766662597655, 0.052563968658447265, 0.052450302124023435, 0.051983360290527345, 0.05299814224243164, 0.05269606399536133, 0.05262335968017578, 0.052147201538085934, 0.05280767822265625, 0.052896766662597655, 0.05286604690551758, 0.05257625579833984, 0.05250559997558594, 0.05259366226196289, 0.05280460739135742, 0.05344255828857422, 0.05301862335205078, 0.05267865753173828, 0.05266124725341797, 0.052847614288330076, 0.05269606399536133, 0.05267763137817383, 0.05209395217895508, 0.052613121032714844, 0.052563968658447265, 0.05275852966308594, 0.05285580825805664, 0.052739070892333983, 0.05248409652709961, 0.053553150177001956, 0.053324798583984374, 0.05259366226196289, 0.0524769287109375, 0.05248102569580078, 0.052759552001953126, 0.05250867080688477, 0.05276774215698242, 0.052209663391113284, 0.053119998931884765, 0.05265919876098633, 0.052313087463378906, 0.05252608108520508, 0.05309542465209961, 0.05207244873046875, 0.05267763137817383, 0.05230899047851562, 0.052552703857421876, 0.05267660903930664, 0.053953536987304686, 0.05299507141113281, 0.11095552062988281, 0.05265817642211914, 0.05082521438598633, 0.05250457763671875, 0.053101566314697264, 0.05249331283569336, 0.05254553604125976, 0.05229363250732422, 0.05200384140014649, 0.0526827507019043, 0.05235609436035156, 0.052596736907958984, 0.0525404167175293, 0.05349785614013672, 0.05272678375244141, 0.05302579116821289, 0.05258342361450195, 0.052636672973632816, 0.05229875183105469, 0.052125694274902344, 0.05250764846801758, 0.052424705505371094, 0.05266739273071289, 0.052673534393310545, 0.052721664428710936, 0.05257011032104492, 0.052569087982177735, 0.052206592559814455, 0.052482048034667966, 0.052787200927734375, 0.05234175872802734, 0.05268479919433594, 0.05274009704589844, 0.052603904724121096, 0.05298995208740234, 0.052951038360595705, 0.05245439910888672, 0.05220454406738281, 0.05257932662963867, 0.05074534225463867, 0.05037977600097656, 0.05015961456298828, 0.05009305572509765, 0.05020569610595703, 0.05031219100952149, 0.050179073333740234, 0.05018316650390625, 0.04992512130737305, 0.05010227203369141, 0.050344959259033206, 0.050164737701416016, 0.05014323043823242, 0.05017497634887695, 0.050383872985839843, 0.052975616455078124, 0.05244723129272461, 0.052585472106933595, 0.05270425415039062, 0.050473983764648435, 0.05030297470092773, 0.0519024658203125, 0.05262335968017578, 0.05269401550292969, 0.11084595489501953, 0.05019647979736328, 0.05005311965942383, 0.05010636901855469, 0.05016371154785156, 0.05063987350463867, 0.05035212707519531, 0.050233345031738284, 0.05042278289794922, 0.05059481430053711, 0.05039308929443359, 0.050361343383789066, 0.050135040283203126, 0.05026303863525391, 0.0501473274230957, 0.05010636901855469, 0.05054873657226563, 0.0503818244934082, 0.050269184112548826, 0.05045145416259766, 0.05013913726806641, 0.0501473274230957, 0.05020467376708984, 0.05027123260498047, 0.05005209732055664, 0.049772544860839846, 0.050026496887207034, 0.05001420974731445, 0.050078720092773435, 0.052018177032470705, 0.052732929229736325, 0.052604927062988284, 0.05247180938720703, 0.052364288330078126, 0.05277798461914063, 0.05299097442626953, 0.052790271759033204, 0.05264588928222656, 0.05258137512207031, 0.05290086364746094, 0.052519935607910156, 0.052245502471923826, 0.05247180938720703, 0.050441215515136716, 0.05024358367919922, 0.05013708877563477, 0.050049022674560545, 0.050336769104003906, 0.05134745788574219, 0.0526295051574707, 0.05250969696044922, 0.05249740982055664, 0.05265919876098633, 0.05255168151855469, 0.05279334259033203, 0.0526561279296875, 0.05274214553833008, 0.05259980773925781, 0.052566017150878906, 0.05260800170898437, 0.0528353271484375, 0.05259468841552734, 0.05210521697998047]",tokens/s,19.015192796629233,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9139,7 +9139,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4666-2943069d50fe47724939fb58;3a5d5c4e-08cc-43a1-b80e-ec60d03ab349) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe98e-5eb8046e1491541074501d34;f63ea8b1-499d-4dbc-a933-dfa57ee777e2) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9210,7 +9210,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9581-5718b4d744d85cb627bbdc29;a7245b02-9df1-4e64-bf8d-e4385a67b15b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe74f-51448b652199aa317a25889e;d53de8fa-9811-464d-914a-9df5beff6fb2) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9281,7 +9281,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9621-2504f957389899fd3d6e9b79;98999aee-1176-4fb5-9d45-707668e14942) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe7f0-22cbf05702821dda2724f8ba;182faf14-65f8-4b95-bfc4-1eddd817395a) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9313,7 +9313,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5028.081664,8173.125632,0.0,7526.678528,6833.744896,s,10,3.26147134399414,0.3261471343994141,0.0004409501692482163,0.3262998352050781,0.3266359405517578,0.32663985443115234,0.326642985534668,"[0.3264150085449219, 0.3254004821777344, 0.326524169921875, 0.32663507080078125, 0.32664376831054687, 0.3256611022949219, 0.32650796508789065, 0.3257033386230469, 0.3257957763671875, 0.32618466186523437]",tokens/s,784.92181288489,kWh,3.8486035547376105e-06,2.1088648073688792e-06,2.01936272659971e-05,2.615109562810359e-05,tokens/kWh,9789264.803302791,MB,5028.081664,8173.125632,0.0,7526.678528,7110.584832,s,10,177.0117890625,17.70117890625,0.0018849237304151378,17.7014580078125,17.703122265625,17.704050390625,17.704792890625,"[17.7017734375, 17.6988359375, 17.701845703125, 17.701814453125, 17.702916015625, 17.704978515625, 17.70087109375, 17.701142578125, 17.69869921875, 17.698912109375]",tokens/s,3.5590849814955385,kWh,0.00020889549900607397,0.00011449215478764955,0.0010727857193390082,0.0013961733731327316,tokens/kWh,45123.3358351769,,s,629,179.6651919555665,0.285636235223476,0.03881150448396202,0.28092312622070315,0.28114656982421876,0.28127252807617187,0.6073619189453126,"[0.2810716247558594, 0.28097842407226564, 0.2807705688476563, 0.28086477661132814, 0.28075418090820314, 0.28107284545898437, 0.28095672607421873, 0.2811484069824219, 0.2811556396484375, 0.2810254821777344, 0.2809948120117188, 0.28088116455078127, 0.28087808227539063, 0.28109619140625, 0.2810828857421875, 0.2810512084960938, 0.28083807373046876, 0.2809333801269531, 0.2807726135253906, 0.2809671630859375, 0.2808340454101563, 0.2806732788085938, 0.28127130126953126, 0.2812733459472656, 0.280880126953125, 0.28080947875976564, 0.28092312622070315, 0.280748046875, 0.2808985595703125, 0.2808719482421875, 0.28096307373046875, 0.28107571411132815, 0.280943603515625, 0.28092312622070315, 0.2808719787597656, 0.28081967163085936, 0.2808186950683594, 0.281025634765625, 0.281087890625, 0.281322509765625, 0.28115252685546877, 0.2808883056640625, 0.28102041625976565, 0.28091290283203124, 0.2809108581542969, 0.2808330383300781, 0.2808821716308594, 0.28100811767578127, 0.2808965148925781, 0.2810798034667969, 0.2807940979003906, 0.28092312622070315, 0.2808330383300781, 0.2809354248046875, 0.28095489501953125, 0.2808739929199219, 0.28079513549804686, 0.2808279113769531, 0.2809620361328125, 0.2811197509765625, 0.280922119140625, 0.28079205322265627, 0.6080081787109375, 0.28076544189453123, 0.280838134765625, 0.28076031494140624, 0.28081253051757815, 0.2807132263183594, 0.28089547729492187, 0.2808545227050781, 0.2807480773925781, 0.2809906921386719, 0.280911865234375, 0.28095590209960936, 0.28086578369140625, 0.28108901977539064, 0.2810091552734375, 0.28119961547851563, 0.2811514892578125, 0.280869873046875, 0.28105319213867186, 0.28088626098632813, 0.28103884887695313, 0.28089959716796875, 0.28075518798828125, 0.28093438720703123, 0.28087295532226564, 0.28100927734375, 0.2809219970703125, 0.28087091064453124, 0.280838134765625, 0.2807982177734375, 0.28106341552734376, 0.2812600402832031, 0.28098046875, 0.28088626098632813, 0.2808821716308594, 0.2809999694824219, 0.2808719177246094, 0.2809323425292969, 0.28095590209960936, 0.28091390991210935, 0.280911865234375, 0.2806722412109375, 0.28075726318359373, 0.28089752197265627, 0.280853515625, 0.28107672119140625, 0.2811269226074219, 0.28113018798828127, 0.28134176635742186, 0.28091290283203124, 0.28086480712890627, 0.2808719177246094, 0.2811617431640625, 0.2807859191894531, 0.2810091552734375, 0.28094259643554687, 0.28108901977539064, 0.2808760375976562, 0.28075418090820314, 0.28086477661132814, 0.2808832092285156, 0.28084735107421877, 0.2807152709960937, 0.6078546142578125, 0.28070706176757815, 0.28101324462890626, 0.2809610290527344, 0.2807173156738281, 0.2806036376953125, 0.2808186950683594, 0.2807275390625, 0.2807459716796875, 0.28081765747070314, 0.28090777587890625, 0.28082278442382813, 0.28080331420898436, 0.2808493957519531, 0.28080435180664065, 0.28077981567382815, 0.2807377624511719, 0.28094259643554687, 0.2809610290527344, 0.28080230712890625, 0.2811473999023438, 0.2808586730957031, 0.2810807800292969, 0.2825359497070313, 0.28085760498046874, 0.28076235961914064, 0.280943603515625, 0.28124774169921873, 0.28127847290039065, 0.28103988647460937, 0.28092620849609373, 0.2809927673339844, 0.28098458862304687, 0.2810777587890625, 0.28099993896484377, 0.28095693969726565, 0.28086578369140625, 0.28130712890625, 0.28111053466796876, 0.2808883056640625, 0.2807859191894531, 0.2808258666992188, 0.2808832092285156, 0.28082073974609373, 0.28086782836914065, 0.28088116455078127, 0.28091595458984375, 0.2825093078613281, 0.2810675048828125, 0.2808412170410156, 0.280953857421875, 0.2808637390136719, 0.28111770629882815, 0.280880126953125, 0.2813460388183594, 0.2811760559082031, 0.28108084106445314, 0.28080953979492185, 0.2807018737792969, 0.28099172973632813, 0.28095281982421877, 0.28108389282226565, 0.28077362060546873, 0.608311279296875, 0.28081048583984375, 0.2809405517578125, 0.2806405029296875, 0.28076031494140624, 0.28073370361328126, 0.28086578369140625, 0.2808135681152344, 0.28088116455078127, 0.280952880859375, 0.28101626586914064, 0.2809241638183594, 0.28168295288085937, 0.2808330383300781, 0.28111053466796876, 0.2809692077636719, 0.2810429382324219, 0.28080947875976564, 0.2810460205078125, 0.28084530639648436, 0.2808330383300781, 0.28081253051757815, 0.2807982177734375, 0.2809302978515625, 0.28086477661132814, 0.2809169921875, 0.2808565673828125, 0.280958984375, 0.2808299560546875, 0.28117913818359375, 0.28106341552734376, 0.28076235961914064, 0.280943603515625, 0.28079000854492187, 0.2809374694824219, 0.2807408752441406, 0.2809620361328125, 0.28110452270507813, 0.28091583251953123, 0.2808842163085937, 0.2811627502441406, 0.2810449829101562, 0.280825927734375, 0.2808196411132812, 0.28327322387695314, 0.28112588500976565, 0.2808965148925781, 0.2809743347167969, 0.2809743347167969, 0.28070501708984374, 0.280880126953125, 0.2807848815917969, 0.28094976806640626, 0.28081253051757815, 0.28092312622070315, 0.28089959716796875, 0.28097125244140625, 0.2808299560546875, 0.28092935180664064, 0.2811114807128906, 0.2811207580566406, 0.2812037048339844, 0.28112896728515624, 0.6081239013671875, 0.28064358520507815, 0.281069580078125, 0.2810849304199219, 0.2808268737792969, 0.2807080993652344, 0.2809927673339844, 0.2810705871582031, 0.2809989013671875, 0.2809968566894531, 0.28100302124023435, 0.2809395141601562, 0.280853515625, 0.2807992248535156, 0.28090777587890625, 0.280985595703125, 0.28104193115234377, 0.2819112854003906, 0.28102859497070315, 0.28089959716796875, 0.28095590209960936, 0.28086782836914065, 0.2809641418457031, 0.2810326843261719, 0.2808555603027344, 0.2808555603027344, 0.2807705688476563, 0.28092312622070315, 0.28085964965820315, 0.2809231567382813, 0.2809599609375, 0.28101223754882815, 0.28098458862304687, 0.280995849609375, 0.28102554321289064, 0.2809692077636719, 0.2812591247558594, 0.28121075439453125, 0.28093646240234377, 0.2810716247558594, 0.28083712768554686, 0.28087808227539063, 0.28104806518554687, 0.2810040283203125, 0.28094873046875, 0.2809456787109375, 0.28094680786132814, 0.28089434814453124, 0.281069580078125, 0.28089447021484376, 0.2810828857421875, 0.2811248779296875, 0.2812682189941406, 0.2808401794433594, 0.28089752197265627, 0.2809354248046875, 0.2809825134277344, 0.28154788208007814, 0.2807090148925781, 0.28145867919921874, 0.28125799560546877, 0.28101119995117185, 0.28075726318359373, 0.607119384765625, 0.28083096313476563, 0.28100607299804686, 0.28079928588867187, 0.28110943603515626, 0.2810675048828125, 0.28095489501953125, 0.2810941467285156, 0.28082891845703123, 0.2809333801269531, 0.28096307373046875, 0.28083917236328126, 0.28091494750976564, 0.28112588500976565, 0.28135220336914063, 0.2809314270019531, 0.2810141906738281, 0.28090060424804686, 0.28100506591796875, 0.28073983764648436, 0.2808986206054688, 0.28088619995117187, 0.2810705871582031, 0.28108697509765623, 0.28090264892578126, 0.28086578369140625, 0.28082891845703123, 0.2808259582519531, 0.2808493041992188, 0.2809292907714844, 0.28081048583984375, 0.28089752197265627, 0.28099172973632813, 0.2809354248046875, 0.28101229858398435, 0.2810070495605469, 0.28085861206054685, 0.28089959716796875, 0.28375860595703123, 0.28100811767578127, 0.2811023254394531, 0.2808770446777344, 0.28098458862304687, 0.2810849304199219, 0.2813665771484375, 0.28089950561523436, 0.28091802978515623, 0.280869873046875, 0.28100607299804686, 0.2809487915039062, 0.28107769775390623, 0.28094772338867186, 0.28119961547851563, 0.2809252014160156, 0.2810787658691406, 0.2808965148925781, 0.28122726440429685, 0.28111770629882815, 0.28107571411132815, 0.2812405700683594, 0.2811535339355469, 0.28105523681640626, 0.28082073974609373, 0.607952880859375, 0.28060671997070313, 0.2809661560058594, 0.2806609802246094, 0.2808463439941406, 0.2806804504394531, 0.28079205322265627, 0.2808760375976562, 0.2808392333984375, 0.28082785034179686, 0.2810624084472656, 0.28129486083984373, 0.28122213745117186, 0.2807715759277344, 0.2808842163085937, 0.28091802978515623, 0.2810572814941406, 0.2808985595703125, 0.28099172973632813, 0.2808565673828125, 0.2809518127441406, 0.28080126953125, 0.28089447021484376, 0.2825502624511719, 0.2808985595703125, 0.28085861206054685, 0.280816650390625, 0.280848388671875, 0.28089447021484376, 0.28097125244140625, 0.28085861206054685, 0.28095590209960936, 0.2809477844238281, 0.2808728942871094, 0.2808985595703125, 0.2808268737792969, 0.28089547729492187, 0.28095077514648437, 0.28085861206054685, 0.2809968566894531, 0.2809774169921875, 0.2807982177734375, 0.28089959716796875, 0.28099172973632813, 0.2807982177734375, 0.2809169921875, 0.281101318359375, 0.28097842407226564, 0.2811545715332031, 0.2810491027832031, 0.28117196655273435, 0.2810101623535156, 0.2810798034667969, 0.280890380859375, 0.2809333801269531, 0.28083917236328126, 0.28120578002929686, 0.28119549560546875, 0.2808186950683594, 0.2811064453125, 0.2810828857421875, 0.2810245056152344, 0.28082073974609373, 0.6070692749023437, 0.2808114624023437, 0.28095693969726565, 0.2807562255859375, 0.28089752197265627, 0.280748046875, 0.2808279113769531, 0.28082891845703123, 0.28091494750976564, 0.28085043334960935, 0.2809415588378906, 0.28091390991210935, 0.28088644409179686, 0.2808758850097656, 0.2809313354492188, 0.2808616943359375, 0.2821949462890625, 0.2810019836425781, 0.28163482666015627, 0.2812569580078125, 0.2809241638183594, 0.281143310546875, 0.2809333801269531, 0.2809456787109375, 0.28096511840820315, 0.2810408935546875, 0.28102041625976565, 0.28100506591796875, 0.28084530639648436, 0.28107571411132815, 0.28093438720703123, 0.28092724609375, 0.280890380859375, 0.28085043334960935, 0.28082073974609373, 0.28086782836914065, 0.280764404296875, 0.28092724609375, 0.280929443359375, 0.2809577941894531, 0.2809302978515625, 0.28088934326171877, 0.28073779296875, 0.28088626098632813, 0.2807859191894531, 0.2808637390136719, 0.28086782836914065, 0.2808883056640625, 0.28105419921875, 0.2811023254394531, 0.2812333984375, 0.28084326171875, 0.2809989013671875, 0.28085147094726565, 0.28113101196289064, 0.2810880126953125, 0.280985595703125, 0.28092620849609373, 0.28081048583984375, 0.2810357666015625, 0.28111666870117186, 0.2810572814941406, 0.2808555603027344, 0.6074562377929688, 0.280658935546875, 0.28088729858398437, 0.28080230712890625, 0.2807451171875, 0.28069366455078126, 0.280728515625, 0.28076031494140624, 0.28095077514648437, 0.2811463623046875, 0.28098458862304687, 0.28073883056640625, 0.2808392333984375, 0.2808042907714844, 0.2809539489746094, 0.28077252197265623, 0.28087808227539063, 0.28191949462890625, 0.28118118286132815, 0.2807510986328125, 0.280943603515625, 0.28087091064453124, 0.2808258666992188, 0.28086782836914065, 0.2809599914550781, 0.28128973388671874, 0.28111566162109375, 0.2807930908203125, 0.2808934326171875, 0.2809313354492188, 0.2811085510253906, 0.28094970703125, 0.280985595703125, 0.2808299560546875, 0.280943603515625, 0.2810040283203125, 0.28105624389648437, 0.28103884887695313, 0.28102859497070315, 0.2809405517578125, 0.280806396484375, 0.28071218872070314, 0.2808770446777344, 0.2809866333007813, 0.280911865234375, 0.28092825317382814, 0.2808463439941406, 0.280869873046875, 0.28091494750976564, 0.2807848815917969, 0.2809037780761719, 0.2807929992675781, 0.2809415588378906, 0.28072653198242187, 0.28112896728515624, 0.2808504638671875, 0.2809825134277344, 0.2808032836914062, 0.2806886291503906, 0.28093954467773435, 0.2810111694335938, 0.28101837158203125, 0.2809968566894531, 0.607963134765625, 0.28109722900390627, 0.28084225463867185, 0.2807828369140625, 0.28078900146484376, 0.28078900146484376, 0.28086383056640624, 0.28084625244140626, 0.2809241638183594, 0.2809098205566406, 0.28088525390625, 0.28086782836914065, 0.28103271484375, 0.28082382202148437, 0.28085964965820315, 0.28094158935546876, 0.2811105041503906, 0.28084225463867185, 0.28108901977539064, 0.2807848815917969, 0.28083917236328126, 0.28080230712890625, 0.28073779296875, 0.28079718017578126, 0.2807562255859375, 0.2808330383300781, 0.2826495971679687, 0.2808985595703125, 0.280853515625, 0.28077566528320314, 0.28073883056640625, 0.2808842163085937, 0.2811269226074219, 0.2812651672363281, 0.28103167724609374, 0.28106341552734376, 0.2807930908203125, 0.2809354248046875, 0.2808637390136719, 0.2806599731445312, 0.28091494750976564, 0.28076446533203125, 0.2810346984863281, 0.2811269226074219, 0.2809743347167969, 0.280764404296875, 0.28090167236328123, 0.280933349609375, 0.2809263000488281, 0.28092303466796875, 0.28106341552734376, 0.28090060424804686, 0.2809794616699219, 0.280806396484375, 0.2810941467285156, 0.28091290283203124, 0.2808883056640625, 0.280875, 0.28072357177734375, 0.28083596801757815, 0.2809948120117188, 0.28083712768554686, 0.28070196533203123]",tokens/s,3.5009563797731067,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9333,9 +9333,9 @@ ChildProcessError: Traceback (most recent call last): return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1736, in generate + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1758, in generate result = self._sample( - File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2375, in _sample + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2397, in _sample outputs = self( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) @@ -9371,7 +9371,7 @@ RuntimeError: shape '[-1, 128, 4672]' is invalid for input of size 21229568 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,,,MB,4753.522688,21498.42944,0.0,20851.982336,20236.731392,s,10,15.791625,1.5791625,0.0006504668109249951,1.579173889160156,1.5798716064453124,1.5800453002929689,1.5801842553710939,"[1.57879541015625, 1.5780565185546875, 1.5798330078125, 1.5797557373046875, 1.580218994140625, 1.57864208984375, 1.5789478759765625, 1.57939990234375, 1.578477294921875, 1.5794981689453125]",tokens/s,162.11124567610997,kWh,1.8641048173109693e-05,1.0215338307862111e-05,9.709499434260027e-05,0.00012595138082357207,tokens/kWh,2032530.3170641307,MB,4754.120704,21498.42944,0.0,20851.982336,20339.707392,s,10,881.8083671875002,88.18083671874999,0.002079438758836812,88.1809140625,88.18330625,88.18352421875001,88.18369859375001,"[88.1769296875, 88.179453125, 88.1832578125, 88.1811796875, 88.18275, 88.178796875, 88.1821796875, 88.1837421875, 88.1794296875, 88.1806484375]",tokens/s,0.7144409414137963,kWh,0.001040938318769137,0.0005705258692486495,0.0053753298280382,0.0069867940160559866,tokens/kWh,9017.011215046985,,s,629,894.6216553955076,1.4222919799610616,0.18759022057657465,1.39966162109375,1.400088818359375,1.4002253417968749,2.978684599609375,"[1.3990267333984374, 1.399194580078125, 1.39926318359375, 1.399057373046875, 1.39915771484375, 1.3993861083984376, 1.3993082275390625, 1.3993277587890625, 1.3993349609375, 1.3990625, 1.3992550048828125, 1.399257080078125, 1.3994239501953125, 1.3993123779296874, 1.3992008056640626, 1.399593017578125, 1.39961962890625, 1.3998203125, 1.3995447998046875, 1.39936865234375, 1.399487548828125, 1.399593994140625, 1.3995611572265625, 1.399668701171875, 1.39974658203125, 1.3998499755859375, 1.399363525390625, 1.39985302734375, 1.399152587890625, 1.399531494140625, 1.39940771484375, 1.399680908203125, 1.39953564453125, 1.399763916015625, 1.3996451416015625, 1.3997568359375, 1.3997650146484375, 1.39991455078125, 1.3995980224609375, 1.3999871826171875, 1.399814208984375, 1.40004248046875, 1.3995816650390625, 1.4001510009765625, 1.39960009765625, 1.400014892578125, 1.4001336669921876, 1.399931884765625, 1.3994854736328124, 1.40003125, 1.3995704345703126, 1.40002197265625, 1.399784423828125, 1.3999564208984374, 1.3997066650390626, 1.400405029296875, 1.3998858642578125, 1.399877685546875, 1.399904296875, 1.3998837890625, 1.3997188720703124, 1.4000732421875, 2.978333740234375, 1.3989600830078126, 1.39924169921875, 1.3988751220703124, 1.399287841796875, 1.3991710205078125, 1.399235595703125, 1.399609375, 1.399541748046875, 1.3992969970703124, 1.39930322265625, 1.399277587890625, 1.39953564453125, 1.399251953125, 1.3993912353515625, 1.399419921875, 1.3995018310546874, 1.3994639892578125, 1.399394287109375, 1.399310302734375, 1.3996768798828125, 1.399478271484375, 1.3996656494140625, 1.3994024658203126, 1.39964208984375, 1.3995662841796874, 1.3995765380859375, 1.39959912109375, 1.399552978515625, 1.3994024658203126, 1.3995858154296874, 1.39936669921875, 1.3997916259765626, 1.3996544189453124, 1.3998448486328126, 1.3994271240234375, 1.3996441650390625, 1.39966162109375, 1.399943115234375, 1.399889892578125, 1.4001827392578126, 1.39974658203125, 1.400034423828125, 1.3994228515625, 1.399857177734375, 1.399530517578125, 1.40224609375, 1.3999288330078126, 1.4000067138671874, 1.400026123046875, 1.400090576171875, 1.3997403564453126, 1.39998828125, 1.3997435302734376, 1.4000753173828124, 1.3996533203125, 1.4001356201171875, 1.39987353515625, 1.4000650634765626, 1.399835693359375, 1.40000048828125, 1.39968408203125, 1.4000137939453126, 2.979962890625, 1.39926220703125, 1.3990328369140625, 1.399584716796875, 1.3994219970703126, 1.3992283935546874, 1.3991546630859375, 1.3993482666015624, 1.3992652587890626, 1.39936669921875, 1.399351318359375, 1.3993861083984376, 1.399299072265625, 1.3993707275390626, 1.3993062744140625, 1.3993953857421875, 1.39961962890625, 1.399415771484375, 1.399525390625, 1.39930517578125, 1.39949365234375, 1.3997568359375, 1.3993184814453126, 1.3997650146484375, 1.3997353515625, 1.399690185546875, 1.4000291748046876, 1.399877685546875, 1.399784423828125, 1.3996329345703125, 1.3998212890625, 1.3994864501953126, 1.4001602783203124, 1.3998140869140625, 1.3999124755859376, 1.3998775634765626, 1.4000035400390625, 1.3994906005859375, 1.399784423828125, 1.39968408203125, 1.3998192138671874, 1.399498779296875, 1.4001407470703124, 1.3995120849609375, 1.3999288330078126, 1.3997454833984375, 1.39981103515625, 1.4027008056640624, 1.3998243408203126, 1.39981103515625, 1.4002513427734375, 1.39980078125, 1.3999759521484374, 1.399794677734375, 1.40002001953125, 1.39987353515625, 1.3999278564453126, 1.399804931640625, 1.400158203125, 1.399793701171875, 1.4000107421875, 1.400045654296875, 1.400123291015625, 2.980041748046875, 1.39945166015625, 1.399203857421875, 1.3994495849609374, 1.3992274169921874, 1.39907177734375, 1.399314453125, 1.399120849609375, 1.39966162109375, 1.399363525390625, 1.3993165283203124, 1.399487548828125, 1.4004613037109375, 1.3995233154296876, 1.399342041015625, 1.399341064453125, 1.3994598388671875, 1.3993349609375, 1.3996513671875, 1.3996400146484376, 1.399520263671875, 1.3995191650390626, 1.3995377197265626, 1.399234619140625, 1.3996871337890624, 1.4000189208984375, 1.3999349365234375, 1.39932470703125, 1.399783447265625, 1.4000445556640626, 1.399647216796875, 1.399568359375, 1.399568359375, 1.399731201171875, 1.399889892578125, 1.39970556640625, 1.39993603515625, 1.3996287841796875, 1.3998807373046875, 1.3998387451171874, 1.399732177734375, 1.3995458984375, 1.3997874755859374, 1.3995592041015625, 1.400785888671875, 1.3997117919921875, 1.3998121337890626, 1.3999656982421875, 1.3996810302734375, 1.3996943359375, 1.39987353515625, 1.399900146484375, 1.4000506591796875, 1.3997332763671875, 1.3998817138671875, 1.3996072998046876, 1.40021044921875, 1.39983251953125, 1.400130615234375, 1.4000435791015624, 1.400005615234375, 1.3997353515625, 1.4001796875, 2.9802998046875, 1.399120849609375, 1.399341064453125, 1.399108642578125, 1.39905126953125, 1.3995653076171874, 1.39922119140625, 1.399256103515625, 1.3993154296875, 1.3994423828125, 1.3993594970703125, 1.399393310546875, 1.3994976806640624, 1.399383056640625, 1.3993502197265626, 1.399436279296875, 1.399673828125, 1.401808837890625, 1.399751708984375, 1.3995008544921874, 1.39957958984375, 1.3996103515625, 1.3996278076171875, 1.3993154296875, 1.39953564453125, 1.399587890625, 1.3997158203125, 1.3994547119140626, 1.399677978515625, 1.39930322265625, 1.3997813720703125, 1.3996236572265626, 1.3997977294921875, 1.3994639892578125, 1.399583740234375, 1.3996553955078126, 1.39995751953125, 1.3994649658203124, 1.3997015380859374, 1.3995806884765625, 1.4000537109375, 1.3995079345703125, 1.40003125, 1.3995447998046875, 1.4000128173828126, 1.3996851806640624, 1.3999759521484374, 1.3998397216796874, 1.40012646484375, 1.399877685546875, 1.400088623046875, 1.3997568359375, 1.4000291748046876, 1.3996800537109375, 1.4001787109375, 1.399815185546875, 1.40000048828125, 1.4016307373046875, 1.400141845703125, 1.4000137939453126, 1.400014892578125, 1.3999698486328125, 1.4001131591796876, 2.978821044921875, 1.399046142578125, 1.3991793212890624, 1.3989376220703125, 1.39928369140625, 1.3996728515625, 1.3994056396484376, 1.399405517578125, 1.3993594970703125, 1.3993656005859374, 1.3992847900390626, 1.3991884765625, 1.3991322021484376, 1.3994342041015626, 1.3993819580078124, 1.3996124267578125, 1.39962060546875, 1.39949365234375, 1.39947216796875, 1.3994957275390625, 1.3995653076171874, 1.399456787109375, 1.3994700927734376, 1.3995662841796874, 1.399562255859375, 1.3996544189453124, 1.399804931640625, 1.39970361328125, 1.3997813720703125, 1.39945068359375, 1.3997384033203124, 1.399419921875, 1.3995765380859375, 1.3994117431640625, 1.400079345703125, 1.3999482421875, 1.399835693359375, 1.399540771484375, 1.399688232421875, 1.399657470703125, 1.3998336181640625, 1.3995345458984374, 1.39985205078125, 1.399656494140625, 1.3999964599609376, 1.39978857421875, 1.3996011962890624, 1.39962060546875, 1.3998602294921876, 1.3998612060546876, 1.4000189208984375, 1.3999698486328125, 1.3998602294921876, 1.399774169921875, 1.40016748046875, 1.399921630859375, 1.4004654541015624, 1.39996875, 1.4002933349609374, 1.3999073486328124, 1.4000814208984376, 1.3998212890625, 1.4002340087890626, 2.97885595703125, 1.3990830078125, 1.3992325439453126, 1.39894482421875, 1.3992386474609375, 1.3991240234375, 1.399192626953125, 1.399552978515625, 1.3998336181640625, 1.3991290283203126, 1.3995643310546875, 1.3993369140625, 1.3992908935546875, 1.3994813232421874, 1.3995396728515626, 1.3994075927734375, 1.399647216796875, 1.3993912353515625, 1.399774169921875, 1.399404541015625, 1.3995263671875, 1.399309326171875, 1.39955810546875, 1.4027039794921874, 1.3995570068359375, 1.3995970458984375, 1.399773193359375, 1.3994434814453125, 1.39992578125, 1.3998408203125, 1.39961962890625, 1.3993533935546876, 1.39991455078125, 1.399616455078125, 1.3995592041015625, 1.3996400146484376, 1.3997864990234374, 1.39940966796875, 1.399609375, 1.39965234375, 1.3998223876953124, 1.39985205078125, 1.4000179443359375, 1.399384033203125, 1.40019921875, 1.3997762451171876, 1.3997117919921875, 1.399884765625, 1.4000496826171875, 1.3995694580078124, 1.4001171875, 1.399783447265625, 1.40018896484375, 1.3996298828125, 1.400130615234375, 1.3999932861328126, 1.400088623046875, 1.4003466796875, 1.400564697265625, 1.3999073486328124, 1.399972900390625, 1.399690185546875, 1.400058837890625, 2.979527587890625, 1.3993922119140625, 1.39902978515625, 1.398983642578125, 1.39919775390625, 1.398992919921875, 1.3992396240234375, 1.3993533935546876, 1.3992376708984375, 1.3990655517578126, 1.399604248046875, 1.3992969970703124, 1.3993492431640624, 1.3994608154296875, 1.39930419921875, 1.3992017822265626, 1.401822265625, 1.39940869140625, 1.3995540771484376, 1.3994659423828124, 1.39966162109375, 1.3993226318359375, 1.3994906005859375, 1.39947216796875, 1.3996871337890624, 1.399563232421875, 1.39978955078125, 1.3996298828125, 1.399773193359375, 1.3996103515625, 1.3997169189453125, 1.3991597900390624, 1.3995765380859375, 1.399605224609375, 1.3999544677734375, 1.399837646484375, 1.3997568359375, 1.3994127197265624, 1.3997230224609376, 1.3997178955078124, 1.400048583984375, 1.399815185546875, 1.4001356201171875, 1.3998172607421875, 1.4001346435546875, 1.3995457763671875, 1.4000435791015624, 1.3997373046875, 1.3998131103515625, 1.3997435302734376, 1.4027857666015624, 1.400364013671875, 1.4001920166015625, 1.39983056640625, 1.4000352783203125, 1.4002769775390624, 1.400194091796875, 1.399773193359375, 1.4002216796875, 1.39990625, 1.4000352783203125, 1.399901123046875, 1.3998612060546876, 2.978325439453125, 1.3991455078125, 1.3992120361328124, 1.3990419921875, 1.3993450927734374, 1.3994271240234375, 1.3989150390625, 1.3993011474609376, 1.3993082275390625, 1.3992078857421875, 1.39926220703125, 1.399310302734375, 1.3993021240234376, 1.399226318359375, 1.3992540283203125, 1.3992283935546874, 1.399677978515625, 1.4018785400390625, 1.399798828125, 1.3992427978515625, 1.3995345458984374, 1.3994066162109375, 1.3994649658203124, 1.39955712890625, 1.3995816650390625, 1.399498779296875, 1.4000189208984375, 1.3997578125, 1.400258544921875, 1.39966162109375, 1.399711669921875, 1.399426025390625, 1.399445556640625, 1.3996800537109375, 1.3996759033203126, 1.3994976806640624, 1.400227783203125, 1.39970458984375, 1.399803955078125, 1.3998212890625, 1.399910400390625, 1.399972900390625, 1.399868408203125, 1.399245849609375, 1.40001171875, 1.3994486083984374, 1.399951416015625, 1.3996124267578125, 1.399826416015625, 1.3997291259765625, 1.400196044921875, 1.3996390380859376, 1.400005615234375, 1.3996943359375, 1.4000947265625, 1.399920654296875, 1.3999810791015626, 1.399846923828125, 1.40007421875, 1.3998878173828124, 1.400089599609375, 1.399841796875, 1.3999400634765624, 2.979295166015625, 1.3990625, 1.399267333984375, 1.39909521484375, 1.399299072265625, 1.3992110595703124, 1.399098388671875, 1.3994393310546875, 1.3993287353515624, 1.39915576171875, 1.39934619140625, 1.3993994140625, 1.399341064453125, 1.399320556640625, 1.3994598388671875, 1.3993870849609376, 1.3995867919921876, 1.399257080078125, 1.3995181884765624, 1.399752685546875, 1.3998980712890625, 1.3995858154296874, 1.39955615234375, 1.3992734375, 1.3995673828125, 1.39959912109375, 1.4034248046875, 1.3997086181640626, 1.39966162109375, 1.3998623046875, 1.399962646484375, 1.3994066162109375, 1.3994639892578125, 1.3996851806640624, 1.399741455078125, 1.3994925537109375, 1.3996871337890624, 1.3995704345703126, 1.40006396484375, 1.39987353515625, 1.399795654296875, 1.3996441650390625, 1.3998203125, 1.3995867919921876, 1.4001162109375, 1.3995396728515626, 1.3999329833984375, 1.3997076416015626, 1.3996749267578126, 1.3995120849609375, 1.3998345947265625, 1.39978955078125, 1.4000609130859376, 1.399593017578125, 1.39999951171875, 1.3997137451171875, 1.400164306640625, 1.4000711669921875, 1.4002811279296874, 1.3999185791015625, 1.399836669921875, 1.3998961181640626, 1.3999913330078124]",tokens/s,0.7030905145280909,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9400,7 +9400,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4a9f-55071c83759c874764b9062d;5fabd602-f57d-47e2-b950-8eb15ba27650) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664feb19-0fe9956b5079556b28cc82d4;f452f0bd-3445-4b82-a975-7d27e5ec5703) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -9485,7 +9485,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e964d-11c1a3c93447acbc0bde4835;edaf7e1a-9900-42bb-9678-07e3be7beb93) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe81c-78374ce56c163e886965c59d;db78d3f0-c8c5-4360-8dc6-c72c1e1c96fd) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9547,7 +9547,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e9096-22cf3beb0e9903a0444ee168;b59e96ce-3912-4975-878b-26ecb76f90ea) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe23e-1e0f86675c8776f53cfb06a6;9178752d-5c96-40df-bd0b-3b581f5ffab3) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -9674,7 +9674,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95fa-7423fa81155ab4197d6dee5c;7baf2efa-12d9-4396-a797-edc8920bea81) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe7c8-2d9a600c090f40756e6111ca;6b39fd15-8050-4a74-8b39-66479a8d85dc) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9747,7 +9747,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e910f-75f1aaa327ab328844cb5db3;39023d13-8b69-499a-9355-095f42f55a09) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe2c1-11a553b62248c9a752be8fba;a5ceb717-59a2-4d29-9f29-2d78ea5c8c34) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9781,7 +9781,7 @@ If this is a private repository, make sure to pass a token having permission to 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1355.640832,1313.34144,0.0,666.894336,583.411712,s,10,0.3753043823242187,0.03753043823242187,0.0017061726371702826,0.03732486343383789,0.038417790222167965,0.04030316772460937,0.0418114697265625,"[0.04218854522705078, 0.03567526245117188, 0.03605964660644531, 0.03759164810180664, 0.03732332611083984, 0.03643900680541992, 0.03732640075683594, 0.037700160980224606, 0.03799881744384766, 0.03700156784057617]",tokens/s,6821.130049551251,kWh,4.375063655553041e-07,2.397299215873419e-07,1.3726420446149002e-06,2.0498783317575464e-06,tokens/kWh,124885460.77781504,MB,1355.640832,1313.34144,0.0,666.894336,644.020224,s,10,23.046878906249997,2.3046878906250003,0.010530589688708434,2.310258544921875,2.3126857177734372,2.313240197753906,2.3136837817382814,"[2.311134521484375, 2.282283935546875, 2.28828759765625, 2.3125625, 2.311003662109375, 2.313794677734375, 2.311868896484375, 2.307242431640625, 2.309513427734375, 2.299187255859375]",tokens/s,27.335588587188372,kWh,2.6983760501499532e-05,1.4787944706284982e-05,7.386716814719235e-05,0.00011563887335497687,tokens/kWh,544799.4966762499,,s,629,23.341892585754387,0.037109527163361515,0.004422100805265259,0.03662847900390625,0.03704197158813476,0.03741410675048828,0.07203065734863284,"[0.035879936218261715, 0.03652710342407227, 0.036673534393310545, 0.03687936019897461, 0.03612268829345703, 0.03798419189453125, 0.03842559814453125, 0.037560382843017576, 0.038442943572998045, 0.036365310668945314, 0.037185535430908204, 0.036550655364990234, 0.03583078384399414, 0.03694694519042969, 0.03779174423217774, 0.037651454925537106, 0.037125118255615236, 0.03663359832763672, 0.03691929626464844, 0.0362608642578125, 0.036383743286132815, 0.03656806564331055, 0.03676160049438477, 0.03685171127319336, 0.03665203094482422, 0.036334590911865236, 0.0366376953125, 0.03631923294067383, 0.036729854583740236, 0.03656294250488281, 0.03668172836303711, 0.03625983810424805, 0.03639603042602539, 0.036574207305908206, 0.036863998413085936, 0.036544513702392575, 0.036560897827148435, 0.036261886596679685, 0.0364288330078125, 0.03603152084350586, 0.036614078521728516, 0.03667967987060547, 0.036686847686767575, 0.03632844924926758, 0.03665407943725586, 0.036160511016845705, 0.03701862335205078, 0.036318206787109376, 0.03648409652709961, 0.03649433517456055, 0.03591987228393555, 0.036713470458984376, 0.03633868789672851, 0.03641856002807617, 0.03654870223999023, 0.036613025665283204, 0.036326400756835936, 0.036843521118164066, 0.036278270721435545, 0.036585472106933595, 0.03630387115478516, 0.03623219299316406, 0.07103282928466798, 0.03485388946533203, 0.0347248649597168, 0.035108863830566404, 0.03509862518310547, 0.03478015899658203, 0.03513241577148438, 0.03518463897705078, 0.03530956649780274, 0.03479449462890625, 0.03498495864868164, 0.035130367279052735, 0.03459481430053711, 0.03486822509765625, 0.03480473709106445, 0.0363059196472168, 0.035458206176757816, 0.03472880172729492, 0.037623809814453124, 0.037416961669921874, 0.036972545623779295, 0.03635200119018555, 0.03666329574584961, 0.03677801513671875, 0.03643286514282226, 0.03608063888549805, 0.03639910507202149, 0.03655379104614258, 0.03633145523071289, 0.03641139221191406, 0.03648409652709961, 0.03645753479003906, 0.03661203384399414, 0.03655372619628906, 0.036703262329101566, 0.03689878463745117, 0.036653057098388675, 0.03629363250732422, 0.0367646713256836, 0.03653222274780273, 0.03689267349243164, 0.03652403259277344, 0.036391937255859375, 0.036729854583740236, 0.036588542938232424, 0.03617996978759765, 0.036708351135253905, 0.03844095993041992, 0.03687833786010742, 0.0367534065246582, 0.03680153656005859, 0.03652710342407227, 0.03646771240234375, 0.03656294250488281, 0.03664896011352539, 0.03679743957519531, 0.03657932662963867, 0.036597824096679686, 0.036764606475830075, 0.03685068893432617, 0.036675582885742186, 0.03703910446166992, 0.03671244812011719, 0.07131033325195313, 0.0348171501159668, 0.034966400146484375, 0.03515084838867188, 0.03504947280883789, 0.035148799896240236, 0.03495116806030273, 0.03530752182006836, 0.03618815994262695, 0.03583180618286133, 0.036429824829101565, 0.036746238708496096, 0.03651686477661133, 0.03638886260986328, 0.03627315139770508, 0.03682099151611328, 0.03655680084228516, 0.03652505493164063, 0.03661209487915039, 0.03641548919677735, 0.036465663909912106, 0.03628851318359375, 0.03663564682006836, 0.03638579177856445, 0.03676774215698242, 0.03656512069702148, 0.0366242561340332, 0.03668377685546875, 0.03650252914428711, 0.036977664947509765, 0.036348926544189454, 0.03617484664916992, 0.0358737907409668, 0.03630499267578125, 0.03663759994506836, 0.036116481781005856, 0.03590655899047852, 0.03667763137817383, 0.036582401275634766, 0.03663564682006836, 0.03637964630126953, 0.03645542526245117, 0.0366110725402832, 0.036580352783203124, 0.03642060852050781, 0.03667251205444336, 0.036450302124023434, 0.036241409301757815, 0.03650559997558594, 0.035737598419189456, 0.03617587280273438, 0.036582401275634766, 0.036708351135253905, 0.03638681411743164, 0.03687014389038086, 0.03681792068481445, 0.037059585571289064, 0.037009471893310546, 0.03683116912841797, 0.03663257598876953, 0.03653836822509766, 0.03680460739135742, 0.03676671981811523, 0.07476121520996094, 0.036734977722167966, 0.03669094467163086, 0.03636019134521484, 0.03656294250488281, 0.03658342361450195, 0.036819969177246094, 0.03642572784423828, 0.036560897827148435, 0.03648819351196289, 0.03648716735839844, 0.036776065826416016, 0.036683647155761716, 0.03653017425537109, 0.03601919937133789, 0.036030494689941406, 0.03646358489990234, 0.03705344009399414, 0.0367564811706543, 0.037028865814208986, 0.03644723129272461, 0.03841535949707031, 0.03753881454467774, 0.03693056106567383, 0.03657932662963867, 0.036241409301757815, 0.036604927062988284, 0.036397056579589845, 0.036435966491699216, 0.03657318496704102, 0.036587520599365236, 0.036606975555419925, 0.03623731231689453, 0.03642367935180664, 0.03631411361694336, 0.036944896697998046, 0.03647999954223633, 0.036850753784179686, 0.036499393463134765, 0.03657011032104492, 0.03669510269165039, 0.03658438491821289, 0.03695001602172852, 0.03663270568847656, 0.03635494232177734, 0.036596736907958984, 0.03721011352539062, 0.03652710342407227, 0.03677798461914063, 0.03775590515136719, 0.037362686157226564, 0.037087230682373046, 0.036805633544921876, 0.036760574340820314, 0.0365865592956543, 0.036950977325439456, 0.03668172836303711, 0.036603904724121096, 0.03699302291870117, 0.036676769256591794, 0.03656585693359375, 0.03657932662963867, 0.03695001602172852, 0.07426457977294922, 0.03659571075439453, 0.03641139221191406, 0.03670943832397461, 0.036706241607666015, 0.03663052749633789, 0.03669401550292969, 0.03667865753173828, 0.036647937774658204, 0.03653222274780273, 0.03715891265869141, 0.03657625579833984, 0.03671142578125, 0.03600588989257812, 0.036898815155029296, 0.036296798706054685, 0.036683681488037106, 0.036462593078613284, 0.03666124725341797, 0.03696844863891602, 0.03650048065185547, 0.03650457763671875, 0.03650559997558594, 0.03646156692504883, 0.03740982437133789, 0.0369837760925293, 0.03661004638671875, 0.03638988876342773, 0.03660595321655274, 0.03637152099609375, 0.0366069107055664, 0.036760574340820314, 0.036805633544921876, 0.03635200119018555, 0.03690496063232422, 0.036939777374267575, 0.036633663177490235, 0.03697657775878906, 0.03657011032104492, 0.03658860778808594, 0.03654137420654297, 0.03666124725341797, 0.037074081420898436, 0.036469600677490235, 0.03672371292114258, 0.03642367935180664, 0.0369172477722168, 0.03655372619628906, 0.03650867080688477, 0.036498432159423826, 0.037064704895019535, 0.036811775207519534, 0.036483070373535154, 0.03680361557006836, 0.03691523361206055, 0.03660486221313477, 0.03713433456420898, 0.0365588493347168, 0.036877311706542966, 0.03642879867553711, 0.036311038970947264, 0.03691929626464844, 0.037266433715820314, 0.07231078338623047, 0.03545600128173828, 0.0361973762512207, 0.036348926544189454, 0.03648614501953125, 0.03696332931518555, 0.03681792068481445, 0.0367534065246582, 0.03653222274780273, 0.03669708633422852, 0.03669305419921875, 0.036628414154052734, 0.03664588928222656, 0.036825088500976565, 0.0366192626953125, 0.03630284881591797, 0.03681075286865235, 0.036746238708496096, 0.03674524688720703, 0.0365700798034668, 0.03599871826171875, 0.036734977722167966, 0.036803585052490234, 0.03670425415039062, 0.036705280303955076, 0.03671654510498047, 0.037028865814208986, 0.03733196640014649, 0.036705280303955076, 0.03640524673461914, 0.03664691162109375, 0.03685273742675781, 0.03743231964111328, 0.03712102508544922, 0.03689574432373047, 0.03719680023193359, 0.03839692687988281, 0.03721420669555664, 0.036805633544921876, 0.036775936126708986, 0.03689267349243164, 0.036429824829101565, 0.03629584121704101, 0.036797279357910155, 0.03661619186401367, 0.03638886260986328, 0.03702783966064453, 0.036763713836669924, 0.03662022399902344, 0.03649228668212891, 0.036782081604003904, 0.03670943832397461, 0.03677484893798828, 0.037125118255615236, 0.036915199279785156, 0.03764534378051758, 0.036789215087890625, 0.03653030395507813, 0.036683647155761716, 0.036675582885742186, 0.03653529739379883, 0.03642060852050781, 0.0367749137878418, 0.07448268890380859, 0.037364734649658206, 0.03663974380493164, 0.036640766143798825, 0.036689918518066404, 0.036332542419433594, 0.036566017150878906, 0.03655475234985352, 0.03644211196899414, 0.03668787384033203, 0.03665206527709961, 0.03657827377319336, 0.03700940704345703, 0.03664486312866211, 0.037817344665527344, 0.036803585052490234, 0.03686707305908203, 0.036827136993408206, 0.03720191955566406, 0.03696640014648438, 0.03661721420288086, 0.0366561279296875, 0.036947967529296875, 0.03726243209838867, 0.036662174224853517, 0.03674726486206055, 0.03662847900390625, 0.03632128143310547, 0.036722686767578124, 0.036751361846923826, 0.0367749137878418, 0.03662847900390625, 0.036789249420166016, 0.03681484985351562, 0.03640729522705078, 0.03652608108520508, 0.036170753479003906, 0.0367534065246582, 0.036201473236083984, 0.03642367935180664, 0.03683430480957031, 0.036587520599365236, 0.03666739273071289, 0.03632844924926758, 0.03636019134521484, 0.0369244155883789, 0.03647180938720703, 0.03628646469116211, 0.03655987167358398, 0.036741119384765625, 0.036441089630126954, 0.03647283172607422, 0.03648614501953125, 0.03655168151855469, 0.03670425415039062, 0.03662335968017578, 0.03659775924682617, 0.03648921585083008, 0.03693056106567383, 0.036877311706542966, 0.03659366226196289, 0.03682918548583984, 0.03728076934814453, 0.07459532928466797, 0.03663872146606445, 0.03683020782470703, 0.036759552001953126, 0.036125694274902344, 0.03650876617431641, 0.036785057067871094, 0.03703705596923828, 0.037082111358642575, 0.0363397102355957, 0.03700838470458984, 0.036347904205322266, 0.03688550567626953, 0.03657011032104492, 0.036395008087158204, 0.03655475234985352, 0.03666124725341797, 0.036416511535644534, 0.036863998413085936, 0.036557823181152346, 0.03657628631591797, 0.036175838470458986, 0.03641139221191406, 0.03650668716430664, 0.03668268966674805, 0.036604927062988284, 0.03665817642211914, 0.03649228668212891, 0.036647937774658204, 0.035955711364746096, 0.036209663391113284, 0.03589324951171875, 0.036765697479248044, 0.03657318496704102, 0.03697049713134765, 0.03668070220947266, 0.036997119903564454, 0.03711795043945312, 0.03787571334838867, 0.037082111358642575, 0.036744190216064454, 0.037059585571289064, 0.0366192626953125, 0.03680255889892578, 0.03654246520996094, 0.03559833526611328, 0.035659774780273434, 0.036397056579589845, 0.03662335968017578, 0.03649228668212891, 0.03666534423828125, 0.03685887908935547, 0.03707392120361328, 0.036880382537841795, 0.036691967010498046, 0.035915775299072264, 0.03681689453125, 0.03593318557739258, 0.03673907089233398, 0.03676160049438477, 0.03672576141357422, 0.03676979064941406, 0.036751361846923826, 0.07499673461914062, 0.03750912094116211, 0.03742105484008789, 0.036563968658447264, 0.036923393249511716, 0.036736000061035154, 0.03690598297119141, 0.03644518280029297, 0.03664281463623047, 0.03659980773925781, 0.03663564682006836, 0.036657150268554685, 0.03683327865600586, 0.036310016632080076, 0.03660083389282227, 0.03654860687255859, 0.0364769287109375, 0.03663872146606445, 0.03670220947265625, 0.03670425415039062, 0.03665817642211914, 0.03661414337158203, 0.036468734741210936, 0.036523006439208985, 0.03629568099975586, 0.03698995208740234, 0.03718860626220703, 0.036592735290527346, 0.036276126861572264, 0.03672780990600586, 0.036506622314453126, 0.03644825744628906, 0.03649945449829101, 0.036478977203369144, 0.036232223510742186, 0.03654038238525391, 0.036674560546875, 0.036603904724121096, 0.036805633544921876, 0.03661312103271484, 0.0366929931640625, 0.03685068893432617, 0.03621478271484375, 0.03703910446166992, 0.03681075286865235, 0.036674560546875, 0.036751361846923826, 0.03650457763671875, 0.03661209487915039, 0.03638988876342773, 0.0363694076538086, 0.036770816802978515, 0.036536319732666016, 0.03663359832763672, 0.03691110229492187, 0.036640766143798825, 0.03666841506958008, 0.036380672454833986, 0.03698995208740234, 0.03657011032104492, 0.0366376953125, 0.03661414337158203, 0.03671039962768555, 0.0741048355102539, 0.03667251205444336, 0.03667660903930664, 0.03663359832763672, 0.036534271240234374, 0.03634483337402344, 0.036893695831298826, 0.036549633026123046, 0.03674009704589844, 0.03668070220947266, 0.03635302352905274, 0.0366561279296875, 0.036705280303955076, 0.036689918518066404, 0.03700121688842774, 0.036915199279785156, 0.03671449661254883, 0.036563968658447264, 0.034473983764648435, 0.03458355331420898, 0.034985984802246094, 0.03528499221801758, 0.03512841415405273, 0.034746273040771485, 0.03503308868408203, 0.03523379135131836, 0.03520409774780273, 0.03676876831054687, 0.036634624481201174, 0.03648614501953125, 0.03609600067138672, 0.03650867080688477, 0.03644416046142578, 0.03657734298706055, 0.03598944091796875, 0.03677798461914063, 0.03691417694091797, 0.03661619186401367, 0.036721664428710936, 0.036603904724121096, 0.03657625579833984, 0.036746238708496096, 0.03682515335083008, 0.03706156921386719, 0.03660083389282227, 0.037147647857666014, 0.0368721923828125, 0.03668172836303711, 0.03661004638671875, 0.03752755355834961, 0.03676979064941406, 0.03687526321411133, 0.03750400161743164, 0.037838848114013675, 0.03719379043579102, 0.03667859268188477, 0.037438465118408204, 0.03693363189697266, 0.03649433517456055, 0.03671551895141602, 0.03680255889892578, 0.03664588928222656, 0.0364400634765625]",tokens/s,26.947257926457944,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,,,MB,4452.151296,23096.459264,0.0,22450.01216,21127.156736,s,10,14.778309326171874,1.4778309326171875,0.002844637726464893,1.4770551147460937,1.4818196166992188,1.482644854736328,1.4833050451660155,"[1.475892822265625, 1.4834700927734374, 1.474785888671875, 1.477040771484375, 1.4770904541015626, 1.475764892578125, 1.4770694580078125, 1.48057080078125, 1.4749879150390626, 1.48163623046875]",tokens/s,173.2268518338785,kWh,1.7427886426448822e-05,9.550359965069218e-06,8.700845849560074e-05,0.00011398670488711878,tokens/kWh,2245875.957669951,MB,4452.151296,23096.459264,0.0,22450.01216,21326.312448,s,10,819.7820859375,81.97820859375,0.002407439463518449,81.9781171875,81.980646875,81.9811671875,81.9815834375,"[81.972390625, 81.9779765625, 81.9774140625, 81.9803828125, 81.9816875, 81.98053125, 81.978328125, 81.978234375, 81.977140625, 81.978]",tokens/s,0.768496910102072,kWh,0.0009677327676614126,0.0005304024149160796,0.004763559699733598,0.0062616948823110904,tokens/kWh,10061.173721187084,,s,629,831.7837823486333,1.3223907509517214,0.17555286511884277,1.3011885986328124,1.3015693115234375,1.301746240234375,2.7772527734375,"[1.300822021484375, 1.3010504150390625, 1.3006878662109376, 1.3009193115234374, 1.300852783203125, 1.3007083740234375, 1.300890625, 1.3009818115234375, 1.3013719482421875, 1.301095458984375, 1.30094384765625, 1.300960205078125, 1.300822021484375, 1.3010032958984374, 1.3008292236328125, 1.301022705078125, 1.30098583984375, 1.3008035888671876, 1.300884521484375, 1.3009510498046875, 1.3010299072265625, 1.300854736328125, 1.30085888671875, 1.3009500732421875, 1.301064697265625, 1.3010103759765625, 1.3010616455078126, 1.3010135498046875, 1.3014302978515624, 1.301043212890625, 1.3013944091796874, 1.301031982421875, 1.30102880859375, 1.3011414794921874, 1.3011148681640625, 1.3013052978515625, 1.3012633056640626, 1.300990966796875, 1.3013411865234374, 1.3011435546875, 1.301085205078125, 1.3012828369140625, 1.30115380859375, 1.3012100830078126, 1.301539794921875, 1.3009417724609376, 1.30134326171875, 1.30136376953125, 1.3012735595703124, 1.3014476318359376, 1.30120703125, 1.3012510986328125, 1.301275634765625, 1.301790771484375, 1.301464111328125, 1.301559326171875, 1.3012930908203124, 1.301392333984375, 1.3017528076171876, 1.301739501953125, 1.3015418701171875, 1.3016575927734375, 2.78485498046875, 1.3009449462890625, 1.3009193115234374, 1.300854736328125, 1.3010831298828125, 1.3009571533203126, 1.3006561279296875, 1.30105859375, 1.30066845703125, 1.30090185546875, 1.301159912109375, 1.3010616455078126, 1.3007667236328124, 1.3009202880859374, 1.30094384765625, 1.3010924072265626, 1.3009111328125, 1.3008302001953125, 1.3007933349609375, 1.3008231201171876, 1.300792236328125, 1.30094287109375, 1.3009132080078125, 1.3007872314453126, 1.300869140625, 1.301265380859375, 1.301338134765625, 1.3013125, 1.3009449462890625, 1.3013892822265625, 1.301149658203125, 1.301179443359375, 1.3008507080078124, 1.3010882568359374, 1.3011435546875, 1.30109130859375, 1.3014149169921876, 1.3013524169921875, 1.3011650390625, 1.301243896484375, 1.3009940185546875, 1.3012838134765625, 1.30125927734375, 1.3011956787109376, 1.301185546875, 1.3012100830078126, 1.305965576171875, 1.301560302734375, 1.3015213623046875, 1.301106689453125, 1.301453857421875, 1.3012919921875, 1.30157373046875, 1.3013677978515625, 1.30126025390625, 1.301613525390625, 1.3015521240234376, 1.3012254638671874, 1.3012889404296875, 1.3014825439453126, 1.3016053466796875, 1.3014722900390625, 1.3013760986328125, 2.77735107421875, 1.3009735107421876, 1.30064697265625, 1.3013421630859374, 1.30083740234375, 1.301045166015625, 1.300828125, 1.3011466064453125, 1.301400634765625, 1.30100537109375, 1.3010565185546874, 1.3010677490234375, 1.30094189453125, 1.3008045654296876, 1.3009930419921876, 1.3008035888671876, 1.300801513671875, 1.3008486328125, 1.3011456298828126, 1.301022705078125, 1.3008670654296874, 1.3008240966796876, 1.300938720703125, 1.302640625, 1.301123046875, 1.30109033203125, 1.3013104248046874, 1.3011885986328124, 1.3009039306640624, 1.301233642578125, 1.301264404296875, 1.30111083984375, 1.30073291015625, 1.300885498046875, 1.3011629638671875, 1.30117431640625, 1.301117919921875, 1.30120703125, 1.3010369873046874, 1.301097412109375, 1.3010616455078126, 1.301327880859375, 1.301392333984375, 1.3012347412109375, 1.301086181640625, 1.30132373046875, 1.301501953125, 1.3046763916015625, 1.3015050048828125, 1.3013463134765626, 1.301180419921875, 1.3013299560546876, 1.30142822265625, 1.3012838134765625, 1.301411865234375, 1.3012796630859376, 1.30123779296875, 1.3011865234375, 1.3015889892578125, 1.3016463623046874, 1.301685302734375, 1.3012254638671874, 1.3016566162109375, 2.77823583984375, 1.3008087158203125, 1.300806640625, 1.300705322265625, 1.300895751953125, 1.3008486328125, 1.3007196044921876, 1.3008138427734375, 1.30092236328125, 1.30109130859375, 1.30102880859375, 1.301179443359375, 1.3025577392578125, 1.300739013671875, 1.3011138916015625, 1.3007125244140625, 1.3010902099609376, 1.3011680908203125, 1.30092236328125, 1.30100537109375, 1.30085791015625, 1.3009244384765626, 1.3010513916015625, 1.3012060546875, 1.3015111083984374, 1.3011845703125, 1.3014527587890625, 1.3011876220703125, 1.3009970703125, 1.301359619140625, 1.301328857421875, 1.3012828369140625, 1.3009879150390624, 1.3010247802734376, 1.3012315673828125, 1.3011578369140624, 1.3013094482421874, 1.301169189453125, 1.3012294921875, 1.3012900390625, 1.30115576171875, 1.3013001708984375, 1.3011610107421876, 1.3011527099609375, 1.3055078125, 1.301476318359375, 1.301550048828125, 1.301517333984375, 1.3014742431640625, 1.301307373046875, 1.3013104248046874, 1.30151318359375, 1.3015111083984374, 1.3015264892578124, 1.301401611328125, 1.3013001708984375, 1.301412841796875, 1.3013453369140624, 1.30142822265625, 1.3013995361328126, 1.3015460205078124, 1.3015255126953125, 1.3015162353515626, 2.778177490234375, 1.3006673583984374, 1.3006939697265625, 1.3005721435546875, 1.300969482421875, 1.3011199951171875, 1.3009674072265625, 1.3007032470703126, 1.300828125, 1.3008507080078124, 1.30096435546875, 1.3010933837890626, 1.3009632568359375, 1.3008660888671875, 1.3010718994140624, 1.3009879150390624, 1.3008035888671876, 1.3034691162109375, 1.301106689453125, 1.3009888916015624, 1.3008782958984375, 1.300886474609375, 1.300864990234375, 1.3010103759765625, 1.3011865234375, 1.30121826171875, 1.3011025390625, 1.301274658203125, 1.3012982177734376, 1.301698486328125, 1.30121826171875, 1.30121630859375, 1.3009090576171876, 1.3011865234375, 1.3013626708984376, 1.3011702880859375, 1.3012960205078126, 1.3011568603515624, 1.3010780029296876, 1.30134326171875, 1.3010616455078126, 1.301508056640625, 1.301390380859375, 1.3011937255859376, 1.3011947021484376, 1.3012296142578126, 1.30364306640625, 1.3014825439453126, 1.3015194091796876, 1.3015162353515626, 1.30155517578125, 1.3015408935546875, 1.301759033203125, 1.3015325927734376, 1.301400634765625, 1.30146923828125, 1.3016771240234375, 1.302951904296875, 1.3014722900390625, 1.3016279296875, 1.3017149658203124, 1.30164013671875, 1.3015224609375, 2.777, 1.300744140625, 1.3008824462890625, 1.3008138427734375, 1.30077587890625, 1.300842529296875, 1.3009520263671874, 1.300822021484375, 1.300927490234375, 1.300853759765625, 1.3014415283203125, 1.30151318359375, 1.301054443359375, 1.30077392578125, 1.3008404541015626, 1.30100634765625, 1.300842529296875, 1.3009100341796875, 1.3009715576171874, 1.30084765625, 1.300958251953125, 1.301233642578125, 1.300926513671875, 1.30098583984375, 1.3013001708984375, 1.30106884765625, 1.3011629638671875, 1.30113330078125, 1.3013309326171876, 1.301391357421875, 1.3012142333984376, 1.3012305908203126, 1.3008731689453126, 1.30104931640625, 1.3010360107421874, 1.3010892333984374, 1.3012510986328125, 1.3012296142578126, 1.3063526611328125, 1.30175390625, 1.3011732177734374, 1.3014620361328124, 1.3012991943359375, 1.3011446533203126, 1.3012264404296876, 1.3013299560546876, 1.3011824951171875, 1.3014456787109374, 1.301423095703125, 1.3016361083984376, 1.3015059814453125, 1.3015654296875, 1.301544921875, 1.3015479736328126, 1.3013831787109376, 1.3015203857421875, 1.3013985595703126, 1.3014793701171874, 1.3015460205078124, 1.3014793701171874, 1.301443603515625, 1.301734375, 1.30144873046875, 2.7785615234375, 1.30075244140625, 1.3009122314453125, 1.3008167724609374, 1.301031982421875, 1.30087939453125, 1.300779052734375, 1.300885498046875, 1.3009141845703125, 1.300916259765625, 1.301181396484375, 1.301142578125, 1.3007769775390625, 1.300885498046875, 1.301053466796875, 1.3011456298828126, 1.3012919921875, 1.300869140625, 1.3009417724609376, 1.300789306640625, 1.3012828369140625, 1.30105859375, 1.3009039306640624, 1.3040076904296876, 1.3013463134765626, 1.3013043212890625, 1.3012510986328125, 1.3011947021484376, 1.3011260986328126, 1.3011292724609376, 1.301284912109375, 1.301201904296875, 1.3010257568359376, 1.30109130859375, 1.3013411865234374, 1.3011353759765625, 1.301443603515625, 1.3012100830078126, 1.301297119140625, 1.3013944091796874, 1.3011395263671874, 1.3012244873046874, 1.301298095703125, 1.3014425048828124, 1.301138427734375, 1.3012275390625, 1.30115380859375, 1.3014599609375, 1.3015889892578125, 1.30134423828125, 1.3014149169921876, 1.3012490234375, 1.301464111328125, 1.3014149169921876, 1.301423095703125, 1.3013770751953124, 1.30163818359375, 1.301612548828125, 1.301464111328125, 1.3013084716796874, 1.3015726318359375, 1.301353515625, 1.3015684814453126, 2.781832275390625, 1.3009500732421875, 1.3008240966796876, 1.301012451171875, 1.3007821044921875, 1.300885498046875, 1.30096533203125, 1.3008824462890625, 1.300654052734375, 1.3009407958984376, 1.30096533203125, 1.3009561767578126, 1.3009090576171876, 1.300969482421875, 1.3010032958984374, 1.3010606689453126, 1.302981689453125, 1.300937744140625, 1.3010401611328124, 1.301053466796875, 1.3011927490234374, 1.3010718994140624, 1.3009766845703126, 1.301338134765625, 1.3011220703125, 1.301360595703125, 1.3014200439453125, 1.300912109375, 1.3011947021484376, 1.3011998291015625, 1.3010667724609375, 1.301264404296875, 1.301211181640625, 1.3012418212890624, 1.3011978759765626, 1.301243896484375, 1.301326904296875, 1.3011978759765626, 1.3012919921875, 1.30178564453125, 1.301286865234375, 1.3013739013671874, 1.301255126953125, 1.301201904296875, 1.301149658203125, 1.3013360595703125, 1.301254150390625, 1.3011046142578124, 1.301750732421875, 1.3015357666015626, 1.3013370361328125, 1.30102783203125, 1.3013729248046875, 1.3013883056640625, 1.3013739013671874, 1.3012767333984374, 1.3015213623046875, 1.3013309326171876, 1.301485595703125, 1.3015357666015626, 1.30161669921875, 1.301986328125, 1.3019422607421876, 2.776386474609375, 1.30092333984375, 1.3011312255859375, 1.3009786376953125, 1.3007410888671875, 1.3013094482421874, 1.30077490234375, 1.3013145751953126, 1.300970458984375, 1.3008353271484374, 1.30104833984375, 1.301074951171875, 1.30100537109375, 1.30090185546875, 1.3010350341796875, 1.301117919921875, 1.3012049560546874, 1.3023876953125, 1.3011998291015625, 1.3009837646484375, 1.30115380859375, 1.3010882568359374, 1.3009541015625, 1.3010042724609374, 1.300970458984375, 1.301360595703125, 1.30148046875, 1.3013258056640624, 1.3009970703125, 1.301338134765625, 1.3016954345703124, 1.3015203857421875, 1.301044189453125, 1.300896728515625, 1.300960205078125, 1.3011158447265625, 1.3012940673828124, 1.3012305908203126, 1.3012520751953125, 1.301458984375, 1.3011732177734374, 1.301149658203125, 1.30106982421875, 1.3015736083984375, 1.3014671630859376, 1.301444580078125, 1.3010584716796876, 1.3011241455078124, 1.3014559326171875, 1.301276611328125, 1.301423095703125, 1.301200927734375, 1.3013197021484375, 1.3013463134765626, 1.3018009033203124, 1.301232666015625, 1.3012919921875, 1.301327880859375, 1.30130224609375, 1.30144970703125, 1.30157470703125, 1.3013277587890626, 1.3014681396484375, 2.7827333984375, 1.3007421875, 1.301612548828125, 1.3009346923828125, 1.300854736328125, 1.3016934814453125, 1.3010504150390625, 1.3013033447265625, 1.301171142578125, 1.30093359375, 1.301065673828125, 1.3010565185546874, 1.3010943603515626, 1.301022705078125, 1.3008394775390626, 1.300715576171875, 1.3006458740234375, 1.301086181640625, 1.3011661376953125, 1.300685791015625, 1.3008701171875, 1.300896728515625, 1.3009193115234374, 1.301139404296875, 1.301137451171875, 1.30161865234375, 1.304037353515625, 1.3010196533203124, 1.3008465576171875, 1.301222412109375, 1.3010892333984374, 1.3011671142578125, 1.301054443359375, 1.301139404296875, 1.3011466064453125, 1.3010994873046875, 1.3012991943359375, 1.3012049560546874, 1.3011732177734374, 1.301221435546875, 1.3011129150390626, 1.301234619140625, 1.3013052978515625, 1.30128076171875, 1.3011927490234374, 1.3010770263671876, 1.30121630859375, 1.301128173828125, 1.3016319580078124, 1.3016514892578126, 1.3016893310546875, 1.3012900390625, 1.3013134765625, 1.3013822021484376, 1.3013883056640625, 1.301622802734375, 1.30130126953125, 1.3013155517578125, 1.30153466796875, 1.301938232421875, 1.30155517578125, 1.3013995361328126, 1.3014169921875]",tokens/s,0.7562061359551271,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-4.5B,facebook/xglm-4.5B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,3047.153664,4364.6976,0.0,3718.250496,3446.424576,s,10,2.1815529632568356,0.2181552963256836,0.0017706188289431455,0.21749915313720702,0.22067105560302736,0.22127264633178712,0.22175391891479493,"[0.22187423706054688, 0.21916972351074218, 0.2163583984375, 0.21842672729492188, 0.21797715759277345, 0.21702114868164063, 0.21676541137695313, 0.216493408203125, 0.22053736877441407, 0.21692938232421874]",tokens/s,1173.4759793217129,kWh,2.562079446535584e-06,1.403905503529847e-06,1.1594441899658822e-05,1.5560426849724254e-05,tokens/kWh,16451990.840118667,MB,3047.153664,4364.6976,0.0,3718.250496,3674.351616,s,10,119.8149814453125,11.98149814453125,0.006852627135568722,11.98012158203125,11.9881150390625,11.99315029296875,11.997178496093749,"[11.998185546875, 11.981759765625, 11.98699609375, 11.9787255859375, 11.9747109375, 11.975310546875, 11.981517578125, 11.974818359375, 11.9780029296875, 11.9849541015625]",tokens/s,5.258107061407448,kWh,0.00014150099731118687,7.755380760533644e-05,0.000639868384825732,0.0008589231897422554,tokens/kWh,73347.65291283492,,s,629,121.58188554382322,0.19329393568175396,0.02592058515520968,0.19005850219726564,0.190696240234375,0.19105648498535155,0.40679104370117186,"[0.18986904907226562, 0.1910292510986328, 0.19020697021484376, 0.1898291473388672, 0.19039741516113282, 0.19016806030273437, 0.1906483154296875, 0.1904742431640625, 0.1901670379638672, 0.1903073272705078, 0.1907947540283203, 0.19089100646972657, 0.19013017272949218, 0.18995199584960937, 0.1898997802734375, 0.190623779296875, 0.1899601593017578, 0.19016499328613282, 0.19018240356445312, 0.1899335632324219, 0.1919129638671875, 0.19052543640136718, 0.19109274291992187, 0.19082240295410155, 0.19045478820800782, 0.19048345947265624, 0.19117465209960938, 0.19038211059570312, 0.19036463928222655, 0.19045068359375, 0.19171839904785157, 0.1905469512939453, 0.1910497283935547, 0.19182591247558595, 0.19049984741210937, 0.19061351013183594, 0.19151872253417968, 0.1904496612548828, 0.19127194213867188, 0.19031654357910155, 0.19062579345703126, 0.1902090301513672, 0.19005850219726564, 0.1901854705810547, 0.19026329040527343, 0.19069952392578124, 0.19007487487792968, 0.18997862243652344, 0.19004928588867187, 0.19016908264160157, 0.19011891174316406, 0.1900728302001953, 0.19013938903808594, 0.1898997802734375, 0.19048243713378907, 0.18986802673339845, 0.19002265930175782, 0.1915115509033203, 0.1902725067138672, 0.1899673614501953, 0.19018240356445312, 0.19001446533203126, 0.40962457275390624, 0.1898219451904297, 0.18988441467285155, 0.1897830352783203, 0.18981272888183592, 0.18987008666992186, 0.1900062713623047, 0.19017625427246093, 0.1898997802734375, 0.1900001220703125, 0.19027456665039064, 0.1896785888671875, 0.19002879333496095, 0.18977792358398438, 0.18995610046386718, 0.1901905975341797, 0.1898997802734375, 0.18986189270019532, 0.18995610046386718, 0.18983628845214845, 0.1908643798828125, 0.18992332458496095, 0.1910609893798828, 0.19052851867675782, 0.18997964477539062, 0.19003904724121093, 0.18978099060058592, 0.1902550964355469, 0.18992025756835937, 0.1900666809082031, 0.18974925231933593, 0.19018342590332032, 0.1899335632324219, 0.19004415893554688, 0.18987213134765624, 0.19002163696289062, 0.18979942321777343, 0.1899673614501953, 0.1898403778076172, 0.19078349304199219, 0.19052134704589843, 0.19099545288085937, 0.19032269287109374, 0.19073741149902343, 0.1903124542236328, 0.19026535034179687, 0.19029913330078124, 0.1919887390136719, 0.19038514709472656, 0.19012403869628905, 0.1904732208251953, 0.19025100708007814, 0.19037901306152344, 0.19070361328125, 0.19044352722167968, 0.19010047912597655, 0.18994688415527344, 0.19050700378417967, 0.19058790588378907, 0.1901957092285156, 0.1900963897705078, 0.19003085327148436, 0.1900349426269531, 0.4065853576660156, 0.19000320434570311, 0.189739013671875, 0.18990284729003906, 0.1898772430419922, 0.19047935485839843, 0.19061964416503907, 0.19013938903808594, 0.19109580993652345, 0.19012197875976564, 0.18996633911132813, 0.19018556213378907, 0.1901444091796875, 0.1901117401123047, 0.18993971252441405, 0.1900421142578125, 0.19017010498046874, 0.1906114501953125, 0.19011378479003907, 0.19067596435546874, 0.18996429443359375, 0.19014247131347656, 0.189844482421875, 0.19014553833007813, 0.18989669799804687, 0.19003187561035156, 0.1898260498046875, 0.19000729370117186, 0.19047219848632813, 0.1906411590576172, 0.1905479736328125, 0.19025613403320313, 0.19160371398925782, 0.191388671875, 0.19063603210449218, 0.19204608154296876, 0.19055616760253907, 0.19106405639648438, 0.19020697021484376, 0.19005746459960937, 0.19004518127441405, 0.18998988342285156, 0.19021311950683595, 0.190160888671875, 0.19023564147949218, 0.1898403778076172, 0.18989773559570314, 0.19060633850097655, 0.1900410919189453, 0.18986189270019532, 0.19028070068359376, 0.18992845153808594, 0.19001651000976563, 0.18993254089355469, 0.19041690063476563, 0.1903626251220703, 0.1898783416748047, 0.1900533142089844, 0.18996121215820314, 0.19023564147949218, 0.18989363098144532, 0.19110911560058594, 0.19087872314453125, 0.40846234130859377, 0.19015168762207033, 0.19006874084472655, 0.18998988342285156, 0.1899304962158203, 0.18986189270019532, 0.18988954162597657, 0.18994688415527344, 0.1900062713623047, 0.19030015563964844, 0.19064218139648437, 0.18984754943847656, 0.19032984924316407, 0.18983833312988282, 0.1903278045654297, 0.1903278045654297, 0.18994586181640624, 0.19019775390625, 0.19005337524414062, 0.19024485778808595, 0.19104051208496095, 0.19011378479003907, 0.18986495971679687, 0.18991615295410155, 0.19060429382324218, 0.18998066711425782, 0.19054591369628907, 0.18988954162597657, 0.18983833312988282, 0.189918212890625, 0.189981689453125, 0.19080601501464844, 0.19011788940429689, 0.18979737854003906, 0.18987110900878906, 0.18992434692382812, 0.1898803253173828, 0.18997042846679688, 0.189918212890625, 0.19009945678710938, 0.18996531677246092, 0.18992845153808594, 0.18997964477539062, 0.1900380096435547, 0.19071487426757813, 0.19046502685546876, 0.19001651000976563, 0.19011993408203126, 0.18998477172851563, 0.1905244140625, 0.18999090576171876, 0.19083673095703124, 0.1906728973388672, 0.19058790588378907, 0.19022950744628905, 0.19023564147949218, 0.19017010498046874, 0.19006361389160156, 0.19003289794921874, 0.18995916748046876, 0.18989260864257812, 0.19008819580078126, 0.1902520294189453, 0.40789605712890625, 0.1898639373779297, 0.19013938903808594, 0.190060546875, 0.18985983276367188, 0.18992230224609374, 0.1899857940673828, 0.18996429443359375, 0.1900943298339844, 0.19123814392089844, 0.19042611694335937, 0.19004722595214843, 0.1900595245361328, 0.1902458953857422, 0.18997145080566405, 0.19007693481445312, 0.18978816223144532, 0.19015679931640625, 0.18976768493652343, 0.18988441467285155, 0.18979942321777343, 0.18995199584960937, 0.18974208068847656, 0.19004928588867187, 0.1898260498046875, 0.19006976318359375, 0.18986495971679687, 0.19006771850585938, 0.189955078125, 0.190202880859375, 0.1897891845703125, 0.19011686706542968, 0.1903278045654297, 0.18998374938964843, 0.18985369873046876, 0.19003085327148436, 0.1897379913330078, 0.1902407684326172, 0.19045068359375, 0.19025100708007814, 0.19082957458496094, 0.19006463623046874, 0.1900912628173828, 0.19036466979980468, 0.18986700439453125, 0.19094937133789064, 0.18996121215820314, 0.1900380096435547, 0.1900789794921875, 0.19005746459960937, 0.18986189270019532, 0.18997862243652344, 0.1900789794921875, 0.18983322143554687, 0.18983322143554687, 0.18982298278808593, 0.18977491760253906, 0.1903492431640625, 0.19014041137695312, 0.18992536926269532, 0.19000729370117186, 0.19089308166503907, 0.19013833618164064, 0.40687103271484376, 0.1902571563720703, 0.1900728302001953, 0.19034112548828125, 0.1899673614501953, 0.18998272705078126, 0.18977894592285155, 0.189844482421875, 0.189876220703125, 0.19003289794921874, 0.18971340942382814, 0.19014247131347656, 0.18985983276367188, 0.18979122924804687, 0.18997042846679688, 0.18992742919921876, 0.18999603271484375, 0.1898455047607422, 0.19000831604003907, 0.18984652709960936, 0.189918212890625, 0.18983526611328125, 0.18977690124511717, 0.19012710571289063, 0.19007180786132813, 0.18988134765625, 0.19060838317871093, 0.19000933837890624, 0.1900595245361328, 0.18997964477539062, 0.1899683837890625, 0.19029093933105468, 0.19010662841796874, 0.18992332458496095, 0.18989260864257812, 0.18987930297851563, 0.1898455047607422, 0.1898751983642578, 0.1901475830078125, 0.18995301818847657, 0.18987930297851563, 0.18991719055175782, 0.18989363098144532, 0.19003289794921874, 0.19037799072265624, 0.19000729370117186, 0.19007487487792968, 0.19001242065429688, 0.19004928588867187, 0.18987110900878906, 0.19108659362792968, 0.19001548767089843, 0.1901957092285156, 0.18998681640625, 0.190455810546875, 0.1903697967529297, 0.1900380096435547, 0.19015679931640625, 0.19044761657714843, 0.19040357971191407, 0.19003596496582031, 0.19056332397460937, 0.19037184143066407, 0.4088299560546875, 0.19015782165527345, 0.19044557189941405, 0.19055410766601563, 0.190603271484375, 0.19021311950683595, 0.19014143371582032, 0.19054899597167968, 0.19041587829589843, 0.19031449890136717, 0.19022950744628905, 0.19018138122558595, 0.18990899658203125, 0.18977690124511717, 0.19039846801757812, 0.19112037658691405, 0.19013017272949218, 0.19058277893066405, 0.1902213134765625, 0.19103538513183593, 0.19020083618164063, 0.18990284729003906, 0.18996018981933593, 0.1905786895751953, 0.1900369873046875, 0.18989260864257812, 0.19007795715332032, 0.18976460266113282, 0.18982298278808593, 0.19013017272949218, 0.18985881042480468, 0.1900052490234375, 0.18992845153808594, 0.19005439758300782, 0.1903697967529297, 0.190171142578125, 0.19014349365234376, 0.19013119506835938, 0.1901107177734375, 0.19025920104980468, 0.18981785583496094, 0.18988236999511718, 0.1902725067138672, 0.1909381103515625, 0.18992950439453124, 0.1899437713623047, 0.18989260864257812, 0.19019673156738282, 0.18974822998046875, 0.1900185546875, 0.190055419921875, 0.1899735107421875, 0.18993766784667968, 0.1899304962158203, 0.1897830352783203, 0.19005133056640625, 0.18983013916015626, 0.19001446533203126, 0.18978201293945313, 0.18987008666992186, 0.18980659484863283, 0.19300352478027344, 0.1902725067138672, 0.40633856201171875, 0.18969190979003905, 0.18986087036132812, 0.19072511291503907, 0.19017727661132813, 0.190708740234375, 0.19001957702636718, 0.18996018981933593, 0.18989363098144532, 0.18967654418945312, 0.18981478881835936, 0.18968678283691406, 0.18957005310058594, 0.18973388671875, 0.1896785888671875, 0.18996940612792967, 0.18995712280273438, 0.19016806030273437, 0.1897891845703125, 0.19005235290527345, 0.18977792358398438, 0.19006259155273436, 0.18984652709960936, 0.1901598663330078, 0.18983322143554687, 0.19014247131347656, 0.1898260498046875, 0.19079782104492188, 0.1899356231689453, 0.19002163696289062, 0.18965811157226561, 0.1905674285888672, 0.19002879333496095, 0.19021005249023437, 0.1899366455078125, 0.190060546875, 0.19000218200683594, 0.19011891174316406, 0.1900482635498047, 0.19024281311035157, 0.18998681640625, 0.19006976318359375, 0.189918212890625, 0.19010662841796874, 0.19004415893554688, 0.19005850219726564, 0.18998783874511718, 0.18996018981933593, 0.19013325500488282, 0.19008511352539062, 0.18990386962890626, 0.18996940612792967, 0.19016294860839844, 0.19002061462402345, 0.19008204650878907, 0.1907578887939453, 0.19060633850097655, 0.19014349365234376, 0.19043122863769532, 0.19075584411621094, 0.19031552124023438, 0.1906524200439453, 0.19047731018066405, 0.4105738220214844, 0.18998988342285156, 0.1901782989501953, 0.19021311950683595, 0.19003187561035156, 0.19004928588867187, 0.18993971252441405, 0.19008409118652345, 0.18992230224609374, 0.1899745330810547, 0.18970419311523437, 0.19005133056640625, 0.18988954162597657, 0.1901598663330078, 0.19045068359375, 0.1913825225830078, 0.19023155212402343, 0.19000933837890624, 0.1899735107421875, 0.18993971252441405, 0.19123712158203124, 0.1901670379638672, 0.1900369873046875, 0.1900247039794922, 0.1898629150390625, 0.18995301818847657, 0.18981484985351563, 0.19017208862304688, 0.1898321990966797, 0.1900298309326172, 0.18986904907226562, 0.19023667907714845, 0.189949951171875, 0.19010354614257813, 0.18994483947753907, 0.18997247314453125, 0.18983628845214845, 0.19002572631835937, 0.1898424377441406, 0.19020390319824218, 0.18978611755371094, 0.1900380096435547, 0.1898567657470703, 0.19012197875976564, 0.19136921691894532, 0.19009536743164063, 0.1905745849609375, 0.19010047912597655, 0.19017216491699218, 0.18993458557128906, 0.18992025756835937, 0.18997247314453125, 0.1900349426269531, 0.18994073486328125, 0.18992236328125, 0.18995500183105468, 0.18991410827636718, 0.18988954162597657, 0.1898751983642578, 0.18996429443359375, 0.18981068420410158, 0.18988543701171876, 0.19000320434570311, 0.41045709228515626, 0.1903831024169922, 0.19026535034179687, 0.19064627075195312, 0.1904117736816406, 0.19001548767089843, 0.18999603271484375, 0.19001446533203126, 0.19034828186035158, 0.19011378479003907, 0.1900175323486328, 0.19005746459960937, 0.18993869018554688, 0.18976358032226562, 0.18989260864257812, 0.1899632568359375, 0.18984754943847656, 0.18984754943847656, 0.1897410583496094, 0.18984141540527344, 0.18965402221679686, 0.18991104125976563, 0.18997247314453125, 0.1900349426269531, 0.18976870727539064, 0.18987008666992186, 0.1900369873046875, 0.19005850219726564, 0.1898219451904297, 0.19005337524414062, 0.1898076171875, 0.19010047912597655, 0.19082546997070313, 0.1910118408203125, 0.19069541931152345, 0.19099545288085937, 0.19047731018066405, 0.19019366455078124, 0.1900912628173828, 0.19030528259277343, 0.19032371520996094, 0.19045376586914062, 0.1901465606689453, 0.19027046203613282, 0.19049574279785156, 0.19036058044433593, 0.19018853759765625, 0.19002265930175782, 0.19018342590332032, 0.19001651000976563, 0.18995199584960937, 0.1901957092285156, 0.19033599853515626, 0.19044248962402344, 0.19031449890136717, 0.190413818359375, 0.1907896270751953, 0.19037593078613282, 0.19011788940429689, 0.1903073272705078, 0.19021209716796875, 0.19087770080566407, 0.19055923461914062]",tokens/s,5.173468047370279,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9820,7 +9820,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4618-278eafba595634dc08dbc87c;8ec471fc-e1df-4ae1-adc4-cc20ddc5bbb2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe93b-1df92d4e18ae163262bd4cc5;e74b20fb-af96-4ffb-9ffc-d0d4bf757626) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9851,7 +9851,7 @@ OSError: B is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -9871,9 +9871,9 @@ ChildProcessError: Traceback (most recent call last): return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1736, in generate + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1758, in generate result = self._sample( - File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2375, in _sample + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2397, in _sample outputs = self( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) @@ -9948,7 +9948,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9439-4f734dfd3e9f774008441c18;8d16cac6-3717-4253-978e-fc9f593a015e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe5fc-30b127d3488d7b63328a2f5d;9fbfc5ff-d36b-428e-b7c4-d05990b8a5a1) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10142,7 +10142,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e93e9-7fd082bd1ca4f04f0f85fdf1;9ba88e99-472c-48d9-9314-8a0c1d4f8d80) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe5ae-214278457c4b3f1f3d7aaccc;42376c3b-dc01-4664-8cac-925eb5ed3955) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10173,7 +10173,7 @@ OSError: m is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10212,7 +10212,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46dc-0574d69847aedc013c7d67fe;101b824b-5fe9-4216-abb3-73aa3b5c249f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fea10-0c543894211f017e25943bdb;0ea9011d-a744-4d15-9612-3f8e84e9e988) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10243,7 +10243,7 @@ OSError: 1 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10282,7 +10282,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d468e-589ee7e43b85fad33f4db78a;bfc33719-09f6-410a-84e4-0d2b2eec5306) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe9b7-5b2bc44729ecca3936bae10f;dad6c1ab-6269-4a00-bfad-9bfa8c58d939) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10314,7 +10314,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,2111.238144,5989.9904,0.0,5343.543296,5028.4416,s,10,3.233366729736328,0.3233366729736328,0.0033406601214661987,0.3222906799316406,0.32357587890625,0.3284577087402344,0.3323631726074219,"[0.3333395385742188, 0.32179965209960937, 0.32208477783203127, 0.3221396789550781, 0.32241891479492185, 0.32226861572265625, 0.32247674560546874, 0.3220350341796875, 0.32249102783203126, 0.322312744140625]",tokens/s,791.7443995623598,kWh,3.809069366090828e-06,2.0868637380544895e-06,1.9419850605315037e-05,2.5315783709460352e-05,tokens/kWh,10112268.41475717,MB,2111.238144,5989.9904,0.0,5343.543296,5239.773184,s,10,177.759330078125,17.7759330078125,0.00539873067631103,17.776914062499998,17.780858398437502,17.78267333984375,17.78412529296875,"[17.78448828125, 17.780142578125, 17.780455078125, 17.77858203125, 17.77524609375, 17.77921875, 17.772978515625, 17.7653359375, 17.77097265625, 17.77191015625]",tokens/s,3.544117767113072,kWh,0.0002097472764841385,0.00011495934078383015,0.0010507334221832726,0.0013754400394512412,tokens/kWh,45803.52337651526,,s,629,180.37106585693388,0.28675845128288324,0.03830342544136477,0.2821273498535156,0.28249927368164063,0.28266066284179686,0.6044239672851563,"[0.28208843994140625, 0.2816143493652344, 0.28174029541015627, 0.2818191223144531, 0.281841796875, 0.28234329223632815, 0.28247866821289064, 0.28213650512695315, 0.2821478271484375, 0.28248574829101564, 0.28239974975585935, 0.28189089965820313, 0.2819306640625, 0.2821160888671875, 0.2824786071777344, 0.2824530029296875, 0.28217037963867186, 0.28215603637695313, 0.2821754760742187, 0.2821601257324219, 0.28203213500976565, 0.28210791015625, 0.2820915222167969, 0.2822574157714844, 0.2826567687988281, 0.28246527099609375, 0.28229937744140626, 0.2824048767089844, 0.2822717590332031, 0.28209765625, 0.282181640625, 0.2824407043457031, 0.2823282470703125, 0.28210977172851565, 0.282134521484375, 0.28219802856445314, 0.2821457824707031, 0.2820218811035156, 0.2823475341796875, 0.28210891723632814, 0.28254104614257813, 0.2825953369140625, 0.28246432495117185, 0.282409912109375, 0.28244174194335936, 0.2821725158691406, 0.28213546752929686, 0.2819011535644531, 0.2821539001464844, 0.28217568969726564, 0.28210360717773436, 0.2820946044921875, 0.28238848876953127, 0.2819573669433594, 0.28211404418945313, 0.2819933166503906, 0.28195733642578125, 0.2820904235839844, 0.2821160888671875, 0.2821734313964844, 0.28238027954101563, 0.28221337890625, 0.6041272583007813, 0.28180377197265627, 0.28192050170898436, 0.28203109741210936, 0.2822051696777344, 0.28212326049804687, 0.2817904663085937, 0.2819010620117188, 0.28161639404296873, 0.28195944213867186, 0.2817013549804688, 0.28198095703125, 0.2819296875, 0.282176513671875, 0.2822379455566406, 0.2823495788574219, 0.282250244140625, 0.28226763916015624, 0.2821877746582031, 0.28203826904296875, 0.2818416748046875, 0.2819532775878906, 0.28200244140625, 0.28207513427734376, 0.282038330078125, 0.28255328369140625, 0.2826588134765625, 0.28228302001953126, 0.2822430725097656, 0.2822901611328125, 0.28207513427734376, 0.2827202453613281, 0.28219091796875, 0.2818016662597656, 0.28206491088867186, 0.2818314208984375, 0.2819798889160156, 0.2825041809082031, 0.2824970092773437, 0.28281240844726563, 0.2824867858886719, 0.2822584228515625, 0.2821826477050781, 0.28282675170898436, 0.2823198852539063, 0.282239990234375, 0.2824007568359375, 0.28243557739257813, 0.28269363403320313, 0.28250933837890624, 0.28244680786132814, 0.2821959533691406, 0.28258917236328124, 0.2825420837402344, 0.28256768798828125, 0.28236703491210935, 0.2824303894042969, 0.28212542724609374, 0.2822542114257813, 0.2821201782226562, 0.28227789306640627, 0.28234445190429686, 0.2823323669433594, 0.6044700927734376, 0.2817679443359375, 0.28177407836914065, 0.2822266845703125, 0.28175668334960935, 0.28163482666015627, 0.2817423400878906, 0.28189901733398437, 0.28187954711914065, 0.28172097778320315, 0.28181591796875, 0.2820700073242188, 0.28192666625976565, 0.2816399230957031, 0.2817362060546875, 0.28181298828125, 0.2816307067871094, 0.28175668334960935, 0.28156927490234374, 0.28206500244140625, 0.282412109375, 0.2820504150390625, 0.2818754577636719, 0.28281857299804686, 0.28224615478515624, 0.2822686767578125, 0.2820690002441406, 0.28191845703125, 0.2822512512207031, 0.28209869384765623, 0.2820894775390625, 0.2821273498535156, 0.2821611633300781, 0.28232498168945314, 0.2822451171875, 0.2823833618164063, 0.28242840576171874, 0.28225433349609375, 0.282576904296875, 0.28276327514648436, 0.282498046875, 0.2824468383789063, 0.28238027954101563, 0.2824765319824219, 0.2825287780761719, 0.28259634399414063, 0.28277658081054685, 0.2829609069824219, 0.28298138427734376, 0.28278271484375, 0.28267111206054685, 0.2823157653808594, 0.28224615478515624, 0.28225637817382815, 0.2821846923828125, 0.2825994262695313, 0.2826752014160156, 0.2823681640625, 0.2823616943359375, 0.2828748779296875, 0.282387451171875, 0.28251852416992185, 0.28242431640625, 0.604927001953125, 0.2821109619140625, 0.28196966552734376, 0.28192666625976565, 0.28178140258789064, 0.2820749816894531, 0.28208743286132815, 0.2823157653808594, 0.2820751647949219, 0.2824621887207031, 0.2823690185546875, 0.2819020690917969, 0.2826956787109375, 0.28207821655273435, 0.2817505187988281, 0.28206796264648437, 0.2821632080078125, 0.282365966796875, 0.2828114013671875, 0.28293734741210935, 0.28228521728515626, 0.2823125305175781, 0.28213861083984376, 0.282239990234375, 0.2822051696777344, 0.28221234130859374, 0.2824673156738281, 0.2822973937988281, 0.2820873718261719, 0.28209356689453124, 0.28235675048828124, 0.28199627685546874, 0.2819942321777344, 0.2818836975097656, 0.2820146484375, 0.28175668334960935, 0.28190618896484376, 0.28169830322265627, 0.2817607727050781, 0.2820966491699219, 0.28200244140625, 0.2818232421875, 0.28187750244140625, 0.2817146911621094, 0.2823198852539063, 0.2825646057128906, 0.2824007568359375, 0.28234649658203126, 0.282166259765625, 0.28203314208984376, 0.28210687255859374, 0.2821058654785156, 0.28243045043945314, 0.2823445129394531, 0.28234454345703125, 0.28226544189453123, 0.28244174194335936, 0.28198910522460935, 0.2824427490234375, 0.2826649475097656, 0.2825635986328125, 0.28235162353515625, 0.28271615600585936, 0.6050764770507813, 0.2824765319824219, 0.282239990234375, 0.28196453857421877, 0.2820085754394531, 0.2821048278808594, 0.28219186401367186, 0.2823638916015625, 0.2818447265625, 0.2817147216796875, 0.28185186767578124, 0.2819164123535156, 0.2817607727050781, 0.281776123046875, 0.28234136962890627, 0.282102783203125, 0.282218505859375, 0.2824847412109375, 0.2821795959472656, 0.2820413513183594, 0.28206491088867186, 0.2821734619140625, 0.2823700256347656, 0.2823157653808594, 0.2824038391113281, 0.28242031860351563, 0.28219381713867187, 0.28210089111328124, 0.2820626831054687, 0.2819921875, 0.28228402709960937, 0.2822102966308594, 0.2819850158691406, 0.28219290161132815, 0.28233319091796877, 0.28209356689453124, 0.28235467529296876, 0.2823075866699219, 0.28208843994140625, 0.28219390869140626, 0.2819031066894531, 0.2821048278808594, 0.2821119995117187, 0.28196148681640626, 0.2821611633300781, 0.28222055053710937, 0.28210891723632814, 0.282287109375, 0.28205056762695313, 0.28214886474609374, 0.28230450439453125, 0.28225637817382815, 0.2821191711425781, 0.2820997009277344, 0.28205978393554687, 0.2820423583984375, 0.28210891723632814, 0.28235162353515625, 0.28193280029296874, 0.2820771789550781, 0.28224920654296876, 0.28226763916015624, 0.28205157470703124, 0.6045245361328125, 0.2818734130859375, 0.28171270751953126, 0.2818794860839844, 0.28213043212890626, 0.2820054931640625, 0.2818191223144531, 0.28199014282226564, 0.282397705078125, 0.2819779052734375, 0.2819819030761719, 0.2818518981933594, 0.28193280029296874, 0.28213253784179687, 0.2824437255859375, 0.28258203125, 0.2823127136230469, 0.2826015625, 0.28251434326171876, 0.2822348937988281, 0.2819430541992187, 0.2821242980957031, 0.2821734313964844, 0.2825594787597656, 0.282144775390625, 0.28232498168945314, 0.28224819946289065, 0.28221746826171873, 0.28216217041015623, 0.28231475830078123, 0.28227685546875, 0.28261376953125, 0.28248779296875, 0.28229428100585935, 0.28233114624023437, 0.2821038208007812, 0.28219610595703126, 0.2821446533203125, 0.2821191711425781, 0.2823833618164063, 0.282498046875, 0.28261376953125, 0.2821549987792969, 0.28219903564453125, 0.2824509582519531, 0.2823638916015625, 0.28233319091796877, 0.28229632568359375, 0.28253387451171874, 0.2824120178222656, 0.28219290161132815, 0.2819850158691406, 0.2822574157714844, 0.2821622619628906, 0.28178012084960935, 0.2820464782714844, 0.282387451171875, 0.2820341796875, 0.28221234130859374, 0.28221951293945313, 0.28213150024414063, 0.28217544555664065, 0.28202410888671875, 0.6049083862304687, 0.2820301513671875, 0.28173715209960937, 0.28217242431640627, 0.28183551025390624, 0.2818027648925781, 0.2818345642089844, 0.28181396484375, 0.28173721313476563, 0.2817689514160156, 0.28174542236328126, 0.2820126647949219, 0.281923583984375, 0.28178738403320314, 0.28179559326171877, 0.2828103637695312, 0.28205874633789063, 0.2821672973632813, 0.281965576171875, 0.28202597045898437, 0.28192462158203124, 0.28203826904296875, 0.28191232299804686, 0.28244479370117187, 0.28209765625, 0.2819430541992187, 0.28198196411132814, 0.2820720520019531, 0.282029052734375, 0.2821376037597656, 0.28201983642578127, 0.2819522705078125, 0.281923583984375, 0.28191436767578126, 0.28214990234375, 0.2819594116210937, 0.28215090942382814, 0.28206491088867186, 0.282461181640625, 0.28227789306640627, 0.28218572998046876, 0.28224203491210936, 0.28214886474609374, 0.2820700073242188, 0.28204544067382814, 0.28215603637695313, 0.28231475830078123, 0.28252056884765625, 0.2822738037109375, 0.28217037963867186, 0.2826618957519531, 0.28251544189453126, 0.28231781005859374, 0.28235467529296876, 0.2827120666503906, 0.28251751708984374, 0.28266702270507815, 0.28215704345703124, 0.28202496337890626, 0.28209869384765623, 0.2819358825683594, 0.28205670166015623, 0.281975830078125, 0.6043053588867188, 0.28190411376953123, 0.28174951171875, 0.28164608764648436, 0.28167678833007814, 0.2816675720214844, 0.28167474365234374, 0.2820045166015625, 0.281871337890625, 0.281625732421875, 0.28212109375, 0.28176809692382815, 0.2816162109375, 0.281734130859375, 0.28169625854492186, 0.28201165771484377, 0.28182119750976564, 0.281744384765625, 0.2821959533691406, 0.281923583984375, 0.28191845703125, 0.2822645874023437, 0.2820546569824219, 0.281965576171875, 0.2817331237792969, 0.28167373657226563, 0.28183450317382813, 0.28208026123046875, 0.28205056762695313, 0.281881591796875, 0.28187136840820315, 0.28195431518554687, 0.2817976379394531, 0.28180889892578126, 0.2819501953125, 0.2818437194824219, 0.28233932495117187, 0.28224716186523435, 0.282144775390625, 0.2821181335449219, 0.28202496337890626, 0.2820997009277344, 0.282092529296875, 0.28205670166015623, 0.28218572998046876, 0.2821898193359375, 0.2820721130371094, 0.2819910888671875, 0.28197479248046875, 0.28180990600585937, 0.28254721069335936, 0.2821273498535156, 0.281965576171875, 0.28245196533203126, 0.28217242431640627, 0.2820670166015625, 0.28219692993164064, 0.2824058837890625, 0.28218673706054687, 0.28217156982421876, 0.2820585632324219, 0.28222772216796876, 0.2820218811035156, 0.604822509765625, 0.2817781677246094, 0.2820003967285156, 0.2817679443359375, 0.28173825073242187, 0.2817177734375, 0.28158770751953127, 0.28191848754882814, 0.28220721435546875, 0.281891845703125, 0.2818559875488281, 0.28181607055664065, 0.2816942138671875, 0.28171160888671876, 0.28265478515625, 0.2822983093261719, 0.282181640625, 0.2823321533203125, 0.28181607055664065, 0.2818529968261719, 0.2817955322265625, 0.28184884643554686, 0.2817362060546875, 0.2819798889160156, 0.28242431640625, 0.2823710632324219, 0.28191949462890625, 0.2820423583984375, 0.28238644409179686, 0.2820577392578125, 0.28205056762695313, 0.28225946044921874, 0.28214373779296875, 0.2821488952636719, 0.28217544555664065, 0.282144775390625, 0.2822738037109375, 0.2819491882324219, 0.28210891723632814, 0.28215090942382814, 0.28207821655273435, 0.2822297668457031, 0.2820556945800781, 0.28251751708984374, 0.2820904846191406, 0.28208126831054686, 0.2820444030761719, 0.28211111450195314, 0.2820811462402344, 0.28227685546875, 0.282419189453125, 0.282071044921875, 0.2823291015625, 0.2822328186035156, 0.2820792236328125, 0.28188980102539063, 0.2820423583984375, 0.28194610595703123, 0.2819584045410156, 0.2820167541503906, 0.28234136962890627, 0.2825953369140625, 0.2823321533203125, 0.6048419799804687, 0.28182119750976564, 0.28181707763671876, 0.28184375, 0.28174026489257814, 0.2820075378417969, 0.28185089111328127, 0.2820351867675781, 0.28181607055664065, 0.2821959533691406, 0.28178738403320314, 0.282038330078125, 0.2819358215332031, 0.2820792236328125, 0.282429443359375, 0.2821683349609375, 0.28193280029296874, 0.28188262939453124, 0.2819737548828125, 0.28205361938476564, 0.2820782775878906, 0.282061767578125, 0.28200140380859373, 0.2819164123535156, 0.28227481079101563, 0.28193704223632815, 0.2823832092285156, 0.28191845703125, 0.2820126647949219, 0.2818959350585937, 0.2817884216308594, 0.28236083984375, 0.2822287292480469, 0.28201165771484377, 0.2821396484375, 0.28211508178710937, 0.28208026123046875, 0.28185089111328127, 0.281965576171875, 0.2819532775878906, 0.28200140380859373, 0.2820137023925781, 0.28240896606445315, 0.282215576171875, 0.28230230712890625, 0.2822021179199219, 0.2822195739746094, 0.2820577392578125, 0.2821047668457031, 0.28217138671875, 0.28273150634765626, 0.28209869384765623, 0.2823219299316406, 0.282203125, 0.28209356689453124, 0.2822574157714844, 0.28228302001953126, 0.2820444030761719, 0.28200454711914064, 0.2824068603515625, 0.28221133422851563, 0.2823987121582031, 0.28231680297851564]",tokens/s,3.4872555473997657,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10382,7 +10382,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e96cb-27be7b3d45588f6b6c2be754;89bf2972-c786-47fc-ae2d-f6f6dc80fd41) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe89f-326bb12f2c8c5b645545ae48;88282cf3-19d9-4b21-a543-52d2e1632fcf) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10414,7 +10414,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-1.4b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1669.656576,2168.979456,0.0,1522.532352,1394.863616,s,10,0.7282600784301758,0.07282600784301758,0.0016046889612723863,0.072286865234375,0.07310013656616211,0.07535923805236816,0.07716651924133301,"[0.07761833953857422, 0.07227785491943359, 0.07229587554931641, 0.07213699340820312, 0.07248140716552734, 0.07220976257324219, 0.07259811401367187, 0.07240211486816406, 0.07212156677246094, 0.07211804962158203]",tokens/s,3515.2276993107866,kWh,8.557824648732095e-07,4.688357626789492e-07,3.919925198287397e-06,5.244543425839556e-06,tokens/kWh,48812638.053238936,MB,1669.656576,2168.979456,0.0,1522.532352,1491.576832,s,10,39.60201782226563,3.960201782226563,0.004287537888854448,3.95872705078125,3.9651301757812503,3.9675232177734374,3.9694376513671874,"[3.969916259765625, 3.95827685546875, 3.961583984375, 3.95917724609375, 3.95727001953125, 3.964598388671875, 3.957877685546875, 3.962222412109375, 3.954532958984375, 3.95656201171875]",tokens/s,15.908280300954566,kWh,4.668280684047466e-05,2.5584906380952142e-05,0.0002084005779913184,0.00028066829121274525,tokens/kWh,224464.25895772563,,s,629,40.188444679260236,0.06389259885414986,0.008649211688047586,0.06279366302490234,0.06310651016235351,0.06346833877563476,0.13534753051757814,"[0.06427442932128906, 0.06367744064331055, 0.06450994873046875, 0.06396723175048828, 0.06460211181640625, 0.06386380767822265, 0.06388838577270507, 0.06376959991455078, 0.06295552062988281, 0.06254079818725586, 0.06283468627929688, 0.06336716842651367, 0.06294425582885742, 0.06263910293579102, 0.06311423873901367, 0.06285107040405273, 0.0626319351196289, 0.0627322883605957, 0.0628479995727539, 0.06280089569091797, 0.06270975875854493, 0.06289823913574219, 0.06288172912597656, 0.0628479995727539, 0.06255206298828125, 0.06285823822021484, 0.06274252700805664, 0.0636497917175293, 0.06330777740478516, 0.06288691329956055, 0.06279782485961914, 0.06272000122070312, 0.0627845115661621, 0.06280089569091797, 0.06273638534545899, 0.06270975875854493, 0.06262476730346679, 0.06267391967773438, 0.06279065704345703, 0.06291155242919921, 0.06272198486328125, 0.06273843383789063, 0.06275788879394531, 0.06283673477172852, 0.06281727981567382, 0.0628223991394043, 0.06280806350708008, 0.0628070411682129, 0.06269132614135742, 0.06270463943481445, 0.06276812744140625, 0.06281625747680664, 0.0628039665222168, 0.06283980941772462, 0.06280806350708008, 0.06286643218994141, 0.06285107040405273, 0.06276812744140625, 0.06268928146362304, 0.06288793563842773, 0.06313164901733398, 0.06272921752929687, 0.13550079345703125, 0.06266368103027344, 0.06356377410888672, 0.06329241561889648, 0.06264422225952149, 0.06284492874145507, 0.0626872329711914, 0.06269440078735351, 0.06265343856811524, 0.06263808059692383, 0.06275481414794921, 0.06283673477172852, 0.06269644927978515, 0.06273023986816406, 0.06265139389038087, 0.06265958404541015, 0.06273740768432617, 0.06273023986816406, 0.06259507369995117, 0.06276095962524414, 0.0627333755493164, 0.06278956985473633, 0.06275788879394531, 0.06280607986450196, 0.06271481704711913, 0.06279987335205078, 0.06269952011108398, 0.06274662399291993, 0.06278963088989258, 0.062740478515625, 0.06303334426879882, 0.06373785781860351, 0.06292582321166992, 0.06271094512939453, 0.06271779251098633, 0.06273843383789063, 0.06266486358642578, 0.06277411270141602, 0.06260326385498047, 0.06282156753540039, 0.06284064102172851, 0.06280192184448243, 0.06282966232299805, 0.0634090576171875, 0.06286438369750977, 0.06295040130615234, 0.0627701759338379, 0.06323712158203125, 0.06280294418334961, 0.06275686264038086, 0.06277939224243165, 0.06276300811767578, 0.06265753555297851, 0.06284697723388671, 0.06265868759155273, 0.06275059127807617, 0.06270156860351563, 0.06265446472167968, 0.06280294418334961, 0.06291558456420898, 0.0630118408203125, 0.06327507019042969, 0.06287968063354492, 0.13579469299316407, 0.0627957763671875, 0.06301900863647461, 0.06312243270874024, 0.06263603210449219, 0.06267289733886719, 0.06265446472167968, 0.06276729583740234, 0.06269830322265625, 0.06266368103027344, 0.06281727981567382, 0.06279167938232422, 0.06278963088989258, 0.06280908966064454, 0.06279270553588867, 0.062740478515625, 0.06272102355957031, 0.06273843383789063, 0.0626688003540039, 0.06272204971313476, 0.06283168029785156, 0.06271788787841796, 0.06267903900146485, 0.06283571243286133, 0.06279996871948242, 0.06272604751586915, 0.06271385574340821, 0.06272614288330078, 0.06299750518798829, 0.0633251838684082, 0.06275174331665039, 0.06292684936523438, 0.06292684936523438, 0.0628592643737793, 0.06273535919189453, 0.06271488189697266, 0.06278963088989258, 0.0627691535949707, 0.06268620681762695, 0.06285823822021484, 0.06286233520507813, 0.06280089569091797, 0.06345727920532226, 0.06384844970703125, 0.06308147048950195, 0.06279372787475586, 0.06315315246582032, 0.06304460906982422, 0.06285414505004883, 0.0628326416015625, 0.06283878326416016, 0.06274457550048829, 0.06277529525756836, 0.0628223991394043, 0.0628930549621582, 0.0629452781677246, 0.06277632141113282, 0.0629227523803711, 0.06294636917114257, 0.06385452651977538, 0.06318899154663087, 0.06286336135864258, 0.06281625747680664, 0.13530624389648438, 0.06260736083984375, 0.06269747161865234, 0.06274252700805664, 0.06294937515258789, 0.06281830215454101, 0.0626351661682129, 0.06279868698120117, 0.06265241622924805, 0.06271692657470704, 0.0626616325378418, 0.06298009490966797, 0.06285414505004883, 0.06270156860351563, 0.06278348922729492, 0.0628961296081543, 0.06269644927978515, 0.06311116790771484, 0.06280499267578125, 0.06280499267578125, 0.06271795272827148, 0.06268518447875976, 0.06274457550048829, 0.06278758239746093, 0.06281932830810547, 0.06269337463378906, 0.06277836990356446, 0.06291660690307617, 0.06279475021362305, 0.06276812744140625, 0.06279167938232422, 0.06280089569091797, 0.06274252700805664, 0.06374297714233398, 0.06347980880737304, 0.06290227127075196, 0.06295040130615234, 0.06277836990356446, 0.06271590423583985, 0.06294425582885742, 0.06265043258666993, 0.06261139297485352, 0.0627691535949707, 0.06273331069946289, 0.06296883010864258, 0.06279987335205078, 0.06273231887817383, 0.06280598449707031, 0.0627066879272461, 0.06277119827270508, 0.06266368103027344, 0.06285619354248047, 0.06290739059448242, 0.06309273529052735, 0.06284697723388671, 0.06271385574340821, 0.06266265487670898, 0.06269644927978515, 0.06273535919189453, 0.06299852752685547, 0.06297702407836914, 0.0630550079345703, 0.06376124954223633, 0.1357189178466797, 0.06272204971313476, 0.06270361709594727, 0.06256435012817382, 0.0626800651550293, 0.06261350250244141, 0.06276812744140625, 0.06274764633178712, 0.06273023986816406, 0.0627845115661621, 0.0631009292602539, 0.06293612670898438, 0.06271072006225586, 0.06259711837768554, 0.06265241622924805, 0.06275686264038086, 0.0627927360534668, 0.06284694290161133, 0.06273440170288086, 0.0627465591430664, 0.0627507209777832, 0.06257459259033203, 0.06266368103027344, 0.06273433685302734, 0.06268928146362304, 0.06268620681762695, 0.0627066879272461, 0.06297804641723632, 0.06287360000610352, 0.06278041458129882, 0.06275174331665039, 0.0631275520324707, 0.06284291076660156, 0.06284489440917969, 0.0627845115661621, 0.06283059310913086, 0.06286131286621094, 0.0627845115661621, 0.06292172622680664, 0.06297907257080078, 0.06292172622680664, 0.06272204971313476, 0.06272927856445312, 0.06301996612548828, 0.06288076782226562, 0.06274969482421874, 0.06268415832519532, 0.06280192184448243, 0.06279372787475586, 0.06269132614135742, 0.06294118499755859, 0.06288076782226562, 0.06280089569091797, 0.06287564849853515, 0.06286131286621094, 0.06286336135864258, 0.06295040130615234, 0.0630312957763672, 0.06286336135864258, 0.06310611343383789, 0.06309062576293946, 0.06276300811767578, 0.06275174331665039, 0.13536358642578125, 0.06267187118530274, 0.06301388931274414, 0.06275382232666016, 0.06270563125610351, 0.06265548706054687, 0.06260940933227539, 0.06260019302368164, 0.06273535919189453, 0.06322793579101563, 0.06315107345581054, 0.06276918411254882, 0.06292272186279296, 0.06296985626220702, 0.06291660690307617, 0.0628326416015625, 0.06267391967773438, 0.06278144073486328, 0.06311116790771484, 0.06270681762695313, 0.06274755096435547, 0.06271996688842774, 0.062814208984375, 0.06287052917480469, 0.06266368103027344, 0.0632616958618164, 0.0640747528076172, 0.0631992301940918, 0.06290848159790038, 0.06279996871948242, 0.06275875091552735, 0.06286336135864258, 0.06276300811767578, 0.06287871932983398, 0.06291455841064453, 0.06292480087280274, 0.06277939224243165, 0.06282956695556641, 0.06310809707641601, 0.06289510345458985, 0.06293299102783204, 0.06299852752685547, 0.06293503952026368, 0.06292172622680664, 0.06282035064697265, 0.0628930549621582, 0.06295244979858398, 0.06293510437011719, 0.06286841583251954, 0.0629719352722168, 0.06285827255249024, 0.06286227035522461, 0.06283683013916015, 0.06287043380737305, 0.06281727981567382, 0.06295449447631836, 0.06283673477172852, 0.06312243270874024, 0.06334975814819335, 0.06338355255126953, 0.0635351676940918, 0.06321964645385743, 0.06300672149658203, 0.13685658264160155, 0.06331391906738282, 0.06269747161865234, 0.06271180725097657, 0.06274457550048829, 0.06270975875854493, 0.06268931198120117, 0.06273737716674804, 0.06266572952270508, 0.06264524841308594, 0.06270880126953125, 0.06279366302490234, 0.06277836990356446, 0.06270463943481445, 0.06258278274536133, 0.06271503829956054, 0.06257852935791015, 0.06262988662719726, 0.06279065704345703, 0.06266777420043945, 0.06298009490966797, 0.0628009910583496, 0.06270147323608398, 0.06304665756225586, 0.06278348922729492, 0.06272716903686523, 0.06272819137573242, 0.06280204772949219, 0.06317350387573242, 0.06309273529052735, 0.06275686264038086, 0.06277119827270508, 0.06265760040283203, 0.06269023895263671, 0.06272000122070312, 0.06273638534545899, 0.06272921752929687, 0.06278348922729492, 0.06273433685302734, 0.06283161544799805, 0.06296371078491211, 0.06279167938232422, 0.06304972839355469, 0.06279884719848633, 0.06280908966064454, 0.06283878326416016, 0.06276095962524414, 0.0628223991394043, 0.06293606567382813, 0.06268928146362304, 0.06282854461669922, 0.06290739059448242, 0.06286454391479492, 0.0629645767211914, 0.06288076782226562, 0.06281833648681641, 0.06281315231323242, 0.06275481414794921, 0.06300262451171874, 0.06286953735351562, 0.0628070068359375, 0.06278860855102539, 0.06281011199951173, 0.135657470703125, 0.06272614288330078, 0.0626063346862793, 0.06272000122070312, 0.062740478515625, 0.06270364761352538, 0.06270153427124023, 0.0626769905090332, 0.06265036773681641, 0.06272614288330078, 0.06274662399291993, 0.06278246307373046, 0.06287052917480469, 0.06273638534545899, 0.06277840042114258, 0.06273942565917968, 0.06270771026611328, 0.0627435531616211, 0.06278246307373046, 0.06273443222045899, 0.06267587280273437, 0.06270774459838867, 0.0627547836303711, 0.06283673477172852, 0.06272716903686523, 0.06282649612426758, 0.06298624038696289, 0.06288179016113281, 0.06275686264038086, 0.06270771026611328, 0.06275584030151367, 0.06277939224243165, 0.06280192184448243, 0.06270566558837891, 0.06274764633178712, 0.06273023986816406, 0.06263808059692383, 0.06282444763183594, 0.06278758239746093, 0.06281011199951173, 0.06376755142211914, 0.06347571182250976, 0.06376038360595702, 0.0629227523803711, 0.06304256057739258, 0.06289510345458985, 0.06300979232788086, 0.0630835189819336, 0.06311014556884766, 0.06291251373291015, 0.06281011199951173, 0.06284595108032226, 0.06297600173950195, 0.0631838722229004, 0.06283980941772462, 0.06276300811767578, 0.06388838577270507, 0.06375936126708984, 0.06314803314208985, 0.06280806350708008, 0.06305177688598633, 0.06278758239746093, 0.06283468627929688, 0.13527244567871094, 0.06263808059692383, 0.06261248016357422, 0.0626431999206543, 0.06266470336914062, 0.06265958404541015, 0.06273331069946289, 0.06268838500976562, 0.06260825729370117, 0.0626984977722168, 0.06257356643676758, 0.06251417541503906, 0.06272000122070312, 0.06271078491210938, 0.06281318283081054, 0.0626165771484375, 0.06258483123779297, 0.06272204971313476, 0.06278656005859375, 0.06273535919189453, 0.06279987335205078, 0.06263296127319336, 0.06265548706054687, 0.06277939224243165, 0.0626688003540039, 0.06253567886352539, 0.06272409439086914, 0.06255615997314454, 0.0626800651550293, 0.06279993438720703, 0.06277011108398438, 0.06278963088989258, 0.06329241561889648, 0.06278553771972656, 0.06273331069946289, 0.06275993728637695, 0.06270361709594727, 0.06275788879394531, 0.06300364685058593, 0.06307123184204101, 0.06295878219604492, 0.06301062393188477, 0.0629452781677246, 0.06261555099487305, 0.06271897506713867, 0.0627435531616211, 0.06307948684692383, 0.06280492782592774, 0.06291151809692383, 0.06285308837890625, 0.06289203262329102, 0.0627957763671875, 0.06280192184448243, 0.0626698226928711, 0.06285516738891601, 0.06284288024902343, 0.06276607894897461, 0.06273740768432617, 0.06280303955078125, 0.06295344161987304, 0.06289913558959961, 0.06276300811767578, 0.06277939224243165, 0.13537791442871094, 0.06255104064941407, 0.0627322883605957, 0.06268739318847656, 0.0626778564453125, 0.06263091278076172, 0.06270771026611328, 0.06290227127075196, 0.06272819137573242, 0.06269132614135742, 0.06260838317871094, 0.06267494583129883, 0.06262080001831055, 0.06272099304199219, 0.06308854293823242, 0.06282035064697265, 0.06273945617675782, 0.06265651321411132, 0.06269337463378906, 0.06272115325927734, 0.06274451065063477, 0.063052734375, 0.0629596176147461, 0.0628930549621582, 0.06275993728637695, 0.06286233520507813, 0.06307123184204101, 0.06278656005859375, 0.06277529525756836, 0.0628039665222168, 0.06286336135864258, 0.06290841674804687, 0.06281523132324218, 0.06282342529296875, 0.062814208984375, 0.06275584030151367, 0.06273331069946289, 0.06286131286621094, 0.06275584030151367, 0.0628848648071289, 0.06284288024902343, 0.06286131286621094, 0.06290537643432617, 0.06282748794555663, 0.06285516738891601, 0.0628922233581543, 0.06275564956665039, 0.0628930549621582, 0.0627619857788086, 0.06279475021362305, 0.06284601593017577, 0.06277318572998047, 0.0627507209777832, 0.0627507209777832, 0.06277222442626954, 0.0628039665222168, 0.0627322883605957, 0.06279065704345703, 0.06281113433837891, 0.06294118499755859, 0.06282547378540039, 0.06285414505004883, 0.06272000122070312]",tokens/s,15.651265059396618,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -10453,7 +10453,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d45f8-3e1a1de308fcccf938ca13cf;28e4513f-4022-4e17-86e2-4d863f7d8ab5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe91a-68243c1118228dc714884850;1447d8bd-3f6e-4326-bc73-ce3087912abf) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10637,7 +10637,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90ee-26c573756a8691fb351afb4b;4f3f14bf-8f55-4159-ac64-53f8babfd4c5) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe29d-5e33a59307ea531f1ee1c9a8;0c933833-7c76-4bc8-bc37-52753cf94b43) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -10726,7 +10726,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e96f4-4eae4c7b789e12783e13b46b;3d9b480b-bcfd-4a3d-9c22-0a08dc158e5a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe8c9-58801f1c5f8837682ef6e0ad;57d6979b-303a-4b70-809b-248405bf2ae5) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10786,7 +10786,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90c5-5215db386e82b32c57680ac7;6622395e-f7d6-45cb-a01a-7874ec3afd5e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe271-52db13875c715d1278e428f3;411e2865-1b8c-4592-b754-4520b40d0dda) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -11087,7 +11087,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e8f31-77cb6c263ec80dcf64376afa;2f9c577a-9de6-43c2-a48c-49e6d217a612) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe0d2-2fa665315e1d572e0d2d5469;10b3970f-4b04-4dc8-9779-f45e93da1c14) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -11172,7 +11172,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e967d-191918497e4f4e0c781e92b9;706ab3ff-7fa0-496d-93b1-0e1cd8043ffb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe84b-334854473453cf842ea2d38e;65f5c0cb-3a8b-4623-9370-e5508106cce3) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11288,7 +11288,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95b0-02e2af6574cae859608a5c42;35ccd9ad-9670-4792-8430-f973d3af90da) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe77e-2060bd2755c6b51a2900fb94;bfa882fd-d9c5-4ab0-a5f2-508981ac4daa) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11393,7 +11393,7 @@ ImportError: This modeling file requires the following packages that were not fo ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1433.575424,1613.234176,0.0,966.787072,857.647616,s,10,0.6057559051513672,0.060575590515136724,0.0005271997959633458,0.0604053611755371,0.060655676651000974,0.06140215816497802,0.06199934337615967,"[0.06214863967895508, 0.060373504638671874, 0.06034431838989258, 0.0604667854309082, 0.06043558502197265, 0.06048979187011719, 0.06032617568969727, 0.06034182357788086, 0.060454143524169925, 0.06037513732910156]",tokens/s,4226.124711669634,kWh,7.124543090262408e-07,3.9038897249671454e-07,3.0336938593495046e-06,4.13653714087246e-06,tokens/kWh,61887513.94747676,MB,1433.575424,1613.234176,0.0,966.787072,888.374272,s,10,32.95503466796875,3.2955034667968746,0.047364368059071654,3.311632568359375,3.3455935791015623,3.347526403808594,3.3490726635742187,"[3.210956298828125, 3.312005126953125, 3.3333193359375, 3.337546142578125, 3.3451640625, 3.311260009765625, 3.2419833984375, 3.244228271484375, 3.349459228515625, 3.26911279296875]",tokens/s,19.116957586220966,kWh,3.7684675597522154e-05,2.065301277971624e-05,0.00015636918366284723,0.00021470687204008565,tokens/kWh,293423.3049990032,,s,629,33.444424747467025,0.053170786561950775,0.007149643566833179,0.052948001861572266,0.053361048889160156,0.053613362121582034,0.11106307861328125,"[0.05055692672729492, 0.050549758911132815, 0.0504268798828125, 0.050601982116699216, 0.050607105255126954, 0.05235609436035156, 0.051211265563964846, 0.05066854476928711, 0.052313087463378906, 0.05290496063232422, 0.051081214904785156, 0.05304012680053711, 0.05316812896728516, 0.05326540756225586, 0.05314048004150391, 0.052894718170166014, 0.053157886505126956, 0.0508006706237793, 0.05070435333251953, 0.05057334518432617, 0.050143199920654295, 0.05045248031616211, 0.050694145202636716, 0.05033267211914062, 0.05023539352416992, 0.05030809783935547, 0.0504453125, 0.05063782501220703, 0.050514942169189454, 0.050050048828125, 0.05061734390258789, 0.050435073852539064, 0.05070950317382812, 0.050601982116699216, 0.050298881530761716, 0.05028966522216797, 0.050678783416748044, 0.05065011215209961, 0.05067468643188477, 0.05041971206665039, 0.050601982116699216, 0.05051084899902344, 0.0503419189453125, 0.05042480087280273, 0.05045248031616211, 0.050813953399658204, 0.05053235244750977, 0.050408447265625, 0.05015654373168945, 0.050729984283447264, 0.05063167953491211, 0.05070336151123047, 0.050759681701660155, 0.05066342544555664, 0.051158016204833984, 0.05045862579345703, 0.050520065307617185, 0.050612224578857425, 0.05101875305175781, 0.050546688079833986, 0.05056716918945312, 0.05059379196166992, 0.11082444763183594, 0.05010227203369141, 0.05011865615844727, 0.050592769622802736, 0.05045555114746094, 0.050351104736328124, 0.05048934555053711, 0.05056512069702149, 0.05039308929443359, 0.05048627090454102, 0.05082009506225586, 0.05061427307128906, 0.05172019195556641, 0.053286911010742184, 0.052977664947509766, 0.053282817840576174, 0.05350297546386719, 0.05334732818603516, 0.052874240875244144, 0.05326438522338867, 0.053250049591064455, 0.052988929748535155, 0.05293363189697266, 0.053048320770263675, 0.053043201446533204, 0.052942848205566405, 0.05296537780761719, 0.05331148910522461, 0.0531599349975586, 0.05266841506958008, 0.05291212844848633, 0.0532940788269043, 0.05293056106567383, 0.05306880187988281, 0.053215232849121094, 0.05328998565673828, 0.05301862335205078, 0.053117950439453124, 0.053152767181396485, 0.05327360153198242, 0.05285273742675781, 0.05313740921020508, 0.05293875122070312, 0.0531599349975586, 0.05350707244873047, 0.05283020782470703, 0.053005313873291014, 0.053250049591064455, 0.05289779281616211, 0.05323263931274414, 0.05325107192993164, 0.053103614807128906, 0.05288857650756836, 0.053174270629882815, 0.05288550567626953, 0.05302374267578125, 0.053031936645507816, 0.053152767181396485, 0.05318656158447266, 0.053048320770263675, 0.052931583404541016, 0.05313433456420898, 0.05313536071777344, 0.11145728302001953, 0.05347020721435547, 0.05130854415893555, 0.05311385726928711, 0.05313024139404297, 0.053664768218994144, 0.05321011352539062, 0.05300940704345703, 0.05283430480957031, 0.05282815933227539, 0.053207038879394535, 0.05295820617675781, 0.05309132766723633, 0.05311078262329102, 0.05315686416625977, 0.05338521575927734, 0.05283225631713867, 0.05321011352539062, 0.05055692672729492, 0.05051801681518555, 0.050525184631347655, 0.05032857513427735, 0.050492416381835936, 0.0506695671081543, 0.05407129669189453, 0.053220352172851565, 0.05303910446166992, 0.05338009643554688, 0.05389209747314453, 0.05305753707885742, 0.053131263732910154, 0.05332787322998047, 0.05344768142700195, 0.053359615325927735, 0.053166080474853515, 0.053114879608154295, 0.05293875122070312, 0.05538611221313477, 0.053528575897216796, 0.05374566268920898, 0.05363507080078125, 0.05336678314208984, 0.05332889556884766, 0.05314355087280274, 0.0530964469909668, 0.0531599349975586, 0.053157886505126956, 0.05327360153198242, 0.05270016098022461, 0.05305344009399414, 0.05307289505004883, 0.052948001861572266, 0.0529571533203125, 0.053146625518798826, 0.054168575286865236, 0.053369857788085937, 0.05245644760131836, 0.05257727813720703, 0.05304422378540039, 0.05309132766723633, 0.0530063362121582, 0.05309235382080078, 0.05296537780761719, 0.11123097229003906, 0.05046988677978516, 0.05231820678710938, 0.05289779281616211, 0.05303705596923828, 0.0531486701965332, 0.05290598297119141, 0.05290188980102539, 0.053070846557617186, 0.052863998413085936, 0.05271859359741211, 0.052961280822753906, 0.05286809539794922, 0.05293056106567383, 0.05316505432128906, 0.05302783966064453, 0.052942848205566405, 0.05327155303955078, 0.05299097442626953, 0.05291212844848633, 0.05306470489501953, 0.05286809539794922, 0.052808704376220705, 0.05311078262329102, 0.05294182586669922, 0.05307699203491211, 0.05300735855102539, 0.05339340972900391, 0.05305446243286133, 0.053215232849121094, 0.05308927917480469, 0.0530964469909668, 0.05355212783813477, 0.05313945770263672, 0.05329715347290039, 0.053356544494628906, 0.05304524612426758, 0.0529090576171875, 0.05321113586425781, 0.052805633544921876, 0.0531671028137207, 0.053059585571289064, 0.053163009643554686, 0.0533043212890625, 0.052972545623779295, 0.05300940704345703, 0.053028865814208986, 0.053128192901611325, 0.05330124664306641, 0.05283737564086914, 0.052929534912109374, 0.05303603363037109, 0.05309439849853516, 0.05319475173950195, 0.052988929748535155, 0.053141502380371096, 0.05310976028442383, 0.05326540756225586, 0.05312409591674805, 0.05306470489501953, 0.05314559936523437, 0.05304115295410156, 0.053269504547119144, 0.11326566314697266, 0.052811775207519535, 0.05257932662963867, 0.052915199279785156, 0.05235302352905274, 0.052964351654052735, 0.052910079956054686, 0.053433345794677733, 0.05347942352294922, 0.05304729461669922, 0.053384193420410155, 0.053310462951660156, 0.05308108901977539, 0.05319680023193359, 0.053157886505126956, 0.05334527969360352, 0.05313024139404297, 0.0531599349975586, 0.05294899368286133, 0.05341286468505859, 0.05349478530883789, 0.05322751998901367, 0.053141502380371096, 0.05327155303955078, 0.05298486328125, 0.05312201690673828, 0.05294079971313476, 0.05312921524047851, 0.05309542465209961, 0.05329715347290039, 0.0529090576171875, 0.05327155303955078, 0.053098495483398435, 0.05312614440917969, 0.05317836761474609, 0.05309337615966797, 0.05302579116821289, 0.05292851257324219, 0.05321830368041992, 0.05315379333496094, 0.0530063362121582, 0.053164031982421874, 0.05276467132568359, 0.05319168090820312, 0.053266433715820315, 0.05325516891479492, 0.05300326538085937, 0.05318041610717773, 0.05311590576171875, 0.05307392120361328, 0.05298688125610351, 0.05311491012573242, 0.05316912078857422, 0.05314252853393555, 0.05291622543334961, 0.05327769470214844, 0.05308108901977539, 0.052838401794433595, 0.05321932983398438, 0.05322854232788086, 0.05291622543334961, 0.05306880187988281, 0.0530513916015625, 0.11380941009521485, 0.05392281723022461, 0.05319782257080078, 0.0529172477722168, 0.05304217529296875, 0.05285171127319336, 0.05305855941772461, 0.053326847076416016, 0.052962303161621094, 0.05067366409301758, 0.05216255950927735, 0.05297868728637695, 0.052967422485351565, 0.05311283111572265, 0.05292748641967773, 0.0527534065246582, 0.05366067123413086, 0.05328076934814453, 0.05293875122070312, 0.05326028823852539, 0.054902782440185545, 0.05330124664306641, 0.05306060791015625, 0.054798336029052735, 0.05343436813354492, 0.05301248168945313, 0.0530247688293457, 0.053008384704589843, 0.053100543975830077, 0.053236736297607425, 0.053103614807128906, 0.05306470489501953, 0.05303705596923828, 0.05449833679199219, 0.05329097747802734, 0.05346406555175781, 0.05122662353515625, 0.05233356857299805, 0.0516577262878418, 0.05175910568237305, 0.05084979248046875, 0.052190208435058595, 0.05241241455078125, 0.05258137512207031, 0.05259366226196289, 0.05246364974975586, 0.05270115280151367, 0.051165184020996096, 0.050356224060058595, 0.050336769104003906, 0.05067366409301758, 0.05067264175415039, 0.05139148712158203, 0.05270220947265625, 0.05225267028808594, 0.052805633544921876, 0.052724735260009765, 0.052673534393310545, 0.052038688659667966, 0.05173859024047851, 0.05067059326171875, 0.05120102310180664, 0.05047500610351562, 0.11094732666015625, 0.05059993743896484, 0.05222707366943359, 0.051961856842041014, 0.05243699264526367, 0.05271244812011719, 0.05230387115478516, 0.05252096176147461, 0.05257625579833984, 0.05247488021850586, 0.05067161560058594, 0.05085388946533203, 0.05221683120727539, 0.05244313430786133, 0.052621311187744144, 0.05067468643188477, 0.050402305603027345, 0.050931713104248044, 0.050408447265625, 0.050628608703613284, 0.05142425537109375, 0.05258342361450195, 0.05256806564331055, 0.052615169525146485, 0.052689918518066405, 0.05127167892456055, 0.05078015899658203, 0.05067264175415039, 0.05039308929443359, 0.050895870208740236, 0.05064089584350586, 0.0510648307800293, 0.05180108642578125, 0.05264691162109375, 0.052613121032714844, 0.05245132827758789, 0.05082624053955078, 0.050563072204589846, 0.05251891326904297, 0.052380672454833986, 0.05247180938720703, 0.05298175811767578, 0.05081497573852539, 0.05067059326171875, 0.05272576141357422, 0.05244211196899414, 0.05257318496704102, 0.05089791870117188, 0.05063577651977539, 0.050754558563232424, 0.05065216064453125, 0.05063987350463867, 0.050726913452148435, 0.05062348937988281, 0.050700286865234374, 0.050685951232910156, 0.050715648651123046, 0.050762752532958984, 0.05066547012329101, 0.05073715209960938, 0.05108736038208008, 0.05073100662231445, 0.05063270568847656, 0.11117158508300781, 0.05055897521972656, 0.05045555114746094, 0.0507852783203125, 0.05078835296630859, 0.05041971206665039, 0.050520065307617185, 0.051056640625, 0.05073920059204102, 0.050677761077880856, 0.05052928161621094, 0.050726913452148435, 0.05092761611938477, 0.0508037109375, 0.050697216033935545, 0.05126348876953125, 0.05082521438598633, 0.05063679885864258, 0.05064089584350586, 0.05107814407348633, 0.05085081481933594, 0.05065420913696289, 0.05098188781738281, 0.05121843338012695, 0.0508671989440918, 0.05059174346923828, 0.05043916702270508, 0.050735103607177735, 0.05065318298339844, 0.05065932846069336, 0.050651134490966795, 0.050902015686035154, 0.05050265502929688, 0.05043711853027344, 0.05080678558349609, 0.05058969497680664, 0.05069823837280273, 0.05122457504272461, 0.050874366760253906, 0.05068902587890625, 0.05118566513061523, 0.05067264175415039, 0.05083750534057617, 0.052819969177246094, 0.05310464096069336, 0.05312102508544922, 0.05280051040649414, 0.053100543975830077, 0.05288140869140625, 0.0530513916015625, 0.052929534912109374, 0.05296025466918945, 0.05301248168945313, 0.05307494354248047, 0.053335041046142576, 0.05313843154907227, 0.05323468780517578, 0.05314355087280274, 0.05298995208740234, 0.05305241775512695, 0.0535470085144043, 0.0530247688293457, 0.05322956848144531, 0.11150131225585938, 0.05239091110229492, 0.05395455932617187, 0.05306675338745117, 0.052983806610107424, 0.05279334259033203, 0.05306777572631836, 0.05310464096069336, 0.05305142211914062, 0.05349577713012695, 0.05253017425537109, 0.052959232330322265, 0.05325823974609375, 0.05369651031494141, 0.05309030532836914, 0.05308313751220703, 0.05314048004150391, 0.05313433456420898, 0.052975616455078124, 0.05286502456665039, 0.053441535949707034, 0.05593600082397461, 0.05402316665649414, 0.054691841125488284, 0.054489086151123044, 0.05355417633056641, 0.05356748962402344, 0.053318656921386716, 0.053172222137451174, 0.05299507141113281, 0.052944896697998046, 0.05327462387084961, 0.053098495483398435, 0.05425971221923828, 0.053751808166503906, 0.053381118774414066, 0.05326131057739258, 0.05306675338745117, 0.053043201446533204, 0.05301248168945313, 0.053032958984375, 0.053318656921386716, 0.052954113006591794, 0.05296025466918945, 0.052831230163574217, 0.0528353271484375, 0.05345075225830078, 0.05315686416625977, 0.05309747314453125, 0.05294387054443359, 0.05311385726928711, 0.05368012619018555, 0.05302783966064453, 0.052789249420166016, 0.05306367874145508, 0.05369241714477539, 0.05337497711181641, 0.05304422378540039, 0.05323878479003906, 0.053119998931884765, 0.053236736297607425, 0.05083852767944336, 0.050754558563232424, 0.11110809326171875, 0.05090611267089844, 0.05059379196166992, 0.05061939239501953, 0.05055692672729492, 0.05063679885864258, 0.05045964813232422, 0.05061939239501953, 0.05209088134765625, 0.053203968048095705, 0.0510115852355957, 0.05065523147583008, 0.05082828903198242, 0.050618366241455076, 0.050678783416748044, 0.05061017608642578, 0.050359294891357424, 0.05069107055664063, 0.05045452880859375, 0.05055692672729492, 0.05055897521972656, 0.05115289688110351, 0.05058867263793945, 0.05075558471679688, 0.05075763320922851, 0.050561023712158204, 0.050651134490966795, 0.050667518615722655, 0.05059788894653321, 0.0521082878112793, 0.053513214111328124, 0.053580799102783204, 0.05332992172241211, 0.053354496002197264, 0.053275646209716795, 0.05306470489501953, 0.053085182189941404, 0.050902015686035154, 0.05086310577392578, 0.05082419204711914, 0.051558399200439455, 0.05321830368041992, 0.05328076934814453, 0.05346815872192383, 0.05288652801513672, 0.05321113586425781, 0.053128192901611325, 0.05309030532836914, 0.05333913421630859, 0.05410815811157227, 0.05075353622436524, 0.05102592086791992, 0.053070846557617186, 0.0530513916015625, 0.053441535949707034, 0.05319168090820312, 0.05306470489501953, 0.05308415985107422, 0.053098495483398435, 0.05348556900024414, 0.05324185562133789, 0.053324798583984374, 0.052944896697998046]",tokens/s,18.807320046598747,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11432,7 +11432,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d466c-7ae192436aed31d34a93895d;ae5f62be-f893-4d4e-891f-3b644203b165) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe994-303ff22678d21c5072fa2321;64ec90b1-f70e-4afd-ac93-64fdff2623f5) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11503,7 +11503,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9587-79b179ae4ba63a9c33d8440b;df2d3b9e-0b06-4fef-a8a8-672d6d029df5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe756-4a5d8c49691ef2887a1dc076;39c221e8-e83e-4636-b324-11af3123b8e1) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11574,7 +11574,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e962b-7bf2d8f649a5f21217917df0;7afda9fa-bb88-438b-bcf1-4c13829e7ac9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe7fa-1d98e61e1f52063a6215ff7e;a526f2b3-bbd0-43eb-aca8-a9b0b1c8541a) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11606,7 +11606,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5008.822272,8173.125632,0.0,7526.678528,6833.744896,s,10,3.2637095642089844,0.3263709564208984,0.0007368089127270137,0.3261606750488281,0.32729655456542966,0.3276723556518555,0.3279729965209961,"[0.32596209716796876, 0.32586428833007813, 0.3263592529296875, 0.32655685424804687, 0.326583251953125, 0.32592489624023435, 0.32721304321289063, 0.32552175903320313, 0.32567596435546875, 0.32804815673828125]",tokens/s,784.3835211545423,kWh,3.849155682816727e-06,2.109159659243408e-06,2.032348400070999e-05,2.6281799342770124e-05,tokens/kWh,9740581.177917834,MB,5008.822272,8173.125632,0.0,7526.678528,7110.584832,s,10,177.045357421875,17.7045357421875,0.0012845826660970417,17.70458984375,17.706225781249998,17.70631015625,17.70637765625,"[17.703931640625, 17.701494140625, 17.704443359375, 17.70496875, 17.70412890625, 17.70639453125, 17.70620703125, 17.704736328125, 17.70418359375, 17.704869140625]",tokens/s,3.558410167733434,kWh,0.00020893763350596564,0.00011451522852683938,0.0010782542004162915,0.0014017070624490963,tokens/kWh,44945.196958574845,,s,629,179.70182278442405,0.285694471835332,0.038836008660610016,0.2809876403808594,0.28119777221679687,0.2813642761230469,0.6074044165039063,"[0.2808002624511719, 0.2810859375, 0.2808913879394531, 0.280943603515625, 0.2809088134765625, 0.2810040283203125, 0.28078695678710935, 0.28094772338867186, 0.2809241638183594, 0.2809722900390625, 0.2809620361328125, 0.2810992736816406, 0.28092007446289063, 0.28101119995117185, 0.2808463439941406, 0.281027587890625, 0.2808821716308594, 0.2812037048339844, 0.2810091552734375, 0.2810931091308594, 0.28097125244140625, 0.2809722900390625, 0.2810142822265625, 0.280943603515625, 0.28114544677734377, 0.28108175659179685, 0.28101937866210935, 0.28090777587890625, 0.2809098205566406, 0.2811453552246094, 0.28107467651367185, 0.281064453125, 0.2811750793457031, 0.28106442260742187, 0.2810439758300781, 0.28124465942382815, 0.2813736877441406, 0.28123956298828123, 0.28094259643554687, 0.28106451416015626, 0.2809650573730469, 0.2809836120605469, 0.2809210205078125, 0.28089547729492187, 0.2809016418457031, 0.280958984375, 0.28140850830078123, 0.28111666870117186, 0.2809743347167969, 0.2808883056640625, 0.28079623413085936, 0.28105413818359376, 0.2809292907714844, 0.2810009460449219, 0.28097329711914065, 0.28109619140625, 0.2809098205566406, 0.28071221923828127, 0.280933349609375, 0.28117401123046876, 0.28113409423828123, 0.28092620849609373, 0.60747265625, 0.28070806884765626, 0.2811351013183594, 0.28077362060546873, 0.28089959716796875, 0.2807675476074219, 0.2810275268554687, 0.28092312622070315, 0.2809046936035156, 0.2808084411621094, 0.2810439758300781, 0.2808105163574219, 0.28097430419921876, 0.2808217468261719, 0.28103271484375, 0.28088729858398437, 0.2810675048828125, 0.28095693969726565, 0.28103167724609374, 0.28106036376953125, 0.28104193115234377, 0.2808299560546875, 0.2809108581542969, 0.28080743408203124, 0.28094873046875, 0.280958984375, 0.2808985595703125, 0.2808197021484375, 0.2809333801269531, 0.28095489501953125, 0.2808002624511719, 0.28090264892578126, 0.2809466857910156, 0.28094158935546876, 0.2808708801269531, 0.28107467651367185, 0.2808638000488281, 0.28104391479492186, 0.2809292907714844, 0.2809661560058594, 0.28094772338867186, 0.28103271484375, 0.28111053466796876, 0.2809088134765625, 0.28127651977539064, 0.2812466125488281, 0.2810214538574219, 0.2809333801269531, 0.2810347595214844, 0.28096408081054686, 0.28108184814453124, 0.2808186950683594, 0.2811372985839844, 0.28084515380859376, 0.28100827026367187, 0.28106539916992185, 0.2810152282714844, 0.2809989013671875, 0.2808842163085937, 0.2811985778808594, 0.2813501586914062, 0.2810224609375, 0.28091802978515623, 0.6079191284179688, 0.2806446228027344, 0.28111154174804687, 0.28073370361328126, 0.2810019836425781, 0.28069903564453125, 0.281046875, 0.2809036865234375, 0.28084225463867185, 0.2807582702636719, 0.28101119995117185, 0.28089959716796875, 0.28108697509765623, 0.28107672119140625, 0.280932373046875, 0.28086679077148435, 0.2810224609375, 0.28107571411132815, 0.2811197509765625, 0.28087091064453124, 0.2810859375, 0.2809395141601562, 0.2808412170410156, 0.28205978393554687, 0.2808739929199219, 0.28123648071289065, 0.28117913818359375, 0.28086273193359373, 0.28080435180664065, 0.2808360900878906, 0.28089242553710936, 0.28096820068359374, 0.2810460205078125, 0.28101544189453126, 0.28080010986328124, 0.2809333801269531, 0.280880126953125, 0.28094464111328127, 0.28085147094726565, 0.2810408935546875, 0.28090060424804686, 0.28079513549804686, 0.28095794677734376, 0.28093438720703123, 0.28107467651367185, 0.28103067016601563, 0.28095693969726565, 0.28259121704101564, 0.2810726318359375, 0.28107366943359374, 0.2812630920410156, 0.2811709289550781, 0.2813378601074219, 0.2809466857910156, 0.28098046875, 0.28088116455078127, 0.2810010681152344, 0.2809774169921875, 0.28094351196289064, 0.28106137084960936, 0.28105523681640626, 0.28106649780273435, 0.28105010986328127, 0.6080983276367188, 0.2805442810058594, 0.2809896545410156, 0.2807879638671875, 0.2809374694824219, 0.28062310791015627, 0.28097537231445313, 0.28071218872070314, 0.280995849609375, 0.2809415588378906, 0.2810777587890625, 0.2808299560546875, 0.28200244140625, 0.28103067016601563, 0.28094259643554687, 0.28076031494140624, 0.28103372192382814, 0.28095693969726565, 0.28099789428710936, 0.28095590209960936, 0.28118731689453125, 0.28101324462890626, 0.2808268737792969, 0.28090060424804686, 0.2810726318359375, 0.28087408447265627, 0.2808984680175781, 0.2808555603027344, 0.2809251708984375, 0.28087100219726563, 0.28081143188476565, 0.28094772338867186, 0.281238525390625, 0.28138494873046876, 0.2810378112792969, 0.28088116455078127, 0.2808719482421875, 0.28103988647460937, 0.28095794677734376, 0.28113714599609374, 0.28095693969726565, 0.2808350830078125, 0.2807224426269531, 0.2808401794433594, 0.283072509765625, 0.2809395141601562, 0.28103167724609374, 0.2809651489257812, 0.2811422424316406, 0.2808913879394531, 0.280922119140625, 0.2810408935546875, 0.28119448852539064, 0.2809671630859375, 0.28113409423828123, 0.28099072265625, 0.28125595092773437, 0.280995849609375, 0.2812262268066406, 0.2814044189453125, 0.28109722900390627, 0.28100811767578127, 0.2810019836425781, 0.6083676147460938, 0.281069580078125, 0.2809661865234375, 0.2808504028320313, 0.2810521545410156, 0.28073675537109377, 0.28086477661132814, 0.2808186950683594, 0.2809518127441406, 0.2808637390136719, 0.28104193115234377, 0.28080743408203124, 0.28095489501953125, 0.2810368041992187, 0.28096408081054686, 0.28084844970703127, 0.2810756530761719, 0.28226763916015624, 0.2809692077636719, 0.28088626098632813, 0.28098355102539063, 0.2808197021484375, 0.2807982177734375, 0.28103277587890624, 0.28112786865234374, 0.280961181640625, 0.2809997863769531, 0.28112399291992185, 0.2811482543945312, 0.2809599914550781, 0.28090060424804686, 0.2811207580566406, 0.28099172973632813, 0.2809938049316406, 0.2809968566894531, 0.28118731689453125, 0.280995849609375, 0.2809426574707031, 0.28101626586914064, 0.2809979553222656, 0.28103265380859377, 0.280943603515625, 0.28089447021484376, 0.28097329711914065, 0.2810726318359375, 0.28112384033203125, 0.28100811767578127, 0.2809580078125, 0.28122515869140624, 0.28084429931640625, 0.2810142822265625, 0.2808360900878906, 0.2810214538574219, 0.2808760375976562, 0.2810133666992187, 0.2810059509277344, 0.28116583251953126, 0.28140237426757814, 0.2807633972167969, 0.28105523681640626, 0.28104702758789063, 0.28084735107421877, 0.2810245056152344, 0.6076272583007812, 0.28068353271484375, 0.28097637939453124, 0.2806927490234375, 0.28078695678710935, 0.28059033203125, 0.28082278442382813, 0.2807767028808594, 0.2808760375976562, 0.28086578369140625, 0.2808350830078125, 0.28088626098632813, 0.2813931579589844, 0.2808126525878906, 0.28106640625, 0.28102655029296875, 0.2811617431640625, 0.28116583251953126, 0.2809968566894531, 0.28094259643554687, 0.28096307373046875, 0.28095489501953125, 0.28097637939453124, 0.28106137084960936, 0.2811781005859375, 0.28100607299804686, 0.28082177734375, 0.28104290771484375, 0.28098150634765623, 0.281091064453125, 0.2810624084472656, 0.28119244384765624, 0.28088729858398437, 0.2811064453125, 0.2811351013183594, 0.2812886962890625, 0.28093438720703123, 0.28102554321289064, 0.28370022583007815, 0.2810654602050781, 0.2810654602050781, 0.280869873046875, 0.2810214538574219, 0.28114227294921873, 0.28109515380859373, 0.28090060424804686, 0.28101324462890626, 0.2810245056152344, 0.28122726440429685, 0.2809302978515625, 0.28107366943359374, 0.28088934326171877, 0.28120883178710937, 0.28091390991210935, 0.2811975708007812, 0.2811269226074219, 0.2810368041992187, 0.2808280639648438, 0.2809659729003906, 0.2812037048339844, 0.28127130126953126, 0.28098968505859373, 0.28096511840820315, 0.60883251953125, 0.28060467529296873, 0.2809794616699219, 0.28092312622070315, 0.2813982849121094, 0.2813009948730469, 0.2809886779785156, 0.2808883056640625, 0.28107366943359374, 0.280922119140625, 0.2809456787109375, 0.2808084411621094, 0.2810163269042969, 0.28097537231445313, 0.2810071105957031, 0.2808258666992188, 0.2810491027832031, 0.28098968505859373, 0.2814863891601562, 0.2812784118652344, 0.281064453125, 0.28083511352539064, 0.28078179931640623, 0.28249298095703124, 0.2810910034179688, 0.2808832092285156, 0.28091494750976564, 0.2809292907714844, 0.28089141845703125, 0.28101425170898436, 0.2809743347167969, 0.2809722900390625, 0.2810408935546875, 0.28118426513671874, 0.2812303466796875, 0.2811094970703125, 0.28078387451171877, 0.28088626098632813, 0.28119451904296877, 0.28095895385742187, 0.28112588500976565, 0.28113006591796874, 0.2810581970214844, 0.280953857421875, 0.28095077514648437, 0.28101223754882815, 0.281027587890625, 0.28109515380859373, 0.2811064453125, 0.28092825317382814, 0.28113818359375, 0.28082891845703123, 0.28119143676757813, 0.28105426025390623, 0.2811329956054687, 0.28106854248046875, 0.2810572814941406, 0.28092724609375, 0.2808821716308594, 0.2811207580566406, 0.2811412353515625, 0.2810296325683594, 0.28099172973632813, 0.60706103515625, 0.2807059936523438, 0.2810726318359375, 0.2809098205566406, 0.28105624389648437, 0.2806220703125, 0.2809917907714844, 0.28088113403320314, 0.28103372192382814, 0.28095281982421877, 0.28107879638671873, 0.2810849304199219, 0.2812733459472656, 0.2809743347167969, 0.2812303466796875, 0.28095281982421877, 0.2824908752441406, 0.2809866333007813, 0.2809456787109375, 0.280880126953125, 0.2809169921875, 0.2809302978515625, 0.2807879638671875, 0.2809661560058594, 0.2810798034667969, 0.28095489501953125, 0.28097537231445313, 0.2809948120117188, 0.2809794616699219, 0.28092724609375, 0.28095794677734376, 0.2811883544921875, 0.28089959716796875, 0.2809692077636719, 0.28097552490234373, 0.2809906005859375, 0.28083914184570313, 0.2809456787109375, 0.2810152893066406, 0.28097125244140625, 0.2810501708984375, 0.2807622985839844, 0.280890380859375, 0.28096511840820315, 0.2809323425292969, 0.28090264892578126, 0.28089752197265627, 0.28092108154296874, 0.2811678771972656, 0.2809251708984375, 0.2810941467285156, 0.2808832092285156, 0.2812651672363281, 0.2808053894042969, 0.281017333984375, 0.28113409423828123, 0.28139724731445315, 0.28102041625976565, 0.2808330383300781, 0.2810378112792969, 0.2812538757324219, 0.28141259765625, 0.2812620849609375, 0.6072289428710937, 0.28075930786132813, 0.28142489624023437, 0.28085861206054685, 0.280919189453125, 0.2806700439453125, 0.28098770141601564, 0.2808155517578125, 0.2809251708984375, 0.28085861206054685, 0.2810214538574219, 0.28077163696289065, 0.280939453125, 0.2808053894042969, 0.2809876403808594, 0.2808883056640625, 0.28102349853515624, 0.2819471435546875, 0.28139111328125, 0.2810654602050781, 0.28137985229492185, 0.28105010986328127, 0.28085861206054685, 0.2810859375, 0.28121600341796876, 0.2809887390136719, 0.28093536376953127, 0.280958984375, 0.28101223754882815, 0.2809333801269531, 0.2809938049316406, 0.28112384033203125, 0.28098458862304687, 0.281059326171875, 0.2811770935058594, 0.28115045166015623, 0.28095794677734376, 0.2810523071289063, 0.28106121826171876, 0.2809232177734375, 0.28085751342773435, 0.28083096313476563, 0.2809395141601562, 0.2809108581542969, 0.2809722900390625, 0.2809190673828125, 0.28105520629882813, 0.28091390991210935, 0.2810091552734375, 0.2810040283203125, 0.28116790771484373, 0.28094573974609377, 0.2811063232421875, 0.280995849609375, 0.2812528686523437, 0.2810511474609375, 0.2810040283203125, 0.2809139709472656, 0.2808759765625, 0.2811085205078125, 0.2810316467285156, 0.28102349853515624, 0.2808760375976562, 0.6096005249023437, 0.28075726318359373, 0.2809169921875, 0.28073370361328126, 0.2812037048339844, 0.2810357666015625, 0.2810040283203125, 0.2807848815917969, 0.281017333984375, 0.2811248779296875, 0.28101223754882815, 0.2807848815917969, 0.28090673828125, 0.28089547729492187, 0.2810491027832031, 0.28088729858398437, 0.2809088134765625, 0.28079617309570315, 0.28105523681640626, 0.28090060424804686, 0.28096820068359374, 0.28085861206054685, 0.28080230712890625, 0.280880126953125, 0.28098355102539063, 0.2810572814941406, 0.2831247253417969, 0.281059326171875, 0.2810521545410156, 0.28110848999023436, 0.2811576232910156, 0.2811760559082031, 0.2811463623046875, 0.280911865234375, 0.2809190368652344, 0.2809876403808594, 0.28097024536132814, 0.2809886779785156, 0.2810624084472656, 0.28105523681640626, 0.28116378784179685, 0.28104193115234377, 0.28093646240234377, 0.2808760375976562, 0.28102859497070315, 0.28101837158203125, 0.2808401794433594, 0.28090573120117185, 0.28106854248046875, 0.28096820068359374, 0.2809661560058594, 0.2807940979003906, 0.28126412963867187, 0.28086273193359373, 0.28110540771484377, 0.2811617431640625, 0.281143310546875, 0.28087295532226564, 0.28085043334960935, 0.2810019836425781, 0.2811822204589844, 0.28111871337890626, 0.28107366943359374]",tokens/s,3.5002427368506384,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11626,9 +11626,9 @@ ChildProcessError: Traceback (most recent call last): return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1736, in generate + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1758, in generate result = self._sample( - File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2375, in _sample + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2397, in _sample outputs = self( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) @@ -11664,7 +11664,7 @@ RuntimeError: shape '[-1, 128, 4672]' is invalid for input of size 21229568 ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4708.687872,21498.42944,0.0,20851.982336,20236.731392,s,10,15.782414306640625,1.5782414306640624,0.0005368111432327376,1.5781434326171875,1.5784585083007812,1.5790858825683594,1.579587781982422,"[1.578103759765625, 1.5775478515625, 1.5778953857421876, 1.5797132568359376, 1.57818310546875, 1.5782744140625, 1.57828564453125, 1.578103759765625, 1.578319091796875, 1.577988037109375]",tokens/s,162.20585458353173,kWh,1.8640190545055602e-05,1.0214868406037568e-05,9.259768518919964e-05,0.0001214527441402928,tokens/kWh,2107815.6925321394,MB,4708.687872,21498.42944,0.0,20851.982336,20339.707392,s,10,881.7457578125,88.17457578125,0.002115503395001207,88.17539453124999,88.17674609375,88.17697460937501,88.17715742187501,"[88.17125, 88.17175, 88.1749609375, 88.175828125, 88.177203125, 88.175890625, 88.1766953125, 88.173328125, 88.176578125, 88.1722734375]",tokens/s,0.714491671117251,kWh,0.001040854390843047,0.0005704774625729078,0.005178987532075602,0.006790319385491556,tokens/kWh,9277.914104395162,,s,629,894.5504398193359,1.4221787596491826,0.18751446646952302,1.3995601806640625,1.4000109375,1.4001393798828126,2.978033017578125,"[1.399194580078125, 1.399298095703125, 1.3991290283203126, 1.3992857666015626, 1.3990655517578126, 1.399204833984375, 1.3991751708984375, 1.3994710693359376, 1.3990738525390625, 1.3990277099609374, 1.39915771484375, 1.399203857421875, 1.3990020751953125, 1.39932470703125, 1.39911376953125, 1.3995806884765625, 1.399319580078125, 1.3996319580078125, 1.399352294921875, 1.3991802978515624, 1.399151611328125, 1.39943017578125, 1.3990972900390626, 1.3995540771484376, 1.3996278076171875, 1.399510009765625, 1.39930322265625, 1.3995018310546874, 1.399568359375, 1.39934619140625, 1.399161865234375, 1.3999493408203125, 1.3994700927734376, 1.399572509765625, 1.39960107421875, 1.399762939453125, 1.3996871337890624, 1.399688232421875, 1.3997169189453125, 1.4000067138671874, 1.399546875, 1.399795654296875, 1.399826416015625, 1.3999605712890626, 1.3998192138671874, 1.399867431640625, 1.399826416015625, 1.399509033203125, 1.399413818359375, 1.3999349365234375, 1.399510986328125, 1.3996646728515625, 1.399531494140625, 1.4000762939453124, 1.3997393798828126, 1.40002099609375, 1.39964208984375, 1.39989306640625, 1.399784423828125, 1.4000015869140625, 1.4001776123046874, 1.4001162109375, 2.977680419921875, 1.3987635498046875, 1.3990216064453125, 1.398865966796875, 1.399183349609375, 1.3990440673828124, 1.399251953125, 1.399235595703125, 1.399419921875, 1.3989498291015625, 1.3994588623046875, 1.3995325927734374, 1.39915576171875, 1.399416748046875, 1.3992857666015626, 1.399309326171875, 1.3996400146484376, 1.39901025390625, 1.3994168701171874, 1.39928369140625, 1.3992008056640626, 1.399120849609375, 1.39950390625, 1.3994332275390624, 1.3995540771484376, 1.399593017578125, 1.3996851806640624, 1.3992059326171875, 1.3998223876953124, 1.3995396728515626, 1.399499755859375, 1.3992294921875, 1.399562255859375, 1.399331787109375, 1.3999698486328125, 1.39949365234375, 1.3995172119140624, 1.3994915771484375, 1.399731201171875, 1.3996185302734374, 1.399804931640625, 1.3992376708984375, 1.3995550537109376, 1.399384033203125, 1.39986328125, 1.3994813232421874, 1.401822265625, 1.399848876953125, 1.400058837890625, 1.399477294921875, 1.399910400390625, 1.3995653076171874, 1.399931884765625, 1.3996185302734374, 1.399846923828125, 1.399761962890625, 1.3999759521484374, 1.3998602294921876, 1.3998929443359376, 1.39964208984375, 1.4000107421875, 1.399794677734375, 1.4000025634765625, 2.978083740234375, 1.398897705078125, 1.3991322021484376, 1.3995980224609375, 1.3994290771484375, 1.399194580078125, 1.399624755859375, 1.399552001953125, 1.3993800048828124, 1.3990799560546876, 1.399405517578125, 1.399300048828125, 1.3992899169921875, 1.3992335205078126, 1.39951513671875, 1.3993482666015624, 1.399700439453125, 1.399189453125, 1.3992960205078124, 1.3992008056640626, 1.3995550537109376, 1.399152587890625, 1.3994906005859375, 1.3997373046875, 1.399525390625, 1.3993953857421875, 1.3996871337890624, 1.3994229736328125, 1.39953662109375, 1.39951513671875, 1.3995233154296876, 1.3995028076171876, 1.399846923828125, 1.399299072265625, 1.3997659912109375, 1.3994556884765625, 1.3997015380859374, 1.39947412109375, 1.3996082763671875, 1.39980078125, 1.3997916259765626, 1.3995919189453125, 1.39965234375, 1.3994854736328124, 1.39974755859375, 1.400369140625, 1.3997813720703125, 1.3998275146484376, 1.3997486572265625, 1.3997486572265625, 1.39992578125, 1.3999349365234375, 1.3997034912109374, 1.3998878173828124, 1.39997900390625, 1.400004638671875, 1.4001182861328125, 1.399736328125, 1.400037353515625, 1.3998233642578124, 1.399794677734375, 1.399687255859375, 1.400122314453125, 2.979734619140625, 1.3991536865234375, 1.399083984375, 1.3987276611328125, 1.39987255859375, 1.3991597900390624, 1.3992386474609375, 1.399719970703125, 1.3995755615234375, 1.3991290283203126, 1.3991075439453124, 1.39930419921875, 1.3992325439453126, 1.3994864501953126, 1.39919775390625, 1.39900830078125, 1.399330810546875, 1.3992427978515625, 1.39966357421875, 1.39919873046875, 1.39934619140625, 1.39936767578125, 1.3994219970703126, 1.399363525390625, 1.3997291259765625, 1.3998028564453124, 1.399531494140625, 1.3992847900390626, 1.399729248046875, 1.3994423828125, 1.399742431640625, 1.3993021240234376, 1.39958984375, 1.3994332275390624, 1.399636962890625, 1.3993062744140625, 1.399445556640625, 1.3994761962890625, 1.399825439453125, 1.3996768798828125, 1.399867431640625, 1.3995806884765625, 1.3999698486328125, 1.3994281005859375, 1.4019144287109375, 1.39976806640625, 1.400162353515625, 1.3998765869140626, 1.3997568359375, 1.3996256103515625, 1.399816162109375, 1.3998919677734376, 1.399845947265625, 1.3995601806640625, 1.3998387451171874, 1.4001397705078125, 1.400004638671875, 1.4000650634765626, 1.40010693359375, 1.39981103515625, 1.3998050537109374, 1.3996778564453125, 1.4001243896484374, 2.97879443359375, 1.399288818359375, 1.3990625, 1.3989530029296875, 1.3993800048828124, 1.3989437255859376, 1.399468017578125, 1.39931640625, 1.3996451416015625, 1.3991168212890626, 1.39941064453125, 1.399319580078125, 1.39913525390625, 1.3994024658203126, 1.399150634765625, 1.3994117431640625, 1.399594970703125, 1.3994075927734375, 1.399793701171875, 1.399268310546875, 1.3994659423828124, 1.3996287841796875, 1.3994556884765625, 1.3993697509765626, 1.3996676025390624, 1.399447509765625, 1.3997608642578125, 1.3995079345703125, 1.399666748046875, 1.3995478515625, 1.3996953125, 1.399482421875, 1.3994608154296875, 1.3996329345703125, 1.3997855224609375, 1.39945166015625, 1.3995386962890626, 1.39945166015625, 1.3997578125, 1.3999656982421875, 1.4002125244140624, 1.3995765380859375, 1.4000814208984376, 1.3997659912109375, 1.39985302734375, 1.399521240234375, 1.399901123046875, 1.39972705078125, 1.400004638671875, 1.399415771484375, 1.3997598876953126, 1.399835693359375, 1.3999913330078124, 1.399677001953125, 1.4000362548828125, 1.4001612548828124, 1.4003128662109374, 1.39993603515625, 1.400158203125, 1.4000721435546875, 1.4003721923828125, 1.399899169921875, 1.3997158203125, 2.978334716796875, 1.39909326171875, 1.3990604248046874, 1.39932568359375, 1.399351318359375, 1.39915771484375, 1.3992540283203125, 1.3992642822265624, 1.399413818359375, 1.3991741943359375, 1.3993359375, 1.399224365234375, 1.3994229736328125, 1.39940966796875, 1.399456787109375, 1.39919873046875, 1.3996123046875, 1.3993431396484375, 1.3997291259765625, 1.3994117431640625, 1.3995386962890626, 1.3994404296875, 1.3996083984375, 1.399300048828125, 1.3998018798828125, 1.4001162109375, 1.399901123046875, 1.3998970947265625, 1.399531494140625, 1.399182373046875, 1.3994619140625, 1.399109619140625, 1.3996697998046874, 1.3993175048828126, 1.399574462890625, 1.3992847900390626, 1.39966162109375, 1.399520263671875, 1.401324462890625, 1.399489501953125, 1.399690185546875, 1.3993994140625, 1.399900146484375, 1.39951611328125, 1.399921630859375, 1.399678955078125, 1.4001673583984375, 1.39990625, 1.3998018798828125, 1.3995079345703125, 1.400015869140625, 1.3994803466796875, 1.399951416015625, 1.3997076416015626, 1.40027294921875, 1.3999246826171876, 1.3999246826171876, 1.399732177734375, 1.3998970947265625, 1.399783447265625, 1.4001387939453125, 1.3993502197265626, 1.4002431640625, 2.9787392578125, 1.398930419921875, 1.3988720703125, 1.3989642333984376, 1.3989478759765626, 1.3991597900390624, 1.3990850830078125, 1.400115234375, 1.3992796630859374, 1.3990533447265625, 1.39926220703125, 1.3995345458984374, 1.39920703125, 1.3993245849609375, 1.399505859375, 1.39932666015625, 1.3994383544921876, 1.399141357421875, 1.3998828125, 1.39945166015625, 1.3994423828125, 1.39908203125, 1.399351318359375, 1.401248779296875, 1.3996380615234374, 1.399615478515625, 1.3997967529296875, 1.3995499267578124, 1.399447509765625, 1.3995377197265626, 1.39965234375, 1.3995089111328125, 1.3995694580078124, 1.39930419921875, 1.3994619140625, 1.399435302734375, 1.399462890625, 1.3994710693359376, 1.39966259765625, 1.3994649658203124, 1.3999195556640625, 1.399552001953125, 1.399803955078125, 1.3996103515625, 1.399816162109375, 1.39976708984375, 1.400056884765625, 1.399909423828125, 1.39981103515625, 1.3995601806640625, 1.3999759521484374, 1.399825439453125, 1.4004111328125, 1.399698486328125, 1.3998018798828125, 1.4003282470703124, 1.4000977783203126, 1.4000067138671874, 1.399920654296875, 1.3997823486328125, 1.400100830078125, 1.3995079345703125, 1.4000343017578125, 2.979209228515625, 1.3990235595703124, 1.398892578125, 1.3989365234375, 1.39919970703125, 1.3989918212890624, 1.3991229248046875, 1.3992652587890626, 1.3993604736328125, 1.3993175048828126, 1.3992017822265626, 1.39951416015625, 1.3994075927734375, 1.399320556640625, 1.39930419921875, 1.39944140625, 1.4017669677734375, 1.3992376708984375, 1.399351318359375, 1.3993123779296874, 1.3993021240234376, 1.39919873046875, 1.399609375, 1.399462890625, 1.399446533203125, 1.3993922119140625, 1.3996553955078126, 1.3996207275390624, 1.399635986328125, 1.399510986328125, 1.3993389892578125, 1.3993994140625, 1.39949462890625, 1.3993338623046876, 1.399525390625, 1.39961962890625, 1.3995970458984375, 1.3995335693359374, 1.39974658203125, 1.3998438720703126, 1.4000506591796875, 1.3993870849609376, 1.40001171875, 1.3994864501953126, 1.3999493408203125, 1.39968408203125, 1.39966162109375, 1.399667724609375, 1.3997752685546876, 1.3997281494140625, 1.3997158203125, 1.399551025390625, 1.39983154296875, 1.39955810546875, 1.4000404052734374, 1.39997998046875, 1.3996800537109375, 1.3997813720703125, 1.3998623046875, 1.3999267578125, 1.3998714599609374, 1.399816162109375, 1.3999677734375, 2.9783203125, 1.3990379638671875, 1.398981689453125, 1.3989110107421876, 1.3995018310546874, 1.399352294921875, 1.39950390625, 1.399098388671875, 1.3995130615234375, 1.399405517578125, 1.3993851318359376, 1.3991854248046875, 1.39917724609375, 1.399246826171875, 1.3993748779296875, 1.3995396728515626, 1.399609375, 1.401608154296875, 1.39957763671875, 1.399351318359375, 1.3994659423828124, 1.3992908935546875, 1.39940966796875, 1.399413818359375, 1.3994700927734376, 1.3996348876953124, 1.3998448486328126, 1.3994188232421876, 1.3997701416015624, 1.3996114501953125, 1.3994290771484375, 1.39945166015625, 1.3996492919921875, 1.39970458984375, 1.39961865234375, 1.3996348876953124, 1.39974755859375, 1.399419921875, 1.3996390380859376, 1.399568359375, 1.40006396484375, 1.3995345458984374, 1.399699462890625, 1.39949462890625, 1.3997711181640624, 1.39957763671875, 1.400057861328125, 1.399718994140625, 1.3997783203125, 1.3994967041015625, 1.3998909912109374, 1.399602294921875, 1.3997977294921875, 1.3996103515625, 1.4001929931640624, 1.400183837890625, 1.4000916748046874, 1.3998070068359374, 1.400088623046875, 1.3998140869140625, 1.399877685546875, 1.3998448486328126, 1.4000291748046876, 2.977902587890625, 1.3989754638671874, 1.3993359375, 1.399014404296875, 1.39926123046875, 1.399205810546875, 1.3993922119140625, 1.39930322265625, 1.39965234375, 1.3993604736328125, 1.3991485595703126, 1.3991854248046875, 1.3994864501953126, 1.399546875, 1.399447509765625, 1.3990963134765626, 1.3993790283203125, 1.399372802734375, 1.399635986328125, 1.3994639892578125, 1.399376953125, 1.398993896484375, 1.399372802734375, 1.3994066162109375, 1.3993277587890625, 1.3993963623046874, 1.3998489990234375, 1.3994404296875, 1.3996298828125, 1.3993543701171876, 1.3996400146484376, 1.3993819580078124, 1.399593994140625, 1.39938818359375, 1.399609375, 1.3995673828125, 1.399614501953125, 1.399294921875, 1.39947412109375, 1.399741455078125, 1.399953369140625, 1.3997691650390625, 1.3998253173828126, 1.399690185546875, 1.399609375, 1.3995694580078124, 1.3996502685546874, 1.3995181884765624, 1.3998612060546876, 1.3997015380859374, 1.399868408203125, 1.399719970703125, 1.3997659912109375, 1.399816162109375, 1.400052734375, 1.3999656982421875, 1.399974853515625, 1.39974755859375, 1.4001602783203124, 1.4000721435546875, 1.3998294677734375, 1.399690185546875, 1.3998980712890625]",tokens/s,0.7031464878906475,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -11693,7 +11693,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4aa7-34acad0a1b0bd2f218242c91;bcc119f6-4e32-4f94-9c96-76fa966e660c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664feb20-20efae812bef9c9442d2c463;04f35eda-f375-4e7d-9041-08f778a7fb83) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -11778,7 +11778,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9654-7ded8777074b471543be2ca4;94821127-18f2-401e-bc3e-a0a9002bc47b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe823-71dd0fbc5bb128851306eebb;a083bdc2-ffc0-401c-8b28-e099c7b7ac89) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11840,7 +11840,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e909c-20a22c86614ba11c6bf9e364;5800029a-0c61-4428-9981-4a9925bc24c7) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe245-79e265ef6b795f99535dbd75;98eb1366-8f9b-4ad5-859f-db491d63110c) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -11967,7 +11967,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9600-6e5328ba106f12e12f8ff5d3;b47c0984-6391-4685-bb9c-317ffade0210) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe7cf-0427125d2489091e17d619e7;b346b7c0-2053-45f8-82eb-600a6ab9637b) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12040,7 +12040,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9116-4f948886668dccee53ced947;33cf9820-9781-49d0-92e6-46abc2497e62) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe2c7-695a031654b00ee648c51200;e875a592-390b-4614-a9fd-249377a7c3a4) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12074,7 +12074,7 @@ If this is a private repository, make sure to pass a token having permission to 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-410m,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1354.99776,1313.34144,0.0,666.894336,583.411712,s,10,0.367520923614502,0.036752092361450196,0.0018078330313943103,0.03598340797424317,0.037524640655517576,0.03974108772277831,0.04151424537658691,"[0.04195753479003906, 0.03595008087158203, 0.035949630737304686, 0.0360167350769043, 0.035681278228759765, 0.0355464973449707, 0.03703209686279297, 0.0357196159362793, 0.036844062805175784, 0.03682339096069336]",tokens/s,6965.589808664122,kWh,4.4016995025242427e-07,2.4119189856989944e-07,1.376317705761802e-06,2.0576795545841257e-06,tokens/kWh,124411986.03041947,MB,1354.99776,1313.34144,0.0,666.894336,644.020224,s,10,22.070285400390624,2.2070285400390626,0.012889624369515166,2.207716430664062,2.217943408203125,2.2274182373046876,2.2349981005859374,"[2.215837890625, 2.19123291015625, 2.190115966796875, 2.2071044921875, 2.196080810546875, 2.203771240234375, 2.21104345703125, 2.209877197265625, 2.208328369140625, 2.23689306640625]",tokens/s,28.54516779329231,kWh,2.7174306134456744e-05,1.489238295772977e-05,7.386624248823808e-05,0.0001159329315804246,tokens/kWh,543417.6393296486,,s,629,22.359374866485588,0.035547495813172646,0.004369599305993048,0.03485388946533203,0.035935641479492186,0.03633192977905273,0.0706462094116211,"[0.035517440795898435, 0.03587686538696289, 0.035958782196044925, 0.036241409301757815, 0.03552358245849609, 0.03649740982055664, 0.036086784362792966, 0.035787841796875, 0.03611334228515625, 0.035814399719238284, 0.036670463562011715, 0.03558399963378906, 0.03552460861206055, 0.03667763137817383, 0.03713945770263672, 0.03656499099731445, 0.035566593170166014, 0.03501260757446289, 0.03496345520019531, 0.03650969696044922, 0.03527679824829102, 0.034767871856689454, 0.03477913665771484, 0.03477196884155274, 0.03492454528808594, 0.0348671989440918, 0.034900993347167966, 0.03502796936035156, 0.034513919830322266, 0.0346429443359375, 0.03457024002075195, 0.0347883529663086, 0.034560001373291016, 0.03474943923950195, 0.03465420913696289, 0.034563072204589845, 0.03450982284545898, 0.0343111686706543, 0.034579456329345705, 0.03442483139038086, 0.034683902740478514, 0.03441356658935547, 0.034772991180419925, 0.034407424926757815, 0.035517440795898435, 0.03611238479614258, 0.03475046539306641, 0.034597888946533206, 0.034735103607177735, 0.03493081665039063, 0.03529203033447265, 0.0354785270690918, 0.034852863311767575, 0.034748416900634765, 0.03462963104248047, 0.034566143035888675, 0.03419750213623047, 0.03429683303833008, 0.03449958419799805, 0.03483647918701172, 0.034385921478271485, 0.034536449432373044, 0.07078399658203124, 0.03447091293334961, 0.03465625762939453, 0.03450470352172851, 0.034754558563232424, 0.03484467315673828, 0.03481702423095703, 0.03438694381713867, 0.034683902740478514, 0.03473715209960938, 0.03455692672729492, 0.03723980712890625, 0.03523583984375, 0.035286014556884765, 0.034473983764648435, 0.0345456657409668, 0.034359294891357424, 0.034477054595947264, 0.03466239929199219, 0.03511808013916016, 0.03498291015625, 0.034813953399658204, 0.03462662506103516, 0.03509958267211914, 0.03489996719360351, 0.03486207962036133, 0.034920448303222655, 0.034597888946533206, 0.03445862579345703, 0.03462246322631836, 0.0344268798828125, 0.03480678558349609, 0.03458969497680664, 0.034900993347167966, 0.034776065826416014, 0.034786304473876956, 0.034699264526367186, 0.03467366409301758, 0.03476889419555664, 0.03478220748901367, 0.03483750534057617, 0.034511871337890625, 0.03454771041870117, 0.03439513778686523, 0.034931713104248044, 0.03477811050415039, 0.034948097229003904, 0.034587646484375, 0.03447296142578125, 0.03463065719604492, 0.034923519134521484, 0.035004417419433595, 0.0352174072265625, 0.03488460922241211, 0.034713600158691404, 0.034785472869873046, 0.03485164642333984, 0.034582527160644534, 0.03461324691772461, 0.034735103607177735, 0.03479142379760742, 0.03497881698608398, 0.03479654312133789, 0.07093452453613282, 0.03465216064453125, 0.0348221435546875, 0.03531161499023437, 0.03479449462890625, 0.034871295928955076, 0.03495219039916992, 0.034723838806152346, 0.03477401733398437, 0.0347064323425293, 0.03492454528808594, 0.034536449432373044, 0.03510784149169922, 0.03465420913696289, 0.03480684661865235, 0.03470124816894531, 0.03458371353149414, 0.03471139144897461, 0.03486003112792969, 0.034661376953125, 0.03493798446655273, 0.03478515243530273, 0.034732032775878906, 0.03485184097290039, 0.03488870239257812, 0.03453235244750977, 0.03489382553100586, 0.03434188842773438, 0.034872318267822264, 0.03492454528808594, 0.034148353576660156, 0.03449651336669922, 0.034813953399658204, 0.03492966461181641, 0.034841598510742186, 0.03460300827026367, 0.034508800506591795, 0.03467264175415039, 0.03484262466430664, 0.034628608703613284, 0.03479859161376953, 0.03498700714111328, 0.034651294708251956, 0.03462236785888672, 0.034640830993652345, 0.03490611267089844, 0.03469823837280273, 0.034902015686035154, 0.034928638458251955, 0.034651294708251956, 0.03462025451660156, 0.03465932846069336, 0.03507712173461914, 0.03481804656982422, 0.03491020965576172, 0.03464089584350586, 0.03476275253295898, 0.035087390899658205, 0.034976734161376956, 0.03449958419799805, 0.03471257781982422, 0.034669567108154296, 0.0345272331237793, 0.071478271484375, 0.034956287384033204, 0.034609153747558595, 0.036604927062988284, 0.037515262603759765, 0.035896320343017575, 0.03503411102294922, 0.03479347229003906, 0.03482828903198242, 0.03454873657226563, 0.03466342544555664, 0.03487334442138672, 0.03464396667480469, 0.034988033294677735, 0.03512319946289062, 0.034746368408203124, 0.03474124908447265, 0.034568286895751955, 0.034476959228515625, 0.03470963287353516, 0.03468991851806641, 0.03481190490722656, 0.034735103607177735, 0.03486617660522461, 0.03487744140625, 0.034931713104248044, 0.035102718353271486, 0.034764801025390625, 0.034582527160644534, 0.03495219039916992, 0.03488870239257812, 0.03491328048706055, 0.03520204925537109, 0.0358809585571289, 0.03488051223754883, 0.03489996719360351, 0.034969600677490234, 0.03485081481933594, 0.03455187225341797, 0.03485996627807617, 0.03504127883911133, 0.036961280822753906, 0.03565055847167969, 0.0350832633972168, 0.034813953399658204, 0.034816001892089846, 0.034685951232910156, 0.034516990661621096, 0.03474534225463867, 0.03615027236938476, 0.03591372680664062, 0.03477196884155274, 0.034917377471923826, 0.034800640106201174, 0.03520204925537109, 0.03498700714111328, 0.03488051223754883, 0.03541401672363281, 0.034797569274902344, 0.034551807403564457, 0.03476377487182617, 0.03484262466430664, 0.03485081481933594, 0.07039282989501953, 0.03467059326171875, 0.03469619369506836, 0.034702335357666016, 0.0346060791015625, 0.0348221435546875, 0.03482217788696289, 0.03479548645019531, 0.03466649627685547, 0.034697216033935545, 0.03539763259887695, 0.035337215423583986, 0.034797569274902344, 0.03480166244506836, 0.03540070343017578, 0.03477503967285156, 0.03475763320922851, 0.034759681701660154, 0.03535257720947266, 0.03613798522949219, 0.035148799896240236, 0.034947071075439456, 0.03492659378051758, 0.03488665771484375, 0.03466239929199219, 0.034715648651123046, 0.034902015686035154, 0.03509145736694336, 0.034800640106201174, 0.03488358306884766, 0.03472281646728516, 0.034729984283447264, 0.03480473709106445, 0.03463679885864258, 0.035009536743164066, 0.03467366409301758, 0.03496448135375976, 0.034678783416748044, 0.034909183502197266, 0.034754558563232424, 0.034869312286376956, 0.03521017456054688, 0.03491123199462891, 0.03483443069458008, 0.034644992828369144, 0.034625537872314455, 0.03480883026123047, 0.035149822235107424, 0.03476070404052734, 0.034825214385986326, 0.035009536743164066, 0.03492454528808594, 0.034746368408203124, 0.034784255981445314, 0.03487948989868164, 0.03450368118286133, 0.03453459167480469, 0.03429254531860351, 0.03462758255004883, 0.034925567626953126, 0.03499724960327148, 0.03464908981323242, 0.035037185668945314, 0.07064678192138672, 0.03441459274291992, 0.034694145202636716, 0.03461734390258789, 0.03445248031616211, 0.03491952133178711, 0.03476166534423828, 0.0344227523803711, 0.034531326293945314, 0.03464191818237305, 0.03439820861816406, 0.03489894485473633, 0.03634380722045898, 0.03576115036010742, 0.03449446487426758, 0.03520307159423828, 0.03449139022827148, 0.03478937530517578, 0.035342334747314456, 0.035163135528564454, 0.034618400573730466, 0.034358238220214846, 0.034702335357666016, 0.03500032043457031, 0.034753631591796875, 0.0347421760559082, 0.03495423889160156, 0.035460094451904296, 0.03466342544555664, 0.03442483139038086, 0.03458457565307617, 0.03460505676269531, 0.03462041473388672, 0.03481292724609375, 0.03470131301879883, 0.03467264175415039, 0.03499728012084961, 0.03446780776977539, 0.03447609710693359, 0.03452105712890625, 0.03455382537841797, 0.034874366760253905, 0.0379422721862793, 0.03952230453491211, 0.036410369873046876, 0.03606425476074219, 0.035922943115234376, 0.03479859161376953, 0.03526860809326172, 0.035707904815673826, 0.034740222930908206, 0.03471155166625976, 0.03473920059204102, 0.034819072723388675, 0.034661376953125, 0.0346951675415039, 0.03463270568847656, 0.034637889862060546, 0.03448416137695313, 0.03467366409301758, 0.034648063659667966, 0.03489279937744141, 0.03479865646362305, 0.07351289367675781, 0.03600896072387695, 0.035197952270507815, 0.03555123138427734, 0.034988094329833984, 0.03467871856689453, 0.03475763320922851, 0.03546121597290039, 0.034708385467529294, 0.03477196884155274, 0.03465932846069336, 0.03476486587524414, 0.03479750442504883, 0.03556249618530274, 0.03582156753540039, 0.03582880020141602, 0.035523521423339845, 0.03475251388549805, 0.03458358383178711, 0.034864097595214846, 0.035884033203125, 0.03502489471435547, 0.03483955383300781, 0.03487334442138672, 0.03500543975830078, 0.035759105682373046, 0.03461324691772461, 0.03523481750488281, 0.034956287384033204, 0.03481190490722656, 0.03480883026123047, 0.034560001373291016, 0.03494911956787109, 0.03489382553100586, 0.03491635131835937, 0.03486617660522461, 0.035081214904785156, 0.034776065826416014, 0.034514942169189454, 0.034531326293945314, 0.03461734390258789, 0.03551846313476562, 0.035914752960205076, 0.035579902648925785, 0.03478220748901367, 0.035248126983642575, 0.03557580947875977, 0.034890750885009765, 0.034685951232910156, 0.0346429443359375, 0.03615334320068359, 0.035279006958007814, 0.034461536407470704, 0.03487846374511719, 0.03565465545654297, 0.03600384140014649, 0.03564851379394531, 0.03565478515625, 0.03515276718139648, 0.03466547012329101, 0.03437875366210937, 0.03461750411987305, 0.0344431037902832, 0.07064473724365235, 0.0347147216796875, 0.034790302276611326, 0.03484377670288086, 0.03461516952514648, 0.034869247436523435, 0.03594137573242188, 0.034958335876464845, 0.03640115356445312, 0.03510476684570313, 0.03486105728149414, 0.03501055908203125, 0.03485388946533203, 0.034310142517089845, 0.03443916702270508, 0.03494911956787109, 0.03459481430053711, 0.03470438385009766, 0.03486617660522461, 0.034344959259033206, 0.03521843338012695, 0.03473715209960938, 0.03472588729858399, 0.03449446487426758, 0.034948097229003904, 0.03617996978759765, 0.03462656021118164, 0.0347064323425293, 0.034874366760253905, 0.03575807952880859, 0.03489177703857422, 0.03488051223754883, 0.034956287384033204, 0.03717324829101563, 0.036507713317871095, 0.03579283142089844, 0.03566592025756836, 0.035716094970703126, 0.03486617660522461, 0.034557952880859374, 0.034680831909179685, 0.035053569793701174, 0.03573657608032227, 0.03503923034667969, 0.03496857452392578, 0.03485184097290039, 0.035326976776123044, 0.03513753509521484, 0.03506380844116211, 0.036201473236083984, 0.034890750885009765, 0.034939903259277344, 0.034902015686035154, 0.035476478576660156, 0.0344002571105957, 0.03499008178710938, 0.03501260757446289, 0.03476688003540039, 0.03490915298461914, 0.0347770881652832, 0.034776065826416014, 0.034857982635498046, 0.03577036666870117, 0.07236412811279297, 0.03441961669921875, 0.034358272552490236, 0.03481497573852539, 0.03492659378051758, 0.0358287353515625, 0.0354703369140625, 0.035351551055908204, 0.03686297607421875, 0.03505152130126953, 0.03466239929199219, 0.0353361930847168, 0.035487743377685545, 0.035620864868164064, 0.03545600128173828, 0.03588608169555664, 0.034941951751708986, 0.03483340835571289, 0.034923519134521484, 0.03481087875366211, 0.034938880920410156, 0.03464089584350586, 0.03494911956787109, 0.034764801025390625, 0.03468288040161133, 0.03474537658691406, 0.034764766693115234, 0.035383296966552735, 0.03480678558349609, 0.03476377487182617, 0.034338878631591796, 0.03467871856689453, 0.03476377487182617, 0.034759681701660154, 0.03488870239257812, 0.034748416900634765, 0.03457740783691406, 0.034948097229003904, 0.03500339126586914, 0.03466035079956055, 0.034597888946533206, 0.03524505615234375, 0.03553484725952148, 0.03542835235595703, 0.03479244613647461, 0.03461529541015625, 0.035593215942382815, 0.03561065673828125, 0.03489891052246094, 0.035195903778076174, 0.03493580627441406, 0.034948097229003904, 0.034941951751708986, 0.03497267150878906, 0.034748416900634765, 0.035342334747314456, 0.036103168487548826, 0.03496550369262695, 0.035093505859375, 0.03604991912841797, 0.034797569274902344, 0.03469823837280273, 0.03488256072998047, 0.07262105560302734, 0.03481190490722656, 0.034570270538330075, 0.03525526428222656, 0.03586457443237305, 0.03491839981079101, 0.03548262405395508, 0.03534131240844727, 0.034533374786376955, 0.03462041473388672, 0.03450582504272461, 0.03459779357910156, 0.03499008178710938, 0.034854911804199216, 0.03489791870117188, 0.034700286865234374, 0.03567718505859375, 0.03510067367553711, 0.034855934143066404, 0.03494604873657227, 0.035383296966552735, 0.03561471939086914, 0.035422206878662106, 0.03482624053955078, 0.03482931137084961, 0.03605408096313477, 0.03592902374267578, 0.035507198333740234, 0.03497369766235352, 0.03558911895751953, 0.03528806304931641, 0.035912704467773435, 0.03631411361694336, 0.03609292984008789, 0.03615129470825195, 0.03599871826171875, 0.035932159423828124, 0.035958782196044925, 0.035934207916259765, 0.03591987228393555, 0.03645542526245117, 0.03650252914428711, 0.03688243103027344, 0.03597107315063477, 0.03595161437988281, 0.035606529235839846, 0.035883007049560545, 0.03639807891845703, 0.035996673583984375, 0.0357314567565918, 0.03626803207397461, 0.0361451530456543, 0.03608473587036133, 0.03609190368652344, 0.0363765754699707, 0.03603148651123047, 0.03598643112182617, 0.03509964752197266, 0.03460403060913086, 0.03459686279296875, 0.034941951751708986, 0.03470131301879883, 0.03467366409301758]",tokens/s,28.131376827659274,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-32B,Qwen/Qwen1.5-32B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.0,,,,,,MB,4191.223808,23096.459264,0.0,22450.01216,21127.156736,s,10,14.774497192382812,1.4774497192382814,0.00380529101888081,1.4756153564453127,1.4841980590820312,1.4849311218261718,1.4855175720214844,"[1.4749822998046875, 1.48403515625, 1.47480615234375, 1.4750469970703124, 1.475837158203125, 1.4753935546875, 1.4776461181640625, 1.4762384033203124, 1.47484716796875, 1.4856641845703125]",tokens/s,173.27154803750895,kWh,1.7430581467019188e-05,9.549013355408533e-06,8.510498475060024e-05,0.00011208457957302798,tokens/kWh,2283989.4745129044,MB,4201.55392,23096.459264,0.0,22450.01216,21326.312448,s,10,819.6256249999999,81.96256249999999,0.004219943407593987,81.96333984374999,81.9675484375,81.9680828125,81.9685103125,"[81.95671875, 81.9641796875, 81.96696875, 81.9674296875, 81.9686171875, 81.9638515625, 81.962828125, 81.9574609375, 81.9571640625, 81.96040625]",tokens/s,0.7686436109168745,kWh,0.0009675532748632961,0.0005303076524099196,0.004793837501733597,0.006291698429006812,tokens/kWh,10013.194483312987,,s,629,831.6245739746099,1.3221376374795062,0.1755437916223313,1.3009254150390626,1.301338330078125,1.301497021484375,2.776785390625,"[1.300578369140625, 1.30064990234375, 1.300453369140625, 1.30045947265625, 1.300621337890625, 1.3005609130859375, 1.3005947265625, 1.3008026123046874, 1.300737060546875, 1.3007452392578125, 1.3005762939453125, 1.3007431640625, 1.3007750244140626, 1.30079638671875, 1.3004246826171875, 1.300674560546875, 1.3005987548828124, 1.3006929931640625, 1.3005220947265625, 1.3009959716796875, 1.300706298828125, 1.3009305419921875, 1.3006796875, 1.30077392578125, 1.3009244384765626, 1.300610107421875, 1.300779052734375, 1.3007779541015625, 1.3009254150390626, 1.30064794921875, 1.3007431640625, 1.3007021484375, 1.3008701171875, 1.3007677001953124, 1.3007923583984375, 1.300869140625, 1.301054443359375, 1.3012991943359375, 1.3009049072265626, 1.3009500732421875, 1.30066943359375, 1.3013421630859374, 1.3010667724609375, 1.3009111328125, 1.3009981689453125, 1.3011220703125, 1.3011927490234374, 1.301360595703125, 1.3012203369140625, 1.30134423828125, 1.3008271484375, 1.301201904296875, 1.301221435546875, 1.301244873046875, 1.3010523681640624, 1.300990966796875, 1.30117626953125, 1.301338134765625, 1.301137451171875, 1.30153271484375, 1.30126025390625, 1.301770263671875, 2.78496142578125, 1.30066943359375, 1.3004676513671876, 1.30016357421875, 1.30035302734375, 1.300717529296875, 1.30077490234375, 1.3006888427734375, 1.3003837890625, 1.3006787109375, 1.3007401123046876, 1.3006304931640624, 1.3006602783203125, 1.300515869140625, 1.30075439453125, 1.3006663818359374, 1.3005926513671875, 1.3008056640625, 1.3007545166015626, 1.300760498046875, 1.300926513671875, 1.3009305419921875, 1.30094287109375, 1.3006510009765626, 1.3011456298828126, 1.301042236328125, 1.30083740234375, 1.3008670654296874, 1.300885498046875, 1.3014405517578125, 1.3009869384765624, 1.3007135009765625, 1.3006090087890625, 1.300738037109375, 1.30071142578125, 1.3009888916015624, 1.3008455810546875, 1.3008138427734375, 1.30094287109375, 1.301116943359375, 1.3008712158203124, 1.3008486328125, 1.30132177734375, 1.3012674560546875, 1.30102880859375, 1.301021728515625, 1.3059317626953124, 1.3011015625, 1.3014088134765625, 1.3012408447265624, 1.3015777587890625, 1.3010166015625, 1.3013145751953126, 1.3013350830078125, 1.3011260986328126, 1.3011947021484376, 1.3012562255859375, 1.3008834228515624, 1.3010831298828125, 1.3012030029296875, 1.301316650390625, 1.3013544921875, 1.301264404296875, 2.777419677734375, 1.300538330078125, 1.301001220703125, 1.3006500244140624, 1.300463623046875, 1.30085986328125, 1.30053125, 1.3006038818359376, 1.3009735107421876, 1.3007288818359375, 1.3006888427734375, 1.301128173828125, 1.3007677001953124, 1.3006806640625, 1.300806640625, 1.3004195556640625, 1.3006346435546876, 1.3007237548828126, 1.30064892578125, 1.3006796875, 1.300641845703125, 1.300473876953125, 1.3006182861328126, 1.302119384765625, 1.30094287109375, 1.3011363525390625, 1.300811767578125, 1.3008507080078124, 1.3006817626953124, 1.3012623291015626, 1.30117626953125, 1.3009305419921875, 1.3007247314453125, 1.3009541015625, 1.3009193115234374, 1.3008465576171875, 1.301043212890625, 1.3010574951171876, 1.3009449462890625, 1.30106982421875, 1.3009664306640625, 1.301012451171875, 1.3013687744140625, 1.30088037109375, 1.300948974609375, 1.3010882568359374, 1.3011353759765625, 1.30454736328125, 1.301443603515625, 1.3011292724609376, 1.301358642578125, 1.3010616455078126, 1.3011578369140624, 1.3011610107421876, 1.301501953125, 1.3013033447265625, 1.301171142578125, 1.3012459716796876, 1.3013411865234374, 1.30141796875, 1.3013453369140624, 1.3010943603515626, 1.30140673828125, 2.778178466796875, 1.300643798828125, 1.300801513671875, 1.3004708251953125, 1.3004769287109375, 1.3006285400390625, 1.3008875732421874, 1.3006510009765626, 1.3005496826171874, 1.3005035400390625, 1.3008302001953125, 1.3009326171875, 1.3024849853515625, 1.3005352783203126, 1.30062646484375, 1.3004461669921874, 1.301075927734375, 1.30092236328125, 1.3006234130859375, 1.3008363037109374, 1.3009613037109375, 1.300926513671875, 1.3007349853515624, 1.3008087158203125, 1.300937744140625, 1.300928466796875, 1.3008936767578125, 1.300842529296875, 1.300822998046875, 1.301306396484375, 1.301065673828125, 1.300947998046875, 1.3007821044921875, 1.3008773193359375, 1.3007369384765626, 1.300990966796875, 1.30096337890625, 1.30093359375, 1.3009888916015624, 1.301021728515625, 1.3011629638671875, 1.3009664306640625, 1.30126953125, 1.3011629638671875, 1.3056778564453124, 1.301190673828125, 1.301339111328125, 1.301127197265625, 1.3016986083984374, 1.3011834716796875, 1.3013453369140624, 1.3007625732421875, 1.300928466796875, 1.3012100830078126, 1.3018818359375, 1.301137451171875, 1.30134326171875, 1.30126025390625, 1.30119580078125, 1.3012305908203126, 1.3012674560546875, 1.3013370361328125, 1.3013780517578124, 2.777077880859375, 1.3004820556640626, 1.30075439453125, 1.30027001953125, 1.300384765625, 1.3006510009765626, 1.300737060546875, 1.300611083984375, 1.301982177734375, 1.3008885498046876, 1.3010308837890625, 1.30088134765625, 1.3007410888671875, 1.3007186279296874, 1.3009940185546875, 1.3006663818359374, 1.3005352783203126, 1.301898193359375, 1.3007135009765625, 1.3007830810546874, 1.3008455810546875, 1.30092333984375, 1.3009459228515625, 1.300763671875, 1.300854736328125, 1.3009571533203126, 1.3007186279296874, 1.3007841796875, 1.301031982421875, 1.3012030029296875, 1.30128173828125, 1.300864013671875, 1.3006080322265625, 1.3009193115234374, 1.30107080078125, 1.30123779296875, 1.3009029541015624, 1.3008353271484374, 1.3009029541015624, 1.300968505859375, 1.301223388671875, 1.3012100830078126, 1.301012451171875, 1.3009561767578126, 1.3011773681640626, 1.3009869384765624, 1.301032958984375, 1.30104736328125, 1.30123876953125, 1.300968505859375, 1.3011661376953125, 1.3010616455078126, 1.3011220703125, 1.3012991943359375, 1.3013094482421874, 1.3011650390625, 1.301307373046875, 1.306968017578125, 1.3015224609375, 1.3011640625, 1.3011395263671874, 1.30132373046875, 1.301569580078125, 2.77629443359375, 1.300684814453125, 1.3006571044921875, 1.3007032470703126, 1.3007728271484376, 1.3006182861328126, 1.30075341796875, 1.300673583984375, 1.300638671875, 1.3007923583984375, 1.300462646484375, 1.3006817626953124, 1.3007318115234374, 1.3004041748046875, 1.300664306640625, 1.300496337890625, 1.3006796875, 1.301011474609375, 1.300559814453125, 1.3005147705078124, 1.3007830810546874, 1.3007288818359375, 1.3005885009765625, 1.300822998046875, 1.300948974609375, 1.3010380859375, 1.3007718505859376, 1.3010247802734376, 1.301080078125, 1.3011260986328126, 1.3008148193359375, 1.3009398193359376, 1.300612060546875, 1.3008292236328125, 1.3007830810546874, 1.3008988037109375, 1.30096435546875, 1.30096337890625, 1.3062010498046874, 1.301222412109375, 1.30113330078125, 1.301170166015625, 1.3014896240234375, 1.30104736328125, 1.3009151611328125, 1.3010350341796875, 1.3011927490234374, 1.3011527099609375, 1.3012142333984376, 1.3013104248046874, 1.3012315673828125, 1.3008927001953126, 1.301086181640625, 1.301221435546875, 1.3014876708984375, 1.301213134765625, 1.3012408447265624, 1.3011661376953125, 1.301044189453125, 1.301147705078125, 1.3015531005859375, 1.3014088134765625, 1.3011251220703124, 2.778526611328125, 1.3008087158203125, 1.30071044921875, 1.3003427734375, 1.30060595703125, 1.300927490234375, 1.3007943115234375, 1.30056396484375, 1.300580322265625, 1.300707275390625, 1.3007625732421875, 1.3007606201171875, 1.300589599609375, 1.3005322265625, 1.300843505859375, 1.300663330078125, 1.300454345703125, 1.3007503662109374, 1.300990966796875, 1.30060693359375, 1.30069091796875, 1.3006510009765626, 1.3008619384765625, 1.30054248046875, 1.3009202880859374, 1.3009776611328125, 1.300906982421875, 1.300621337890625, 1.3008834228515624, 1.3012684326171875, 1.301033935546875, 1.300828125, 1.30052099609375, 1.300696044921875, 1.3007615966796875, 1.300918212890625, 1.3008670654296874, 1.3007861328125, 1.3010574951171876, 1.3008834228515624, 1.3008619384765625, 1.3009827880859375, 1.30100634765625, 1.3008638916015625, 1.301022705078125, 1.300989990234375, 1.30661376953125, 1.3012244873046874, 1.3013668212890626, 1.3011302490234375, 1.301508056640625, 1.3010247802734376, 1.3012562255859375, 1.301317626953125, 1.3011998291015625, 1.3013790283203126, 1.3010667724609375, 1.3010595703125, 1.30128173828125, 1.3011824951171875, 1.3012562255859375, 1.301348388671875, 1.3013197021484375, 2.776976318359375, 1.300622314453125, 1.3006234130859375, 1.3007462158203125, 1.300918212890625, 1.3006182861328126, 1.300453369140625, 1.3005947265625, 1.3005772705078125, 1.3006673583984374, 1.3008045654296876, 1.300738037109375, 1.3006612548828125, 1.3005496826171874, 1.300828125, 1.30060595703125, 1.3006458740234375, 1.300716552734375, 1.3007769775390625, 1.30071044921875, 1.3006878662109376, 1.3007288818359375, 1.3008394775390626, 1.300548583984375, 1.3008157958984374, 1.3008148193359375, 1.3009132080078125, 1.3008609619140625, 1.30098486328125, 1.3010780029296876, 1.30096435546875, 1.300960205078125, 1.30081787109375, 1.30079541015625, 1.3009622802734375, 1.30109033203125, 1.3008292236328125, 1.301412841796875, 1.3011568603515624, 1.3007943115234375, 1.300947998046875, 1.300906005859375, 1.3009725341796874, 1.300641845703125, 1.300947998046875, 1.3010985107421875, 1.3011005859375, 1.3011015625, 1.3012275390625, 1.3011414794921874, 1.301274658203125, 1.301185546875, 1.30088134765625, 1.3012347412109375, 1.3010789794921875, 1.3010626220703125, 1.3011650390625, 1.3011650390625, 1.3010504150390625, 1.301147705078125, 1.30126953125, 1.30136376953125, 1.301601318359375, 2.775970703125, 1.3004820556640626, 1.3007205810546876, 1.3006285400390625, 1.300580322265625, 1.300664306640625, 1.3006397705078125, 1.300610107421875, 1.300527099609375, 1.300705322265625, 1.30071240234375, 1.3007933349609375, 1.30050048828125, 1.300528076171875, 1.3007093505859375, 1.3008260498046875, 1.3007298583984375, 1.3006990966796874, 1.3004923095703125, 1.300674560546875, 1.3005322265625, 1.3008824462890625, 1.300791259765625, 1.30073291015625, 1.3007401123046876, 1.300757568359375, 1.3008895263671876, 1.3009417724609376, 1.3009029541015624, 1.30100537109375, 1.3008721923828126, 1.30088037109375, 1.30073193359375, 1.3008731689453126, 1.300918212890625, 1.3009224853515624, 1.3008629150390625, 1.3010103759765625, 1.301017578125, 1.3010155029296875, 1.301011474609375, 1.300653076171875, 1.301284912109375, 1.3010789794921875, 1.3012305908203126, 1.3012296142578126, 1.3009049072265626, 1.3011405029296874, 1.3012039794921875, 1.301171142578125, 1.301518310546875, 1.3013084716796874, 1.3012459716796876, 1.3011865234375, 1.3014415283203125, 1.3010924072265626, 1.3013197021484375, 1.3008885498046876, 1.301032958984375, 1.3011640625, 1.301353515625, 1.3011036376953125, 1.30125830078125, 2.786734130859375, 1.30047900390625, 1.300717529296875, 1.3006090087890625, 1.3007882080078126, 1.300664306640625, 1.3003203125, 1.30054345703125, 1.30045849609375, 1.3009674072265625, 1.3010166015625, 1.3007298583984375, 1.30047998046875, 1.3004779052734374, 1.30075146484375, 1.300571044921875, 1.3005394287109375, 1.3007462158203125, 1.3005875244140626, 1.3007073974609376, 1.300906005859375, 1.3008302001953125, 1.300822998046875, 1.3005301513671874, 1.300739013671875, 1.3010103759765625, 1.3040128173828125, 1.30119580078125, 1.300763671875, 1.3010780029296876, 1.3009725341796874, 1.30075244140625, 1.3008199462890624, 1.300832275390625, 1.3009295654296875, 1.3008568115234376, 1.3009305419921875, 1.3010032958984374, 1.301105712890625, 1.3009949951171875, 1.3009141845703125, 1.3007862548828124, 1.301212158203125, 1.300886474609375, 1.3010155029296875, 1.30094287109375, 1.30104931640625, 1.300917236328125, 1.30115283203125, 1.3011905517578124, 1.3014036865234375, 1.300979736328125, 1.3011773681640626, 1.3011097412109376, 1.3011650390625, 1.3012838134765625, 1.30128173828125, 1.3013780517578124, 1.3014814453125, 1.301169189453125, 1.3013974609375, 1.30119580078125, 1.3010616455078126]",tokens/s,0.7563509060269835,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/xglm-4.5B,facebook/xglm-4.5B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,3040.452608,4364.6976,0.0,3718.250496,3446.424576,s,10,2.190915100097656,0.21909151000976562,0.001943963189092213,0.21872471618652345,0.2219030990600586,0.22202302169799806,0.22211895980834961,"[0.22095942687988282, 0.21912918090820313, 0.21651104736328125, 0.2170089874267578, 0.217830810546875, 0.21724432373046876, 0.21832025146484374, 0.2198916778564453, 0.2221429443359375, 0.22187644958496094]",tokens/s,1168.4615254538583,kWh,2.5576309003728505e-06,1.4014560809416853e-06,1.1737255252680539e-05,1.5696342233995073e-05,tokens/kWh,16309532.258130576,MB,3040.452608,4364.6976,0.0,3718.250496,3674.351616,s,10,119.729416015625,11.9729416015625,0.004946861890726124,11.972722167968751,11.9776060546875,11.980676074218751,11.98313208984375,"[11.98374609375, 11.9704267578125, 11.976736328125, 11.97025390625, 11.972646484375, 11.96654296875, 11.9727978515625, 11.976923828125, 11.972966796875, 11.966375]",tokens/s,5.261864802863345,kWh,0.00014142660262138004,7.751303716739788e-05,0.000647636494468118,0.0008665761342568958,tokens/kWh,72699.90195843969,,s,629,121.50664608764652,0.19317431810436644,0.026001706272005144,0.18995712280273438,0.19051458740234375,0.1908127716064453,0.4075341943359375,"[0.19015679931640625, 0.19102310180664062, 0.19044044494628906, 0.18992025756835937, 0.18986598205566407, 0.1898137664794922, 0.1900943298339844, 0.18970828247070312, 0.1896294403076172, 0.18981887817382812, 0.1907466278076172, 0.19037081909179687, 0.18979225158691407, 0.18958950805664063, 0.19006259155273436, 0.1900298309326172, 0.19079270935058593, 0.19074560546875, 0.18997555541992187, 0.18996531677246092, 0.19174400329589844, 0.1903124542236328, 0.19008511352539062, 0.1898741760253906, 0.19113267517089844, 0.1904127960205078, 0.1901844482421875, 0.18970930480957032, 0.19028378295898438, 0.1904332733154297, 0.1909452819824219, 0.1907271728515625, 0.19070361328125, 0.1898803253173828, 0.1908449249267578, 0.19039949035644532, 0.19103948974609375, 0.19014553833007813, 0.19077529907226562, 0.190060546875, 0.19027456665039064, 0.19001548767089843, 0.190455810546875, 0.19010047912597655, 0.19035340881347657, 0.19053977966308594, 0.19030630493164064, 0.19005850219726564, 0.18978509521484374, 0.18953829956054688, 0.18995916748046876, 0.18988645935058593, 0.19023052978515625, 0.19020083618164063, 0.19000422668457032, 0.18960281372070312, 0.1903585205078125, 0.18995916748046876, 0.19028172302246094, 0.1897523193359375, 0.18992332458496095, 0.1898455047607422, 0.40921908569335935, 0.18996018981933593, 0.18974832153320312, 0.1898474578857422, 0.18964889526367187, 0.18998783874511718, 0.19081932067871094, 0.19014860534667968, 0.18967141723632813, 0.18989669799804687, 0.1900595245361328, 0.1903585205078125, 0.18998988342285156, 0.19033395385742188, 0.19040357971191407, 0.19020390319824218, 0.1897584686279297, 0.18977996826171875, 0.18946560668945311, 0.19001036071777344, 0.19018649291992187, 0.19096882629394532, 0.18982400512695313, 0.18978509521484374, 0.1906298828125, 0.1900963897705078, 0.1901527099609375, 0.1908326416015625, 0.18979942321777343, 0.189844482421875, 0.18977894592285155, 0.18990284729003906, 0.18995712280273438, 0.1897830352783203, 0.18991513061523438, 0.1899315185546875, 0.1898455047607422, 0.189918212890625, 0.18978713989257812, 0.19004928588867187, 0.19029402160644532, 0.18976153564453124, 0.18995610046386718, 0.1902571563720703, 0.1899745330810547, 0.18970828247070312, 0.19011686706542968, 0.18995610046386718, 0.1899683837890625, 0.1897840576171875, 0.18966630554199218, 0.18977587890625, 0.18983833312988282, 0.18982400512695313, 0.18963661193847656, 0.18981990051269532, 0.18968678283691406, 0.18995610046386718, 0.18979942321777343, 0.19080294799804687, 0.18968576049804686, 0.18998374938964843, 0.18979225158691407, 0.4081162109375, 0.1903513641357422, 0.19143679809570313, 0.1900615692138672, 0.18989158630371095, 0.19010662841796874, 0.18985574340820313, 0.18982809448242188, 0.1897031707763672, 0.19044557189941405, 0.18989260864257812, 0.18974412536621094, 0.19069541931152345, 0.1899069366455078, 0.19015884399414062, 0.1899304962158203, 0.18987213134765624, 0.18976358032226562, 0.18973388671875, 0.1898403778076172, 0.18985369873046876, 0.18991410827636718, 0.1899683837890625, 0.18998886108398438, 0.18977996826171875, 0.1899673614501953, 0.1898577880859375, 0.18981272888183592, 0.18985983276367188, 0.18999501037597658, 0.18983322143554687, 0.19034010314941407, 0.1899120635986328, 0.1901158447265625, 0.19043431091308594, 0.1899735107421875, 0.19016908264160157, 0.1900789794921875, 0.18976255798339844, 0.1899304962158203, 0.18983833312988282, 0.19030528259277343, 0.18996121215820314, 0.18991719055175782, 0.18978201293945313, 0.1901844482421875, 0.1902550964355469, 0.19099853515625, 0.19009024047851564, 0.18987110900878906, 0.18990899658203125, 0.19039231872558593, 0.19038514709472656, 0.18986802673339845, 0.19011378479003907, 0.19073741149902343, 0.19014247131347656, 0.1900963897705078, 0.19028787231445313, 0.19005439758300782, 0.19058277893066405, 0.19075276184082032, 0.19105075073242186, 0.4071434326171875, 0.19111117553710938, 0.19029708862304687, 0.18971852111816406, 0.18985983276367188, 0.18970828247070312, 0.1896478729248047, 0.189770751953125, 0.18967654418945312, 0.19015884399414062, 0.18960794067382813, 0.18983013916015626, 0.18978099060058592, 0.1898629150390625, 0.18977177429199218, 0.18993766784667968, 0.18974412536621094, 0.18983731079101562, 0.18981990051269532, 0.19044557189941405, 0.19042201232910155, 0.19000218200683594, 0.1902540740966797, 0.1899622344970703, 0.18967245483398437, 0.1898946533203125, 0.18965913391113282, 0.18988134765625, 0.18992640686035156, 0.19032882690429687, 0.1897891845703125, 0.18988748168945313, 0.18968678283691406, 0.1902223358154297, 0.19056536865234375, 0.19039129638671876, 0.1901107177734375, 0.19016294860839844, 0.1900933074951172, 0.1902888946533203, 0.1896058807373047, 0.18996940612792967, 0.18971034240722656, 0.19055718994140625, 0.19048037719726563, 0.1900062713623047, 0.18988645935058593, 0.18974003601074219, 0.18984754943847656, 0.19016908264160157, 0.18987213134765624, 0.19173785400390625, 0.19008921813964844, 0.1908520965576172, 0.1898629150390625, 0.18972467041015625, 0.1896468505859375, 0.1895045166015625, 0.18976666259765626, 0.18975949096679687, 0.18985267639160155, 0.1897902069091797, 0.19000422668457032, 0.4076861572265625, 0.18977177429199218, 0.18957107543945312, 0.189949951171875, 0.19010252380371093, 0.18991513061523438, 0.18986189270019532, 0.19002061462402345, 0.1899120635986328, 0.1900011444091797, 0.18968576049804686, 0.190202880859375, 0.1896796112060547, 0.19047628784179688, 0.18981887817382812, 0.18989158630371095, 0.18972877502441407, 0.19000933837890624, 0.18988748168945313, 0.18985574340820313, 0.18974822998046875, 0.18976153564453124, 0.18987110900878906, 0.1898567657470703, 0.1904496612548828, 0.189876220703125, 0.18986189270019532, 0.18975436401367188, 0.1897584686279297, 0.1897891845703125, 0.1898424377441406, 0.1898772430419922, 0.19119410705566406, 0.18992434692382812, 0.18991719055175782, 0.1897461700439453, 0.18983013916015626, 0.1899857940673828, 0.18987826538085936, 0.19075276184082032, 0.1902172088623047, 0.18990797424316405, 0.19089511108398438, 0.19005850219726564, 0.1903626251220703, 0.19016294860839844, 0.19013119506835938, 0.19052032470703126, 0.1904496612548828, 0.19014247131347656, 0.19082444763183593, 0.1898086395263672, 0.18996018981933593, 0.18961817932128905, 0.1898219451904297, 0.19011686706542968, 0.18992845153808594, 0.1902581787109375, 0.19024485778808595, 0.19038412475585936, 0.19026431274414063, 0.19060838317871093, 0.19010867309570312, 0.40711065673828123, 0.1900789794921875, 0.19005439758300782, 0.1898639373779297, 0.18978099060058592, 0.1897512969970703, 0.18972262573242188, 0.18973695373535157, 0.18979327392578124, 0.18976153564453124, 0.1896785888671875, 0.18964480590820312, 0.18986189270019532, 0.19021311950683595, 0.18974003601074219, 0.1898629150390625, 0.18963967895507813, 0.19023564147949218, 0.19015577697753908, 0.19044146728515626, 0.19060838317871093, 0.1899632568359375, 0.18974412536621094, 0.1899192352294922, 0.18979942321777343, 0.189739013671875, 0.18984959411621094, 0.18971238708496094, 0.19051315307617187, 0.18992536926269532, 0.19039949035644532, 0.19006976318359375, 0.1899745330810547, 0.1899192352294922, 0.1896898498535156, 0.18990386962890626, 0.18992536926269532, 0.18979635620117188, 0.18993766784667968, 0.19039334106445313, 0.1896407012939453, 0.18960794067382813, 0.18975027465820313, 0.1897512969970703, 0.1898209228515625, 0.19032575988769532, 0.1900349426269531, 0.18992332458496095, 0.1897840576171875, 0.1898250274658203, 0.19001957702636718, 0.18980557250976562, 0.19014041137695312, 0.1902192687988281, 0.18999398803710937, 0.19036058044433593, 0.18978816223144532, 0.18988851928710937, 0.19006976318359375, 0.19062681579589844, 0.189955078125, 0.18988134765625, 0.18972160339355468, 0.40814080810546877, 0.19043020629882812, 0.18990489196777344, 0.18998886108398438, 0.18958131408691406, 0.1901045684814453, 0.18983935546875, 0.18997042846679688, 0.19002572631835937, 0.19087257385253906, 0.18973388671875, 0.18977792358398438, 0.18963250732421874, 0.18960691833496093, 0.1898567657470703, 0.18984652709960936, 0.18974822998046875, 0.1904117736816406, 0.18991001892089843, 0.19027967834472656, 0.1899366455078125, 0.1898772430419922, 0.18979942321777343, 0.18984346008300781, 0.18987315368652344, 0.18983833312988282, 0.18985267639160155, 0.1900421142578125, 0.1898997802734375, 0.18981683349609374, 0.18969088745117188, 0.18982911682128906, 0.1898987579345703, 0.18990489196777344, 0.18993254089355469, 0.18979533386230468, 0.18988645935058593, 0.1900943298339844, 0.19071078491210938, 0.19044248962402344, 0.1901854705810547, 0.19031552124023438, 0.190413818359375, 0.19061862182617187, 0.1901045684814453, 0.190487548828125, 0.18996940612792967, 0.18995916748046876, 0.19043533325195314, 0.1902458953857422, 0.19001242065429688, 0.18984754943847656, 0.18988543701171876, 0.1902755889892578, 0.19110809326171874, 0.19005746459960937, 0.18990591430664064, 0.18971034240722656, 0.19021209716796875, 0.18991719055175782, 0.19015884399414062, 0.19030630493164064, 0.19037799072265624, 0.40975668334960935, 0.18986904907226562, 0.1896826934814453, 0.189981689453125, 0.18960076904296874, 0.1903206329345703, 0.18970112609863282, 0.19098828125, 0.19000729370117186, 0.19016908264160157, 0.18983322143554687, 0.1904148406982422, 0.19014451599121093, 0.18970008850097655, 0.19014041137695312, 0.19007693481445312, 0.1901230010986328, 0.18999090576171876, 0.189802490234375, 0.19023564147949218, 0.18995916748046876, 0.19001242065429688, 0.1904680938720703, 0.19021209716796875, 0.19009024047851564, 0.19066879272460938, 0.190376953125, 0.19032882690429687, 0.18982707214355468, 0.19041690063476563, 0.18962124633789063, 0.1903308868408203, 0.19029093933105468, 0.19019775390625, 0.18998066711425782, 0.19013119506835938, 0.19030323791503906, 0.1901158447265625, 0.18965708923339844, 0.1903687744140625, 0.18969088745117188, 0.18983833312988282, 0.1913559112548828, 0.19069541931152345, 0.19020799255371093, 0.18968882751464844, 0.19023564147949218, 0.19037388610839845, 0.18994790649414062, 0.1905786895751953, 0.19024281311035157, 0.19003187561035156, 0.1898137664794922, 0.18969088745117188, 0.18964582824707032, 0.18987826538085936, 0.18970112609863282, 0.190171142578125, 0.19045989990234374, 0.190166015625, 0.19074458312988282, 0.18998886108398438, 0.18979942321777343, 0.41182516479492187, 0.19034010314941407, 0.19052032470703126, 0.19043431091308594, 0.19040562438964845, 0.18998783874511718, 0.1902335968017578, 0.19058790588378907, 0.18992025756835937, 0.18989056396484374, 0.19020492553710938, 0.18997555541992187, 0.18963967895507813, 0.19021516418457032, 0.19030015563964844, 0.19045274353027344, 0.1899673614501953, 0.19038925170898438, 0.189802490234375, 0.19024281311035157, 0.18988236999511718, 0.18972671508789063, 0.18969395446777343, 0.19045376586914062, 0.19055616760253907, 0.1903390655517578, 0.1896959991455078, 0.18977894592285155, 0.1897164764404297, 0.18966835021972656, 0.18990899658203125, 0.1901782989501953, 0.18996429443359375, 0.18969190979003905, 0.189802490234375, 0.18997964477539062, 0.1899192352294922, 0.18988645935058593, 0.19035749816894532, 0.19015577697753908, 0.19002163696289062, 0.18990591430664064, 0.1898577880859375, 0.18996429443359375, 0.19007078552246093, 0.1899735107421875, 0.19000729370117186, 0.19146957397460937, 0.1904332733154297, 0.190276611328125, 0.18994073486328125, 0.18992845153808594, 0.18996940612792967, 0.18989158630371095, 0.19016192626953124, 0.18999501037597658, 0.18971340942382814, 0.18979942321777343, 0.18969804382324218, 0.18984141540527344, 0.18973695373535157, 0.18978509521484374, 0.1900011444091797, 0.41165618896484374, 0.1897943115234375, 0.19018751525878907, 0.1898987579345703, 0.18995916748046876, 0.18978099060058592, 0.18968780517578124, 0.18963763427734376, 0.18988236999511718, 0.18971238708496094, 0.1893079071044922, 0.18949632263183594, 0.1894246368408203, 0.18974412536621094, 0.18988645935058593, 0.18984857177734374, 0.1897584686279297, 0.1898772430419922, 0.18988851928710937, 0.1902899169921875, 0.18975027465820313, 0.18976051330566407, 0.18977587890625, 0.18996633911132813, 0.1901598663330078, 0.190023681640625, 0.1906237487792969, 0.19034725952148437, 0.18987315368652344, 0.18974003601074219, 0.18974310302734376, 0.18989363098144532, 0.19047731018066405, 0.18996018981933593, 0.19003904724121093, 0.1902090301513672, 0.19019879150390626, 0.1905479736328125, 0.18968576049804686, 0.1903308868408203, 0.18991410827636718, 0.1900738525390625, 0.18983424377441407, 0.18995814514160156, 0.19019879150390626, 0.1899069366455078, 0.18994175720214843, 0.1899130859375, 0.18976460266113282, 0.18975334167480468, 0.18988134765625, 0.18966835021972656, 0.1902899169921875, 0.189770751953125, 0.18975334167480468, 0.18957618713378907, 0.18963046264648437, 0.18963661193847656, 0.18959666442871093, 0.18997862243652344, 0.18970623779296875, 0.18991001892089843, 0.18985472106933593]",tokens/s,5.176671566971595,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12113,7 +12113,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d461f-7df667e649c655c6333785ca;5092a282-e965-48f1-baeb-f2863e5da726) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe942-2a2dd1150ce350a449711122;7048f38b-af90-4026-a0c4-ce4cd9fac1e2) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12144,7 +12144,7 @@ OSError: B is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12164,9 +12164,9 @@ ChildProcessError: Traceback (most recent call last): return self.pretrained_model.generate(**inputs, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context return func(*args, **kwargs) - File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1736, in generate + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1758, in generate result = self._sample( - File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2375, in _sample + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2397, in _sample outputs = self( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) @@ -12241,7 +12241,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9440-007114206a39c66045f3cfad;37b8ebbd-28db-4891-ba43-8ed6f0631124) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe603-603e0a506aff01b07f2d2f6c;bb70b357-519a-4bf8-ad4c-3323c639349c) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12435,7 +12435,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e93f0-126b3ee51c12a0ca6fdc2983;09d3c397-3591-4076-99fe-2f527671876b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe5b4-0fcc6e004cf024d011d4eec2;12a8e41b-0084-4710-ba4a-a9d5af95cab6) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12466,7 +12466,7 @@ OSError: m is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12505,7 +12505,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46e3-601eb589507c38ac1b38a847;2df42ab0-a98b-4b4a-a9d0-300e698205f6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fea19-44b5bb282e5063d9454c8a61;17d1ae6b-1307-470f-8be9-082085e21505) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12536,7 +12536,7 @@ OSError: 1 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12575,7 +12575,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4695-5bcd078e11234c5378b66406;f2a9b307-b66f-4271-a200-9db9a181acdf) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe9bd-3418aaf05c378021548ba469;25031947-37cf-486f-89c1-75ccf372c8d9) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12607,7 +12607,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-eager,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-6.7b,,cuda,0,42,,,,,True,True,,float16,True,False,,eager,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1985.257472,5989.9904,0.0,5343.543296,5028.4416,s,10,3.2313330993652345,0.3231333099365234,0.0032608708937781655,0.3219995422363281,0.3235000213623047,0.32819869537353513,0.33195763458251953,"[0.3328973693847656, 0.32206170654296873, 0.32237094116210935, 0.32195501708984375, 0.3217920227050781, 0.32245587158203126, 0.32184100341796873, 0.3220360717773437, 0.3219600830078125, 0.3219630126953125]",tokens/s,792.2426816668602,kWh,3.8074948255800542e-06,2.086324617886248e-06,1.9453253409809634e-05,2.5347072853275936e-05,tokens/kWh,10099785.544543214,MB,1985.257472,5989.9904,0.0,5343.543296,5239.773184,s,10,177.66795703125,17.766795703125002,0.0032013312193415297,17.7664296875,17.77143515625,17.771705859375,17.771922421875,"[17.769552734375, 17.76445703125, 17.763001953125, 17.764943359375, 17.7653984375, 17.767556640625, 17.771375, 17.7719765625, 17.762234375, 17.7674609375]",tokens/s,3.5459404752945365,kWh,0.0002097305886416386,0.00011494983980204666,0.0010577858306029947,0.0013824662590466799,tokens/kWh,45570.73244119788,,s,629,180.27804629516564,0.2866105664470048,0.038278268305440656,0.28196453857421877,0.2823047058105469,0.2824747009277344,0.6040556909179687,"[0.28233099365234376, 0.2819727478027344, 0.281860107421875, 0.2821601257324219, 0.281987060546875, 0.28179559326171877, 0.282102783203125, 0.2820362243652344, 0.28177304077148435, 0.2817556457519531, 0.2817996826171875, 0.2817679443359375, 0.28180377197265627, 0.28161227416992185, 0.281786376953125, 0.281818115234375, 0.28176177978515626, 0.28187646484375, 0.2820137634277344, 0.2819573059082031, 0.2816624755859375, 0.28175155639648436, 0.2818252868652344, 0.2817096252441406, 0.281872314453125, 0.28213565063476564, 0.2819685363769531, 0.2818385925292969, 0.28178225708007815, 0.2817433471679687, 0.28216421508789064, 0.2820137023925781, 0.28190411376953123, 0.28185498046875, 0.2819000244140625, 0.28225332641601564, 0.2818887634277344, 0.28175973510742186, 0.2823813171386719, 0.2819471435546875, 0.2820556945800781, 0.2818559875488281, 0.2817208251953125, 0.2819082336425781, 0.28190411376953123, 0.2822379455566406, 0.2823198852539063, 0.2819471435546875, 0.28203314208984376, 0.2821549987792969, 0.28195635986328127, 0.28193484497070315, 0.2820054931640625, 0.2822307739257813, 0.28209869384765623, 0.2821160888671875, 0.2820208740234375, 0.2819665832519531, 0.2820179138183594, 0.2818005676269531, 0.28207308959960936, 0.28215704345703124, 0.6044856567382813, 0.281807861328125, 0.2821396484375, 0.2819686279296875, 0.2818385925292969, 0.2819450988769531, 0.2816942138671875, 0.28203009033203125, 0.28204031372070315, 0.2819020690917969, 0.2819594116210937, 0.28254925537109377, 0.2820771789550781, 0.28199322509765623, 0.28177099609375, 0.28218060302734377, 0.2819430541992187, 0.28173004150390624, 0.281881591796875, 0.28195635986328127, 0.2817556457519531, 0.28178738403320314, 0.28190618896484376, 0.28194610595703123, 0.2817659301757813, 0.2822686462402344, 0.28204544067382814, 0.28183551025390624, 0.2816849975585938, 0.2819368896484375, 0.28181607055664065, 0.2819020690917969, 0.2819522705078125, 0.2820147094726563, 0.28198809814453124, 0.28182833862304685, 0.28189901733398437, 0.2818006896972656, 0.2820351867675781, 0.2822287292480469, 0.2821826477050781, 0.28221234130859374, 0.28221133422851563, 0.2820782470703125, 0.28194403076171876, 0.28187954711914065, 0.28224615478515624, 0.2821264038085938, 0.28191738891601564, 0.2819686279296875, 0.28202392578125, 0.2817894287109375, 0.2819092407226563, 0.28182833862304685, 0.2819440612792969, 0.28205056762695313, 0.2820198669433594, 0.28188668823242186, 0.2817976379394531, 0.2819850158691406, 0.28190411376953123, 0.28225537109375, 0.2821099548339844, 0.6047026977539063, 0.2818478088378906, 0.2815784912109375, 0.281923583984375, 0.28182833862304685, 0.2820904846191406, 0.2820085754394531, 0.2820711059570313, 0.28191653442382814, 0.282148681640625, 0.2819225769042969, 0.28175970458984373, 0.28166656494140624, 0.2818662414550781, 0.2819635314941406, 0.2820208740234375, 0.2818027648925781, 0.2819850158691406, 0.28203009033203125, 0.281776123046875, 0.2816849975585938, 0.28174642944335937, 0.2818006896972656, 0.28222055053710937, 0.28184063720703123, 0.2817054748535156, 0.281849853515625, 0.28192767333984375, 0.2818109436035156, 0.28192562866210935, 0.28192767333984375, 0.2817986450195312, 0.2819717102050781, 0.281965576171875, 0.2819368896484375, 0.28174951171875, 0.28181707763671876, 0.28224920654296876, 0.28189901733398437, 0.2819686279296875, 0.28199322509765623, 0.28184884643554686, 0.2817525634765625, 0.28185906982421877, 0.281818115234375, 0.28234136962890627, 0.2822901611328125, 0.28211306762695315, 0.28201162719726564, 0.2825216064453125, 0.2822717590332031, 0.28177407836914065, 0.2818089904785156, 0.282315673828125, 0.28198196411132814, 0.282071044921875, 0.281952392578125, 0.28193572998046873, 0.282166259765625, 0.2820823059082031, 0.28184268188476563, 0.2819573669433594, 0.2820147094726563, 0.6040350952148438, 0.2816143493652344, 0.28170443725585936, 0.28171878051757815, 0.28220928955078123, 0.28202597045898437, 0.2818529357910156, 0.2819317626953125, 0.28184268188476563, 0.2818948974609375, 0.2818580627441406, 0.2819522705078125, 0.28203213500976565, 0.2819112854003906, 0.2817433471679687, 0.28199014282226564, 0.2820044860839844, 0.2821058654785156, 0.28183551025390624, 0.28186215209960935, 0.28192562866210935, 0.28206695556640626, 0.2817556457519531, 0.28182833862304685, 0.28229428100585935, 0.2819522705078125, 0.28187033081054685, 0.28178125, 0.2819993591308594, 0.28177203369140624, 0.2818109436035156, 0.2818242492675781, 0.2818437194824219, 0.2818191223144531, 0.28180685424804686, 0.2818682861328125, 0.2819368896484375, 0.28164608764648436, 0.28166043090820314, 0.28221951293945313, 0.28204544067382814, 0.28195431518554687, 0.28177203369140624, 0.281891845703125, 0.28230450439453125, 0.2819798889160156, 0.28213861083984376, 0.2822297668457031, 0.28192050170898436, 0.2819102783203125, 0.28218368530273436, 0.2817392578125, 0.28209869384765623, 0.2821652526855469, 0.28196249389648437, 0.2819584045410156, 0.28251760864257813, 0.28237710571289065, 0.2821396484375, 0.28233319091796877, 0.2826168212890625, 0.2827724914550781, 0.28209561157226565, 0.6040617065429688, 0.2819317626953125, 0.28161639404296873, 0.2820362243652344, 0.2817914733886719, 0.2817003479003906, 0.28167678833007814, 0.2820157470703125, 0.2818049011230469, 0.28160296630859377, 0.2821549987792969, 0.2819921875, 0.2818201599121094, 0.281712646484375, 0.28182733154296874, 0.2818877563476562, 0.2817576904296875, 0.28202496337890626, 0.2818006896972656, 0.28176690673828125, 0.28174642944335937, 0.2816860046386719, 0.28168295288085937, 0.2819491882324219, 0.28186419677734376, 0.2817986450195312, 0.28165325927734375, 0.282060791015625, 0.2818304138183594, 0.28190213012695314, 0.2818467102050781, 0.2823157653808594, 0.28200344848632813, 0.28196966552734376, 0.2821939392089844, 0.28190921020507814, 0.2821949462890625, 0.28206491088867186, 0.28202597045898437, 0.28245709228515625, 0.28211624145507813, 0.2821169738769531, 0.2823270568847656, 0.282124267578125, 0.28217459106445314, 0.2820258483886719, 0.282281005859375, 0.28206179809570314, 0.28189706420898436, 0.2817832946777344, 0.28202587890625, 0.2818365478515625, 0.28199234008789065, 0.2823606872558594, 0.2821099548339844, 0.282166259765625, 0.28222772216796876, 0.28192050170898436, 0.2819942321777344, 0.2828697509765625, 0.28225442504882814, 0.28219586181640627, 0.28221951293945313, 0.6048604125976562, 0.2817424011230469, 0.2820597229003906, 0.28221337890625, 0.28190618896484376, 0.2818539428710938, 0.2816881103515625, 0.28207305908203123, 0.28197891235351563, 0.28253488159179685, 0.28218170166015627, 0.2823607788085937, 0.28224102783203125, 0.28216421508789064, 0.281860107421875, 0.28207940673828125, 0.28228182983398437, 0.2823638916015625, 0.282292236328125, 0.28184063720703123, 0.2817525634765625, 0.28177304077148435, 0.2818385925292969, 0.2820157470703125, 0.282102783203125, 0.28252774047851564, 0.28196453857421877, 0.2820823059082031, 0.28203109741210936, 0.2817914733886719, 0.2819010620117188, 0.2821314697265625, 0.28197994995117187, 0.28188156127929687, 0.2820915222167969, 0.2819164123535156, 0.28194097900390624, 0.28187237548828126, 0.28200140380859373, 0.2821672973632813, 0.2818785400390625, 0.28180990600585937, 0.28187237548828126, 0.2819092407226563, 0.2823270263671875, 0.28201165771484377, 0.28220928955078123, 0.28206796264648437, 0.2821273498535156, 0.2819850158691406, 0.28201165771484377, 0.2816839599609375, 0.2819512329101562, 0.28215399169921873, 0.28188980102539063, 0.2819092407226563, 0.28194097900390624, 0.28182220458984375, 0.28202291870117185, 0.2819368896484375, 0.2820485229492187, 0.2821109619140625, 0.282060791015625, 0.6040402221679687, 0.281923583984375, 0.28196249389648437, 0.2817413024902344, 0.28170443725585936, 0.2818191223144531, 0.28186932373046875, 0.2819584045410156, 0.2815528869628906, 0.2819287109375, 0.281913330078125, 0.281891845703125, 0.2820075378417969, 0.28173004150390624, 0.2819450988769531, 0.28215399169921873, 0.28208639526367185, 0.28201779174804686, 0.28188980102539063, 0.2818805847167969, 0.281860107421875, 0.2819942932128906, 0.2818794860839844, 0.2821693420410156, 0.2824816589355469, 0.28214169311523435, 0.2819840087890625, 0.28209765625, 0.2820618286132813, 0.2821314697265625, 0.28188467407226564, 0.2822000732421875, 0.2822072448730469, 0.281975830078125, 0.2820024108886719, 0.28200244140625, 0.2819942321777344, 0.282322998046875, 0.2823403015136719, 0.28221746826171873, 0.28208331298828127, 0.28273159790039065, 0.28245086669921876, 0.28214886474609374, 0.2823987121582031, 0.28251339721679686, 0.2822686767578125, 0.28212530517578127, 0.28193997192382814, 0.2821396484375, 0.28218572998046876, 0.28209561157226565, 0.2820413513183594, 0.28223898315429685, 0.2822881164550781, 0.2821160888671875, 0.28215090942382814, 0.28214273071289064, 0.2820474853515625, 0.2826158447265625, 0.28230551147460936, 0.28215090942382814, 0.28208639526367185, 0.6041384887695312, 0.28187750244140625, 0.28196148681640626, 0.28208026123046875, 0.28184884643554686, 0.2818877563476562, 0.2818262939453125, 0.28184268188476563, 0.281744384765625, 0.28251239013671875, 0.2820126647949219, 0.2819420166015625, 0.28238644409179686, 0.2822102966308594, 0.28228607177734377, 0.2822328186035156, 0.2827069396972656, 0.28225537109375, 0.2819993591308594, 0.28178533935546873, 0.28187033081054685, 0.281933837890625, 0.2818836364746094, 0.28196453857421877, 0.2820075378417969, 0.2818457641601563, 0.2820843505859375, 0.28167788696289064, 0.28201669311523436, 0.2820167541503906, 0.2818252868652344, 0.28246426391601565, 0.2820495300292969, 0.28203729248046877, 0.28202801513671877, 0.28208331298828127, 0.2821468200683594, 0.28205056762695313, 0.28212838745117186, 0.2823260192871094, 0.2821396484375, 0.2821181335449219, 0.2820208740234375, 0.28194207763671875, 0.28242529296875, 0.28203729248046877, 0.28214779663085937, 0.2822318115234375, 0.2823741455078125, 0.28212939453125, 0.28218572998046876, 0.282777587890625, 0.2825062255859375, 0.28214886474609374, 0.28201983642578127, 0.28207623291015627, 0.28205050659179687, 0.28201983642578127, 0.2820823059082031, 0.28218572998046876, 0.2821652526855469, 0.28215090942382814, 0.28215090942382814, 0.6043412475585938, 0.2818334655761719, 0.28201983642578127, 0.2822164611816406, 0.2819512329101562, 0.28177203369140624, 0.28185906982421877, 0.28215090942382814, 0.2816911315917969, 0.2819389343261719, 0.2819573669433594, 0.2817331237792969, 0.2818611145019531, 0.28189080810546874, 0.28189797973632813, 0.28195550537109376, 0.28243743896484375, 0.28220928955078123, 0.28171673583984375, 0.28182119750976564, 0.28188577270507814, 0.2818887023925781, 0.2818551025390625, 0.2821610107421875, 0.2817925109863281, 0.28167987060546873, 0.2816778259277344, 0.28177203369140624, 0.28190618896484376, 0.28177716064453123, 0.28174542236328126, 0.28187136840820315, 0.28183551025390624, 0.28193997192382814, 0.28182220458984375, 0.28168295288085937, 0.28177716064453123, 0.2818785400390625, 0.281913330078125, 0.28198196411132814, 0.28183551025390624, 0.28177203369140624, 0.28184884643554686, 0.2818437194824219, 0.2817843322753906, 0.2818785400390625, 0.281786376953125, 0.28183551025390624, 0.28170443725585936, 0.28180996704101563, 0.28188360595703127, 0.2819215393066406, 0.2824488830566406, 0.2821846923828125, 0.28226968383789064, 0.28262713623046876, 0.28212628173828125, 0.28186932373046875, 0.28233319091796877, 0.2823638916015625, 0.28208743286132815, 0.2820280456542969, 0.2818959350585937, 0.6041456909179688, 0.2816860046386719, 0.28180889892578126, 0.28184268188476563, 0.28177938842773437, 0.2817175598144531, 0.281786376953125, 0.28196249389648437, 0.2816092224121094, 0.28214886474609374, 0.281965576171875, 0.2818734130859375, 0.2816501770019531, 0.28167169189453123, 0.2818754577636719, 0.28181402587890625, 0.28172186279296874, 0.2816583557128906, 0.28184268188476563, 0.2817156982421875, 0.2817392578125, 0.2816337890625, 0.28177920532226564, 0.28253289794921876, 0.2824201965332031, 0.28236697387695314, 0.28277862548828125, 0.28257177734375, 0.2826495971679687, 0.2826649475097656, 0.28213555908203125, 0.28208126831054686, 0.28189797973632813, 0.28176589965820314, 0.28208639526367185, 0.28214273071289064, 0.2819154052734375, 0.28188568115234375, 0.28198809814453124, 0.28205056762695313, 0.28182937622070314, 0.2822102966308594, 0.28220620727539064, 0.28197994995117187, 0.2824550170898438, 0.2820362243652344, 0.28224102783203125, 0.28213360595703124, 0.28185589599609373, 0.2819082336425781, 0.28213351440429685, 0.2818447265625, 0.28206491088867186, 0.28210891723632814, 0.2819420166015625, 0.2819020690917969, 0.28194815063476564, 0.28186520385742186, 0.28187954711914065, 0.2821846923828125, 0.2821048278808594, 0.2829885559082031, 0.28226763916015624]",tokens/s,3.48905489562578,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,.,.,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12675,7 +12675,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e96d9-6d852c41315dadbb59bea7dc;29665440-1f78-4277-a524-9c802d8ff4d8) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe8ae-171a7c9b2eb5e3410696ac5e;57e83651-947d-47ee-bc54-c1e52e75acb1) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12735,7 +12735,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,2,2,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -12774,7 +12774,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d4605-3020e19f1154ec3a1188dfd0;e91ec65d-9f8d-41b5-be82-221446c928f3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe928-2663266e281753c84e1e2cb1;cf3c69a8-16b4-45a6-9233-a5056b52d671) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12984,7 +12984,7 @@ ChildProcessError: Traceback (most recent call last): TypeError: DeciCoderAttention.forward() got an unexpected keyword argument 'cache_position' ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,togethercomputer/RedPajama-INCITE-Base-7B-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13042,7 +13042,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90fc-4b524f3a04bda484064ba5b2;6298507e-ebdc-4ff5-902a-43bdad03ec03) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe2ac-7b3c965754f981b448410b67;d1568324-bceb-4f68-8dd8-efbd15f74ccd) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -13150,7 +13150,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: OPTForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-7b,stabilityai/stablelm-base-alpha-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -13247,7 +13247,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95e5-509bc59d3ef3eed90b2683c3;fb999e39-0300-467f-b9dc-756376b2ea38) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe8d7-751105d4094d459c6a0171ce;2c8475d3-ad97-4801-ad1a-08c55a242153) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13307,7 +13307,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90d2-44d81d7700956fed66316435;37f895ed-f090-4632-9de6-c54fe5ab7420) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe27f-6d906cf24b6dee1a19b5115c;95981a7f-93f2-4d5e-8ce4-22f9b3e603a0) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -13832,7 +13832,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e8f3e-5a79cc970494f0ea58cfdf11;9b9ecf3e-a36a-426b-8f41-576f42ab713b) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe0e0-3fc2a1b5093319b04e674cd4;cb091e2c-8cd6-4745-8a38-c59c209eab99) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -13917,7 +13917,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e968a-2cdda90a62d381d870295e1e;17605c8e-2baf-4b91-a572-c4c648fd7768) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe858-019180ce1692294271d4f495;97b4d578-3d5b-43ef-ba87-12fd392f5529) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14061,7 +14061,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e95bd-3196a9b07be8ba7a063360fe;5ea3614c-42c8-4e30-9834-445fc4341de6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe78c-669943bd4edaeba432c727cb;d9789f28-6da0-497a-a2d0-f01eb401a5b1) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14135,7 +14135,7 @@ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 112.00 MiB. G ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-1.8B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,1857.773568,3395.813376,0.0,2749.366272,2585.357312,s,10,0.7326278381347655,0.07326278381347656,0.0010202652967553705,0.07289332580566407,0.0744323989868164,0.07474466476440429,0.0749944773864746,"[0.07307369232177735, 0.07505693054199218, 0.07257308959960937, 0.07189148712158203, 0.07271295928955078, 0.07210118103027344, 0.072621826171875, 0.07393405151367187, 0.07429961395263672, 0.07436300659179687]",tokens/s,3494.2707152892713,kWh,8.526288467250187e-07,4.671987187216898e-07,4.050754639481646e-06,5.370582204928354e-06,tokens/kWh,47667085.28641079,MB,1857.773568,3395.813376,0.0,2749.366272,2628.768768,s,10,38.62795434570313,3.8627954345703124,0.0013787928567573532,3.86274462890625,3.8645461914062498,3.86462001953125,3.86467908203125,"[3.862239990234375, 3.863882568359375, 3.86469384765625, 3.86452978515625, 3.863611328125, 3.862596923828125, 3.862892333984375, 3.8602333984375, 3.86094921875, 3.862324951171875]",tokens/s,16.309432137197284,kWh,4.555938902852252e-05,2.4969072835359344e-05,0.00021447355906873103,0.00028500202093261287,tokens/kWh,221051.06410770328,,s,629,39.2255651473999,0.062361788787599216,0.008780807090293767,0.06126895904541016,0.061400267791748044,0.06155386810302734,0.1344632177734375,"[0.06139596939086914, 0.06134783935546875, 0.06130278396606445, 0.06122700881958008, 0.06143606567382812, 0.061158241271972655, 0.061230079650878906, 0.06156902313232422, 0.06164582443237305, 0.06149529647827148, 0.06133760070800781, 0.06120243072509766, 0.06127718353271484, 0.0615997428894043, 0.0614205436706543, 0.06143385696411133, 0.06148198318481445, 0.061162494659423826, 0.061267967224121096, 0.06151065444946289, 0.06127414321899414, 0.061287391662597654, 0.06125568008422851, 0.06121779251098633, 0.06121574401855469, 0.061246463775634766, 0.06121270370483398, 0.061201377868652346, 0.06132428741455078, 0.061178878784179686, 0.06119833755493164, 0.06125260925292969, 0.06123519897460938, 0.06116454315185547, 0.06141747283935547, 0.06122089767456055, 0.06124540710449219, 0.06128332901000977, 0.061276161193847656, 0.0612044792175293, 0.0612935676574707, 0.06123110580444336, 0.061246463775634766, 0.06124249649047851, 0.06127705764770508, 0.06118608093261719, 0.06124233627319336, 0.06122598266601562, 0.061317119598388675, 0.0612044792175293, 0.0613201904296875, 0.06122393417358398, 0.06131302261352539, 0.06132735824584961, 0.061297664642333986, 0.061230079650878906, 0.061290496826171874, 0.06120345687866211, 0.06123519897460938, 0.06125568008422851, 0.061211647033691405, 0.06121062469482422, 0.13685043334960936, 0.061385726928710936, 0.06136422348022461, 0.061391902923583985, 0.0613969612121582, 0.06137651062011719, 0.06124031829833984, 0.061290496826171874, 0.06117171096801758, 0.06124748611450195, 0.06125056076049805, 0.061244415283203124, 0.06116352081298828, 0.06120243072509766, 0.061238273620605466, 0.061238273620605466, 0.061163585662841795, 0.06130579376220703, 0.061200382232666016, 0.06135603332519531, 0.06155263900756836, 0.061240478515625, 0.061153247833251954, 0.061241214752197265, 0.06119116973876953, 0.06119014358520508, 0.06122102355957031, 0.06126063919067383, 0.06113894271850586, 0.06125363159179688, 0.061257728576660155, 0.06135398483276367, 0.06126182556152344, 0.06128844833374023, 0.06118092727661133, 0.06130790328979492, 0.061292545318603515, 0.06130585479736328, 0.06130585479736328, 0.06157823944091797, 0.0615546875, 0.06145843124389649, 0.061260799407958984, 0.06128742218017578, 0.061306880950927733, 0.06136832046508789, 0.06350233459472657, 0.061344768524169924, 0.061143039703369144, 0.061300830841064455, 0.061209503173828124, 0.06128025436401367, 0.061257728576660155, 0.06132035064697266, 0.06122991943359375, 0.06137548828125, 0.061472766876220705, 0.061297664642333986, 0.06125673675537109, 0.06129660797119141, 0.0612935676574707, 0.06125158309936524, 0.06124448013305664, 0.134383544921875, 0.06131916809082031, 0.0611409912109375, 0.0613488655090332, 0.06120044708251953, 0.06135084915161133, 0.06120755386352539, 0.06130380630493164, 0.0612751350402832, 0.06122905731201172, 0.06122700881958008, 0.06135295867919922, 0.061241344451904295, 0.061300735473632816, 0.06121577453613281, 0.061286369323730466, 0.0612751350402832, 0.06132940673828125, 0.06118809509277344, 0.06131814575195312, 0.06130790328979492, 0.0613488655090332, 0.061249534606933595, 0.06230227279663086, 0.06123923110961914, 0.06123929595947265, 0.06117478561401367, 0.061221889495849606, 0.06119833755493164, 0.061257728576660155, 0.061176830291748044, 0.06125158309936524, 0.061233150482177735, 0.06127718353271484, 0.061344768524169924, 0.06139494323730469, 0.06125574493408203, 0.061348800659179685, 0.06131097412109375, 0.061314048767089846, 0.061292545318603515, 0.06137855911254883, 0.06121267318725586, 0.06131609725952149, 0.0612935676574707, 0.0612977294921875, 0.06115423965454102, 0.062395423889160154, 0.06123737716674805, 0.06142038345336914, 0.06166016006469727, 0.06140723037719727, 0.061308929443359375, 0.06141439819335937, 0.061271041870117185, 0.061306880950927733, 0.06125568008422851, 0.061437950134277344, 0.061267967224121096, 0.06134067153930664, 0.06126396942138672, 0.061334430694580076, 0.06136627197265625, 0.1344952697753906, 0.061371360778808594, 0.061200382232666016, 0.06137036895751953, 0.06131916809082031, 0.061306880950927733, 0.06132940673828125, 0.06137651062011719, 0.06126182556152344, 0.06136217498779297, 0.06131097412109375, 0.06130995178222656, 0.06125056076049805, 0.061431808471679686, 0.06132223892211914, 0.06132940673828125, 0.061211647033691405, 0.0613570556640625, 0.06121984100341797, 0.061379585266113285, 0.061233150482177735, 0.06123724746704102, 0.061211647033691405, 0.061295616149902345, 0.06129151916503906, 0.06138675308227539, 0.06125260925292969, 0.06131302261352539, 0.061254657745361325, 0.06133452987670898, 0.06119116973876953, 0.061399040222167967, 0.06132940673828125, 0.06128025436401367, 0.06132223892211914, 0.06143590545654297, 0.06123519897460938, 0.06140108871459961, 0.061257728576660155, 0.06137753677368164, 0.061502464294433595, 0.06135910415649414, 0.061244415283203124, 0.06137753677368164, 0.061966335296630856, 0.06137651062011719, 0.06124748611450195, 0.06135910415649414, 0.06123724746704102, 0.061400062561035154, 0.06116761779785156, 0.06129983901977539, 0.06128524780273437, 0.061574142456054685, 0.06160192108154297, 0.06146547317504883, 0.061254657745361325, 0.0613939208984375, 0.06138777542114258, 0.061290496826171874, 0.06130380630493164, 0.061260799407958984, 0.0612782096862793, 0.13449420166015624, 0.0613201904296875, 0.06128128051757813, 0.06128128051757813, 0.06118608093261719, 0.06126895904541016, 0.061241344451904295, 0.061216766357421876, 0.06117171096801758, 0.061315071105957034, 0.06134272003173828, 0.06174822235107422, 0.06120140838623047, 0.061178878784179686, 0.061195297241210936, 0.0614153938293457, 0.06117068862915039, 0.0621393928527832, 0.06123110580444336, 0.06123519897460938, 0.0611860466003418, 0.061274112701416014, 0.06114713668823242, 0.06122716903686523, 0.061267807006835935, 0.061276161193847656, 0.06118297576904297, 0.06133555221557617, 0.06133452987670898, 0.061262847900390625, 0.061224960327148435, 0.06130278396606445, 0.06118092727661133, 0.06130694580078125, 0.06124435043334961, 0.06132940673828125, 0.061295616149902345, 0.06131513595581055, 0.06114809417724609, 0.061308929443359375, 0.06128947067260742, 0.06133350372314453, 0.06126387023925781, 0.06140825653076172, 0.061385726928710936, 0.06133248138427734, 0.06132121658325195, 0.061254657745361325, 0.06115532684326172, 0.06133964920043945, 0.061189121246337894, 0.061254657745361325, 0.061254657745361325, 0.06132940673828125, 0.061244415283203124, 0.06130278396606445, 0.06126387023925781, 0.06329753494262695, 0.06128639984130859, 0.061312000274658204, 0.06132735824584961, 0.06121062469482422, 0.061227104187011716, 0.13391145324707032, 0.06119424057006836, 0.061162494659423826, 0.06187417602539062, 0.06159462356567383, 0.061346817016601565, 0.061265918731689455, 0.061241344451904295, 0.061262847900390625, 0.06127718353271484, 0.06129663848876953, 0.06137446212768555, 0.061271041870117185, 0.06124236679077148, 0.061238273620605466, 0.061267967224121096, 0.06116556930541992, 0.061265918731689455, 0.06117478561401367, 0.06121779251098633, 0.06124544143676758, 0.06132735824584961, 0.061238273620605466, 0.061195392608642575, 0.06117772674560547, 0.06121984100341797, 0.06117478561401367, 0.061249534606933595, 0.06135398483276367, 0.06128025436401367, 0.061143039703369144, 0.061276161193847656, 0.061193214416503904, 0.06121984100341797, 0.061216766357421876, 0.06132428741455078, 0.06124544143676758, 0.061282302856445314, 0.06337228775024414, 0.06132940673828125, 0.061230079650878906, 0.06132326507568359, 0.06124249649047851, 0.06130777740478516, 0.06127836990356445, 0.061315937042236326, 0.06120755386352539, 0.061228031158447264, 0.061224960327148435, 0.061297664642333986, 0.061158401489257816, 0.06132428741455078, 0.061193214416503904, 0.06124560165405273, 0.06118076705932617, 0.06130176162719726, 0.06120550537109375, 0.06126694488525391, 0.06125363159179688, 0.061369342803955076, 0.061314048767089846, 0.06129971313476563, 0.06124236679077148, 0.13450445556640625, 0.061317119598388675, 0.061284351348876956, 0.06129971313476563, 0.06124860763549805, 0.06139587020874023, 0.06129663848876953, 0.06127628707885742, 0.061279102325439455, 0.06135603332519531, 0.0612147216796875, 0.061383678436279294, 0.06137241744995117, 0.06130483245849609, 0.061230079650878906, 0.06132326507568359, 0.061260799407958984, 0.061249534606933595, 0.061295616149902345, 0.06130278396606445, 0.06117171096801758, 0.061284351348876956, 0.061284351348876956, 0.0623636474609375, 0.061454334259033204, 0.061413375854492185, 0.061369342803955076, 0.06145024108886719, 0.061213695526123046, 0.061308929443359375, 0.06139801788330078, 0.06125568008422851, 0.06120243072509766, 0.06139596939086914, 0.06120652770996094, 0.061292545318603515, 0.061486080169677736, 0.061244415283203124, 0.06117478561401367, 0.061338623046875, 0.061300735473632816, 0.06128025436401367, 0.061260799407958984, 0.061298686981201174, 0.06127308654785156, 0.06131302261352539, 0.061295616149902345, 0.061325313568115235, 0.061228031158447264, 0.06130380630493164, 0.06117174530029297, 0.06120867156982422, 0.06118182373046875, 0.061200382232666016, 0.06121267318725586, 0.061249534606933595, 0.06119731140136719, 0.061259777069091796, 0.06121779251098633, 0.0612853775024414, 0.06123519897460938, 0.061363201141357425, 0.06137651062011719, 0.13581520080566406, 0.06123721694946289, 0.061230079650878906, 0.06120243072509766, 0.06118809509277344, 0.06125888061523437, 0.06128627014160156, 0.061183998107910156, 0.06116966247558594, 0.061292545318603515, 0.06119116973876953, 0.06121267318725586, 0.06119424057006836, 0.0612567024230957, 0.06120345687866211, 0.0611778564453125, 0.062005279541015625, 0.061244384765625, 0.06121267318725586, 0.06127001571655273, 0.061156352996826174, 0.06120140838623047, 0.06130585479736328, 0.06121984100341797, 0.06111231994628906, 0.0612567024230957, 0.06117068862915039, 0.061274112701416014, 0.06124031829833984, 0.0612567024230957, 0.06120959854125976, 0.06128332901000977, 0.06125568008422851, 0.06122393417358398, 0.061233150482177735, 0.0613089599609375, 0.06115020751953125, 0.06122288131713867, 0.06142156982421875, 0.06144102478027344, 0.061252639770507815, 0.06141129684448242, 0.061379585266113285, 0.061317119598388675, 0.06144524765014649, 0.06125657653808594, 0.061233150482177735, 0.061290496826171874, 0.0611778564453125, 0.06124851226806641, 0.06120755386352539, 0.061276161193847656, 0.06118809509277344, 0.06129151916503906, 0.06121267318725586, 0.06132326507568359, 0.061453311920166016, 0.06124748611450195, 0.06111743927001953, 0.06126387023925781, 0.06130790328979492, 0.061262847900390625, 0.06124748611450195, 0.13622784423828124, 0.06128128051757813, 0.06116864013671875, 0.06118195343017578, 0.06121881484985352, 0.06123212814331055, 0.061211742401123044, 0.061231006622314454, 0.06120550537109375, 0.061178878784179686, 0.061222911834716794, 0.061193214416503904, 0.06135910415649414, 0.061623294830322264, 0.061295616149902345, 0.06120652770996094, 0.06120761489868164, 0.06214854431152344, 0.061200382232666016, 0.06128128051757813, 0.06116147232055664, 0.06121062469482422, 0.06120755386352539, 0.06122393417358398, 0.06119424057006836, 0.06125475311279297, 0.061219745635986325, 0.06122598266601562, 0.061143039703369144, 0.06130585479736328, 0.06125363159179688, 0.06124748611450195, 0.06132735824584961, 0.06121881484985352, 0.06116966247558594, 0.061338623046875, 0.0612567024230957, 0.061250625610351564, 0.061302753448486326, 0.061341663360595704, 0.061221889495849606, 0.06131820678710938, 0.061247425079345705, 0.061279232025146485, 0.061300735473632816, 0.06132735824584961, 0.0612935676574707, 0.061274112701416014, 0.06122393417358398, 0.06123622512817383, 0.061162494659423826, 0.06129663848876953, 0.06119424057006836, 0.061271041870117185, 0.061315071105957034, 0.06127001571655273, 0.061238273620605466, 0.06128128051757813, 0.06128844833374023, 0.061314048767089846, 0.06154751968383789, 0.06133248138427734, 0.061330432891845706, 0.13624729919433592, 0.061328384399414064, 0.061246463775634766, 0.06124031829833984, 0.061189121246337894, 0.061243392944335937, 0.061193214416503904, 0.06124851226806641, 0.061246463775634766, 0.06123219299316406, 0.061181888580322266, 0.06125158309936524, 0.06116761779785156, 0.06128742218017578, 0.061243392944335937, 0.06128639984130859, 0.06118406295776367, 0.06136620712280273, 0.06136115264892578, 0.06126396942138672, 0.061195167541503906, 0.0612751350402832, 0.06120243072509766, 0.06128332901000977, 0.06121779251098633, 0.061230079650878906, 0.0632248306274414, 0.06133350372314453, 0.06123212814331055, 0.06128639984130859, 0.061257728576660155, 0.06139596939086914, 0.061276161193847656, 0.06122700881958008, 0.06123519897460938, 0.06130380630493164, 0.06121881484985352, 0.06122905731201172, 0.06126387023925781, 0.06125568008422851, 0.061244415283203124, 0.06136627197265625, 0.06116147232055664, 0.061262847900390625, 0.06133145523071289, 0.06130294418334961, 0.0612022705078125, 0.06134783935546875, 0.06124236679077148, 0.06139801788330078, 0.06120969772338867, 0.061249439239501956, 0.06118204879760742, 0.06134364700317383, 0.06126489639282227, 0.06138374328613281, 0.06129779052734375, 0.06130259323120117, 0.06122086334228516, 0.06133555221557617, 0.061274112701416014, 0.061267967224121096, 0.061358081817626954]",tokens/s,16.035460486964933,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-rw-1b,tiiuae/falcon-rw-1b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14194,7 +14194,7 @@ ImportError: This modeling file requires the following packages that were not fo ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/rho-math-1b-v0.1,microsoft/rho-math-1b-v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1395.191808,1596.45696,0.0,950.009856,857.516544,s,10,0.5807296676635743,0.058072966766357426,0.00034456144070889745,0.05796723365783692,0.058138173675537114,0.05861479835510254,0.05899609809875488,"[0.05909142303466797, 0.05802451324462891, 0.05794847869873047, 0.05801567840576172, 0.05795558547973633, 0.0579788818359375, 0.05782329559326172, 0.057954433441162106, 0.05803225708007813, 0.05790512084960937]",tokens/s,4408.247318067187,kWh,6.839628269753551e-07,3.747758243667504e-07,2.899796796392844e-06,3.958535447734949e-06,tokens/kWh,64670382.10974761,MB,1395.48672,1596.45696,0.0,950.009856,888.373248,s,10,31.92044140625,3.1920441406249997,0.02836712052909163,3.19824169921875,3.2185572265625,3.2194255859375,3.2201202734375003,"[3.2202939453125, 3.201953125, 3.19358837890625, 3.211521728515625, 3.208235595703125, 3.1945302734375, 3.194171142578125, 3.2183642578125, 3.14763525390625, 3.130147705078125]",tokens/s,19.736569177788578,kWh,3.7041113467632236e-05,2.0300253596832975e-05,0.0001558331635232091,0.0002131745305876743,tokens/kWh,295532.49080142524,,s,629,32.391628852844285,0.051497025203249974,0.006842897447827928,0.05099008178710938,0.05155389404296875,0.051830375671386715,0.10754776977539063,"[0.05087539291381836, 0.050931713104248044, 0.051484672546386716, 0.04997324752807617, 0.05004800033569336, 0.05108019256591797, 0.05008281707763672, 0.04997123336791992, 0.05227721786499023, 0.05145907211303711, 0.04942233657836914, 0.04940595245361328, 0.04943462371826172, 0.0513361930847168, 0.050933761596679686, 0.051097599029541016, 0.0513966064453125, 0.05129011154174805, 0.05063987350463867, 0.05122867202758789, 0.051329025268554686, 0.05156147384643555, 0.05144985580444336, 0.05113651275634765, 0.051576831817626956, 0.05129318237304688, 0.051111934661865234, 0.05128192138671875, 0.05123072052001953, 0.05204377746582031, 0.05131267166137695, 0.05174576187133789, 0.05175296020507812, 0.05199155044555664, 0.05171200180053711, 0.051622913360595706, 0.05135974502563476, 0.05151232147216797, 0.051296257019042966, 0.05145907211303711, 0.051465217590332034, 0.05126758575439453, 0.05104742431640625, 0.05098188781738281, 0.05071257781982422, 0.051335166931152344, 0.05114879989624024, 0.0511016960144043, 0.05123276901245117, 0.05228953552246094, 0.05026816177368164, 0.049719295501708984, 0.04957388687133789, 0.05164851379394531, 0.05176422500610352, 0.05173964691162109, 0.051716094970703126, 0.05203046417236328, 0.051297279357910154, 0.050490367889404295, 0.05096038436889649, 0.051574783325195314, 0.1075937271118164, 0.04953500747680664, 0.049406944274902345, 0.04939059066772461, 0.04952473449707031, 0.04942540740966797, 0.04949606323242187, 0.04940288162231445, 0.04940083312988281, 0.04941619110107422, 0.05014015960693359, 0.05142323303222656, 0.05121023941040039, 0.05096550369262695, 0.05170687866210937, 0.051394561767578124, 0.051438591003417966, 0.04949708938598633, 0.04952166366577149, 0.050817024230957034, 0.05314048004150391, 0.05183692932128906, 0.05128192138671875, 0.050988033294677736, 0.05128499221801758, 0.05126144027709961, 0.051432449340820315, 0.050939903259277344, 0.051111934661865234, 0.0511016960144043, 0.051282943725585936, 0.05147238540649414, 0.051297279357910154, 0.05123379135131836, 0.051912704467773435, 0.05106687927246094, 0.05095935821533203, 0.05077811050415039, 0.0510648307800293, 0.051111934661865234, 0.051996673583984375, 0.05130854415893555, 0.051154945373535154, 0.05133824157714844, 0.05151641464233398, 0.051163135528564455, 0.051399681091308595, 0.05112934494018555, 0.05120307159423828, 0.05120614242553711, 0.05113446426391602, 0.05104127883911133, 0.049634304046630856, 0.051745792388916016, 0.05118873596191406, 0.05128806304931641, 0.049495040893554686, 0.049503231048583986, 0.04944486236572266, 0.05134950256347656, 0.05124607849121094, 0.05094707107543945, 0.05131161499023437, 0.10751385498046875, 0.05064908981323242, 0.05106995010375977, 0.051059711456298826, 0.05166592025756836, 0.050991104125976565, 0.0494571533203125, 0.051135486602783206, 0.05144166564941406, 0.04943667221069336, 0.049957889556884766, 0.051266559600830076, 0.05106892776489258, 0.05137100982666016, 0.05138534545898438, 0.05170995330810547, 0.04946944046020508, 0.05125632095336914, 0.051884033203125, 0.05146112060546875, 0.05174272155761719, 0.05139046478271484, 0.05128704071044922, 0.051454975128173826, 0.051369983673095705, 0.051119102478027346, 0.0512911376953125, 0.049459201812744144, 0.051471359252929685, 0.0514252815246582, 0.05107712173461914, 0.04943155288696289, 0.050402305603027345, 0.051643390655517575, 0.0514150390625, 0.05107712173461914, 0.05109452819824219, 0.051187713623046874, 0.05120409774780273, 0.05143961715698242, 0.051040256500244144, 0.05125836944580078, 0.05112319946289062, 0.04944179153442383, 0.04944793701171875, 0.04944793701171875, 0.04944179153442383, 0.04945817565917969, 0.04944486236572266, 0.049503231048583986, 0.04945612716674805, 0.0494284782409668, 0.04946636962890625, 0.04945407867431641, 0.049446910858154294, 0.04992409515380859, 0.050895870208740236, 0.051127296447753906, 0.051241985321044924, 0.0507852783203125, 0.05137919998168945, 0.051171329498291014, 0.05095116806030273, 0.1091809310913086, 0.05078835296630859, 0.05091020965576172, 0.05135974502563476, 0.05077503967285156, 0.05084979248046875, 0.05113958358764648, 0.049410049438476565, 0.049459201812744144, 0.05127167892456055, 0.05082009506225586, 0.050969600677490234, 0.051326976776123044, 0.0510382080078125, 0.050971649169921876, 0.05126348876953125, 0.0494202880859375, 0.051111934661865234, 0.051113983154296876, 0.051282943725585936, 0.05088460922241211, 0.051090431213378903, 0.05146419143676758, 0.05098700714111328, 0.04940697479248047, 0.052152320861816405, 0.05123891067504883, 0.05101670455932617, 0.050612224578857425, 0.051097599029541016, 0.050923519134521485, 0.05028761672973633, 0.05097369766235352, 0.05110067367553711, 0.051351551055908204, 0.052222976684570314, 0.05116211318969727, 0.05158911895751953, 0.051664894104003906, 0.05117337417602539, 0.05112934494018555, 0.05126860809326172, 0.04941209411621094, 0.04944384002685547, 0.050871295928955076, 0.0512542724609375, 0.050945022583007815, 0.050836479187011716, 0.0509409294128418, 0.05099827194213867, 0.05123481750488281, 0.050836479187011716, 0.05200998306274414, 0.051138561248779295, 0.050885631561279294, 0.05298175811767578, 0.05130752182006836, 0.050971649169921876, 0.05079654312133789, 0.050852863311767575, 0.05079347229003906, 0.05081292724609375, 0.050971649169921876, 0.10760294342041016, 0.050948097229003904, 0.05115596771240234, 0.05077299118041992, 0.05086515045166016, 0.05120307159423828, 0.05077503967285156, 0.04940697479248047, 0.05072793579101562, 0.05072793579101562, 0.05127884674072265, 0.05120716857910156, 0.05086310577392578, 0.05097062301635742, 0.052278270721435545, 0.05176729583740235, 0.051165184020996096, 0.05074739074707031, 0.05151027297973633, 0.05100953674316406, 0.05159628677368164, 0.05121945571899414, 0.051151870727539066, 0.050964481353759764, 0.05584588623046875, 0.05127475357055664, 0.051141632080078124, 0.050985984802246094, 0.05130547332763672, 0.05127987289428711, 0.049448959350585936, 0.04943462371826172, 0.04945510482788086, 0.04943564987182617, 0.049438720703125, 0.04944179153442383, 0.04946432113647461, 0.049432575225830076, 0.04937932968139649, 0.04948787307739258, 0.05092966461181641, 0.05106585693359375, 0.05104435348510742, 0.05101772689819336, 0.05253836822509766, 0.05126863861083984, 0.05154198455810547, 0.051120128631591794, 0.05107712173461914, 0.051194881439208986, 0.05142323303222656, 0.051672065734863284, 0.05144985580444336, 0.05123072052001953, 0.050907135009765625, 0.05097983932495117, 0.05097369766235352, 0.05148672103881836, 0.05092966461181641, 0.05118361663818359, 0.05110067367553711, 0.05105049514770508, 0.050887680053710936, 0.107578369140625, 0.04943974304199219, 0.04941823959350586, 0.04940492630004883, 0.04948070526123047, 0.051163135528564455, 0.05143756866455078, 0.05128806304931641, 0.0494837760925293, 0.051394561767578124, 0.05057843017578125, 0.051040256500244144, 0.050754558563232424, 0.051299327850341796, 0.05107199859619141, 0.0514785270690918, 0.051664894104003906, 0.051194881439208986, 0.05088460922241211, 0.05104844665527344, 0.050895870208740236, 0.05107097625732422, 0.05106790542602539, 0.05088460922241211, 0.05089996719360351, 0.051507198333740234, 0.05130035018920898, 0.05121023941040039, 0.05156454467773437, 0.05015654373168945, 0.05126144027709961, 0.051227649688720706, 0.05155635070800781, 0.05124710464477539, 0.05112319946289062, 0.05108531188964844, 0.050948097229003904, 0.05116108703613281, 0.051225601196289064, 0.05076070404052734, 0.05082316970825195, 0.051023872375488284, 0.050912254333496096, 0.051323902130126955, 0.05098188781738281, 0.05105868911743164, 0.049426433563232425, 0.05039718246459961, 0.05149491119384766, 0.05088051223754883, 0.05091430282592774, 0.05080678558349609, 0.04967833709716797, 0.049463294982910154, 0.04943462371826172, 0.04944179153442383, 0.04941721725463867, 0.04939263916015625, 0.049465343475341796, 0.05128908920288086, 0.05101772689819336, 0.05122048187255859, 0.049426433563232425, 0.10817330932617188, 0.05112527847290039, 0.05111395263671875, 0.049396736145019535, 0.04938751983642578, 0.049442817687988284, 0.049432575225830076, 0.04941209411621094, 0.04939571380615235, 0.04947763061523437, 0.052209663391113284, 0.05124505615234375, 0.05123788833618164, 0.05096345520019531, 0.05117542266845703, 0.05103923034667969, 0.05078220748901367, 0.04946944046020508, 0.04948070526123047, 0.049442817687988284, 0.04948992156982422, 0.04941823959350586, 0.04941926574707031, 0.0494571533203125, 0.051059711456298826, 0.051053569793701174, 0.05084774398803711, 0.050769920349121096, 0.05172326278686523, 0.05092147064208984, 0.051792896270751954, 0.05111603164672852, 0.05066649627685547, 0.05086822509765625, 0.05097574234008789, 0.04945510482788086, 0.05036441421508789, 0.05094400024414063, 0.050974720001220705, 0.050918399810791014, 0.05118463897705078, 0.05108736038208008, 0.050945022583007815, 0.051176448822021485, 0.05111500930786133, 0.050874366760253906, 0.05081190490722656, 0.051553279876708984, 0.050920448303222655, 0.051211265563964846, 0.051043327331542966, 0.05098188781738281, 0.050902015686035154, 0.05135257720947266, 0.05083443069458008, 0.05088460922241211, 0.051383296966552736, 0.050955265045166016, 0.05091328048706055, 0.0528988151550293, 0.05152767944335938, 0.050667518615722655, 0.051163135528564455, 0.10756095886230468, 0.04947353744506836, 0.05100851058959961, 0.05118873596191406, 0.05105561447143555, 0.05143756866455078, 0.05097369766235352, 0.05106790542602539, 0.05102284622192383, 0.04944384002685547, 0.04942438507080078, 0.05131468963623047, 0.05090508651733398, 0.05114470291137695, 0.050956287384033204, 0.051062782287597655, 0.050941951751708986, 0.05073920059204102, 0.05122867202758789, 0.052566017150878906, 0.05120000076293945, 0.050985984802246094, 0.05074943923950195, 0.051637248992919924, 0.05092454528808594, 0.051023872375488284, 0.05079040145874023, 0.05143142318725586, 0.052466686248779294, 0.054403072357177736, 0.05147750473022461, 0.05167923355102539, 0.051525630950927735, 0.051020801544189455, 0.0509224967956543, 0.049452030181884765, 0.05125836944580078, 0.05112627029418945, 0.051133438110351564, 0.05150003051757813, 0.050966529846191405, 0.05105561447143555, 0.0518205451965332, 0.05099008178710938, 0.05095219039916992, 0.05101772689819336, 0.051143680572509766, 0.05095219039916992, 0.05204070281982422, 0.05122969436645508, 0.051095550537109374, 0.05101260757446289, 0.05126348876953125, 0.05101875305175781, 0.05103615951538086, 0.05110681533813476, 0.051062782287597655, 0.04968550491333008, 0.05125734329223633, 0.051046398162841795, 0.051046398162841795, 0.05125120162963867, 0.05106995010375977, 0.10760601806640625, 0.04944489669799805, 0.049420257568359376, 0.04939878463745117, 0.04941823959350586, 0.04941823959350586, 0.04943564987182617, 0.04941312026977539, 0.04940390396118164, 0.04941209411621094, 0.049438720703125, 0.04937625503540039, 0.04954214477539062, 0.04943564987182617, 0.049430526733398435, 0.049414142608642575, 0.04938854217529297, 0.04948070526123047, 0.04946636962890625, 0.049479679107666014, 0.049426433563232425, 0.050918399810791014, 0.05172019195556641, 0.051244033813476565, 0.05097881698608398, 0.05135974502563476, 0.05120000076293945, 0.051197952270507815, 0.05117337417602539, 0.05151948928833008, 0.05120716857910156, 0.05105868911743164, 0.05154304122924805, 0.051465217590332034, 0.05145804977416992, 0.0494202880859375, 0.05116723251342774, 0.05136896133422852, 0.04941107177734375, 0.04948992156982422, 0.049481727600097655, 0.04981145477294922, 0.05076172637939453, 0.050971649169921876, 0.049465343475341796, 0.049459201812744144, 0.049426433563232425, 0.049438720703125, 0.04942438507080078, 0.049438720703125, 0.049460224151611325, 0.049430526733398435, 0.04942233657836914, 0.049468414306640625, 0.04944588851928711, 0.04944793701171875, 0.049460224151611325, 0.04948992156982422, 0.04944384002685547, 0.04944384002685547, 0.04941312026977539, 0.04947763061523437, 0.04947353744506836, 0.10748416137695313, 0.0501923828125, 0.04940492630004883, 0.049432575225830076, 0.04942950439453125, 0.049426433563232425, 0.04947148895263672, 0.049470462799072266, 0.04944591903686523, 0.04973360061645508, 0.05122867202758789, 0.05102592086791992, 0.051074047088623044, 0.051043327331542966, 0.05106585693359375, 0.049470462799072266, 0.04944076919555664, 0.049438720703125, 0.049522689819335934, 0.049442817687988284, 0.049408000946044923, 0.04944588851928711, 0.049459201812744144, 0.0494284782409668, 0.04943667221069336, 0.049470462799072266, 0.04946124649047851, 0.049465343475341796, 0.04946739196777344, 0.049460224151611325, 0.04947251129150391, 0.04943462371826172, 0.049438720703125, 0.04942131042480469, 0.04940185546875, 0.049460224151611325, 0.04942745590209961, 0.04943974304199219, 0.049442817687988284, 0.04940902328491211, 0.04944998550415039, 0.049495040893554686, 0.04942438507080078, 0.049438720703125, 0.049442817687988284, 0.049465343475341796, 0.04943360137939453, 0.04944486236572266, 0.049479679107666014, 0.049498111724853515, 0.04940902328491211, 0.04941209411621094, 0.04937625503540039, 0.04946432113647461, 0.050477054595947264, 0.050948097229003904, 0.050958335876464846, 0.05120819091796875, 0.049452030181884765, 0.0494284782409668, 0.04939059066772461, 0.04947148895263672, 0.04942950439453125]",tokens/s,19.418597405445666,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,v,v,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14233,7 +14233,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d467a-1eb50b0f339c76d03ab9bb90;72474fca-b4bf-4d03-8c00-24fcbc912629) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe9a3-6532ff563b125d084de04fbd;631e7196-fbd4-41de-9557-4320a841456f) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14332,7 +14332,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9595-716eba2b3ba2a34875369e3e;5370016d-7d1d-4462-b478-545e6329e5d0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe764-2d2705423086c1ad3d3dcc8e;00b31656-a694-4d23-abf8-62672d38a947) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14431,7 +14431,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9639-4361cc7b27ebaa953292b6ea;b887b6e2-1626-4409-96fd-5a04c212c292) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe807-7b44f7c855e364a040858d45;f9e7a507-8639-44f7-8fb1-ddb88b66e9df) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14463,7 +14463,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-7B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,5005.856768,8173.125632,0.0,7526.678528,6833.613824,s,10,3.237135192871093,0.32371351928710934,0.0011265830391930785,0.32340528869628904,0.3252778289794922,0.32533260955810545,0.3253764340209961,"[0.3226540832519531, 0.32538739013671875, 0.3225751647949219, 0.32281320190429685, 0.323228515625, 0.3222604675292969, 0.32358206176757814, 0.32464373779296873, 0.32526565551757813, 0.32472491455078123]",tokens/s,790.8227020106239,kWh,3.805233683023188e-06,2.085111826499997e-06,1.9868748533876518e-05,2.57590940433997e-05,tokens/kWh,9938237.717859311,MB,5005.856768,8173.125632,0.0,7526.678528,7110.583808,s,10,176.72496484375003,17.672496484375003,0.0017080130721937633,17.6719814453125,17.674641601562502,17.67482177734375,17.674965917968752,"[17.67114453125, 17.670958984375, 17.673966796875, 17.671052734375, 17.6746015625, 17.675001953125, 17.67048828125, 17.67075, 17.672818359375, 17.674181640625]",tokens/s,3.5648613683811425,kWh,0.00020860287739584842,0.00011433028744598232,0.00106939106731773,0.0013923242321595608,tokens/kWh,45248.081262138214,,s,629,179.35808807373044,0.2851479937579181,0.038540759752562596,0.2804643859863281,0.2806615966796875,0.280817041015625,0.6038926318359376,"[0.2805963134765625, 0.28048486328125, 0.28035687255859376, 0.28048077392578125, 0.2802708435058594, 0.28063641357421876, 0.280880126953125, 0.2805626831054687, 0.28056576538085937, 0.28044699096679687, 0.28062106323242186, 0.2806651000976563, 0.2804029541015625, 0.28052276611328125, 0.28037017822265625, 0.28041830444335936, 0.28037835693359375, 0.28043264770507814, 0.2805278625488281, 0.28084326171875, 0.280748046875, 0.28032305908203126, 0.2804572143554688, 0.2806138916015625, 0.2806814575195313, 0.2805565490722656, 0.28059237670898435, 0.28052276611328125, 0.2805063781738281, 0.2803814697265625, 0.28054525756835935, 0.2805565490722656, 0.28050546264648435, 0.280545166015625, 0.2803732604980469, 0.2802689514160156, 0.2804836730957031, 0.28037631225585935, 0.28042648315429686, 0.2804510803222656, 0.28022171020507813, 0.28037530517578124, 0.2806087646484375, 0.2803619689941406, 0.2803292236328125, 0.28030975341796877, 0.28040704345703127, 0.2805073852539062, 0.2803589172363281, 0.2804869079589844, 0.28044699096679687, 0.2804418640136719, 0.28030670166015625, 0.28054425048828124, 0.2804132080078125, 0.2802564697265625, 0.28043878173828124, 0.28012237548828123, 0.2805473327636719, 0.28069381713867186, 0.28056365966796876, 0.28044390869140623, 0.606454833984375, 0.28040594482421877, 0.28052581787109376, 0.2802616271972656, 0.28048382568359376, 0.2802401428222656, 0.28023602294921873, 0.28026571655273436, 0.2802964477539063, 0.28036309814453125, 0.28032196044921875, 0.28032000732421875, 0.28035482788085936, 0.28039883422851564, 0.2804223937988281, 0.2805524597167969, 0.28039678955078123, 0.280553466796875, 0.28060159301757814, 0.28041012573242186, 0.2802832336425781, 0.2802912292480469, 0.2802391052246094, 0.2806220703125, 0.2804019165039062, 0.28050433349609377, 0.28030157470703126, 0.2803660888671875, 0.2802759704589844, 0.28037631225585935, 0.28027188110351564, 0.2805688171386719, 0.2806599731445312, 0.280518798828125, 0.28085235595703123, 0.2807992248535156, 0.2804234313964844, 0.28058111572265626, 0.2803384399414062, 0.2803363952636719, 0.2804286193847656, 0.28044281005859373, 0.2804090881347656, 0.2804664306640625, 0.2803998718261719, 0.2803660888671875, 0.2832220153808594, 0.2804951171875, 0.28065899658203125, 0.28072747802734377, 0.2807408752441406, 0.280563720703125, 0.2804142150878906, 0.28033535766601564, 0.28055755615234373, 0.280489990234375, 0.2803722229003906, 0.28031402587890625, 0.28017034912109373, 0.2805073852539062, 0.28048895263671875, 0.2806773681640625, 0.2802821044921875, 0.6037001953125, 0.280263671875, 0.28055755615234373, 0.2806200256347656, 0.28048486328125, 0.28029132080078123, 0.2803927001953125, 0.28048077392578125, 0.28041522216796877, 0.2804715576171875, 0.28054425048828124, 0.2804725646972656, 0.28047360229492185, 0.2803722229003906, 0.280458251953125, 0.2804316101074219, 0.28048895263671875, 0.2804490356445313, 0.28044699096679687, 0.28046234130859377, 0.28052377319335936, 0.280511474609375, 0.28041830444335936, 0.28209765625, 0.2803117980957031, 0.2805626831054687, 0.2804756774902344, 0.28051861572265624, 0.28057907104492186, 0.2806241149902344, 0.2804572143554688, 0.28048281860351565, 0.2805565490722656, 0.28042245483398437, 0.28039572143554686, 0.28050433349609377, 0.2804633483886719, 0.2805053405761719, 0.28054629516601565, 0.2807029724121094, 0.2806098022460938, 0.2805504150390625, 0.2805821533203125, 0.28057089233398436, 0.28033444213867187, 0.28032501220703127, 0.28029541015625, 0.2819450988769531, 0.2804808349609375, 0.28046636962890625, 0.28050225830078124, 0.2803804016113281, 0.2805422058105469, 0.28032211303710936, 0.28040594482421877, 0.28035174560546877, 0.28062924194335936, 0.28064666748046874, 0.28025140380859376, 0.2804951171875, 0.2808965148925781, 0.2807439270019531, 0.2803025817871094, 0.6039674682617188, 0.2802984924316406, 0.2803036193847656, 0.2803732604980469, 0.2804992065429687, 0.2802586059570312, 0.28033941650390626, 0.2803660888671875, 0.28037530517578124, 0.28076544189453123, 0.2807685241699219, 0.2807357482910156, 0.2816512145996094, 0.2805882873535156, 0.28059341430664064, 0.28039678955078123, 0.2803189697265625, 0.280300537109375, 0.2803343505859375, 0.28028619384765624, 0.28039883422851564, 0.2803978271484375, 0.28032101440429685, 0.2804572143554688, 0.2802882690429688, 0.280342529296875, 0.2803343505859375, 0.28048794555664064, 0.2802626647949219, 0.28064364624023436, 0.2803271179199219, 0.2806200256347656, 0.28050433349609377, 0.280521728515625, 0.2802626647949219, 0.2803947448730469, 0.28026776123046876, 0.2804962463378906, 0.28034445190429685, 0.2803445739746094, 0.2804142150878906, 0.2802759704589844, 0.28033740234375, 0.28036300659179686, 0.2823341979980469, 0.2804653930664063, 0.28042547607421875, 0.2806005859375, 0.28057089233398436, 0.280553466796875, 0.28051046752929687, 0.28042752075195315, 0.28049099731445315, 0.28036505126953126, 0.2804695129394531, 0.28058724975585936, 0.28055960083007814, 0.28043777465820313, 0.2803456115722656, 0.28051864624023437, 0.28044082641601564, 0.2805299072265625, 0.28034048461914063, 0.6044927978515625, 0.28044390869140623, 0.28047564697265626, 0.2803916931152344, 0.2804715576171875, 0.2804223937988281, 0.28053607177734374, 0.2804244384765625, 0.28046029663085936, 0.2805473327636719, 0.28041830444335936, 0.2804009094238281, 0.28044287109375, 0.28053707885742185, 0.2805196838378906, 0.28043060302734374, 0.2806947937011719, 0.28244479370117187, 0.2806507568359375, 0.2805596923828125, 0.2804827575683594, 0.2806190185546875, 0.2803179626464844, 0.280352783203125, 0.2803558349609375, 0.28041830444335936, 0.28029541015625, 0.2804592590332031, 0.2803466186523437, 0.2804244384765625, 0.2802903137207031, 0.28051763916015626, 0.28042752075195315, 0.28032205200195315, 0.28036300659179686, 0.2805196838378906, 0.28048178100585935, 0.28048178100585935, 0.2802964477539063, 0.28041317749023437, 0.2806640625, 0.2805135498046875, 0.28036300659179686, 0.28051776123046873, 0.2803956604003906, 0.28043777465820313, 0.2803589172363281, 0.28044082641601564, 0.28042547607421875, 0.28058111572265626, 0.2805688171386719, 0.2803609619140625, 0.2805853576660156, 0.28058303833007814, 0.28053094482421875, 0.2804971618652344, 0.28056781005859377, 0.28342782592773436, 0.2802606201171875, 0.280511474609375, 0.28044699096679687, 0.28046234130859377, 0.28032101440429685, 0.6034566040039062, 0.2803240051269531, 0.28044287109375, 0.28026470947265625, 0.28046646118164065, 0.2802431640625, 0.2804776916503906, 0.28053912353515625, 0.280553466796875, 0.2805084228515625, 0.28041317749023437, 0.280406005859375, 0.28044699096679687, 0.2805145568847656, 0.28044000244140627, 0.2804590759277344, 0.2803117980957031, 0.28043060302734374, 0.2803947448730469, 0.28041317749023437, 0.2804346923828125, 0.28048281860351565, 0.2803292236328125, 0.28043060302734374, 0.2803916931152344, 0.28090573120117185, 0.281034912109375, 0.2807693786621094, 0.2806179809570313, 0.2805002136230469, 0.2804552001953125, 0.28051556396484373, 0.2805329895019531, 0.28052890014648435, 0.2803896179199219, 0.2805248107910156, 0.28049612426757814, 0.28070196533203123, 0.2834462585449219, 0.28048492431640626, 0.28036190795898436, 0.28037118530273436, 0.2805299072265625, 0.28084225463867185, 0.28050128173828126, 0.2805862121582031, 0.2804244384765625, 0.2804561767578125, 0.2806599731445312, 0.28053607177734374, 0.28061489868164063, 0.28050335693359374, 0.28050936889648437, 0.2803722229003906, 0.2803978271484375, 0.28043777465820313, 0.2805206909179688, 0.2805084228515625, 0.2803292236328125, 0.28083203125, 0.28075106811523437, 0.2806098022460938, 0.2802431945800781, 0.6047293701171875, 0.2803046264648438, 0.28038552856445315, 0.2803200378417969, 0.28035171508789064, 0.280205322265625, 0.28043878173828124, 0.280405029296875, 0.2803721923828125, 0.2802923583984375, 0.2804090881347656, 0.2803650817871094, 0.2803547973632812, 0.2803897705078125, 0.28044378662109376, 0.28044082641601564, 0.28047360229492185, 0.28044699096679687, 0.28070501708984374, 0.2806456298828125, 0.280531982421875, 0.2805053405761719, 0.2803169250488281, 0.2821160888671875, 0.2803343505859375, 0.2804653930664063, 0.2802472839355469, 0.28062310791015627, 0.2803384399414062, 0.28051559448242186, 0.28032513427734373, 0.2805145568847656, 0.28086477661132814, 0.2805626831054687, 0.28067840576171876, 0.28092825317382814, 0.28034866333007813, 0.280458251953125, 0.28032205200195315, 0.28049615478515627, 0.28045614624023435, 0.2804019165039062, 0.28041317749023437, 0.2805063781738281, 0.28031900024414064, 0.2803783264160156, 0.2802769775390625, 0.2803916931152344, 0.2805688171386719, 0.2805893249511719, 0.2807664489746094, 0.28051864624023437, 0.280453125, 0.28029440307617187, 0.28048794555664064, 0.2804981689453125, 0.28035687255859376, 0.28037530517578124, 0.2801438598632813, 0.280474609375, 0.2804223937988281, 0.28052276611328125, 0.2804583129882813, 0.605824951171875, 0.28028619384765624, 0.2806384582519531, 0.2806241149902344, 0.2806609802246094, 0.280263671875, 0.2804490356445313, 0.2804356994628906, 0.2804572143554688, 0.2802821044921875, 0.28043264770507814, 0.2805053405761719, 0.28048794555664064, 0.28051864624023437, 0.280521728515625, 0.28047052001953127, 0.28164608764648436, 0.2804326782226563, 0.2803670654296875, 0.2803937377929687, 0.2804633483886719, 0.2804572143554688, 0.2804356994628906, 0.2803651123046875, 0.2801980895996094, 0.2803189697265625, 0.28040606689453124, 0.2805523986816406, 0.28038861083984373, 0.2804223937988281, 0.28030670166015625, 0.28060159301757814, 0.28061489868164063, 0.28049099731445315, 0.2803456115722656, 0.2804869079589844, 0.2803640441894531, 0.28059854125976563, 0.280479736328125, 0.28065484619140624, 0.2805299072265625, 0.2803129577636719, 0.28043557739257813, 0.2806098022460938, 0.28044595336914063, 0.28049612426757814, 0.28042752075195315, 0.28031488037109376, 0.2805555114746094, 0.28026571655273436, 0.28036813354492185, 0.28044287109375, 0.280352783203125, 0.28028005981445314, 0.28044287109375, 0.28049612426757814, 0.2805616760253906, 0.28052276611328125, 0.280416259765625, 0.28058624267578125, 0.28044082641601564, 0.28082891845703123, 0.28066815185546873, 0.6064496459960937, 0.28037631225585935, 0.28064358520507815, 0.2803916931152344, 0.28039883422851564, 0.28029336547851563, 0.2804223937988281, 0.2804695129394531, 0.28044287109375, 0.2806394958496094, 0.28072970581054685, 0.28055029296875, 0.28057598876953127, 0.2806005859375, 0.28058624267578125, 0.280553466796875, 0.2804592590332031, 0.2818191223144531, 0.2805145568847656, 0.28053912353515625, 0.28071633911132815, 0.2805329895019531, 0.28040093994140625, 0.2804130859375, 0.28032000732421875, 0.2806302795410156, 0.28032205200195315, 0.2805473327636719, 0.28040499877929687, 0.2805749816894531, 0.2803865661621094, 0.2806026306152344, 0.28049099731445315, 0.2805350341796875, 0.2804643859863281, 0.2805053405761719, 0.28049099731445315, 0.2805206909179688, 0.28050125122070313, 0.28054937744140623, 0.2804787292480469, 0.28038861083984373, 0.28054425048828124, 0.2805473327636719, 0.280352783203125, 0.28046847534179686, 0.2804111328125, 0.28038348388671874, 0.2803814392089844, 0.2805504150390625, 0.28066815185546873, 0.2804643859863281, 0.28052581787109376, 0.28044595336914063, 0.2804920349121094, 0.2805177001953125, 0.2804632873535156, 0.280416259765625, 0.28042034912109376, 0.2804696044921875, 0.2805472412109375, 0.2804971618652344, 0.28035687255859376, 0.6061209716796875, 0.28034356689453127, 0.28048382568359376, 0.2802329711914063, 0.28037835693359375, 0.28030975341796877, 0.2804234313964844, 0.2804244384765625, 0.2806826477050781, 0.2807540283203125, 0.28043264770507814, 0.280531982421875, 0.2804561767578125, 0.28045208740234373, 0.2804736633300781, 0.2804694519042969, 0.28052276611328125, 0.2806302795410156, 0.2805452880859375, 0.2804142150878906, 0.28073370361328126, 0.2807664489746094, 0.28032818603515625, 0.280453125, 0.2803537902832031, 0.28041934204101565, 0.28231680297851564, 0.2805032958984375, 0.2803619689941406, 0.280521728515625, 0.2803240966796875, 0.28040704345703127, 0.2804725646972656, 0.2804316101074219, 0.2803804016113281, 0.2805718994140625, 0.2806169738769531, 0.28048281860351565, 0.2804500427246094, 0.28053912353515625, 0.28075314331054685, 0.2806179809570313, 0.2803937377929687, 0.28062310791015627, 0.2805350341796875, 0.2806814575195313, 0.2804490356445313, 0.2805248107910156, 0.2805606384277344, 0.2803978271484375, 0.28048281860351565, 0.28046234130859377, 0.2805616760253906, 0.28059033203125, 0.28056781005859377, 0.28047564697265626, 0.28059033203125, 0.28059442138671875, 0.28028619384765624, 0.28051763916015626, 0.28055960083007814, 0.28064358520507815, 0.28044390869140623]",tokens/s,3.5069508532084197,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-7b,tiiuae/falcon-7b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14493,7 +14493,7 @@ ValueError: FalconForCausalLM does not support an attention implementation throu ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4660.158464,21498.42944,0.0,20851.982336,20236.60032,s,10,15.67305651855469,1.5673056518554689,0.002935213598468352,1.5664246826171875,1.5689432861328125,1.5722443481445312,1.5748851977539062,"[1.5660777587890624, 1.5652294921875, 1.5653187255859375, 1.56764404296875, 1.5667716064453125, 1.565794677734375, 1.568209716796875, 1.5673927001953125, 1.57554541015625, 1.5650723876953125]",tokens/s,163.3376359594774,kWh,1.8488901803890867e-05,1.0131977631626796e-05,9.191171241820228e-05,0.00012053259185371994,tokens/kWh,2123906.8708543596,MB,4660.158464,21498.42944,0.0,20851.982336,20339.706368,s,10,881.3746874999999,88.13746875000001,0.00269603704029487,88.13719531250001,88.13986015625,88.141726953125,88.14322039062499,"[88.1357109375, 88.1359765625, 88.13715625, 88.137234375, 88.1394453125, 88.1378203125, 88.1392890625, 88.1333984375, 88.14359375, 88.1350625]",tokens/s,0.7147924814893211,kWh,0.001040502367599143,0.0005702873234177241,0.005180215005279792,0.006791004696296659,tokens/kWh,9276.97782838463,,s,629,894.0823742675777,1.4214346172775487,0.18625247039114828,1.3989376220703125,1.3992808349609374,1.3994753662109376,2.9653164257812503,"[1.39905224609375, 1.3987666015625, 1.3988515625, 1.3986263427734376, 1.398855712890625, 1.3987164306640625, 1.3989693603515625, 1.39875537109375, 1.3986160888671875, 1.3986785888671875, 1.3988690185546875, 1.39879833984375, 1.398836181640625, 1.3988291015625, 1.398834228515625, 1.39947412109375, 1.3986580810546876, 1.398688720703125, 1.398931396484375, 1.3984696044921876, 1.398876220703125, 1.398803466796875, 1.399109619140625, 1.3990972900390626, 1.3988321533203125, 1.3986181640625, 1.398677490234375, 1.3991065673828125, 1.3987225341796874, 1.39873486328125, 1.398603759765625, 1.3990164794921875, 1.398830078125, 1.3991229248046875, 1.3992918701171875, 1.4000015869140625, 1.3992899169921875, 1.3990687255859375, 1.3990062255859375, 1.399046142578125, 1.3990379638671875, 1.3989140625, 1.3989693603515625, 1.398824951171875, 1.3988116455078126, 1.3989478759765626, 1.3991649169921876, 1.3989530029296875, 1.399182373046875, 1.3990277099609374, 1.399468017578125, 1.399045166015625, 1.3990164794921875, 1.3987860107421874, 1.399066650390625, 1.3993809814453124, 1.3991322021484376, 1.399140380859375, 1.3991536865234375, 1.3990225830078125, 1.3993870849609376, 1.3994906005859375, 2.964931640625, 1.3986201171875, 1.3987666015625, 1.39865087890625, 1.3985853271484374, 1.398708251953125, 1.39879833984375, 1.398951904296875, 1.398729736328125, 1.399015380859375, 1.3991639404296874, 1.39902978515625, 1.3989161376953125, 1.3987215576171874, 1.3987850341796875, 1.3988013916015625, 1.3988004150390625, 1.3990830078125, 1.3986212158203124, 1.3988321533203125, 1.398856689453125, 1.3988372802734375, 1.398662109375, 1.3987337646484375, 1.398878173828125, 1.39877783203125, 1.398846435546875, 1.398877197265625, 1.4004705810546876, 1.3986488037109375, 1.399041015625, 1.398792236328125, 1.3986795654296875, 1.398814697265625, 1.3986632080078125, 1.398877197265625, 1.399088134765625, 1.398846435546875, 1.398697998046875, 1.398998046875, 1.398794189453125, 1.3989376220703125, 1.3990277099609374, 1.398767578125, 1.398866943359375, 1.398981689453125, 1.4009364013671874, 1.39877587890625, 1.399131103515625, 1.3991177978515625, 1.3994281005859375, 1.399003173828125, 1.398744140625, 1.3990062255859375, 1.398962158203125, 1.3992396240234375, 1.3992283935546874, 1.3991383056640625, 1.3989990234375, 1.39947314453125, 1.399541748046875, 1.3994761962890625, 1.39911474609375, 2.96489990234375, 1.3986221923828126, 1.3986396484375, 1.3987462158203126, 1.398728759765625, 1.3984818115234374, 1.3987225341796874, 1.3987666015625, 1.3989918212890624, 1.3991875, 1.398984619140625, 1.399055419921875, 1.39894482421875, 1.3992437744140624, 1.3988853759765625, 1.3988741455078124, 1.3987921142578126, 1.3989744873046874, 1.4005791015625, 1.3989959716796876, 1.3986417236328126, 1.39873583984375, 1.3984593505859375, 1.3988751220703124, 1.398887451171875, 1.3985894775390626, 1.3989530029296875, 1.3985628662109375, 1.398551513671875, 1.398708251953125, 1.3987327880859375, 1.3988291015625, 1.3988648681640625, 1.3988126220703125, 1.3988331298828125, 1.3991240234375, 1.3989898681640625, 1.3990860595703125, 1.3987808837890625, 1.3990594482421874, 1.3988106689453126, 1.398973388671875, 1.401247802734375, 1.398792236328125, 1.3992130126953124, 1.399266357421875, 1.3992427978515625, 1.399146484375, 1.398971435546875, 1.398993896484375, 1.3989744873046874, 1.3990277099609374, 1.3990123291015626, 1.3992764892578125, 1.39936767578125, 1.39896826171875, 1.399108642578125, 1.3990491943359376, 1.399204833984375, 1.3993646240234374, 1.399277587890625, 1.399076904296875, 1.39888330078125, 2.967499755859375, 1.3984256591796875, 1.3987880859375, 1.3985330810546874, 1.39865087890625, 1.3985106201171875, 1.3986744384765626, 1.3990687255859375, 1.398646728515625, 1.3988494873046875, 1.3989488525390625, 1.3988157958984375, 1.4003507080078126, 1.3992181396484376, 1.3986282958984375, 1.3989796142578126, 1.3991424560546875, 1.3989110107421876, 1.39889453125, 1.398781982421875, 1.398676513671875, 1.39867236328125, 1.3985843505859374, 1.3988526611328125, 1.398802490234375, 1.3987880859375, 1.3986365966796874, 1.398794189453125, 1.3989488525390625, 1.3988382568359374, 1.398718505859375, 1.3987451171875, 1.3987327880859375, 1.398834228515625, 1.3988956298828126, 1.3989713134765625, 1.3988956298828126, 1.3992642822265624, 1.398856689453125, 1.399077880859375, 1.39888232421875, 1.39892529296875, 1.3987440185546876, 1.3989969482421876, 1.4024007568359376, 1.399172119140625, 1.3992745361328125, 1.3992008056640626, 1.3990318603515626, 1.398909912109375, 1.3992008056640626, 1.3993646240234374, 1.39901953125, 1.399173095703125, 1.3989775390625, 1.399046142578125, 1.399246826171875, 1.3991126708984376, 1.399083984375, 1.399047119140625, 1.3991341552734375, 1.399214111328125, 1.3988863525390625, 2.96650537109375, 1.3986048583984374, 1.3984266357421875, 1.39875537109375, 1.3986641845703125, 1.398703125, 1.3987225341796874, 1.39884130859375, 1.3988424072265624, 1.398752197265625, 1.398993896484375, 1.399047119140625, 1.3987225341796874, 1.39919775390625, 1.3986856689453124, 1.399119873046875, 1.3985965576171875, 1.400858642578125, 1.398835205078125, 1.3988424072265624, 1.3985096435546875, 1.3990072021484374, 1.3985545654296876, 1.3986754150390626, 1.3990328369140625, 1.3986468505859375, 1.3989324951171875, 1.3988331298828125, 1.3989283447265626, 1.398687744140625, 1.39865087890625, 1.3986221923828126, 1.3990799560546876, 1.3989017333984375, 1.39886083984375, 1.39890283203125, 1.39919873046875, 1.3989652099609375, 1.39868359375, 1.398814697265625, 1.3989857177734375, 1.3991802978515624, 1.3991793212890624, 1.3992017822265626, 1.398865966796875, 1.39908203125, 1.39892529296875, 1.399057373046875, 1.3991065673828125, 1.39901953125, 1.4028165283203125, 1.399258056640625, 1.39905029296875, 1.399616455078125, 1.3996502685546874, 1.399204833984375, 1.399150634765625, 1.399098388671875, 1.3990655517578126, 1.3990645751953126, 1.3991474609375, 1.3992633056640624, 1.39909423828125, 2.965466064453125, 1.398435791015625, 1.3989212646484375, 1.3987410888671874, 1.398793212890625, 1.398793212890625, 1.3986417236328126, 1.398877197265625, 1.398681640625, 1.398823974609375, 1.398592529296875, 1.398836181640625, 1.3988443603515626, 1.398813720703125, 1.398644775390625, 1.398635498046875, 1.3989150390625, 1.3987430419921876, 1.399067626953125, 1.399125, 1.398625244140625, 1.3987430419921876, 1.3987952880859376, 1.3990604248046874, 1.398804443359375, 1.3987532958984374, 1.39879736328125, 1.3987962646484375, 1.398583251953125, 1.398846435546875, 1.398750244140625, 1.3987994384765625, 1.3991485595703126, 1.3989007568359375, 1.3991126708984376, 1.3992816162109376, 1.3991004638671876, 1.3990328369140625, 1.4033602294921874, 1.398982666015625, 1.3990400390625, 1.3989600830078126, 1.3988065185546874, 1.3990594482421874, 1.398823974609375, 1.3988914794921874, 1.398825927734375, 1.3992816162109376, 1.3990482177734376, 1.3992335205078126, 1.3989591064453124, 1.399271484375, 1.3991065673828125, 1.399330810546875, 1.398982666015625, 1.3989150390625, 1.39902978515625, 1.3993175048828126, 1.399251953125, 1.3993543701171876, 1.399087158203125, 1.399087158203125, 1.39928271484375, 2.967877685546875, 1.399, 1.3988577880859374, 1.398720458984375, 1.3989058837890624, 1.3989591064453124, 1.398603759765625, 1.398509521484375, 1.3988433837890626, 1.398835205078125, 1.398892578125, 1.3990687255859375, 1.3989376220703125, 1.3988812255859375, 1.3986109619140625, 1.398836181640625, 1.3987666015625, 1.3988116455078126, 1.398899658203125, 1.398931396484375, 1.3985228271484376, 1.398972412109375, 1.3989959716796876, 1.3989703369140625, 1.3989365234375, 1.3988116455078126, 1.398856689453125, 1.39875537109375, 1.3987542724609374, 1.398794189453125, 1.3987860107421874, 1.3986048583984374, 1.3989365234375, 1.398892578125, 1.3988424072265624, 1.398866943359375, 1.4031943359375, 1.399446533203125, 1.3989171142578125, 1.3989058837890624, 1.39900927734375, 1.3987225341796874, 1.399125, 1.39943115234375, 1.3992969970703124, 1.398981689453125, 1.399066650390625, 1.39924169921875, 1.39911474609375, 1.3990235595703124, 1.3991434326171874, 1.3992274169921874, 1.398771728515625, 1.3991884765625, 1.3990921630859374, 1.3990921630859374, 1.3991424560546875, 1.399141357421875, 1.399182373046875, 1.399150634765625, 1.399173095703125, 1.3995130615234375, 1.3992161865234376, 2.966830078125, 1.39867236328125, 1.3989918212890624, 1.3987451171875, 1.399013427734375, 1.3989908447265624, 1.3986611328125, 1.3989530029296875, 1.3987686767578125, 1.3985955810546875, 1.3984910888671875, 1.398908935546875, 1.3986058349609376, 1.3991700439453125, 1.3991044921875, 1.3990963134765626, 1.398783935546875, 1.3988414306640624, 1.39865087890625, 1.398973388671875, 1.398529052734375, 1.398624267578125, 1.3986068115234376, 1.39877783203125, 1.3988494873046875, 1.398686767578125, 1.3989161376953125, 1.3988074951171876, 1.398720458984375, 1.3985545654296876, 1.398549560546875, 1.398677490234375, 1.399013427734375, 1.3988321533203125, 1.3990225830078125, 1.3991875, 1.399309326171875, 1.3990697021484375, 1.3987850341796875, 1.3989110107421876, 1.3988515625, 1.39879833984375, 1.3990164794921875, 1.3991669921875, 1.3989970703125, 1.398814697265625, 1.3990440673828124, 1.3993902587890625, 1.39905029296875, 1.3986529541015624, 1.399108642578125, 1.3995069580078126, 1.3990020751953125, 1.3991290283203126, 1.399109619140625, 1.3991793212890624, 1.3990748291015624, 1.3990533447265625, 1.399342041015625, 1.3992509765625, 1.3991875, 1.3994813232421874, 1.3992847900390626, 2.975549560546875, 1.3990369873046875, 1.398686767578125, 1.3988699951171875, 1.3988116455078126, 1.398876220703125, 1.398803466796875, 1.398488037109375, 1.398929443359375, 1.398972412109375, 1.398601806640625, 1.3989036865234374, 1.3985341796875, 1.398978515625, 1.398749267578125, 1.398830078125, 1.3989969482421876, 1.398941650390625, 1.3987880859375, 1.398920166015625, 1.3986058349609376, 1.3989754638671874, 1.3987789306640626, 1.3987738037109374, 1.398740966796875, 1.39875537109375, 1.398782958984375, 1.398865966796875, 1.3988935546875, 1.3988218994140624, 1.39879931640625, 1.398435791015625, 1.39890283203125, 1.39907275390625, 1.3990400390625, 1.39919970703125, 1.3989857177734375, 1.39919873046875, 1.399015380859375, 1.3991219482421875, 1.3990953369140624, 1.3992161865234376, 1.39879736328125, 1.399045166015625, 1.3990635986328126, 1.3989110107421876, 1.3989007568359375, 1.3991126708984376, 1.3991710205078125, 1.3991290283203126, 1.3995233154296876, 1.399393310546875, 1.4048031005859376, 1.3995601806640625, 1.39926318359375, 1.399203857421875, 1.399208984375, 1.3992806396484374, 1.3992008056640626, 1.3996329345703125, 1.399182373046875, 1.3990205078125, 1.399141357421875, 2.966223876953125, 1.398709228515625, 1.399329833984375, 1.3987266845703126, 1.398455322265625, 1.398873046875, 1.39873583984375, 1.39871533203125, 1.39875537109375, 1.3985904541015626, 1.398762451171875, 1.398834228515625, 1.3986702880859374, 1.3991434326171874, 1.3988095703125, 1.3991116943359374, 1.398962158203125, 1.3986590576171876, 1.398698974609375, 1.398835205078125, 1.39871435546875, 1.3988116455078126, 1.398834228515625, 1.398676513671875, 1.3988095703125, 1.398972412109375, 1.399234619140625, 1.3989007568359375, 1.3986314697265625, 1.3986949462890625, 1.398951904296875, 1.3991597900390624, 1.3991392822265625, 1.3987696533203124, 1.3990225830078125, 1.398940673828125, 1.3991015625, 1.3990347900390625, 1.39875634765625, 1.3991424560546875, 1.3989007568359375, 1.3987020263671874, 1.3989140625, 1.399394287109375, 1.3992960205078124, 1.3988515625, 1.3991485595703126, 1.399234619140625, 1.398983642578125, 1.3988310546875, 1.3990318603515626, 1.3990963134765626, 1.3992110595703124, 1.399140380859375, 1.3990972900390626, 1.39926123046875, 1.3992274169921874, 1.3991915283203125, 1.3993380126953125, 1.3993133544921874, 1.3990963134765626, 1.399372802734375, 1.3989632568359376]",tokens/s,0.7035145956380915,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-180B,tiiuae/falcon-180B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14522,7 +14522,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664d4ab5-171b328a778f712479b07012;1bdc15d3-d38d-4d92-8329-1bff2dac1900) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664feb2d-4f3d76b84f29f7785fbcbe65;0fb8acaa-e82d-4330-8732-1de2018eca7b) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -14607,7 +14607,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9662-445666f41c0fd312284cb909;9aab9292-c910-4680-b99a-71409dcb346d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe830-7c78dfb35b0dfee21c0bcc8a;b2b512a2-0365-472d-a85f-7c0827ce6e27) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14639,7 +14639,7 @@ If this is a private repository, make sure to pass a token having permission to ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-3b-4e1t,stabilityai/stablelm-3b-4e1t,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2108.735488,2846.359552,0.0,2199.912448,2075.149824,s,10,1.3464691162109377,0.13464691162109377,0.0003254535509591645,0.1345770721435547,0.1350561508178711,0.1350975486755371,0.13513066696166992,"[0.13456275939941406, 0.13459138488769531, 0.13411830139160155, 0.1348402862548828, 0.13513894653320313, 0.1344079284667969, 0.13428579711914063, 0.1350469512939453, 0.13498921203613282, 0.134487548828125]",tokens/s,1901.2690073456915,kWh,1.5852142130887069e-06,8.686249076854436e-07,7.373461454319348e-06,9.827300575093499e-06,tokens/kWh,26049879.928249203,MB,2108.735488,2846.359552,0.0,2199.912448,2180.296192,s,10,72.99818603515625,7.299818603515625,0.000639301859605335,7.299923828124999,7.300465576171875,7.300673217773437,7.3008393310546875,"[7.300880859375, 7.300095703125, 7.3002060546875, 7.299826171875, 7.30041943359375, 7.2990068359375, 7.299837890625, 7.300009765625, 7.29865869140625, 7.29924462890625]",tokens/s,8.630351440467154,kWh,8.617964071476902e-05,4.7232711682958325e-05,0.00039865630781368215,0.0005320686602114095,tokens/kWh,118405.77111790025,,s,629,74.09423864746091,0.11779688179246572,0.016030014937327634,0.11585330963134766,0.11598766174316406,0.11602022247314453,0.25049972717285157,"[0.11587891387939453, 0.11590962982177734, 0.11583795166015624, 0.11630899047851563, 0.11594445037841797, 0.11586867523193359, 0.1157201919555664, 0.11584307098388671, 0.11569664001464844, 0.11587174224853515, 0.11594035339355468, 0.11572121429443359, 0.11581439971923828, 0.11586867523193359, 0.11578368377685547, 0.11574374389648437, 0.11594956970214844, 0.11572633361816406, 0.11587686157226562, 0.11594547271728516, 0.11587174224853515, 0.1157550048828125, 0.1159710693359375, 0.11576319885253906, 0.11601510620117188, 0.11585740661621094, 0.115810302734375, 0.11578470611572265, 0.11592806243896485, 0.11579596710205078, 0.11591372680664062, 0.11584716796875, 0.11584307098388671, 0.11578265380859375, 0.11594342041015625, 0.115810302734375, 0.11592294311523438, 0.11594137573242187, 0.11582054138183594, 0.1157918701171875, 0.11591577911376953, 0.11587071990966796, 0.11590348815917968, 0.11592396545410157, 0.11587686157226562, 0.11579289245605469, 0.11591577911376953, 0.11579289245605469, 0.11597618865966797, 0.11596595001220703, 0.11586764526367188, 0.11596185302734376, 0.115989501953125, 0.11585536193847656, 0.11587174224853515, 0.11595673370361329, 0.11591577911376953, 0.11621683502197265, 0.1159557113647461, 0.11586764526367188, 0.11588607788085938, 0.11599359893798829, 0.2506895294189453, 0.1157550048828125, 0.11599974060058593, 0.11574476623535156, 0.1163345947265625, 0.11596288299560546, 0.11580108642578125, 0.1157754898071289, 0.11592806243896485, 0.11585945892333985, 0.11574374389648437, 0.11587174224853515, 0.11579084777832031, 0.11582975769042969, 0.11603250885009765, 0.11579596710205078, 0.11580108642578125, 0.11589939117431641, 0.11582975769042969, 0.11591680145263672, 0.11590656280517578, 0.11581542205810547, 0.11584614562988281, 0.11589324951171875, 0.11580416107177735, 0.11581132507324218, 0.1158656005859375, 0.11576934051513672, 0.11585945892333985, 0.11589324951171875, 0.11582566070556641, 0.11580313873291016, 0.1158656005859375, 0.11583283233642579, 0.11586150360107422, 0.1159004135131836, 0.11584921264648437, 0.11579392242431641, 0.11597004699707031, 0.1157918701171875, 0.11587071990966796, 0.11592601776123047, 0.11588915252685547, 0.11577241516113282, 0.11601203155517578, 0.11582975769042969, 0.11592601776123047, 0.11604889678955078, 0.11587993621826172, 0.1158440933227539, 0.11598336029052735, 0.11578060913085937, 0.11591782379150391, 0.1158809585571289, 0.11587993621826172, 0.11579392242431641, 0.11591680145263672, 0.11579084777832031, 0.11587481689453125, 0.11583692932128906, 0.11589631652832032, 0.11585740661621094, 0.1159925765991211, 0.2502563781738281, 0.11579801940917969, 0.11596697235107421, 0.11578265380859375, 0.11583590698242187, 0.11592704010009766, 0.11574169921875, 0.11581951904296875, 0.11580313873291016, 0.11575091552734375, 0.11572121429443359, 0.11591168212890625, 0.11579904174804688, 0.11585330963134766, 0.11578777313232422, 0.11579289245605469, 0.1157550048828125, 0.11602227020263672, 0.11578163146972656, 0.11585842895507813, 0.11581337738037109, 0.1158656005859375, 0.11577651214599609, 0.115957763671875, 0.115810302734375, 0.11594751739501953, 0.11644927978515625, 0.11578060913085937, 0.11589427185058594, 0.1160110092163086, 0.11574988555908203, 0.11592396545410157, 0.11583078765869141, 0.11586252593994141, 0.11576627349853516, 0.11596697235107421, 0.11582975769042969, 0.11590758514404297, 0.11587891387939453, 0.11583999633789062, 0.11587071990966796, 0.11605811309814452, 0.11584921264648437, 0.11585126495361328, 0.11589734649658204, 0.1158287353515625, 0.11585126495361328, 0.11597209930419922, 0.11613593292236328, 0.11593215942382812, 0.11590758514404297, 0.1158287353515625, 0.1158502426147461, 0.11597721862792969, 0.11583590698242187, 0.11585638427734375, 0.11585638427734375, 0.11588813018798828, 0.11583795166015624, 0.11590348815917968, 0.1158287353515625, 0.11585433959960938, 0.11587891387939453, 0.2509680633544922, 0.11578880310058594, 0.115957763671875, 0.1157734375, 0.11582566070556641, 0.11597721862792969, 0.11578982543945313, 0.1157201919555664, 0.1157918701171875, 0.11576627349853516, 0.1157396469116211, 0.11594547271728516, 0.11575296020507812, 0.11577855682373046, 0.11581644439697265, 0.11581337738037109, 0.11573350524902344, 0.1160099868774414, 0.11578982543945313, 0.11580825805664062, 0.11585433959960938, 0.11591372680664062, 0.115884033203125, 0.11605503845214844, 0.11580108642578125, 0.1159188461303711, 0.1158318099975586, 0.11580620574951171, 0.1157570571899414, 0.11603456115722656, 0.11580210876464844, 0.11576525115966797, 0.11579392242431641, 0.11586252593994141, 0.11581132507324218, 0.11600691223144531, 0.11584307098388671, 0.11583283233642579, 0.11587789154052734, 0.1159362564086914, 0.11580518341064452, 0.11600588989257812, 0.11585330963134766, 0.1157734375, 0.11589836883544923, 0.11585945892333985, 0.11578470611572265, 0.11602227020263672, 0.11591680145263672, 0.1159925765991211, 0.11588198089599609, 0.11585433959960938, 0.11585536193847656, 0.1160417251586914, 0.11592396545410157, 0.11581439971923828, 0.11594751739501953, 0.11595673370361329, 0.11584204864501953, 0.11601407623291016, 0.1159024658203125, 0.11579904174804688, 0.11590758514404297, 0.25138380432128904, 0.1158287353515625, 0.1159024658203125, 0.1157754898071289, 0.11582771301269532, 0.11591680145263672, 0.11577139282226563, 0.11580416107177735, 0.11595366668701172, 0.11575193786621094, 0.11614924621582032, 0.11594649505615234, 0.11574988555908203, 0.11585945892333985, 0.11589017486572266, 0.11571916961669922, 0.11579801940917969, 0.11601203155517578, 0.1157734375, 0.11581132507324218, 0.11593011474609374, 0.11580928039550781, 0.11586969757080078, 0.11597618865966797, 0.11573554992675782, 0.11589836883544923, 0.11588505554199219, 0.11572223663330078, 0.1158502426147461, 0.11597516632080078, 0.11578572845458984, 0.11587276458740234, 0.11590860748291015, 0.11581337738037109, 0.11589222717285157, 0.11598847961425782, 0.11580620574951171, 0.11578572845458984, 0.11595366668701172, 0.11580210876464844, 0.11586048126220704, 0.11598540496826172, 0.11580416107177735, 0.11589222717285157, 0.11598131561279297, 0.11580518341064452, 0.11586764526367188, 0.11596288299560546, 0.11581747436523437, 0.11583078765869141, 0.11598643493652344, 0.11577855682373046, 0.11591577911376953, 0.11600179290771484, 0.11580928039550781, 0.11583795166015624, 0.1159362564086914, 0.11586457824707032, 0.11584102630615234, 0.11597209930419922, 0.11591986846923828, 0.11613593292236328, 0.11600077056884765, 0.25054617309570315, 0.11577855682373046, 0.1159331817626953, 0.11573760223388672, 0.11583385467529297, 0.11588915252685547, 0.11579084777832031, 0.11581747436523437, 0.1157918701171875, 0.11576525115966797, 0.11576012420654297, 0.1158809585571289, 0.11572428894042969, 0.11585433959960938, 0.11582361602783203, 0.11580825805664062, 0.11569356536865234, 0.11601203155517578, 0.1157232666015625, 0.11584614562988281, 0.11582361602783203, 0.11634585571289062, 0.11577241516113282, 0.11597824096679688, 0.11574681854248046, 0.11588198089599609, 0.11579596710205078, 0.11583590698242187, 0.11573452758789063, 0.11592908477783204, 0.11574272155761718, 0.1158656005859375, 0.11583897399902343, 0.11584716796875, 0.1157918701171875, 0.11604377746582031, 0.11584102630615234, 0.11585433959960938, 0.11581849670410156, 0.11589529418945313, 0.11578163146972656, 0.11593113708496093, 0.1158287353515625, 0.11595468902587891, 0.1159331817626953, 0.11587891387939453, 0.11582463836669922, 0.11596902465820312, 0.1157949447631836, 0.11587891387939453, 0.11583999633789062, 0.11583487701416016, 0.11583487701416016, 0.11596800231933593, 0.11584921264648437, 0.11585740661621094, 0.11589119720458985, 0.11587276458740234, 0.11581542205810547, 0.11589734649658204, 0.11579289245605469, 0.11591168212890625, 0.11589222717285157, 0.25038029479980467, 0.11576217651367188, 0.11597516632080078, 0.11575091552734375, 0.11582669067382813, 0.11594035339355468, 0.11582054138183594, 0.11579392242431641, 0.11578060913085937, 0.11574476623535156, 0.11573043060302735, 0.11592499542236329, 0.11576319885253906, 0.11574476623535156, 0.11584307098388671, 0.11583385467529297, 0.1157918701171875, 0.1160099868774414, 0.11584716796875, 0.1158123550415039, 0.1158809585571289, 0.1158502426147461, 0.11576012420654297, 0.11603865814208984, 0.115810302734375, 0.11580313873291016, 0.11578265380859375, 0.11580210876464844, 0.11578060913085937, 0.11608678436279297, 0.11582259368896484, 0.11578060913085937, 0.11583999633789062, 0.11587583923339843, 0.11582054138183594, 0.11598540496826172, 0.11584614562988281, 0.11591986846923828, 0.11590962982177734, 0.11589734649658204, 0.11587071990966796, 0.11598745727539063, 0.1159004135131836, 0.11628543853759765, 0.11594445037841797, 0.11589119720458985, 0.1157949447631836, 0.11596288299560546, 0.11586252593994141, 0.1158502426147461, 0.11587686157226562, 0.11586457824707032, 0.11581439971923828, 0.11600691223144531, 0.11587481689453125, 0.11581644439697265, 0.11586252593994141, 0.11589631652832032, 0.11584921264648437, 0.11597516632080078, 0.11587993621826172, 0.11580928039550781, 0.11593830108642578, 0.2512076873779297, 0.11580006408691407, 0.11591782379150391, 0.11583897399902343, 0.11577855682373046, 0.11591065979003906, 0.11580108642578125, 0.11576319885253906, 0.11591986846923828, 0.11575193786621094, 0.1157232666015625, 0.1160099868774414, 0.11578777313232422, 0.11589529418945313, 0.11587789154052734, 0.11572428894042969, 0.11579084777832031, 0.11599565124511718, 0.11584307098388671, 0.115810302734375, 0.11587071990966796, 0.11580108642578125, 0.11590860748291015, 0.11597824096679688, 0.11575910186767578, 0.11585330963134766, 0.11588710021972656, 0.11573862457275391, 0.11581951904296875, 0.1159731216430664, 0.115736572265625, 0.11588505554199219, 0.11591372680664062, 0.11578470611572265, 0.11581644439697265, 0.11596800231933593, 0.115810302734375, 0.1157918701171875, 0.11598643493652344, 0.11578265380859375, 0.11588607788085938, 0.11595878601074219, 0.11590144348144531, 0.11588915252685547, 0.11598336029052735, 0.11578368377685547, 0.11589427185058594, 0.11601715087890625, 0.11588301086425781, 0.11580518341064452, 0.11592704010009766, 0.11583078765869141, 0.11587891387939453, 0.11599769592285156, 0.11583078765869141, 0.1158635482788086, 0.11601612854003907, 0.11580723571777343, 0.11590553283691406, 0.11595878601074219, 0.11587481689453125, 0.11597824096679688, 0.11600383758544922, 0.25114930725097656, 0.11628749084472656, 0.11598847961425782, 0.1157570571899414, 0.11583590698242187, 0.11586867523193359, 0.11570381164550782, 0.11576422119140625, 0.115810302734375, 0.11578982543945313, 0.11569868469238281, 0.11589836883544923, 0.11571405029296875, 0.11586048126220704, 0.11577037048339844, 0.11576934051513672, 0.11577241516113282, 0.11599052429199219, 0.11569868469238281, 0.11587071990966796, 0.11583692932128906, 0.11585638427734375, 0.11574476623535156, 0.11591986846923828, 0.11577651214599609, 0.11584204864501953, 0.11575296020507812, 0.11586457824707032, 0.11572633361816406, 0.1159393310546875, 0.11578880310058594, 0.11586764526367188, 0.11578777313232422, 0.11592192077636719, 0.11583795166015624, 0.11593113708496093, 0.11576422119140625, 0.11589734649658204, 0.11585330963134766, 0.11582259368896484, 0.11577753448486328, 0.11593830108642578, 0.11579084777832031, 0.11583897399902343, 0.11586457824707032, 0.11587379455566406, 0.11583795166015624, 0.11592704010009766, 0.11577446746826171, 0.11584921264648437, 0.11588607788085938, 0.11581951904296875, 0.11581439971923828, 0.11600588989257812, 0.1158809585571289, 0.11587891387939453, 0.11592806243896485, 0.11589836883544923, 0.1158123550415039, 0.11596083068847657, 0.11583590698242187, 0.11587071990966796, 0.11587276458740234, 0.2509967346191406, 0.11641139221191406, 0.11601715087890625, 0.11583795166015624, 0.11582361602783203, 0.11592192077636719, 0.11578470611572265, 0.11565670776367187, 0.11581542205810547, 0.11581542205810547, 0.11570175933837891, 0.1159024658203125, 0.11577241516113282, 0.11574988555908203, 0.11580723571777343, 0.11583999633789062, 0.11574066925048829, 0.11599769592285156, 0.11575910186767578, 0.11590553283691406, 0.1158287353515625, 0.11586048126220704, 0.1157734375, 0.11626496124267578, 0.11583590698242187, 0.11575603485107422, 0.11579392242431641, 0.11578572845458984, 0.11572428894042969, 0.11596800231933593, 0.11580006408691407, 0.11583795166015624, 0.11580620574951171, 0.11584819030761719, 0.11577855682373046, 0.11593215942382812, 0.11581439971923828, 0.11576319885253906, 0.11582771301269532, 0.11587071990966796, 0.1157570571899414, 0.11595673370361329, 0.11582771301269532, 0.11582259368896484, 0.1158666229248047, 0.1158440933227539, 0.11580928039550781, 0.1159557113647461, 0.11582669067382813, 0.11585330963134766, 0.11585638427734375, 0.11581951904296875, 0.11580723571777343, 0.11601612854003907, 0.11587481689453125, 0.11576627349853516, 0.1158502426147461, 0.11590451049804687, 0.11581644439697265, 0.11592806243896485, 0.11587276458740234, 0.11577651214599609, 0.11585228729248047]",tokens/s,8.489189058177258,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2-large,openai-community/gpt2-large,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14697,7 +14697,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664e90aa-7123c50a39edde740f9ca8ba;1bf1d582-66d9-4922-a597-060de59096ad) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-664fe256-60fd6feb7cfde5ff49b8cec1;c0e8dee4-ba52-43d2-a67e-84c2d159eeaf) 403 Forbidden: Authorization error.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -14812,7 +14812,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: GPTNeoXForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,stabilityai/stablelm-base-alpha-3b,stabilityai/stablelm-base-alpha-3b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -14880,7 +14880,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e960e-33d8c75c00967e927750cb1f;f5b76649-0224-49fd-aefb-683ae0484ad2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe7dc-5250df9e009a24ef34843c9a;823a9bf3-3f27-4edb-ac91-1780bf8dd866) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15009,7 +15009,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e9123-453eab76428e3f7f28cce01e;3da72a95-bfcf-4630-82b2-037ae7c81d9b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe2d5-4e664c342a4bcfce023c5a3d;5d440d77-2b86-4a90-a2e2-99692ed7d875) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15099,7 +15099,7 @@ ChildProcessError: Traceback (most recent call last): ValueError: XGLMForCausalLM does not support an attention implementation through torch.nn.functional.scaled_dot_product_attention yet. Please request the support for this architecture: https://github.com/huggingface/transformers/issues/28005. If you believe this error is a bug, please open an issue in Transformers GitHub repository and load your model with the argument `attn_implementation=""eager""` meanwhile. Example: `model = AutoModel.from_pretrained(""openai/whisper-tiny"", attn_implementation=""eager"")` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15138,7 +15138,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d462c-5d8d0f2434749b5f32d831fa;38d66fc7-9b20-46a9-9a70-c9cb9e6ff37c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe94f-2aed67c8200d10582937f68b;dac7f488-b642-4901-9db7-68b9bfc5f4fe) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15169,7 +15169,7 @@ OSError: B is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,tiiuae/falcon-40b,tiiuae/falcon-40b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15238,7 +15238,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e944d-20c01463040e59262a293a9f;7260186d-311b-49b3-95ce-06171876ff8b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe617-6276ab744d54e1654c3e30c2;00de2ee0-be8d-4109-91ee-50b5b266dfaf) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15269,7 +15269,7 @@ OSError: s is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,togethercomputer/RedPajama-INCITE-Base-3B-v1,togethercomputer/RedPajama-INCITE-Base-3B-v1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15416,7 +15416,7 @@ ImportError: This modeling file requires the following packages that were not fo ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-13b-hf,meta-llama/Llama-2-13b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,2625.98656,9028.763648,0.0,8382.316544,7939.283456,s,10,6.165609191894531,0.6165609191894531,0.001439653397024848,0.6159833984375,0.6172576660156249,0.6189813110351562,0.6203602270507812,"[0.6157730102539063, 0.6207049560546875, 0.6154873046875, 0.6165646362304688, 0.6165377197265625, 0.6158151245117187, 0.6168746337890625, 0.6159981689453125, 0.6159686279296875, 0.615885009765625]",tokens/s,415.20633571220213,kWh,7.2687008314662515e-06,3.98295215025712e-06,3.516605754459804e-05,4.641771052632141e-05,tokens/kWh,5515136.293825475,MB,2625.98656,9028.763648,0.0,8382.316544,8233.862144,s,10,338.1217890625,33.81217890625,0.0010735847560921045,33.812240234375,33.813633203125,33.813722851562495,33.813794570312496,"[33.81140234375, 33.81193359375, 33.81028125, 33.81291015625, 33.81206640625, 33.81083203125, 33.8124140625, 33.81361328125, 33.8138125, 33.8125234375]",tokens/s,1.8632339600082617,kWh,0.00039918773012028797,0.00021878949164742085,0.0019440407758196127,0.002562017997587322,tokens/kWh,24589.99119417886,,s,629,343.13429852294934,0.5455235270635123,0.07330267162846767,0.536690673828125,0.5369712402343749,0.5370589111328126,1.153123212890625,"[0.536447998046875, 0.53654833984375, 0.5363824462890625, 0.536595458984375, 0.536395751953125, 0.536637451171875, 0.5366241455078125, 0.536605712890625, 0.5362606201171874, 0.5368657836914063, 0.5363804321289063, 0.5366886596679687, 0.5363906860351563, 0.5366435546875, 0.5362196655273438, 0.536585205078125, 0.5362493286132812, 0.5370194091796875, 0.536542236328125, 0.5368770751953125, 0.5363814697265625, 0.5366333618164062, 0.5366722412109375, 0.53677978515625, 0.53680126953125, 0.5368053588867188, 0.536647705078125, 0.5366046752929687, 0.5369467163085937, 0.53671630859375, 0.53665380859375, 0.5370398559570313, 0.536537109375, 0.5368299560546875, 0.5365708618164062, 0.5366610107421875, 0.5365626831054687, 0.536722412109375, 0.5366077270507813, 0.5365964965820312, 0.5365596313476563, 0.5367767333984375, 0.5364439086914062, 0.536932373046875, 0.5368739624023438, 0.53686376953125, 0.5364971313476562, 0.5370101928710938, 0.5367367553710938, 0.536869873046875, 0.5364766845703125, 0.536900634765625, 0.5366435546875, 0.536932373046875, 0.5366773681640625, 0.5370142822265624, 0.536685546875, 0.5368391723632813, 0.5365176391601563, 0.5370234985351563, 0.5368289184570313, 0.5368719482421875, 1.1580068359375, 0.5365903930664062, 0.5371329956054688, 0.5363855590820312, 0.5365852661132813, 0.5363424682617187, 0.536585205078125, 0.5364224243164063, 0.5366988525390625, 0.5369517822265625, 0.537153564453125, 0.5364326171875, 0.536721435546875, 0.536453125, 0.536690673828125, 0.5363517456054687, 0.536563720703125, 0.5363988647460938, 0.5367879638671875, 0.5364132080078124, 0.536784912109375, 0.5365596313476563, 0.5364940795898437, 0.5364992065429688, 0.5367828369140625, 0.5366251220703125, 0.53686376953125, 0.5364090576171875, 0.5365841674804688, 0.5366661376953125, 0.5367070922851562, 0.536742919921875, 0.5366917114257812, 0.5364141845703125, 0.5367572631835937, 0.5365964965820312, 0.5365135498046875, 0.536501220703125, 0.536722412109375, 0.5365883178710937, 0.536563720703125, 0.5368657836914063, 0.5368831787109375, 0.536742919921875, 0.536732666015625, 0.53655859375, 0.537112548828125, 0.5369077758789063, 0.536669189453125, 0.536580078125, 0.5368289184570313, 0.5365380859375, 0.5366087646484375, 0.5367357177734375, 0.5367705688476563, 0.5368289184570313, 0.536911865234375, 0.53677978515625, 0.5371351318359375, 0.5367470092773438, 0.5369292602539063, 0.5371084594726563, 0.5370203857421875, 1.1526328125, 0.5366445922851563, 0.5366138916015625, 0.536363037109375, 0.5367664794921875, 0.5365811157226562, 0.5364561767578125, 0.5363773193359375, 0.5364439086914062, 0.5363814697265625, 0.5367193603515625, 0.5364623413085937, 0.53661181640625, 0.5364971313476562, 0.5363722534179688, 0.536489990234375, 0.5365944213867188, 0.5363978271484375, 0.5366681518554688, 0.5365657348632813, 0.5366773681640625, 0.5365667724609375, 0.536595458984375, 0.5366016235351563, 0.5367060546875, 0.5365811157226562, 0.5367408447265625, 0.5364060668945313, 0.5367572021484375, 0.5365862426757813, 0.5367152709960937, 0.5366753540039062, 0.5366527709960938, 0.5364623413085937, 0.5371474609375, 0.5365206298828125, 0.5368186645507812, 0.5366773681640625, 0.5368606567382812, 0.5365309448242187, 0.5366364135742188, 0.5366814575195312, 0.5368237915039062, 0.5366456298828125, 0.5367255249023437, 0.536848388671875, 0.5368125610351563, 0.5366128540039062, 0.5368760375976562, 0.5368555297851563, 0.5368391723632813, 0.536580078125, 0.5368125610351563, 0.5367971801757813, 0.536869873046875, 0.53675927734375, 0.536859619140625, 0.5367306518554688, 0.5370460205078125, 0.536543212890625, 0.536858642578125, 0.5368361206054687, 0.5368463134765625, 1.1538544921875, 0.5365811157226562, 0.5366527709960938, 0.53648486328125, 0.53648486328125, 0.5364992065429688, 0.5365022583007812, 0.5365626831054687, 0.5365975341796875, 0.5362677612304687, 0.5365288696289062, 0.5363916625976562, 0.5365759887695313, 0.5365391235351562, 0.5367060546875, 0.5364602661132812, 0.5364459228515625, 0.5363937377929687, 0.5367633666992188, 0.5369948120117187, 0.536901611328125, 0.5364592895507813, 0.536431640625, 0.5367992553710937, 0.536722412109375, 0.5369375, 0.5371658325195312, 0.536816650390625, 0.5367275390625, 0.5365729370117187, 0.5365442504882812, 0.53661181640625, 0.5367203979492188, 0.5365104370117187, 0.5370091552734375, 0.536647705078125, 0.5368258666992187, 0.5366620483398438, 0.5370664672851563, 0.5366456298828125, 0.5367869262695313, 0.5366251831054687, 0.5368422241210937, 0.5367409057617187, 0.53673779296875, 0.5369200439453125, 0.5369835815429688, 0.5367101440429688, 0.5367756958007812, 0.536605712890625, 0.5369313354492188, 0.5366343383789063, 0.5368842163085937, 0.5366947631835938, 0.5370408935546875, 0.5366026000976563, 0.536848388671875, 0.536748046875, 0.5367879638671875, 0.5366241455078125, 0.5369999389648438, 0.5369251708984375, 0.5369354248046875, 1.153869873046875, 0.5364306030273438, 0.5364418334960938, 0.5363456420898437, 0.536542236328125, 0.5364735717773438, 0.5364735717773438, 0.53648388671875, 0.5364367065429687, 0.5362861938476563, 0.536732666015625, 0.5364490356445313, 0.5367275390625, 0.5363240966796875, 0.53640087890625, 0.53631591796875, 0.5365862426757813, 0.5364346923828125, 0.5366824951171875, 0.5368053588867188, 0.5367879638671875, 0.536599609375, 0.5366302490234375, 0.5369886474609376, 0.536711181640625, 0.536721435546875, 0.536796142578125, 0.5365575561523438, 0.5366405029296875, 0.5367551879882813, 0.5367510986328125, 0.5366394653320312, 0.536648681640625, 0.5365442504882812, 0.5370214233398437, 0.5365883178710937, 0.5368555297851563, 0.5365698852539063, 0.5368135375976563, 0.5365309448242187, 0.5367337036132812, 0.5365657348632813, 0.5366886596679687, 0.537196533203125, 0.53675830078125, 0.5367787475585938, 0.5369528198242187, 0.5366651000976562, 0.5369630737304687, 0.5365892944335937, 0.5368842163085937, 0.5366292724609375, 0.5370153198242188, 0.5369415893554688, 0.5370337524414063, 0.5372006225585938, 0.536859619140625, 0.5367398681640625, 0.5367623901367188, 0.5365073852539063, 0.5370162963867188, 0.5367234497070312, 0.5370582885742188, 1.1531458740234375, 0.5364705200195312, 0.5366681518554688, 0.5363804321289063, 0.5363886108398438, 0.5364756469726563, 0.5364664306640625, 0.5365401611328126, 0.536416259765625, 0.5363035888671875, 0.5367296142578125, 0.5364285278320312, 0.5365032958984375, 0.536479736328125, 0.536352783203125, 0.5366026000976563, 0.5365340576171875, 0.5364674072265625, 0.5365104370117187, 0.5365411987304688, 0.5367285766601563, 0.5363517456054687, 0.53652685546875, 0.5369180297851562, 0.5368514404296875, 0.5370685424804688, 0.5368442993164062, 0.5365104370117187, 0.5370828857421875, 0.536679443359375, 0.5365964965820312, 0.53659033203125, 0.5366548461914062, 0.5363896484375, 0.53697021484375, 0.536563720703125, 0.5366937866210938, 0.5366200561523438, 0.5369671630859375, 0.5365391235351562, 0.5367275390625, 0.536658935546875, 0.536690673828125, 0.5367152709960937, 0.5369057006835938, 0.5368923950195312, 0.5368975219726563, 0.5367337036132812, 0.5367684936523438, 0.5366927490234376, 0.536722412109375, 0.536542236328125, 0.5369354248046875, 0.5368811645507813, 0.53682275390625, 0.5366046752929687, 0.5369548950195312, 0.5367941284179687, 0.536922119140625, 0.5365626831054687, 0.536975341796875, 0.5367193603515625, 0.5370153198242188, 1.1540572509765625, 0.5365452880859375, 0.5367817993164062, 0.5364510498046875, 0.5364029541015625, 0.536352783203125, 0.5365350341796875, 0.5363701782226562, 0.5369231567382813, 0.5363783569335937, 0.5364838256835938, 0.5365411987304688, 0.536494140625, 0.5365811157226562, 0.5364940795898437, 0.536395751953125, 0.5365975341796875, 0.5363773193359375, 0.5367418823242187, 0.536500244140625, 0.5367091064453124, 0.5365442504882812, 0.5366876220703125, 0.5366896362304687, 0.5367890014648438, 0.536721435546875, 0.5368831787109375, 0.5366046752929687, 0.5366005859375, 0.53669677734375, 0.5367347412109374, 0.5368207397460938, 0.5366753540039062, 0.536352783203125, 0.5369262084960937, 0.5364080810546875, 0.5368033447265625, 0.5368494262695312, 0.5367664794921875, 0.5367347412109374, 0.536669189453125, 0.5365504150390625, 0.5367367553710938, 0.5366998901367187, 0.5368668212890625, 0.536669189453125, 0.537059326171875, 0.5366917114257812, 0.5368176879882812, 0.5366077270507813, 0.5368545532226563, 0.5367203979492188, 0.536932373046875, 0.536816650390625, 0.5371740112304687, 0.5367347412109374, 0.5368494262695312, 0.5372958984375, 0.5368975219726563, 0.5367756958007812, 0.536911865234375, 0.536816650390625, 0.5369989013671875, 1.1533507080078125, 0.5366333618164062, 0.536700927734375, 0.5364121704101562, 0.5365677490234375, 0.536543212890625, 0.5365463256835937, 0.53659033203125, 0.5365555419921875, 0.5362913208007812, 0.5365339965820313, 0.5364132080078124, 0.5366364135742188, 0.5365258178710938, 0.5364510498046875, 0.5363507080078125, 0.5366098022460938, 0.5363865356445312, 0.536679443359375, 0.5367337036132812, 0.5368115234375, 0.53640087890625, 0.5366077270507813, 0.5367276000976563, 0.5367551879882813, 0.5366753540039062, 0.5368023071289062, 0.5366896362304687, 0.5367562255859375, 0.5366886596679687, 0.5366261596679688, 0.5366763305664063, 0.5368207397460938, 0.5363517456054687, 0.5370572509765625, 0.5365883178710937, 0.5371094970703125, 0.5369569091796875, 0.5369425659179687, 0.536500244140625, 0.5366784057617188, 0.5366405029296875, 0.5368790893554688, 0.53669580078125, 0.5369569091796875, 0.5366865844726563, 0.5374371948242187, 0.5367757568359375, 0.5369190063476562, 0.536690673828125, 0.5367654418945312, 0.53671630859375, 0.5369190673828125, 0.53690673828125, 0.5369476928710938, 0.5368565673828125, 0.537027587890625, 0.5368350830078125, 0.53691796875, 0.5366415405273437, 0.5371309814453125, 0.5368033447265625, 0.5369467163085937, 1.1533577880859376, 0.5367510986328125, 0.5366128540039062, 0.536431640625, 0.5363916625976562, 0.5363978271484375, 0.5365330200195313, 0.5364408569335938, 0.5366784057617188, 0.5363814697265625, 0.5370460205078125, 0.53640087890625, 0.5365524291992188, 0.5364510498046875, 0.5364623413085937, 0.5364613037109375, 0.5364879150390625, 0.5363118286132813, 0.5367357177734375, 0.5368207397460938, 0.5369169921875, 0.5366005859375, 0.5368955078125, 0.5365463256835937, 0.5368115234375, 0.536748046875, 0.5370255126953125, 0.53661181640625, 0.5370132446289062, 0.5365913696289063, 0.5367347412109374, 0.5369497680664063, 0.5367306518554688, 0.5366149291992187, 0.5370972290039062, 0.536642578125, 0.5368780517578124, 0.5365924072265625, 0.5369425659179687, 0.5370839233398438, 0.53680126953125, 0.5367603149414063, 0.5367490844726562, 0.5366896362304687, 0.5368125610351563, 0.53688525390625, 0.5368094482421875, 0.5367756958007812, 0.5368790893554688, 0.536732666015625, 0.5368176879882812, 0.5366179809570313, 0.5369610595703125, 0.5367265014648438, 0.53696923828125, 0.5366077270507813, 0.537112548828125, 0.5368370971679688, 0.536774658203125, 0.5365411987304688, 0.5368115234375, 0.5366661376953125, 0.5367930908203125, 1.15306494140625, 0.53646435546875, 0.5366159057617188, 0.5365698852539063, 0.5364930419921875, 0.536580078125, 0.5365565185546874, 0.536474609375, 0.5366661376953125, 0.5363200073242187, 0.5365155639648438, 0.5367122192382813, 0.536537109375, 0.5364735717773438, 0.536426513671875, 0.5363875732421876, 0.5367296142578125, 0.5363353881835937, 0.53671728515625, 0.5365411987304688, 0.5367490844726562, 0.5365330200195313, 0.5366384887695312, 0.5367664794921875, 0.5367367553710938, 0.5367357177734375, 0.5367675170898437, 0.5364070434570313, 0.5367142333984375, 0.5366456298828125, 0.5367705688476563, 0.5365892944335937, 0.5366998901367187, 0.5364541625976562, 0.5369794311523437, 0.53646337890625, 0.5367705688476563, 0.5365473022460937, 0.5370245361328125, 0.536875, 0.536679443359375, 0.5365883178710937, 0.5366005859375, 0.5366456298828125, 0.5369630737304687, 0.53676953125, 0.5369354248046875, 0.5367982177734375, 0.5370009765625, 0.5367684936523438, 0.5370101928710938, 0.5370214233398437, 0.5369425659179687, 0.5367664794921875, 0.5368330078125, 0.5366927490234376, 0.537059326171875, 0.53688525390625, 0.5369047241210938, 0.536648681640625, 0.5369476928710938, 0.5367716064453125, 0.537112548828125]",tokens/s,1.8331015077990864,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,meta-llama/Llama-2-7b-hf,meta-llama/Llama-2-7b-hf,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.289728,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,4108.419072,5287.44448,0.0,4640.997376,4330.668544,s,10,3.1997629089355466,0.31997629089355467,0.0010652547133575853,0.31959405517578127,0.3214676086425781,0.32149762878417965,0.32152164489746093,"[0.31874813842773436, 0.3210158386230469, 0.3191706848144531, 0.3194938354492188, 0.31946047973632813, 0.31844869995117187, 0.31969427490234376, 0.32074237060546873, 0.3214609375, 0.32152764892578123]",tokens/s,800.0592771580148,kWh,3.7604298649562735e-06,2.060550245869101e-06,1.8770327516251517e-05,2.4591307627076894e-05,tokens/kWh,10410182.487332419,MB,4108.419072,5287.44448,0.0,4640.997376,4515.667456,s,10,175.235673828125,17.5235673828125,0.0014736007130784414,17.5240908203125,17.525130859375,17.525280273437502,17.525399804687503,"[17.52509765625, 17.52158984375, 17.524423828125, 17.523767578125, 17.5244140625, 17.521322265625, 17.524623046875, 17.5235234375, 17.5254296875, 17.521482421875]",tokens/s,3.5951583729344847,kWh,0.00020692485703776285,0.00011341188782000648,0.001005294630623929,0.0013256313754816985,tokens/kWh,47524.523910055694,,s,629,177.83855282592782,0.28273219845139547,0.038060612034765025,0.2781388854980469,0.2783006591796875,0.2783752197265625,0.5978882275390625,"[0.27858328247070313, 0.27810714721679686, 0.27806002807617186, 0.2781603698730469, 0.2780487670898438, 0.2779842529296875, 0.27810406494140627, 0.27796890258789064, 0.27810302734375, 0.2779637756347656, 0.27821774291992185, 0.2781737060546875, 0.2781051025390625, 0.278034423828125, 0.2780057678222656, 0.27806719970703125, 0.27839590454101565, 0.27814501953125, 0.27807232666015624, 0.27819009399414063, 0.2782167053222656, 0.2782105712890625, 0.27820748901367187, 0.278297607421875, 0.27826995849609376, 0.2780712890625, 0.2780426330566406, 0.2781808776855469, 0.27829556274414063, 0.27816754150390627, 0.2781163635253906, 0.2781972351074219, 0.2782863464355469, 0.2780477294921875, 0.2781614074707031, 0.2781419372558594, 0.2782822265625, 0.27810302734375, 0.2781747131347656, 0.278181884765625, 0.27818597412109375, 0.27807949829101564, 0.2783580017089844, 0.27806002807617186, 0.27818496704101564, 0.278265869140625, 0.27827813720703126, 0.27825357055664063, 0.27825457763671874, 0.2781265869140625, 0.2781214599609375, 0.2780579833984375, 0.278192138671875, 0.27812353515625, 0.2781327209472656, 0.2782576599121094, 0.278287353515625, 0.27825357055664063, 0.27815731811523436, 0.27832012939453127, 0.2783068237304688, 0.2782637939453125, 0.599541748046875, 0.27804568481445313, 0.27802215576171874, 0.27826483154296877, 0.278076416015625, 0.2781112365722656, 0.277928955078125, 0.2780845947265625, 0.2779053955078125, 0.27808050537109374, 0.27793508911132814, 0.27789413452148437, 0.27800677490234377, 0.2780477294921875, 0.27790643310546875, 0.2779893798828125, 0.2780702819824219, 0.27814093017578123, 0.2780211181640625, 0.27843890380859376, 0.2781225280761719, 0.2781531982421875, 0.2780579833984375, 0.27811019897460937, 0.278002685546875, 0.27822796630859375, 0.2780057678222656, 0.2780579833984375, 0.27817266845703126, 0.27817266845703126, 0.27806002807617186, 0.27816754150390627, 0.27811224365234377, 0.278287353515625, 0.2780968933105469, 0.2781757507324219, 0.278150146484375, 0.2783088684082031, 0.2780928039550781, 0.27805490112304687, 0.2782443542480469, 0.278150146484375, 0.27803750610351563, 0.27807437133789065, 0.27804571533203126, 0.2782361145019531, 0.27828326416015625, 0.2781214599609375, 0.27816244506835935, 0.2782443542480469, 0.27803237915039064, 0.2781583251953125, 0.27821774291992185, 0.2781603698730469, 0.27803955078125, 0.27840408325195315, 0.27810202026367187, 0.27821978759765625, 0.2781163635253906, 0.27807333374023435, 0.2780528564453125, 0.27824639892578124, 0.27811431884765625, 0.59765966796875, 0.2781757507324219, 0.27797607421875, 0.2782167053222656, 0.2780559387207031, 0.2779996032714844, 0.2779862976074219, 0.2779637756347656, 0.27797503662109374, 0.27801702880859375, 0.2779002990722656, 0.27802215576171874, 0.27807537841796875, 0.27807537841796875, 0.2780313720703125, 0.2780508117675781, 0.2779525146484375, 0.2781388854980469, 0.27807232666015624, 0.27817779541015625, 0.2780928039550781, 0.2780682373046875, 0.27811737060546876, 0.2781368408203125, 0.27811737060546876, 0.2782392272949219, 0.2781337585449219, 0.278392822265625, 0.2782310485839844, 0.278265869140625, 0.27814093017578123, 0.27818496704101564, 0.27824026489257814, 0.27826483154296877, 0.27806719970703125, 0.27855975341796874, 0.27825665283203127, 0.2781685791015625, 0.2782085266113281, 0.2781481018066406, 0.27810919189453126, 0.27824331665039065, 0.278034423828125, 0.2783047790527344, 0.2781388854980469, 0.27817166137695315, 0.2782822265625, 0.2782945251464844, 0.27814910888671873, 0.27833242797851565, 0.27806207275390626, 0.2781890563964844, 0.27830374145507814, 0.27817984008789065, 0.27814093017578123, 0.2780979309082031, 0.2780887145996094, 0.2783211669921875, 0.27862014770507815, 0.27816549682617187, 0.2781839294433594, 0.2782105712890625, 0.27824639892578124, 0.5981255493164063, 0.278150146484375, 0.27785626220703125, 0.2780334167480469, 0.2780712890625, 0.27800167846679685, 0.2779176940917969, 0.27832321166992186, 0.2780078125, 0.2780569458007813, 0.2779176940917969, 0.27797708129882814, 0.2779504699707031, 0.2781634521484375, 0.27796990966796875, 0.277923828125, 0.2779791259765625, 0.2781951904296875, 0.2780712890625, 0.2781603698730469, 0.2780784606933594, 0.27812762451171874, 0.2781521911621094, 0.2781368408203125, 0.27850137329101565, 0.2782167053222656, 0.27812045288085935, 0.2780579833984375, 0.2780405883789063, 0.27810919189453126, 0.2783846435546875, 0.2782300109863281, 0.2781163635253906, 0.278108154296875, 0.2781839294433594, 0.27849114990234375, 0.27811328125, 0.27817880249023436, 0.2781163635253906, 0.27801190185546876, 0.2782033996582031, 0.27817266845703126, 0.27813580322265624, 0.27843585205078125, 0.27829144287109375, 0.2781808776855469, 0.2781808776855469, 0.2782023620605469, 0.2781583251953125, 0.27829656982421874, 0.2781194152832031, 0.27825357055664063, 0.2780968933105469, 0.27834878540039065, 0.278118408203125, 0.2780631103515625, 0.27816244506835935, 0.27819314575195314, 0.27817880249023436, 0.27821466064453126, 0.2786324462890625, 0.2782525329589844, 0.278255615234375, 0.5979771118164062, 0.27813784790039064, 0.27801089477539065, 0.278066162109375, 0.27801190185546876, 0.278086669921875, 0.2779729919433594, 0.27828326416015625, 0.2779862976074219, 0.27811328125, 0.2778961791992188, 0.2779228210449219, 0.2780139465332031, 0.2780231628417969, 0.27794329833984377, 0.2781051025390625, 0.2780784606933594, 0.2783068237304688, 0.2780641174316406, 0.27816754150390627, 0.27818801879882815, 0.2781562805175781, 0.2779955139160156, 0.27814501953125, 0.27816448974609376, 0.27828121948242185, 0.27810919189453126, 0.2782033996582031, 0.27807537841796875, 0.2782801818847656, 0.27810406494140627, 0.27813784790039064, 0.27835186767578124, 0.27854437255859377, 0.278118408203125, 0.27813479614257813, 0.27823822021484373, 0.2782248840332031, 0.2781470642089844, 0.2781808776855469, 0.27812762451171874, 0.278139892578125, 0.27823614501953126, 0.27833139038085936, 0.27812249755859375, 0.2782443542480469, 0.2781890563964844, 0.27816549682617187, 0.27817266845703126, 0.27839590454101565, 0.27812762451171874, 0.2781603698730469, 0.27825665283203127, 0.2782945251464844, 0.2781419372558594, 0.2781552734375, 0.2782054443359375, 0.27824639892578124, 0.2780979309082031, 0.27818292236328124, 0.27818292236328124, 0.27840103149414064, 0.27840103149414064, 0.59683837890625, 0.2780917663574219, 0.27800372314453126, 0.277970947265625, 0.27823410034179685, 0.2782843017578125, 0.27795455932617186, 0.278044677734375, 0.27804568481445313, 0.2779945068359375, 0.27793508911132814, 0.27800167846679685, 0.27796685791015624, 0.27805490112304687, 0.2780139465332031, 0.27813067626953125, 0.2780088195800781, 0.2781214599609375, 0.27797402954101563, 0.27808154296875, 0.27807537841796875, 0.278118408203125, 0.2779258728027344, 0.27817062377929686, 0.2780487670898438, 0.2783273010253906, 0.2781265869140625, 0.2780579833984375, 0.2780211181640625, 0.27823822021484373, 0.2781255798339844, 0.27804568481445313, 0.2780559387207031, 0.27821978759765625, 0.27810302734375, 0.27798016357421873, 0.27811019897460937, 0.27810302734375, 0.2781695861816406, 0.27807232666015624, 0.2781419372558594, 0.27825457763671874, 0.27811019897460937, 0.2782085266113281, 0.2781265869140625, 0.2782320556640625, 0.2781665344238281, 0.27819622802734373, 0.27819827270507813, 0.27824844360351564, 0.2780887145996094, 0.27829556274414063, 0.2781214599609375, 0.27816549682617187, 0.2781255798339844, 0.27803237915039064, 0.2779996032714844, 0.27817984008789065, 0.2780989379882812, 0.2781603698730469, 0.27807949829101564, 0.2785577087402344, 0.278192138671875, 0.59820849609375, 0.27816754150390627, 0.2782525329589844, 0.27804672241210937, 0.27811737060546876, 0.27817266845703126, 0.2780159912109375, 0.27801089477539065, 0.2779852905273438, 0.27800473022460936, 0.27799755859375, 0.2780313720703125, 0.2780928039550781, 0.2781419372558594, 0.2780712890625, 0.2780354614257812, 0.27805389404296876, 0.27812762451171874, 0.27807745361328123, 0.2782248840332031, 0.27809588623046877, 0.2781470642089844, 0.27810098266601563, 0.27829351806640623, 0.2781583251953125, 0.278287353515625, 0.278150146484375, 0.2780835876464844, 0.2781767578125, 0.27830987548828123, 0.27811019897460937, 0.278029296875, 0.2782105712890625, 0.2782822265625, 0.2781481018066406, 0.278034423828125, 0.2781163635253906, 0.27817779541015625, 0.2781245422363281, 0.27811532592773436, 0.27822592163085935, 0.2782637939453125, 0.27812762451171874, 0.27823309326171874, 0.2781532287597656, 0.2782105712890625, 0.27814297485351563, 0.2782689208984375, 0.278319091796875, 0.27834060668945315, 0.2782003173828125, 0.2781808776855469, 0.278192138671875, 0.2784429931640625, 0.27812249755859375, 0.2782085266113281, 0.27821875, 0.278245361328125, 0.2782423095703125, 0.2782637939453125, 0.278245361328125, 0.2781634521484375, 0.278361083984375, 0.5991997680664063, 0.27812249755859375, 0.27820748901367187, 0.27797503662109374, 0.2780415954589844, 0.27798834228515623, 0.27795559692382815, 0.2780201110839844, 0.27805490112304687, 0.27803955078125, 0.2779053955078125, 0.27794842529296876, 0.278066162109375, 0.2780702819824219, 0.2779914245605469, 0.27794638061523436, 0.27797503662109374, 0.2781532287597656, 0.2780139465332031, 0.27818801879882815, 0.2781051025390625, 0.278108154296875, 0.27811224365234377, 0.2780989379882812, 0.27801803588867186, 0.27859457397460935, 0.27808767700195314, 0.278076416015625, 0.2780999755859375, 0.2781941833496094, 0.2783006591796875, 0.2781562805175781, 0.2780907592773437, 0.27820440673828123, 0.2780057678222656, 0.27811328125, 0.27811019897460937, 0.27817779541015625, 0.27850445556640624, 0.2780948486328125, 0.27814093017578123, 0.27819314575195314, 0.2780907592773437, 0.2782013549804688, 0.27804672241210937, 0.27818496704101564, 0.2781634521484375, 0.2782720031738281, 0.2782156677246094, 0.27832525634765626, 0.2783416442871094, 0.2781368408203125, 0.2780928039550781, 0.2781245422363281, 0.27811224365234377, 0.278624267578125, 0.27811328125, 0.2781767578125, 0.27811328125, 0.2783078308105469, 0.27821466064453126, 0.27869287109375, 0.27829248046875, 0.5999595947265625, 0.278151123046875, 0.2783006591796875, 0.2781767578125, 0.2781562805175781, 0.278287353515625, 0.27801190185546876, 0.27812762451171874, 0.27812249755859375, 0.2780569458007813, 0.27806515502929685, 0.27807745361328123, 0.27812249755859375, 0.27806207275390626, 0.2780282897949219, 0.27807232666015624, 0.27803851318359374, 0.278192138671875, 0.2781460571289063, 0.2781552734375, 0.27821875, 0.2780917663574219, 0.27806002807617186, 0.27822283935546877, 0.278044677734375, 0.27817984008789065, 0.27797607421875, 0.2783211669921875, 0.2781972351074219, 0.2782423095703125, 0.2780641174316406, 0.2781972351074219, 0.27817880249023436, 0.2783529052734375, 0.27826791381835936, 0.27813067626953125, 0.2781532287597656, 0.2781481018066406, 0.27818701171875, 0.2780784606933594, 0.27830169677734373, 0.2783303527832031, 0.2781839294433594, 0.2782423095703125, 0.27816244506835935, 0.27821875, 0.2782740478515625, 0.27820645141601563, 0.2781583251953125, 0.278361083984375, 0.27816754150390627, 0.2782095336914063, 0.2782668762207031, 0.278319091796875, 0.2781327209472656, 0.2781767578125, 0.27830374145507814, 0.27822900390625, 0.2781603698730469, 0.27811737060546876, 0.27818801879882815, 0.2782023620605469, 0.2782637939453125, 0.6001541137695312, 0.27810202026367187, 0.27797708129882814, 0.27800167846679685, 0.2780498046875, 0.27807949829101564, 0.2780989379882812, 0.2780559387207031, 0.27792486572265623, 0.2780498046875, 0.2779361267089844, 0.27795559692382815, 0.2781327209472656, 0.27822592163085935, 0.2779873352050781, 0.27796786499023435, 0.2779811706542969, 0.278044677734375, 0.27802215576171874, 0.2781317138671875, 0.27807949829101564, 0.2781245422363281, 0.2779422607421875, 0.27812966918945314, 0.2780569458007813, 0.2781808776855469, 0.2780313720703125, 0.27810714721679686, 0.2781757507324219, 0.27822796630859375, 0.2781767578125, 0.27811431884765625, 0.27811532592773436, 0.27820440673828123, 0.27810098266601563, 0.27806207275390626, 0.27838668823242185, 0.2781460571289063, 0.27807232666015624, 0.27805389404296876, 0.27812249755859375, 0.2782218322753906, 0.27806207275390626, 0.27811328125, 0.27808050537109374, 0.278255615234375, 0.2782023620605469, 0.27814910888671873, 0.27822900390625, 0.2783365173339844, 0.27817266845703126, 0.27810406494140627, 0.2780968933105469, 0.27818597412109375, 0.27826791381835936, 0.27807232666015624, 0.27808563232421873, 0.27834368896484374, 0.27808050537109374, 0.27809381103515624, 0.27811224365234377, 0.2781562805175781, 0.2782167053222656]",tokens/s,3.5369158711928956,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,openai-community/gpt2,openai-community/gpt2,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15619,7 +15619,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664e93fd-7fda79c0111fd1ca6a2ad484;ec8833c3-e551-4213-955f-286ca64d23b7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe5c1-01e1d6ff0dd96c270768a1ac;502dc5c0-80f7-4213-89fc-50aea5f885d6) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15650,7 +15650,7 @@ OSError: m is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,1,1,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15689,7 +15689,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46f0-4f381d0120c10a524a28701f;51a09c2b-cf60-49a5-b893-d6c0be84a1ab) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fea28-454d26a17516df60789e35c5;d3e68f3d-4da7-4d92-92b4-391a807b98a5) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15720,7 +15720,7 @@ OSError: 1 is not a local folder and is not a valid model identifier listed on ' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `huggingface-cli login` or by passing `token=` ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,0,0,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.1,,0.30.1,,,,1.19.2,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch @@ -15759,7 +15759,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664d46a2-0e1566c026efc0785a00e8ae;526ed273-5f97-4bb0-a5d9-e29103d1197a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-664fe9cb-04fe4598153ba53006b7cf45;ea6cc55c-d0ce-405d-b6cd-6326a24ee272) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`.