|
{ |
|
"config": { |
|
"name": "pytorch_generate", |
|
"backend": { |
|
"name": "pytorch", |
|
"version": "2.3.0+cu121", |
|
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", |
|
"task": "text-generation", |
|
"library": "transformers", |
|
"model": "google/gemma-2b", |
|
"processor": "google/gemma-2b", |
|
"device": "cuda", |
|
"device_ids": "0", |
|
"seed": 42, |
|
"inter_op_num_threads": null, |
|
"intra_op_num_threads": null, |
|
"model_kwargs": {}, |
|
"processor_kwargs": {}, |
|
"hub_kwargs": {}, |
|
"no_weights": true, |
|
"device_map": null, |
|
"torch_dtype": "float16", |
|
"eval_mode": true, |
|
"to_bettertransformer": false, |
|
"low_cpu_mem_usage": null, |
|
"attn_implementation": null, |
|
"cache_implementation": null, |
|
"autocast_enabled": false, |
|
"autocast_dtype": null, |
|
"torch_compile": false, |
|
"torch_compile_target": "forward", |
|
"torch_compile_config": { |
|
"backend": "inductor", |
|
"mode": "reduce-overhead", |
|
"fullgraph": true |
|
}, |
|
"quantization_scheme": null, |
|
"quantization_config": {}, |
|
"deepspeed_inference": false, |
|
"deepspeed_inference_config": {}, |
|
"peft_type": null, |
|
"peft_config": {} |
|
}, |
|
"scenario": { |
|
"name": "inference", |
|
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", |
|
"iterations": 2, |
|
"duration": 0, |
|
"warmup_runs": 10, |
|
"input_shapes": { |
|
"batch_size": 1, |
|
"num_choices": 2, |
|
"sequence_length": 7 |
|
}, |
|
"new_tokens": null, |
|
"latency": true, |
|
"memory": true, |
|
"energy": false, |
|
"forward_kwargs": {}, |
|
"generate_kwargs": { |
|
"max_new_tokens": 128, |
|
"min_new_tokens": 128, |
|
"do_sample": false |
|
}, |
|
"call_kwargs": {} |
|
}, |
|
"launcher": { |
|
"name": "process", |
|
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", |
|
"device_isolation": true, |
|
"device_isolation_action": "warn", |
|
"start_method": "spawn" |
|
}, |
|
"environment": { |
|
"cpu": " AMD EPYC 7R32", |
|
"cpu_count": 16, |
|
"cpu_ram_mb": 66697.29792, |
|
"system": "Linux", |
|
"machine": "x86_64", |
|
"platform": "Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.29", |
|
"processor": "x86_64", |
|
"python_version": "3.8.10", |
|
"gpu": [ |
|
"NVIDIA A10G" |
|
], |
|
"gpu_count": 1, |
|
"gpu_vram_mb": 24146608128, |
|
"optimum_benchmark_version": "0.2.1", |
|
"optimum_benchmark_commit": null, |
|
"transformers_version": "4.42.0.dev0", |
|
"transformers_commit": "0ed3ffcb4461a244b87781a24e5ebd0a78f98142", |
|
"accelerate_version": "0.32.0.dev0", |
|
"accelerate_commit": null, |
|
"diffusers_version": null, |
|
"diffusers_commit": null, |
|
"optimum_version": "1.21.0.dev0", |
|
"optimum_commit": null, |
|
"timm_version": "0.9.16", |
|
"timm_commit": null, |
|
"peft_version": "0.11.2.dev0", |
|
"peft_commit": null |
|
} |
|
}, |
|
"report": { |
|
"prefill": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 1708.326912, |
|
"max_global_vram": 6793.199616, |
|
"max_process_vram": 0.0, |
|
"max_reserved": 6146.752512, |
|
"max_allocated": 5034.576896 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"count": 2, |
|
"total": 0.03926543998718262, |
|
"mean": 0.01963271999359131, |
|
"stdev": 0.0005087194442749028, |
|
"p50": 0.01963271999359131, |
|
"p90": 0.020039695549011232, |
|
"p95": 0.02009056749343872, |
|
"p99": 0.020131265048980714, |
|
"values": [ |
|
0.020141439437866213, |
|
0.019124000549316407 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 356.54764099345397 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"decode": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 1708.40064, |
|
"max_global_vram": 6795.296768, |
|
"max_process_vram": 0.0, |
|
"max_reserved": 6148.849664, |
|
"max_allocated": 5034.577408 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"count": 2, |
|
"total": 4.689177001953125, |
|
"mean": 2.3445885009765624, |
|
"stdev": 0.010478149414062443, |
|
"p50": 2.3445885009765624, |
|
"p90": 2.3529710205078125, |
|
"p95": 2.3540188354492186, |
|
"p99": 2.3548570874023436, |
|
"values": [ |
|
2.355066650390625, |
|
2.3341103515625 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 54.16728775522966 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"per_token": { |
|
"memory": null, |
|
"latency": { |
|
"unit": "s", |
|
"count": 253, |
|
"total": 4.68916942024231, |
|
"mean": 0.01853426648317118, |
|
"stdev": 0.0012692450445622333, |
|
"p50": 0.018335744857788085, |
|
"p90": 0.019147596740722656, |
|
"p95": 0.019481190490722656, |
|
"p99": 0.019787325820922852, |
|
"values": [ |
|
0.01946009635925293, |
|
0.019739648818969727, |
|
0.01983897590637207, |
|
0.019528703689575197, |
|
0.019710975646972655, |
|
0.019612672805786133, |
|
0.01861222457885742, |
|
0.01821183967590332, |
|
0.01824358367919922, |
|
0.01841152000427246, |
|
0.018721792221069337, |
|
0.018892799377441406, |
|
0.018497535705566406, |
|
0.018299903869628906, |
|
0.018135040283203126, |
|
0.018108415603637695, |
|
0.018354175567626953, |
|
0.01815449523925781, |
|
0.018242559432983398, |
|
0.018198528289794923, |
|
0.01803775978088379, |
|
0.01828556823730469, |
|
0.018257919311523436, |
|
0.018136064529418947, |
|
0.01803059196472168, |
|
0.017957887649536132, |
|
0.01800396728515625, |
|
0.01802956771850586, |
|
0.018283519744873047, |
|
0.01804697608947754, |
|
0.018342912673950194, |
|
0.018059263229370116, |
|
0.017937408447265626, |
|
0.018713600158691408, |
|
0.018386943817138672, |
|
0.018373632431030275, |
|
0.018473983764648438, |
|
0.018167808532714845, |
|
0.01847500801086426, |
|
0.01820364761352539, |
|
0.018440191268920898, |
|
0.018330623626708984, |
|
0.018148416519165038, |
|
0.018210752487182617, |
|
0.018231296539306642, |
|
0.01843097686767578, |
|
0.01862553596496582, |
|
0.01838387107849121, |
|
0.018339839935302735, |
|
0.018372608184814454, |
|
0.018358272552490236, |
|
0.018357248306274415, |
|
0.018321407318115233, |
|
0.01859174346923828, |
|
0.01859174346923828, |
|
0.018732032775878905, |
|
0.019372032165527343, |
|
0.019527679443359376, |
|
0.01843916893005371, |
|
0.018568191528320312, |
|
0.01843302345275879, |
|
0.01843097686767578, |
|
0.018531328201293946, |
|
0.018404352188110353, |
|
0.018495487213134765, |
|
0.01822719955444336, |
|
0.018331647872924805, |
|
0.018552831649780274, |
|
0.018509824752807616, |
|
0.018381824493408205, |
|
0.018525184631347655, |
|
0.018408447265625, |
|
0.018231296539306642, |
|
0.018489343643188477, |
|
0.018451616287231444, |
|
0.01834582328796387, |
|
0.018335775375366212, |
|
0.018428895950317385, |
|
0.01842278480529785, |
|
0.018380800247192384, |
|
0.02051481628417969, |
|
0.019514368057250975, |
|
0.019304447174072266, |
|
0.019526655197143555, |
|
0.018280448913574218, |
|
0.018405376434326173, |
|
0.018109439849853515, |
|
0.01807257652282715, |
|
0.018669567108154296, |
|
0.018519039154052733, |
|
0.018528255462646484, |
|
0.01866854476928711, |
|
0.018490432739257812, |
|
0.01820460891723633, |
|
0.01803980827331543, |
|
0.01808076858520508, |
|
0.018165760040283203, |
|
0.01817804718017578, |
|
0.018172927856445312, |
|
0.018077695846557617, |
|
0.018273279190063475, |
|
0.01827020835876465, |
|
0.018239488601684572, |
|
0.01823641586303711, |
|
0.01820876884460449, |
|
0.01883443260192871, |
|
0.019412992477416992, |
|
0.019272703170776367, |
|
0.01946112060546875, |
|
0.019148799896240236, |
|
0.019171327590942384, |
|
0.01946112060546875, |
|
0.01947750473022461, |
|
0.019586048126220702, |
|
0.019207168579101562, |
|
0.01902079963684082, |
|
0.01825279998779297, |
|
0.018214912414550782, |
|
0.018543615341186523, |
|
0.018507776260375978, |
|
0.018371583938598633, |
|
0.018274303436279296, |
|
0.01820057678222656, |
|
0.018229248046875, |
|
0.01824051284790039, |
|
0.018396160125732423, |
|
0.03759308624267578, |
|
0.018299903869628906, |
|
0.01824870491027832, |
|
0.018372608184814454, |
|
0.018330623626708984, |
|
0.018152448654174806, |
|
0.018428928375244142, |
|
0.018324480056762696, |
|
0.0182108154296875, |
|
0.018241535186767577, |
|
0.018168832778930662, |
|
0.018120704650878908, |
|
0.01843302345275879, |
|
0.01830297660827637, |
|
0.018541568756103514, |
|
0.018552831649780274, |
|
0.018148351669311523, |
|
0.018112512588500978, |
|
0.01815449523925781, |
|
0.018142208099365235, |
|
0.018291744232177734, |
|
0.01809916877746582, |
|
0.018268159866333008, |
|
0.018274303436279296, |
|
0.01816166305541992, |
|
0.018076671600341796, |
|
0.01820467185974121, |
|
0.018274400711059572, |
|
0.018233247756958008, |
|
0.018486272811889647, |
|
0.018306047439575195, |
|
0.018142208099365235, |
|
0.018258943557739257, |
|
0.018185216903686522, |
|
0.01818726348876953, |
|
0.018300928115844727, |
|
0.018335744857788085, |
|
0.018347007751464844, |
|
0.018157567977905274, |
|
0.018168832778930662, |
|
0.01798860740661621, |
|
0.018136064529418947, |
|
0.01820057678222656, |
|
0.018759679794311524, |
|
0.018800640106201173, |
|
0.019127296447753905, |
|
0.019204095840454103, |
|
0.018190336227416993, |
|
0.018486272811889647, |
|
0.018397184371948243, |
|
0.01847603225708008, |
|
0.018299903869628906, |
|
0.018198528289794923, |
|
0.018211999893188478, |
|
0.018141151428222656, |
|
0.018199424743652343, |
|
0.018469888687133788, |
|
0.018311168670654295, |
|
0.018336767196655272, |
|
0.019356672286987304, |
|
0.018765823364257812, |
|
0.01845145606994629, |
|
0.018329599380493163, |
|
0.01823641586303711, |
|
0.018257919311523436, |
|
0.01884774398803711, |
|
0.018264064788818358, |
|
0.018386943817138672, |
|
0.018465791702270508, |
|
0.018326528549194337, |
|
0.01848422431945801, |
|
0.018533376693725585, |
|
0.018473983764648438, |
|
0.01824460792541504, |
|
0.018241600036621095, |
|
0.018298816680908204, |
|
0.01823232078552246, |
|
0.01838489532470703, |
|
0.018358272552490236, |
|
0.018480127334594726, |
|
0.018496511459350586, |
|
0.01860915184020996, |
|
0.018506752014160157, |
|
0.018496511459350586, |
|
0.018350080490112306, |
|
0.018912256240844725, |
|
0.018318336486816408, |
|
0.018249727249145507, |
|
0.018317312240600587, |
|
0.01821183967590332, |
|
0.018362367630004883, |
|
0.018076671600341796, |
|
0.018126848220825196, |
|
0.018367488861083983, |
|
0.01823846435546875, |
|
0.01860403251647949, |
|
0.019486719131469727, |
|
0.018938880920410156, |
|
0.01884671974182129, |
|
0.018721792221069337, |
|
0.018304000854492186, |
|
0.018447359085083007, |
|
0.01843097686767578, |
|
0.01832467269897461, |
|
0.01815839958190918, |
|
0.018305023193359374, |
|
0.018324480056762696, |
|
0.018453504562377928, |
|
0.018299903869628906, |
|
0.018206720352172853, |
|
0.01842995262145996, |
|
0.018295808792114256, |
|
0.018177024841308592, |
|
0.01797222328186035, |
|
0.01805619239807129, |
|
0.018041856765747072, |
|
0.017958911895751953, |
|
0.01815769577026367, |
|
0.017997695922851564, |
|
0.0179814395904541, |
|
0.018223104476928712, |
|
0.018231296539306642, |
|
0.019142784118652344, |
|
0.018554752349853515, |
|
0.019607551574707033, |
|
0.018931711196899414, |
|
0.019009536743164062 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 53.954117952711215 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
} |
|
} |
|
} |