|
{ |
|
"config": { |
|
"name": "pytorch_generate", |
|
"backend": { |
|
"name": "pytorch", |
|
"version": "2.3.0+cu121", |
|
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", |
|
"task": "text-generation", |
|
"library": "transformers", |
|
"model": "google/gemma-2b", |
|
"processor": "google/gemma-2b", |
|
"device": "cuda", |
|
"device_ids": "0", |
|
"seed": 42, |
|
"inter_op_num_threads": null, |
|
"intra_op_num_threads": null, |
|
"model_kwargs": {}, |
|
"processor_kwargs": {}, |
|
"hub_kwargs": {}, |
|
"no_weights": true, |
|
"device_map": null, |
|
"torch_dtype": "float16", |
|
"eval_mode": true, |
|
"to_bettertransformer": false, |
|
"low_cpu_mem_usage": null, |
|
"attn_implementation": null, |
|
"cache_implementation": "static", |
|
"autocast_enabled": false, |
|
"autocast_dtype": null, |
|
"torch_compile": false, |
|
"torch_compile_target": "forward", |
|
"torch_compile_config": { |
|
"backend": "inductor", |
|
"mode": "reduce-overhead", |
|
"fullgraph": true |
|
}, |
|
"quantization_scheme": null, |
|
"quantization_config": {}, |
|
"deepspeed_inference": false, |
|
"deepspeed_inference_config": {}, |
|
"peft_type": null, |
|
"peft_config": {} |
|
}, |
|
"scenario": { |
|
"name": "inference", |
|
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", |
|
"iterations": 2, |
|
"duration": 0, |
|
"warmup_runs": 10, |
|
"input_shapes": { |
|
"batch_size": 1, |
|
"num_choices": 2, |
|
"sequence_length": 7 |
|
}, |
|
"new_tokens": null, |
|
"latency": true, |
|
"memory": true, |
|
"energy": false, |
|
"forward_kwargs": {}, |
|
"generate_kwargs": { |
|
"max_new_tokens": 128, |
|
"min_new_tokens": 128, |
|
"do_sample": false |
|
}, |
|
"call_kwargs": {} |
|
}, |
|
"launcher": { |
|
"name": "process", |
|
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", |
|
"device_isolation": true, |
|
"device_isolation_action": "warn", |
|
"start_method": "spawn" |
|
}, |
|
"environment": { |
|
"cpu": " AMD EPYC 7R32", |
|
"cpu_count": 16, |
|
"cpu_ram_mb": 66697.29792, |
|
"system": "Linux", |
|
"machine": "x86_64", |
|
"platform": "Linux-5.10.216-204.855.amzn2.x86_64-x86_64-with-glibc2.29", |
|
"processor": "x86_64", |
|
"python_version": "3.8.10", |
|
"gpu": [ |
|
"NVIDIA A10G" |
|
], |
|
"gpu_count": 1, |
|
"gpu_vram_mb": 24146608128, |
|
"optimum_benchmark_version": "0.2.1", |
|
"optimum_benchmark_commit": null, |
|
"transformers_version": "4.42.0.dev0", |
|
"transformers_commit": "a51a3fbc4827db26e896292b6409246fe8152cd4", |
|
"accelerate_version": "0.31.0.dev0", |
|
"accelerate_commit": null, |
|
"diffusers_version": null, |
|
"diffusers_commit": null, |
|
"optimum_version": "1.21.0.dev0", |
|
"optimum_commit": null, |
|
"timm_version": "0.9.16", |
|
"timm_commit": null, |
|
"peft_version": "0.11.2.dev0", |
|
"peft_commit": null |
|
} |
|
}, |
|
"report": { |
|
"prefill": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 1753.206784, |
|
"max_global_vram": 6793.199616, |
|
"max_process_vram": 0.0, |
|
"max_reserved": 6146.752512, |
|
"max_allocated": 5035.912192 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"count": 2, |
|
"total": 0.04486032104492188, |
|
"mean": 0.02243016052246094, |
|
"stdev": 0.0003491039276123046, |
|
"p50": 0.02243016052246094, |
|
"p90": 0.022709443664550782, |
|
"p95": 0.022744354057312014, |
|
"p99": 0.022772282371520996, |
|
"values": [ |
|
0.022779264450073243, |
|
0.022081056594848634 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 312.0797995623078 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"decode": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 1753.227264, |
|
"max_global_vram": 6793.199616, |
|
"max_process_vram": 0.0, |
|
"max_reserved": 6146.752512, |
|
"max_allocated": 5035.912704 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"count": 2, |
|
"total": 5.246615478515625, |
|
"mean": 2.6233077392578124, |
|
"stdev": 0.01454187011718755, |
|
"p50": 2.6233077392578124, |
|
"p90": 2.634941235351562, |
|
"p95": 2.636395422363281, |
|
"p99": 2.6375587719726563, |
|
"values": [ |
|
2.637849609375, |
|
2.608765869140625 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 48.4121622863549 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"per_token": { |
|
"memory": null, |
|
"latency": { |
|
"unit": "s", |
|
"count": 253, |
|
"total": 5.24776754379273, |
|
"mean": 0.020742164204714326, |
|
"stdev": 0.0014777911921905315, |
|
"p50": 0.020716543197631835, |
|
"p90": 0.02107084846496582, |
|
"p95": 0.021373951721191407, |
|
"p99": 0.021995028305053704, |
|
"values": [ |
|
0.021323776245117186, |
|
0.021770240783691407, |
|
0.021406719207763672, |
|
0.02088243293762207, |
|
0.02087116813659668, |
|
0.021331968307495116, |
|
0.020982784271240236, |
|
0.02126335906982422, |
|
0.021222400665283202, |
|
0.020972543716430665, |
|
0.020864000320434572, |
|
0.020743167877197266, |
|
0.02089369583129883, |
|
0.02146816062927246, |
|
0.02123161506652832, |
|
0.021181440353393553, |
|
0.02081279945373535, |
|
0.020891647338867187, |
|
0.0210513916015625, |
|
0.021132287979125978, |
|
0.020974592208862306, |
|
0.020703231811523438, |
|
0.020632575988769532, |
|
0.020570112228393556, |
|
0.020771839141845702, |
|
0.021409791946411134, |
|
0.021580799102783203, |
|
0.020961280822753905, |
|
0.02089574432373047, |
|
0.02101759910583496, |
|
0.0208855037689209, |
|
0.02071347236633301, |
|
0.021361663818359376, |
|
0.021392383575439454, |
|
0.02101043128967285, |
|
0.021163007736206055, |
|
0.02067967987060547, |
|
0.021167104721069335, |
|
0.020644863128662108, |
|
0.020971519470214844, |
|
0.021176319122314453, |
|
0.02223411178588867, |
|
0.02451968002319336, |
|
0.021395456314086913, |
|
0.02105548858642578, |
|
0.020931583404541015, |
|
0.02082713508605957, |
|
0.020758527755737305, |
|
0.02107084846496582, |
|
0.020722688674926756, |
|
0.020327423095703127, |
|
0.020794368743896483, |
|
0.02082815933227539, |
|
0.02082611274719238, |
|
0.02083839988708496, |
|
0.020709375381469726, |
|
0.020758527755737305, |
|
0.020716543197631835, |
|
0.020755456924438476, |
|
0.020714496612548827, |
|
0.020739072799682616, |
|
0.02062131118774414, |
|
0.020801536560058592, |
|
0.02025267219543457, |
|
0.020739072799682616, |
|
0.02083839988708496, |
|
0.02081177520751953, |
|
0.02072166442871094, |
|
0.02066227149963379, |
|
0.020196352005004883, |
|
0.019991552352905274, |
|
0.01999564743041992, |
|
0.019998720169067383, |
|
0.02008678436279297, |
|
0.020344831466674804, |
|
0.020782079696655274, |
|
0.020695039749145508, |
|
0.02030182456970215, |
|
0.0202608642578125, |
|
0.020701183319091796, |
|
0.019976192474365235, |
|
0.02028441619873047, |
|
0.020746240615844725, |
|
0.020685823440551757, |
|
0.020612096786499022, |
|
0.02063871955871582, |
|
0.02087014389038086, |
|
0.019937280654907227, |
|
0.019949567794799804, |
|
0.019953664779663087, |
|
0.019940351486206053, |
|
0.020728832244873048, |
|
0.020632575988769532, |
|
0.0206878719329834, |
|
0.02061516761779785, |
|
0.02063974380493164, |
|
0.019948543548583983, |
|
0.020717567443847656, |
|
0.020714496612548827, |
|
0.02068889617919922, |
|
0.020726783752441406, |
|
0.02066022491455078, |
|
0.020603904724121092, |
|
0.020617216110229493, |
|
0.020617216110229493, |
|
0.020273151397705077, |
|
0.02082815933227539, |
|
0.020702207565307617, |
|
0.020653055191040038, |
|
0.019981311798095702, |
|
0.019960832595825196, |
|
0.020455423355102538, |
|
0.02007756805419922, |
|
0.01982054328918457, |
|
0.020579328536987306, |
|
0.020322303771972656, |
|
0.020708351135253905, |
|
0.020668415069580077, |
|
0.02072166442871094, |
|
0.021144575119018554, |
|
0.021774335861206053, |
|
0.02085478401184082, |
|
0.020724735260009765, |
|
0.02065920066833496, |
|
0.02035916709899902, |
|
0.02011955261230469, |
|
0.04295475387573242, |
|
0.020708351135253905, |
|
0.019990528106689453, |
|
0.019962879180908204, |
|
0.020368383407592772, |
|
0.0208721923828125, |
|
0.02165350341796875, |
|
0.02107084846496582, |
|
0.020855808258056642, |
|
0.020730880737304686, |
|
0.020884479522705078, |
|
0.020740095138549804, |
|
0.020786176681518553, |
|
0.020734975814819336, |
|
0.020716543197631835, |
|
0.020745216369628908, |
|
0.020724735260009765, |
|
0.020797439575195312, |
|
0.02026188850402832, |
|
0.020767744064331056, |
|
0.020502527236938475, |
|
0.020767744064331056, |
|
0.02068992042541504, |
|
0.02082508850097656, |
|
0.02068070411682129, |
|
0.02004991912841797, |
|
0.02003763198852539, |
|
0.020716543197631835, |
|
0.020716543197631835, |
|
0.020710399627685547, |
|
0.020730880737304686, |
|
0.020674560546875, |
|
0.02023423957824707, |
|
0.02072985649108887, |
|
0.02068070411682129, |
|
0.020832256317138673, |
|
0.020892671585083008, |
|
0.020739072799682616, |
|
0.020015104293823242, |
|
0.019982336044311523, |
|
0.019965951919555663, |
|
0.019992576599121094, |
|
0.019947519302368166, |
|
0.02070425605773926, |
|
0.020677631378173827, |
|
0.02068992042541504, |
|
0.020695039749145508, |
|
0.02064691162109375, |
|
0.02000588798522949, |
|
0.020946943283081054, |
|
0.020817920684814452, |
|
0.020716543197631835, |
|
0.020719615936279297, |
|
0.020744192123413087, |
|
0.019956735610961913, |
|
0.019956735610961913, |
|
0.020151296615600587, |
|
0.020695039749145508, |
|
0.019978239059448243, |
|
0.020025344848632814, |
|
0.02002739143371582, |
|
0.020912128448486327, |
|
0.020751359939575196, |
|
0.021551103591918946, |
|
0.020757503509521484, |
|
0.020754432678222655, |
|
0.02068070411682129, |
|
0.02090188789367676, |
|
0.020797439575195312, |
|
0.020725759506225586, |
|
0.020625408172607423, |
|
0.0206878719329834, |
|
0.0208035831451416, |
|
0.020732927322387695, |
|
0.02066534423828125, |
|
0.020711423873901368, |
|
0.020676607131958007, |
|
0.020807680130004884, |
|
0.020740095138549804, |
|
0.020807680130004884, |
|
0.020742143630981445, |
|
0.020723712921142577, |
|
0.020755456924438476, |
|
0.02080460739135742, |
|
0.020802560806274413, |
|
0.02086195182800293, |
|
0.020714496612548827, |
|
0.020738048553466795, |
|
0.020720640182495118, |
|
0.020718591690063477, |
|
0.020718591690063477, |
|
0.020011007308959963, |
|
0.020006912231445313, |
|
0.019958784103393554, |
|
0.019991552352905274, |
|
0.019952640533447266, |
|
0.019936256408691407, |
|
0.020339712142944336, |
|
0.020789247512817383, |
|
0.020711423873901368, |
|
0.02081279945373535, |
|
0.020736000061035157, |
|
0.02068889617919922, |
|
0.020039680480957032, |
|
0.020188159942626953, |
|
0.0200447998046875, |
|
0.01997209548950195, |
|
0.019960832595825196, |
|
0.020025344848632814, |
|
0.020165632247924805, |
|
0.020776960372924806, |
|
0.02082099151611328, |
|
0.020731903076171874, |
|
0.020715520858764647, |
|
0.020684799194335936, |
|
0.020039680480957032, |
|
0.020714496612548827, |
|
0.020782079696655274, |
|
0.01992192077636719, |
|
0.019974143981933593, |
|
0.019974143981933593, |
|
0.020953088760375976, |
|
0.020746240615844725, |
|
0.020777984619140624, |
|
0.02002124786376953, |
|
0.019961856842041017, |
|
0.019958784103393554 |
|
] |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 48.21097693232598 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
} |
|
} |
|
} |