|
{ |
|
"config": { |
|
"name": "pytorch-llama", |
|
"backend": { |
|
"name": "pytorch", |
|
"version": "2.5.1", |
|
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", |
|
"task": "text-generation", |
|
"library": "transformers", |
|
"model_type": "llama", |
|
"model": "meta-llama/Llama-2-7b-chat-hf", |
|
"processor": "meta-llama/Llama-2-7b-chat-hf", |
|
"device": "cuda", |
|
"device_ids": "0", |
|
"seed": 42, |
|
"inter_op_num_threads": null, |
|
"intra_op_num_threads": null, |
|
"model_kwargs": {}, |
|
"processor_kwargs": {}, |
|
"no_weights": true, |
|
"device_map": null, |
|
"torch_dtype": "float16", |
|
"eval_mode": true, |
|
"to_bettertransformer": false, |
|
"low_cpu_mem_usage": null, |
|
"attn_implementation": null, |
|
"cache_implementation": null, |
|
"autocast_enabled": false, |
|
"autocast_dtype": null, |
|
"torch_compile": false, |
|
"torch_compile_target": "forward", |
|
"torch_compile_config": {}, |
|
"quantization_scheme": "awq", |
|
"quantization_config": { |
|
"bits": 4, |
|
"version": "gemm" |
|
}, |
|
"deepspeed_inference": false, |
|
"deepspeed_inference_config": {}, |
|
"peft_type": null, |
|
"peft_config": {} |
|
}, |
|
"scenario": { |
|
"name": "inference", |
|
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", |
|
"iterations": 10, |
|
"duration": 10, |
|
"warmup_runs": 10, |
|
"input_shapes": { |
|
"batch_size": 1, |
|
"num_choices": 2, |
|
"sequence_length": 128 |
|
}, |
|
"new_tokens": null, |
|
"memory": true, |
|
"latency": true, |
|
"energy": false, |
|
"forward_kwargs": {}, |
|
"generate_kwargs": { |
|
"max_new_tokens": 32, |
|
"min_new_tokens": 32 |
|
}, |
|
"call_kwargs": {} |
|
}, |
|
"launcher": { |
|
"name": "process", |
|
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", |
|
"device_isolation": true, |
|
"device_isolation_action": "warn", |
|
"numactl": false, |
|
"numactl_kwargs": {}, |
|
"start_method": "spawn" |
|
}, |
|
"environment": { |
|
"cpu": " Intel(R) Xeon(R) Platinum 8339HC CPU @ 1.80GHz", |
|
"cpu_count": 22, |
|
"cpu_ram_mb": 189584.162816, |
|
"system": "Linux", |
|
"machine": "x86_64", |
|
"platform": "Linux-5.19.0-0_fbk12_hardened_11583_g0bef9520ca2b-x86_64-with-glibc2.34", |
|
"processor": "x86_64", |
|
"python_version": "3.10.15", |
|
"gpu": [ |
|
"NVIDIA PG509-210" |
|
], |
|
"gpu_count": 1, |
|
"gpu_vram_mb": 85899345920, |
|
"optimum_benchmark_version": "0.5.0.dev0", |
|
"optimum_benchmark_commit": "78d7da546ad060a586060d0d3680bec33f32a45b", |
|
"transformers_version": "4.47.0.dev0", |
|
"transformers_commit": null, |
|
"accelerate_version": "1.1.0", |
|
"accelerate_commit": null, |
|
"diffusers_version": "0.31.0", |
|
"diffusers_commit": null, |
|
"optimum_version": "1.24.0.dev0", |
|
"optimum_commit": null, |
|
"timm_version": "1.0.11", |
|
"timm_commit": null, |
|
"peft_version": "0.13.2", |
|
"peft_commit": null |
|
}, |
|
"print_report": true, |
|
"log_report": true |
|
}, |
|
"report": { |
|
"load": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 6349.856768, |
|
"max_global_vram": 23709.220864, |
|
"max_process_vram": 5286.920192, |
|
"max_reserved": 4211.081216, |
|
"max_allocated": 3939.656192 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
6.64453759765625 |
|
], |
|
"count": 1, |
|
"total": 6.64453759765625, |
|
"mean": 6.64453759765625, |
|
"p50": 6.64453759765625, |
|
"p90": 6.64453759765625, |
|
"p95": 6.64453759765625, |
|
"p99": 6.64453759765625, |
|
"stdev": 0, |
|
"stdev_": 0 |
|
}, |
|
"throughput": null, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"prefill": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 4635.348992, |
|
"max_global_vram": 24478.875648, |
|
"max_process_vram": 5668.601856, |
|
"max_reserved": 4330.61888, |
|
"max_allocated": 4048.256512 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
0.05503110504150391, |
|
0.05503148651123047, |
|
0.056178112030029294, |
|
0.05623561477661133, |
|
0.05539152145385742, |
|
0.05583244705200195, |
|
0.054368606567382814, |
|
0.0565338249206543, |
|
0.05576860809326172, |
|
0.05855392074584961 |
|
], |
|
"count": 10, |
|
"total": 0.5589252471923828, |
|
"mean": 0.05589252471923828, |
|
"p50": 0.05580052757263183, |
|
"p90": 0.05673583450317383, |
|
"p95": 0.05764487762451172, |
|
"p99": 0.058372112121582026, |
|
"stdev": 0.0010848636677924092, |
|
"stdev_": 1.9409816844773837 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 2290.1094671062915 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"decode": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 4635.348992, |
|
"max_global_vram": 24642.453504, |
|
"max_process_vram": 5668.601856, |
|
"max_reserved": 4330.61888, |
|
"max_allocated": 4048.258048 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
1.717583984375, |
|
1.7140399169921876, |
|
1.71884130859375, |
|
1.7282589111328126, |
|
1.7339022216796875, |
|
1.7213660888671876, |
|
1.7198265380859374, |
|
1.69797265625, |
|
1.7129921875, |
|
1.8204344482421875 |
|
], |
|
"count": 10, |
|
"total": 17.28521826171875, |
|
"mean": 1.728521826171875, |
|
"p50": 1.7193339233398437, |
|
"p90": 1.7425554443359375, |
|
"p95": 1.7814949462890624, |
|
"p99": 1.8126465478515625, |
|
"stdev": 0.03193940582522888, |
|
"stdev_": 1.8477872446635217 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 17.93439893591342 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"per_token": { |
|
"memory": null, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
0.05596720123291016, |
|
0.05637516784667969, |
|
0.05543753433227539, |
|
0.05507980728149414, |
|
0.05437231826782227, |
|
0.055271968841552735, |
|
0.056059070587158207, |
|
0.056449600219726566, |
|
0.05596623992919922, |
|
0.055463455200195313, |
|
0.05580003356933594, |
|
0.05514332962036133, |
|
0.055476062774658205, |
|
0.05491804885864258, |
|
0.054805694580078126, |
|
0.054357696533203125, |
|
0.055393089294433595, |
|
0.054887840270996094, |
|
0.05500073623657226, |
|
0.05470105743408203, |
|
0.05624684906005859, |
|
0.055674560546875, |
|
0.054724830627441406, |
|
0.0550332145690918, |
|
0.055050785064697266, |
|
0.05457702255249024, |
|
0.054798465728759765, |
|
0.05624467086791992, |
|
0.0560634880065918, |
|
0.05558377456665039, |
|
0.05559833526611328, |
|
0.05444054412841797, |
|
0.05430252838134766, |
|
0.05439487838745117, |
|
0.055029857635498045, |
|
0.055554271697998044, |
|
0.05528985595703125, |
|
0.05566563034057617, |
|
0.0559892807006836, |
|
0.05581078338623047, |
|
0.055621471405029294, |
|
0.05533734512329101, |
|
0.056597633361816405, |
|
0.055871902465820314, |
|
0.05509036636352539, |
|
0.05506854248046875, |
|
0.05513759994506836, |
|
0.055596576690673825, |
|
0.05539929580688477, |
|
0.0551014404296875, |
|
0.055277217864990236, |
|
0.05462879943847656, |
|
0.05518960189819336, |
|
0.055173408508300784, |
|
0.05521807861328125, |
|
0.05475705718994141, |
|
0.05470764923095703, |
|
0.05541244888305664, |
|
0.055156097412109376, |
|
0.05496499252319336, |
|
0.05455446243286133, |
|
0.056657470703125, |
|
0.05545353698730469, |
|
0.055706527709960936, |
|
0.055823455810546874, |
|
0.05598230361938476, |
|
0.055120574951171876, |
|
0.05478438568115234, |
|
0.05467168045043945, |
|
0.05517049789428711, |
|
0.05529232025146484, |
|
0.05538246536254883, |
|
0.054830142974853516, |
|
0.05468191909790039, |
|
0.05496223831176758, |
|
0.05508172988891601, |
|
0.057529472351074216, |
|
0.05543775939941406, |
|
0.05504735946655273, |
|
0.056652385711669924, |
|
0.05581151962280274, |
|
0.05540220642089844, |
|
0.055823711395263674, |
|
0.055400318145751956, |
|
0.05468521499633789, |
|
0.054793087005615235, |
|
0.05524348831176758, |
|
0.05561004638671875, |
|
0.0554854736328125, |
|
0.05575203323364258, |
|
0.05517225646972656, |
|
0.055634910583496094, |
|
0.05538291168212891, |
|
0.055080894470214845, |
|
0.05647171020507812, |
|
0.05586511993408203, |
|
0.056236286163330075, |
|
0.0558397102355957, |
|
0.054919456481933596, |
|
0.05532627105712891, |
|
0.05551532745361328, |
|
0.05585120010375977, |
|
0.05585951995849609, |
|
0.05586428833007812, |
|
0.05486710357666016, |
|
0.05557219314575195, |
|
0.05467580795288086, |
|
0.05549046325683594, |
|
0.05546985626220703, |
|
0.05542700958251953, |
|
0.055618526458740235, |
|
0.05524524688720703, |
|
0.05667580795288086, |
|
0.055995777130126954, |
|
0.05630582427978516, |
|
0.055726558685302734, |
|
0.05565529632568359, |
|
0.055659648895263675, |
|
0.05626460647583008, |
|
0.05617097473144531, |
|
0.05771820831298828, |
|
0.055132320404052734, |
|
0.055727775573730466, |
|
0.05502320098876953, |
|
0.05532460784912109, |
|
0.056063232421875, |
|
0.05610335922241211, |
|
0.05521129608154297, |
|
0.05461382293701172, |
|
0.05621088027954101, |
|
0.05493718338012695, |
|
0.055665790557861326, |
|
0.05619779205322266, |
|
0.058518657684326174, |
|
0.055929889678955076, |
|
0.055870975494384766, |
|
0.055893535614013674, |
|
0.05570566558837891, |
|
0.05599852752685547, |
|
0.05544598388671875, |
|
0.05592911911010742, |
|
0.05619827270507813, |
|
0.05598297500610352, |
|
0.05601696014404297, |
|
0.05518793487548828, |
|
0.05504118347167969, |
|
0.0560975341796875, |
|
0.05725628662109375, |
|
0.056258049011230465, |
|
0.0561228141784668, |
|
0.05579196929931641, |
|
0.05590393447875976, |
|
0.055871585845947265, |
|
0.055774593353271486, |
|
0.05570595169067383, |
|
0.05606550216674805, |
|
0.05573023986816406, |
|
0.055296417236328124, |
|
0.05504934310913086, |
|
0.05574620819091797, |
|
0.05592668914794922, |
|
0.05649148941040039, |
|
0.0556003532409668, |
|
0.05506803131103516, |
|
0.055417152404785154, |
|
0.05743433761596679, |
|
0.056299358367919924, |
|
0.05565270233154297, |
|
0.05635913467407227, |
|
0.05638422393798828, |
|
0.054519424438476564, |
|
0.05492627334594727, |
|
0.0552775993347168, |
|
0.05675699234008789, |
|
0.05520207977294922, |
|
0.054873950958251955, |
|
0.054984222412109374, |
|
0.055158687591552735, |
|
0.055889854431152346, |
|
0.056033313751220705, |
|
0.054382305145263675, |
|
0.053922527313232424, |
|
0.0544095344543457, |
|
0.05602323150634766, |
|
0.05501222229003906, |
|
0.054485694885253906, |
|
0.05455244827270508, |
|
0.05586511993408203, |
|
0.05601020812988281, |
|
0.055950656890869144, |
|
0.053898048400878903, |
|
0.05543491363525391, |
|
0.05637590408325195, |
|
0.055580577850341796, |
|
0.05497836685180664, |
|
0.05409417724609375, |
|
0.054698112487792966, |
|
0.05414665603637695, |
|
0.054037246704101566, |
|
0.05524428939819336, |
|
0.056038494110107424, |
|
0.05577017593383789, |
|
0.055400928497314456, |
|
0.054889118194580075, |
|
0.05563926315307617, |
|
0.05477939224243164, |
|
0.055285247802734375, |
|
0.05561484909057617, |
|
0.05592176055908203, |
|
0.05637519836425781, |
|
0.05623830413818359, |
|
0.05617081451416016, |
|
0.0559967041015625, |
|
0.055958625793457034, |
|
0.05606281661987305, |
|
0.05589920043945312, |
|
0.05584953689575195, |
|
0.05704137420654297, |
|
0.05657097625732422, |
|
0.05636627197265625, |
|
0.05633766555786133, |
|
0.05681817626953125, |
|
0.05622713470458984, |
|
0.05593119812011719, |
|
0.05513379287719727, |
|
0.057726593017578126, |
|
0.055895999908447264, |
|
0.05436476898193359, |
|
0.052256446838378906, |
|
0.051722496032714844, |
|
0.05475353622436523, |
|
0.05436435317993164, |
|
0.052705726623535155, |
|
0.05655100631713867, |
|
0.0550648307800293, |
|
0.05304572677612305, |
|
0.05447673416137695, |
|
0.053489505767822264, |
|
0.05376393508911133, |
|
0.05360259246826172, |
|
0.05315206527709961, |
|
0.05345993423461914, |
|
0.053002559661865234, |
|
0.053725791931152345, |
|
0.053625598907470706, |
|
0.05417843246459961, |
|
0.05624614334106445, |
|
0.05539561462402344, |
|
0.055847137451171876, |
|
0.05501161575317383, |
|
0.05341702270507812, |
|
0.054609630584716795, |
|
0.055186721801757814, |
|
0.05576544189453125, |
|
0.05407526397705078, |
|
0.05470758438110351, |
|
0.055508926391601564, |
|
0.05403196716308594, |
|
0.0524271354675293, |
|
0.052520736694335934, |
|
0.052684127807617186, |
|
0.05460601425170898, |
|
0.054717952728271485, |
|
0.054216064453125, |
|
0.05506659317016602, |
|
0.05477875137329102, |
|
0.054787647247314455, |
|
0.05401033782958985, |
|
0.054600383758544924, |
|
0.05515737533569336, |
|
0.054726718902587894, |
|
0.05671171188354492, |
|
0.056615550994873046, |
|
0.056598335266113284, |
|
0.055347328186035154, |
|
0.05720006561279297, |
|
0.05708249664306641, |
|
0.059014110565185546, |
|
0.06099126434326172, |
|
0.059362590789794924, |
|
0.057707679748535155, |
|
0.05730239868164062, |
|
0.05657676696777344, |
|
0.05653670501708984, |
|
0.057211135864257814, |
|
0.05927750396728516, |
|
0.05773324966430664, |
|
0.056075424194335935, |
|
0.059159454345703126, |
|
0.05695606231689453, |
|
0.056573982238769534, |
|
0.05758566284179688, |
|
0.05676591873168945, |
|
0.05843027114868164, |
|
0.05655945587158203, |
|
0.05572092819213867, |
|
0.05647289657592774, |
|
0.05808297729492187, |
|
0.05623801422119141, |
|
0.06520060729980469, |
|
0.06345920181274414, |
|
0.05951456069946289, |
|
0.06055311965942383, |
|
0.061469566345214846, |
|
0.062056705474853514, |
|
0.057702430725097655, |
|
0.05922835159301758, |
|
0.06045548629760742, |
|
0.06728828430175782, |
|
0.05621558380126953 |
|
], |
|
"count": 310, |
|
"total": 17.275037555694567, |
|
"mean": 0.05572592759901478, |
|
"p50": 0.05558217620849609, |
|
"p90": 0.056964593505859376, |
|
"p95": 0.05827398891448975, |
|
"p99": 0.06200386295318605, |
|
"stdev": 0.0016283729962815897, |
|
"stdev_": 2.9221101674588885 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 17.944968223690545 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
} |
|
} |
|
} |