|
{ |
|
"config": { |
|
"name": "pytorch-llama", |
|
"backend": { |
|
"name": "pytorch", |
|
"version": "2.5.1", |
|
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", |
|
"task": "text-generation", |
|
"library": "transformers", |
|
"model_type": "llama", |
|
"model": "meta-llama/Llama-2-7b-chat-hf", |
|
"processor": "meta-llama/Llama-2-7b-chat-hf", |
|
"device": "cuda", |
|
"device_ids": "0", |
|
"seed": 42, |
|
"inter_op_num_threads": null, |
|
"intra_op_num_threads": null, |
|
"model_kwargs": {}, |
|
"processor_kwargs": {}, |
|
"no_weights": true, |
|
"device_map": null, |
|
"torch_dtype": "bfloat16", |
|
"eval_mode": true, |
|
"to_bettertransformer": false, |
|
"low_cpu_mem_usage": null, |
|
"attn_implementation": null, |
|
"cache_implementation": null, |
|
"autocast_enabled": false, |
|
"autocast_dtype": null, |
|
"torch_compile": false, |
|
"torch_compile_target": "forward", |
|
"torch_compile_config": {}, |
|
"quantization_scheme": "torchao", |
|
"quantization_config": { |
|
"quant_type": "int4_weight_only", |
|
"group_size": 128 |
|
}, |
|
"deepspeed_inference": false, |
|
"deepspeed_inference_config": {}, |
|
"peft_type": null, |
|
"peft_config": {} |
|
}, |
|
"scenario": { |
|
"name": "inference", |
|
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", |
|
"iterations": 10, |
|
"duration": 10, |
|
"warmup_runs": 10, |
|
"input_shapes": { |
|
"batch_size": 1, |
|
"num_choices": 2, |
|
"sequence_length": 128 |
|
}, |
|
"new_tokens": null, |
|
"memory": true, |
|
"latency": true, |
|
"energy": false, |
|
"forward_kwargs": {}, |
|
"generate_kwargs": { |
|
"max_new_tokens": 32, |
|
"min_new_tokens": 32 |
|
}, |
|
"call_kwargs": {} |
|
}, |
|
"launcher": { |
|
"name": "process", |
|
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", |
|
"device_isolation": true, |
|
"device_isolation_action": "warn", |
|
"numactl": false, |
|
"numactl_kwargs": {}, |
|
"start_method": "spawn" |
|
}, |
|
"environment": { |
|
"cpu": " Intel(R) Xeon(R) Platinum 8339HC CPU @ 1.80GHz", |
|
"cpu_count": 22, |
|
"cpu_ram_mb": 189584.162816, |
|
"system": "Linux", |
|
"machine": "x86_64", |
|
"platform": "Linux-5.19.0-0_fbk12_hardened_11583_g0bef9520ca2b-x86_64-with-glibc2.34", |
|
"processor": "x86_64", |
|
"python_version": "3.10.15", |
|
"gpu": [ |
|
"NVIDIA PG509-210" |
|
], |
|
"gpu_count": 1, |
|
"gpu_vram_mb": 85899345920, |
|
"optimum_benchmark_version": "0.5.0.dev0", |
|
"optimum_benchmark_commit": "78d7da546ad060a586060d0d3680bec33f32a45b", |
|
"transformers_version": "4.47.0.dev0", |
|
"transformers_commit": null, |
|
"accelerate_version": "1.1.0", |
|
"accelerate_commit": null, |
|
"diffusers_version": "0.31.0", |
|
"diffusers_commit": null, |
|
"optimum_version": "1.24.0.dev0", |
|
"optimum_commit": null, |
|
"timm_version": "1.0.11", |
|
"timm_commit": null, |
|
"peft_version": "0.13.2", |
|
"peft_commit": null |
|
}, |
|
"print_report": true, |
|
"log_report": true |
|
}, |
|
"report": { |
|
"load": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 3085.971456, |
|
"max_global_vram": 15467.413504, |
|
"max_process_vram": 14554.23488, |
|
"max_reserved": 13478.395904, |
|
"max_allocated": 13476.848128 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
5.966353515625 |
|
], |
|
"count": 1, |
|
"total": 5.966353515625, |
|
"mean": 5.966353515625, |
|
"p50": 5.966353515625, |
|
"p90": 5.966353515625, |
|
"p95": 5.966353515625, |
|
"p99": 5.966353515625, |
|
"stdev": 0, |
|
"stdev_": 0 |
|
}, |
|
"throughput": null, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"prefill": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 4351.102976, |
|
"max_global_vram": 15922.495488, |
|
"max_process_vram": 15009.316864, |
|
"max_reserved": 13679.722496, |
|
"max_allocated": 13565.197824 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
0.0414266242980957, |
|
0.04142636871337891, |
|
0.043187839508056644, |
|
0.04308838272094727, |
|
0.04203615951538086, |
|
0.040791263580322264, |
|
0.04282470321655273, |
|
0.04198252868652344, |
|
0.042815521240234376, |
|
0.042053089141845704 |
|
], |
|
"count": 10, |
|
"total": 0.42163248062133785, |
|
"mean": 0.042163248062133785, |
|
"p50": 0.04204462432861328, |
|
"p90": 0.04309832839965821, |
|
"p95": 0.04314308395385742, |
|
"p99": 0.0431788883972168, |
|
"stdev": 0.0007614156637492785, |
|
"stdev_": 1.8058752556900264 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 3035.8192474017433 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"decode": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 4351.102976, |
|
"max_global_vram": 15922.495488, |
|
"max_process_vram": 15009.316864, |
|
"max_reserved": 13679.722496, |
|
"max_allocated": 13579.885056 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
1.2708951416015626, |
|
1.2522698974609374, |
|
1.25271875, |
|
1.278862548828125, |
|
1.2760985107421876, |
|
1.2531329345703126, |
|
1.2651630859375, |
|
1.2697864990234375, |
|
1.271283935546875, |
|
1.2574981689453124 |
|
], |
|
"count": 10, |
|
"total": 12.647709472656251, |
|
"mean": 1.264770947265625, |
|
"p50": 1.2674747924804688, |
|
"p90": 1.2763749145507812, |
|
"p95": 1.2776187316894532, |
|
"p99": 1.2786137854003905, |
|
"stdev": 0.009602028862915314, |
|
"stdev_": 0.7591911313012405 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 24.5103669300916 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"per_token": { |
|
"memory": null, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
0.0401976318359375, |
|
0.040500736236572264, |
|
0.03963894271850586, |
|
0.0408045425415039, |
|
0.04022505569458008, |
|
0.04094451141357422, |
|
0.04028652954101562, |
|
0.03987968063354492, |
|
0.03972876739501953, |
|
0.04102735900878906, |
|
0.03927689743041992, |
|
0.038796287536621094, |
|
0.03880176162719726, |
|
0.039893054962158205, |
|
0.042005599975585936, |
|
0.041323265075683596, |
|
0.0407031364440918, |
|
0.04084348678588867, |
|
0.04144950485229492, |
|
0.04140662384033203, |
|
0.0421802864074707, |
|
0.04241731262207031, |
|
0.04222982406616211, |
|
0.042391136169433595, |
|
0.04235331344604492, |
|
0.042299072265625, |
|
0.041650463104248046, |
|
0.04126144027709961, |
|
0.04216086578369141, |
|
0.04156169509887695, |
|
0.041647838592529296, |
|
0.0394257926940918, |
|
0.03950249481201172, |
|
0.04053414535522461, |
|
0.04100012969970703, |
|
0.040489376068115236, |
|
0.04085747146606445, |
|
0.041715774536132816, |
|
0.04107148742675781, |
|
0.041091678619384765, |
|
0.04071059036254883, |
|
0.04041411209106445, |
|
0.04057107162475586, |
|
0.04092681503295898, |
|
0.04048652648925781, |
|
0.040594783782958985, |
|
0.04086592102050781, |
|
0.04105219268798828, |
|
0.03997222518920898, |
|
0.04036175918579102, |
|
0.038977855682373046, |
|
0.040046142578125, |
|
0.04067155075073242, |
|
0.03990975952148437, |
|
0.04017107009887695, |
|
0.040495361328125, |
|
0.040779937744140626, |
|
0.04123270416259766, |
|
0.03907385635375977, |
|
0.03942886352539063, |
|
0.03925279998779297, |
|
0.039649185180664064, |
|
0.04047062301635742, |
|
0.04043833541870117, |
|
0.04056195068359375, |
|
0.03920675277709961, |
|
0.04018156814575195, |
|
0.03922252655029297, |
|
0.039454944610595705, |
|
0.03938899230957031, |
|
0.04092544174194336, |
|
0.041861312866210934, |
|
0.04115097427368164, |
|
0.03951171112060547, |
|
0.039311519622802736, |
|
0.03943756866455078, |
|
0.04027667236328125, |
|
0.0399692497253418, |
|
0.03905260848999023, |
|
0.039489822387695314, |
|
0.0403172492980957, |
|
0.041360416412353516, |
|
0.039731006622314456, |
|
0.0407435188293457, |
|
0.040927806854248044, |
|
0.04049945449829102, |
|
0.04109635162353516, |
|
0.04051635360717774, |
|
0.041217758178710935, |
|
0.041164703369140625, |
|
0.04132793426513672, |
|
0.04140889739990234, |
|
0.04156195068359375, |
|
0.04175334548950195, |
|
0.04056777572631836, |
|
0.04142803192138672, |
|
0.04089683151245117, |
|
0.040890113830566406, |
|
0.04128550338745117, |
|
0.04154127883911133, |
|
0.04208550262451172, |
|
0.04127667236328125, |
|
0.041409889221191404, |
|
0.04119113540649414, |
|
0.04024716949462891, |
|
0.04074355316162109, |
|
0.04013951873779297, |
|
0.04112214279174805, |
|
0.04130985641479492, |
|
0.042137409210205076, |
|
0.041546302795410155, |
|
0.04130144119262695, |
|
0.042115329742431644, |
|
0.04159372711181641, |
|
0.04161011123657227, |
|
0.040951713562011716, |
|
0.04055036926269531, |
|
0.04089923095703125, |
|
0.04099647903442383, |
|
0.04173875045776367, |
|
0.041740222930908205, |
|
0.04061587142944336, |
|
0.041406944274902345, |
|
0.04077532958984375, |
|
0.04054735946655273, |
|
0.04026124954223633, |
|
0.039889984130859375, |
|
0.039836769104003904, |
|
0.040271678924560544, |
|
0.04051667022705078, |
|
0.04025843048095703, |
|
0.041102657318115236, |
|
0.04170012664794922, |
|
0.04115087890625, |
|
0.03971779251098633, |
|
0.04065865707397461, |
|
0.04112899017333985, |
|
0.04167340850830078, |
|
0.04179584121704102, |
|
0.04180348968505859, |
|
0.04165439987182617, |
|
0.04130879974365234, |
|
0.041522430419921874, |
|
0.04147289657592773, |
|
0.04177433776855469, |
|
0.04116348648071289, |
|
0.043388256072998045, |
|
0.04378895950317383, |
|
0.04129987335205078, |
|
0.041887454986572266, |
|
0.04039408111572266, |
|
0.04118435287475586, |
|
0.0407973747253418, |
|
0.04055510330200195, |
|
0.040620670318603516, |
|
0.04092355346679687, |
|
0.041842784881591794, |
|
0.04060460662841797, |
|
0.04089462280273438, |
|
0.04106710433959961, |
|
0.03994182586669922, |
|
0.04090476989746094, |
|
0.03931772613525391, |
|
0.04022332763671875, |
|
0.04055683135986328, |
|
0.03997577667236328, |
|
0.039663646697998045, |
|
0.03931644821166992, |
|
0.039621631622314454, |
|
0.03991455841064453, |
|
0.04005836868286133, |
|
0.0400975341796875, |
|
0.040369407653808594, |
|
0.039888801574707033, |
|
0.04047494506835937, |
|
0.03972719955444336, |
|
0.04018918228149414, |
|
0.040294143676757814, |
|
0.040176033020019535, |
|
0.04035347366333008, |
|
0.041140640258789066, |
|
0.04213587188720703, |
|
0.04063116836547852, |
|
0.03985286331176758, |
|
0.04016323089599609, |
|
0.04180486297607422, |
|
0.04106159973144531, |
|
0.04055104064941406, |
|
0.04075049591064453, |
|
0.04144265747070312, |
|
0.042190944671630856, |
|
0.04093769454956055, |
|
0.04136153411865234, |
|
0.04203952026367187, |
|
0.04138134384155273, |
|
0.04055936050415039, |
|
0.04028268814086914, |
|
0.04082495880126953, |
|
0.040946720123291015, |
|
0.04052201461791992, |
|
0.03910755157470703, |
|
0.039964927673339846, |
|
0.03910614395141602, |
|
0.039076030731201174, |
|
0.0416732177734375, |
|
0.04074822235107422, |
|
0.041765151977539064, |
|
0.040593727111816406, |
|
0.0396712646484375, |
|
0.04136105728149414, |
|
0.040004928588867186, |
|
0.0393787841796875, |
|
0.03986854553222656, |
|
0.04215033721923828, |
|
0.042046432495117185, |
|
0.04165488052368164, |
|
0.041177631378173825, |
|
0.041282142639160156, |
|
0.04197859191894531, |
|
0.04058841705322266, |
|
0.04110255813598633, |
|
0.041939712524414065, |
|
0.041769920349121095, |
|
0.04141471862792969, |
|
0.04044083023071289, |
|
0.04058028793334961, |
|
0.039788063049316404, |
|
0.04175974273681641, |
|
0.04062607955932617, |
|
0.040643840789794924, |
|
0.041989952087402346, |
|
0.041531360626220704, |
|
0.04089692687988281, |
|
0.041573856353759764, |
|
0.040997920989990236, |
|
0.04107161712646484, |
|
0.04183990478515625, |
|
0.04095897674560547, |
|
0.04107731246948242, |
|
0.03973628616333008, |
|
0.04064271926879883, |
|
0.04028361511230469, |
|
0.04032780838012695, |
|
0.04030451202392578, |
|
0.03970716857910156, |
|
0.04026736068725586, |
|
0.040434719085693356, |
|
0.041146785736083984, |
|
0.04140016174316406, |
|
0.04091072082519531, |
|
0.04113446426391602, |
|
0.04292873764038086, |
|
0.04137481689453125, |
|
0.04076559829711914, |
|
0.04198960113525391, |
|
0.04107952117919922, |
|
0.041350048065185545, |
|
0.042510944366455077, |
|
0.041451454162597656, |
|
0.0411693115234375, |
|
0.04067270278930664, |
|
0.04105206298828125, |
|
0.040296638488769534, |
|
0.04047183990478516, |
|
0.04077145767211914, |
|
0.041627552032470705, |
|
0.04033446502685547, |
|
0.04019507217407227, |
|
0.04038185501098633, |
|
0.040734046936035155, |
|
0.03979292678833008, |
|
0.04035673522949219, |
|
0.039920993804931644, |
|
0.03981913757324219, |
|
0.039256862640380856, |
|
0.04180863952636719, |
|
0.04201779174804687, |
|
0.04097087860107422, |
|
0.041655937194824216, |
|
0.04192448043823242, |
|
0.04117097473144531, |
|
0.041486625671386716, |
|
0.041059616088867185, |
|
0.040226593017578124, |
|
0.03967728042602539, |
|
0.03971267318725586, |
|
0.040346847534179685, |
|
0.040519073486328126, |
|
0.04071187210083008, |
|
0.040181121826171874, |
|
0.039830718994140625, |
|
0.04020953750610352, |
|
0.04101318359375, |
|
0.041192031860351565, |
|
0.04000604629516601, |
|
0.03978236770629883, |
|
0.03962108612060547, |
|
0.03952025604248047, |
|
0.039889152526855466, |
|
0.04150806427001953, |
|
0.04086486434936523, |
|
0.039819713592529296, |
|
0.04043600082397461, |
|
0.03956934356689453, |
|
0.04144131088256836, |
|
0.04124975967407227, |
|
0.04144188690185547, |
|
0.04072371292114258, |
|
0.04142940902709961, |
|
0.040022014617919925 |
|
], |
|
"count": 310, |
|
"total": 12.637822704315191, |
|
"mean": 0.040767170013919954, |
|
"p50": 0.04076852798461914, |
|
"p90": 0.04180524063110352, |
|
"p95": 0.042101907539367676, |
|
"p99": 0.04250251750946045, |
|
"stdev": 0.0008450281427833556, |
|
"stdev_": 2.0728153131424643 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 24.529541777330873 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
} |
|
} |
|
} |