|
{ |
|
"config": { |
|
"name": "pytorch-llama", |
|
"backend": { |
|
"name": "pytorch", |
|
"version": "2.5.1", |
|
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", |
|
"task": "text-generation", |
|
"library": "transformers", |
|
"model_type": "llama", |
|
"model": "meta-llama/Llama-2-7b-chat-hf", |
|
"processor": "meta-llama/Llama-2-7b-chat-hf", |
|
"device": "cuda", |
|
"device_ids": "0", |
|
"seed": 42, |
|
"inter_op_num_threads": null, |
|
"intra_op_num_threads": null, |
|
"model_kwargs": {}, |
|
"processor_kwargs": {}, |
|
"no_weights": true, |
|
"device_map": null, |
|
"torch_dtype": "float16", |
|
"eval_mode": true, |
|
"to_bettertransformer": false, |
|
"low_cpu_mem_usage": null, |
|
"attn_implementation": null, |
|
"cache_implementation": null, |
|
"autocast_enabled": false, |
|
"autocast_dtype": null, |
|
"torch_compile": false, |
|
"torch_compile_target": "forward", |
|
"torch_compile_config": {}, |
|
"quantization_scheme": "awq", |
|
"quantization_config": { |
|
"bits": 4, |
|
"version": "gemm" |
|
}, |
|
"deepspeed_inference": false, |
|
"deepspeed_inference_config": {}, |
|
"peft_type": null, |
|
"peft_config": {} |
|
}, |
|
"scenario": { |
|
"name": "inference", |
|
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", |
|
"iterations": 10, |
|
"duration": 10, |
|
"warmup_runs": 10, |
|
"input_shapes": { |
|
"batch_size": 1, |
|
"num_choices": 2, |
|
"sequence_length": 128 |
|
}, |
|
"new_tokens": null, |
|
"memory": true, |
|
"latency": true, |
|
"energy": false, |
|
"forward_kwargs": {}, |
|
"generate_kwargs": { |
|
"max_new_tokens": 32, |
|
"min_new_tokens": 32 |
|
}, |
|
"call_kwargs": {} |
|
}, |
|
"launcher": { |
|
"name": "process", |
|
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", |
|
"device_isolation": true, |
|
"device_isolation_action": "warn", |
|
"numactl": false, |
|
"numactl_kwargs": {}, |
|
"start_method": "spawn" |
|
}, |
|
"environment": { |
|
"cpu": " Intel(R) Xeon(R) Platinum 8339HC CPU @ 1.80GHz", |
|
"cpu_count": 22, |
|
"cpu_ram_mb": 189584.162816, |
|
"system": "Linux", |
|
"machine": "x86_64", |
|
"platform": "Linux-5.19.0-0_fbk12_hardened_11583_g0bef9520ca2b-x86_64-with-glibc2.34", |
|
"processor": "x86_64", |
|
"python_version": "3.10.15", |
|
"gpu": [ |
|
"NVIDIA PG509-210" |
|
], |
|
"gpu_count": 1, |
|
"gpu_vram_mb": 85899345920, |
|
"optimum_benchmark_version": "0.5.0.dev0", |
|
"optimum_benchmark_commit": "78d7da546ad060a586060d0d3680bec33f32a45b", |
|
"transformers_version": "4.47.0.dev0", |
|
"transformers_commit": null, |
|
"accelerate_version": "1.1.0", |
|
"accelerate_commit": null, |
|
"diffusers_version": "0.31.0", |
|
"diffusers_commit": null, |
|
"optimum_version": "1.24.0.dev0", |
|
"optimum_commit": null, |
|
"timm_version": "1.0.11", |
|
"timm_commit": null, |
|
"peft_version": "0.13.2", |
|
"peft_commit": null |
|
}, |
|
"print_report": true, |
|
"log_report": true |
|
}, |
|
"report": { |
|
"load": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 6349.7216, |
|
"max_global_vram": 6200.098816, |
|
"max_process_vram": 5286.920192, |
|
"max_reserved": 4211.081216, |
|
"max_allocated": 3939.656192 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
6.58368017578125 |
|
], |
|
"count": 1, |
|
"total": 6.58368017578125, |
|
"mean": 6.58368017578125, |
|
"p50": 6.58368017578125, |
|
"p90": 6.58368017578125, |
|
"p95": 6.58368017578125, |
|
"p99": 6.58368017578125, |
|
"stdev": 0, |
|
"stdev_": 0 |
|
}, |
|
"throughput": null, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"prefill": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 4591.88224, |
|
"max_global_vram": 6581.78048, |
|
"max_process_vram": 5668.601856, |
|
"max_reserved": 4330.61888, |
|
"max_allocated": 4048.256512 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
0.05557324981689453, |
|
0.055142337799072266, |
|
0.05312239837646485, |
|
0.052617057800292966, |
|
0.055080799102783205, |
|
0.055220096588134766, |
|
0.05571369552612305, |
|
0.055056800842285154, |
|
0.05303116989135742, |
|
0.055473793029785154 |
|
], |
|
"count": 10, |
|
"total": 0.5460313987731933, |
|
"mean": 0.05460313987731934, |
|
"p50": 0.05511156845092774, |
|
"p90": 0.05558729438781738, |
|
"p95": 0.05565049495697021, |
|
"p99": 0.055701055412292484, |
|
"stdev": 0.0011245455833491904, |
|
"stdev_": 2.0594888606695236 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 2344.18753733918 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"decode": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 4591.88224, |
|
"max_global_vram": 6581.78048, |
|
"max_process_vram": 5668.601856, |
|
"max_reserved": 4330.61888, |
|
"max_allocated": 4048.258048 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
1.723550048828125, |
|
1.7150679931640624, |
|
1.6791304931640625, |
|
1.7016610107421875, |
|
1.7005169677734375, |
|
1.7107308349609376, |
|
1.7058489990234376, |
|
1.6905662841796876, |
|
1.6915482177734376, |
|
1.7029310302734375 |
|
], |
|
"count": 10, |
|
"total": 17.021551879882814, |
|
"mean": 1.7021551879882815, |
|
"p50": 1.7022960205078124, |
|
"p90": 1.7159161987304687, |
|
"p95": 1.7197331237792968, |
|
"p99": 1.7227866638183593, |
|
"stdev": 0.01221533281197518, |
|
"stdev_": 0.7176391963656417 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 18.21220545503717 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"per_token": { |
|
"memory": null, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
0.05595344161987305, |
|
0.05630144119262695, |
|
0.05638419342041016, |
|
0.056025310516357424, |
|
0.05608524703979492, |
|
0.05604048156738281, |
|
0.056223297119140626, |
|
0.055261409759521485, |
|
0.05402851104736328, |
|
0.05467363357543945, |
|
0.05407878494262695, |
|
0.053210239410400394, |
|
0.05310246276855469, |
|
0.0552031364440918, |
|
0.05591427230834961, |
|
0.05553513717651367, |
|
0.05638947296142578, |
|
0.05665884780883789, |
|
0.056774559020996096, |
|
0.05695423889160156, |
|
0.05661772918701172, |
|
0.056368576049804685, |
|
0.056631969451904296, |
|
0.05589382553100586, |
|
0.05577628707885742, |
|
0.05579734420776367, |
|
0.055099422454833985, |
|
0.05440447998046875, |
|
0.05558729553222656, |
|
0.05425529479980469, |
|
0.05518668746948242, |
|
0.056249504089355466, |
|
0.05475804901123047, |
|
0.05513875198364258, |
|
0.056074943542480465, |
|
0.05527878570556641, |
|
0.05644416046142578, |
|
0.055677791595458985, |
|
0.05474860763549805, |
|
0.05430595016479492, |
|
0.05509481430053711, |
|
0.055217918395996095, |
|
0.05657657623291015, |
|
0.055500606536865234, |
|
0.05603456115722656, |
|
0.05509852981567383, |
|
0.05502892684936524, |
|
0.055677345275878906, |
|
0.054931774139404296, |
|
0.055991840362548825, |
|
0.054562015533447264, |
|
0.05484684753417969, |
|
0.05457289505004883, |
|
0.055000160217285154, |
|
0.05583244705200195, |
|
0.05615126419067383, |
|
0.056124702453613284, |
|
0.05578102493286133, |
|
0.055998847961425784, |
|
0.05513017654418945, |
|
0.054037246704101566, |
|
0.05224281692504883, |
|
0.0551421127319336, |
|
0.05612515258789062, |
|
0.05596931076049805, |
|
0.05493721771240234, |
|
0.054672927856445314, |
|
0.05311308670043945, |
|
0.052114303588867185, |
|
0.05333135986328125, |
|
0.052935009002685544, |
|
0.05347971343994141, |
|
0.05590726470947266, |
|
0.05462992095947266, |
|
0.05473334503173828, |
|
0.05381894302368164, |
|
0.05392294311523437, |
|
0.05412188720703125, |
|
0.05325907135009766, |
|
0.053782337188720705, |
|
0.05433737564086914, |
|
0.05394620895385742, |
|
0.05448883056640625, |
|
0.05434579086303711, |
|
0.0543004150390625, |
|
0.05484207916259766, |
|
0.05356006240844727, |
|
0.053646846771240236, |
|
0.05356854248046875, |
|
0.053737312316894534, |
|
0.053435359954833984, |
|
0.05438911819458008, |
|
0.0535404167175293, |
|
0.053569183349609376, |
|
0.05353859329223633, |
|
0.05488636779785156, |
|
0.05522751998901367, |
|
0.055496574401855465, |
|
0.054357566833496095, |
|
0.05569007873535156, |
|
0.05462636947631836, |
|
0.054997089385986325, |
|
0.05580419158935547, |
|
0.054647903442382816, |
|
0.054636703491210935, |
|
0.05408620834350586, |
|
0.05409894561767578, |
|
0.05452304077148438, |
|
0.054459102630615236, |
|
0.05547500610351563, |
|
0.054643009185791014, |
|
0.05426655960083008, |
|
0.05395276641845703, |
|
0.05488409423828125, |
|
0.053593921661376956, |
|
0.055316608428955076, |
|
0.05464921569824219, |
|
0.05564089584350586, |
|
0.05609260940551758, |
|
0.05472192001342773, |
|
0.05603631973266601, |
|
0.055739646911621095, |
|
0.055740222930908204, |
|
0.05520828628540039, |
|
0.055530689239501954, |
|
0.0557729606628418, |
|
0.054949825286865234, |
|
0.055590305328369144, |
|
0.055293216705322265, |
|
0.055336097717285156, |
|
0.054789054870605466, |
|
0.055515102386474606, |
|
0.056732288360595705, |
|
0.054702144622802734, |
|
0.05589670562744141, |
|
0.05359260940551758, |
|
0.05416019058227539, |
|
0.05523471832275391, |
|
0.05573945617675781, |
|
0.054663551330566405, |
|
0.054610977172851564, |
|
0.053763294219970705, |
|
0.05361865615844726, |
|
0.052618721008300784, |
|
0.053403297424316404, |
|
0.05291356658935547, |
|
0.05414217758178711, |
|
0.05394787216186524, |
|
0.054357662200927734, |
|
0.05496761703491211, |
|
0.055395553588867184, |
|
0.0555340461730957, |
|
0.05574969482421875, |
|
0.0550912971496582, |
|
0.055879905700683595, |
|
0.05594278335571289, |
|
0.0550203857421875, |
|
0.054569793701171876, |
|
0.054833438873291014, |
|
0.05490607833862305, |
|
0.05536956787109375, |
|
0.05504108810424805, |
|
0.05481353759765625, |
|
0.054422622680664064, |
|
0.055524864196777345, |
|
0.05428435134887695, |
|
0.05532134246826172, |
|
0.054091358184814455, |
|
0.054521183013916015, |
|
0.05431523132324219, |
|
0.05521104049682617, |
|
0.054943199157714846, |
|
0.05521731185913086, |
|
0.05486444854736328, |
|
0.05471327972412109, |
|
0.05461270523071289, |
|
0.055639198303222656, |
|
0.055633281707763674, |
|
0.05520358276367188, |
|
0.055136703491210935, |
|
0.056127105712890625, |
|
0.05615753555297852, |
|
0.05627926254272461, |
|
0.056114017486572264, |
|
0.05580563354492187, |
|
0.055107166290283206, |
|
0.0564747200012207, |
|
0.05642041778564453, |
|
0.05618156814575195, |
|
0.05563382339477539, |
|
0.055429855346679685, |
|
0.05520320129394531, |
|
0.05494992065429687, |
|
0.05411705780029297, |
|
0.05400239944458008, |
|
0.05362188720703125, |
|
0.05409116744995117, |
|
0.054063968658447266, |
|
0.05398668670654297, |
|
0.05413904190063477, |
|
0.054786430358886716, |
|
0.056085536956787106, |
|
0.05582169723510742, |
|
0.05553190231323242, |
|
0.055452159881591793, |
|
0.05561862564086914, |
|
0.055433792114257814, |
|
0.05554512023925781, |
|
0.05649542236328125, |
|
0.05561177444458008, |
|
0.05471219253540039, |
|
0.05381017684936523, |
|
0.05379654312133789, |
|
0.05397766494750977, |
|
0.05456870269775391, |
|
0.05510851287841797, |
|
0.05408432006835937, |
|
0.05473548889160156, |
|
0.05548799896240234, |
|
0.0550302734375, |
|
0.05437615966796875, |
|
0.0535530891418457, |
|
0.05485023880004883, |
|
0.054630878448486325, |
|
0.05511056137084961, |
|
0.05665027236938477, |
|
0.055129695892333984, |
|
0.05487206268310547, |
|
0.05597065734863281, |
|
0.05607372665405273, |
|
0.054868000030517575, |
|
0.05577104187011719, |
|
0.05377679824829101, |
|
0.055522590637207034, |
|
0.05484198379516601, |
|
0.054145278930664065, |
|
0.054066944122314456, |
|
0.053343551635742184, |
|
0.05281884765625, |
|
0.05389769744873047, |
|
0.053199199676513674, |
|
0.05374358367919922, |
|
0.05295315170288086, |
|
0.05480575942993164, |
|
0.05520339202880859, |
|
0.05356464004516601, |
|
0.05247532653808594, |
|
0.05415043258666992, |
|
0.05406252670288086, |
|
0.05373948669433594, |
|
0.05339779281616211, |
|
0.05365603256225586, |
|
0.05324240112304687, |
|
0.052718017578125, |
|
0.05449702453613281, |
|
0.054787006378173825, |
|
0.05510543823242187, |
|
0.054537120819091796, |
|
0.05466003036499024, |
|
0.05448390579223633, |
|
0.05517379379272461, |
|
0.055890113830566406, |
|
0.05599193572998047, |
|
0.05495305633544922, |
|
0.055145729064941404, |
|
0.05483929443359375, |
|
0.05498681640625, |
|
0.054416351318359375, |
|
0.05498831939697266, |
|
0.05397849655151367, |
|
0.0533004150390625, |
|
0.05413475036621094, |
|
0.05480147171020508, |
|
0.05536652755737305, |
|
0.05512454223632812, |
|
0.05611062240600586, |
|
0.054006240844726563, |
|
0.05435820770263672, |
|
0.05408432006835937, |
|
0.054825023651123045, |
|
0.055166400909423825, |
|
0.054605377197265624, |
|
0.05434207916259766, |
|
0.05438684844970703, |
|
0.0544598388671875, |
|
0.05455369567871094, |
|
0.05488566589355469, |
|
0.05537984085083008, |
|
0.055301280975341795, |
|
0.05533103942871094, |
|
0.055293792724609374, |
|
0.055651775360107424, |
|
0.0558702392578125, |
|
0.05602703857421875, |
|
0.054728382110595705, |
|
0.05501891326904297, |
|
0.056121055603027346, |
|
0.056034912109375, |
|
0.054680641174316404, |
|
0.0540752944946289, |
|
0.05418393707275391, |
|
0.05439299011230469, |
|
0.05420982360839844, |
|
0.052660736083984375, |
|
0.05462825775146484, |
|
0.05379900741577148, |
|
0.05542063903808594, |
|
0.055786113739013675, |
|
0.054292289733886716, |
|
0.05584230422973633 |
|
], |
|
"count": 310, |
|
"total": 17.011372692108157, |
|
"mean": 0.05487539578099404, |
|
"p50": 0.05488601684570313, |
|
"p90": 0.05607597389221191, |
|
"p95": 0.056291460800170896, |
|
"p99": 0.05665807601928711, |
|
"stdev": 0.0009412214148377797, |
|
"stdev_": 1.7151974968784998 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 18.22310319165566 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
} |
|
} |
|
} |