|
{ |
|
"config": { |
|
"name": "pytorch-llama", |
|
"backend": { |
|
"name": "pytorch", |
|
"version": "2.5.1", |
|
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", |
|
"task": "text-generation", |
|
"library": "transformers", |
|
"model_type": "llama", |
|
"model": "meta-llama/Llama-2-7b-chat-hf", |
|
"processor": "meta-llama/Llama-2-7b-chat-hf", |
|
"device": "cuda", |
|
"device_ids": "0", |
|
"seed": 42, |
|
"inter_op_num_threads": null, |
|
"intra_op_num_threads": null, |
|
"model_kwargs": {}, |
|
"processor_kwargs": {}, |
|
"no_weights": true, |
|
"device_map": null, |
|
"torch_dtype": "float16", |
|
"eval_mode": true, |
|
"to_bettertransformer": false, |
|
"low_cpu_mem_usage": null, |
|
"attn_implementation": null, |
|
"cache_implementation": null, |
|
"autocast_enabled": false, |
|
"autocast_dtype": null, |
|
"torch_compile": false, |
|
"torch_compile_target": "forward", |
|
"torch_compile_config": {}, |
|
"quantization_scheme": "gptq", |
|
"quantization_config": { |
|
"bits": 4, |
|
"use_exllama ": true, |
|
"version": 2, |
|
"model_seqlen": 256 |
|
}, |
|
"deepspeed_inference": false, |
|
"deepspeed_inference_config": {}, |
|
"peft_type": null, |
|
"peft_config": {} |
|
}, |
|
"scenario": { |
|
"name": "inference", |
|
"_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", |
|
"iterations": 10, |
|
"duration": 10, |
|
"warmup_runs": 10, |
|
"input_shapes": { |
|
"batch_size": 1, |
|
"num_choices": 2, |
|
"sequence_length": 128 |
|
}, |
|
"new_tokens": null, |
|
"memory": true, |
|
"latency": true, |
|
"energy": false, |
|
"forward_kwargs": {}, |
|
"generate_kwargs": { |
|
"max_new_tokens": 32, |
|
"min_new_tokens": 32 |
|
}, |
|
"call_kwargs": {} |
|
}, |
|
"launcher": { |
|
"name": "process", |
|
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", |
|
"device_isolation": true, |
|
"device_isolation_action": "warn", |
|
"numactl": false, |
|
"numactl_kwargs": {}, |
|
"start_method": "spawn" |
|
}, |
|
"environment": { |
|
"cpu": " Intel(R) Xeon(R) Platinum 8339HC CPU @ 1.80GHz", |
|
"cpu_count": 22, |
|
"cpu_ram_mb": 189584.162816, |
|
"system": "Linux", |
|
"machine": "x86_64", |
|
"platform": "Linux-5.19.0-0_fbk12_hardened_11583_g0bef9520ca2b-x86_64-with-glibc2.34", |
|
"processor": "x86_64", |
|
"python_version": "3.10.15", |
|
"gpu": [ |
|
"NVIDIA PG509-210" |
|
], |
|
"gpu_count": 1, |
|
"gpu_vram_mb": 85899345920, |
|
"optimum_benchmark_version": "0.5.0.dev0", |
|
"optimum_benchmark_commit": "78d7da546ad060a586060d0d3680bec33f32a45b", |
|
"transformers_version": "4.47.0.dev0", |
|
"transformers_commit": null, |
|
"accelerate_version": "1.1.0", |
|
"accelerate_commit": null, |
|
"diffusers_version": "0.31.0", |
|
"diffusers_commit": null, |
|
"optimum_version": "1.24.0.dev0", |
|
"optimum_commit": null, |
|
"timm_version": "1.0.11", |
|
"timm_commit": null, |
|
"peft_version": "0.13.2", |
|
"peft_commit": null |
|
}, |
|
"print_report": true, |
|
"log_report": true |
|
}, |
|
"report": { |
|
"load": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 6375.329792, |
|
"max_global_vram": 72107.29472, |
|
"max_process_vram": 5381.292032, |
|
"max_reserved": 4303.355904, |
|
"max_allocated": 4034.388992 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
14.6606484375 |
|
], |
|
"count": 1, |
|
"total": 14.6606484375, |
|
"mean": 14.6606484375, |
|
"p50": 14.6606484375, |
|
"p90": 14.6606484375, |
|
"p95": 14.6606484375, |
|
"p99": 14.6606484375, |
|
"stdev": 0, |
|
"stdev_": 0 |
|
}, |
|
"throughput": null, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"prefill": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 6414.938112, |
|
"max_global_vram": 77369.049088, |
|
"max_process_vram": 5750.390784, |
|
"max_reserved": 4401.922048, |
|
"max_allocated": 4122.738688 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
0.04282547378540039, |
|
0.04379910278320313, |
|
0.0430645751953125, |
|
0.04397065734863281, |
|
0.04328860855102539, |
|
0.04283321762084961, |
|
0.04336835098266602, |
|
0.04369612884521484, |
|
0.04341609573364258, |
|
0.062448383331298825 |
|
], |
|
"count": 10, |
|
"total": 0.4527105941772461, |
|
"mean": 0.045271059417724614, |
|
"p50": 0.0433922233581543, |
|
"p90": 0.04581842994689941, |
|
"p95": 0.0541334066390991, |
|
"p99": 0.06078538799285889, |
|
"stdev": 0.005737441942605104, |
|
"stdev_": 12.673531426920329 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 2827.4133993401797 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"decode": { |
|
"memory": { |
|
"unit": "MB", |
|
"max_ram": 6414.938112, |
|
"max_global_vram": 84677.623808, |
|
"max_process_vram": 5750.390784, |
|
"max_reserved": 4401.922048, |
|
"max_allocated": 4137.581568 |
|
}, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
1.1446434326171875, |
|
1.1312457275390626, |
|
1.1492806396484374, |
|
1.1636923828125, |
|
1.1359462890625, |
|
1.12259521484375, |
|
1.118836181640625, |
|
1.1177918701171874, |
|
1.147646728515625, |
|
1.291679443359375 |
|
], |
|
"count": 10, |
|
"total": 11.523357910156248, |
|
"mean": 1.1523357910156249, |
|
"p50": 1.1402948608398438, |
|
"p90": 1.1764910888671876, |
|
"p95": 1.234085266113281, |
|
"p99": 1.2801606079101562, |
|
"stdev": 0.04852786208204437, |
|
"stdev_": 4.211260507605492 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 26.901880720616845 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
}, |
|
"per_token": { |
|
"memory": null, |
|
"latency": { |
|
"unit": "s", |
|
"values": [ |
|
0.03595731353759766, |
|
0.03624009704589844, |
|
0.03713497543334961, |
|
0.036713951110839846, |
|
0.03598102569580078, |
|
0.0356715202331543, |
|
0.03654713439941406, |
|
0.03701628875732422, |
|
0.0365937271118164, |
|
0.036367935180664064, |
|
0.03673350524902344, |
|
0.03692073440551758, |
|
0.03674390411376953, |
|
0.03694230270385742, |
|
0.03666019058227539, |
|
0.03697779083251953, |
|
0.036823486328125, |
|
0.03659756851196289, |
|
0.03698876953125, |
|
0.03725027084350586, |
|
0.03755724716186523, |
|
0.03737756729125977, |
|
0.0375269775390625, |
|
0.0378520622253418, |
|
0.03748668670654297, |
|
0.03696822357177734, |
|
0.037523681640625, |
|
0.037523006439208986, |
|
0.037192192077636715, |
|
0.03669868850708008, |
|
0.03700940704345703, |
|
0.03728604888916016, |
|
0.03772364807128906, |
|
0.036850334167480465, |
|
0.036908287048339844, |
|
0.03730115127563476, |
|
0.037047550201416014, |
|
0.03660595321655274, |
|
0.03694559860229492, |
|
0.03658291244506836, |
|
0.0366712646484375, |
|
0.0362498893737793, |
|
0.035942848205566404, |
|
0.035322017669677734, |
|
0.03507583999633789, |
|
0.03516364669799805, |
|
0.035625537872314456, |
|
0.035464031219482425, |
|
0.03500419235229492, |
|
0.03528540802001953, |
|
0.037079296112060546, |
|
0.03623779296875, |
|
0.0368109130859375, |
|
0.03728329467773438, |
|
0.036510688781738285, |
|
0.03571551895141602, |
|
0.03624691009521484, |
|
0.03654790496826172, |
|
0.03718291091918945, |
|
0.036795745849609374, |
|
0.03788579177856445, |
|
0.0369071044921875, |
|
0.03748160171508789, |
|
0.0377050895690918, |
|
0.03723452758789063, |
|
0.037270240783691407, |
|
0.03725990295410156, |
|
0.03632831954956055, |
|
0.03682374572753906, |
|
0.03615884780883789, |
|
0.03704444885253906, |
|
0.03667324829101563, |
|
0.0367913932800293, |
|
0.0373804817199707, |
|
0.03712713623046875, |
|
0.037198848724365234, |
|
0.03745049667358399, |
|
0.0372237434387207, |
|
0.037007198333740235, |
|
0.03718291091918945, |
|
0.03761852645874023, |
|
0.036934593200683596, |
|
0.03726793670654297, |
|
0.03669001770019531, |
|
0.037087841033935545, |
|
0.03685628890991211, |
|
0.03717059326171875, |
|
0.03747612762451172, |
|
0.03693795013427734, |
|
0.036179073333740236, |
|
0.03682166290283203, |
|
0.03697062301635742, |
|
0.03689651107788086, |
|
0.03706435012817383, |
|
0.037270465850830076, |
|
0.035971839904785155, |
|
0.036920799255371096, |
|
0.03681465530395508, |
|
0.0369854736328125, |
|
0.03704467010498047, |
|
0.03700086212158203, |
|
0.03643961715698242, |
|
0.0370840950012207, |
|
0.037169761657714843, |
|
0.03741856002807617, |
|
0.036993568420410156, |
|
0.03796236801147461, |
|
0.03702950286865234, |
|
0.03688911819458008, |
|
0.036421279907226566, |
|
0.0377540168762207, |
|
0.03744473648071289, |
|
0.04109030532836914, |
|
0.04965795135498047, |
|
0.03727587127685547, |
|
0.03747747039794922, |
|
0.03727571105957031, |
|
0.03713622283935547, |
|
0.036595935821533206, |
|
0.03686832046508789, |
|
0.036631904602050784, |
|
0.036168128967285156, |
|
0.03665676879882813, |
|
0.03609823989868164, |
|
0.03664368057250977, |
|
0.036477184295654295, |
|
0.0366478385925293, |
|
0.036935649871826175, |
|
0.036280990600585934, |
|
0.03671769714355469, |
|
0.036772960662841796, |
|
0.03645542526245117, |
|
0.03598780822753906, |
|
0.035988353729248045, |
|
0.035874847412109376, |
|
0.03614518356323242, |
|
0.036284576416015624, |
|
0.03758396911621094, |
|
0.03668572616577148, |
|
0.03693961715698242, |
|
0.03611155319213867, |
|
0.03728643035888672, |
|
0.037300159454345706, |
|
0.03694675064086914, |
|
0.03692620849609375, |
|
0.03698134231567383, |
|
0.036760032653808596, |
|
0.03694249725341797, |
|
0.036695552825927735, |
|
0.03729305648803711, |
|
0.037698944091796874, |
|
0.036593311309814455, |
|
0.03610140609741211, |
|
0.035955230712890626, |
|
0.035017822265625, |
|
0.036146846771240235, |
|
0.03579935836791992, |
|
0.03700652694702149, |
|
0.03609513473510742, |
|
0.0357050895690918, |
|
0.036111358642578126, |
|
0.035346206665039064, |
|
0.03659628677368164, |
|
0.03786441421508789, |
|
0.03727798461914063, |
|
0.03694675064086914, |
|
0.03681289672851563, |
|
0.03643088150024414, |
|
0.03525331115722656, |
|
0.036265247344970705, |
|
0.03562035369873047, |
|
0.03675743865966797, |
|
0.03684204864501953, |
|
0.036423904418945316, |
|
0.036361759185791015, |
|
0.035707073211669924, |
|
0.0352196159362793, |
|
0.03612031936645508, |
|
0.03643833541870117, |
|
0.03616476821899414, |
|
0.03556515121459961, |
|
0.03600313568115234, |
|
0.03593075180053711, |
|
0.03558707046508789, |
|
0.03574892807006836, |
|
0.03536566543579102, |
|
0.035780353546142576, |
|
0.03572681427001953, |
|
0.03538473510742188, |
|
0.035416030883789064, |
|
0.037334465026855466, |
|
0.036534656524658204, |
|
0.03557791900634766, |
|
0.03659225463867188, |
|
0.03616390228271484, |
|
0.03587062454223633, |
|
0.03605587387084961, |
|
0.03600467300415039, |
|
0.036198238372802734, |
|
0.03658566284179687, |
|
0.03599353790283203, |
|
0.03611094284057617, |
|
0.035894046783447264, |
|
0.03659619140625, |
|
0.03686988830566406, |
|
0.03661088180541992, |
|
0.036827136993408206, |
|
0.03656438446044922, |
|
0.03548425674438477, |
|
0.03565628814697266, |
|
0.03530723190307617, |
|
0.035811328887939455, |
|
0.03646108627319336, |
|
0.03669027328491211, |
|
0.0351959342956543, |
|
0.035250526428222656, |
|
0.03528268814086914, |
|
0.036346240997314455, |
|
0.03538345718383789, |
|
0.0360588493347168, |
|
0.03625296020507812, |
|
0.03606617736816406, |
|
0.03607494354248047, |
|
0.036089889526367186, |
|
0.03624774551391602, |
|
0.03634748840332031, |
|
0.036305374145507815, |
|
0.03629427337646484, |
|
0.036154815673828125, |
|
0.03588390350341797, |
|
0.03697488021850586, |
|
0.036380062103271486, |
|
0.03564672088623047, |
|
0.03618310546875, |
|
0.03602627182006836, |
|
0.03591683197021484, |
|
0.03659167861938477, |
|
0.03637334442138672, |
|
0.03517212677001953, |
|
0.035636062622070315, |
|
0.03653926467895508, |
|
0.035418529510498044, |
|
0.03555548858642578, |
|
0.035576831817626955, |
|
0.03670748901367187, |
|
0.036157184600830075, |
|
0.03516700744628906, |
|
0.035265632629394535, |
|
0.035632511138916016, |
|
0.036483806610107423, |
|
0.03661891174316406, |
|
0.03655871963500976, |
|
0.03703827285766602, |
|
0.037617919921875, |
|
0.03706230545043945, |
|
0.03730966567993164, |
|
0.03732236862182617, |
|
0.03743148803710938, |
|
0.03719247817993164, |
|
0.03726732635498047, |
|
0.037321281433105466, |
|
0.037644287109375, |
|
0.03730979156494141, |
|
0.03729123306274414, |
|
0.03712403106689453, |
|
0.03692931365966797, |
|
0.037281822204589844, |
|
0.03759795379638672, |
|
0.03726166534423828, |
|
0.03703315353393555, |
|
0.03741417694091797, |
|
0.03737984085083008, |
|
0.03714316940307617, |
|
0.03710857772827148, |
|
0.036980159759521486, |
|
0.03651663970947266, |
|
0.03553305435180664, |
|
0.03653798294067383, |
|
0.03568384170532227, |
|
0.04555145645141601, |
|
0.042402015686035154, |
|
0.040898239135742184, |
|
0.0382880973815918, |
|
0.04074873733520508, |
|
0.04209692764282227, |
|
0.04165305709838867, |
|
0.03827654266357422, |
|
0.04032793426513672, |
|
0.040399806976318356, |
|
0.046360767364501954, |
|
0.041255329132080076, |
|
0.046764129638671874, |
|
0.03941721725463867, |
|
0.04514297485351562, |
|
0.03980563354492188, |
|
0.044795169830322265, |
|
0.040840190887451173, |
|
0.04167273712158203, |
|
0.0367586555480957, |
|
0.041850208282470706, |
|
0.03750924682617188, |
|
0.044384449005126954, |
|
0.039292991638183596, |
|
0.04272025680541992, |
|
0.03939622497558594, |
|
0.04575481414794922, |
|
0.038351905822753905, |
|
0.04231804656982422, |
|
0.041536479949951174, |
|
0.04459820938110352 |
|
], |
|
"count": 310, |
|
"total": 11.513670848846436, |
|
"mean": 0.03714087370595624, |
|
"p50": 0.03678217697143554, |
|
"p90": 0.037993785476684576, |
|
"p95": 0.0416005973815918, |
|
"p99": 0.04573651195526124, |
|
"stdev": 0.001997879281315144, |
|
"stdev_": 5.379193007499838 |
|
}, |
|
"throughput": { |
|
"unit": "tokens/s", |
|
"value": 26.924514698199765 |
|
}, |
|
"energy": null, |
|
"efficiency": null |
|
} |
|
} |
|
} |