performance comparison of LLM between AIX vs ppclinux (CPU based execution) #1301
Replies: 3 comments 1 reply
-
@kunal-vaishnavi @baijumeswani @yufenglee |
Beta Was this translation helpful? Give feedback.
-
@snnn @tianleiwu @yuslepukhin |
Beta Was this translation helpful? Give feedback.
-
You can use ONNX Runtime's LLM tools to profile. You can find more information at the end of this section. A JSON file containing profiling data for each op in the ONNX model should get generated. The generated file can then help determine which kernels are slow in AIX. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi
I am trying to compare LLM performance between AIX vs ppclinux using model_benchmark tool.
Attach sheet is having the results. All mentioned models are FP32 .
As per sheet, in ppclinux , avg (tokens/s) is 5-6 times better than AIX.
I am wondering , which API's or functions I can profile to find the root cause in AIX.
Llama-3.2-1B-Instruct model seems somewhat OKAY. I read from some google article that this model is having only 16 layers.
Could be because of less layers, faster processing while inferencing.
onnxruntime-genai-AIX-VS-Linux.xlsx
Beta Was this translation helpful? Give feedback.
All reactions