Hi my model is taking almost 10 x inference time on DSP and AIP as compared to GPU . What am I missing ? In order to run model on DSP are there any constraints or further optimization which is required. I have followed the entire setup guide provided on Qualcomm developers and all layers are compatible with DSP but still its taking too much time. Any idea or suggestions are welcome . Thanks!
Timing Numbers of AIP, DSP and GPU
Posted: Thu, 2020-11-05 10:21
Hi Amitshuklaacer,
The statement "The higher processing power of DSP will help my model perform better compared to GPU", is not always true.
We worked with Face Expression Recognition(FER) model built using Keras and converted to the DLC file,
Comparison of Total Inference Time for GPU and DSP draws to the conclusion that DSP performance is 60% to that of GPU. Before drawing that conclusion we also have an account of the time consumed for RPC Execute ( acts as a communicator between CPU/GPU and DSP), SNPE Accelerator and Accelerator. On considering these mentioned parameters, it looks GPU is performing better than DSP for a single/lesser number of predictions. We can choose DSP as a run time only if we required to make a higher number of predictions using the FER model.
Request you to check the beanchmarking application from SNPE with multiple itterations and compare the results.
Here you can find the instructions on the usage of Bencharking Tool.