- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi, I am performing classification using 2 models on a single image on the edge CPU, GPU and HETERO:FPGA,CPU aiming to compare the inference speed between this 3 devices. I have calculated the inference timing with reference with the sample codes given.
Since I am evaluating 2 models at one time, I obtained an unexpected result where the inference timing for on FPGA is longest for model1 while lowest for model2.
Initially I suspected it may be the model's problem that affects the inference timing, but when I switched to evaluate model2 first then only model1, the results shows the same where the first evaluated model has a significantly higher inference timing on FPGA where it is expected to be the lowest among the 3 devices. Is there any cause or explanation for this scenario?
*For reference I am putting the coding part of inference and also the inference result at below.
Inference part code:
inf_start_1 = time.time()
res = exec_net.infer(inputs={input_blob: images})
inf_time_1 = (time.time() - inf_start_1)*1000
log.info("Inference time Model 1 (ms): {:.3f}".format(inf_time_1))
inf_start_2 = time.time()
res2 = exec_net2.infer(inputs={input_blob2: images})
inf_time_2 = (time.time() - inf_start_2)*1000
log.info("Inference time Model 2 (ms): {:.3f}".format(inf_time_2))
Inference result:
Expected first graph should be having the same trend as the second graph.
Link Copied

- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page