Intel® Distribution of OpenVINO™ Toolkit
Community assistance about the Intel® Distribution of OpenVINO™ toolkit, OpenCV, and all aspects of computer vision-related on Intel® platforms.

inference engine on remote CPU-device

Bodensteiner__Stpeha
494 Views

Hello,

Is it possible to run the inference engine on a different process or remote CPU-machine? E.g. pass images and collect inference results from a client-process/machine?

Thanks a lot in advance

Stephan  

0 Kudos
3 Replies
Shubha_R_Intel
Employee
494 Views

Dearest Stephan:

Current architecture implies application execution on a CPU host and then forwarding of the execution to a chosen device only (using the corresponding plugin). So, it’s not possible to delegate CPU execution to some other process or host at runtime (at least until you have written your own plugin). When you start your application it executes on CPU: a network is read, then you load read network to the chosen plugin, create infer requests here and then ask to start inference.

Of course you can build your own application and just put it on a remote machine together with the required libraries (inference engine libs and plugins).  And then you can start that app remotely from your local laptop using TCP/IP networking.  But I don't think this is quite what you were asking...

Hope it helps !

Shubha

0 Kudos
Bodensteiner__Stpeha
494 Views

Hello  Shubha,

Thank you very much for the Information! One more short question:

What if there is a process with several threads and each threads wants to make inferences using the same network. For memory efficiency it would be preferable if the network is only loaded/instantiated once. Is there some support (like synchronization, queuing, …?) so that multiple threads of the same process could use a single network/plugin?

 

thanks

Stephan

0 Kudos
Shubha_R_Intel
Employee
494 Views

Dear Stephan:

Please read this documentation here and see if it will address your needs (I think it will). Search for StartAsync.

https://docs.openvinotoolkit.org/latest/_docs_IE_DG_Integrate_with_customer_application_new_API.html

Also have a look at our Async samples under \inference_engine\samples. R5 has at least three different Async examples you can look at.

Hope it helps,

Shubha

 

0 Kudos
Reply