Intel® Distribution of OpenVINO™ Toolkit
Community assistance about the Intel® Distribution of OpenVINO™ toolkit, OpenCV, and all aspects of computer vision-related on Intel® platforms.

Is parallelism possible in OpenVINO?

jic5760
New Contributor I
214 Views

Hello,

 

I want to run large LLM models using multiple Intel GPUs.

Is there a way to split a large model and run it on multiple GPUs?

 

Ref: https://huggingface.co/docs/transformers/v4.13.0/parallelism

 

Thank you

0 Kudos
1 Reply
Zulkifli_Intel
Moderator
163 Views

Hi jic5760,

Thank you for reaching out.

 

To distribute inference across multiple GPUs, you can use the heterogeneous plugin in OpenVINO, which allows you to simultaneously leverage multiple inference devices (e.g., CPU, GPU, NPU) in one model. You can refer to pipeline-parallelism for multiple devices' execution in OpenVINO.

 

 

Regards,

Zul


0 Kudos
Reply