Intel® Distribution of OpenVINO™ Toolkit
Community assistance about the Intel® Distribution of OpenVINO™ toolkit, OpenCV, and all aspects of computer vision-related on Intel® platforms.
6572 ディスカッション

unsupported layer type "FakeQuantize"

Simardeep
ビギナー
1,107件の閲覧回数

Hi, I tried compressing a model using POT, with default quantization and got the error above. I understand that Int8 is not supported by VPU devices yet. Also, I have tried different methods such as converting the model to FP16 and reducing the input shape.

 

Is there any other alternative to compress the model?

 

 

Regards,

Simardeep Singh Sethi

0 件の賞賛
2 返答(返信)
IntelSupport
管理者
1,071件の閲覧回数

Hi Simardeep,

 

Thanks for reaching out.

 

Firstly, the "unsupported layer type "FakeQuantize" is obviously because VPU plugin doesn't support INT8 model format. The relevant information is available in Supported Model Format documentation.

 

Meanwhile, In OpenVINO, there are 2 ways to enhance the performance:

  1. During development: Post-training Optimization tool (POT), Neural Network Compression Framework (NNCF), Model Optimizer
  2. During deployment: tuning inference parameters and optimizing model execution

 

It's also possible to combine both approaches.

 

Since you have tried with POT, another alternative way is to retrain the model with NNCF. Refer to Introducing a Training Add-on for OpenVINO toolkit: Neural Network Compression Framework, for the steps to implement the NNCF optimization methods using supported training samples and through integration into the custom training code.

 

The training samples are available at this GitHub repository.

 

 

Regards,

Aznie


IntelSupport
管理者
999件の閲覧回数

Hi Simardeep,


This thread will no longer be monitored since we have provided information. If you need any additional information from Intel, please submit a new question.



Regards,

Aznie


返信