- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
How can we modify the input activations of the intermediate layers? For example, if we want to use input activations as int4 instead of FP32 or int8, how can we do it.
Link Copied
1 Reply
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi Sudi,
I think there isn't a easy way to perform what you are looking for, since the openvino converter (model optimizer) only provides reduced precisions of FP16 (half floats). If you run the model optimizer with -h flag you will find:
--data_type {FP16,FP32,half,float} Data type for all intermediate tensors and weights. If original model is in FP32 and --data_type=FP16 is specified, all model weights and biases are quantized to FP16.
Hope it helps,
César.

Reply
Topic Options
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page