- 新着としてマーク
- ブックマーク
- 購読
- ミュート
- RSS フィードを購読する
- ハイライト
- 印刷
- 不適切なコンテンツを報告
I found expanation on Neural Network Compression Framework
https://docs.openvino.ai/latest/docs_nncf_introduction.html
I'm confusing the difference between pruning and sparcity. I know there is unstructure and structure pruning, is this "sparcity" in NNCF?
Is it possible to perfom structure pruning to make inference fast? I
checked these pages, but, I'm not sure which page gives corresponding information...
https://github.com/openvinotoolkit/nncf/tree/develop/examples/torch/classification/configs/sparsity
https://github.com/openvinotoolkit/nncf/blob/develop/docs/compression_algorithms/Sparsity.md
- 新着としてマーク
- ブックマーク
- 購読
- ミュート
- RSS フィードを購読する
- ハイライト
- 印刷
- 不適切なコンテンツを報告
Hi Timosy,
Thank you for reaching out to us.
Sparsity is an approach to compressing CNN. There are two types of sparsification methods:
- Structured sparsification (also known as pruning). As a result of structured sparsity, we get a new neural network that is smaller than the original network (fewer channels, filters, etc.).
- Unstructured sparsity. As a result of unstructured sparsity, we get a new network the same size as the original once, but weight tensors are sparse now. Using unstructured sparsification, we can remove more weights than via pruning.
The main idea of all sparsification algorithms is based on the fact that many modern DNNs are over-parameterized. It means that DNN contains more weights than it is needed to solve the problem (or more than we can effectively train). Thus the target of any sparsification algorithm is to find a subset of weights that contribute maximum result in accuracy and remove all other weights. So the contributions of sparsity algorithms are as follows:
- Minimization of the physical size of weights, using some sparse data representation methods.
- Improve the inference time using an implementation of sparse arithmetics (software or hardware).
The objective of NNCF is to prepare the model for accelerated inference by simulating the compression at train time. You can refer to Introducing a Training Add-on for OpenVINO™ toolkit: Neural Network Compression Framework in the Sparsity section for more detail.
Sincerely,
Zulkifli
コピーされたリンク
- 新着としてマーク
- ブックマーク
- 購読
- ミュート
- RSS フィードを購読する
- ハイライト
- 印刷
- 不適切なコンテンツを報告
Hi Timosy,
Thank you for reaching out to us.
Sparsity is an approach to compressing CNN. There are two types of sparsification methods:
- Structured sparsification (also known as pruning). As a result of structured sparsity, we get a new neural network that is smaller than the original network (fewer channels, filters, etc.).
- Unstructured sparsity. As a result of unstructured sparsity, we get a new network the same size as the original once, but weight tensors are sparse now. Using unstructured sparsification, we can remove more weights than via pruning.
The main idea of all sparsification algorithms is based on the fact that many modern DNNs are over-parameterized. It means that DNN contains more weights than it is needed to solve the problem (or more than we can effectively train). Thus the target of any sparsification algorithm is to find a subset of weights that contribute maximum result in accuracy and remove all other weights. So the contributions of sparsity algorithms are as follows:
- Minimization of the physical size of weights, using some sparse data representation methods.
- Improve the inference time using an implementation of sparse arithmetics (software or hardware).
The objective of NNCF is to prepare the model for accelerated inference by simulating the compression at train time. You can refer to Introducing a Training Add-on for OpenVINO™ toolkit: Neural Network Compression Framework in the Sparsity section for more detail.
Sincerely,
Zulkifli
- 新着としてマーク
- ブックマーク
- 購読
- ミュート
- RSS フィードを購読する
- ハイライト
- 印刷
- 不適切なコンテンツを報告
Dear Zulkifli_Intel
Thanks for your explantion.
I understand that,
Structured sparsification here is same with structure pruning that is mentioned in several Web page as one of the compression methods, and Unstructured sparsity is same with unstructure pruning.
- 新着としてマーク
- ブックマーク
- 購読
- ミュート
- RSS フィードを購読する
- ハイライト
- 印刷
- 不適切なコンテンツを報告
Hi Timosy,
This thread will no longer be monitored since this issue has been resolved. If you need any additional information from Intel, please submit a new question.
Sincerely,
Zulkifli