<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic FP32 inference is on GPU. in Intel® Distribution of OpenVINO™ Toolkit</title>
    <link>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151350#M12494</link>
    <description>&lt;P&gt;FP32 inference is on GPU.&lt;/P&gt;&lt;P&gt;FP16 Inference device is Arria 10 on Mustang-F100-A10.&lt;/P&gt;&lt;P&gt;The topology is GoogleNet.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Thu, 14 Feb 2019 11:32:16 GMT</pubDate>
    <dc:creator>Kiyoshi_F_Intel</dc:creator>
    <dc:date>2019-02-14T11:32:16Z</dc:date>
    <item>
      <title>FP16 overflow</title>
      <link>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151348#M12492</link>
      <description>&lt;P&gt;Hello, Everyone&lt;/P&gt;&lt;P&gt;My customer is now developing their product using OpenVINO R4. He is using his own dataset.&lt;/P&gt;&lt;P&gt;During the training,&amp;nbsp;he trained by no scale value and no mean value on Caffe.&lt;/P&gt;&lt;P&gt;He converted the training model from FP32 to FP16 by using mo.py script.&lt;/P&gt;&lt;P&gt;However, he encountered a lot of inference errors when using FP16.&lt;/P&gt;&lt;P&gt;He used "cross-check tool" included in the OpenVINO package to compare the result between FP32 and FP16, and found overflow in some layers when using FP16 mode.&lt;/P&gt;&lt;P&gt;The overflow is not surprising because FP16 range is much smaller than FP32.&lt;/P&gt;&lt;P&gt;In this case, could you let me know what my customer has to do?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 12 Feb 2019 11:03:38 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151348#M12492</guid>
      <dc:creator>Kiyoshi_F_Intel</dc:creator>
      <dc:date>2019-02-12T11:03:38Z</dc:date>
    </item>
    <item>
      <title>&gt; errors when using FP16.</title>
      <link>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151349#M12493</link>
      <description>&lt;P&gt;&amp;gt;&amp;nbsp;errors when using FP16.&lt;/P&gt;&lt;P&gt;Assuming FP32 is on CPU.&amp;nbsp;May I ask what inference device is using for FP16 ? Also what topology?&lt;/P&gt;</description>
      <pubDate>Wed, 13 Feb 2019 17:29:06 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151349#M12493</guid>
      <dc:creator>nikos1</dc:creator>
      <dc:date>2019-02-13T17:29:06Z</dc:date>
    </item>
    <item>
      <title>FP32 inference is on GPU.</title>
      <link>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151350#M12494</link>
      <description>&lt;P&gt;FP32 inference is on GPU.&lt;/P&gt;&lt;P&gt;FP16 Inference device is Arria 10 on Mustang-F100-A10.&lt;/P&gt;&lt;P&gt;The topology is GoogleNet.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 14 Feb 2019 11:32:16 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151350#M12494</guid>
      <dc:creator>Kiyoshi_F_Intel</dc:creator>
      <dc:date>2019-02-14T11:32:16Z</dc:date>
    </item>
    <item>
      <title> </title>
      <link>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151351#M12495</link>
      <description>&lt;P style="margin-left:0in; margin-right:0in"&gt;When MO (Model Optimizer) converts weights of a model from FP32 to FP16 it checks for maximum value overflow (in fact MO uses numpy function &lt;A href="https://docs.scipy.org/doc/numpy/reference/generated/numpy.ndarray.astype.html"&gt;astype&lt;/A&gt;&amp;nbsp;which performs the values conversion).&lt;/P&gt;&lt;P style="margin-left:0in; margin-right:0in"&gt;If the value overflow occurs then the following error is printed (however the IR is generated):&lt;/P&gt;&lt;P style="margin-left:0in; margin-right:0in"&gt;[ ERROR ]&amp;nbsp; 83 elements of 189 were clipped to infinity while converting a blob for node [['conv2d_transpose']] to &amp;lt;class 'numpy.float16'&amp;gt;.&lt;/P&gt;&lt;P style="margin-left:0in; margin-right:0in"&gt;But the MO &lt;STRONG&gt;cannot&lt;/STRONG&gt; guarantee that the overflow will not occur &lt;STRONG&gt;during inference&lt;/STRONG&gt;. For example, you can create a network that will sum 2 values. Even though both of them are below float16 max value the sum of them will be more than the limit.&lt;/P&gt;&lt;P style="margin-left:0in; margin-right:0in"&gt;It is not possible to normalize weights values before converting because it will significantly decrease prediction results (or most probably completely break the topology) so there is no such feature in MO.&lt;/P&gt;&lt;P style="margin-left:0in; margin-right:0in"&gt;The recommendation to the customer would be to re-train the model with scaled input values to, for example, [0, 1] or [-1, 1] segment.&lt;/P&gt;</description>
      <pubDate>Thu, 14 Feb 2019 15:38:00 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151351#M12495</guid>
      <dc:creator>Shubha_R_Intel</dc:creator>
      <dc:date>2019-02-14T15:38:00Z</dc:date>
    </item>
    <item>
      <title>Hello, Shubha</title>
      <link>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151352#M12496</link>
      <description>&lt;P&gt;Hello, Shubha&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you so much for your answer. It is helpful and I reported it to my customer.&lt;/P&gt;</description>
      <pubDate>Mon, 18 Feb 2019 07:40:59 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151352#M12496</guid>
      <dc:creator>Kiyoshi_F_Intel</dc:creator>
      <dc:date>2019-02-18T07:40:59Z</dc:date>
    </item>
    <item>
      <title>Hello Shubba, please specify</title>
      <link>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151353#M12497</link>
      <description>&lt;P&gt;Hello Shubba, please specify how to "re-train the model with scaled input values".&lt;/P&gt;</description>
      <pubDate>Mon, 08 Apr 2019 21:12:20 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/FP16-overflow/m-p/1151353#M12497</guid>
      <dc:creator>Leini__Mikk</dc:creator>
      <dc:date>2019-04-08T21:12:20Z</dc:date>
    </item>
  </channel>
</rss>

