<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Hi Simon, in Intel® MPI Library</title>
    <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027375#M4120</link>
    <description>&lt;P&gt;Hi Simon,&lt;/P&gt;

&lt;P&gt;I just contacted an expert. Could you describe in details how to reproduce the issue, please? Thank you.&lt;/P&gt;</description>
    <pubDate>Fri, 19 Dec 2014 20:52:51 GMT</pubDate>
    <dc:creator>Loc_N_Intel</dc:creator>
    <dc:date>2014-12-19T20:52:51Z</dc:date>
    <item>
      <title>Performance issue with multi-rail support in Intel MPI 5.0</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027373#M4118</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;

&lt;P&gt;I am experiencing a severy performance loss when using multiple rails in Intel MPI 5.0 and the KNC and an mlx5 adapter (which has 2 ports). With Intel MPI 4.1 it was much better.&lt;/P&gt;

&lt;P&gt;Let me give an example of the performance of our application (per KNC):&lt;/P&gt;

&lt;UL&gt;
	&lt;LI&gt;Intel MPI 4.1, single-rail (I_MPI_DAPL_PROVIDER_LIST=ofa-v2-mlx5_0-1u): &lt;STRONG&gt;220&lt;/STRONG&gt; Gflop/s&lt;/LI&gt;
	&lt;LI&gt;Intel MPI 4.1, dual-rail (-IB I_MPI_OFA_ADAPTER_NAME=mlx5_0&amp;nbsp; I_MPI_OFA_NUM_PORTS=2): &lt;STRONG&gt;270&lt;/STRONG&gt; Gflop/s&lt;/LI&gt;
	&lt;LI&gt;Intel MPI 5.0, single-rail (I_MPI_DAPL_PROVIDER_LIST=ofa-v2-mlx5_0-1u): &lt;STRONG&gt;220&lt;/STRONG&gt; Gflop/s&lt;/LI&gt;
	&lt;LI&gt;Intel MPI 5.0, dual-rail (-IB I_MPI_OFA_ADAPTER_NAME=mlx5_0&amp;nbsp; I_MPI_OFA_NUM_PORTS=2): &lt;STRONG&gt;150&lt;/STRONG&gt; Gflop/s&lt;/LI&gt;
	&lt;LI&gt;Intel MPI 5.0, single-rail (-IB I_MPI_OFA_ADAPTER_NAME=mlx5_0&amp;nbsp; I_MPI_OFA_NUM_PORTS=1): &lt;STRONG&gt;150&lt;/STRONG&gt; Gflop/s&lt;/LI&gt;
&lt;/UL&gt;

&lt;P&gt;With DAPL the performance is unchanged, but apparently there is no way to use it with dual-rail support. With OFA I got the best performance in v4.1, but with v5.0 it is extremely low. In particular it is the same for 1 or 2 ports.&lt;/P&gt;

&lt;P&gt;Is there anything I am overlooking in the documentation?&lt;/P&gt;

&lt;P&gt;Thanks,&lt;/P&gt;

&lt;P&gt;Simon&lt;/P&gt;</description>
      <pubDate>Wed, 17 Dec 2014 09:42:36 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027373#M4118</guid>
      <dc:creator>Simon_H_2</dc:creator>
      <dc:date>2014-12-17T09:42:36Z</dc:date>
    </item>
    <item>
      <title>The HCA card in my system has</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027374#M4119</link>
      <description>&lt;P&gt;The&amp;nbsp;HCA card in my system has one port only, I cannot reproduce the dual-rail issue that you saw. But let me ask&amp;nbsp;experts around here and get back to you. Thank you.&lt;/P&gt;</description>
      <pubDate>Fri, 19 Dec 2014 18:39:58 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027374#M4119</guid>
      <dc:creator>Loc_N_Intel</dc:creator>
      <dc:date>2014-12-19T18:39:58Z</dc:date>
    </item>
    <item>
      <title>Hi Simon,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027375#M4120</link>
      <description>&lt;P&gt;Hi Simon,&lt;/P&gt;

&lt;P&gt;I just contacted an expert. Could you describe in details how to reproduce the issue, please? Thank you.&lt;/P&gt;</description>
      <pubDate>Fri, 19 Dec 2014 20:52:51 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027375#M4120</guid>
      <dc:creator>Loc_N_Intel</dc:creator>
      <dc:date>2014-12-19T20:52:51Z</dc:date>
    </item>
    <item>
      <title>Hi,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027376#M4121</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;

&lt;P&gt;thanks for your reply. To reproduce, you can use, for example, the OSU bandwidth benchmark: &lt;A href="http://mvapich.cse.ohio-state.edu/benchmarks/" target="_blank"&gt;http://mvapich.cse.ohio-state.edu/benchmarks/&lt;/A&gt;. My original tests were done on the KNC, but the same problem shows up on the Xeon (Haswell) host.&lt;/P&gt;

&lt;P&gt;You can see the result in the attached figure. You can see that for message sizes around 100 kB and above Intel MPI 4.1 with "dual rail" is by far the best (blue solid squares). Intel MPI 5.0 is much much worse.&lt;/P&gt;

&lt;P&gt;&lt;A href="https://community.intel.com/legacyfs/online/drupal_files/444821"&gt;444821&lt;/A&gt;&lt;/P&gt;

&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 22 Dec 2014 07:40:52 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027376#M4121</guid>
      <dc:creator>Simon_H_2</dc:creator>
      <dc:date>2014-12-22T07:40:52Z</dc:date>
    </item>
    <item>
      <title>Are there any new</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027377#M4122</link>
      <description>Are there any new developments on this? Can you confirm the issue now?

Thanks,
Simon</description>
      <pubDate>Mon, 19 Jan 2015 07:03:00 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027377#M4122</guid>
      <dc:creator>Simon_H_2</dc:creator>
      <dc:date>2015-01-19T07:03:00Z</dc:date>
    </item>
    <item>
      <title>Hi Simon,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027378#M4123</link>
      <description>&lt;P&gt;Hi Simon,&lt;/P&gt;

&lt;P&gt;Sorry for the delayed answer. This issue was forwarded to the development team for investigation. I will let you know when I have an update.&lt;/P&gt;

&lt;P&gt;Thank you.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 21 Jan 2015 19:41:32 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027378#M4123</guid>
      <dc:creator>Loc_N_Intel</dc:creator>
      <dc:date>2015-01-21T19:41:32Z</dc:date>
    </item>
    <item>
      <title>Hi,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027379#M4124</link>
      <description>Hi,
Are there any updates regarding this issue?

Thanks,
Simon</description>
      <pubDate>Wed, 04 Mar 2015 10:17:25 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027379#M4124</guid>
      <dc:creator>Simon_H_2</dc:creator>
      <dc:date>2015-03-04T10:17:25Z</dc:date>
    </item>
    <item>
      <title>Hi Simon,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027380#M4125</link>
      <description>&lt;P&gt;Hi Simon,&lt;/P&gt;

&lt;P&gt;Could you please specify the exact versions of Intel MPI Library (4.x, 5.x) and OS/MPSS/OFED/DAPL.&lt;BR /&gt;
	Also could you please provide test scenarios you used. Which compute nodes were involved in each run (MPI ranks only on HOST, or only on KNC, or both on HOST and KNC).&lt;/P&gt;

&lt;P&gt;Regarding to DAPL - try to run the same scenarios with default DAPL provider (without I_MPI_DAPL_PROVIDER_LIST).&lt;/P&gt;</description>
      <pubDate>Tue, 10 Mar 2015 09:28:55 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027380#M4125</guid>
      <dc:creator>Artem_R_Intel1</dc:creator>
      <dc:date>2015-03-10T09:28:55Z</dc:date>
    </item>
    <item>
      <title>Hi Artem,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027381#M4126</link>
      <description>&lt;P&gt;Hi Artem,&lt;/P&gt;

&lt;P&gt;I used two scenarios, the issue shows up in both cases:&lt;/P&gt;

&lt;OL&gt;
	&lt;LI&gt;HOST &amp;lt;-&amp;gt; HOST&lt;/LI&gt;
	&lt;LI&gt;KNC &amp;lt;-&amp;gt; KNC&lt;/LI&gt;
&lt;/OL&gt;

&lt;P&gt;Versions:&lt;/P&gt;

&lt;UL&gt;
	&lt;LI&gt;Intel MPI 4.1.3.045 and 5.0.2.044&lt;/LI&gt;
	&lt;LI&gt;OS is Linux (CentOS)&lt;/LI&gt;
	&lt;LI&gt;OFED 3.5.2&lt;/LI&gt;
	&lt;LI&gt;DAPL 2.1.2&lt;/LI&gt;
	&lt;LI&gt;MPSS 3.3.3 (I guess this is irrelevant, since the issue shows up also if only HOSTS are involved)&lt;/LI&gt;
&lt;/UL&gt;

&lt;P&gt;I think I had tried in the past to run without I_MPI_DAPL_PROVIDER_LIST, but Intel MPI tried to default to an mlx4 device (which does not exist on our system), and would not use the mlx5 device, so using I_MPI_DAPL_PROVIDER_LIST was mandatory. I will try again.&lt;/P&gt;

&lt;P&gt;Remark: Should this topic be moved to the general forum, since by now we know that it is not MIC-specific?&lt;/P&gt;

&lt;P&gt;Simon&lt;/P&gt;</description>
      <pubDate>Tue, 17 Mar 2015 08:18:00 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027381#M4126</guid>
      <dc:creator>Simon_H_2</dc:creator>
      <dc:date>2015-03-17T08:18:00Z</dc:date>
    </item>
    <item>
      <title>Hi Simon,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027382#M4127</link>
      <description>&lt;P&gt;Hi Simon,&lt;/P&gt;

&lt;P&gt;FYI, we submitted an internal bug report (&lt;SPAN style="color: rgb(31, 73, 125); font-family: &amp;quot;Calibri&amp;quot;,sans-serif; font-size: 11pt; mso-fareast-font-family: Calibri; mso-fareast-theme-font: minor-latin; mso-bidi-font-family: &amp;quot;Times New Roman&amp;quot;; mso-ansi-language: EN-US; mso-fareast-language: EN-US; mso-bidi-language: AR-SA;"&gt;DPD200368369) &lt;/SPAN&gt;a while ago but there is no update yet.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 27 Mar 2015 23:06:26 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027382#M4127</guid>
      <dc:creator>Loc_N_Intel</dc:creator>
      <dc:date>2015-03-27T23:06:26Z</dc:date>
    </item>
    <item>
      <title>Hi Simon,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027383#M4128</link>
      <description>&lt;P&gt;Hi Simon,&lt;/P&gt;

&lt;P&gt;Loc and I have been communicating internally about this since you initially submitted it.&lt;/P&gt;

&lt;P&gt;Just as an FYI, I'm moving this issue over to the regular Intel® Clusters and HPC Technology forum since it's not Phi-specific.&amp;nbsp; That way I can keep track of the internal bug I submitted and update you on current status.&lt;/P&gt;

&lt;P&gt;Thanks,&lt;BR /&gt;
	~Gergana&lt;/P&gt;</description>
      <pubDate>Mon, 30 Mar 2015 15:51:09 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027383#M4128</guid>
      <dc:creator>Gergana_S_Intel</dc:creator>
      <dc:date>2015-03-30T15:51:09Z</dc:date>
    </item>
    <item>
      <title>Hey Simon,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027384#M4129</link>
      <description>&lt;P&gt;Hey Simon,&lt;/P&gt;

&lt;P&gt;We've made several fixes to the Intel MPI Library in regards to multi-rail support.&amp;nbsp; Just wondering if you've tried the latest Intel MPI 5.1.2 with any better success regarding performance?&lt;/P&gt;

&lt;P&gt;~Gergana&lt;/P&gt;</description>
      <pubDate>Mon, 23 Nov 2015 17:06:46 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Performance-issue-with-multi-rail-support-in-Intel-MPI-5-0/m-p/1027384#M4129</guid>
      <dc:creator>Gergana_S_Intel</dc:creator>
      <dc:date>2015-11-23T17:06:46Z</dc:date>
    </item>
  </channel>
</rss>

