Intel® MPI Library
Get help with building, analyzing, optimizing, and scaling high-performance computing (HPC) applications.
2154 Discussions

Severe Memory Leak with 2019 impi

L__D__Marks
New Contributor II
1,154 Views

Both 2019 impi releases have a severe memory leak which goes away when I regress to the 2015 version (i.e. source /opt/intel/comp2015/impi/5.0.2.044/intel64/bin/mpivars.sh). I am attaching two valgrind outputs, lapw1.vg.285276 from 2019 impi and lapw1.vg.5451 from 2015 impi which show it quite clearly.

For reference, the entries in the valgrind logs with "init_parallel_ (in /opt/Wien2k_18.1F/lapw1Q_mpi)" are the mpi initialization, so these are almost certainly not real as the initialization is only done once. The entries associated with the scalapack pdsygst call are probably the culprit.

If needed I can provide a package to reproduce this. It is part of a large code, so decomposing into a small test code is not feasible.

0 Kudos
7 Replies
Khang_N_Intel
Employee
1,153 Views

Hi Marks,

Can you send me the package so that I can reproduce the issue?

 

Thanks,

Khang

0 Kudos
L__D__Marks
New Contributor II
1,153 Views

Where do I send it to?

0 Kudos
L__D__Marks
New Contributor II
1,153 Views

Unwrap the attached tarball and look at README. The OUTPUT I obtained a few minutes ago came from "sh README > OUTPUT"

0 Kudos
Khang_N_Intel
Employee
1,153 Views

Hi Marks,

With this kind of issue, it would be better if you can file a ticket in the online service center.

 

Best Regards,

Khang

0 Kudos
L__D__Marks
New Contributor II
1,153 Views

Can I ask for this issue to be escalated.  I reported it to Intel on  11/26/2018 and it is request 03837257. Beyond confirmation that it is probably in impi (which I already had verified) it does not appear to be going anywhere.

I strongly suspect that this bug is a very significant roadblock in the use of impi for the Wien2k software package (www.wien2k.at) which has about 12000 citations on Google scholar and has been licenced to more than 3000 users groups around the world. So far I have myself documented it on my cluster, one in Romania and more recently when testing a cluster from a vendor in the US.

0 Kudos
James_T_Intel
Moderator
1,153 Views

This should be resolved with Intel® MPI Library 2019 Update 4.

0 Kudos
L__D__Marks
New Contributor II
1,153 Views

On 04/03/2019 the engineering team response to 03837257 (the ticket related to this) indicated that it might be solved with Upgrade 4. They have not confirmed with me that it is.

0 Kudos
Reply