We actually don't include MPICH in Intel® Parallel Studio XE. We include our own implementation, Intel® MPI Library. Can you please confirm which MPI you are using?
Our engineering team has implemented a fix for the bug mentioned above, and expects you could be encountering a different problem.
Can you also provide the exact configuration and tests which are failing? Our engineering team tested with the problem test reported in the mpich-maillist and did not get any errors.
$ mpirun -n 2 src/ksp/ksp/examples/tutorials/ex2 Norm of error 0.000411674 iterations 7 $ I_MPI_FABRICS=ofi mpirun -n 2 src/ksp/ksp/examples/tutorials/ex2 Norm of error 0.000411674 iterations 7
please help me understand:
Why the mpi.h file distributed is the MPICH one?
When trying to identify the MPI flavor at *compile time*, it is impossible! (see https://software.intel.com/en-us/forums/intel-clusters-and-hpc-technology/topic/797705)
Ok, I reinstalled everything...
to confirm the version I used:
which mpirun /opt/intel/compilers_and_libraries_2019.0.117/linux/mpi/intel64/bin/mpirun
-bash-4.4$ mpirun -v
[mpiexec@thorin] main (../../../../../src/pm/hydra2/mpiexec/mpiexec.c:1645): assert (pg->total_proc_count * sizeof(int)) failed
-bash-4.4$ mpirun -V
Intel(R) MPI Library for Linux* OS, Version 2019 Build 20180829 (id: 15f5d6c0c) Copyright 2003-2018, Intel Corporation.
Also, I tried the PETSc example, and it works! :/
So my problem is different then the one I thinked...
I have an error returned by a "simple" MPI_Isend. I am trying to produce a MWE to give it to you.
this is still a "tag" bug.
I have simple MWE attached, which gives the following explicit error:
Abort(67744516) on node 0 (rank 0 in comm 0): Fatal error in PMPI_Irecv: Invalid tag, error stack:
PMPI_Irecv(156): MPI_Irecv(buf=0x7ffff2268900, count=10, MPI_INT, src=0, tag=123454321, MPI_COMM_WORLD, request=0x7ffff2268974) failed
PMPI_Irecv(100): Invalid tag, value is 123454321
If you compile the attached example with OpenMPI or MPICH, it works!
Hope this help!