- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi:
My environment: linux64, mpicxx for MVAPICH2 version 2.0b, icpc version 13.1.3 (gcc version 4.7.0 compatibility). In order not to confuse with the mkl library in icpc version 13.1.3, I put the mkl 11.3 in /home/intel.
I use the following command:
mpic++ cluster_sparse_solverc/source/cl_solver_unsym_c.c -Wl,-rpath=/home/intel/mkl/lib/intel64 -Wl,-rpath=/home/intel/compiler/lib/intel64 -L/home/intel/mkl/lib/intel64 -L/home/intel/compiler/lib/intel64 -lmkl_intel_lp64 -lmkl_core -lmkl_intel_thread -lmkl_blacs_intelmpi_lp64 -liomp5
to compile and cause segmentation fault. But in mkl 11.2.4, it is totally correct. So is it a bug in mkl 11.3?
Link Copied
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi,
Can you please provide us a reproducer for your case, so that we can investigate it more carefully?
Best regards,
Maria.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Chaowen, thanks for the case, we will check and let you know the results, I only see you use MPICH beta and pretty aged version of compiler, We validated and check this functionality with the latest version of icc.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
I use the example in mkl examples, that is mkl/examples/examples_cluster.tgz and then extract it and then use cluster_sparse_solverc/source/cl_solver_unsym_c.c as the source code
I use the following command to run it
export MV2_DEBUG_SHOW_BACKTRACE=1
export MV2_DEBUG_CORESIZE=unlimited
mpiexec -n 2 ./a.out
print out:
=== CPARDISO: solving a real nonsymmetric system ===
1-based array indexing is turned ON
CPARDISO double precision computation is turned ON
METIS algorithm at reorder step is turned ON
Scaling is turned ON
Matching is turned ON
Summary: ( reordering phase )
================
Times:
======
Time spent in calculations of symmetric matrix portrait (fulladj): 0.000024 s
Time spent in reordering of the initial matrix (reorder) : 0.000355 s
Time spent in symbolic factorization (symbfct) : 0.000255 s
Time spent in data preparations for factorization (parlist) : 0.000010 s
Time spent in allocation of internal data structures (malloc) : 0.000415 s
Time spent in additional calculations : 0.000037 s
Total time spent : 0.001096 s
Statistics:
===========
Parallel Direct Factorization is running on 2 MPI and 1 OpenMP per MPI process
< Linear system Ax = b >
number of equations: 5
number of non-zeros in A: 13
number of non-zeros in A (%): 52.000000
number of right-hand sides: 1
< Factors L and U >
number of columns for each panel: 128
number of independent subgraphs: 0
< Preprocessing with state of the art partitioning metis>
number of supernodes: 2
size of largest supernode: 4
number of non-zeros in L: 19
number of non-zeros in U: 2
number of non-zeros in L+U: 21
Reordering completed ... [compute-1-1.local:mpi_rank_0][error_sighandler] Caught error: Segmentation fault (signal 11)
[compute-1-1.local:mpi_rank_0][print_backtrace] 0: /lustre/work/prog/mvapich2-2.0b-composer/lib/libmpich.so.10(print_backtrace+0x17) [0x2aabf5e0cfe7]
[compute-1-1.local:mpi_rank_0][print_backtrace] 1: /lustre/work/prog/mvapich2-2.0b-composer/lib/libmpich.so.10(error_sighandler+0x5a) [0x2aabf5e0cfca]
[compute-1-1.local:mpi_rank_0][print_backtrace] 2: /lib64/libc.so.6() [0x36b8e32920]
===================================================================================
= BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
= EXIT CODE: 139
= CLEANING UP REMAINING PROCESSES
= YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
===================================================================================
YOUR APPLICATION TERMINATED WITH THE EXIT STRING: Segmentation fault (signal 11)
This typically refers to a problem with your application.
Please see the FAQ page for debugging suggestions
that means it just finish phase 1, but phase 2 cause segmentation fault.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Chaowen, I checked how this example works when I used Intel MPI instead of MVAPICH2 ( which, actually, officially is not supported by MKL - You may find the list of suppored versions of MPI into RN) .The test passed and the obtained results were corrected.
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page