<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Quote:Hearns, John wrote: in Intel® MPI Library</title>
    <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140752#M5901</link>
    <description>&lt;P&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;Hearns, John wrote:&lt;BR /&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Also please tell us a little about the HPC cluster which you are using.&lt;/P&gt;&lt;P&gt;I think the answer is to install CICE on a shared sotrage are which you have access to.&lt;/P&gt;&lt;P&gt;Runnign codes from the /home directory is normally a bad idea on HPC clusters&lt;/P&gt;&lt;P&gt;(This depends of course - if the /home is on a fast parallel filesystem what I said does not apply)&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hey John.&amp;nbsp;&lt;/P&gt;&lt;P&gt;What I am currently using is a small server. And here is what I got from '&lt;STRONG&gt;lscpu&lt;/STRONG&gt;'&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;&lt;P&gt;Architecture: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;x86_64&lt;BR /&gt;CPU op-mode(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;32-bit, 64-bit&lt;BR /&gt;Byte Order: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;Little Endian&lt;BR /&gt;CPU(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;112&lt;BR /&gt;On-line CPU(s) list: &amp;nbsp; 0-111&lt;BR /&gt;Thread(s) per core: &amp;nbsp; &amp;nbsp;2&lt;BR /&gt;Core(s) per socket: &amp;nbsp; &amp;nbsp;14&lt;BR /&gt;Socket(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 4&lt;BR /&gt;NUMA node(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;4&lt;BR /&gt;Vendor ID: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; GenuineIntel&lt;BR /&gt;CPU family: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;6&lt;BR /&gt;Model: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 79&lt;BR /&gt;Model name: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;Intel(R) Xeon(R) CPU E7-4830 v4 @ 2.00GHz&lt;BR /&gt;Stepping: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;1&lt;BR /&gt;CPU MHz: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 2001.000&lt;BR /&gt;BogoMIPS: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;3999.91&lt;BR /&gt;Virtualization: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;VT-x&lt;BR /&gt;L1d cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 32K&lt;BR /&gt;L1i cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 32K&lt;BR /&gt;L2 cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;256K&lt;BR /&gt;L3 cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;35840K&lt;BR /&gt;NUMA node0 CPU(s): &amp;nbsp; &amp;nbsp; 0-13,56-69&lt;BR /&gt;NUMA node1 CPU(s): &amp;nbsp; &amp;nbsp; 14-27,70-83&lt;BR /&gt;NUMA node2 CPU(s): &amp;nbsp; &amp;nbsp; 28-41,84-97&lt;BR /&gt;NUMA node3 CPU(s): &amp;nbsp; &amp;nbsp; 42-55,98-111&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;&lt;P&gt;Below is what I got from 'pbsnodes'&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;&lt;P&gt;node01&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;state = free&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;np = 112&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;ntype = cluster&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;status = rectime=1584627510,varattr=,jobs=,state=free,netload=1049588601,gres=,loadave=62.00,ncpus=112,physmem=264605084kb,availmem=309521516kb,totmem=331724648kb,idletime=687823,nusers=4,nsessions=8,sessions=11365 11369 42732 42765 42843 59303 90588 101768,uname=Linux node01 2.6.32-696.el6.x86_64 #1 SMP Tue Mar 21 19:29:05 UTC 2017 x86_64,opsys=linux&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_service_port = 15002&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_manager_port = 15003&lt;/P&gt;&lt;P&gt;node02&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;state = free&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;np = 36&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;ntype = cluster&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;status = rectime=1584627470,varattr=,jobs=,state=free,netload=465920164,gres=,loadave=22.01,ncpus=36,physmem=132250420kb,availmem=192346396kb,totmem=199369952kb,idletime=1483530,nusers=3,nsessions=4,sessions=5401 5405 25630 30630,uname=Linux node02 2.6.32-573.el6.x86_64 #1 SMP Thu Jul 23 15:44:03 UTC 2015 x86_64,opsys=linux&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_service_port = 15002&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_manager_port = 15003&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;</description>
    <pubDate>Thu, 19 Mar 2020 14:28:00 GMT</pubDate>
    <dc:creator>xiong__wang</dc:creator>
    <dc:date>2020-03-19T14:28:00Z</dc:date>
    <item>
      <title>Can't run job with impi</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140745#M5894</link>
      <description>&lt;P&gt;Dear all:&lt;/P&gt;&lt;P&gt;I am currently using INTEL COMPILER and IMPI to run the CICE numerical model. But I failed every single time. According to the runlog, the model returns the error message "&lt;STRONG&gt;rank 0 in job 2 &amp;nbsp;node01_44414 &amp;nbsp; caused collective abort of all ranks exit status of rank 0: killed by signal 11&lt;/STRONG&gt;" I treid to searched this error meassage but unfortunately, I do not know much about MPI and have no idea about how to debug the error. One thing I&amp;nbsp;can make sure is that CICE is a widely used numerical model and I don't think there is any major bug in the code which brings this error.&amp;nbsp;&lt;/P&gt;&lt;P&gt;So is there anyone may provide some insight about this error or tell me which information i should provide&amp;nbsp; to locate this error??&lt;/P&gt;&lt;P&gt;THANKS&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 12 Mar 2020 17:59:59 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140745#M5894</guid>
      <dc:creator>xiong__wang</dc:creator>
      <dc:date>2020-03-12T17:59:59Z</dc:date>
    </item>
    <item>
      <title>Hello Wang.  I think the</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140746#M5895</link>
      <description>&lt;P&gt;Hello Wang.&amp;nbsp; I think the problem is that the CICE code is not available on your compute nodes.&lt;/P&gt;&lt;P&gt;Please do the following:&lt;/P&gt;&lt;P&gt;Log into the cluster login node or head node&lt;/P&gt;&lt;P&gt;Run the command&amp;nbsp; which CICE&amp;nbsp; &amp;nbsp; then&amp;nbsp; ldd `which CICE`&lt;/P&gt;&lt;P&gt;the ldd command will list the libraries which an eecutable code needs. if any ibraries are unavilable we need to investigate&lt;/P&gt;&lt;P&gt;Now log into node01&amp;nbsp; or any compute node. Run which CICE and&amp;nbsp; &amp;nbsp;ldd `which CICE`&lt;/P&gt;&lt;P&gt;Do you have the code available on node01 and are all the libraries available?&lt;/P&gt;</description>
      <pubDate>Mon, 16 Mar 2020 09:12:10 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140746#M5895</guid>
      <dc:creator>Hearns__John</dc:creator>
      <dc:date>2020-03-16T09:12:10Z</dc:date>
    </item>
    <item>
      <title>Also please tell us a little</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140747#M5896</link>
      <description>&lt;P&gt;Also please tell us a little about the HPC cluster which you are using.&lt;/P&gt;&lt;P&gt;I think the answer is to install CICE on a shared sotrage are which you have access to.&lt;/P&gt;&lt;P&gt;Runnign codes from the /home directory is normally a bad idea on HPC clusters&lt;/P&gt;&lt;P&gt;(This depends of course - if the /home is on a fast parallel filesystem what I said does not apply)&lt;/P&gt;</description>
      <pubDate>Mon, 16 Mar 2020 12:05:25 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140747#M5896</guid>
      <dc:creator>Hearns__John</dc:creator>
      <dc:date>2020-03-16T12:05:25Z</dc:date>
    </item>
    <item>
      <title>Quote:Hearns, John wrote:</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140748#M5897</link>
      <description>&lt;P&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;Hearns, John wrote:&lt;BR /&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Hello Wang.&amp;nbsp; I think the problem is that the CICE code is not available on your compute nodes.&lt;/P&gt;&lt;P&gt;Please do the following:&lt;/P&gt;&lt;P&gt;Log into the cluster login node or head node&lt;/P&gt;&lt;P&gt;Run the command&amp;nbsp; which CICE&amp;nbsp; &amp;nbsp; then&amp;nbsp; ldd `which CICE`&lt;/P&gt;&lt;P&gt;the ldd command will list the libraries which an eecutable code needs. if any ibraries are unavilable we need to investigate&lt;/P&gt;&lt;P&gt;Now log into node01&amp;nbsp; or any compute node. Run which CICE and&amp;nbsp; &amp;nbsp;ldd `which CICE`&lt;/P&gt;&lt;P&gt;Do you have the code available on node01 and are all the libraries available?&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Hello John. Thank you for your reply.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I logged into the node01 and changed to the directory which CICE exists. I typed into the which CICE and I got&lt;STRONG&gt; "no CICE in (...... a lot of path)". &lt;/STRONG&gt;And I typed in "ldd 'which CICE'". I got &lt;STRONG&gt;"which : ldd: ./which: No such file or directory, CICE: ldd: ./CICE: No such file or directory" &lt;/STRONG&gt;I tried to add the CICE path into the PATH&lt;STRONG&gt; by "export PATH=/home/wangxiong/CICE/mycase8:$PATH". &lt;/STRONG&gt;(/home/wangxiong/CICE/mycase8 is where CICE exists) After doing that. I typed in "which cice" and I got "~/CICE/mycase8/cice".&lt;/P&gt;&lt;P&gt;I typed into "ldd 'which cice'", I got&lt;/P&gt;&lt;P&gt;"cice:&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; linux-vdso.so.1 =&amp;gt; &amp;nbsp;(0x00007ffc77fe8000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libnetcdf.so.15 =&amp;gt; /usr/local/netcdf-intel-mpi/netcdf/lib/libnetcdf.so.15 (0x00007f77a18a9000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libnetcdff.so.7 =&amp;gt; /usr/local/netcdf-intel-mpi/netcdf/lib/libnetcdff.so.7 (0x00007f77a1409000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libpnetcdf.so.4 =&amp;gt; /usr/local/netcdf-intel-mpi/pnetcdf/lib/libpnetcdf.so.4 (0x00007f77a0a3b000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libmpifort.so.12 =&amp;gt; /opt/impi/5.0.1.035/intel64/lib/libmpifort.so.12 (0x00007f77a07af000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libmpi.so.12 =&amp;gt; /opt/impi/5.0.1.035/intel64/lib/libmpi.so.12 (0x00007f779fde3000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libdl.so.2 =&amp;gt; /lib64/libdl.so.2 (0x000000318ac00000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; librt.so.1 =&amp;gt; /lib64/librt.so.1 (0x000000318b800000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libpthread.so.0 =&amp;gt; /lib64/libpthread.so.0 (0x000000318a800000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libm.so.6 =&amp;gt; /lib64/libm.so.6 (0x000000318b000000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libc.so.6 =&amp;gt; /lib64/libc.so.6 (0x000000318a400000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libgcc_s.so.1 =&amp;gt; /lib64/libgcc_s.so.1 (0x000000318d800000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libhdf5_hl.so.8 =&amp;gt; /usr/local/netcdf-intel-mpi/hdf5/lib/libhdf5_hl.so.8 (0x00007f779fb71000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libhdf5.so.8 =&amp;gt; /usr/local/netcdf-intel-mpi/hdf5/lib/libhdf5.so.8 (0x00007f779f595000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libsz.so.2 =&amp;gt; /usr/local/netcdf-intel-mpi/szip/lib/libsz.so.2 (0x00007f779f376000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libz.so.1 =&amp;gt; /lib64/libz.so.1 (0x000000318b400000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libifport.so.5 =&amp;gt; /opt/intel/icc/composer_xe_2013.3.163/compiler/lib/intel64/libifport.so.5 (0x00007f779f147000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libifcore.so.5 =&amp;gt; /opt/intel/icc/composer_xe_2013.3.163/compiler/lib/intel64/libifcore.so.5 (0x00007f779ee10000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libimf.so =&amp;gt; /opt/intel/icc/composer_xe_2013.3.163/compiler/lib/intel64/libimf.so (0x00007f779e954000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libsvml.so =&amp;gt; /opt/intel/icc/composer_xe_2013.3.163/compiler/lib/intel64/libsvml.so (0x00007f779df8a000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libirc.so =&amp;gt; /opt/intel/icc/composer_xe_2013.3.163/compiler/lib/intel64/libirc.so (0x00007f779dd3b000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libirng.so =&amp;gt; /opt/intel/icc/composer_xe_2013.3.163/compiler/lib/intel64/libirng.so (0x00007f779db34000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libintlc.so.5 =&amp;gt; /opt/intel/icc/composer_xe_2013.3.163/compiler/lib/intel64/libintlc.so.5 (0x00007f779d8e6000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; /lib64/ld-linux-x86-64.so.2 (0x0000003189c00000)&lt;BR /&gt;"&lt;/P&gt;&lt;P&gt;And after that, I recompiled and rerun the model. Again I got the error message&lt;STRONG&gt; "rank 3 in job 11 &amp;nbsp;node01_44414 &amp;nbsp; caused collective abort of all ranks&amp;nbsp; exit status of rank 3: killed by signal 11" The error message looks like poping up at where the calculations should begin.&amp;nbsp;&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;FYI: I don't konw if this information would be helpful but I still want to mention it. About two to three years ago. This server has been used to run MITgcm(another numerical model) successfully(also use&amp;nbsp;intel compiler and impi). And after that, this server hasn't been used to run model in parallel. I switched into the MITgcm directory and typed into which 'which mitgcmuv' . This time I got "&lt;STRONG&gt;./mitgcmuv&lt;/STRONG&gt;" When typed into "ldd 'which mitgcmuv'", I also got some results like "&lt;/P&gt;&lt;P&gt;mitgcmuv:&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; linux-vdso.so.1 =&amp;gt; &amp;nbsp;(0x00007ffd1a797000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libdl.so.2 =&amp;gt; /lib64/libdl.so.2 (0x000000318ac00000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libmpi_dbg.so.4 =&amp;gt; /opt/impi/5.0.1.035/intel64/lib/libmpi_dbg.so.4 (0x00007fef40ee5000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libmpigf.so.4 =&amp;gt; /opt/impi/5.0.1.035/intel64/lib/libmpigf.so.4 (0x00007fef40c59000)&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libpthread.so.0 =&amp;gt; /lib64/libpthread.so.0 (0x000000318a800000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; librt.so.1 =&amp;gt; /lib64/librt.so.1 (0x000000318b800000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libm.so.6 =&amp;gt; /lib64/libm.so.6 (0x000000318b000000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libc.so.6 =&amp;gt; /lib64/libc.so.6 (0x000000318a400000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; libgcc_s.so.1 =&amp;gt; /lib64/libgcc_s.so.1 (0x000000318d800000)&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; /lib64/ld-linux-x86-64.so.2 (0x0000003189c00000)&lt;BR /&gt;"&lt;/P&gt;&lt;P&gt;I even tried to rerun the mitgcm case and it succeed. But I just can't make the CICE to run successfully which really makes me&amp;nbsp;frustrating.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 16 Mar 2020 15:10:34 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140748#M5897</guid>
      <dc:creator>xiong__wang</dc:creator>
      <dc:date>2020-03-16T15:10:34Z</dc:date>
    </item>
    <item>
      <title>I would try running the code</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140749#M5898</link>
      <description>&lt;P&gt;I would try running the code with 2 processes but only on the login node - ie run it in an interactive session with np-2&lt;/P&gt;&lt;P&gt;Also set I_MPI_DEBUG=5&amp;nbsp; before you run the code&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The next step if that work is to run between two compute nodes&lt;/P&gt;</description>
      <pubDate>Tue, 17 Mar 2020 08:25:53 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140749#M5898</guid>
      <dc:creator>Hearns__John</dc:creator>
      <dc:date>2020-03-17T08:25:53Z</dc:date>
    </item>
    <item>
      <title>Hi Wang,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140750#M5899</link>
      <description>&lt;P&gt;Hi Wang,&lt;/P&gt;&lt;P&gt;Could you please provide the GitHub link of the CICE code you are working, so that we can reproduce from our side.&lt;/P&gt;&lt;P&gt;Also please provide the steps you are following and the environmental details (like OS, Compiler, Hardware)&amp;nbsp;of your system.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Prasanth&lt;/P&gt;</description>
      <pubDate>Tue, 17 Mar 2020 10:27:03 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140750#M5899</guid>
      <dc:creator>PrasanthD_intel</dc:creator>
      <dc:date>2020-03-17T10:27:03Z</dc:date>
    </item>
    <item>
      <title>Quote:Hearns, John wrote:</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140752#M5901</link>
      <description>&lt;P&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;Hearns, John wrote:&lt;BR /&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Also please tell us a little about the HPC cluster which you are using.&lt;/P&gt;&lt;P&gt;I think the answer is to install CICE on a shared sotrage are which you have access to.&lt;/P&gt;&lt;P&gt;Runnign codes from the /home directory is normally a bad idea on HPC clusters&lt;/P&gt;&lt;P&gt;(This depends of course - if the /home is on a fast parallel filesystem what I said does not apply)&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hey John.&amp;nbsp;&lt;/P&gt;&lt;P&gt;What I am currently using is a small server. And here is what I got from '&lt;STRONG&gt;lscpu&lt;/STRONG&gt;'&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;&lt;P&gt;Architecture: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;x86_64&lt;BR /&gt;CPU op-mode(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;32-bit, 64-bit&lt;BR /&gt;Byte Order: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;Little Endian&lt;BR /&gt;CPU(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;112&lt;BR /&gt;On-line CPU(s) list: &amp;nbsp; 0-111&lt;BR /&gt;Thread(s) per core: &amp;nbsp; &amp;nbsp;2&lt;BR /&gt;Core(s) per socket: &amp;nbsp; &amp;nbsp;14&lt;BR /&gt;Socket(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 4&lt;BR /&gt;NUMA node(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;4&lt;BR /&gt;Vendor ID: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; GenuineIntel&lt;BR /&gt;CPU family: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;6&lt;BR /&gt;Model: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 79&lt;BR /&gt;Model name: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;Intel(R) Xeon(R) CPU E7-4830 v4 @ 2.00GHz&lt;BR /&gt;Stepping: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;1&lt;BR /&gt;CPU MHz: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 2001.000&lt;BR /&gt;BogoMIPS: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;3999.91&lt;BR /&gt;Virtualization: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;VT-x&lt;BR /&gt;L1d cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 32K&lt;BR /&gt;L1i cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 32K&lt;BR /&gt;L2 cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;256K&lt;BR /&gt;L3 cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;35840K&lt;BR /&gt;NUMA node0 CPU(s): &amp;nbsp; &amp;nbsp; 0-13,56-69&lt;BR /&gt;NUMA node1 CPU(s): &amp;nbsp; &amp;nbsp; 14-27,70-83&lt;BR /&gt;NUMA node2 CPU(s): &amp;nbsp; &amp;nbsp; 28-41,84-97&lt;BR /&gt;NUMA node3 CPU(s): &amp;nbsp; &amp;nbsp; 42-55,98-111&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;&lt;P&gt;Below is what I got from 'pbsnodes'&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;&lt;P&gt;node01&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;state = free&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;np = 112&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;ntype = cluster&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;status = rectime=1584627510,varattr=,jobs=,state=free,netload=1049588601,gres=,loadave=62.00,ncpus=112,physmem=264605084kb,availmem=309521516kb,totmem=331724648kb,idletime=687823,nusers=4,nsessions=8,sessions=11365 11369 42732 42765 42843 59303 90588 101768,uname=Linux node01 2.6.32-696.el6.x86_64 #1 SMP Tue Mar 21 19:29:05 UTC 2017 x86_64,opsys=linux&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_service_port = 15002&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_manager_port = 15003&lt;/P&gt;&lt;P&gt;node02&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;state = free&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;np = 36&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;ntype = cluster&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;status = rectime=1584627470,varattr=,jobs=,state=free,netload=465920164,gres=,loadave=22.01,ncpus=36,physmem=132250420kb,availmem=192346396kb,totmem=199369952kb,idletime=1483530,nusers=3,nsessions=4,sessions=5401 5405 25630 30630,uname=Linux node02 2.6.32-573.el6.x86_64 #1 SMP Thu Jul 23 15:44:03 UTC 2015 x86_64,opsys=linux&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_service_port = 15002&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_manager_port = 15003&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;</description>
      <pubDate>Thu, 19 Mar 2020 14:28:00 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140752#M5901</guid>
      <dc:creator>xiong__wang</dc:creator>
      <dc:date>2020-03-19T14:28:00Z</dc:date>
    </item>
    <item>
      <title>Quote:Hearns, John wrote:</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140753#M5902</link>
      <description>&lt;P&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;Hearns, John wrote:&lt;BR /&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I would try running the code with 2 processes but only on the login node - ie run it in an interactive session with np-2&lt;/P&gt;&lt;P&gt;Also set I_MPI_DEBUG=5&amp;nbsp; before you run the code&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The next step if that work is to run between two compute nodes&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hello John.&lt;/P&gt;&lt;P&gt;I followed your advice and tried to run the job with an interactive session and 2 processors.&lt;/P&gt;&lt;P&gt;First I used the setting "I_MPI_DEBUG=5" . But I almost got the same result as the previous run (which is wired I think). &lt;STRONG&gt;"Finished writing ./history/iceh_ic.1998-01-01-00000.nc APPLICATION TERMINATED WITH THE EXIT STRING: Segmentation fault (signal 11)".&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;After finish wirting the initial condition file, the job terminated and computation wan't start. And no more error message was given.&lt;/P&gt;&lt;P&gt;So I even increased the debug parameter to 20. But apart from some MPI related sentence I am not very familiar about. I still havn't discover some problem which may cause this run-time error.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I also uploaded these two logs in case you want to see them. The one named "&lt;STRONG&gt;run-output-dubug5&lt;/STRONG&gt;" means it is log from the run &amp;nbsp;with "&lt;STRONG&gt;I_MPI_DEBUG=5&lt;/STRONG&gt;"&lt;/P&gt;&lt;P&gt;And&amp;nbsp;"&lt;STRONG&gt;run-output-dubug20&lt;/STRONG&gt;" manes it is the log from the run with "&lt;STRONG&gt;I_MPI_DEBUG=20&lt;/STRONG&gt;".&lt;/P&gt;&lt;P&gt;Thanks again.&lt;/P&gt;&lt;P&gt;Have a good day!&lt;/P&gt;</description>
      <pubDate>Thu, 19 Mar 2020 14:48:26 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140753#M5902</guid>
      <dc:creator>xiong__wang</dc:creator>
      <dc:date>2020-03-19T14:48:26Z</dc:date>
    </item>
    <item>
      <title>Quote:Dwadasi, Prasanth</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140754#M5903</link>
      <description>&lt;P&gt;&lt;/P&gt;&lt;BLOCKQUOTE&gt;Dwadasi, Prasanth (Intel) wrote:&lt;BR /&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Hi Wang,&lt;/P&gt;&lt;P&gt;Could you please provide the GitHub link of the CICE code you are working, so that we can reproduce from our side.&lt;/P&gt;&lt;P&gt;Also please provide the steps you are following and the environmental details (like OS, Compiler, Hardware)&amp;nbsp;of your system.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Prasanth&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hello&amp;nbsp;Prasanth. Thank you for your reply.&lt;/P&gt;&lt;P&gt;1. This is the main page of CICE project on GitHub.&lt;/P&gt;&lt;P&gt;&lt;A href="https://github.com/CICE-Consortium/CICE"&gt;https://github.com/CICE-Consortium/CICE&lt;/A&gt;&lt;/P&gt;&lt;P&gt;2. This is the CICE version index page. What I am currently trying to run is CICE Version 6.1.0&lt;/P&gt;&lt;P&gt;&lt;A href="https://github.com/CICE-Consortium/CICE/wiki/CICE-Version-Index"&gt;https://github.com/CICE-Consortium/CICE/wiki/CICE-Version-Index&lt;/A&gt;&lt;/P&gt;&lt;P&gt;3. Icepack version index page. Icepack is one part of CICE project. One may need to put the Icepack code in the CICE code directory before compile and run.&amp;nbsp; CICE Version 6.1.0 corresponds to Icepack Version 1.2.0&lt;/P&gt;&lt;P&gt;&lt;A href="https://github.com/CICE-Consortium/Icepack/wiki/Icepack-Version-Index"&gt;https://github.com/CICE-Consortium/Icepack/wiki/Icepack-Version-Index&lt;/A&gt;&lt;/P&gt;&lt;P&gt;4. We also need to download some forcing data in order to do some test run.&lt;/P&gt;&lt;P&gt;&lt;A href="https://github.com/CICE-Consortium/CICE/wiki/CICE-Input-Data"&gt;https://github.com/CICE-Consortium/CICE/wiki/CICE-Input-Data&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://github.com/CICE-Consortium/Icepack/wiki/Icepack-Input-Data"&gt;https://github.com/CICE-Consortium/Icepack/wiki/Icepack-Input-Data&lt;/A&gt;&lt;/P&gt;&lt;P&gt;5. CICE documentation for CICE Version 6.1.0&lt;/P&gt;&lt;P&gt;&lt;A href="https://cice-consortium-cice.readthedocs.io/en/cice6.1.0/"&gt;https://cice-consortium-cice.readthedocs.io/en/cice6.1.0/&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Here are the steps for CICE compile and run.&lt;/P&gt;&lt;P&gt;1) Once downloaded the CICE and Icepack code and forcing data. Uncompress the code and forcing data.&lt;/P&gt;&lt;P&gt;Copy the Icepack code into the icepack directory under the CICE directory.&lt;/P&gt;&lt;P&gt;2) Porting (tells the cice model about the compiler information on this server)&lt;/P&gt;&lt;P&gt;this includes:&lt;/P&gt;&lt;P&gt;cd to configuration/scripts/machines/&lt;/P&gt;&lt;P&gt;Copy an existing env and a Macros file to new names for your new machine&lt;/P&gt;&lt;P&gt;Edit your env and Macros files&lt;/P&gt;&lt;P&gt;cd .. to configuration/scripts/&lt;/P&gt;&lt;P&gt;Edit the cice.batch.csh script to add a section for your machine with batch settings&lt;/P&gt;&lt;P&gt;Edit the cice.batch.csh script to add a section for your machine with job launch settings&lt;/P&gt;&lt;P&gt;Change the variable&amp;nbsp;ICE_MACHINE_INPUTDATA in the env file accordint to where you put the forcing data.&lt;/P&gt;&lt;P&gt;3) setup the CICE run directory&lt;/P&gt;&lt;P&gt;cd to the CICE main directory&lt;/P&gt;&lt;P&gt;./cice.setup -c ~/mycase1 -g gx3 -m testmachine &amp;nbsp;(gx3 means you choose the gx3 grid rather than gx1 grid ,&amp;nbsp;"testmachine" may change&amp;nbsp;to the name you set for your Macros file in step 2&amp;nbsp;)&lt;/P&gt;&lt;P&gt;4) Compile the code&lt;/P&gt;&lt;P&gt;cd mycase1 ("mycase1" is specified by -c option in step 3)&lt;/P&gt;&lt;P&gt;./cice.build&lt;/P&gt;&lt;P&gt;5) Run the model&lt;/P&gt;&lt;P&gt;./cice.submit&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The server is installed with&amp;nbsp;CentOS release 6.9 (Final) and kernel version&amp;nbsp;2.6.32-696.el6.x86_64.&lt;/P&gt;&lt;P&gt;The IMPI version is :&amp;nbsp;Intel(R) MPI Library, Version 5.0 Update 1 &amp;nbsp;Build 20140709&lt;/P&gt;&lt;P&gt;Intel Compiler Version:&amp;nbsp;&amp;nbsp;Intel(R) C Intel(R) 64 Compiler XE for applications running on Intel(R) 64, Version 13.1.1.163 Build 20130313&lt;/P&gt;&lt;P&gt;Below are the hardware&amp;nbsp; and nodes information:&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;&lt;P&gt;Architecture: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;x86_64&lt;BR /&gt;CPU op-mode(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;32-bit, 64-bit&lt;BR /&gt;Byte Order: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;Little Endian&lt;BR /&gt;CPU(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;112&lt;BR /&gt;On-line CPU(s) list: &amp;nbsp; 0-111&lt;BR /&gt;Thread(s) per core: &amp;nbsp; &amp;nbsp;2&lt;BR /&gt;Core(s) per socket: &amp;nbsp; &amp;nbsp;14&lt;BR /&gt;Socket(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 4&lt;BR /&gt;NUMA node(s): &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;4&lt;BR /&gt;Vendor ID: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; GenuineIntel&lt;BR /&gt;CPU family: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;6&lt;BR /&gt;Model: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 79&lt;BR /&gt;Model name: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;Intel(R) Xeon(R) CPU E7-4830 v4 @ 2.00GHz&lt;BR /&gt;Stepping: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;1&lt;BR /&gt;CPU MHz: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 2001.000&lt;BR /&gt;BogoMIPS: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;3999.91&lt;BR /&gt;Virtualization: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;VT-x&lt;BR /&gt;L1d cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 32K&lt;BR /&gt;L1i cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; 32K&lt;BR /&gt;L2 cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;256K&lt;BR /&gt;L3 cache: &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;35840K&lt;BR /&gt;NUMA node0 CPU(s): &amp;nbsp; &amp;nbsp; 0-13,56-69&lt;BR /&gt;NUMA node1 CPU(s): &amp;nbsp; &amp;nbsp; 14-27,70-83&lt;BR /&gt;NUMA node2 CPU(s): &amp;nbsp; &amp;nbsp; 28-41,84-97&lt;BR /&gt;NUMA node3 CPU(s): &amp;nbsp; &amp;nbsp; 42-55,98-111&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;&lt;P&gt;——————————————————————————————————————————————————&lt;/P&gt;&lt;P&gt;"&lt;/P&gt;&lt;P&gt;node01&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;state = free&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;np = 112&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;ntype = cluster&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;status = rectime=1584627510,varattr=,jobs=,state=free,netload=1049588601,gres=,loadave=62.00,ncpus=112,physmem=264605084kb,availmem=309521516kb,totmem=331724648kb,idletime=687823,nusers=4,nsessions=8,sessions=11365 11369 42732 42765 42843 59303 90588 101768,uname=Linux node01 2.6.32-696.el6.x86_64 #1 SMP Tue Mar 21 19:29:05 UTC 2017 x86_64,opsys=linux&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_service_port = 15002&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_manager_port = 15003&lt;/P&gt;&lt;P&gt;node02&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;state = free&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;np = 36&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;ntype = cluster&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;status = rectime=1584627470,varattr=,jobs=,state=free,netload=465920164,gres=,loadave=22.01,ncpus=36,physmem=132250420kb,availmem=192346396kb,totmem=199369952kb,idletime=1483530,nusers=3,nsessions=4,sessions=5401 5405 25630 30630,uname=Linux node02 2.6.32-573.el6.x86_64 #1 SMP Thu Jul 23 15:44:03 UTC 2015 x86_64,opsys=linux&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_service_port = 15002&lt;BR /&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;mom_manager_port = 15003&lt;/P&gt;&lt;P&gt;"&lt;BR /&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Many thanks.&lt;/P&gt;&lt;P&gt;Have a nice day.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 19 Mar 2020 15:46:00 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140754#M5903</guid>
      <dc:creator>xiong__wang</dc:creator>
      <dc:date>2020-03-19T15:46:00Z</dc:date>
    </item>
    <item>
      <title>Here is another suggestion.</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140755#M5904</link>
      <description>&lt;P&gt;Here is another suggestion.&lt;/P&gt;&lt;P&gt;Can you create a simple MPI 'Hello World' program?&amp;nbsp; Just Google for this.&lt;/P&gt;&lt;P&gt;compiel and line the program then submit it as a batch job. Set I_MPI_DEBUG=5 in the script&lt;/P&gt;&lt;P&gt;If a 'hello world' runs fine we can see that the cluster is working OK&lt;/P&gt;</description>
      <pubDate>Thu, 19 Mar 2020 16:35:21 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140755#M5904</guid>
      <dc:creator>Hearns__John</dc:creator>
      <dc:date>2020-03-19T16:35:21Z</dc:date>
    </item>
    <item>
      <title>Hi Wang,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140756#M5905</link>
      <description>&lt;P&gt;Hi Wang,&lt;/P&gt;&lt;P&gt;As John suggested please&amp;nbsp;try to run a sample Hello World program.&lt;/P&gt;&lt;P&gt;You can compile using&amp;nbsp;&lt;/P&gt;&lt;P&gt;mpiicc &amp;lt;Foo.c &amp;gt;&lt;/P&gt;&lt;P&gt;and&lt;/P&gt;&lt;P&gt;run using&lt;/P&gt;&lt;P&gt;mpirun -np &amp;lt;number of processes&amp;gt; ./&amp;lt;a.out&amp;gt;&lt;/P&gt;&lt;P&gt;tell us whether you can run it successfully.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Prasanth&lt;/P&gt;</description>
      <pubDate>Thu, 26 Mar 2020 13:33:21 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140756#M5905</guid>
      <dc:creator>PrasanthD_intel</dc:creator>
      <dc:date>2020-03-26T13:33:21Z</dc:date>
    </item>
    <item>
      <title>Hi Wang,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140757#M5906</link>
      <description>&lt;P&gt;Hi&amp;nbsp;Wang,&lt;/P&gt;&lt;P&gt;The IMPI version you are using&amp;nbsp;5.0 which is an outdated version and currently not supported.&lt;/P&gt;&lt;P&gt;Can you update to the latest IMPI&amp;nbsp;&amp;nbsp;version and check if the error persists.&lt;/P&gt;&lt;P&gt;Also if possible could you check with other MPI vendors and see whether you are getting any errors.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Prasanth&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 31 Mar 2020 09:23:00 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140757#M5906</guid>
      <dc:creator>PrasanthD_intel</dc:creator>
      <dc:date>2020-03-31T09:23:00Z</dc:date>
    </item>
    <item>
      <title>Hi Wang,</title>
      <link>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140758#M5907</link>
      <description>&lt;P&gt;Hi Wang,&lt;/P&gt;&lt;P&gt;We are closing this thread considering your issue got resolved.&lt;/P&gt;&lt;P&gt;Please raise a new thread for any further questions.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Regards&lt;/P&gt;&lt;P&gt;Prasanth&lt;/P&gt;</description>
      <pubDate>Mon, 27 Apr 2020 10:49:46 GMT</pubDate>
      <guid>https://community.intel.com/t5/Intel-MPI-Library/Can-t-run-job-with-impi/m-p/1140758#M5907</guid>
      <dc:creator>PrasanthD_intel</dc:creator>
      <dc:date>2020-04-27T10:49:46Z</dc:date>
    </item>
  </channel>
</rss>

