I cannot. Because I get the SAME errors. Following is the LSF script I used to launch the job.<br><br>#!/bin/bash<br>#BSUB -L /bin/bash<br>#BSUB -n 8<br>#BSUB -N<br>#BSUB -o /data1/visitor/cgaurik/testmpi/helloworld.mympi.mpiexec.%J.out<br>
<br>cd /data1/visitor/cgaurik/testmpi<br>/data1/visitor/cgaurik/mympi/bin/mpiexec -np 8 ./helloworld.mympi<br><br clear="all">The job is NOT parallelized i.e. every process is rank 0. And errors are same. Of course, if I change (as, I think, Dave pointed out) to srun ./helloworld.mympi in the last line of the script, everything is all rosy. My question is, (maybe it's obvious...) that if my mpich2 is configured with the options "--with-pmi=slurm --with-pm=no --with-slurm=/path/to/slurm/lib", can I still use mpiexec?<br>
<br>Gauri.<br>---------<br>
<br><br><div class="gmail_quote">On Wed, Apr 1, 2009 at 11:42 PM, Rajeev Thakur <span dir="ltr"><<a href="mailto:thakur@mcs.anl.gov">thakur@mcs.anl.gov</a>></span> wrote:<br><blockquote class="gmail_quote" style="border-left: 1px solid rgb(204, 204, 204); margin: 0pt 0pt 0pt 0.8ex; padding-left: 1ex;">
<div>
<div dir="ltr" align="left"><span><font color="#0000ff" face="Arial" size="2">You need to use the mpicc and mpiexec from the MPICH2
installation that was built to use MPD.</font></span></div>
<div dir="ltr" align="left"><span><font color="#0000ff" face="Arial" size="2"></font></span> </div>
<div dir="ltr" align="left"><span><font color="#0000ff" face="Arial" size="2">Rajeev</font></span></div>
<div dir="ltr" align="left"><span><font color="#0000ff" face="Arial" size="2"></font></span> </div><br>
<blockquote style="border-left: 2px solid rgb(0, 0, 255); padding-left: 5px; margin-left: 5px; margin-right: 0px;">
<div dir="ltr" align="left" lang="en-us">
<hr>
<font face="Tahoma" size="2"><b>From:</b> <a href="mailto:mpich-discuss-bounces@mcs.anl.gov" target="_blank">mpich-discuss-bounces@mcs.anl.gov</a>
[mailto:<a href="mailto:mpich-discuss-bounces@mcs.anl.gov" target="_blank">mpich-discuss-bounces@mcs.anl.gov</a>] <b>On Behalf Of </b>Gauri
Kulkarni<br><b>Sent:</b> Wednesday, April 01, 2009 8:56 AM<br><b>To:</b>
<a href="mailto:mpich-discuss@mcs.anl.gov" target="_blank">mpich-discuss@mcs.anl.gov</a><br><b>Subject:</b> [mpich-discuss] What do these
errors mean?<br></font><br></div><div><div></div><div class="h5">
<div></div>Hi,<br><br>I am using MPICH2-1.0.7 (I cannot go to 1.0.8 right now)
which is configured to be used with SLURM. That is, the process manager is
SLURM and NOT mpd. When I submit my job through bsub (bsub [options] srun
./helloworld.mympi), it works perfectly. I cannot use mpiexec as it is not the
one spawning jobs, I must use srun. My question is, can I still use mpiexec
from command-line? Well.. I tried. Here is the output:<br><br>mpiexec -n 2
./helloworld.mympi<br>mpiexec_n53: cannot connect to local mpd
(/tmp/mpd2.console_cgaurik); possible causes:<br> 1. no mpd is running
on this host<br> 2. an mpd is running but was started without a
"console" (-n option)<br>In case 1, you can start an mpd on this host
with:<br> mpd &<br>and you will be able to run jobs just
on this host.<br>For more details on starting mpds on a set of hosts,
see<br>the MPICH2 Installation Guide.<br><br>Then:<br><br>mpd &<br>mpiexec
-n 2 ./helloworld.mympi<br><br><b>Hello world! I'm 0 of 2 on n53</b><br>Fatal
error in MPI_Finalize: Other MPI error, error
stack:<br>MPI_Finalize(255)...................: MPI_Finalize
failed<br>MPI_Finalize(154)...................:<br>MPID_Finalize(94)...................:<br>MPI_Barrier(406)....................:
MPI_Barrier(comm=0x44000002)
failed<br>MPIR_Barrier(77)....................:<br>MPIC_Sendrecv(120)..................:<br>MPID_Isend(103).....................:
failure occurred while attempting to send an eager
message<br>MPIDI_CH3_iSend(172)................:<br>MPIDI_CH3I_VC_post_sockconnect(1090):<br>MPIDI_PG_SetConnInfo(615)...........:
PMI_KVS_Get failedFatal error in MPI_Finalize: Other MPI error, error
stack:<br>MPI_Finalize(255)...................: MPI_Finalize
failed<br>MPI_Finalize(154)...................:<br>MPID_Finalize(94)...................:<br>MPI_Barrier(406)....................:
MPI_Barrier(comm=0x44000002)
failed<br>MPIR_Barrier(77)....................:<br>MPIC_Sendrecv(120)..................:<br>MPID_Isend(103).....................:
failure occurred while attempting to send an eager message<br>MP<b>Hello
world! I'm 1 of 2 on
n53</b><br>IDI_CH3_iSend(172)................:<br>MPIDI_CH3I_VC_post_sockconnect(1090):<br>MPIDI_PG_SetConnInfo(615)...........:
PMI_KVS_Get failed<br><br>The bold text shows that the job gets executed but
there is a lot of other garbage. It seems to me that I can either configure
MPICH2 to be used with cluster job scheduler or to be used from command line.
I cannot have both.<br><br>Am I
right?<br><br>-Gauri.<br>----------<br></div></div></blockquote></div>
</blockquote></div><br>