<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
<HTML><HEAD>
<META HTTP-EQUIV="Content-Type" CONTENT="text/html; charset=gb2312">
<TITLE>邮件</TITLE>
<META name=GENERATOR content="MSHTML 8.00.6001.18939"></HEAD>
<BODY>
<DIV><SPAN class=093414909-03092010><FONT color=#0000ff size=2 face=宋体>Hi,
all,</FONT></SPAN></DIV>
<DIV><SPAN class=093414909-03092010><FONT color=#0000ff size=2
face=宋体></FONT></SPAN> </DIV>
<DIV><SPAN class=093414909-03092010><FONT color=#0000ff size=2 face=宋体>I allways
recieved many emails from <A
href="mailto:mpich-discuss@mcs.anl.gov">mpich-discuss@mcs.anl.gov</A> every
day. I am therefore disturbed. Can anyone tell me how to deleted my user ID
in the MPI forum.</FONT></SPAN></DIV>
<DIV><SPAN class=093414909-03092010></SPAN> </DIV>
<DIV><SPAN class=093414909-03092010><FONT color=#0000ff size=2
face=宋体>Thanks,</FONT></SPAN></DIV>
<DIV><SPAN class=093414909-03092010></SPAN> </DIV>
<DIV><SPAN class=093414909-03092010><FONT color=#0000ff size=2 face=宋体>Qian-Lin
Tang</FONT> </SPAN></DIV>
<BLOCKQUOTE style="MARGIN-RIGHT: 0px">
<DIV></DIV>
<DIV dir=ltr lang=zh-cn class=OutlookMessageHeader align=left><FONT size=2
face=宋体>-----邮件原件-----<BR><B>发件人:</B> mpich-discuss-bounces@mcs.anl.gov
[mailto:mpich-discuss-bounces@mcs.anl.gov] <B>代表 </B>Fabio
F.Gervasi<BR><B>发送时间:</B> 2010年9月3日 15:57<BR><B>收件人:</B>
mpich-discuss@mcs.anl.gov<BR><B>主题:</B> Re: [mpich-discuss] MPI Fatal
error,but only with more cluster nodes!<BR><BR></FONT></DIV>Thank you Rajeev
and Gus!<BR><BR>I have no problems about network connection because<BR>I tried
to run the "WRF Gnu gcc/gfortran" buildings version<BR>and everything is ok (I
can run also 5 pc <SPAN class=number></SPAN><SPAN
class=definition>together).<BR><BR>So, </SPAN>on equal other terms, I have
problems (on more then 1 pc) only when I use the WRF Intel compiled version
(wrf stop with mpi errors).<BR>Besides, before start wrf.exe, I need to start
another executable (real_nmm.exe) <SPAN class=number></SPAN><SPAN
class=definition>which run correctly both wrf compiled
version.</SPAN><BR><BR>About ulimit, I just set "ulimit -s unlimited" for each
machine:<BR>infact I put this condition on them .bash_profile and also<BR>I
verified it with ulimit -a for each pc.<BR><BR>So, <B>it seems only an "WRF
Intel version <-> Mpi" problem..</B><BR>For this reason I'm going very
crazy! :-(<BR><BR>Thank you!<BR>Fabio.<BR><BR><BR>
<DIV class=gmail_quote>2010/9/2 Gus Correa <SPAN dir=ltr><<A
href="mailto:gus@ldeo.columbia.edu">gus@ldeo.columbia.edu</A>></SPAN><BR>
<BLOCKQUOTE
style="BORDER-LEFT: rgb(204,204,204) 1px solid; MARGIN: 0pt 0pt 0pt 0.8ex; PADDING-LEFT: 1ex"
class=gmail_quote>Hi Fabio<BR><BR>Besides Rajeev's suggestion.<BR><BR>You
mentioned some "error stack" and that to<BR>run on one node successfully you
did "ulimit -s unlimited".<BR>This may be needed on *all nodes* running
WRF.<BR>It doesn't propagate to the other nodes, if you do it in the command
line, or if you put the ulimit command in your job script, for
instance.<BR>It can be done in the resource manager (Torque, SGE, SLURM)
startup<BR>script, or in the Linux limit configuration files.<BR>Maybe your
system administrator can help you with this.<BR><BR>FYI, a number of large
atmosphere/ocean/climate models we<BR>run produce a large program stack,
often times larger than the<BR>default limit set by Linux, and do require
the change above on<BR>all nodes.<BR>(I haven't run WRF,
though.)<BR><BR>Also, to sort out if your network has a problem, you may
want to try<BR>something simpler than WRF.<BR>The cpi.c program in the
MPICH2 'examples' directory is good for this.<BR>Compile with mpicc, run
with mpirun on all nodes.<BR><BR>I hope it helps.<BR><FONT color=#888888>Gus
Correa<BR>---------------------------------------------------------------------<BR>Gustavo
Correa<BR>Lamont-Doherty Earth Observatory - Columbia
University<BR>Palisades, NY, 10964-8000 -
USA<BR>---------------------------------------------------------------------</FONT>
<DIV>
<DIV></DIV>
<DIV class=h5><BR><BR><BR>Rajeev Thakur wrote:<BR>
<BLOCKQUOTE
style="BORDER-LEFT: rgb(204,204,204) 1px solid; MARGIN: 0pt 0pt 0pt 0.8ex; PADDING-LEFT: 1ex"
class=gmail_quote>Try running the cpi example from the MPICH2 examples
directory across two machines. There could be a connection issue between
the two machines.<BR><BR>Rajeev<BR><BR>On Sep 2, 2010, at 8:20 AM, Fabio
F.Gervasi wrote:<BR><BR>
<BLOCKQUOTE
style="BORDER-LEFT: rgb(204,204,204) 1px solid; MARGIN: 0pt 0pt 0pt 0.8ex; PADDING-LEFT: 1ex"
class=gmail_quote>Hi,<BR><BR>I have a "strange" MPI problem when I run a
WRF-NMM model, compiled with Intel v11.1.072 (by GNU run
ok!).<BR>Mpich2-1.2 also is compiled by Intel.<BR><BR>If I run on a
single Quad-core machine everything is ok, but when I try on two or more
Quad-core machine,<BR>initially the wrf.exe processes seem start on
every pc, but after few second wrf stop and I get the error:<BR>Fatal
error in MPI_Allreduce other mpi error error stack.. and so
on...<BR><BR>I just set: "ulimit -s unlimited", otherwise wrf crash also
with a single machine...<BR><BR>This probably is an MPI problem, but how
can I fix it?<BR><BR>Thank you very
much<BR>Fabio.<BR>_______________________________________________<BR>mpich-discuss
mailing list<BR><A href="mailto:mpich-discuss@mcs.anl.gov"
target=_blank>mpich-discuss@mcs.anl.gov</A><BR><A
href="https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss"
target=_blank>https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss</A><BR></BLOCKQUOTE><BR>_______________________________________________<BR>mpich-discuss
mailing list<BR><A href="mailto:mpich-discuss@mcs.anl.gov"
target=_blank>mpich-discuss@mcs.anl.gov</A><BR><A
href="https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss"
target=_blank>https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss</A><BR></BLOCKQUOTE><BR>_______________________________________________<BR>mpich-discuss
mailing list<BR><A href="mailto:mpich-discuss@mcs.anl.gov"
target=_blank>mpich-discuss@mcs.anl.gov</A><BR><A
href="https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss"
target=_blank>https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss</A><BR></DIV></DIV></BLOCKQUOTE></DIV><BR></BLOCKQUOTE></BODY></HTML>