Okay, it wasn't clear to me that you were even using MPICH2, since POE often includes IBM-MPI.<div><br></div><div>It appears to me that this is an application failure, rather than anything with MPI. Have you verified that this application runs in serial with the identical input? Have you run e.g. cpi to verify that MPI works?<br>
<br>Jeff<br><br><div class="gmail_quote">On Wed, Dec 14, 2011 at 9:17 AM, <span dir="ltr"><<a href="mailto:aiswarya.pawar@gmail.com">aiswarya.pawar@gmail.com</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
<u></u><div><div class="im">Am using mipch2-1.2.7 version.<div>Sent from my BlackBerry® on Reliance Mobile, India's No. 1 Network. Go for it!</div><hr></div><div><b>From: </b> aiswarya pawar <<a href="mailto:aiswarya.pawar@gmail.com" target="_blank">aiswarya.pawar@gmail.com</a>>
</div><div><b>Date: </b>Wed, 14 Dec 2011 19:13:17 +0530</div><div><b>To: </b><<a href="mailto:mpich-discuss@mcs.anl.gov" target="_blank">mpich-discuss@mcs.anl.gov</a>></div><div><b>Subject: </b>submission error on IBM cluster</div>
<div><div class="h5"><div><br></div>Hi users,<br><div class="gmail_quote"><br>I have a submission script for gromacs software to be used on IBM cluster, but i get an error while running it. the script goes like this=<br><br>
#!/bin/sh<br># @ error = job1.$(Host).$(Cluster).$(Process).err<br>
# @ output = job1.$(Host).$(Cluster).$(Process).out<br># @ class = ptask32<br># @ job_type = parallel<br># @ node = 1<br># @ tasks_per_node = 4<br># @ queue<br><br>echo "_____________________________________"<br>
echo "LOADL_STEP_ID=$LOADL_STEP_ID"<br>echo "_____________________________________"<br><br>machine_file="/tmp/machinelist.$LOADL_STEP_ID"<br>rm -f $machine_file<br>for node in $LOADL_PROCESSOR_LIST<br>
do<br>echo $node >> $machine_file<br>done<br>machine_count=`cat /tmp/machinelist.$LOADL_STEP_ID|wc -l`<br>echo $machine_count<br>echo MachineList:<br>cat /tmp/machinelist.$LOADL_STEP_ID<br>echo "_____________________________________"<br>
unset LOADLBATCH<br>env |grep LOADLBATCH<br>cd /home/staff/1adf/<br>/usr/bin/poe /home/gromacs-4.5.5/bin/mdrun -deffnm /home/staff/1adf/md -procs $machine_count -hostfile /tmp/machinelist.$LOADL_STEP_ID<br>rm /tmp/machinelist.$LOADL_STEP_ID<br>
<br><br>i get an out file as=<br>_____________________________________<br>LOADL_STEP_ID=cnode39.97541.0<br>_____________________________________<br>4<br>MachineList:<br>cnode62<br>cnode7<br>cnode4<br>cnode8<br>_____________________________________<br>
p0_25108: p4_error: interrupt SIGx: 4<br>p0_2890: p4_error: interrupt SIGx: 4<br>p0_2901: p4_error: interrupt SIGx: 15<br>p0_22760: p4_error: interrupt SIGx: 15<br><br><br>an error file =<br><br>Reading file /home/staff/1adf/md.tpr, VERSION 4.5.4 (single precision)<br>
Sorry couldn't backup /home/staff/1adf/md.log to /home/staff/1adf/#md.log.14#<br><br>Back Off! I just backed up /home/staff/1adf/md.log to /home/staff/1adf/#md.log.14#<br>ERROR: 0031-300 Forcing all remote tasks to exit due to exit code 1 in task 0<br>
<br>Please anyone can help with this error.<br><br>Thanks<br><br><br>
<br>
</div><br>
</div></div></div><br>_______________________________________________<br>
mpich-discuss mailing list <a href="mailto:mpich-discuss@mcs.anl.gov">mpich-discuss@mcs.anl.gov</a><br>
To manage subscription options or unsubscribe:<br>
<a href="https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss" target="_blank">https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss</a><br>
<br></blockquote></div><br><br clear="all"><div><br></div>-- <br>Jeff Hammond<br>Argonne Leadership Computing Facility<br>University of Chicago Computation Institute<br><a href="mailto:jhammond@alcf.anl.gov" target="_blank">jhammond@alcf.anl.gov</a> / (630) 252-5381<br>
<a href="http://www.linkedin.com/in/jeffhammond" target="_blank">http://www.linkedin.com/in/jeffhammond</a><br><a href="https://wiki-old.alcf.anl.gov/index.php/User:Jhammond" target="_blank">https://wiki-old.alcf.anl.gov/index.php/User:Jhammond</a><br>
</div>