[mpich-discuss] mpiexec Hangs, cannot run examples

Rajeev Thakur thakur at mcs.anl.gov
Sat May 7 20:47:50 CDT 2011


You don't need to run "mpd&" any more. (Hydra doesn't need it.) First see if you can run cpi with multiple processes on a single machine.

Rajeev


On May 7, 2011, at 1:12 AM, Michael Morrison wrote:

> Hey all, I'm new to MPI so if anything doesn't make sense let me know and I'll do my best to clarify. I'm attempting to set up MPI on an Amazon EC2 machine. I've gone through and followed the instructions in the instruction manual but whenever I try to run one of the built in examples (examples/cpi), the call simply hangs. Below is some information about my system, my MPI set up and the output returned from mpiexec. Right now i'm not even trying to run anything on a cluster, I simply want to run an MPI application on one machine. Let me know what other information I can furnish to help troubleshoot this problem. Thanks in advance for any help. Mike
> 
> 1) First I ran mpd& to start the process manager
> 
> 2) mpiexec -info yields the following information
> 
> HYDRA build details:
>     Version:                                 1.4rc2
>     Release Date:                            Thu Apr  7 11:17:20 CDT 2011
>     CC:                              gcc    
>     CXX:                             c++    
>     F77:                             gfortran   
>     F90:                             f95   
>     Configure options:                       '--enable-timing=log_detailed' '--with-logging=rlog' '--enable-timer-type=linux86_cycle' '--enable-totalview' '--with-pm=hydra:mpd:gforker' '--with-python=/root/Python-2.7.1/python' '--prefix=/root/mpi' '--disable-option-checking' 'CC=gcc' 'CFLAGS= -O2' 'LDFLAGS= ' 'LIBS=-lrt -lpthread ' 'CPPFLAGS= -I/usr/local/mpich2-1.4rc2/src/mpl/include -I/usr/local/mpich2-1.4rc2/src/mpl/include -I/usr/local/mpich2-1.4rc2/src/openpa/src -I/usr/local/mpich2-1.4rc2/src/openpa/src -I/usr/local/mpich2-1.4rc2/src/mpid/ch3/include -I/usr/local/mpich2-1.4rc2/src/mpid/ch3/include -I/usr/local/mpich2-1.4rc2/src/mpid/common/datatype -I/usr/local/mpich2-1.4rc2/src/mpid/common/datatype -I/usr/local/mpich2-1.4rc2/src/mpid/common/locks -I/usr/local/mpich2-1.4rc2/src/mpid/common/locks -I/usr/local/mpich2-1.4rc2/src/mpid/ch3/channels/nemesis/include -I/usr/local/mpich2-1.4rc2/src/mpid/ch3/channels/nemesis/include -I/usr/local/mpich2-1.4rc2/src/mpid/ch3/channels/nemesis/nemesis/include -I/usr/local/mpich2-1.4rc2/src/mpid/ch3/channels/nemesis/nemesis/include -I/usr/local/mpich2-1.4rc2/src/mpid/ch3/channels/nemesis/nemesis/utils/monitor -I/usr/local/mpich2-1.4rc2/src/mpid/ch3/channels/nemesis/nemesis/utils/monitor -I/usr/local/mpich2-1.4rc2/src/util/wrappers -I/usr/local/mpich2-1.4rc2/src/util/wrappers'
>     Process Manager:                         pmi
>     Launchers available:                     ssh rsh fork slurm ll lsf sge none persist
>     Binding libraries available:             hwloc plpa
>     Resource management kernels available:   none slurm ll lsf sge pbs
>     Checkpointing libraries available:       
>     Demux engines available:                 poll select
> 
> 3) Next I run mpiexec -verbose -profile -n 1 /usr/local/mpich2-1.4rc2/examples/cpi the follow is output and then the program just hangs
> 
> host: localhost
> 
> ==================================================================================================
> mpiexec options:
> ----------------
>   Base path: /root/mpi/bin/
>   Launcher: (null)
>   Debug level: 1
>   Enable X: -1
> 
>   Global environment:
>   -------------------
>     HOSTNAME=domU-12-31-39-0E-D4-53
>     TERM=xterm
>     SHELL=/bin/bash
>     HISTSIZE=1000
>     SSH_CLIENT=76.91.23.64 65438 22
>     SSH_TTY=/dev/pts/0
>     USER=root
>     LD_LIBRARY_PATH=/root/mpi/lib:/root/mpi/lib/trace_rlog
>     LS_COLORS=no=00:fi=00:di=00;34:ln=00;36:pi=40;33:so=00;35:bd=40;33;01:cd=40;33;01:or=01;05;37;41:mi=01;05;37;41:ex=00;32:*.cmd=00;32:*.exe=00;32:*.com=00;32:*.btm=00;32:*.bat=00;32:*.sh=00;32:*.csh=00;32:*.tar=00;31:*.tgz=00;31:*.arj=00;31:*.taz=00;31:*.lzh=00;31:*.zip=00;31:*.z=00;31:*.Z=00;31:*.gz=00;31:*.bz2=00;31:*.bz=00;31:*.tz=00;31:*.rpm=00;31:*.cpio=00;31:*.jpg=00;35:*.gif=00;35:*.bmp=00;35:*.xbm=00;35:*.xpm=00;35:*.png=00;35:*.tif=00;35:
>     EC2_HOME=/home/ec2
>     MAIL=/var/spool/mail/root
>     PATH=/root/Python-2.7.1:/root/mpi/bin:/usr/kerberos/sbin:/usr/kerberos/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin:/home/ec2/bin:/home/ec2/bin:/home/ec2/bin:/root/bin
>     INPUTRC=/etc/inputrc
>     PWD=/usr/local/mpich2-1.4rc2/examples
>     JAVA_HOME=/usr/java/default
>     LANG=en_US.UTF-8
>     SSH_ASKPASS=/usr/libexec/openssh/gnome-ssh-askpass
>     SHLVL=1
>     HOME=/root
>     LOGNAME=root
>     CVS_RSH=ssh
>     SSH_CONNECTION=76.91.23.64 65438 10.192.215.161 22
>     PKG_CONFIG_PATH=/usr/lib/pkgconfig:/usr/local/lib/pkgconfig
>     LESSOPEN=|/usr/bin/lesspipe.sh %s
>     G_BROKEN_FILENAMES=1
>     _=/root/mpi/bin/mpiexec
>     OLDPWD=/root
> 
>   Hydra internal environment:
>   ---------------------------
>     GFORTRAN_UNBUFFERED_PRECONNECTED=y
> 
> 
>     Proxy information:
>     *********************
>       [1] proxy: localhost (1 cores)
>       Exec list: /usr/local/mpich2-1.4rc2/examples/cpi (1 processes); 
> 
> 
> ==================================================================================================
> 
> [mpiexec at domU-12-31-39-0E-D4-53] Timeout set to -1 (-1 means infinite)
> [mpiexec at domU-12-31-39-0E-D4-53] Got a control port string of domU-12-31-39-0E-D4-53:44113
> 
> Proxy launch args: /root/mpi/bin/hydra_pmi_proxy --control-port domU-12-31-39-0E-D4-53:44113 --debug --demux poll --pgid 0 --proxy-id 
> 
> [mpiexec at domU-12-31-39-0E-D4-53] PMI FD: (null); PMI PORT: (null); PMI ID/RANK: -1
> Arguments being passed to proxy 0:
> --version 1.4rc2 --interface-env-name MPICH_INTERFACE_HOSTNAME --hostname localhost --global-core-map 0,1,0 --filler-process-map 0,1,0 --global-process-count 1 --auto-cleanup 1 --pmi-rank -1 --pmi-kvsname kvs_1549_0 --pmi-process-mapping (vector,(0,1,1)) --ckpoint-num -1 --global-inherited-env 27 'HOSTNAME=domU-12-31-39-0E-D4-53' 'TERM=xterm' 'SHELL=/bin/bash' 'HISTSIZE=1000' 'SSH_CLIENT=76.91.23.64 65438 22' 'SSH_TTY=/dev/pts/0' 'USER=root' 'LD_LIBRARY_PATH=/root/mpi/lib:/root/mpi/lib/trace_rlog' 'LS_COLORS=no=00:fi=00:di=00;34:ln=00;36:pi=40;33:so=00;35:bd=40;33;01:cd=40;33;01:or=01;05;37;41:mi=01;05;37;41:ex=00;32:*.cmd=00;32:*.exe=00;32:*.com=00;32:*.btm=00;32:*.bat=00;32:*.sh=00;32:*.csh=00;32:*.tar=00;31:*.tgz=00;31:*.arj=00;31:*.taz=00;31:*.lzh=00;31:*.zip=00;31:*.z=00;31:*.Z=00;31:*.gz=00;31:*.bz2=00;31:*.bz=00;31:*.tz=00;31:*.rpm=00;31:*.cpio=00;31:*.jpg=00;35:*.gif=00;35:*.bmp=00;35:*.xbm=00;35:*.xpm=00;35:*.png=00;35:*.tif=00;35:' 'EC2_HOME=/home/ec2' 'MAIL=/var/spool/mail/root' 'PATH=/root/Python-2.7.1:/root/mpi/bin:/usr/kerberos/sbin:/usr/kerberos/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin:/home/ec2/bin:/home/ec2/bin:/home/ec2/bin:/root/bin' 'INPUTRC=/etc/inputrc' 'PWD=/usr/local/mpich2-1.4rc2/examples' 'JAVA_HOME=/usr/java/default' 'LANG=en_US.UTF-8' 'SSH_ASKPASS=/usr/libexec/openssh/gnome-ssh-askpass' 'SHLVL=1' 'HOME=/root' 'LOGNAME=root' 'CVS_RSH=ssh' 'SSH_CONNECTION=76.91.23.64 65438 10.192.215.161 22' 'PKG_CONFIG_PATH=/usr/lib/pkgconfig:/usr/local/lib/pkgconfig' 'LESSOPEN=|/usr/bin/lesspipe.sh %s' 'G_BROKEN_FILENAMES=1' '_=/root/mpi/bin/mpiexec' 'OLDPWD=/root' --global-user-env 0 --global-system-env 1 'GFORTRAN_UNBUFFERED_PRECONNECTED=y' --proxy-core-count 1 --exec --exec-appnum 0 --exec-proc-count 1 --exec-local-env 0 --exec-wdir /usr/local/mpich2-1.4rc2/examples --exec-args 1 /usr/local/mpich2-1.4rc2/examples/cpi 
> 
> [mpiexec at domU-12-31-39-0E-D4-53] Launch arguments: /root/mpi/bin/hydra_pmi_proxy --control-port domU-12-31-39-0E-D4-53:44113 --debug --demux poll --pgid 0 --proxy-id 0 
> 
> 4) Here's some system info
> 
> Linux version 2.6.21.7-2.fc8xen (mockbuild at xenbuilder1.fedora.redhat.com) (gcc version 4.1.2 20070925 (Red Hat 4.1.2-33)) #1 SMP Fri Feb 15 12:39:36 EST 2008
> 
> 
> _______________________________________________
> mpich-discuss mailing list
> mpich-discuss at mcs.anl.gov
> https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss



More information about the mpich-discuss mailing list