[MPICH] any way to ask nemesis to turn-off and turn of active polling ?

chong tan chong_guan_tan at yahoo.com
Mon Dec 17 16:48:05 CST 2007


I am running RedHat enterprise 5.  sysctl complains that sched_compat_yield is not known for kernel.

BTW, I run the test, and both master and slaves utilize 100% of CPU.  

any suggestion ?

thanks
tan



----- Original Message ----
From: Darius Buntinas <buntinas at mcs.anl.gov>
To: chong tan <chong_guan_tan at yahoo.com>
Cc: mpich-discuss at mcs.anl.gov
Sent: Monday, December 17, 2007 1:23:46 PM
Subject: Re: [MPICH] any way to ask nemesis to turn-off and turn of active polling ?



On 12/17/2007 01:00 PM, chong tan wrote:
> Thanks,
> I don;t have root access to the box.  I will see if I can ask sys-admin 
> to do it.  I am running
> Linux snowwhite 2.6.18-8.el5 #1 SMP
>  
> DO you know if the broken yield got into this version ? 

I don't know, but you can try the master/slave programs from the 
discussion we had on sched_yield a few months ago:

Master:

int main() {
  while (1) {
    sched_yield();
  }
  return 0;
}

Slave:

int main() {
  while (1);
  return 0;
}

Start 4 slaves first, THEN one master, and check 'top'.  If it shows 
that the master is taking more than 1% or so, you have a kernel with the 
'broken' yield.

> FYI : the 'yield' people said it is not 'broken', it is in fact the 
> 'right yield'.

Maybe, but Linus is on my side :-)

-d

>  
> tan
> 
> 
>  
> ----- Original Message ----
> From: Darius Buntinas <buntinas at mcs.anl.gov>
> To: chong tan <chong_guan_tan at yahoo.com>
> Cc: mpich-discuss at mcs.anl.gov
> Sent: Monday, December 17, 2007 10:50:12 AM
> Subject: Re: [MPICH] any way to ask nemesis to turn-off and turn of 
> active polling ?
> 
> 
> Try setting the processor affinity for the "average" processes (map each
> one to its own processor).  If you have a kernel with the "broken"
> sched_yield implementations, that may not help.
> 
> If you have a "broken" sched_yield implementation, you can try doing
> this as root:
>  sysctl kernel.sched_compat_yield=1
> or
>  echo "1">/proc/sys/kernel/sched_compat_yield
> 
> -d
> 
> 
> On 12/17/2007 11:35 AM, chong tan wrote:
>  > Yes, in a very subtle way which has major impact on performance.   I will
>  > try to decribe it a litle here:
>  > 
>  > system has 32G, total image 35G.  Load is a litle offbalance
>  > mathematically, 4X dual core, running 5 processes.
>  > 4 processes are the same size, each runs on a CPU.  the last process is
>  > very small, about10% of others, run
>  > on a core of one of the CPU.  SO 1 CPU runs 2 procs: average (P1)one and
>  > light one (P2).
>  > 
>  > All proc do first MPI comm in a fixed algorithmic point.  The 'useful'
>  > image is about 29G at that point, and should
>  > fit into the physical memory.  P2 get there in a heart beat, then
>  > others., followed by P1  which took another 60+ minutes
>  > to get there.  If I combine P1 and P2 into 1 process, then I don;t no
>  > see this extra delay.
>  > 
>  > tan
>  >
>  >
>  > 
>  > ----- Original Message ----
>  > From: Darius Buntinas <buntinas at mcs.anl.gov 
> <mailto:buntinas at mcs.anl.gov>>
>  > To: chong tan <chong_guan_tan at yahoo.com 
> <mailto:chong_guan_tan at yahoo.com>>
>  > Cc: mpich-discuss at mcs.anl.gov <mailto:mpich-discuss at mcs.anl.gov>
>  > Sent: Monday, December 17, 2007 8:02:17 AM
>  > Subject: Re: [MPICH] any way to ask nemesis to turn-off and turn of
>  > active polling ?
>  >
>  >
>  > No, there's no way to do that.  Even MPI_Barrier will do active polling.
>  >
>  > Are you having issues where an MPI process that is waiting in a blocking
>  > call is taking CPU time away from other processes?
>  >
>  > -d
>  >
>  > On 12/14/2007 04:53 PM, chong tan wrote:
>  >  > My issue is like this :
>  >  >
>  >  > among all the processess, some will get to the point of first MPI
>  >  > communication points faster than
>  >  > than other.  Is there a way that I tell nemesis to start without doing
>  >  > active polling, and then turn
>  >  > on active polling with some function ?
>  >  >
>  >  > Or should I just use MPI_Barrier() on that ?
>  >  >
>  >  > thanks
>  >  > tan
>  >  >
>  >  >
>  >  > 
> ------------------------------------------------------------------------
>  >  > Be a better friend, newshound, and know-it-all with Yahoo! Mobile. Try
>  >  > it now.
>  >  >
>  > 
> <http://us.rd.yahoo.com/evt=51733/*http://mobile.yahoo.com/;_ylt=Ahu06i62sR8HDtDypao8Wcj9tAcJ 
> 
>  >
>  >  >  >
>  >
>  >
>  > ------------------------------------------------------------------------
>  > Be a better friend, newshound, and know-it-all with Yahoo! Mobile. Try
>  > it now.
>  > 
> <http://us.rd.yahoo.com/evt=51733/*http://mobile.yahoo.com/;_ylt=Ahu06i62sR8HDtDypao8Wcj9tAcJ 
> 
>  >  >
> 
> 
> ------------------------------------------------------------------------
> Be a better friend, newshound, and know-it-all with Yahoo! Mobile. Try 
> it now. 
> <http://us.rd.yahoo.com/evt=51733/*http://mobile.yahoo.com/;_ylt=Ahu06i62sR8HDtDypao8Wcj9tAcJ 
>  >


      ____________________________________________________________________________________
Be a better friend, newshound, and 
know-it-all with Yahoo! Mobile.  Try it now.  http://mobile.yahoo.com/;_ylt=Ahu06i62sR8HDtDypao8Wcj9tAcJ 
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/mpich-discuss/attachments/20071217/39bbedb9/attachment.htm>


More information about the mpich-discuss mailing list