[Swift-devel] Error 521 provider-staging files to PADS nodes

Mihael Hategan hategan at mcs.anl.gov
Wed Mar 9 14:33:05 CST 2011


Funny thing.

I reran this and I don't see the dramatic performance drop when using
multiple nodes any more.

Mihael

On Wed, 2011-01-26 at 11:06 -0800, Mihael Hategan wrote:
> First we ignore wpn, or rather said we calculate the total worker
> throughput (for all 4 concurrent jobs per worker). In any event that
> stays a constant, so when I say 1 worker I mean 1 worker with 4
> concurrent jobs. I'm doing that to remove the job life-cycle latencies
> from the picture, and keep I/O at maximum.
> 
> That said, here's the summary:
> a. 1 worker (clearly on one node): 80 MB/s in/80 out aggregate
> b. 2 workers on the same node: 80 MB/s in/80 out aggregate
> c. 2 workers on different nodes: 20 MB/s in/20 out aggregate
> 
> I ran these a sufficiently large number of times to not believe that the
> difference can be attributed to statistical variation.
> 
> If what you say were true (job scheduled along other jobs on the same
> node), then I believe that (a) would also have 20 MB/s.
> 
> Mihael
> 
> On Wed, 2011-01-26 at 11:02 -0600, Allan Espinosa wrote:
> > Shouldn't we use ppn=4 to guarantee different nodes?
> > 
> > It might be the case that the 3 other cores got assigned to other jobs
> > by PBS. 
> > 
> > -Allan (mobile)
> > On Jan 24, 2011 6:55 PM, "Mihael Hategan" <hategan at mcs.anl.gov> wrote:
> > >
> > > And then here's the funny thing:
> > > 2 workers, 4 wpn.
> > > When running with ppn=2 (so both on the same node):
> > > [IN]: Total transferred: 7.99 GB, current rate: 13.07 MB/s, average
> > > rate: 85.23 MB/s
> > > [OUT] Total transferred: 8 GB, current rate: 42 B/s, average rate:
> > 85.38
> > > MB/s
> > >
> > > Same situation, but with ppn=1 (so the two are on different nodes):
> > > [IN]: Total transferred: 5.83 GB, current rate: 20.79 MB/s, average
> > > rate: 20.31 MB/s
> > > [OUT] Total transferred: 5.97 GB, current rate: 32.01 MB/s, average
> > > rate: 20.8 MB/s
> > >
> > > This, to me, looks fine because it's the opposite of what I'm
> > expecting.
> > > The service itself should see no difference between the two, and I
> > > suspect it doesn't. But something else is going on. Any ideas?
> > >
> > > Mihael
> > >
> > >
> > > On Mon, 2011-01-24 at 01:41 -0800, Mihael Hategan wrote:
> > > > Play with buffer sizes and ye shall be rewarded.
> > > >
> > > > Turns out that setting TCP buffer sizes to obscene numbers, like
> > 2M,
> > > > gives you quite a bit: 70MB/s in + 70MB/s out on average. Those
> > pads
> > > > nodes must have some fast disks (though maybe it's just the
> > cache).
> > > >
> > > > This is with 1 worker and 4wpn. I'm assuming that with many
> > workers, the
> > > > fact that each worker connection has its separate buffer will
> > > > essentially achieve a similar effect. But then there should be an
> > option
> > > > for setting the buffer size.
> > > >
> > > > The numbers are attached. This all goes from head node local disk
> > to
> > > > worker node local disk directly, so there is no nfs. I'd be
> > curious to
> > > > know how that compares, but I am done for the day.
> > > >
> > > > Mihael
> > > >
> > > > On Sun, 2011-01-23 at 13:32 -0800, Mihael Hategan wrote:
> > > > > I'm trying to run tests on pads. The queues aren't quite empty.
> > In the
> > > > > mean time, I committed a bit of a patch to trunk to measure
> > aggregate
> > > > > traffic on TCP channels (those are only used by the workers).
> > You can
> > > > > enable it by setting the "tcp.channel.log.io.performance" system
> > > > > property to "true".
> > > > >
> > > > > Mihael
> > > > >
> > > > > On Wed, 2011-01-19 at 13:12 -0800, Mihael Hategan wrote:
> > > > > > might be due to one of the recent patches.
> > > > > >
> > > > > > you could try to set IOBLOCKSZ to 1 in worker.pl and rerun.
> > > > > >
> > > > > > On Wed, 2011-01-19 at 13:37 -0600, Michael Wilde wrote:
> > > > > > > An interesting observation on the returned output files:
> >  there are exactly 33 files in the output dir from this run: the same
> > as the number of jobs Swift reports as Finished successfully.   But of
> > those 33, the last 4 are only of partial length, and one of the 4 is
> > length zero (see below).
> > > > > > >
> > > > > > > Its surprising and perhaps a bug that the jobs are reported
> > finished before the output file is fully written???
> > > > > > >
> > > > > > > Also this 3-partial plus 1-zero file looks to me like one
> > worker staging op hung (the oldest of the 4 incomplete output files)
> > and then perhaps 3 were cut short when the coaster service data
> > protocol froze?
> > > > > > >
> > > > > > > - Mike
> > > > > > >
> > > > > > > login1$ pwd
> > > > > > > /scratch/local/wilde/lab
> > > > > > > login1$ cd outdir
> > > > > > > login1$ ls -lt | grep 10:48
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0023.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  2686976 Jan 19 10:48 f.0125.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  2621440 Jan 19 10:48 f.0167.out
> > > > > > > -rw-r--r-- 1 wilde ci-users        0 Jan 19 10:48 f.0259.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0336.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0380.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0015.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0204.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0379.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0066.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0221.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0281.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0403.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0142.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0187.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0067.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0081.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0134.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0136.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0146.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0254.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0362.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0312.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0370.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0389.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0027.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0094.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0183.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0363.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0016.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0025.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0429.out
> > > > > > > -rw-r--r-- 1 wilde ci-users  3010301 Jan 19 10:48 f.0239.out
> > > > > > > login1$ ls -lt | grep 10:48 | wc -l
> > > > > > > 33
> > > > > > > login1$
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > ----- Original Message -----
> > > > > > > > Mihael,
> > > > > > > >
> > > > > > > > The following test on pads failed/hung with an error 521
> > from
> > > > > > > > worker.pl:
> > > > > > > >
> > > > > > > > ---
> > > > > > > > sub getFileCBDataInIndirect {
> > > > > > > > ...
> > > > > > > > elsif ($timeout) {
> > > > > > > > queueCmd((nullCB(), "JOBSTATUS", $jobid, FAILED, "521",
> > "Timeout
> > > > > > > > staging in file"));
> > > > > > > > delete($JOBDATA{$jobid});
> > > > > > > > ---
> > > > > > > >
> > > > > > > > single foreach loop, doing 1,000 "mv" commands
> > > > > > > >
> > > > > > > > throttle was 200 jobs to this coaster pool (1 4-node
> > 32-core PBS job):
> > > > > > > >
> > > > > > > > <pool handle="localhost" sysinfo="INTEL32::LINUX">
> > > > > > > > <execution provider="coaster" url=""
> > jobmanager="local:pbs"/>
> > > > > > > > <profile namespace="globus"
> > key="workersPerNode">8</profile>
> > > > > > > > <profile namespace="globus" key="maxTime">3500</profile>
> > > > > > > > <profile namespace="globus" key="slots">1</profile>
> > > > > > > > <profile namespace="globus"
> > key="nodeGranularity">4</profile>
> > > > > > > > <profile namespace="globus" key="maxNodes">4</profile>
> > > > > > > > <profile namespace="globus" key="queue">short</profile>
> > > > > > > > <profile namespace="karajan"
> > key="jobThrottle">2.0</profile>
> > > > > > > > <profile namespace="karajan"
> > key="initialScore">10000</profile>
> > > > > > > > <filesystem provider="local"/>
> > > > > > > >
> > <workdirectory>/scratch/local/wilde/test/swiftwork</workdirectory>
> > > > > > > > <profile namespace="swift"
> > key="stagingMethod">file</profile>
> > > > > > > > <scratch>/scratch/local/wilde/swiftscratch</scratch>
> > > > > > > > </pool>
> > > > > > > >
> > > > > > > > Ran 33 jobs - 1 job over 1 "wave" of 32 and then one or
> > more workers
> > > > > > > > timed out. Note that the hang may have happened earlier,
> > as no new
> > > > > > > > jobs were starting as the jobs in the first wave were
> > finishing.
> > > > > > > >
> > > > > > > > time swift -tc.file tc -sites.file pbscoasters.xml -config
> > cf.ps
> > > > > > > > mvn.swift -n=1000 >& out &
> > > > > > > >
> > > > > > > >
> > > > > > > > The log is in ~wilde/mvn-20110119-0956-s3s8h9h2.log on CI
> > net.
> > > > > > > >
> > > > > > > > Swift stdout showed the following after waiting a while
> > for a 4-node
> > > > > > > > PADS coaster allocation to start:
> > > > > > > >
> > > > > > > > Progress: Selecting site:799 Submitted:201
> > > > > > > > Progress: Selecting site:799 Submitted:201
> > > > > > > > Progress: Selecting site:799 Submitted:200 Active:1
> > > > > > > > Progress: Selecting site:798 Submitted:177 Active:24
> > Finished
> > > > > > > > successfully:1
> > > > > > > > Progress: Selecting site:796 Submitted:172 Active:28
> > Finished
> > > > > > > > successfully:4
> > > > > > > > Progress: Selecting site:792 Submitted:176 Active:24
> > Finished
> > > > > > > > successfully:8
> > > > > > > > Progress: Selecting site:788 Submitted:180 Active:20
> > Finished
> > > > > > > > successfully:12
> > > > > > > > Progress: Selecting site:784 Submitted:184 Active:16
> > Finished
> > > > > > > > successfully:16
> > > > > > > > Progress: Selecting site:780 Submitted:188 Active:12
> > Finished
> > > > > > > > successfully:20
> > > > > > > > Progress: Selecting site:777 Submitted:191 Active:9
> > Finished
> > > > > > > > successfully:23
> > > > > > > > Progress: Selecting site:773 Submitted:195 Active:5
> > Finished
> > > > > > > > successfully:27
> > > > > > > > Progress: Selecting site:770 Submitted:197 Active:3
> > Finished
> > > > > > > > successfully:30
> > > > > > > > Progress: Selecting site:767 Submitted:200 Finished
> > successfully:33
> > > > > > > > Progress: Selecting site:766 Submitted:201 Finished
> > successfully:33
> > > > > > > > Progress: Selecting site:766 Submitted:201 Finished
> > successfully:33
> > > > > > > > Progress: Selecting site:766 Submitted:201 Finished
> > successfully:33
> > > > > > > > Progress: Selecting site:766 Submitted:201 Finished
> > successfully:33
> > > > > > > > Progress: Selecting site:766 Submitted:201 Finished
> > successfully:33
> > > > > > > > Progress: Selecting site:766 Submitted:201 Finished
> > successfully:33
> > > > > > > > Progress: Selecting site:766 Submitted:200 Active:1
> > Finished
> > > > > > > > successfully:33
> > > > > > > > Execution failed:
> > > > > > > > Job failed with an exit code of 521
> > > > > > > > login1$
> > > > > > > > login1$
> > > > > > > > login1$ pwd
> > > > > > > > /scratch/local/wilde/lab
> > > > > > > > login1$ ls -lt | head
> > > > > > > > total 51408
> > > > > > > > -rw-r--r-- 1 wilde ci-users 5043350 Jan 19 10:51
> > > > > > > > mvn-20110119-0956-s3s8h9h2.log
> > > > > > > >
> > > > > > > > (copied to ~wilde)
> > > > > > > >
> > > > > > > > script was:
> > > > > > > >
> > > > > > > > login1$ cat mvn.swift
> > > > > > > > type file;
> > > > > > > >
> > > > > > > > app (file o) mv (file i)
> > > > > > > > {
> > > > > > > > mv @i @o;
> > > > > > > > }
> > > > > > > >
> > > > > > > > file out[]<simple_mapper; location="outdir",
> > > > > > > > prefix="f.",suffix=".out">;
> > > > > > > > foreach j in [1:@toint(@arg("n","1"))] {
> > > > > > > > file data<"data.txt">;
> > > > > > > > out[j] = mv(data);
> > > > > > > > }
> > > > > > > >
> > > > > > > >
> > > > > > > > data.txt was 3MB
> > > > > > > >
> > > > > > > > A look at the outdir gives a clue to where things hung:
> > The files of
> > > > > > > > <= ~3MB from time 10:48 are from this job. Files from
> > 10:39 and
> > > > > > > > earlier are from other manual runs executed on login1,
> > Note that 3 of
> > > > > > > > the 3MB output files have length 0 or <3MB, and were
> > likely in transit
> > > > > > > > back from the worker:
> > > > > > > >
> > > > > > > > -rw-r--r-- 1 wilde ci-users 2686976 Jan 19 10:48
> > f.0125.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 2621440 Jan 19 10:48
> > f.0167.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 0 Jan 19 10:48 f.0259.out
> > > > > > > >
> > > > > > > >
> > > > > > > > login1$ pwd
> > > > > > > > /scratch/local/wilde/lab
> > > > > > > > login1$ cd outdir
> > > > > > > > login1$ ls -lt | head -40
> > > > > > > > total 2772188
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0023.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 2686976 Jan 19 10:48
> > f.0125.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 2621440 Jan 19 10:48
> > f.0167.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 0 Jan 19 10:48 f.0259.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0336.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0380.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0015.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0204.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0379.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0066.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0221.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0281.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0403.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0142.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0187.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0067.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0081.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0134.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0136.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0146.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0254.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0362.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0312.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0370.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0389.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0027.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0094.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0183.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0363.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0016.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0025.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0429.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 3010301 Jan 19 10:48
> > f.0239.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 30103010 Jan 19 10:39
> > f.0024.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 30103010 Jan 19 10:39
> > f.0037.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 30103010 Jan 19 10:39
> > f.0001.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 30103010 Jan 19 10:39
> > f.0042.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 30103010 Jan 19 10:39
> > f.0033.out
> > > > > > > > -rw-r--r-- 1 wilde ci-users 30103010 Jan 19 10:39
> > f.0051.out
> > > > > > > > l
> > > > > > > >
> > > > > > > > - Mike
> > > > > > > >
> > > > > > > >
> > > > > > > > --
> > > > > > > > Michael Wilde
> > > > > > > > Computation Institute, University of Chicago
> > > > > > > > Mathematics and Computer Science Division
> > > > > > > > Argonne National Laboratory
> > > > > > > >
> > > > > > > > _______________________________________________
> > > > > > > > Swift-devel mailing list
> > > > > > > > Swift-devel at ci.uchicago.edu
> > > > > > > > http://mail.ci.uchicago.edu/mailman/listinfo/swift-devel
> > > > > > >
> > > > > >
> > > > > >
> > > > > > _______________________________________________
> > > > > > Swift-devel mailing list
> > > > > > Swift-devel at ci.uchicago.edu
> > > > > > http://mail.ci.uchicago.edu/mailman/listinfo/swift-devel
> > > > >
> > > > >
> > > > > _______________________________________________
> > > > > Swift-devel mailing list
> > > > > Swift-devel at ci.uchicago.edu
> > > > > http://mail.ci.uchicago.edu/mailman/listinfo/swift-devel
> > > >
> > > > _______________________________________________
> > > > Swift-devel mailing list
> > > > Swift-devel at ci.uchicago.edu
> > > > http://mail.ci.uchicago.edu/mailman/listinfo/swift-devel
> > >
> > >
> > > _______________________________________________
> > > Swift-devel mailing list
> > > Swift-devel at ci.uchicago.edu
> > > http://mail.ci.uchicago.edu/mailman/listinfo/swift-devel
> > >
> > 
> > 
> 
> 
> _______________________________________________
> Swift-devel mailing list
> Swift-devel at ci.uchicago.edu
> http://mail.ci.uchicago.edu/mailman/listinfo/swift-devel





More information about the Swift-devel mailing list