[Swift-devel] First tests with swift faster
Lorenzo Pesce
lpesce at uchicago.edu
Wed Feb 20 14:17:15 CST 2013
Trying with a different project. I have no idea here either.
Could not initialize shared directory on pbs
exception @ swift-int.k, line: 380
Caused by: null
Caused by: org.globus.cog.abstraction.impl.common.task.TaskSubmissionException: java.lang.NullPointerException
Caused by: java.lang.NullPointerException
at org.globus.cog.abstraction.impl.common.task.ServiceImpl.hashCode(ServiceImpl.java:171)
at java.util.HashMap.getEntry(HashMap.java:344)
at java.util.HashMap.containsKey(HashMap.java:335)
at org.globus.cog.abstraction.impl.file.FileResourceCache.getResource(FileResourceCache.java:77)
at org.globus.cog.abstraction.impl.file.CachingDelegatedFileOperationHandler.getResource(CachingDelegatedFileOperationHandler.java:75)
at org.globus.cog.abstraction.impl.file.CachingDelegatedFileOperationHandler.submit(CachingDelegatedFileOperationHandler.java:40)
at org.globus.cog.abstraction.impl.common.task.CachingFileOperationTaskHandler.submit(CachingFileOperationTaskHandler.java:28)
at org.globus.cog.karajan.scheduler.submitQueue.NonBlockingSubmit.run(NonBlockingSubmit.java:113)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)
parallelFor @ swift.k, line: 206
Thanks!
Lorenzo
On Feb 19, 2013, at 1:52 PM, Mihael Hategan wrote:
> The way the site catalog is parsed has changed. I committed a fix in
> svn. A workaround is to make sure you specify a url:
>
> ...
> <execution provider="coaster" jobmanager="local:pbs" url="localhost"/>
> ...
> <filesystem provider="local" url="localhost"/>
> ...
>
> The URL is actually required for the coaster provider.
>
> Mihael
>
> On Tue, 2013-02-19 at 13:26 -0600, Lorenzo Pesce wrote:
>> This is the content of the file where we have the first complaint from swift (see attached):
>> <config>
>> <pool handle="pbs">
>> <execution provider="coaster" jobmanager="local:pbs"/>
>> <!-- replace with your project -->
>> <profile namespace="globus" key="project">CI-DEB000002</profile>
>>
>> <profile namespace="globus" key="providerAttributes">pbs.aprun;pbs.mpp;depth=24</profile>
>>
>>
>> <profile namespace="globus" key="jobsPerNode">24</profile>
>> <profile namespace="globus" key="maxTime">172800</profile>
>> <profile namespace="globus" key="maxwalltime">0:10:00</profile>
>> <profile namespace="globus" key="lowOverallocation">100</profile>
>> <profile namespace="globus" key="highOverallocation">100</profile>
>>
>> <profile namespace="globus" key="slots">200</profile>
>> <profile namespace="globus" key="nodeGranularity">1</profile>
>> <profile namespace="globus" key="maxNodes">1</profile>
>>
>> <profile namespace="karajan" key="jobThrottle">47.99</profile>
>> <profile namespace="karajan" key="initialScore">10000</profile>
>>
>> <filesystem provider="local"/>
>> <!-- replace this with your home on lustre -->
>> <workdirectory>/lustre/beagle/samseaver/GS/swift.workdir</workdirectory>
>> </pool>
>> </config>
>>
>> Any ideas?
>>
>> Begin forwarded message:
>>
>>> From: Sam Seaver <samseaver at gmail.com>
>>> Date: February 19, 2013 1:16:28 PM CST
>>> To: Lorenzo Pesce <lpesce at uchicago.edu>
>>> Subject: Re: How are things going?
>>>
>>> I got this error. I suspect using the new SWIFT_HOME directory means that there's possibly a missing parameter someplace:
>>>
>>> should we resume a previous calculation? [y/N] y
>>> rlog files displayed in reverse time order
>>> should I use GS-20130203-0717-jgeppt98.0.rlog ?[y/n]
>>> y
>>> Using GS-20130203-0717-jgeppt98.0.rlog
>>> [Error] GS_sites.xml:1:9: cvc-elt.1: Cannot find the declaration of element 'config'.
>>>
>>> Execution failed:
>>> Failed to parse site catalog
>>> swift:siteCatalog @ scheduler.k, line: 31
>>> Caused by: Invalid pool entry 'pbs':
>>> swift:siteCatalog @ scheduler.k, line: 31
>>> Caused by: java.lang.IllegalArgumentException: Missing URL
>>> at org.griphyn.vdl.karajan.lib.SiteCatalog.execution(SiteCatalog.java:173)
>>> at org.griphyn.vdl.karajan.lib.SiteCatalog.pool(SiteCatalog.java:100)
>>> at org.griphyn.vdl.karajan.lib.SiteCatalog.buildResources(SiteCatalog.java:60)
>>> at org.griphyn.vdl.karajan.lib.SiteCatalog.function(SiteCatalog.java:48)
>>> at org.globus.cog.karajan.compiled.nodes.functions.AbstractFunction.runBody(AbstractFunction.java:38)
>>> at org.globus.cog.karajan.compiled.nodes.InternalFunction.run(InternalFunction.java:154)
>>> at org.globus.cog.karajan.compiled.nodes.CompoundNode.runChild(CompoundNode.java:87)
>>> at org.globus.cog.karajan.compiled.nodes.InternalFunction.run(InternalFunction.java:147)
>>> at org.globus.cog.karajan.compiled.nodes.CompoundNode.runChild(CompoundNode.java:87)
>>> at org.globus.cog.karajan.compiled.nodes.InternalFunction.run(InternalFunction.java:147)
>>> at org.globus.cog.karajan.compiled.nodes.CompoundNode.runChild(CompoundNode.java:87)
>>> at org.globus.cog.karajan.compiled.nodes.FramedInternalFunction.run(FramedInternalFunction.java:63)
>>> at org.globus.cog.karajan.compiled.nodes.Import.runBody(Import.java:269)
>>> at org.globus.cog.karajan.compiled.nodes.InternalFunction.run(InternalFunction.java:154)
>>> at org.globus.cog.karajan.compiled.nodes.CompoundNode.runChild(CompoundNode.java:87)
>>> at org.globus.cog.karajan.compiled.nodes.FramedInternalFunction.run(FramedInternalFunction.java:63)
>>> at org.globus.cog.karajan.compiled.nodes.Main.run(Main.java:79)
>>> at k.thr.LWThread.run(LWThread.java:243)
>>> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
>>> at java.lang.Thread.run(Thread.java:722)
>>>
>>>
>>> On Tue, Feb 19, 2013 at 1:13 PM, Sam Seaver <samseaver at gmail.com> wrote:
>>> OK, it got to the point where it really did hang. I'm retrying, but with your suggestions. The other three finished fine!
>>>
>>> Progress: time: Tue, 19 Feb 2013 19:08:53 +0000 Selecting site:18147 Submitted:174 Active:96 Failed:2 Finished successfully:132323 Failed but can retry:183
>>> Progress: time: Tue, 19 Feb 2013 19:09:23 +0000 Selecting site:18147 Submitted:174 Active:96 Failed:2 Finished successfully:132323 Failed but can retry:183
>>> Progress: time: Tue, 19 Feb 2013 19:09:53 +0000 Selecting site:18147 Submitted:174 Active:96 Failed:2 Finished successfully:132323 Failed but can retry:183
>>> Progress: time: Tue, 19 Feb 2013 19:10:23 +0000 Selecting site:18147 Submitted:174 Active:96 Failed:2 Finished successfully:132323 Failed but can retry:183
>>> Progress: time: Tue, 19 Feb 2013 19:10:53 +0000 Selecting site:18147 Submitted:174 Active:96 Failed:2 Finished successfully:132323 Failed but can retry:183
>>> Progress: time: Tue, 19 Feb 2013 19:11:23 +0000 Selecting site:18147 Submitted:174 Active:96 Failed:2 Finished successfully:132323 Failed but can retry:183
>>>
>>>
>>> On Tue, Feb 19, 2013 at 8:51 AM, Lorenzo Pesce <lpesce at uchicago.edu> wrote:
>>> Hmm...
>>>
>>> foreach.max.threads=100
>>>
>>> maybe you should increase this number a bit and see what happens.
>>>
>>> Also, I would try to replace
>>>
>>> SWIFT_HOME=/home/wilde/swift/rev/swift-r6151-cog-r3552
>>>
>>> with
>>>
>>> SWIFT_HOME=/soft/swift/fast
>>>
>>> Keep me posted. Let's get this rolling.
>>>
>>> if it doesn't work, I can redo the packing.
>>>
>>>
>>>
>>>
>>> On Feb 19, 2013, at 1:07 AM, Sam Seaver wrote:
>>>
>>>> Actually, the ten agents job does seem to be stuck in a strange loop. It is incrementing the number of jobs that has finished successfully, and at a fast pace, but the number of jobs its starting is decrementing much more slowly, its almost as its repeatedly attempting the same set of parameters multiple times...
>>>>
>>>> I'll see what it's doing in the morning
>>>> S
>>>>
>>>>
>>>> On Tue, Feb 19, 2013 at 1:00 AM, Sam Seaver <samseaver at gmail.com> wrote:
>>>> Seems to have worked overall this time!
>>>>
>>>> I resume four jobs, each were for a different number of agents (10,100,1000,10000) it made it easier for me to decide on the app time. Two of them have already finished i.e.:
>>>>
>>>> Progress: time: Mon, 18 Feb 2013 23:50:12 +0000 Active:4 Checking status:1 Finished in previous run:148098 Finished successfully:37897
>>>> Progress: time: Mon, 18 Feb 2013 23:50:15 +0000 Active:2 Checking status:1 Finished in previous run:148098 Finished successfully:37899
>>>> Final status: Mon, 18 Feb 2013 23:50:15 +0000 Finished in previous run:148098 Finished successfully:37902
>>>>
>>>> and the only one that is showing any failure (50/110000), is the ten agents version which is so short I can understand why, but its still actively trying to run jobs and is actively finishing jobs, so that's good.
>>>>
>>>> Yay!
>>>>
>>>>
>>>>
>>>> On Mon, Feb 18, 2013 at 1:09 PM, Lorenzo Pesce <lpesce at uchicago.edu> wrote:
>>>> Good. Keep me posted, I would really like to solve your problems in running on Beagle this week, I wish that Swift would have been friendlier.
>>>>
>>>> On Feb 18, 2013, at 1:01 PM, Sam Seaver wrote:
>>>>
>>>>> I just resumed the jobs that I'd killed before the system went down, lets see how it does. I always did a mini-review of the data I've got and it seems to be working as expected.
>>>>>
>>>>>
>>>>> On Mon, Feb 18, 2013 at 12:28 PM, Lorenzo Pesce <lpesce at uchicago.edu> wrote:
>>>>> I have lost track a bit of what's up. I am happy to try and go over it with you when you are ready.
>>>>>
>>>>> Some of the problems of swift might have improved with a new version and the new system.
>>>>>
>>>>>
>>>>> On Feb 18, 2013, at 12:22 PM, Sam Seaver wrote:
>>>>>
>>>>>> They're not, I've not looked since Beagle came back up. Will do so later today.
>>>>>> S
>>>>>>
>>>>>>
>>>>>> On Mon, Feb 18, 2013 at 12:20 PM, Lorenzo Pesce <lpesce at uchicago.edu> wrote:
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Postdoctoral Fellow
>>>>>> Mathematics and Computer Science Division
>>>>>> Argonne National Laboratory
>>>>>> 9700 S. Cass Avenue
>>>>>> Argonne, IL 60439
>>>>>>
>>>>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>>>>> samseaver at gmail.com
>>>>>> (773) 796-7144
>>>>>>
>>>>>> "We shall not cease from exploration
>>>>>> And the end of all our exploring
>>>>>> Will be to arrive where we started
>>>>>> And know the place for the first time."
>>>>>> --T. S. Eliot
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Postdoctoral Fellow
>>>>> Mathematics and Computer Science Division
>>>>> Argonne National Laboratory
>>>>> 9700 S. Cass Avenue
>>>>> Argonne, IL 60439
>>>>>
>>>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>>>> samseaver at gmail.com
>>>>> (773) 796-7144
>>>>>
>>>>> "We shall not cease from exploration
>>>>> And the end of all our exploring
>>>>> Will be to arrive where we started
>>>>> And know the place for the first time."
>>>>> --T. S. Eliot
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Postdoctoral Fellow
>>>> Mathematics and Computer Science Division
>>>> Argonne National Laboratory
>>>> 9700 S. Cass Avenue
>>>> Argonne, IL 60439
>>>>
>>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>>> samseaver at gmail.com
>>>> (773) 796-7144
>>>>
>>>> "We shall not cease from exploration
>>>> And the end of all our exploring
>>>> Will be to arrive where we started
>>>> And know the place for the first time."
>>>> --T. S. Eliot
>>>>
>>>>
>>>>
>>>> --
>>>> Postdoctoral Fellow
>>>> Mathematics and Computer Science Division
>>>> Argonne National Laboratory
>>>> 9700 S. Cass Avenue
>>>> Argonne, IL 60439
>>>>
>>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>>> samseaver at gmail.com
>>>> (773) 796-7144
>>>>
>>>> "We shall not cease from exploration
>>>> And the end of all our exploring
>>>> Will be to arrive where we started
>>>> And know the place for the first time."
>>>> --T. S. Eliot
>>>
>>>
>>>
>>>
>>> --
>>> Postdoctoral Fellow
>>> Mathematics and Computer Science Division
>>> Argonne National Laboratory
>>> 9700 S. Cass Avenue
>>> Argonne, IL 60439
>>>
>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>> samseaver at gmail.com
>>> (773) 796-7144
>>>
>>> "We shall not cease from exploration
>>> And the end of all our exploring
>>> Will be to arrive where we started
>>> And know the place for the first time."
>>> --T. S. Eliot
>>>
>>>
>>>
>>> --
>>> Postdoctoral Fellow
>>> Mathematics and Computer Science Division
>>> Argonne National Laboratory
>>> 9700 S. Cass Avenue
>>> Argonne, IL 60439
>>>
>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>> samseaver at gmail.com
>>> (773) 796-7144
>>>
>>> "We shall not cease from exploration
>>> And the end of all our exploring
>>> Will be to arrive where we started
>>> And know the place for the first time."
>>> --T. S. Eliot
>>
>> This is the content of the file where we have the first complaint from
>> swift (see attached):
>> <config>
>> <pool handle="pbs">
>> <execution provider="coaster" jobmanager="local:pbs"/>
>> <!-- replace with your project -->
>> <profile namespace="globus" key="project">CI-DEB000002</profile>
>>
>>
>> <profile namespace="globus"
>> key="providerAttributes">pbs.aprun;pbs.mpp;depth=24</profile>
>>
>>
>>
>>
>> <profile namespace="globus" key="jobsPerNode">24</profile>
>> <profile namespace="globus" key="maxTime">172800</profile>
>> <profile namespace="globus" key="maxwalltime">0:10:00</profile>
>> <profile namespace="globus" key="lowOverallocation">100</profile>
>> <profile namespace="globus" key="highOverallocation">100</profile>
>>
>>
>> <profile namespace="globus" key="slots">200</profile>
>> <profile namespace="globus" key="nodeGranularity">1</profile>
>> <profile namespace="globus" key="maxNodes">1</profile>
>>
>>
>> <profile namespace="karajan" key="jobThrottle">47.99</profile>
>> <profile namespace="karajan" key="initialScore">10000</profile>
>>
>>
>> <filesystem provider="local"/>
>> <!-- replace this with your home on lustre -->
>>
>> <workdirectory>/lustre/beagle/samseaver/GS/swift.workdir</workdirectory>
>> </pool>
>> </config>
>>
>>
>> Any ideas?
>>
>> Begin forwarded message:
>>
>>> From: Sam Seaver <samseaver at gmail.com>
>>>
>>> Date: February 19, 2013 1:16:28 PM CST
>>>
>>> To: Lorenzo Pesce <lpesce at uchicago.edu>
>>>
>>> Subject: Re: How are things going?
>>>
>>>
>>> I got this error. I suspect using the new SWIFT_HOME directory
>>> means that there's possibly a missing parameter someplace:
>>>
>>>
>>> should we resume a previous calculation? [y/N] y
>>> rlog files displayed in reverse time order
>>> should I use GS-20130203-0717-jgeppt98.0.rlog ?[y/n]
>>> y
>>> Using GS-20130203-0717-jgeppt98.0.rlog
>>> [Error] GS_sites.xml:1:9: cvc-elt.1: Cannot find the declaration of
>>> element 'config'.
>>>
>>>
>>> Execution failed:
>>> Failed to parse site catalog
>>> swift:siteCatalog @ scheduler.k, line: 31
>>> Caused by: Invalid pool entry 'pbs':
>>> swift:siteCatalog @ scheduler.k, line: 31
>>> Caused by: java.lang.IllegalArgumentException: Missing URL
>>> at
>>> org.griphyn.vdl.karajan.lib.SiteCatalog.execution(SiteCatalog.java:173)
>>> at
>>> org.griphyn.vdl.karajan.lib.SiteCatalog.pool(SiteCatalog.java:100)
>>> at
>>> org.griphyn.vdl.karajan.lib.SiteCatalog.buildResources(SiteCatalog.java:60)
>>> at
>>> org.griphyn.vdl.karajan.lib.SiteCatalog.function(SiteCatalog.java:48)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.functions.AbstractFunction.runBody(AbstractFunction.java:38)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.InternalFunction.run(InternalFunction.java:154)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.CompoundNode.runChild(CompoundNode.java:87)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.InternalFunction.run(InternalFunction.java:147)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.CompoundNode.runChild(CompoundNode.java:87)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.InternalFunction.run(InternalFunction.java:147)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.CompoundNode.runChild(CompoundNode.java:87)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.FramedInternalFunction.run(FramedInternalFunction.java:63)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.Import.runBody(Import.java:269)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.InternalFunction.run(InternalFunction.java:154)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.CompoundNode.runChild(CompoundNode.java:87)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.FramedInternalFunction.run(FramedInternalFunction.java:63)
>>> at
>>> org.globus.cog.karajan.compiled.nodes.Main.run(Main.java:79)
>>> at k.thr.LWThread.run(LWThread.java:243)
>>> at
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
>>> at java.util.concurrent.ThreadPoolExecutor
>>> $Worker.run(ThreadPoolExecutor.java:603)
>>> at java.lang.Thread.run(Thread.java:722)
>>>
>>>
>>> On Tue, Feb 19, 2013 at 1:13 PM, Sam Seaver <samseaver at gmail.com>
>>> wrote:
>>> OK, it got to the point where it really did hang. I'm
>>> retrying, but with your suggestions. The other three
>>> finished fine!
>>>
>>>
>>> Progress: time: Tue, 19 Feb 2013 19:08:53 +0000 Selecting
>>> site:18147 Submitted:174 Active:96 Failed:2 Finished
>>> successfully:132323 Failed but can retry:183
>>> Progress: time: Tue, 19 Feb 2013 19:09:23 +0000 Selecting
>>> site:18147 Submitted:174 Active:96 Failed:2 Finished
>>> successfully:132323 Failed but can retry:183
>>> Progress: time: Tue, 19 Feb 2013 19:09:53 +0000 Selecting
>>> site:18147 Submitted:174 Active:96 Failed:2 Finished
>>> successfully:132323 Failed but can retry:183
>>> Progress: time: Tue, 19 Feb 2013 19:10:23 +0000 Selecting
>>> site:18147 Submitted:174 Active:96 Failed:2 Finished
>>> successfully:132323 Failed but can retry:183
>>> Progress: time: Tue, 19 Feb 2013 19:10:53 +0000 Selecting
>>> site:18147 Submitted:174 Active:96 Failed:2 Finished
>>> successfully:132323 Failed but can retry:183
>>> Progress: time: Tue, 19 Feb 2013 19:11:23 +0000 Selecting
>>> site:18147 Submitted:174 Active:96 Failed:2 Finished
>>> successfully:132323 Failed but can retry:183
>>>
>>>
>>> On Tue, Feb 19, 2013 at 8:51 AM, Lorenzo Pesce
>>> <lpesce at uchicago.edu> wrote:
>>> Hmm...
>>>
>>>
>>> foreach.max.threads=100
>>>
>>>
>>> maybe you should increase this number a bit and see
>>> what happens.
>>>
>>>
>>> Also, I would try to replace
>>>
>>>
>>> SWIFT_HOME=/home/wilde/swift/rev/swift-r6151-cog-r3552
>>>
>>>
>>> with
>>>
>>>
>>> SWIFT_HOME=/soft/swift/fast
>>>
>>>
>>> Keep me posted. Let's get this rolling.
>>>
>>>
>>> if it doesn't work, I can redo the packing.
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> On Feb 19, 2013, at 1:07 AM, Sam Seaver wrote:
>>>
>>>> Actually, the ten agents job does seem to be stuck
>>>> in a strange loop. It is incrementing the number
>>>> of jobs that has finished successfully, and at a
>>>> fast pace, but the number of jobs its starting is
>>>> decrementing much more slowly, its almost as its
>>>> repeatedly attempting the same set of parameters
>>>> multiple times...
>>>>
>>>>
>>>> I'll see what it's doing in the morning
>>>> S
>>>>
>>>>
>>>> On Tue, Feb 19, 2013 at 1:00 AM, Sam Seaver
>>>> <samseaver at gmail.com> wrote:
>>>> Seems to have worked overall this time!
>>>>
>>>>
>>>> I resume four jobs, each were for a
>>>> different number of agents
>>>> (10,100,1000,10000) it made it easier for
>>>> me to decide on the app time. Two of them
>>>> have already finished i.e.:
>>>>
>>>>
>>>> Progress: time: Mon, 18 Feb 2013 23:50:12
>>>> +0000 Active:4 Checking status:1
>>>> Finished in previous run:148098 Finished
>>>> successfully:37897
>>>> Progress: time: Mon, 18 Feb 2013 23:50:15
>>>> +0000 Active:2 Checking status:1
>>>> Finished in previous run:148098 Finished
>>>> successfully:37899
>>>> Final status: Mon, 18 Feb 2013 23:50:15
>>>> +0000 Finished in previous run:148098
>>>> Finished successfully:37902
>>>>
>>>>
>>>> and the only one that is showing any
>>>> failure (50/110000), is the ten agents
>>>> version which is so short I can understand
>>>> why, but its still actively trying to run
>>>> jobs and is actively finishing jobs, so
>>>> that's good.
>>>>
>>>>
>>>> Yay!
>>>>
>>>>
>>>>
>>>>
>>>> On Mon, Feb 18, 2013 at 1:09 PM, Lorenzo
>>>> Pesce <lpesce at uchicago.edu> wrote:
>>>> Good. Keep me posted, I would
>>>> really like to solve your problems
>>>> in running on Beagle this week, I
>>>> wish that Swift would have been
>>>> friendlier.
>>>>
>>>> On Feb 18, 2013, at 1:01 PM, Sam
>>>> Seaver wrote:
>>>>
>>>>> I just resumed the jobs that I'd
>>>>> killed before the system went
>>>>> down, lets see how it does. I
>>>>> always did a mini-review of the
>>>>> data I've got and it seems to be
>>>>> working as expected.
>>>>>
>>>>>
>>>>> On Mon, Feb 18, 2013 at 12:28
>>>>> PM, Lorenzo Pesce
>>>>> <lpesce at uchicago.edu> wrote:
>>>>> I have lost track a bit
>>>>> of what's up. I am happy
>>>>> to try and go over it
>>>>> with you when you are
>>>>> ready.
>>>>>
>>>>>
>>>>> Some of the problems of
>>>>> swift might have
>>>>> improved with a new
>>>>> version and the new
>>>>> system.
>>>>>
>>>>>
>>>>>
>>>>> On Feb 18, 2013, at
>>>>> 12:22 PM, Sam Seaver
>>>>> wrote:
>>>>>
>>>>>> They're not, I've not
>>>>>> looked since Beagle
>>>>>> came back up. Will do
>>>>>> so later today.
>>>>>> S
>>>>>>
>>>>>>
>>>>>> On Mon, Feb 18, 2013
>>>>>> at 12:20 PM, Lorenzo
>>>>>> Pesce
>>>>>> <lpesce at uchicago.edu>
>>>>>> wrote:
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Postdoctoral Fellow
>>>>>> Mathematics and
>>>>>> Computer Science
>>>>>> Division
>>>>>> Argonne National
>>>>>> Laboratory
>>>>>> 9700 S. Cass Avenue
>>>>>> Argonne, IL 60439
>>>>>>
>>>>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>>>>> samseaver at gmail.com
>>>>>> (773) 796-7144
>>>>>>
>>>>>> "We shall not cease
>>>>>> from exploration
>>>>>> And the end of all our
>>>>>> exploring
>>>>>> Will be to arrive
>>>>>> where we started
>>>>>> And know the place for
>>>>>> the first time."
>>>>>> --T. S. Eliot
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Postdoctoral Fellow
>>>>> Mathematics and Computer Science
>>>>> Division
>>>>> Argonne National Laboratory
>>>>> 9700 S. Cass Avenue
>>>>> Argonne, IL 60439
>>>>>
>>>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>>>> samseaver at gmail.com
>>>>> (773) 796-7144
>>>>>
>>>>> "We shall not cease from
>>>>> exploration
>>>>> And the end of all our exploring
>>>>> Will be to arrive where we
>>>>> started
>>>>> And know the place for the first
>>>>> time."
>>>>> --T. S. Eliot
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Postdoctoral Fellow
>>>> Mathematics and Computer Science Division
>>>> Argonne National Laboratory
>>>> 9700 S. Cass Avenue
>>>> Argonne, IL 60439
>>>>
>>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>>> samseaver at gmail.com
>>>> (773) 796-7144
>>>>
>>>> "We shall not cease from exploration
>>>> And the end of all our exploring
>>>> Will be to arrive where we started
>>>> And know the place for the first time."
>>>> --T. S. Eliot
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Postdoctoral Fellow
>>>> Mathematics and Computer Science Division
>>>> Argonne National Laboratory
>>>> 9700 S. Cass Avenue
>>>> Argonne, IL 60439
>>>>
>>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>>> samseaver at gmail.com
>>>> (773) 796-7144
>>>>
>>>> "We shall not cease from exploration
>>>> And the end of all our exploring
>>>> Will be to arrive where we started
>>>> And know the place for the first time."
>>>> --T. S. Eliot
>>>
>>>
>>>
>>>
>>>
>>>
>>> --
>>> Postdoctoral Fellow
>>> Mathematics and Computer Science Division
>>> Argonne National Laboratory
>>> 9700 S. Cass Avenue
>>> Argonne, IL 60439
>>>
>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>> samseaver at gmail.com
>>> (773) 796-7144
>>>
>>> "We shall not cease from exploration
>>> And the end of all our exploring
>>> Will be to arrive where we started
>>> And know the place for the first time."
>>> --T. S. Eliot
>>>
>>>
>>>
>>>
>>> --
>>> Postdoctoral Fellow
>>> Mathematics and Computer Science Division
>>> Argonne National Laboratory
>>> 9700 S. Cass Avenue
>>> Argonne, IL 60439
>>>
>>> http://www.linkedin.com/pub/sam-seaver/0/412/168
>>> samseaver at gmail.com
>>> (773) 796-7144
>>>
>>> "We shall not cease from exploration
>>> And the end of all our exploring
>>> Will be to arrive where we started
>>> And know the place for the first time."
>>> --T. S. Eliot
>>
>>
>> _______________________________________________
>> Swift-devel mailing list
>> Swift-devel at ci.uchicago.edu
>> https://lists.ci.uchicago.edu/cgi-bin/mailman/listinfo/swift-devel
>
>
More information about the Swift-devel
mailing list