What the error? It always shows the error when it cannot pop up the window.<br>
Sounds like a problem with some batch environment being different from the<br>
interactive node. Computer centers are the worst run thing in the world.<br>
<br>
Matt<br><br><div><span class="gmail_quote">On 5/27/06, <b class="gmail_sendername">Randall Mackie</b> <<a href="mailto:randy@geosystem.us">randy@geosystem.us</a>> wrote:</span><blockquote class="gmail_quote" style="border-left: 1px solid rgb(204, 204, 204); margin: 0pt 0pt 0pt 0.8ex; padding-left: 1ex;">
I can't seem to get the debugger to pop up on my screen.<br><br>When I'm logged into the cluster I'm working on, I can<br>type xterm &, and an xterm pops up on my display. So I know<br>I can get something from the remote cluster.
<br><br>Now, when I try this using PETSc, I'm getting the following error<br>message, for example:<br><br>------------------------------------------------------------------------<br>[17]PETSC ERROR: PETSC: Attaching gdb to /home/randy/d3inv/PETSC_V3.3/d3inv_3_3_petsc of pid 3628 on display
<a href="http://24.5.142.138:0">24.5.142.138:0</a>.0 on<br>machine compute-0-23.local<br>------------------------------------------------------------------------<br><br>I'm using this in my command file:<br><br>source ~/.bashrc
<br>time /opt/mpich/intel/bin/mpirun -np 20 -nolocal -machinefile machines \<br> /home/randy/d3inv/PETSC_V3.3/d3inv_3_3_petsc \<br> -start_in_debugger \<br> -debugger_node 1 \<br> -display
<a href="http://24.5.142.138:0">24.5.142.138:0</a>.0 \<br> -em_ksp_type bcgs \<br> -em_sub_pc_type ilu \<br> -em_sub_pc_factor_levels 8 \<br> -em_sub_pc_factor_fill 4 \<br> -em_sub_pc_factor_reuse_ordering \
<br> -em_sub_pc_factor_reuse_fill \<br> -em_sub_pc_factor_mat_ordering_type rcm \<br> -divh_ksp_type cr \<br> -divh_sub_pc_type icc \<br> -ppc_sub_pc_type ilu \<br><< EOF
<br>...<br><br><br>Randy<br><br><br>Matthew Knepley wrote:<br>> 1) Make sure ssh is forwarding X (-Y I think)<br>><br>> 2) -start_in_debugger<br>><br>> 3) -display <your machine>:0.0<br>><br>> should do it.
<br>><br>> Matt<br>><br>> On 5/27/06, *Randall Mackie* <<a href="mailto:randy@geosystem.us">randy@geosystem.us</a><br>> <mailto:<a href="mailto:randy@geosystem.us">randy@geosystem.us</a>>> wrote:
<br>><br>> This is a stupid question, but how do I start in the debugger if I'm<br>> running<br>> on a cluster half-way around the world and I'm working on that cluster<br>> via ssh?<br>>
<br>> Randy<br>><br>><br>> Matthew Knepley wrote:<br>> > The best thing to do here is get a stack trace from the debugger.<br>> From the<br>> > description, it is hard to tell what statement is trying to
<br>> access which<br>> > illegal<br>> > memory.<br>> ><br>> > Matt<br>> ><br>> > On 5/27/06, *Randall Mackie* < <a href="mailto:randy@geosystem.us">
randy@geosystem.us</a><br>> <mailto:<a href="mailto:randy@geosystem.us">randy@geosystem.us</a>><br>> > <mailto:<a href="mailto:randy@geosystem.us">randy@geosystem.us</a> <mailto:<a href="mailto:randy@geosystem.us">
randy@geosystem.us</a>>>> wrote:<br>> ><br>> > In my PETSc based modeling code, I write out intermediate<br>> results to<br>> > a scratch<br>> >
file, and then read them back later. This has worked fine up<br>> until<br>> > today,<br>> > when for a large model, this seems to be causing my program<br>> to crash<br>> > with
<br>> > errors like:<br>> ><br>> ><br>> ------------------------------------------------------------------------<br>> > [9]PETSC ERROR: Caught signal number 11 SEGV: Segmentation
<br>> > Violation, probably memory access out of range<br>> ><br>> ><br>> > I've tracked down the offending code to:<br>> ><br>> > IF
(rank == 0) THEN<br>> > irec=(iper-1)*2+ipol<br>> > write(7,rec=irec)
(xvec(i),i=1,np)<br>> > END
IF<br>> ><br>> >
It writes out xvec for the first record, but then on the second<br>> > record my program is crashing.<br>> ><br>> >
The record length (from an inquire statement) is recl<br>> 22626552<br>> ><br>> >
The size of the scratch file when my program crashes is 98M.<br>> ><br>> >
PETSc is compiled using the intel compilers ( v9.0 for fortran),<br>> >
and the users manual says that you can have record lengths of<br>> > up to 2 billion bytes.<br>> ><br>> > I'm kind of stuck as to what might be the cause. Any ideas<br>> from anyone
<br>> > would be greatly appreciated.<br>> ><br>> > Randy Mackie<br>> ><br>> > ps. I've tried both the optimized and debugging versions of<br>> the PETSc
<br>> > libraries, with the same result.<br>> ><br>> ><br>> > --<br>> > Randall Mackie<br>> > GSY-USA, Inc.<br>> > PMB# 643<br>
> > 2261 Market St.,<br>> > San Francisco, CA 94114-1600<br>> > Tel (415) 469-8649<br>> > Fax (415) 469-5044<br>> ><br>> > California Registered Geophysicist
<br>> > License No. GP 1034<br>> ><br>> ><br>> ><br>> ><br>> > --<br>> > "Failure has a thousand explanations. Success doesn't need one"
<br>> -- Sir<br>> > Alec Guiness<br>><br>> --<br>> Randall Mackie<br>> GSY-USA, Inc.<br>> PMB# 643<br>> 2261 Market St.,<br>> San Francisco, CA 94114-1600<br>
> Tel (415) 469-8649<br>> Fax (415) 469-5044<br>><br>> California Registered Geophysicist<br>> License No. GP 1034<br>><br>><br>><br>><br>> --<br>> "Failure has a thousand explanations. Success doesn't need one" -- Sir
<br>> Alec Guiness<br><br>--<br>Randall Mackie<br>GSY-USA, Inc.<br>PMB# 643<br>2261 Market St.,<br>San Francisco, CA 94114-1600<br>Tel (415) 469-8649<br>Fax (415) 469-5044<br><br>California Registered Geophysicist<br>License No. GP 1034
<br><br></blockquote></div><br><br clear="all"><br>-- <br>"Failure has a thousand explanations. Success doesn't need one" -- Sir Alec Guiness