On Thu, Oct 6, 2011 at 10:27 AM, Rongliang Chen <span dir="ltr"><<a href="mailto:rongliang.chan@gmail.com">rongliang.chan@gmail.com</a>></span> wrote:<br><div class="gmail_quote"><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex;">
<div class="im"> That is an overflow somewhere. You can probably get the right answer by<br></div><div class="gmail_quote"><div class="im"><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
using -snes_view. I will try and track down this overflow.<br>
<br>
Matt<br>
<br>
<br></blockquote><div> </div></div><div>Hi Matt,<br><br>Thank you for your reply. <br>The -snes_view and -log_summary output is followed. But I did not find any unusual results in the -snes_view output.<br></div></div></blockquote>
<div><br></div><div>Yes, there was no overflow for individual matrices, so this output is correct. The -log_summary output is for all</div><div>matrices, and that is the problem.</div><div> </div><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex;">
<div class="gmail_quote"><div>I have another question. what the number 23 mean in the following -log_summary output:<div class="im"><br>
" Object Type Creations Destructions Memory Descendants' Mem.<br>Reports information only for process 0.<br><br>--- Event Stage 0: Main Stage<br><br></div> Matrix 23 23 18446744073922240512 0 "<br>
<br>Does it mean that I created 23 matrices in my code? But I think I have not created so many <br>matrices. Thanks.<br></div></div></blockquote><div><br></div><div>You might have created them, but we are counting all matrices, for instance those created for</div>
<div>subdomain preconditioning.</div><div><br></div><div> Matt</div><div> </div><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex;"><div class="gmail_quote"><div>Best,<br>
Rongliang<div class="im"><br><br>------------------------------------------------------------------------------------------------------------<br>
Starting to load grid...<br> Nodes on moving boundary: coarse 199, fine 799, Gridratio 0.250000.<br>Setupping Interpolation matrix......<br></div>Interpolation matrix done......Time spent: 0.207468 <br><div class="im">finished.<br>
Grid has 32000 elements, 1096658 degrees of freedom.<br>
Coarse grid has 2000 elements, 70170 degrees of freedom.<br></div> [0] has 9234 degrees of freedom (matrix), 9234 degrees of freedom (including shared points).<br> [0] coarse grid has 484 degrees of freedom (matrix), 484 degrees of freedom (including shared points).<br>
[127] has 7876 degrees of freedom (matrix), 9100 degrees of freedom (including shared points).<br> [127] coarse grid has 588 degrees of freedom (matrix), 912 degrees of freedom (including shared points).<br>Time spend on the load grid and create matrix etc.: 3.719866.<div class="im">
<br>
Solving Shape Optimization problem (steady-state problem) <br>Solving coarse problem......<br></div> 0 SNES norm 3.4998054301e+01, 0 KSP its last norm 0.0000000000e+00.<br> 1 SNES norm 3.1501179205e+01, 34 KSP its last norm 3.3927102450e-01.<br>
2 SNES norm 2.1246874435e+01, 57 KSP its last norm 3.1177722630e-01.<br> 3 SNES norm 1.7390263296e+01, 141 KSP its last norm 1.9452289323e-01.<br> 4 SNES norm 1.1644760718e+01, 160 KSP its last norm 1.6835316856e-01.<br>
5 SNES norm 1.0601030093e+01, 181 KSP its last norm 1.1003156828e-01.<br> 6 SNES norm 1.0145938759e+00, 126 KSP its last norm 1.0556059459e-01.<br> 7 SNES norm 1.9267547420e-01, 203 KSP its last norm 9.9489004947e-03.<br>
8 SNES norm 1.8901340973e-03, 195 KSP its last norm 1.8359299380e-03.<br>Coarse solver done......<br>Optimized value of object function (Energy dissipation) (Coarse): 29.9743671231<br>The reduction of the energy dissipation (Coarse): -inf% <br>
The optimized curve (Coarse):<br> a = (4.500000, -0.042961, -0.002068, 0.043750, -0.018783, 0.001816) <br>Solving moving mesh equation......<br> KSP norm 2.2906632201e-07, KSP its. 741. Time spent 2.772948 <br>Moving mesh solver done.<br>
0 SNES norm 4.7914118974e+02, 0 KSP its last norm 0.0000000000e+00.<br> 1 SNES norm 1.0150289152e+02, 63 KSP its last norm 4.6576374323e-01.<br> 2 SNES norm 1.8326417396e+00, 90 KSP its last norm 9.9707541310e-02.<br>
3 SNES norm 3.7711809663e-03, 348 KSP its last norm 1.8059473115e-03.<br> 4 SNES norm 9.7342448527e-06, 484 KSP its last norm 3.6343704479e-06.<br>SNES Object: 128 MPI processes<br> type: ls<br> line search variant: SNESLineSearchCubic<br>
alpha=1.000000000000e-04, maxstep=1.000000000000e+08, minlambda=1.000000000000e-12<br> maximum iterations=20, maximum function evaluations=10000<br> tolerances: relative=1e-06, absolute=1e-10, solution=1e-08<br> total number of linear solver iterations=985<br>
total number of function evaluations=5<br> KSP Object: 128 MPI processes <br> type: gmres<br> GMRES: restart=600, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement<br>
GMRES: happy breakdown tolerance 1e-30<br>
maximum iterations=3000, initial guess is zero<br> tolerances: relative=0.001, absolute=1e-08, divergence=10000<br> right preconditioning<br> using UNPRECONDITIONED norm type for convergence test<br> PC Object: 128 MPI processes <br>
type: asm<br> Additive Schwarz: total subdomain blocks = 128, user-defined overlap<br> Additive Schwarz: restriction/interpolation type - BASIC<br> Local solve is same for all blocks, in the following KSP and PC objects:<br>
KSP Object: (sub_) 1 MPI processes <br> type: preonly<br> maximum iterations=10000, initial guess is zero<br> tolerances: relative=1e-05, absolute=1e-50, divergence=10000<br> left preconditioning<br>
using NONE norm type for convergence test<br> PC Object: (sub_) 1 MPI processes <br> type: lu<br> LU: out-of-place factorization<br> tolerance for zero pivot 1e-12<br> using diagonal shift to prevent zero pivot<br>
matrix ordering: qmd<br> factor fill ratio given 5, needed 5.26731<br> Factored matrix follows:<br> Matrix Object: 1 MPI processes <br> type: seqaij<br>
rows=25170, cols=25170<br> package used to perform factorization: petsc<br> total: nonzeros=11090981, allocated nonzeros=11090981<br> total number of mallocs used during MatSetValues calls =0<br>
using I-node routines: found 12872 nodes, limit used is 5<br> linear system matrix = precond matrix:<br> Matrix Object: 1 MPI processes <br> type: seqaij<br> rows=25170, cols=25170<br>
total: nonzeros=2105626, allocated nonzeros=2105626<br> total number of mallocs used during MatSetValues calls =0<br> using I-node routines: found 13453 nodes, limit used is 5<br> linear system matrix = precond matrix:<br>
Matrix Object: 128 MPI processes <br> type: mpiaij<br> rows=1096658, cols=1096658<br> total: nonzeros=94170314, allocated nonzeros=223806957<br> total number of mallocs used during MatSetValues calls =6185057<br>
not using I-node (on process 0) routines<br>Optimized value of object function (Energy dissipation) (Fine): 33.2754475059<br>Solution time of 395.289169 sec.<div class="im"><br>Number of unknowns = 1096658<br>Parameters: kinematic viscosity = 0.01<br>
inlet velocity: u = 5, v = 0 <br>Total number of nonlinear iterations = 4<br></div>Total number of linear iterations = 985<br>Average number of linear iterations = 246.250000<br>Time computing: 395.289169 sec, Time outputting: 0.000000 sec.<br>
Time spent in coarse nonlinear solve: 13.134366 sec, 0.033227 fraction of total compute time.<br>The optimized curve (fine):<br> a = (4.500000, -0.046466, -0.001962, 0.045734, -0.019141, 0.001789) <br>The reduction of the energy dissipation (Fine): -inf% <br>
Time spend on fixed mesh solving: 0.013564 <br>Time spend on shape opt. solving: 395.324390 <br><div class="im">Latex command line: <br> np Newton GMRES Time(Total) Time(Coarse) Ratio <br></div> 128 & 4 & 246.25 & 395.29 & 13.13 & 3.3\% <br>
<br>Running finished on: Wed Oct 5 19:02:01 2011<br>Total running time: 395.376442 <br><div class="im">************************************************************************************************************************<br>
*** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document ***<br>
************************************************************************************************************************<br><br>---------------------------------------------- PETSc Performance Summary: ----------------------------------------------<br>
<br></div>./joab on a Janus-nod named node0844 with 128 processors, by ronglian Wed Oct 5 19:02:01 2011<div class="im"><br>Using Petsc Release Version 3.2.0, Patch 1, Mon Sep 12 16:01:51 CDT 2011 <br><br> Max Max/Min Avg Total <br>
</div>
Time (sec): 3.991e+02 1.00013 3.991e+02<br>Objects: 1.066e+03 1.00000 1.066e+03<br>Flops: 7.938e+10 2.52133 5.615e+10 7.187e+12<br>Flops/sec: 1.989e+08 2.52100 1.407e+08 1.801e+10<br>
MPI Messages: 2.724e+05 8.91400 6.158e+04 7.883e+06<br>MPI Message Lengths: 8.340e+08 2.63753 1.025e+04 8.083e+10<br>MPI Reductions: 6.537e+03 1.00000<div class="im"><br><br>Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract)<br>
e.g., VecAXPY() for real vectors of length N --> 2N flops<br> and VecAXPY() for complex vectors of length N --> 8N flops<br><br>Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions --<br>
Avg %Total Avg %Total counts %Total Avg %Total counts %Total <br></div> 0: Main Stage: 3.9910e+02 100.0% 7.1875e+12 100.0% 7.883e+06 100.0% 1.025e+04 100.0% 6.536e+03 100.0% <br>
<div class="im">
<br>------------------------------------------------------------------------------------------------------------------------<br>See the 'Profiling' chapter of the users' manual for details on interpreting output.<br>
Phase summary info:<br> Count: number of times phase was executed<br> Time and Flops: Max - maximum over all processors<br> Ratio - ratio of maximum to minimum over all processors<br> Mess: number of messages sent<br>
Avg. len: average message length<br> Reduct: number of global reductions<br> Global: entire computation<br> Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop().<br> %T - percent time in this phase %F - percent flops in this phase<br>
%M - percent messages in this phase %L - percent message lengths in this phase<br> %R - percent reductions in this phase<br> Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors)<br>
------------------------------------------------------------------------------------------------------------------------<br>Event Count Time (sec) Flops --- Global --- --- Stage --- Total<br>
Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s<br>------------------------------------------------------------------------------------------------------------------------<br>
<br>--- Event Stage 0: Main Stage<br><br></div>MatMult 2879 1.0 4.3698e+0116.7 1.86e+09 1.3 2.1e+06 1.4e+03 0.0e+00 4 3 27 4 0 4 3 27 4 0 4839<br>MatMultTranspose 3 1.0 3.0989e-0226.2 9.81e+05 1.2 2.0e+03 7.3e+02 0.0e+00 0 0 0 0 0 0 0 0 0 0 3646<br>
MatSolve 2860 1.0 7.2956e+01 2.3 3.95e+10 2.5 0.0e+00 0.0e+00 0.0e+00 14 52 0 0 0 14 52 0 0 0 50895<br>MatLUFactorSym 2 1.0 1.3847e+00 4.7 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>
MatLUFactorNum 13 1.0 4.9187e+01 5.8 3.56e+10 4.8 0.0e+00 0.0e+00 0.0e+00 5 33 0 0 0 5 33 0 0 0 48174<br>MatILUFactorSym 1 1.0 3.9380e-03 3.3 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>
MatAssemblyBegin 78 1.0 8.7529e+0129.0 0.00e+00 0.0 4.3e+04 5.2e+04 1.3e+02 11 0 1 3 2 11 0 1 3 2 0<br>MatAssemblyEnd 78 1.0 7.2215e+00 1.0 0.00e+00 0.0 8.5e+03 3.6e+02 1.1e+02 2 0 0 0 2 2 0 0 0 2 0<br>
MatGetRowIJ 3 1.0 4.3902e-02 8.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>MatGetSubMatrice 13 1.0 1.9239e+00 2.9 0.00e+00 0.0 7.9e+04 1.8e+05 5.1e+01 0 0 1 17 1 0 0 1 17 1 0<br>
MatGetOrdering 3 1.0 4.1121e-01 3.5 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 0 0 0 0 0 0 0<br>MatPartitioning 1 1.0 2.2540e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>
MatZeroEntries 32 1.0 3.8607e-02 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>MatView 3 3.0 1.6980e-0323.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>
VecDot 12 1.0 6.8195e-0317.5 8.36e+04 1.2 0.0e+00 0.0e+00 1.2e+01 0 0 0 0 0 0 0 0 0 0 1451<br>VecMDot 2823 1.0 4.1334e+01 7.2 3.82e+09 1.2 0.0e+00 0.0e+00 2.8e+03 4 6 0 0 43 4 6 0 0 43 10682<br>
VecNorm 2888 1.0 2.5551e+00 3.1 3.47e+07 1.2 0.0e+00 0.0e+00 2.9e+03 0 0 0 0 44 0 0 0 0 44 1575<br>VecScale 2860 1.0 2.0850e-02 1.9 1.73e+07 1.2 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 96028<br>
VecCopy 117 1.0 2.1448e-03 2.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>VecSet 5795 1.0 2.2957e-01 2.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>
VecAXPY 116 1.0 1.9181e-03 1.6 1.36e+06 1.2 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 82341<br>VecWAXPY 16 1.0 2.6107e-04 1.5 4.61e+04 1.2 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 21104<br>
VecMAXPY 2860 1.0 5.2077e+00 1.5 3.85e+09 1.2 0.0e+00 0.0e+00 0.0e+00 1 6 0 0 0 1 6 0 0 0 85546<br>VecAssemblyBegin 60 1.0 3.4554e-0110.6 0.00e+00 0.0 1.8e+04 3.4e+02 1.8e+02 0 0 0 0 3 0 0 0 0 3 0<br>
VecAssemblyEnd 60 1.0 1.9860e-04 3.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>VecScatterBegin 8648 1.0 1.1008e+00 3.0 0.00e+00 0.0 7.7e+06 8.4e+03 0.0e+00 0 0 98 80 0 0 0 98 80 0 0<br>
VecScatterEnd 8648 1.0 8.4154e+0135.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 10 0 0 0 0 10 0 0 0 0 0<br>VecReduceArith 4 1.0 2.6989e-04 2.3 4.00e+04 1.2 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 17292<br>
VecReduceComm 2 1.0 2.7108e-04 6.3 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>VecNormalize 2860 1.0 2.5307e+00 3.1 5.17e+07 1.2 0.0e+00 0.0e+00 2.8e+03 0 0 0 0 44 0 0 0 0 44 2370<br>
SNESSolve 2 1.0 3.9251e+02 1.0 7.85e+10 2.6 6.7e+06 9.7e+03 4.7e+03 98 98 84 80 72 98 98 84 80 72 18034<br>SNESLineSearch 12 1.0 3.0610e+00 1.0 1.26e+07 1.2 6.1e+04 1.1e+04 2.9e+02 1 0 1 1 4 1 0 1 1 4 473<br>
SNESFunctionEval 18 1.0 6.5305e+00 1.0 6.01e+06 1.2 6.2e+04 1.3e+04 2.9e+02 2 0 1 1 4 2 0 1 1 4 106<br>SNESJacobianEval 12 1.0 2.4492e+02 1.0 0.00e+00 0.0 2.5e+04 5.9e+04 9.0e+01 61 0 0 2 1 61 0 0 2 1 0<br>
KSPGMRESOrthog 2823 1.0 4.6476e+01 4.3 7.64e+09 1.2 0.0e+00 0.0e+00 2.8e+03 5 12 0 0 43 5 12 0 0 43 19001<br>KSPSetup 26 1.0 1.3622e-02 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>
KSPSolve 13 1.0 1.4371e+02 1.0 7.94e+10 2.5 7.8e+06 1.0e+04 5.8e+03 36100 98 97 88 36100 98 97 88 50001<br>PCSetUp 26 1.0 5.2544e+01 5.3 3.56e+10 4.8 8.5e+04 1.7e+05 9.8e+01 6 33 1 17 1 6 33 1 17 1 45097<br>
PCSetUpOnBlocks 13 1.0 5.0695e+01 5.7 3.56e+10 4.8 0.0e+00 0.0e+00 1.7e+01 6 33 0 0 0 6 33 0 0 0 46742<br>PCApply 2860 1.0 1.1268e+02 1.8 3.95e+10 2.5 5.6e+06 1.1e+04 0.0e+00 20 52 71 76 0 20 52 71 76 0 32953<div class="im">
<br>
------------------------------------------------------------------------------------------------------------------------<br><br>Memory usage is given in bytes:<br><br>Object Type Creations Destructions Memory Descendants' Mem.<br>
Reports information only for process 0.<br><br>--- Event Stage 0: Main Stage<br><br></div> Matrix 23 23 18446744073922240512 0<div class="im"><br> Matrix Partitioning 1 1 640 0<br>
</div>
Index Set 168 168 922496 0<br> IS L to G Mapping 2 2 78872 0<br> Vector 828 828 44121632 0<br> Vector Scatter 23 23 24196 0<div class="im">
<br>
Application Order 2 2 9335968 0<br></div> SNES 3 2 2544 0<br> Krylov Solver 7 6 16141840 0<br> Preconditioner 7 6 5456 0<br>
Viewer 2 1 712 0<br>========================================================================================================================<br>Average time to get PetscTime(): 9.53674e-08<br>
Average time for MPI_Barrier(): 2.36034e-05<br>Average time for zero size MPI_Send(): 2.78279e-06<br>#PETSc Option Table entries:<br>-coarse_ksp_rtol 1.0e-1<br>-coarsegrid /scratch/stmp00/ronglian/input/Cannula/Cannula_Nest2_E2000_N8241_D70170.fsi<br>
-f /scratch/stmp00/ronglian/input/Cannula/Cannula_Nest2_E32000_N128961_D1096650.fsi<br>-geometric_asm<br>-geometric_asm_overlap 8<br>-inletu 5.0<br>-ksp_atol 1e-8<br>-ksp_gmres_restart 600<br>-ksp_max_it 3000<br>-ksp_pc_side right<br>
-ksp_rtol 1.e-3<br>-ksp_type gmres<br>-log_summary<br>-mat_partitioning_type parmetis<br>-nest_geometric_asm_overlap 4<br>-nest_ksp_atol 1e-8<br>-nest_ksp_gmres_restart 800<br>-nest_ksp_max_it 1000<br>-nest_ksp_pc_side right<br>
-nest_ksp_rtol 1.e-2<br>-nest_ksp_type gmres<br>-nest_pc_asm_type basic<br>-nest_pc_type asm<br>-nest_snes_atol 1.e-10<br>-nest_snes_max_it 20<br>-nest_snes_rtol 1.e-4<br>-nest_sub_pc_factor_mat_ordering_type qmd<br>-nest_sub_pc_factor_shift_amount 1e-8<br>
-nest_sub_pc_factor_shift_type nonzero<br>-nest_sub_pc_type lu<br>-nested<br>-noboundaryreduce<br>-pc_asm_type basic<br>-pc_type asm<br>-shapebeta 10.0<br>-snes_atol 1.e-10<br>-snes_max_it 20<br>-snes_rtol 1.e-6<br>-snes_view<br>
-sub_pc_factor_mat_ordering_type qmd<br>-sub_pc_factor_shift_amount 1e-8<br>-sub_pc_factor_shift_type nonzero<br>-sub_pc_type lu<br>-viscosity 0.01<br>#End of PETSc Option Table entries<br>Compiled without FORTRAN kernels<br>
Compiled with full precision matrices (default)<br>sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8<br>Configure run at: Tue Sep 13 13:28:48 2011<br>Configure options: --known-level1-dcache-size=32768 --known-level1-dcache-linesize=32 --known-level1-dcache-assoc=0 --known-memcmp-ok=1 --known-sizeof-char=1 --known-sizeof-void-p=8 --known-sizeof-short=2 --known-sizeof-int=4 --known-sizeof-long=8 --known-sizeof-long-long=8 --known-sizeof-float=4 --known-sizeof-double=8 --known-sizeof-size_t=8 --known-bits-per-byte=8 --known-sizeof-MPI_Comm=8 --known-sizeof-MPI_Fint=4 --known-mpi-long-double=1 --with-batch=1 --with-mpi-shared-libraries=1 --known-mpi-shared-libraries=0 --download-f-blas-lapack=1 --download-hypre=1 --download-superlu=1 --download-parmetis=1 --download-superlu_dist=1 --download-blacs=1 --download-scalapack=1 --download-mumps=1 --with-debugging=0<br>
-----------------------------------------<br>Libraries compiled on Tue Sep 13 13:28:48 2011 on node1367 <br>Machine characteristics: Linux-2.6.18-238.12.1.el5-x86_64-with-redhat-5.6-Tikanga<br>Using PETSc directory: /home/ronglian/soft/petsc-3.2-p1<br>
Using PETSc arch: Janus-nodebug<br>-----------------------------------------<br><br>Using C compiler: mpicc -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS}<br>Using Fortran compiler: mpif90 -Wall -Wno-unused-variable -O ${FOPTFLAGS} ${FFLAGS} <br>
-----------------------------------------<br><br>Using include paths: -I/home/ronglian/soft/petsc-3.2-p1/Janus-nodebug/include -I/home/ronglian/soft/petsc-3.2-p1/include -I/home/ronglian/soft/petsc-3.2-p1/include -I/home/ronglian/soft/petsc-3.2-p1/Janus-nodebug/include -I/curc/tools/free/redhat_5_x86_64/openmpi-1.4.3_ib/include<br>
-----------------------------------------<br><br>Using C linker: mpicc<br>Using Fortran linker: mpif90<br>Using libraries: -Wl,-rpath,/home/ronglian/soft/petsc-3.2-p1/Janus-nodebug/lib -L/home/ronglian/soft/petsc-3.2-p1/Janus-nodebug/lib -lpetsc -lX11 -Wl,-rpath,/home/ronglian/soft/petsc-3.2-p1/Janus-nodebug/lib -L/home/ronglian/soft/petsc-3.2-p1/Janus-nodebug/lib -lsuperlu_dist_2.5 -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lparmetis -lmetis -lHYPRE -lmpi_cxx -lstdc++ -lscalapack -lblacs -lsuperlu_4.2 -lflapack -lfblas -L/curc/tools/free/redhat_5_x86_64/openmpi-1.4.3_ib/lib -L/usr/lib/gcc/x86_64-redhat-linux/4.1.2 -ldl -lmpi -lopen-rte -lopen-pal -lnsl -lutil -lgcc_s -lpthread -lmpi_f90 -lmpi_f77 -lgfortran -lm -lm -lm -lm -lmpi_cxx -lstdc++ -lmpi_cxx -lstdc++ -ldl -lmpi -lopen-rte -lopen-pal -lnsl -lutil -lgcc_s -lpthread -ldl <br>
----------------------------------------- <br></div></div>
</blockquote></div><br><br clear="all"><div><br></div>-- <br>What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead.<br>
-- Norbert Wiener<br>