From bisheshkh at gmail.com Fri Nov 1 00:08:41 2013 From: bisheshkh at gmail.com (Bishesh Khanal) Date: Fri, 1 Nov 2013 06:08:41 +0100 Subject: [petsc-users] using petsc tools to solve isolated irregular domains with finite difference In-Reply-To: References: Message-ID: On Thu, Oct 31, 2013 at 11:34 PM, Matthew Knepley wrote: > On Thu, Oct 31, 2013 at 5:02 PM, Bishesh Khanal wrote: > >> >> >> >> On Thu, Oct 31, 2013 at 7:43 PM, Matthew Knepley wrote: >> >>> On Thu, Oct 31, 2013 at 1:25 PM, Bishesh Khanal wrote: >>> >>>> I did not call DMCreateMatrix() before when using just dmda (and it is >>>> working). In any case, I added a call to DMCreateMatrix() now but still >>>> there are problems. Here are the code snippets and the error I get: >>>> >>> >>> Then you were allowing it to be called automatically by PETSc, and it >>> would have been this time as well. >>> >>> >>>> //Setting up the section and linking it to DM: >>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>> PetscInt point; >>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>> &point);CHKERRQ(ierr); >>>> ierr = PetscSectionSetDof(s, point, testPoisson.mDof); >>>> // No. of dofs associated with the point. >>>> } >>>> } >>>> } >>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>> >>>> //Set up KSP: >>>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>>> ierr = >>>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>> ierr = >>>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>> ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); >>>> ierr = KSPSolve(ksp,NULL,NULL);CHKERRQ(ierr); >>>> >>>> ------------------------------------------------------ >>>> The computeMatrix2dSection function has: >>>> >>>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>>> for(PetscInt j = info.ys; j>>> for(PetscInt i = info.xs; i>>> if (isPosInDomain(ctx,i,j)) { >>>> ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>> ierr = PetscSectionGetDof(gs,point,&dof); >>>> for(PetscInt cDof = 0; cDof < dof; ++cDof) { >>>> num = 0; >>>> row+=cDof; >>>> col[num] = row; //(i,j) position >>>> v[num++] = -4; >>>> if(isPosInDomain(ctx,i+1,j)) { >>>> ierr = >>>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>>> ierr = >>>> PetscSectionGetOffset(gs,point,&col[num]); >>>> col[num] += cDof; >>>> v[num++] = 1; >>>> } >>>> if(isPosInDomain(ctx,i-1,j)) { >>>> ierr = >>>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>>> ierr = >>>> PetscSectionGetOffset(gs,point,&col[num]); >>>> col[num] += cDof; >>>> v[num++] = 1; >>>> } >>>> if(isPosInDomain(ctx,i,j+1)) { >>>> ierr = >>>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>>> ierr = >>>> PetscSectionGetOffset(gs,point,&col[num]); >>>> col[num] += cDof; >>>> v[num++] = 1; >>>> } >>>> if(isPosInDomain(ctx,i,j-1)) { >>>> ierr = >>>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>>> ierr = >>>> PetscSectionGetOffset(gs,point,&col[num]); >>>> col[num] += cDof; >>>> v[num++] = 1; >>>> } >>>> ierr = >>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>> } >>>> } >>>> } >>>> } >>>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>> >>>> But this is not working. For e.g. for a 6X6 grid size I get the >>>> following error: >>>> >>> >>> Okay, here is how to debug this: >>> >>> 0) Go to a single scalar equations to make things easier >>> >>> 1) Use MatView(A, PETSC_VIEWER_STDOUT_WORLD), which should output a 36 >>> row matrix with >>> the preallocated nonzero pattern, filled with zeros >>> >>> 2) Make sure this is the pattern you want >>> >>> 3) Run in the debugger with -start_in_debugger >>> >>> 4) When you get the error, see >>> >>> a) If the (i, j) is one that should be setting a value >>> >>> b) Why this (i, j) was not preallocated >>> >> >> Up to 4 (a), it is correct. There is a problem in the way >> DMDAGetCellPoint and PetscSectionGetOffset works in this case scenario. I >> will try to explain it below for the case of 4X4 grid. >> First case: >> If I set the computational domain to be all the points of the dmda grid, >> (i.e. isPosInDomain(..,i,j,..) returns true for all i and j in the dmda >> grid), the program runs fine and does not give any error. >> >> Second case: >> I want the computational domain to be some part of the whole grid. There >> is a problem in this case. >> The following test is in a case where, >> isPosInDomain(..,i,j,..) returns true ONLY for (i,j) pairs of (2,1) (1,2) >> (2,2) (3,2) and (3,2). The grid with its corresponding point number in the >> petscsection is shown below: >> >> 12 13 14 15 >> 8 9 10 11 >> 4 5 6 7 >> 0 1 2 3 >> >> of which 6, 9, 10, 11 and 14 correspond to the (i,j) points that returns >> true for isPosInDomain(..,i,j,..) >> MatView gives me the 16-row matrix with the star stencil non-zero >> structure as expected. >> The error I get is: new non-zero at (0,2) caused a malloc. >> >> This error is for the value of (i,j) = (2,1), i.e. point 6. >> The loop to set values in the matrix starts as: >> for(PetscInt j = info.ys; j> for(PetscInt i = info.xs; i> if (isPosInDomain(ctx,i,j)) { >> ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >> ierr = PetscSectionGetOffset(gs,point,&row); >> >> Now, what I wanted from the beginning is to create the matrix containing >> the rows corresponding to only those points (i,j) which has true values for >> isPosInDomain(ctx,i,j), that means in this case 5 row matrix. In the >> current test, the first (i,j) that reaches DMDAGetCellPoint is (2,1), which >> gives sets point variable to 6. >> The line PetscSectionGetOffset(gs,point,&row) sets the value of row to 0. >> So I think this where the inconsistency lies. >> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES) expects row variable and >> col[] array to correspond to (i,j) based on DMDA, but PetsSectionGetOffset >> gives result based on how we masked away some of the points from dmda grid. >> Or am I missing something very basic here ? >> > > You are missing something basic. The row is correctly identified as 0, and > that means the 2 is > point 10. > The problem must be preallocation. First, you should send the result of > MatView(). > MatView shows that the preallocation and the matrix created is for the full dmda grid which is probably not what we want here, right ? Here is the matview result: Mat Object: 1 MPI processes type: seqaij row 0: (0, 0) (1, 0) (4, 0) row 1: (0, 0) (1, 0) (2, 0) (5, 0) row 2: (1, 0) (2, 0) (3, 0) (6, 0) row 3: (2, 0) (3, 0) (7, 0) row 4: (0, 0) (4, 0) (5, 0) (8, 0) row 5: (1, 0) (4, 0) (5, 0) (6, 0) (9, 0) row 6: (2, 0) (5, 0) (6, 0) (7, 0) (10, 0) row 7: (3, 0) (6, 0) (7, 0) (11, 0) row 8: (4, 0) (8, 0) (9, 0) (12, 0) row 9: (5, 0) (8, 0) (9, 0) (10, 0) (13, 0) row 10: (6, 0) (9, 0) (10, 0) (11, 0) (14, 0) row 11: (7, 0) (10, 0) (11, 0) (15, 0) row 12: (8, 0) (12, 0) (13, 0) row 13: (9, 0) (12, 0) (13, 0) (14, 0) row 14: (10, 0) (13, 0) (14, 0) (15, 0) row 15: (11, 0) (14, 0) (15, 0) [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Argument out of range! [0]PETSC ERROR: New nonzero at (0,2) caused a malloc! > Matt > > >> >>> Thanks, >>> >>> Matt >>> >>> >>>> [0]PETSC ERROR: --------------------- Error Message >>>> ------------------------------------ >>>> [0]PETSC ERROR: Argument out of range! >>>> [0]PETSC ERROR: New nonzero at (0,6) caused a malloc! >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: Petsc Development GIT revision: >>>> 61e5e40bb2c5bf2423e94b71a15fef47e411b0da GIT Date: 2013-10-25 21:47:45 >>>> -0500 >>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: build/poissonIrregular on a arch-linux2-cxx-debug named >>>> edwards by bkhanal Thu Oct 31 19:23:58 2013 >>>> [0]PETSC ERROR: Libraries linked from >>>> /home/bkhanal/Documents/softwares/petsc/arch-linux2-cxx-debug/lib >>>> [0]PETSC ERROR: Configure run at Sat Oct 26 16:35:15 2013 >>>> [0]PETSC ERROR: Configure options --download-mpich >>>> -download-f-blas-lapack=1 --download-metis --download-parmetis >>>> --download-superlu_dist --download-scalapack --download-mumps >>>> --download-hypre --with-clanguage=cxx >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: MatSetValues_SeqAIJ() line 413 in >>>> src/mat/impls/aij/seq/aij.c >>>> [0]PETSC ERROR: MatSetValues() line 1130 in src/mat/interface/matrix.c >>>> [0]PETSC ERROR: computeMatrix2dSection() line 318 in >>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>> [0]PETSC ERROR: KSPSetUp() line 228 in src/ksp/ksp/interface/itfunc.c >>>> [0]PETSC ERROR: KSPSolve() line 399 in src/ksp/ksp/interface/itfunc.c >>>> [0]PETSC ERROR: main() line 598 in >>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>> [unset]: aborting job: >>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>> >>>> >>>>> Matt >>>>> >>>>> >>>>>> However, now since I do not want the rows in the matrix corresponding >>>>>> to the points in the grid which do not lie in the computational domain. >>>>>> This means the size of the matrix will not necessarily equal the total >>>>>> number of cells in DMDA. So in the following code: >>>>>> >>>>>> ierr = DMDAGetLocalInfo(dm,&info);CHKERRQ(ierr); >>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>> PetscInt point; >>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>> &point);CHKERRQ(ierr); >>>>>> ierr = PetscSectionSetDof(s, point, >>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>> } >>>>>> >>>>>> } >>>>>> } >>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>> >>>>>> should I myself compute proper nC (i.e. total no. of points for which >>>>>> I want the matrix to have a corresponding row) before calling >>>>>> PetscSectionSetChart() or is the masking out of the points inside the loop >>>>>> with if(isPosInDomain(&testPoisson,i,j,0)) sufficient ? >>>>>> And, when you write: >>>>>> PetscSectionGetOffset(gs, p, &ind); >>>>>> row = ind < 0 ? -(ind+1) : ind; >>>>>> >>>>>> it seems you allow the possibility of getting a -ve ind when using >>>>>> PetscSectionGetOffset. When would an offset to a particular dof and point? >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>> >>>>>>>> for(PetscInt j = info.ys; j>>>>>>> for(PetscInt i = info.xs; i>>>>>>> num = 0; >>>>>>>> /*ierr = >>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr);*/ /*Get the PetscPoint*/ >>>>>>>> /*But I did now understand how I would emulate the >>>>>>>> row.c and col.c info with PetscSection*/ >>>>>>>> row.i = i; row.j = j; >>>>>>>> >>>>>>> >>>>>>> Here you would call >>>>>>> >>>>>>> ierr = DMDAGetCellPoint(da, i, j, 0, &cell);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionGetOffset(da, cell, &row);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionGetDof(da, cell, &dof);CHKERRQ(ierr); >>>>>>> >>>>>>> This means that this cell has global rows = [row, row+dof). >>>>>>> >>>>>>> >>>>>>>> col[num].i = i; col[num].j = j; >>>>>>>> if (isPosInDomain(ctx,i,j)) { /*put the >>>>>>>> operator for only the values for only the points lying in the domain*/ >>>>>>>> v[num++] = -4; >>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>> col[num].i = i+1; col[num].j = j; >>>>>>>> v[num++] = 1; >>>>>>>> } >>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>> col[num].i = i-1; col[num].j = j; >>>>>>>> v[num++] = 1; >>>>>>>> } >>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>> col[num].i = i; col[num].j = j+1; >>>>>>>> v[num++] = 1; >>>>>>>> } >>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>> col[num].i = i; col[num].j = j-1; >>>>>>>> v[num++] = 1; >>>>>>>> } >>>>>>>> } else { >>>>>>>> v[num++] = dScale; /*set Dirichlet identity >>>>>>>> rows for points in the rectangle but outside the computational domain*/ >>>>>>>> } >>>>>>>> >>>>>>> >>>>>>> You do the same thing for the columns, and then call >>>>>>> >>>>>>> ierr = MatSetValues(A, dof, rows, num*dof, cols, v, >>>>>>> INSERT_VALUES);CHKERRQ(ierr); >>>>>>> >>>>>>> >>>>>>>> ierr = >>>>>>>> MatSetValuesStencil(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>> } >>>>>>>> } >>>>>>>> } >>>>>>>> >>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> >>>>>>>>> Matt >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>>> Matt >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> Matt >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> If not, just put the identity into >>>>>>>>>>>>>>>>>>> the rows you do not use on the full cube. It will not >>>>>>>>>>>>>>>>>>> hurt scalability or convergence. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> In the case of Poisson with Dirichlet condition this >>>>>>>>>>>>>>>>>> might be the case. But is it always true that having identity rows in the >>>>>>>>>>>>>>>>>> system matrix will not hurt convergence ? I thought otherwise for the >>>>>>>>>>>>>>>>>> following reasons: >>>>>>>>>>>>>>>>>> 1) Having read Jed's answer here : >>>>>>>>>>>>>>>>>> http://scicomp.stackexchange.com/questions/3426/why-is-pinning-a-point-to-remove-a-null-space-bad/3427#3427 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Jed is talking about a constraint on a the pressure at a >>>>>>>>>>>>>>>>> point. This is just decoupling these unknowns from the rest >>>>>>>>>>>>>>>>> of the problem. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> 2) Some observation I am getting (but I am still doing >>>>>>>>>>>>>>>>>> more experiments to confirm) while solving my staggered-grid 3D stokes flow >>>>>>>>>>>>>>>>>> with schur complement and using -pc_type gamg for A00 matrix. Putting the >>>>>>>>>>>>>>>>>> identity rows for dirichlet boundaries and for ghost cells seemed to have >>>>>>>>>>>>>>>>>> effects on its convergence. I'm hoping once I know how to use PetscSection, >>>>>>>>>>>>>>>>>> I can get rid of using ghost cells method for the staggered grid and get >>>>>>>>>>>>>>>>>> rid of the identity rows too. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> It can change the exact iteration, but it does not make >>>>>>>>>>>>>>>>> the matrix conditioning worse. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Anyway please provide me with some pointers so that I >>>>>>>>>>>>>>>>>> can start trying with petscsection on top of a dmda, in the beginning for >>>>>>>>>>>>>>>>>> non-staggered case. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>>>> experiments lead. >>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>> experiments lead. >>>>>>>>> -- Norbert Wiener >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> What most experimenters take for granted before they begin their >>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>> experiments lead. >>>>>>> -- Norbert Wiener >>>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> What most experimenters take for granted before they begin their >>>>> experiments is infinitely more interesting than any results to which their >>>>> experiments lead. >>>>> -- Norbert Wiener >>>>> >>>> >>>> >>> >>> >>> -- >>> What most experimenters take for granted before they begin their >>> experiments is infinitely more interesting than any results to which their >>> experiments lead. >>> -- Norbert Wiener >>> >> >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From rongliang.chan at gmail.com Fri Nov 1 00:27:17 2013 From: rongliang.chan at gmail.com (Rongliang Chen) Date: Fri, 01 Nov 2013 13:27:17 +0800 Subject: [petsc-users] Debug AOCreateBasic In-Reply-To: References: Message-ID: <52733BB5.5030203@gmail.com> Hi there, My code died in the AOCreateBasic and the error messages are followed. Do you have any suggestions to debug this? Notes: 1. In this case, it has about 30,000,000 unstructured mesh and use 96 processors (I also tried 1024 processors and it has the same problem). 2. My code works well for a smaller case (about 25,000,000 unstructured meshes) . 3. I also check the memory usage of this case and it is very small because the solution stage does not start yet. Best, Rongliang -------------------------------------------------------------------------- MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD with errorcode 59. NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. You may or may not see output from other processes, depending on exactly when Open MPI kills them. -------------------------------------------------------------------------- [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the batch system) has told this process to end [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger [0]PETSC ERROR: or see http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSC ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory corruption errors [0]PETSC ERROR: likely location of problem given in stack below [0]PETSC ERROR: --------------------- Stack Frames ------------------------------------ [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available, [0]PETSC ERROR: INSTEAD the line number of the start of the function [0]PETSC ERROR: is given. [0]PETSC ERROR: [0] AOCreate_Basic line 203 src/vec/is/ao/impls/basic/aobasic.c [0]PETSC ERROR: [0] AOSetType line 35 src/vec/is/ao/interface/aoreg.c [0]PETSC ERROR: [0] AOCreateBasicIS line 380 src/vec/is/ao/impls/basic/aobasic.c [0]PETSC ERROR: [0] AOCreateBasic line 335 src/vec/is/ao/impls/basic/aobasic.c [0]PETSC ERROR: [0] DataPartitionVertices_Block line 1634 /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/readbinary3d.c [0]PETSC ERROR: [0] ReadBinary line 184 /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/readbinary3d.c [0]PETSC ERROR: [0] LoadGrid line 720 /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/loadgrid3d.c [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Signal received! [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Petsc Development GIT revision: ee17fca9fd6ac48e6579ef235144daafbb22b801 GIT Date: 2013-10-23 14:21:20 -0500 [0]PETSC ERROR: See docs/changes/index.html for recent updates. [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [0]PETSC ERROR: See docs/index.html for manual pages. [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: ./fsi3d on a Janus-debug-64bit named node0880 by ronglian Thu Oct 31 21:58:20 2013 [0]PETSC ERROR: Libraries linked from /projects/ronglian/soft/petsc-dev-latest/Janus-debug-64bit/lib [0]PETSC ERROR: Configure run at Thu Oct 24 21:24:31 2013 [0]PETSC ERROR: Configure options --known-level1-dcache-size=32768 --known-level1-dcache-linesize=64 --known-level1-dcache-assoc=8 --known-memcmp-ok=1 --known-sizeof-char=1 --known-sizeof-void-p=8 --known-sizeof-short=2 --known-sizeof-int=4 --known-sizeof-long=8 --known-sizeof-long-long=8 --known-sizeof-float=4 --known-sizeof-double=8 --known-sizeof-size_t=8 --known-bits-per-byte=8 --known-sizeof-MPI_Comm=8 --known-sizeof-MPI_Fint=4 --known-mpi-long-double=1 --known-mpi-c-double-complex=0 --download-blacs=1 --download-f-blas-lapack=1 --download-metis=1 --download-parmetis=1 --download-scalapack=1 --download-superlu_dist=1 --known-mpi-shared-libraries=0 --with-64-bit-indices --with-batch=1 --with-mpi-shared=1 --download-exodusii=1 --download-hdf5=1 --download-netcdf=1 --known-64-bit-blas-indices --with-debugging=1 COPTFLAGS="-O0 -g" [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: User provided function() line 0 in unknown file [1]PETSC ERROR: ------------------------------------------------------------------------ [1]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the batch system) has told this process to end [1]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger [1]PETSC ERROR: or see http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[1]PETSC ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory corruption errors [1]PETSC ERROR: likely location of problem given in stack below From jedbrown at mcs.anl.gov Fri Nov 1 00:35:25 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 31 Oct 2013 23:35:25 -0600 Subject: [petsc-users] Debug AOCreateBasic In-Reply-To: <52733BB5.5030203@gmail.com> References: <52733BB5.5030203@gmail.com> Message-ID: <8761scheg2.fsf@mcs.anl.gov> Rongliang Chen writes: > Hi there, > > My code died in the AOCreateBasic and the error messages are followed. > Do you have any suggestions to debug this? 1. Make sure your code is valgrind-clean for small sizes (to provide more evidence that it is getting the right answer for the right reason). 2. Try MPICH instead to see if you error in the same place. 3. Set up your system to dump core on selected ranks. > Notes: > 1. In this case, it has about 30,000,000 unstructured mesh and use 96 > processors (I also tried 1024 processors and it has the same problem). > 2. My code works well for a smaller case (about 25,000,000 unstructured > meshes) . > 3. I also check the memory usage of this case and it is very small > because the solution stage does not start yet. > > Best, > Rongliang > > -------------------------------------------------------------------------- > MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD > with errorcode 59. > > NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. > You may or may not see output from other processes, depending on > exactly when Open MPI kills them. > -------------------------------------------------------------------------- > [0]PETSC ERROR: > ------------------------------------------------------------------------ > [0]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the > batch system) has told this process to end > [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger > [0]PETSC ERROR: or see > http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSC > ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to > find memory corruption errors > [0]PETSC ERROR: likely location of problem given in stack below > [0]PETSC ERROR: --------------------- Stack Frames > ------------------------------------ > [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available, > [0]PETSC ERROR: INSTEAD the line number of the start of the function > [0]PETSC ERROR: is given. > [0]PETSC ERROR: [0] AOCreate_Basic line 203 > src/vec/is/ao/impls/basic/aobasic.c > [0]PETSC ERROR: [0] AOSetType line 35 src/vec/is/ao/interface/aoreg.c > [0]PETSC ERROR: [0] AOCreateBasicIS line 380 > src/vec/is/ao/impls/basic/aobasic.c > [0]PETSC ERROR: [0] AOCreateBasic line 335 > src/vec/is/ao/impls/basic/aobasic.c > [0]PETSC ERROR: [0] DataPartitionVertices_Block line 1634 > /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/readbinary3d.c > [0]PETSC ERROR: [0] ReadBinary line 184 > /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/readbinary3d.c > [0]PETSC ERROR: [0] LoadGrid line 720 > /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/loadgrid3d.c > [0]PETSC ERROR: --------------------- Error Message > ------------------------------------ > [0]PETSC ERROR: Signal received! > [0]PETSC ERROR: > ------------------------------------------------------------------------ > [0]PETSC ERROR: Petsc Development GIT revision: > ee17fca9fd6ac48e6579ef235144daafbb22b801 GIT Date: 2013-10-23 14:21:20 > -0500 > [0]PETSC ERROR: See docs/changes/index.html for recent updates. > [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > [0]PETSC ERROR: See docs/index.html for manual pages. > [0]PETSC ERROR: > ------------------------------------------------------------------------ > [0]PETSC ERROR: ./fsi3d on a Janus-debug-64bit named node0880 by > ronglian Thu Oct 31 21:58:20 2013 > [0]PETSC ERROR: Libraries linked from > /projects/ronglian/soft/petsc-dev-latest/Janus-debug-64bit/lib > [0]PETSC ERROR: Configure run at Thu Oct 24 21:24:31 2013 > [0]PETSC ERROR: Configure options --known-level1-dcache-size=32768 > --known-level1-dcache-linesize=64 --known-level1-dcache-assoc=8 > --known-memcmp-ok=1 --known-sizeof-char=1 --known-sizeof-void-p=8 > --known-sizeof-short=2 --known-sizeof-int=4 --known-sizeof-long=8 > --known-sizeof-long-long=8 --known-sizeof-float=4 > --known-sizeof-double=8 --known-sizeof-size_t=8 --known-bits-per-byte=8 > --known-sizeof-MPI_Comm=8 --known-sizeof-MPI_Fint=4 > --known-mpi-long-double=1 --known-mpi-c-double-complex=0 > --download-blacs=1 --download-f-blas-lapack=1 --download-metis=1 > --download-parmetis=1 --download-scalapack=1 --download-superlu_dist=1 > --known-mpi-shared-libraries=0 --with-64-bit-indices --with-batch=1 > --with-mpi-shared=1 --download-exodusii=1 --download-hdf5=1 > --download-netcdf=1 --known-64-bit-blas-indices --with-debugging=1 > COPTFLAGS="-O0 -g" > [0]PETSC ERROR: > ------------------------------------------------------------------------ > [0]PETSC ERROR: User provided function() line 0 in unknown file > [1]PETSC ERROR: > ------------------------------------------------------------------------ > [1]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the > batch system) has told this process to end > [1]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger > [1]PETSC ERROR: or see > http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[1]PETSC > ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to > find memory corruption errors > [1]PETSC ERROR: likely location of problem given in stack below -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From knepley at gmail.com Fri Nov 1 07:30:51 2013 From: knepley at gmail.com (Matthew Knepley) Date: Fri, 1 Nov 2013 07:30:51 -0500 Subject: [petsc-users] using petsc tools to solve isolated irregular domains with finite difference In-Reply-To: References: Message-ID: On Fri, Nov 1, 2013 at 12:08 AM, Bishesh Khanal wrote: > On Thu, Oct 31, 2013 at 11:34 PM, Matthew Knepley wrote: > >> On Thu, Oct 31, 2013 at 5:02 PM, Bishesh Khanal wrote: >> >>> On Thu, Oct 31, 2013 at 7:43 PM, Matthew Knepley wrote: >>> >>>> On Thu, Oct 31, 2013 at 1:25 PM, Bishesh Khanal wrote: >>>> >>>>> I did not call DMCreateMatrix() before when using just dmda (and it is >>>>> working). In any case, I added a call to DMCreateMatrix() now but still >>>>> there are problems. Here are the code snippets and the error I get: >>>>> >>>> >>>> Then you were allowing it to be called automatically by PETSc, and it >>>> would have been this time as well. >>>> >>>> >>>>> //Setting up the section and linking it to DM: >>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>> PetscInt point; >>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>> &point);CHKERRQ(ierr); >>>>> ierr = PetscSectionSetDof(s, point, testPoisson.mDof); >>>>> // No. of dofs associated with the point. >>>>> } >>>>> } >>>>> } >>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>>> >>>>> //Set up KSP: >>>>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>>>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>>>> ierr = >>>>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>> ierr = >>>>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>> ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); >>>>> ierr = KSPSolve(ksp,NULL,NULL);CHKERRQ(ierr); >>>>> >>>>> ------------------------------------------------------ >>>>> The computeMatrix2dSection function has: >>>>> >>>>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>>>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>>>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>>>> for(PetscInt j = info.ys; j>>>> for(PetscInt i = info.xs; i>>>> if (isPosInDomain(ctx,i,j)) { >>>>> ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>> ierr = PetscSectionGetDof(gs,point,&dof); >>>>> for(PetscInt cDof = 0; cDof < dof; ++cDof) { >>>>> num = 0; >>>>> row+=cDof; >>>>> col[num] = row; //(i,j) position >>>>> v[num++] = -4; >>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>> ierr = >>>>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>>>> ierr = >>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>> col[num] += cDof; >>>>> v[num++] = 1; >>>>> } >>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>> ierr = >>>>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>>>> ierr = >>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>> col[num] += cDof; >>>>> v[num++] = 1; >>>>> } >>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>> ierr = >>>>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>>>> ierr = >>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>> col[num] += cDof; >>>>> v[num++] = 1; >>>>> } >>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>> ierr = >>>>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>>>> ierr = >>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>> col[num] += cDof; >>>>> v[num++] = 1; >>>>> } >>>>> ierr = >>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>> } >>>>> } >>>>> } >>>>> } >>>>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>> >>>>> But this is not working. For e.g. for a 6X6 grid size I get the >>>>> following error: >>>>> >>>> >>>> Okay, here is how to debug this: >>>> >>>> 0) Go to a single scalar equations to make things easier >>>> >>>> 1) Use MatView(A, PETSC_VIEWER_STDOUT_WORLD), which should output a >>>> 36 row matrix with >>>> the preallocated nonzero pattern, filled with zeros >>>> >>>> 2) Make sure this is the pattern you want >>>> >>>> 3) Run in the debugger with -start_in_debugger >>>> >>>> 4) When you get the error, see >>>> >>>> a) If the (i, j) is one that should be setting a value >>>> >>>> b) Why this (i, j) was not preallocated >>>> >>> >>> Up to 4 (a), it is correct. There is a problem in the way >>> DMDAGetCellPoint and PetscSectionGetOffset works in this case scenario. I >>> will try to explain it below for the case of 4X4 grid. >>> First case: >>> If I set the computational domain to be all the points of the dmda grid, >>> (i.e. isPosInDomain(..,i,j,..) returns true for all i and j in the dmda >>> grid), the program runs fine and does not give any error. >>> >>> Second case: >>> I want the computational domain to be some part of the whole grid. There >>> is a problem in this case. >>> The following test is in a case where, >>> isPosInDomain(..,i,j,..) returns true ONLY for (i,j) pairs of (2,1) >>> (1,2) (2,2) (3,2) and (3,2). The grid with its corresponding point number >>> in the petscsection is shown below: >>> >>> 12 13 14 15 >>> 8 9 10 11 >>> 4 5 6 7 >>> 0 1 2 3 >>> >>> of which 6, 9, 10, 11 and 14 correspond to the (i,j) points that returns >>> true for isPosInDomain(..,i,j,..) >>> MatView gives me the 16-row matrix with the star stencil non-zero >>> structure as expected. >>> The error I get is: new non-zero at (0,2) caused a malloc. >>> >>> This error is for the value of (i,j) = (2,1), i.e. point 6. >>> The loop to set values in the matrix starts as: >>> for(PetscInt j = info.ys; j>> for(PetscInt i = info.xs; i>> if (isPosInDomain(ctx,i,j)) { >>> ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>> ierr = PetscSectionGetOffset(gs,point,&row); >>> >>> Now, what I wanted from the beginning is to create the matrix containing >>> the rows corresponding to only those points (i,j) which has true values for >>> isPosInDomain(ctx,i,j), that means in this case 5 row matrix. In the >>> current test, the first (i,j) that reaches DMDAGetCellPoint is (2,1), which >>> gives sets point variable to 6. >>> The line PetscSectionGetOffset(gs,point,&row) sets the value of row to 0. >>> So I think this where the inconsistency lies. >>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES) expects row variable and >>> col[] array to correspond to (i,j) based on DMDA, but PetsSectionGetOffset >>> gives result based on how we masked away some of the points from dmda grid. >>> Or am I missing something very basic here ? >>> >> >> You are missing something basic. The row is correctly identified as 0, >> and that means the 2 is >> point 10. >> > The problem must be preallocation. First, you should send the result of >> MatView(). >> > > MatView shows that the preallocation and the matrix created is for the > full dmda grid which is probably not what we want here, right ? Here is the > matview result: > Thats right. Did you call DMSetDefaultSection(da, s) after you made the new section and before you called DMGetMatrix()? Matt > Mat Object: 1 MPI processes > > type: seqaij > > row 0: (0, 0) (1, 0) (4, 0) > > row 1: (0, 0) (1, 0) (2, 0) (5, 0) > > row 2: (1, 0) (2, 0) (3, 0) (6, 0) > > row 3: (2, 0) (3, 0) (7, 0) > > row 4: (0, 0) (4, 0) (5, 0) (8, 0) > > row 5: (1, 0) (4, 0) (5, 0) (6, 0) (9, 0) > > row 6: (2, 0) (5, 0) (6, 0) (7, 0) (10, 0) > > row 7: (3, 0) (6, 0) (7, 0) (11, 0) > > row 8: (4, 0) (8, 0) (9, 0) (12, 0) > > row 9: (5, 0) (8, 0) (9, 0) (10, 0) (13, 0) > > row 10: (6, 0) (9, 0) (10, 0) (11, 0) (14, 0) > > row 11: (7, 0) (10, 0) (11, 0) (15, 0) > > row 12: (8, 0) (12, 0) (13, 0) > > row 13: (9, 0) (12, 0) (13, 0) (14, 0) > > row 14: (10, 0) (13, 0) (14, 0) (15, 0) > > row 15: (11, 0) (14, 0) (15, 0) > > [0]PETSC ERROR: --------------------- Error Message > ------------------------------------ > > [0]PETSC ERROR: Argument out of range! > > [0]PETSC ERROR: New nonzero at (0,2) caused a malloc! > > > >> Matt >> >> >>> >>>> Thanks, >>>> >>>> Matt >>>> >>>> >>>>> [0]PETSC ERROR: --------------------- Error Message >>>>> ------------------------------------ >>>>> [0]PETSC ERROR: Argument out of range! >>>>> [0]PETSC ERROR: New nonzero at (0,6) caused a malloc! >>>>> [0]PETSC ERROR: >>>>> ------------------------------------------------------------------------ >>>>> [0]PETSC ERROR: Petsc Development GIT revision: >>>>> 61e5e40bb2c5bf2423e94b71a15fef47e411b0da GIT Date: 2013-10-25 21:47:45 >>>>> -0500 >>>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>>> [0]PETSC ERROR: >>>>> ------------------------------------------------------------------------ >>>>> [0]PETSC ERROR: build/poissonIrregular on a arch-linux2-cxx-debug >>>>> named edwards by bkhanal Thu Oct 31 19:23:58 2013 >>>>> [0]PETSC ERROR: Libraries linked from >>>>> /home/bkhanal/Documents/softwares/petsc/arch-linux2-cxx-debug/lib >>>>> [0]PETSC ERROR: Configure run at Sat Oct 26 16:35:15 2013 >>>>> [0]PETSC ERROR: Configure options --download-mpich >>>>> -download-f-blas-lapack=1 --download-metis --download-parmetis >>>>> --download-superlu_dist --download-scalapack --download-mumps >>>>> --download-hypre --with-clanguage=cxx >>>>> [0]PETSC ERROR: >>>>> ------------------------------------------------------------------------ >>>>> [0]PETSC ERROR: MatSetValues_SeqAIJ() line 413 in >>>>> src/mat/impls/aij/seq/aij.c >>>>> [0]PETSC ERROR: MatSetValues() line 1130 in src/mat/interface/matrix.c >>>>> [0]PETSC ERROR: computeMatrix2dSection() line 318 in >>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>> [0]PETSC ERROR: KSPSetUp() line 228 in src/ksp/ksp/interface/itfunc.c >>>>> [0]PETSC ERROR: KSPSolve() line 399 in src/ksp/ksp/interface/itfunc.c >>>>> [0]PETSC ERROR: main() line 598 in >>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>> [unset]: aborting job: >>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>> >>>>> >>>>>> Matt >>>>>> >>>>>> >>>>>>> However, now since I do not want the rows in the matrix >>>>>>> corresponding to the points in the grid which do not lie in the >>>>>>> computational domain. This means the size of the matrix will not >>>>>>> necessarily equal the total number of cells in DMDA. So in the following >>>>>>> code: >>>>>>> >>>>>>> ierr = DMDAGetLocalInfo(dm,&info);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>> PetscInt point; >>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>> &point);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>> } >>>>>>> >>>>>>> } >>>>>>> } >>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>> >>>>>>> should I myself compute proper nC (i.e. total no. of points for >>>>>>> which I want the matrix to have a corresponding row) before calling >>>>>>> PetscSectionSetChart() or is the masking out of the points inside the loop >>>>>>> with if(isPosInDomain(&testPoisson,i,j,0)) sufficient ? >>>>>>> And, when you write: >>>>>>> PetscSectionGetOffset(gs, p, &ind); >>>>>>> row = ind < 0 ? -(ind+1) : ind; >>>>>>> >>>>>>> it seems you allow the possibility of getting a -ve ind when using >>>>>>> PetscSectionGetOffset. When would an offset to a particular dof and point? >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>> >>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>> num = 0; >>>>>>>>> /*ierr = >>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr);*/ /*Get the PetscPoint*/ >>>>>>>>> /*But I did now understand how I would emulate the >>>>>>>>> row.c and col.c info with PetscSection*/ >>>>>>>>> row.i = i; row.j = j; >>>>>>>>> >>>>>>>> >>>>>>>> Here you would call >>>>>>>> >>>>>>>> ierr = DMDAGetCellPoint(da, i, j, 0, &cell);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionGetOffset(da, cell, &row);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionGetDof(da, cell, &dof);CHKERRQ(ierr); >>>>>>>> >>>>>>>> This means that this cell has global rows = [row, row+dof). >>>>>>>> >>>>>>>> >>>>>>>>> col[num].i = i; col[num].j = j; >>>>>>>>> if (isPosInDomain(ctx,i,j)) { /*put the >>>>>>>>> operator for only the values for only the points lying in the domain*/ >>>>>>>>> v[num++] = -4; >>>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>>> col[num].i = i+1; col[num].j = j; >>>>>>>>> v[num++] = 1; >>>>>>>>> } >>>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>>> col[num].i = i-1; col[num].j = j; >>>>>>>>> v[num++] = 1; >>>>>>>>> } >>>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>>> col[num].i = i; col[num].j = j+1; >>>>>>>>> v[num++] = 1; >>>>>>>>> } >>>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>>> col[num].i = i; col[num].j = j-1; >>>>>>>>> v[num++] = 1; >>>>>>>>> } >>>>>>>>> } else { >>>>>>>>> v[num++] = dScale; /*set Dirichlet identity >>>>>>>>> rows for points in the rectangle but outside the computational domain*/ >>>>>>>>> } >>>>>>>>> >>>>>>>> >>>>>>>> You do the same thing for the columns, and then call >>>>>>>> >>>>>>>> ierr = MatSetValues(A, dof, rows, num*dof, cols, v, >>>>>>>> INSERT_VALUES);CHKERRQ(ierr); >>>>>>>> >>>>>>>> >>>>>>>>> ierr = >>>>>>>>> MatSetValuesStencil(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>> } >>>>>>>>> } >>>>>>>>> } >>>>>>>>> >>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> >>>>>>>>>> Matt >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> Matt >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> Matt >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> If not, just put the identity into >>>>>>>>>>>>>>>>>>>> the rows you do not use on the full cube. It will not >>>>>>>>>>>>>>>>>>>> hurt scalability or convergence. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> In the case of Poisson with Dirichlet condition this >>>>>>>>>>>>>>>>>>> might be the case. But is it always true that having identity rows in the >>>>>>>>>>>>>>>>>>> system matrix will not hurt convergence ? I thought otherwise for the >>>>>>>>>>>>>>>>>>> following reasons: >>>>>>>>>>>>>>>>>>> 1) Having read Jed's answer here : >>>>>>>>>>>>>>>>>>> http://scicomp.stackexchange.com/questions/3426/why-is-pinning-a-point-to-remove-a-null-space-bad/3427#3427 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Jed is talking about a constraint on a the pressure at a >>>>>>>>>>>>>>>>>> point. This is just decoupling these unknowns from the rest >>>>>>>>>>>>>>>>>> of the problem. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> 2) Some observation I am getting (but I am still doing >>>>>>>>>>>>>>>>>>> more experiments to confirm) while solving my staggered-grid 3D stokes flow >>>>>>>>>>>>>>>>>>> with schur complement and using -pc_type gamg for A00 matrix. Putting the >>>>>>>>>>>>>>>>>>> identity rows for dirichlet boundaries and for ghost cells seemed to have >>>>>>>>>>>>>>>>>>> effects on its convergence. I'm hoping once I know how to use PetscSection, >>>>>>>>>>>>>>>>>>> I can get rid of using ghost cells method for the staggered grid and get >>>>>>>>>>>>>>>>>>> rid of the identity rows too. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> It can change the exact iteration, but it does not make >>>>>>>>>>>>>>>>>> the matrix conditioning worse. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Anyway please provide me with some pointers so that I >>>>>>>>>>>>>>>>>>> can start trying with petscsection on top of a dmda, in the beginning for >>>>>>>>>>>>>>>>>>> non-staggered case. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>> their experiments lead. >>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>>> experiments lead. >>>>>>>>>> -- Norbert Wiener >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> What most experimenters take for granted before they begin their >>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>> experiments lead. >>>>>>>> -- Norbert Wiener >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> What most experimenters take for granted before they begin their >>>>>> experiments is infinitely more interesting than any results to which their >>>>>> experiments lead. >>>>>> -- Norbert Wiener >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> What most experimenters take for granted before they begin their >>>> experiments is infinitely more interesting than any results to which their >>>> experiments lead. >>>> -- Norbert Wiener >>>> >>> >>> >> >> >> -- >> What most experimenters take for granted before they begin their >> experiments is infinitely more interesting than any results to which their >> experiments lead. >> -- Norbert Wiener >> > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From tsariysk at craft-tech.com Fri Nov 1 09:12:51 2013 From: tsariysk at craft-tech.com (Ted Sariyski) Date: Fri, 01 Nov 2013 10:12:51 -0400 Subject: [petsc-users] How to pass data to PETSc In-Reply-To: <87mwlpi2ly.fsf@mcs.anl.gov> References: <5272C1F1.2030008@craft-tech.com> <87mwlpi2ly.fsf@mcs.anl.gov> Message-ID: <5273B6E3.40402@craft-tech.com> On 10/31/2013 04:53 PM, Jed Brown wrote: > Ted Sariyski writes: > >> Hi, >> I am about to use PETSc linear algebra. Two questions. >> >> Q1: I have the stiffness matrix and the rhs vector assembled (outside) >> and in PETSc order. Do I still need to create a scatter:"ierr = >> VecScatterCreate(bb,isglobal,localX,islocal,&scatter)"? > What do you want to do with the Vec? Use KSP for solving a linear system. > >> Q2: I have PETSc vectors and matrix set: >> >> ierr = VecCreate(MPI_COMM_WORLD,&bb);CHKERRQ(ierr); >> ierr = VecSetFromOptions(bb);CHKERRQ(ierr);CHKPT >> ierr = VecSetSizes(bb,Nvlocal,Nvglobal);CHKERRQ(ierr); >> ierr = MatCreate(PETSC_COMM_WORLD,&A);CHKERRQ(ierr); >> ierr = MatSetSizes(A,Nvlocal,Nvlocal,Nvglobal,Nvglobal);CHKERRQ(ierr); >> >> I passed the rhs myData to bb using VecSetValues: >> ierr = VecSetValues(bb,Nvlocal,vertices,myData,INSERT_VALUES);CHKERRQ(ierr); >> >> The matrix is in HYPRE sparse format. What is the recommended way to >> pass it to A? > Best is to use MatSetValues while you are generating the matrix entries > (so that you don't store the Hypre format). You can always do this > afterward. (The Hypre matrix can't be used directly because Hypre does > not expose a complete set of functionality. There is not a conversion > routine in PETSc, though it would be easy to add.) I'll do that. Thanks a lot. --Ted From jedbrown at mcs.anl.gov Fri Nov 1 09:17:02 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 01 Nov 2013 08:17:02 -0600 Subject: [petsc-users] How to pass data to PETSc In-Reply-To: <5273B6E3.40402@craft-tech.com> References: <5272C1F1.2030008@craft-tech.com> <87mwlpi2ly.fsf@mcs.anl.gov> <5273B6E3.40402@craft-tech.com> Message-ID: <87txfwfbq9.fsf@mcs.anl.gov> Ted Sariyski writes: > On 10/31/2013 04:53 PM, Jed Brown wrote: >> Ted Sariyski writes: >>> Q1: I have the stiffness matrix and the rhs vector assembled (outside) >>> and in PETSc order. Do I still need to create a scatter:"ierr = >>> VecScatterCreate(bb,isglobal,localX,islocal,&scatter)"? >> What do you want to do with the Vec? > Use KSP for solving a linear system. It looks like your vectors are already distributed, so you don't need a VecScatter. If you were computing nonlinear residuals, you would probably need to communicate with neighbors; VecScatter is a flexible way to do that. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From jedbrown at mcs.anl.gov Fri Nov 1 23:52:55 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 01 Nov 2013 22:52:55 -0600 Subject: [petsc-users] Debug AOCreateBasic In-Reply-To: <52734F7B.2060002@gmail.com> References: <52733BB5.5030203@gmail.com> <8761scheg2.fsf@mcs.anl.gov> <52734F7B.2060002@gmail.com> Message-ID: <8761sbpfq0.fsf@mcs.anl.gov> Please keep replies on the list. Rongliang Chen writes: > Hi Jed, > > Thank you for your suggestions. > > The valgrind results are attached. Can you help me to check it? (I do > not understand the result). There is a ton of noise from Open MPI. You can use suppression files to hide it. And then you have a number of places that are not inside the MPI stack. I would install locally (on your laptop/workstation) using MPICH and check that it is valgrind-clean for small sizes. > I am using a supercomputer where they do not have mpich (Are there other > ways to check if the bug comes from openmpi?). Easiest is to try a different implementation. > For your third > suggestion, I have no idea how to do that. Check the docs or ask the staff at your computing center. > Best, > Rongliang > > On 11/01/2013 01:35 PM, Jed Brown wrote: >> Rongliang Chen writes: >> >>> Hi there, >>> >>> My code died in the AOCreateBasic and the error messages are followed. >>> Do you have any suggestions to debug this? >> 1. Make sure your code is valgrind-clean for small sizes (to provide >> more evidence that it is getting the right answer for the right reason). >> >> 2. Try MPICH instead to see if you error in the same place. >> >> 3. Set up your system to dump core on selected ranks. >> >>> Notes: >>> 1. In this case, it has about 30,000,000 unstructured mesh and use 96 >>> processors (I also tried 1024 processors and it has the same problem). >>> 2. My code works well for a smaller case (about 25,000,000 unstructured >>> meshes) . >>> 3. I also check the memory usage of this case and it is very small >>> because the solution stage does not start yet. >>> >>> Best, >>> Rongliang >>> >>> -------------------------------------------------------------------------- >>> MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD >>> with errorcode 59. >>> >>> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. >>> You may or may not see output from other processes, depending on >>> exactly when Open MPI kills them. >>> -------------------------------------------------------------------------- >>> [0]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [0]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the >>> batch system) has told this process to end >>> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger >>> [0]PETSC ERROR: or see >>> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSC >>> ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to >>> find memory corruption errors >>> [0]PETSC ERROR: likely location of problem given in stack below >>> [0]PETSC ERROR: --------------------- Stack Frames >>> ------------------------------------ >>> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available, >>> [0]PETSC ERROR: INSTEAD the line number of the start of the function >>> [0]PETSC ERROR: is given. >>> [0]PETSC ERROR: [0] AOCreate_Basic line 203 >>> src/vec/is/ao/impls/basic/aobasic.c >>> [0]PETSC ERROR: [0] AOSetType line 35 src/vec/is/ao/interface/aoreg.c >>> [0]PETSC ERROR: [0] AOCreateBasicIS line 380 >>> src/vec/is/ao/impls/basic/aobasic.c >>> [0]PETSC ERROR: [0] AOCreateBasic line 335 >>> src/vec/is/ao/impls/basic/aobasic.c >>> [0]PETSC ERROR: [0] DataPartitionVertices_Block line 1634 >>> /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/readbinary3d.c >>> [0]PETSC ERROR: [0] ReadBinary line 184 >>> /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/readbinary3d.c >>> [0]PETSC ERROR: [0] LoadGrid line 720 >>> /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/loadgrid3d.c >>> [0]PETSC ERROR: --------------------- Error Message >>> ------------------------------------ >>> [0]PETSC ERROR: Signal received! >>> [0]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [0]PETSC ERROR: Petsc Development GIT revision: >>> ee17fca9fd6ac48e6579ef235144daafbb22b801 GIT Date: 2013-10-23 14:21:20 >>> -0500 >>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>> [0]PETSC ERROR: See docs/index.html for manual pages. >>> [0]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [0]PETSC ERROR: ./fsi3d on a Janus-debug-64bit named node0880 by >>> ronglian Thu Oct 31 21:58:20 2013 >>> [0]PETSC ERROR: Libraries linked from >>> /projects/ronglian/soft/petsc-dev-latest/Janus-debug-64bit/lib >>> [0]PETSC ERROR: Configure run at Thu Oct 24 21:24:31 2013 >>> [0]PETSC ERROR: Configure options --known-level1-dcache-size=32768 >>> --known-level1-dcache-linesize=64 --known-level1-dcache-assoc=8 >>> --known-memcmp-ok=1 --known-sizeof-char=1 --known-sizeof-void-p=8 >>> --known-sizeof-short=2 --known-sizeof-int=4 --known-sizeof-long=8 >>> --known-sizeof-long-long=8 --known-sizeof-float=4 >>> --known-sizeof-double=8 --known-sizeof-size_t=8 --known-bits-per-byte=8 >>> --known-sizeof-MPI_Comm=8 --known-sizeof-MPI_Fint=4 >>> --known-mpi-long-double=1 --known-mpi-c-double-complex=0 >>> --download-blacs=1 --download-f-blas-lapack=1 --download-metis=1 >>> --download-parmetis=1 --download-scalapack=1 --download-superlu_dist=1 >>> --known-mpi-shared-libraries=0 --with-64-bit-indices --with-batch=1 >>> --with-mpi-shared=1 --download-exodusii=1 --download-hdf5=1 >>> --download-netcdf=1 --known-64-bit-blas-indices --with-debugging=1 >>> COPTFLAGS="-O0 -g" >>> [0]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [0]PETSC ERROR: User provided function() line 0 in unknown file >>> [1]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [1]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the >>> batch system) has told this process to end >>> [1]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger >>> [1]PETSC ERROR: or see >>> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[1]PETSC >>> ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to >>> find memory corruption errors >>> [1]PETSC ERROR: likely location of problem given in stack below > > --15577-- WARNING: Serious error when reading debug info > --15577-- When reading debug info from /projects/ronglian/soft/petsc-dev-latest/Janus-debug-64bit/lib/libnetcdf.so.7.2.0: > --15577-- DWARF line info appears to be corrupt - the section is too small > --15577-- Warning: DWARF2 CFI reader: unhandled DW_OP_ opcode 0x8 > --15577-- Warning: DWARF2 CFI reader: unhandled DW_OP_ opcode 0x8 > --15577-- Warning: DWARF2 CFI reader: unhandled DW_OP_ opcode 0x8 > --15577-- Warning: DWARF2 CFI reader: unhandled DW_OP_ opcode 0x8 > --15577-- Warning: DWARF2 CFI reader: unhandled DW_OP_ opcode 0x8 > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0278: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9755825: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9755967: rdma_create_event_channel (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9091663: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:313) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xAF13D2B: _itoa_word (in /lib64/libc-2.12.so) > ==15577== by 0xAF150F5: vfprintf (in /lib64/libc-2.12.so) > ==15577== by 0xAFCF2BF: __vsnprintf_chk (in /lib64/libc-2.12.so) > ==15577== by 0xAFCF1F9: __snprintf_chk (in /lib64/libc-2.12.so) > ==15577== by 0xF1C2F06: mlx4_query_device (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C02CE: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9755825: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9755967: rdma_create_event_channel (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9091663: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:313) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xAF13D35: _itoa_word (in /lib64/libc-2.12.so) > ==15577== by 0xAF150F5: vfprintf (in /lib64/libc-2.12.so) > ==15577== by 0xAFCF2BF: __vsnprintf_chk (in /lib64/libc-2.12.so) > ==15577== by 0xAFCF1F9: __snprintf_chk (in /lib64/libc-2.12.so) > ==15577== by 0xF1C2F06: mlx4_query_device (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C02CE: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9755825: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9755967: rdma_create_event_channel (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9091663: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:313) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0306: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9755825: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9755967: rdma_create_event_channel (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9091663: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:313) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x975656A: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x90916AE: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:330) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffdfac is on thread 1's stack > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x9755C0E: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9755C5D: rdma_destroy_id (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9091766: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:370) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffdf80 is on thread 1's stack > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9755C62: rdma_destroy_id (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9091766: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:370) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9755C8A: rdma_destroy_id (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9091766: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:370) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x9756207: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x90916AE: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:330) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffde70 is on thread 1's stack > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x975621C: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x90916AE: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:330) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9756368: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x90916AE: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:330) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x975637F: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x90916AE: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:330) > ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) > ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) > ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0278: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x90813A7: btl_openib_component_init (btl_openib_component.c:1663) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0306: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x90813A7: btl_openib_component_init (btl_openib_component.c:1663) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90815CC: btl_openib_component_init (btl_openib_component.c:1248) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9081829: btl_openib_component_init (btl_openib_component.c:1685) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x908E768: ompi_btl_openib_ini_query (btl_openib_ini.c:176) > ==15577== by 0x9081854: btl_openib_component_init (btl_openib_component.c:1694) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x908E76E: ompi_btl_openib_ini_query (btl_openib_ini.c:176) > ==15577== by 0x9081854: btl_openib_component_init (btl_openib_component.c:1694) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9081AC8: btl_openib_component_init (btl_openib_component.c:1896) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9081AB8: btl_openib_component_init (btl_openib_component.c:1905) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9081B01: btl_openib_component_init (btl_openib_component.c:1910) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x902DA00: init_one_port (btl_openib_component.c:681) > ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x902DB0E: init_one_port (btl_openib_component.c:712) > ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x902DBBF: init_one_port (btl_openib_component.c:758) > ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x902E062: init_one_port (btl_openib_component.c:778) > ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xAF13D2B: _itoa_word (in /lib64/libc-2.12.so) > ==15577== by 0xAF150F5: vfprintf (in /lib64/libc-2.12.so) > ==15577== by 0xAF391C8: vsprintf (in /lib64/libc-2.12.so) > ==15577== by 0xAF1F167: sprintf (in /lib64/libc-2.12.so) > ==15577== by 0x902DE57: init_one_port (btl_openib_component.c:824) > ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xAF13D35: _itoa_word (in /lib64/libc-2.12.so) > ==15577== by 0xAF150F5: vfprintf (in /lib64/libc-2.12.so) > ==15577== by 0xAF391C8: vsprintf (in /lib64/libc-2.12.so) > ==15577== by 0xAF1F167: sprintf (in /lib64/libc-2.12.so) > ==15577== by 0x902DE57: init_one_port (btl_openib_component.c:824) > ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x902DF1D: init_one_port (btl_openib_component.c:856) > ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x902DF80: init_one_port (btl_openib_component.c:888) > ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9092BC4: oob_component_query (btl_openib_connect_oob.c:131) > ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) > ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x975656A: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9098B22: rdmacm_component_query (btl_openib_connect_rdmacm.c:1859) > ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) > ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffdf7c is on thread 1's stack > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x9756207: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9098B22: rdmacm_component_query (btl_openib_connect_rdmacm.c:1859) > ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) > ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffde40 is on thread 1's stack > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x975621C: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9098B22: rdmacm_component_query (btl_openib_connect_rdmacm.c:1859) > ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) > ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9756368: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9098B22: rdmacm_component_query (btl_openib_connect_rdmacm.c:1859) > ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) > ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9091545: mca_btl_openib_rdma_get_ipv4addr (btl_openib_ip.c:138) > ==15577== by 0x9098B5E: rdmacm_component_query (btl_openib_connect_rdmacm.c:1702) > ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) > ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x9756470: rdma_listen (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9098D41: rdmacm_component_query (btl_openib_connect_rdmacm.c:1881) > ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) > ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffdf98 is on thread 1's stack > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x9756207: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x975648C: rdma_listen (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9098D41: rdmacm_component_query (btl_openib_connect_rdmacm.c:1881) > ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) > ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffde80 is on thread 1's stack > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x975621C: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x975648C: rdma_listen (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9098D41: rdmacm_component_query (btl_openib_connect_rdmacm.c:1881) > ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) > ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9756368: ??? (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x975648C: rdma_listen (in /usr/lib64/librdmacm.so.1.0.0) > ==15577== by 0x9098D41: rdmacm_component_query (btl_openib_connect_rdmacm.c:1881) > ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) > ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x9082D96: btl_openib_component_init (btl_openib_component.c:1085) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0xda015cc is 268 bytes inside a block of size 8,784 alloc'd > ==15577== at 0x4C2694F: calloc (vg_replace_malloc.c:593) > ==15577== by 0xF1C0126: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x90813A7: btl_openib_component_init (btl_openib_component.c:1663) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x908EA9B: btl_openib_async_command_done (btl_openib_async.c:483) > ==15577== by 0x9082DAE: btl_openib_component_init (btl_openib_component.c:1092) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9082DBF: btl_openib_component_init (btl_openib_component.c:1110) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x995EEB6: ibv_cmd_reg_mr (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1C2D79: mlx4_reg_mr (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965782: ibv_reg_mr (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9080016: openib_reg_mr (btl_openib_component.c:602) > ==15577== by 0x90ED046: register_cache_bypass (mpool_rdma_module.c:172) > ==15577== by 0x90ED728: mca_mpool_rdma_alloc (mpool_rdma_module.c:115) > ==15577== by 0x902FAA7: ompi_free_list_grow (ompi_free_list.c:203) > ==15577== by 0x9082EA4: btl_openib_component_init (btl_openib_component.c:1150) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffdf18 is on thread 1's stack > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x907E6A4: mca_btl_openib_get_transport_type (btl_openib.c:436) > ==15577== by 0x9080188: btl_openib_modex_send (btl_openib_component.c:386) > ==15577== by 0x90829AD: btl_openib_component_init (btl_openib_component.c:2925) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x907E6CA: mca_btl_openib_get_transport_type (btl_openib.c:459) > ==15577== by 0x9080188: btl_openib_modex_send (btl_openib_component.c:386) > ==15577== by 0x90829AD: btl_openib_component_init (btl_openib_component.c:2925) > ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) > ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) > ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) > ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) > ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) > ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param writev(vector[...]) points to uninitialised byte(s) > ==15577== at 0xAFB0377: writev (in /lib64/libc-2.12.so) > ==15577== by 0x918AFE2: mca_oob_tcp_msg_send_handler (oob_tcp_msg.c:249) > ==15577== by 0x918BD4C: mca_oob_tcp_peer_send (oob_tcp_peer.c:204) > ==15577== by 0x918EEFB: mca_oob_tcp_send_nb (oob_tcp_send.c:167) > ==15577== by 0x91A8EEE: orte_rml_oob_send (rml_oob_send.c:136) > ==15577== by 0x91A9468: orte_rml_oob_send_buffer (rml_oob_send.c:270) > ==15577== by 0x917549F: modex (grpcomm_bad_module.c:573) > ==15577== by 0x9044943: ompi_mpi_init (ompi_mpi_init.c:541) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0xdb504c1 is 161 bytes inside a block of size 512 alloc'd > ==15577== at 0x4C2884E: realloc (vg_replace_malloc.c:662) > ==15577== by 0x91C0B17: opal_dss_buffer_extend (dss_internal_functions.c:63) > ==15577== by 0x91C0E4D: opal_dss_copy_payload (dss_load_unload.c:164) > ==15577== by 0x916F255: orte_grpcomm_base_pack_modex_entries (grpcomm_base_modex.c:861) > ==15577== by 0x917537E: modex (grpcomm_bad_module.c:563) > ==15577== by 0x9044943: ompi_mpi_init (ompi_mpi_init.c:541) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x907E6A4: mca_btl_openib_get_transport_type (btl_openib.c:436) > ==15577== by 0x907EE11: mca_btl_openib_add_procs (btl_openib.c:475) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x907E6CA: mca_btl_openib_get_transport_type (btl_openib.c:459) > ==15577== by 0x907EE11: mca_btl_openib_add_procs (btl_openib.c:475) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C2B16: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C2B9C: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C2BA6: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param mmap(length) contains uninitialised byte(s) > ==15577== at 0xAFB4C6A: mmap (in /lib64/libc-2.12.so) > ==15577== by 0xF1BF01A: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x996412E: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9964218: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9964224: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x99639A3: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99642FA: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x99639B1: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99642FA: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param madvise(length) contains uninitialised byte(s) > ==15577== at 0xAFB4D57: madvise (in /lib64/libc-2.12.so) > ==15577== by 0x9964449: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x996426D: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) > ==15577== by 0xF1BF1CC: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) > ==15577== by 0xF1BF1CC: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B11D: memset (mc_replace_strmem.c:1007) > ==15577== by 0xF1BF1CC: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C2B16: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907EDBA: mca_btl_openib_add_procs (btl_openib.c:410) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C2B9C: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907EDBA: mca_btl_openib_add_procs (btl_openib.c:410) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C2BA6: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) > ==15577== by 0x907EDBA: mca_btl_openib_add_procs (btl_openib.c:410) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x9960C21: ibv_cmd_create_srq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1C2869: mlx4_create_srq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99652BB: ibv_create_srq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907F359: mca_btl_openib_add_procs (btl_openib.c:255) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffe138 is on thread 1's stack > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x9960766: ibv_cmd_modify_srq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1C2731: mlx4_modify_srq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x907F92A: mca_btl_openib_add_procs (btl_openib.c:279) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffe1a8 is on thread 1's stack > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x995FDD1: ibv_cmd_destroy_srq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1C266D: mlx4_destroy_srq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x907F3FC: mca_btl_openib_add_procs (btl_openib.c:285) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== Address 0x7feffe160 is on thread 1's stack > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x995FE0B: ibv_cmd_destroy_srq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1C266D: mlx4_destroy_srq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x907F3FC: mca_btl_openib_add_procs (btl_openib.c:285) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x99641A4: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C1A79: mlx4_alloc_srq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2807: mlx4_create_srq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99652BB: ibv_create_srq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907F54E: mca_btl_openib_add_procs (btl_openib.c:334) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x99641BB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C1A79: mlx4_alloc_srq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2807: mlx4_create_srq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99652BB: ibv_create_srq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907F54E: mca_btl_openib_add_procs (btl_openib.c:334) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x99641CD: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C1A79: mlx4_alloc_srq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C2807: mlx4_create_srq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x99652BB: ibv_create_srq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x907F54E: mca_btl_openib_add_procs (btl_openib.c:334) > ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) > ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) > ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1BF64A: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9174254: barrier (grpcomm_bad_module.c:277) > ==15577== by 0x90450C1: ompi_mpi_init (ompi_mpi_init.c:783) > ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) > ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C2F5B: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C2F60: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C2FCA: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C2FCF: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x9960AA8: ibv_cmd_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1C31A5: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== Address 0x7feffaec8 is on thread 1's stack > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C089F: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C26906: calloc (vg_replace_malloc.c:593) > ==15577== by 0xF1C0902: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C0906: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C091A: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C08B3: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C08BE: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C08C3: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x995F923: ibv_cmd_modify_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1C24BF: mlx4_modify_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9964C73: ibv_modify_qp (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9092FA7: qp_create_all (btl_openib_connect_oob.c:514) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== Address 0x7feffaef8 is on thread 1's stack > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0B91: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C0C4E: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9087EE4: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0C62: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9087EE4: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0CD4: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9087EE4: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C0CAF: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9087EE4: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C0D0D: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9087EE4: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0B91: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C0C4E: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x908824E: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0C62: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x908824E: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C0CAF: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x908824E: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0CD4: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x908824E: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C0D0D: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x908824E: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Syscall param write(buf) points to uninitialised byte(s) > ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) > ==15577== by 0x9960766: ibv_cmd_modify_srq (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0xF1C2731: mlx4_modify_srq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x908446D: mca_btl_openib_post_srr (btl_openib_component.c:3760) > ==15577== by 0x9087E5F: mca_btl_openib_endpoint_post_recvs (btl_openib_endpoint.c:493) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== Address 0x7feffadf8 is on thread 1's stack > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x99641A4: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9965746: ibv_reg_mr (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9080016: openib_reg_mr (btl_openib_component.c:602) > ==15577== by 0x90ED046: register_cache_bypass (mpool_rdma_module.c:172) > ==15577== by 0x90ED728: mca_mpool_rdma_alloc (mpool_rdma_module.c:115) > ==15577== by 0x902FAA7: ompi_free_list_grow (ompi_free_list.c:203) > ==15577== by 0x9084273: mca_btl_openib_post_srr (ompi_free_list.h:253) > ==15577== by 0x9087E5F: mca_btl_openib_endpoint_post_recvs (btl_openib_endpoint.c:493) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x99641BB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9965746: ibv_reg_mr (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9080016: openib_reg_mr (btl_openib_component.c:602) > ==15577== by 0x90ED046: register_cache_bypass (mpool_rdma_module.c:172) > ==15577== by 0x90ED728: mca_mpool_rdma_alloc (mpool_rdma_module.c:115) > ==15577== by 0x902FAA7: ompi_free_list_grow (ompi_free_list.c:203) > ==15577== by 0x9084273: mca_btl_openib_post_srr (ompi_free_list.h:253) > ==15577== by 0x9087E5F: mca_btl_openib_endpoint_post_recvs (btl_openib_endpoint.c:493) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x99641CD: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9965746: ibv_reg_mr (in /usr/lib64/libibverbs.so.1.0.0) > ==15577== by 0x9080016: openib_reg_mr (btl_openib_component.c:602) > ==15577== by 0x90ED046: register_cache_bypass (mpool_rdma_module.c:172) > ==15577== by 0x90ED728: mca_mpool_rdma_alloc (mpool_rdma_module.c:115) > ==15577== by 0x902FAA7: ompi_free_list_grow (ompi_free_list.c:203) > ==15577== by 0x9084273: mca_btl_openib_post_srr (ompi_free_list.h:253) > ==15577== by 0x9087E5F: mca_btl_openib_endpoint_post_recvs (btl_openib_endpoint.c:493) > ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) > ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== > ==15577== > ==15577== More than 100 errors detected. Subsequent errors > ==15577== will still be recorded, but in less detail than before. > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9093D3C: rml_recv_cb (btl_openib_connect_oob.c:784) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9093D42: rml_recv_cb (btl_openib_connect_oob.c:785) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1BF6BB: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1BF6CE: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C07EF: mlx4_find_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BFA2F: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C07FD: mlx4_find_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BFA2F: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C0802: mlx4_find_qp (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1BFA2F: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1BF8D0: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1BF909: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1BF9C8: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1BF659: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9084584: btl_openib_handle_incoming (btl_openib_component.c:1374) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x90845A2: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9111AF0: mca_pml_ob1_recv_frag_callback_match (opal_pointer_array.h:130) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111B34: mca_pml_ob1_recv_frag_callback_match (opal_pointer_array.h:134) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111B5B: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:157) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9111B61: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:157) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111B67: mca_pml_ob1_recv_frag_callback_match (opal_list.h:322) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111B78: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:167) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91115EF: match_incomming.isra.2 (opal_list.h:197) > ==15577== by 0x9111B93: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:470) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91116A5: match_incomming.isra.2 (pml_ob1_recvfrag.c:445) > ==15577== by 0x9111B93: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:470) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91116F0: match_incomming.isra.2 (opal_list.h:378) > ==15577== by 0x9111B93: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:470) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111BA7: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:474) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90845D2: btl_openib_handle_incoming (btl_openib_component.c:3104) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90845DE: btl_openib_handle_incoming (btl_openib_component.c:3119) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9084614: btl_openib_handle_incoming (btl_openib_component.c:1382) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90848DF: btl_openib_handle_incoming (btl_openib_component.c:3170) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1BFC00: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1BF710: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x902E678: append_frag_to_list (opal_list.h:431) > ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66986: PetscOptionsInsertFile (options.c:471) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66A34: PetscOptionsInsertFile (options.c:473) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66A3E: PetscOptionsInsertFile (options.c:473) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66B04: PetscOptionsInsertFile (options.c:479) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66DB3: PetscOptionsInsertFile (options.c:493) > ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115B7B: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:896) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115C39: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1020) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115C41: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1020) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115C49: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1020) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) > ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66986: PetscOptionsInsertFile (options.c:471) > ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66A34: PetscOptionsInsertFile (options.c:473) > ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66A3E: PetscOptionsInsertFile (options.c:473) > ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66B04: PetscOptionsInsertFile (options.c:479) > ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66DB3: PetscOptionsInsertFile (options.c:493) > ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66986: PetscOptionsInsertFile (options.c:471) > ==15577== by 0x4F68144: PetscOptionsInsert (options.c:626) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66A34: PetscOptionsInsertFile (options.c:473) > ==15577== by 0x4F68144: PetscOptionsInsert (options.c:626) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66A3E: PetscOptionsInsertFile (options.c:473) > ==15577== by 0x4F68144: PetscOptionsInsert (options.c:626) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66B04: PetscOptionsInsertFile (options.c:479) > ==15577== by 0x4F68144: PetscOptionsInsert (options.c:626) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F66DB3: PetscOptionsInsertFile (options.c:493) > ==15577== by 0x4F68144: PetscOptionsInsert (options.c:626) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9115C1C: mca_pml_ob1_recv_req_start (opal_list.h:376) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4F682EE: PetscOptionsInsert (options.c:638) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F6834E: PetscOptionsInsert (options.c:639) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F68415: PetscOptionsInsert (options.c:643) > ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1BF814: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4EF1345: PetscWorldIsSingleHost (pdisplay.c:94) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1BFA5F: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x4EF1345: PetscWorldIsSingleHost (pdisplay.c:94) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C29BDC: strcmp (mc_replace_strmem.c:725) > ==15577== by 0x4EE8F19: PetscStrcmp (str.c:440) > ==15577== by 0x4EF13BA: PetscWorldIsSingleHost (pdisplay.c:95) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C29BF0: strcmp (mc_replace_strmem.c:725) > ==15577== by 0x4EE8F19: PetscStrcmp (str.c:440) > ==15577== by 0x4EF13BA: PetscWorldIsSingleHost (pdisplay.c:95) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE8F21: PetscStrcmp (str.c:441) > ==15577== by 0x4EF13BA: PetscWorldIsSingleHost (pdisplay.c:95) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0B91: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0xF1C0E28: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x907D977: mca_btl_openib_sendi (verbs.h:1111) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C13DC: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x907D977: mca_btl_openib_sendi (verbs.h:1111) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C13DC: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x908701E: mca_btl_openib_endpoint_post_send (verbs.h:1111) > ==15577== by 0x9088458: mca_btl_openib_endpoint_connected (btl_openib_endpoint.c:689) > ==15577== by 0x9093E0C: rml_recv_cb (btl_openib_connect_oob.c:914) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C144B: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x908701E: mca_btl_openib_endpoint_post_send (verbs.h:1111) > ==15577== by 0x9088458: mca_btl_openib_endpoint_connected (btl_openib_endpoint.c:689) > ==15577== by 0x9093E0C: rml_recv_cb (btl_openib_connect_oob.c:914) > ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) > ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) > ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) > ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) > ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1BF728: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1BF7DE: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x90852FB: poll_device (verbs.h:976) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) > ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) > ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C144B: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x907D977: mca_btl_openib_sendi (verbs.h:1111) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4FA3DC7: PetscLogBegin_Private (plog.c:248) > ==15577== by 0x4FA4103: PetscLogBegin (plog.c:292) > ==15577== by 0x4F7A5FF: PetscOptionsCheckInitial_Private (init.c:491) > ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) > ==15577== by 0x405831: main (fsi3d.c:27) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x441A29: DataRead (readbinary3d.c:678) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x441B80: DataRead (readbinary3d.c:696) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x441B98: DataRead (readbinary3d.c:697) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x441BD9: DataRead (readbinary3d.c:697) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x441BD9: DataRead (readbinary3d.c:697) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x441BD9: DataRead (readbinary3d.c:697) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x441C05: DataRead (readbinary3d.c:698) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x441C46: DataRead (readbinary3d.c:698) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F231A2: PetscMallocValidate (mtr.c:139) > ==15577== by 0x4F235E5: PetscTrMallocDefault (mtr.c:182) > ==15577== by 0x441C46: DataRead (readbinary3d.c:698) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x441C46: DataRead (readbinary3d.c:698) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x441C46: DataRead (readbinary3d.c:698) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x905061F: PMPI_Bcast (pbcast.c:75) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90506A3: PMPI_Bcast (pbcast.c:101) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90A0E8D: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:251) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115BA2: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.h:200) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DB968: opal_convertor_prepare_for_recv (opal_convertor.c:523) > ==15577== by 0x9115BD7: mca_pml_ob1_recv_req_start (opal_convertor.h:258) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90A92FA: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:224) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9111BE8: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:194) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DB703: opal_convertor_unpack (opal_convertor.c:282) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3B4: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A435: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A43B: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A450: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A456: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A47E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9111C64: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvreq.h:168) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) > ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x905061F: PMPI_Bcast (pbcast.c:75) > ==15577== by 0x441FF9: DataRead (readbinary3d.c:713) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90506A3: PMPI_Bcast (pbcast.c:101) > ==15577== by 0x441FF9: DataRead (readbinary3d.c:713) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91139FC: mca_pml_ob1_recv_request_progress_match (opal_convertor.h:284) > ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441FF9: DataRead (readbinary3d.c:713) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DB703: opal_convertor_unpack (opal_convertor.c:282) > ==15577== by 0x9113A26: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.c:632) > ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441FF9: DataRead (readbinary3d.c:713) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9113BE8: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.h:168) > ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) > ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) > ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) > ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) > ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) > ==15577== by 0x441FF9: DataRead (readbinary3d.c:713) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44205D: DataRead (readbinary3d.c:718) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x442208: DataRead (readbinary3d.c:719) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44208D: DataRead (readbinary3d.c:731) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4420E8: DataRead (readbinary3d.c:733) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x442143: DataRead (readbinary3d.c:735) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x445163: DataReadAndSplitGeneric (readbinary3d.c:1068) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x445163: DataReadAndSplitGeneric (readbinary3d.c:1068) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x445163: DataReadAndSplitGeneric (readbinary3d.c:1068) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x905D67C: PMPI_Recv (precv.c:53) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x905D7EC: PMPI_Recv (precv.c:54) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115BA2: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.h:200) > ==15577== by 0x910F5C1: mca_pml_ob1_recv (pml_ob1_irecv.c:104) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9111793: mca_pml_ob1_recv_frag_match (opal_pointer_array.h:130) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91117EC: mca_pml_ob1_recv_frag_match (opal_pointer_array.h:134) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x911180E: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:606) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9111816: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:607) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111836: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:618) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91115E0: match_incomming.isra.2 (opal_list.h:322) > ==15577== by 0x9111854: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:470) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91115EF: match_incomming.isra.2 (opal_list.h:197) > ==15577== by 0x9111854: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:470) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91116A5: match_incomming.isra.2 (pml_ob1_recvfrag.c:445) > ==15577== by 0x9111854: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:470) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91116F0: match_incomming.isra.2 (opal_list.h:378) > ==15577== by 0x9111854: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:470) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x911186A: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:474) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9112D10: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:258) > ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9112DEB: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:265) > ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9112E16: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:271) > ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9112E2E: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:282) > ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9112D3C: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:302) > ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91145B8: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9114645: mca_pml_ob1_recv_request_progress_rndv (opal_convertor.h:284) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DB703: opal_convertor_unpack (opal_convertor.c:282) > ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9114526: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.h:185) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91118A2: mca_pml_ob1_recv_frag_match (opal_list.h:322) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9115644: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115655: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9115678: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:284) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115689: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:284) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x911568F: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:294) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115696: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:294) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91DB694: opal_convertor_unpack (opal_convertor.c:266) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91DB6B0: opal_convertor_unpack (opal_convertor.c:266) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91DB6CC: opal_convertor_unpack (opal_convertor.c:276) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DB703: opal_convertor_unpack (opal_convertor.c:282) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3A0: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3B4: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A435: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A456: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91DB754: opal_convertor_unpack (opal_convertor.c:290) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91156B3: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:465) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91156CA: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:185) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91156D1: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:185) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91156EB: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:467) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3A0: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91DB785: opal_convertor_unpack (opal_convertor.c:296) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91156D3: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:58) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91157D8: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:152) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x911581F: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:161) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9115834: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:166) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115854: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:168) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9115861: mca_pml_ob1_recv_request_progress_frag (request.h:395) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9115876: mca_pml_ob1_recv_request_progress_frag (request.h:399) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44223B: DataRead (readbinary3d.c:756) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) > ==15577== by 0x4422BE: DataRead (readbinary3d.c:757) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x445106: DataReadAndSplitGeneric (readbinary3d.c:1062) > ==15577== by 0x4422BE: DataRead (readbinary3d.c:757) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) > ==15577== by 0x4422BE: DataRead (readbinary3d.c:757) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) > ==15577== by 0x442341: DataRead (readbinary3d.c:758) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) > ==15577== by 0x442341: DataRead (readbinary3d.c:758) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) > ==15577== by 0x4423C4: DataRead (readbinary3d.c:759) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) > ==15577== by 0x4423C4: DataRead (readbinary3d.c:759) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4426CC: DataRead (readbinary3d.c:785) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) > ==15577== by 0x442702: DataRead (readbinary3d.c:786) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) > ==15577== by 0x442702: DataRead (readbinary3d.c:786) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) > ==15577== by 0x442788: DataRead (readbinary3d.c:787) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) > ==15577== by 0x442788: DataRead (readbinary3d.c:787) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x445106: DataReadAndSplitGeneric (readbinary3d.c:1062) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084581: btl_openib_handle_incoming (btl_openib_component.c:1374) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9084584: btl_openib_handle_incoming (btl_openib_component.c:1374) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x908458A: btl_openib_handle_incoming (btl_openib_component.c:3096) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x90845A2: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9111F8B: mca_pml_ob1_recv_frag_callback_rndv (pml_ob1_recvfrag.c:254) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x911177D: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:568) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91117F5: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:585) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111841: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:470) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91144C0: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:564) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9112D09: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:258) > ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9112DE0: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:266) > ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9112E26: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:282) > ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9112D52: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:307) > ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9114502: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:574) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91145F7: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A43B: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A450: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A47E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x911454E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:597) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084E80: btl_openib_handle_incoming (btl_openib_component.c:3099) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90845D2: btl_openib_handle_incoming (btl_openib_component.c:3104) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x90845D8: btl_openib_handle_incoming (btl_openib_component.c:3119) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90845DE: btl_openib_handle_incoming (btl_openib_component.c:3119) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90848DF: btl_openib_handle_incoming (btl_openib_component.c:3170) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084A3F: btl_openib_handle_incoming (btl_openib_endpoint.h:464) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x902E6AE: opal_convertor_set_position.part.8 (opal_convertor.h:297) > ==15577== by 0x9115A4E: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:156) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x902E6D8: opal_convertor_set_position.part.8 (opal_convertor.h:303) > ==15577== by 0x9115A4E: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:156) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91156F9: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:467) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) > ==15577== by 0x44280E: DataRead (readbinary3d.c:791) > ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x43D2C0: ReadBinary (readbinary3d.c:214) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44631F: DataPartitionElements (readbinary3d.c:1190) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x446355: DataPartitionElements (readbinary3d.c:1190) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x446355: DataPartitionElements (readbinary3d.c:1190) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x446355: DataPartitionElements (readbinary3d.c:1190) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4463C6: DataPartitionElements (readbinary3d.c:1191) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x4463FC: DataPartitionElements (readbinary3d.c:1191) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x4463FC: DataPartitionElements (readbinary3d.c:1191) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4463FC: DataPartitionElements (readbinary3d.c:1191) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x446597: DataPartitionElements (readbinary3d.c:1196) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4464ED: DataPartitionElements (readbinary3d.c:1202) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44650F: DataPartitionElements (readbinary3d.c:1203) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x446531: DataPartitionElements (readbinary3d.c:1204) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x446553: DataPartitionElements (readbinary3d.c:1205) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4EE790A: PetscStrallocpy (str.c:188) > ==15577== by 0x4FBF60A: PetscClassRegLogRegister (classlog.c:253) > ==15577== by 0x4FBBCCE: PetscClassIdRegister (plog.c:2570) > ==15577== by 0x5481173: MatMFFDInitializePackage (mffd.c:57) > ==15577== by 0x52A42A0: MatInitializePackage (dlregismat.c:88) > ==15577== by 0x577B3AF: MatCreate (gcreate.c:60) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F0997C: PetscFunctionListAdd_Private (reg.c:180) > ==15577== by 0x548274A: MatMFFDRegister (mffd.c:205) > ==15577== by 0x548EBF7: MatMFFDRegisterAll (mfregis.c:28) > ==15577== by 0x54811CD: MatMFFDInitializePackage (mffd.c:59) > ==15577== by 0x52A42A0: MatInitializePackage (dlregismat.c:88) > ==15577== by 0x577B3AF: MatCreate (gcreate.c:60) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F09C8C: PetscFunctionListAdd_Private (reg.c:210) > ==15577== by 0x548274A: MatMFFDRegister (mffd.c:205) > ==15577== by 0x548EC62: MatMFFDRegisterAll (mfregis.c:29) > ==15577== by 0x54811CD: MatMFFDInitializePackage (mffd.c:59) > ==15577== by 0x52A42A0: MatInitializePackage (dlregismat.c:88) > ==15577== by 0x577B3AF: MatCreate (gcreate.c:60) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x529D004: MatRegisterBaseName (matreg.c:170) > ==15577== by 0x529D970: MatRegisterAll (matregis.c:97) > ==15577== by 0x52A457C: MatInitializePackage (dlregismat.c:97) > ==15577== by 0x577B3AF: MatCreate (gcreate.c:60) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x577B42F: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x577B47F: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x577B4D6: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9037693: ompi_comm_nextcid (comm_cid.c:236) > ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) > ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) > ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90376EA: ompi_comm_nextcid (comm_cid.c:255) > ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) > ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) > ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91E6821: opal_pointer_array_set_item (opal_pointer_array.c:172) > ==15577== by 0x903778C: ompi_comm_nextcid (comm_cid.c:272) > ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) > ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) > ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91E6837: opal_pointer_array_set_item (opal_pointer_array.c:189) > ==15577== by 0x903778C: ompi_comm_nextcid (comm_cid.c:272) > ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) > ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) > ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91E6844: opal_pointer_array_set_item (opal_pointer_array.c:193) > ==15577== by 0x903778C: ompi_comm_nextcid (comm_cid.c:272) > ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) > ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) > ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91E6846: opal_pointer_array_set_item (opal_pointer_array.c:205) > ==15577== by 0x903778C: ompi_comm_nextcid (comm_cid.c:272) > ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) > ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) > ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xAF16CBD: vfprintf (in /lib64/libc-2.12.so) > ==15577== by 0xAF3F201: vsnprintf (in /lib64/libc-2.12.so) > ==15577== by 0xAF1F0D2: snprintf (in /lib64/libc-2.12.so) > ==15577== by 0x9034558: ompi_comm_dup (comm.c:687) > ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) > ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xAF151DC: vfprintf (in /lib64/libc-2.12.so) > ==15577== by 0xAF3F201: vsnprintf (in /lib64/libc-2.12.so) > ==15577== by 0xAF1F0D2: snprintf (in /lib64/libc-2.12.so) > ==15577== by 0x9034558: ompi_comm_dup (comm.c:687) > ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) > ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x910CD57: mca_pml_ob1_add_comm (pml_ob1.c:191) > ==15577== by 0x9037A6F: ompi_comm_activate (comm_cid.c:427) > ==15577== by 0x9034576: ompi_comm_dup (comm.c:691) > ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) > ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F51A35: PetscCommDuplicate (tagm.c:151) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F2A82D: PetscThreadCommCreate (threadcomm.c:146) > ==15577== by 0x4F32F1D: PetscThreadCommWorldInitialize (threadcomm.c:1228) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F2A8F9: PetscThreadCommCreate (threadcomm.c:150) > ==15577== by 0x4F32F1D: PetscThreadCommWorldInitialize (threadcomm.c:1228) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F231A2: PetscMallocValidate (mtr.c:139) > ==15577== by 0x4F23DA4: PetscTrFreeDefault (mtr.c:266) > ==15577== by 0x4F8CD1D: PetscOptionsEnd_Private (aoptions.c:486) > ==15577== by 0x4F2C03B: PetscThreadCommSetNThreads (threadcomm.c:340) > ==15577== by 0x4F32F93: PetscThreadCommWorldInitialize (threadcomm.c:1230) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F2C6E9: PetscThreadCommSetAffinities (threadcomm.c:423) > ==15577== by 0x4F32FF9: PetscThreadCommWorldInitialize (threadcomm.c:1231) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F3307C: PetscThreadCommWorldInitialize (threadcomm.c:1232) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F333DC: PetscThreadCommWorldInitialize (threadcomm.c:1240) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F334C0: PetscThreadCommWorldInitialize (threadcomm.c:1241) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F3668F: PetscThreadCommReductionCreate (threadcommred.c:432) > ==15577== by 0x4F336BD: PetscThreadCommWorldInitialize (threadcomm.c:1251) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F3674F: PetscThreadCommReductionCreate (threadcommred.c:435) > ==15577== by 0x4F336BD: PetscThreadCommWorldInitialize (threadcomm.c:1251) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F367F2: PetscThreadCommReductionCreate (threadcommred.c:436) > ==15577== by 0x4F336BD: PetscThreadCommWorldInitialize (threadcomm.c:1251) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F368AC: PetscThreadCommReductionCreate (threadcommred.c:440) > ==15577== by 0x4F336BD: PetscThreadCommWorldInitialize (threadcomm.c:1251) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F36A3A: PetscThreadCommReductionCreate (threadcommred.c:448) > ==15577== by 0x4F336BD: PetscThreadCommWorldInitialize (threadcomm.c:1251) > ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) > ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) > ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F579F1: PetscHeaderCreate_Private (inherit.c:68) > ==15577== by 0x577B546: MatCreate (gcreate.c:62) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x509452A: PetscLayoutCreate (pmap.c:53) > ==15577== by 0x577B60E: MatCreate (gcreate.c:63) > ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x577BBC1: MatSetSizes (gcreate.c:110) > ==15577== by 0x5498760: MatCreateMPIAdj (mpiadj.c:657) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x577BCB3: MatSetSizes (gcreate.c:110) > ==15577== by 0x5498760: MatCreateMPIAdj (mpiadj.c:657) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x577BD73: MatSetSizes (gcreate.c:112) > ==15577== by 0x5498760: MatCreateMPIAdj (mpiadj.c:657) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x577BD80: MatSetSizes (gcreate.c:112) > ==15577== by 0x5498760: MatCreateMPIAdj (mpiadj.c:657) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5497EEB: MatCreate_MPIAdj (mpiadj.c:576) > ==15577== by 0x529C6B9: MatSetType (matreg.c:74) > ==15577== by 0x54987CB: MatCreateMPIAdj (mpiadj.c:658) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5094D7F: PetscLayoutSetUp (pmap.c:146) > ==15577== by 0x549681D: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:450) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5094F50: PetscLayoutSetUp (pmap.c:150) > ==15577== by 0x549681D: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:450) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF6740: PetscSplitOwnership (psplit.c:81) > ==15577== by 0x5094FD4: PetscLayoutSetUp (pmap.c:152) > ==15577== by 0x549681D: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:450) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50950BE: PetscLayoutSetUp (pmap.c:156) > ==15577== by 0x549681D: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:450) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5094F87: PetscLayoutSetUp (pmap.c:151) > ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF6733: PetscSplitOwnership (psplit.c:81) > ==15577== by 0x5094FD4: PetscLayoutSetUp (pmap.c:152) > ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF6790: PetscSplitOwnership (psplit.c:83) > ==15577== by 0x5094FD4: PetscLayoutSetUp (pmap.c:152) > ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9111ADC: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:114) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111ADE: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:120) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111B3D: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:136) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111B85: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:470) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9111BD3: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:686) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5496A42: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:455) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x5496B24: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:458) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5496A6E: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:459) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5496A98: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:459) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x5496B81: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:468) > ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) > ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) > ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x520E2D2: MatPartitioningCreate (partition.c:403) > ==15577== by 0x44663A: DataPartitionElements (readbinary3d.c:1220) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x520E322: MatPartitioningCreate (partition.c:403) > ==15577== by 0x44663A: DataPartitionElements (readbinary3d.c:1220) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x520E379: MatPartitioningCreate (partition.c:403) > ==15577== by 0x44663A: DataPartitionElements (readbinary3d.c:1220) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084A50: btl_openib_handle_incoming (btl_openib_endpoint.h:465) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x520E3E9: MatPartitioningCreate (partition.c:403) > ==15577== by 0x44663A: DataPartitionElements (readbinary3d.c:1220) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084A5A: btl_openib_handle_incoming (btl_openib_endpoint.h:465) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x520E3E9: MatPartitioningCreate (partition.c:403) > ==15577== by 0x44663A: DataPartitionElements (readbinary3d.c:1220) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50CB3A1: ISCreate (isreg.c:41) > ==15577== by 0x50B113E: ISCreateStride (stride.c:417) > ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) > ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) > ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50CB3F7: ISCreate (isreg.c:41) > ==15577== by 0x50B113E: ISCreateStride (stride.c:417) > ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) > ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) > ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50CB454: ISCreate (isreg.c:41) > ==15577== by 0x50B113E: ISCreateStride (stride.c:417) > ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) > ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) > ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50B1517: ISCreate_Stride (stride.c:432) > ==15577== by 0x50CBC08: ISSetType (isreg.c:87) > ==15577== by 0x50B11A9: ISCreateStride (stride.c:418) > ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) > ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) > ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B093F: ISStrideSetStride (stride.c:352) > ==15577== by 0x50B1219: ISCreateStride (stride.c:419) > ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) > ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) > ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B0E84: ISStrideSetStride_Stride (stride.c:373) > ==15577== by 0x50B0A23: ISStrideSetStride (stride.c:353) > ==15577== by 0x50B1219: ISCreateStride (stride.c:419) > ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) > ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) > ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B0EA5: ISStrideSetStride_Stride (stride.c:374) > ==15577== by 0x50B0A23: ISStrideSetStride (stride.c:353) > ==15577== by 0x50B1219: ISCreateStride (stride.c:419) > ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) > ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) > ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50AE960: ISGetIndices_Stride (stride.c:160) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x50AE997: ISGetIndices_Stride (stride.c:160) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x50AE997: ISGetIndices_Stride (stride.c:160) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50AE997: ISGetIndices_Stride (stride.c:160) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50AEA0A: ISGetIndices_Stride (stride.c:161) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50AEA72: ISGetIndices_Stride (stride.c:163) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D2463: ISPartitioningToNumbering (iscoloring.c:330) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D2533: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x50D2566: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x50D2566: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50D2566: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D2584: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x50D25B7: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x50D25B7: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50D25B7: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D25D1: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x50D2604: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x50D2604: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50D2604: ISPartitioningToNumbering (iscoloring.c:339) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50CEF3C: PetscMemzero (petscsys.h:1716) > ==15577== by 0x50D2690: ISPartitioningToNumbering (iscoloring.c:340) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) > ==15577== by 0x50CEF99: PetscMemzero (petscsys.h:1735) > ==15577== by 0x50D2690: ISPartitioningToNumbering (iscoloring.c:340) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) > ==15577== by 0x50CEF99: PetscMemzero (petscsys.h:1735) > ==15577== by 0x50D2690: ISPartitioningToNumbering (iscoloring.c:340) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D2724: ISPartitioningToNumbering (iscoloring.c:341) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x904F6AD: PMPI_Allreduce (pallreduce.c:87) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x904F6F2: PMPI_Allreduce (pallreduce.c:96) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90A0D42: ompi_coll_tuned_allreduce_intra_dec_fixed (coll_tuned_decision_fixed.c:60) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DF202: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:87) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91DF240: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:104) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DF07B: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3B4: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3BA: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3E0: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A427: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DF0A4: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) > ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x910FCA8: mca_pml_ob1_isend (pml_ob1_isend.c:76) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DBB78: opal_convertor_prepare_for_send (opal_convertor.c:559) > ==15577== by 0x91100C0: mca_pml_ob1_isend (opal_convertor.h:237) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x910FDAA: mca_pml_ob1_isend (pml_ob1_sendreq.h:360) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x911002C: mca_pml_ob1_isend (pml_ob1_sendreq.h:372) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90797D8: mca_btl_sm_sendi (btl_sm.c:815) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x907984E: mca_btl_sm_sendi (btl_sm.c:840) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DB575: opal_convertor_pack (opal_convertor.c:232) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A333: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A35E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A368: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9111BE8: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:194) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9111C64: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvreq.h:168) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116204: mca_pml_ob1_send_request_free (pml_ob1_sendreq.c:117) > ==15577== by 0x9043BC9: ompi_request_default_wait_all (request.h:350) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90F04C8: ompi_op_base_sum_int64_t (op_base_functions.c:264) > ==15577== by 0x90A3533: ompi_coll_tuned_allreduce_intra_recursivedoubling (op.h:498) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90F04E3: ompi_op_base_sum_int64_t (op_base_functions.c:264) > ==15577== by 0x90A3533: ompi_coll_tuned_allreduce_intra_recursivedoubling (op.h:498) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90F04C8: ompi_op_base_sum_int64_t (op_base_functions.c:264) > ==15577== by 0x90A3573: ompi_coll_tuned_allreduce_intra_recursivedoubling (op.h:498) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90F04E3: ompi_op_base_sum_int64_t (op_base_functions.c:264) > ==15577== by 0x90A3573: ompi_coll_tuned_allreduce_intra_recursivedoubling (op.h:498) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x907D557: mca_btl_openib_sendi (btl_openib_frag.h:396) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x907D568: mca_btl_openib_sendi (btl_openib_frag.h:396) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x907D5E5: mca_btl_openib_sendi (btl_openib.c:1533) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x907D6AB: mca_btl_openib_sendi (btl_openib.c:1573) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DB575: opal_convertor_pack (opal_convertor.c:232) > ==15577== by 0x907DB50: mca_btl_openib_sendi (btl_openib.c:1582) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A37E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907DB50: mca_btl_openib_sendi (btl_openib.c:1582) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x907D777: mca_btl_openib_sendi (btl_openib_endpoint.h:305) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0F5A: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x907D977: mca_btl_openib_sendi (verbs.h:1111) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x905EDD2: PMPI_Scan (pscan.c:83) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x905EC62: PMPI_Scan (pscan.c:92) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DF202: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:87) > ==15577== by 0x90BDBED: mca_coll_basic_scan_intra (ompi_datatype.h:290) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91DF240: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:104) > ==15577== by 0x90BDBED: mca_coll_basic_scan_intra (ompi_datatype.h:290) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DF07B: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) > ==15577== by 0x90BDBED: mca_coll_basic_scan_intra (ompi_datatype.h:290) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DF0A4: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) > ==15577== by 0x90BDBED: mca_coll_basic_scan_intra (ompi_datatype.h:290) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90F04C8: ompi_op_base_sum_int64_t (op_base_functions.c:264) > ==15577== by 0x90BDCFB: mca_coll_basic_scan_intra (op.h:498) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90F04E3: ompi_op_base_sum_int64_t (op_base_functions.c:264) > ==15577== by 0x90BDCFB: mca_coll_basic_scan_intra (op.h:498) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91103AA: mca_pml_ob1_send (pml_ob1_isend.c:108) > ==15577== by 0x90BDB8E: mca_coll_basic_scan_intra (coll_basic_scan.c:122) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DBB78: opal_convertor_prepare_for_send (opal_convertor.c:559) > ==15577== by 0x9110A69: mca_pml_ob1_send (opal_convertor.h:237) > ==15577== by 0x90BDB8E: mca_coll_basic_scan_intra (coll_basic_scan.c:122) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91104AF: mca_pml_ob1_send (pml_ob1_sendreq.h:360) > ==15577== by 0x90BDB8E: mca_coll_basic_scan_intra (coll_basic_scan.c:122) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9110840: mca_pml_ob1_send (pml_ob1_sendreq.h:372) > ==15577== by 0x90BDB8E: mca_coll_basic_scan_intra (coll_basic_scan.c:122) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116204: mca_pml_ob1_send_request_free (pml_ob1_sendreq.c:117) > ==15577== by 0x91109F0: mca_pml_ob1_send (request.h:350) > ==15577== by 0x90BDB8E: mca_coll_basic_scan_intra (coll_basic_scan.c:122) > ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) > ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) > ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D28B4: ISPartitioningToNumbering (iscoloring.c:344) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D2947: ISPartitioningToNumbering (iscoloring.c:345) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D2958: ISPartitioningToNumbering (iscoloring.c:353) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x50D298B: ISPartitioningToNumbering (iscoloring.c:353) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x50D298B: ISPartitioningToNumbering (iscoloring.c:353) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50D298B: ISPartitioningToNumbering (iscoloring.c:353) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D2A41: ISPartitioningToNumbering (iscoloring.c:354) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x50D2A6E: ISPartitioningToNumbering (iscoloring.c:355) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x50D2A6E: ISPartitioningToNumbering (iscoloring.c:355) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F22C01: PetscMemzero (petscsys.h:1716) > ==15577== by 0x4F2417A: PetscTrFreeDefault (mtr.c:308) > ==15577== by 0x50D2A6E: ISPartitioningToNumbering (iscoloring.c:355) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) > ==15577== by 0x4F22C5E: PetscMemzero (petscsys.h:1735) > ==15577== by 0x4F2417A: PetscTrFreeDefault (mtr.c:308) > ==15577== by 0x50D2A6E: ISPartitioningToNumbering (iscoloring.c:355) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) > ==15577== by 0x4F22C5E: PetscMemzero (petscsys.h:1735) > ==15577== by 0x4F2417A: PetscTrFreeDefault (mtr.c:308) > ==15577== by 0x50D2A6E: ISPartitioningToNumbering (iscoloring.c:355) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x50D2AA6: ISPartitioningToNumbering (iscoloring.c:355) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x50D2AA6: ISPartitioningToNumbering (iscoloring.c:355) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x50D2ADE: ISPartitioningToNumbering (iscoloring.c:355) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x50D2ADE: ISPartitioningToNumbering (iscoloring.c:355) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x50AECB3: ISRestoreIndices_Stride (stride.c:175) > ==15577== by 0x50C47D1: ISRestoreIndices (index.c:452) > ==15577== by 0x50D2B63: ISPartitioningToNumbering (iscoloring.c:357) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x50AECB3: ISRestoreIndices_Stride (stride.c:175) > ==15577== by 0x50C47D1: ISRestoreIndices (index.c:452) > ==15577== by 0x50D2B63: ISPartitioningToNumbering (iscoloring.c:357) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50BFC1B: ISCreate_General (general.c:510) > ==15577== by 0x50CBC08: ISSetType (isreg.c:87) > ==15577== by 0x50BEEC1: ISCreateGeneral (general.c:438) > ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50BF554: ISGeneralSetIndices_General (general.c:480) > ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) > ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) > ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50BF59E: ISGeneralSetIndices_General (general.c:481) > ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) > ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) > ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50BE9DD: ISCreateGeneral_Private (general.c:378) > ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) > ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) > ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) > ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50BE9C5: ISCreateGeneral_Private (general.c:379) > ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) > ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) > ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) > ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50BE9E6: ISCreateGeneral_Private (general.c:381) > ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) > ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) > ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) > ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50BEA90: ISCreateGeneral_Private (general.c:383) > ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) > ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) > ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) > ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50BEA31: ISCreateGeneral_Private (general.c:384) > ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) > ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) > ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) > ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50BEA67: ISCreateGeneral_Private (general.c:385) > ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) > ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) > ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) > ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x5493F54: MatDestroy_MPIAdj (mpiadj.c:73) > ==15577== by 0x523BDF2: MatDestroy (matrix.c:1029) > ==15577== by 0x4468C1: DataPartitionElements (readbinary3d.c:1246) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x5493F54: MatDestroy_MPIAdj (mpiadj.c:73) > ==15577== by 0x523BDF2: MatDestroy (matrix.c:1029) > ==15577== by 0x4468C1: DataPartitionElements (readbinary3d.c:1246) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x5493FFB: MatDestroy_MPIAdj (mpiadj.c:74) > ==15577== by 0x523BDF2: MatDestroy (matrix.c:1029) > ==15577== by 0x4468C1: DataPartitionElements (readbinary3d.c:1246) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x5493FFB: MatDestroy_MPIAdj (mpiadj.c:74) > ==15577== by 0x523BDF2: MatDestroy (matrix.c:1029) > ==15577== by 0x4468C1: DataPartitionElements (readbinary3d.c:1246) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x446C58: DataMoveGeneric (readbinary3d.c:1297) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D30A5: ISPartitioningCount (iscoloring.c:415) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2BC: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A32D: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A350: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A353: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A35E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D3172: ISPartitioningCount (iscoloring.c:418) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50D3207: ISPartitioningCount (iscoloring.c:427) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D331A: ISPartitioningCount (iscoloring.c:429) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x902E5D0: append_frag_to_list (pml_ob1_recvfrag.c:72) > ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x902E608: append_frag_to_list (pml_ob1_recvfrag.c:72) > ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x906F575: mca_allocator_bucket_alloc (allocator_bucket_alloc.c:91) > ==15577== by 0x902E638: append_frag_to_list (pml_ob1_recvfrag.c:72) > ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x906F591: mca_allocator_bucket_alloc (allocator_bucket_alloc.c:91) > ==15577== by 0x902E638: append_frag_to_list (pml_ob1_recvfrag.c:72) > ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x902E664: append_frag_to_list (pml_ob1_recvfrag.c:72) > ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x902E664: append_frag_to_list (pml_ob1_recvfrag.c:72) > ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9113982: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.c:632) > ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A3243: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:216) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91139B7: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.c:632) > ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A3243: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:216) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115C7B: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1037) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A3243: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:216) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) > ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4FC852D: EventRegLogRegister (eventlog.c:310) > ==15577== by 0x4FA6657: PetscLogEventRegister (plog.c:753) > ==15577== by 0x517D85C: VecInitializePackage (dlregisvec.c:127) > ==15577== by 0x517AB60: VecCreate (veccreate.c:35) > ==15577== by 0x446D7D: DataMoveGeneric (readbinary3d.c:1306) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4FC7EE7: EventPerfLogEnsureSize (eventlog.c:211) > ==15577== by 0x4FA66E5: PetscLogEventRegister (plog.c:755) > ==15577== by 0x517D85C: VecInitializePackage (dlregisvec.c:127) > ==15577== by 0x517AB60: VecCreate (veccreate.c:35) > ==15577== by 0x446D7D: DataMoveGeneric (readbinary3d.c:1306) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x517ABE0: VecCreate (veccreate.c:37) > ==15577== by 0x446D7D: DataMoveGeneric (readbinary3d.c:1306) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x517AC30: VecCreate (veccreate.c:37) > ==15577== by 0x446D7D: DataMoveGeneric (readbinary3d.c:1306) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x517AC87: VecCreate (veccreate.c:37) > ==15577== by 0x446D7D: DataMoveGeneric (readbinary3d.c:1306) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x517B71C: VecSetType (vecreg.c:49) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x514FDC7: VecCreate_MPI_Private (pbvec.c:185) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5094D7F: PetscLayoutSetUp (pmap.c:146) > ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5094F50: PetscLayoutSetUp (pmap.c:150) > ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084E1A: btl_openib_handle_incoming (btl_openib_component.c:1374) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9084E1D: btl_openib_handle_incoming (btl_openib_component.c:1374) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9083BD8: btl_openib_control (btl_openib_component.c:478) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9083BDD: btl_openib_control (btl_openib_component.c:478) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9083DD0: btl_openib_control (btl_openib_component.c:507) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x514FFC2: VecCreate_MPI_Private (pbvec.c:197) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x514FFF5: VecCreate_MPI_Private (pbvec.c:197) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x514FFF5: VecCreate_MPI_Private (pbvec.c:197) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x514FFF5: VecCreate_MPI_Private (pbvec.c:197) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5150068: VecCreate_MPI_Private (pbvec.c:198) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x514E9F4: PetscMemzero (petscsys.h:1716) > ==15577== by 0x515010F: VecCreate_MPI_Private (pbvec.c:199) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) > ==15577== by 0x514EA51: PetscMemzero (petscsys.h:1735) > ==15577== by 0x515010F: VecCreate_MPI_Private (pbvec.c:199) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) > ==15577== by 0x514EA51: PetscMemzero (petscsys.h:1735) > ==15577== by 0x515010F: VecCreate_MPI_Private (pbvec.c:199) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084E65: btl_openib_handle_incoming (btl_openib_component.c:3106) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) > ==15577== by 0x51F6263: VecStashCreate_Private (vecstash.c:31) > ==15577== by 0x51501C0: VecCreate_MPI_Private (pbvec.c:211) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90848C9: btl_openib_handle_incoming (btl_openib_component.c:3167) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) > ==15577== by 0x51F6263: VecStashCreate_Private (vecstash.c:31) > ==15577== by 0x51501C0: VecCreate_MPI_Private (pbvec.c:211) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9084903: btl_openib_handle_incoming (btl_openib_component.c:3173) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) > ==15577== by 0x51F6263: VecStashCreate_Private (vecstash.c:31) > ==15577== by 0x51501C0: VecCreate_MPI_Private (pbvec.c:211) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51F6454: VecStashCreate_Private (vecstash.c:37) > ==15577== by 0x51501C0: VecCreate_MPI_Private (pbvec.c:211) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x446FE3: DataMoveGeneric (readbinary3d.c:1319) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x446FE3: DataMoveGeneric (readbinary3d.c:1319) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x446FE3: DataMoveGeneric (readbinary3d.c:1319) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B5E51: ISCreateBlock (block.c:376) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B5E9B: ISCreateBlock (block.c:377) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084A66: btl_openib_handle_incoming (btl_openib_endpoint.h:475) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084A76: btl_openib_handle_incoming (btl_openib_endpoint.h:475) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088709: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:768) > ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088C4F: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:773) > ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x908872E: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.h:526) > ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088747: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:789) > ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088769: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:799) > ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088773: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:799) > ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088792: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:802) > ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088822: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.h:322) > ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C13DC: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9088991: mca_btl_openib_endpoint_send_credits (verbs.h:1111) > ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C144B: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9088991: mca_btl_openib_endpoint_send_credits (verbs.h:1111) > ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088EC2: mca_btl_openib_endpoint_credits (btl_openib_endpoint.h:455) > ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088ECF: mca_btl_openib_endpoint_credits (btl_openib_endpoint.c:745) > ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088EDA: mca_btl_openib_endpoint_credits (btl_openib_endpoint.c:745) > ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088F04: mca_btl_openib_endpoint_credits (btl_openib_endpoint.c:748) > ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088F80: mca_btl_openib_endpoint_credits (btl_openib_endpoint.c:748) > ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088F95: mca_btl_openib_endpoint_credits (btl_openib_endpoint.c:748) > ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088F19: mca_btl_openib_endpoint_credits (btl_openib_endpoint.h:465) > ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9088F1D: mca_btl_openib_endpoint_credits (btl_openib_endpoint.h:465) > ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50B7D85: ISCreate_Block (block.c:568) > ==15577== by 0x50CBC08: ISSetType (isreg.c:87) > ==15577== by 0x50B5FD0: ISCreateBlock (block.c:380) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x907D5F4: mca_btl_openib_sendi (btl_openib_endpoint.h:526) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50B5718: ISBlockSetIndices_Block (block.c:307) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B57D3: ISBlockSetIndices_Block (block.c:308) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B57BB: ISBlockSetIndices_Block (block.c:309) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B57DC: ISBlockSetIndices_Block (block.c:311) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B5886: ISBlockSetIndices_Block (block.c:313) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B5827: ISBlockSetIndices_Block (block.c:314) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B585D: ISBlockSetIndices_Block (block.c:315) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B589D: ISBlockSetIndices_Block (block.c:318) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x50B58D4: ISBlockSetIndices_Block (block.c:318) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x50B58D4: ISBlockSetIndices_Block (block.c:318) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50B58D4: ISBlockSetIndices_Block (block.c:318) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B5948: ISBlockSetIndices_Block (block.c:319) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B1DDE: PetscMemcpy (petscsys.h:1656) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B1E2F: PetscMemcpy (petscsys.h:1657) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B1EA8: PetscMemcpy (petscsys.h:1663) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B1EBF: PetscMemcpy (petscsys.h:1663) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) > ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3E0: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) > ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) > ==15577== by 0x50B6046: ISCreateBlock (block.c:381) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x447225: DataMoveGeneric (readbinary3d.c:1329) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x447225: DataMoveGeneric (readbinary3d.c:1329) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5174EA5: VecSetSizes (vector.c:1430) > ==15577== by 0x514CF9A: VecCreateSeq (vseqcr.c:38) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5174F91: VecSetSizes (vector.c:1430) > ==15577== by 0x514CF9A: VecCreateSeq (vseqcr.c:38) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5174FE8: VecSetSizes (vector.c:1431) > ==15577== by 0x514CF9A: VecCreateSeq (vseqcr.c:38) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5174FF2: VecSetSizes (vector.c:1431) > ==15577== by 0x514CF9A: VecCreateSeq (vseqcr.c:38) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x517B71C: VecSetType (vecreg.c:49) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x514D594: VecCreate_Seq (bvec3.c:38) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x514D5C7: VecCreate_Seq (bvec3.c:38) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x514D5C7: VecCreate_Seq (bvec3.c:38) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x514D5C7: VecCreate_Seq (bvec3.c:38) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x514D637: VecCreate_Seq (bvec3.c:39) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5141A03: VecCreate_Seq_Private (bvec2.c:1245) > ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5094D7F: PetscLayoutSetUp (pmap.c:146) > ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) > ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5094D90: PetscLayoutSetUp (pmap.c:146) > ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) > ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5094F50: PetscLayoutSetUp (pmap.c:150) > ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) > ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5094F87: PetscLayoutSetUp (pmap.c:151) > ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) > ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF685C: PetscSplitOwnership (psplit.c:85) > ==15577== by 0x5094FD4: PetscLayoutSetUp (pmap.c:152) > ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) > ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF6A4B: PetscSplitOwnership (psplit.c:93) > ==15577== by 0x5094FD4: PetscLayoutSetUp (pmap.c:152) > ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) > ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5145D6F: PetscMemzero (petscsys.h:1716) > ==15577== by 0x5149E73: VecSet_Seq (dvec2.c:729) > ==15577== by 0x5185F76: VecSet (rvector.c:571) > ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) > ==15577== by 0x5145DCC: PetscMemzero (petscsys.h:1735) > ==15577== by 0x5149E73: VecSet_Seq (dvec2.c:729) > ==15577== by 0x5185F76: VecSet (rvector.c:571) > ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) > ==15577== by 0x5145DCC: PetscMemzero (petscsys.h:1735) > ==15577== by 0x5149E73: VecSet_Seq (dvec2.c:729) > ==15577== by 0x5185F76: VecSet (rvector.c:571) > ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F8811A: PetscObjectComposedDataIncreaseReal (state.c:169) > ==15577== by 0x518619F: VecSet (rvector.c:577) > ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4F8822D: PetscObjectComposedDataIncreaseReal (state.c:171) > ==15577== by 0x518619F: VecSet (rvector.c:577) > ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x83FA903: sqrt (in /lib64/libm-2.12.so) > ==15577== by 0x518635A: VecSet (rvector.c:579) > ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x83FA90D: sqrt (in /lib64/libm-2.12.so) > ==15577== by 0x518635A: VecSet (rvector.c:579) > ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) > ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51B46C3: VecScatterCreate (vscat.c:1028) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51B4713: VecScatterCreate (vscat.c:1028) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51B476A: VecScatterCreate (vscat.c:1028) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3BA: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x51BAC0B: VecScatterCreate (vscat.c:1548) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3D1: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x51BAC0B: VecScatterCreate (vscat.c:1548) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3DA: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x51BAC0B: VecScatterCreate (vscat.c:1548) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B27A4: ISGetIndices_Block (block.c:45) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x51BB45D: VecScatterCreate (vscat.c:1583) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x50B27DC: ISGetIndices_Block (block.c:45) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x51BB45D: VecScatterCreate (vscat.c:1583) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x50B27DC: ISGetIndices_Block (block.c:45) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x51BB45D: VecScatterCreate (vscat.c:1583) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x50B27DC: ISGetIndices_Block (block.c:45) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x51BB45D: VecScatterCreate (vscat.c:1583) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B28CC: ISGetIndices_Block (block.c:49) > ==15577== by 0x50C3F71: ISGetIndices (index.c:374) > ==15577== by 0x51BB45D: VecScatterCreate (vscat.c:1583) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51BB4C4: VecScatterCreate (vscat.c:1584) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9084E1D: btl_openib_handle_incoming (btl_openib_component.c:1374) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) > ==15577== by 0x4F51037: PetscObjectGetNewTag (tagm.c:47) > ==15577== by 0x51E1339: VecScatterCreate_PtoS (vpscat.c:1736) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9084F20: btl_openib_handle_incoming (btl_openib_component.c:1382) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) > ==15577== by 0x4F51037: PetscObjectGetNewTag (tagm.c:47) > ==15577== by 0x51E1339: VecScatterCreate_PtoS (vpscat.c:1736) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Invalid read of size 4 > ==15577== at 0x902E5D0: append_frag_to_list (pml_ob1_recvfrag.c:72) > ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) > ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) > ==15577== by 0x4F51037: PetscObjectGetNewTag (tagm.c:47) > ==15577== by 0x51E1339: VecScatterCreate_PtoS (vpscat.c:1736) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== Address 0x145854be is 209,982 bytes inside a block of size 209,984 alloc'd > ==15577== at 0x4C266CA: memalign (vg_replace_malloc.c:727) > ==15577== by 0x4C26765: posix_memalign (vg_replace_malloc.c:876) > ==15577== by 0x90ED708: mca_mpool_rdma_alloc (mpool_rdma_module.c:103) > ==15577== by 0x908903C: mca_btl_openib_endpoint_connect_eager_rdma (btl_openib_endpoint.c:947) > ==15577== by 0x908592B: poll_device (btl_openib_component.c:3424) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) > ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) > ==15577== by 0x50BE922: ISCreateGeneral_Private (general.c:377) > ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) > ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) > ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) > ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) > ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) > ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E1615: VecScatterCreate_PtoS (vpscat.c:1745) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E1635: VecScatterCreate_PtoS (vpscat.c:1745) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x51E166E: VecScatterCreate_PtoS (vpscat.c:1745) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x51E166E: VecScatterCreate_PtoS (vpscat.c:1745) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E166E: VecScatterCreate_PtoS (vpscat.c:1745) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E18C9: VecScatterCreate_PtoS (vpscat.c:1750) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E17B3: VecScatterCreate_PtoS (vpscat.c:1752) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E17DD: VecScatterCreate_PtoS (vpscat.c:1754) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4EF806D: PetscGatherNumberOfMessages (mpimesg.c:45) > ==15577== by 0x51E1938: VecScatterCreate_PtoS (vpscat.c:1767) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4EF80BD: PetscGatherNumberOfMessages (mpimesg.c:45) > ==15577== by 0x51E1938: VecScatterCreate_PtoS (vpscat.c:1767) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF86B1: PetscGatherMessageLengths (mpimesg.c:112) > ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF8715: PetscGatherMessageLengths (mpimesg.c:112) > ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF87EB: PetscGatherMessageLengths (mpimesg.c:116) > ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF8998: PetscGatherMessageLengths (mpimesg.c:117) > ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF8AEA: PetscGatherMessageLengths (mpimesg.c:130) > ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF8BC3: PetscGatherMessageLengths (mpimesg.c:133) > ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EF8CA6: PetscGatherMessageLengths (mpimesg.c:134) > ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE4C2B: PetscSortMPIIntWithArray (sorti.c:476) > ==15577== by 0x51E1A36: VecScatterCreate_PtoS (vpscat.c:1769) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE4D72: PetscSortMPIIntWithArray (sorti.c:477) > ==15577== by 0x51E1A36: VecScatterCreate_PtoS (vpscat.c:1769) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E1AC9: VecScatterCreate_PtoS (vpscat.c:1770) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E1B31: VecScatterCreate_PtoS (vpscat.c:1773) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E1B8D: VecScatterCreate_PtoS (vpscat.c:1773) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E1DA8: VecScatterCreate_PtoS (vpscat.c:1775) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E1DBC: VecScatterCreate_PtoS (vpscat.c:1784) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x51E1DF5: VecScatterCreate_PtoS (vpscat.c:1784) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x51E1DF5: VecScatterCreate_PtoS (vpscat.c:1784) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E1DF5: VecScatterCreate_PtoS (vpscat.c:1784) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E1EAE: VecScatterCreate_PtoS (vpscat.c:1784) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E2038: VecScatterCreate_PtoS (vpscat.c:1788) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E23A6: VecScatterCreate_PtoS (vpscat.c:1803) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E2438: VecScatterCreate_PtoS (vpscat.c:1814) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E2509: VecScatterCreate_PtoS (vpscat.c:1817) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E268E: VecScatterCreate_PtoS (vpscat.c:1818) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x51E26D0: VecScatterCreate_PtoS (vpscat.c:1818) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x51E26D0: VecScatterCreate_PtoS (vpscat.c:1818) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E26D0: VecScatterCreate_PtoS (vpscat.c:1818) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E26F5: VecScatterCreate_PtoS (vpscat.c:1818) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E27D6: VecScatterCreate_PtoS (vpscat.c:1819) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E2877: VecScatterCreate_PtoS (vpscat.c:1819) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E297A: VecScatterCreate_PtoS (vpscat.c:1824) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E2D8D: VecScatterCreate_PtoS (vpscat.c:1841) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E2F1B: VecScatterCreate_PtoS (vpscat.c:1845) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E2F9F: VecScatterCreate_PtoS (vpscat.c:1845) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E304F: VecScatterCreate_PtoS (vpscat.c:1845) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E318D: VecScatterCreate_PtoS (vpscat.c:1849) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E31E7: VecScatterCreate_PtoS (vpscat.c:1849) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E3488: VecScatterCreate_PtoS (vpscat.c:1859) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x51E35A4: VecScatterCreate_PtoS (vpscat.c:1866) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x51E35A4: VecScatterCreate_PtoS (vpscat.c:1866) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x51E393C: VecScatterCreate_PtoS (vpscat.c:1874) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x51E393C: VecScatterCreate_PtoS (vpscat.c:1874) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E3A25: VecScatterCreate_PtoS (vpscat.c:1879) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51E3AD6: VecScatterCreate_PtoS (vpscat.c:1880) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E3CC2: VecScatterCreate_PtoS (vpscat.c:1882) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E3B97: VecScatterCreate_PtoS (vpscat.c:1884) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E3BBF: VecScatterCreate_PtoS (vpscat.c:1884) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E3C6D: VecScatterCreate_PtoS (vpscat.c:1887) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E5BAA: VecScatterCreateCommon_PtoS (vpscat.c:2070) > ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E66EB: VecScatterCreateCommon_PtoS (vpscat.c:2100) > ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E69AA: VecScatterCreateCommon_PtoS (vpscat.c:2113) > ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E6E80: VecScatterCreateCommon_PtoS (vpscat.c:2129) > ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51E6E93: VecScatterCreateCommon_PtoS (vpscat.c:2129) > ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51C32DC: VecScatterLocalOptimizeCopy_Private (vpscat.c:263) > ==15577== by 0x51E716C: VecScatterCreateCommon_PtoS (vpscat.c:2175) > ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x50B2B31: ISRestoreIndices_Block (block.c:65) > ==15577== by 0x50C47D1: ISRestoreIndices (index.c:452) > ==15577== by 0x51BB627: VecScatterCreate (vscat.c:1587) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x50B2B31: ISRestoreIndices_Block (block.c:65) > ==15577== by 0x50C47D1: ISRestoreIndices (index.c:452) > ==15577== by 0x51BB627: VecScatterCreate (vscat.c:1587) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x50B21FC: ISDestroy_Block (block.c:24) > ==15577== by 0x50C29BD: ISDestroy (index.c:220) > ==15577== by 0x4474FD: DataMoveGeneric (readbinary3d.c:1349) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x50B21FC: ISDestroy_Block (block.c:24) > ==15577== by 0x50C29BD: ISDestroy (index.c:220) > ==15577== by 0x4474FD: DataMoveGeneric (readbinary3d.c:1349) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51BC86C: VecScatterBegin (vscat.c:1718) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51BC880: VecScatterBegin (vscat.c:1718) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51BCA5A: VecScatterBegin (vscat.c:1725) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51BCACA: VecScatterBegin (vscat.c:1726) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51D1012: VecScatterBegin_1 (vpscat.h:81) > ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51BFE32: PetscMemcpy (petscsys.h:1657) > ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) > ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51BFE7F: PetscMemcpy (petscsys.h:1661) > ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) > ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51BFE8D: PetscMemcpy (petscsys.h:1663) > ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) > ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51BFEBB: PetscMemcpy (petscsys.h:1663) > ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) > ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) > ==15577== by 0x51BFF31: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) > ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) > ==15577== by 0x51BFF31: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) > ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x51BFF31: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) > ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x51BFF31: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) > ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x51BFF31: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) > ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) > ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51D242A: VecScatterEnd_1 (vpscat.h:189) > ==15577== by 0x51BD974: VecScatterEnd (vscat.c:1780) > ==15577== by 0x4475E5: DataMoveGeneric (readbinary3d.c:1351) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x51C2E20: VecScatterDestroy_PtoP (vpscat.c:234) > ==15577== by 0x51BE0FD: VecScatterDestroy (vscat.c:1813) > ==15577== by 0x447646: DataMoveGeneric (readbinary3d.c:1352) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x51C2E20: VecScatterDestroy_PtoP (vpscat.c:234) > ==15577== by 0x51BE0FD: VecScatterDestroy (vscat.c:1813) > ==15577== by 0x447646: DataMoveGeneric (readbinary3d.c:1352) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x5140F02: VecDestroy_Seq (bvec2.c:1132) > ==15577== by 0x516CA2A: VecDestroy (vector.c:547) > ==15577== by 0x4476A7: DataMoveGeneric (readbinary3d.c:1353) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x5140F02: VecDestroy_Seq (bvec2.c:1132) > ==15577== by 0x516CA2A: VecDestroy (vector.c:547) > ==15577== by 0x4476A7: DataMoveGeneric (readbinary3d.c:1353) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x44774D: DataMoveGeneric (readbinary3d.c:1355) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x44774D: DataMoveGeneric (readbinary3d.c:1355) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x447841: DataMoveGeneric (readbinary3d.c:1364) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x447841: DataMoveGeneric (readbinary3d.c:1364) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x447841: DataMoveGeneric (readbinary3d.c:1364) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x51554A7: VecDestroy_MPI (pdvec.c:20) > ==15577== by 0x516CA2A: VecDestroy (vector.c:547) > ==15577== by 0x447A21: DataMoveGeneric (readbinary3d.c:1376) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x51554A7: VecDestroy_MPI (pdvec.c:20) > ==15577== by 0x516CA2A: VecDestroy (vector.c:547) > ==15577== by 0x447A21: DataMoveGeneric (readbinary3d.c:1376) > ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90848C9: btl_openib_handle_incoming (btl_openib_component.c:3167) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9084903: btl_openib_handle_incoming (btl_openib_component.c:3173) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) > ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) > ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) > ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) > ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) > ==15577== by 0x517B76F: VecSetType (vecreg.c:53) > ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) > ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) > ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9085177: btl_openib_handle_incoming (btl_openib_component.c:3150) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x90846B5: btl_openib_handle_incoming (btl_openib_endpoint.h:391) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x90846C9: btl_openib_handle_incoming (btl_openib_endpoint.h:398) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084C3B: btl_openib_handle_incoming (btl_openib_endpoint.h:409) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084C50: btl_openib_handle_incoming (btl_openib_endpoint.h:410) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0C62: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9084F8D: btl_openib_handle_incoming (verbs.h:1120) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C0CAF: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9084F8D: btl_openib_handle_incoming (verbs.h:1120) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0xF1C0CD4: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9084F8D: btl_openib_handle_incoming (verbs.h:1120) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0xF1C0D0D: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) > ==15577== by 0x9084F8D: btl_openib_handle_incoming (verbs.h:1120) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084FA8: btl_openib_handle_incoming (btl_openib_endpoint.h:416) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084FBA: btl_openib_handle_incoming (btl_openib_endpoint.h:417) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084A3F: btl_openib_handle_incoming (btl_openib_endpoint.h:464) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084A50: btl_openib_handle_incoming (btl_openib_endpoint.h:465) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084A5A: btl_openib_handle_incoming (btl_openib_endpoint.h:465) > ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) > ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) > ==15577== by 0x43D586: ReadBinary (readbinary3d.c:235) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) > ==15577== by 0x43D586: ReadBinary (readbinary3d.c:235) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) > ==15577== by 0x43D586: ReadBinary (readbinary3d.c:235) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) > ==15577== by 0x43D586: ReadBinary (readbinary3d.c:235) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) > ==15577== by 0x43D586: ReadBinary (readbinary3d.c:235) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) > ==15577== by 0x43D5F9: ReadBinary (readbinary3d.c:236) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) > ==15577== by 0x43D5F9: ReadBinary (readbinary3d.c:236) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) > ==15577== by 0x43D5F9: ReadBinary (readbinary3d.c:236) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) > ==15577== by 0x43D5F9: ReadBinary (readbinary3d.c:236) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) > ==15577== by 0x43D5F9: ReadBinary (readbinary3d.c:236) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x43D75A: ReadBinary (readbinary3d.c:255) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) > ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) > ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) > ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9087052: mca_btl_openib_endpoint_post_send (btl_openib_endpoint.h:526) > ==15577== by 0x90886AC: mca_btl_openib_endpoint_send (btl_openib_endpoint.c:714) > ==15577== by 0x9089275: mca_btl_openib_endpoint_connect_eager_rdma (btl_openib_endpoint.c:901) > ==15577== by 0x908592B: poll_device (btl_openib_component.c:3424) > ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) > ==15577== by 0x4F51037: PetscObjectGetNewTag (tagm.c:47) > ==15577== by 0x51E410A: VecScatterCreateCommon_PtoS (vpscat.c:1929) > ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) > ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) > ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) > ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) > ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) > ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) > ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) > ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) > ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084E32: btl_openib_handle_incoming (btl_openib_component.c:3111) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084E40: btl_openib_handle_incoming (btl_openib_component.c:3113) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x9084E52: btl_openib_handle_incoming (btl_openib_component.c:3113) > ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) > ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) > ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) > ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) > ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) > ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) > ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) > ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) > ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) > ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) > ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B0ED3: ISStrideSetStride_Stride (stride.c:377) > ==15577== by 0x50B0A23: ISStrideSetStride (stride.c:353) > ==15577== by 0x50B1219: ISCreateStride (stride.c:419) > ==15577== by 0x43D898: ReadBinary (readbinary3d.c:263) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50B0EE4: ISStrideSetStride_Stride (stride.c:377) > ==15577== by 0x50B0A23: ISStrideSetStride (stride.c:353) > ==15577== by 0x50B1219: ISCreateStride (stride.c:419) > ==15577== by 0x43D898: ReadBinary (readbinary3d.c:263) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5119745: AOCreate (ao.c:577) > ==15577== by 0x5114195: AOCreateBasicIS (aobasic.c:382) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5119795: AOCreate (ao.c:577) > ==15577== by 0x5114195: AOCreateBasicIS (aobasic.c:382) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51197EC: AOCreate (ao.c:577) > ==15577== by 0x5114195: AOCreateBasicIS (aobasic.c:382) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5118EAF: AOSetIS (ao.c:498) > ==15577== by 0x5114201: AOCreateBasicIS (aobasic.c:383) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x51122DE: AOCreate_Basic (aobasic.c:205) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510FD16: PetscMPIIntCast (petscsys.h:1983) > ==15577== by 0x5112568: AOCreate_Basic (aobasic.c:213) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5112740: AOCreate_Basic (aobasic.c:219) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5112790: AOCreate_Basic (aobasic.c:219) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510FD16: PetscMPIIntCast (petscsys.h:1983) > ==15577== by 0x51128F5: AOCreate_Basic (aobasic.c:223) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51129A1: AOCreate_Basic (aobasic.c:230) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5112B3A: AOCreate_Basic (aobasic.c:242) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x5112B70: AOCreate_Basic (aobasic.c:242) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x5112B70: AOCreate_Basic (aobasic.c:242) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5112B70: AOCreate_Basic (aobasic.c:242) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5112B8D: AOCreate_Basic (aobasic.c:242) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x5112BC3: AOCreate_Basic (aobasic.c:242) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x5112BC3: AOCreate_Basic (aobasic.c:242) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5112BC3: AOCreate_Basic (aobasic.c:242) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x904F223: PMPI_Allgatherv (pallgatherv.c:93) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x90A1442: ompi_coll_tuned_allgatherv_intra_dec_fixed (coll_tuned_decision_fixed.c:643) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x904DFD9: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:54) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DF202: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:87) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DF223: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:98) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x91DF225: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:99) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DF07B: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2BC: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2EC: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A316: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A323: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DF0A4: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) > ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x904E491: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:62) > ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x910FCA8: mca_pml_ob1_isend (pml_ob1_isend.c:76) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x910FDAA: mca_pml_ob1_isend (pml_ob1_sendreq.h:360) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DB575: opal_convertor_pack (opal_convertor.c:232) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) > ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) > ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) > ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) > ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3A0: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) > ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3B4: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) > ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3BA: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) > ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) > ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) > ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2BC: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) > ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x911668C: mca_pml_ob1_send_request_copy_in_out (pml_ob1_sendreq.c:880) > ==15577== by 0x9112002: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_recvfrag.c:304) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x911204E: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:273) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116C0F: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:983) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116C40: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:998) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116C66: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1011) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116C81: mca_pml_ob1_send_request_schedule_once (opal_convertor.h:284) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116C8F: mca_pml_ob1_send_request_schedule_once (opal_convertor.h:294) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9079B0B: mca_btl_sm_prepare_src (btl_sm.c:696) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9079B45: mca_btl_sm_prepare_src (btl_sm.c:705) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3A0: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3B4: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3BA: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3E0: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A427: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116CED: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1041) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116D84: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1079) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9112290: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:273) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115689: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:284) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115696: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2EC: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A316: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A323: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9115854: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:168) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x911A65F: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:273) > ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116204: mca_pml_ob1_send_request_free (pml_ob1_sendreq.c:117) > ==15577== by 0x9043B1D: ompi_request_default_wait_all (request.h:350) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9049798: ompi_datatype_create_indexed (ompi_datatype_create_indexed.c:52) > ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DE0F7: opal_datatype_add (opal_datatype_add.c:153) > ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) > ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DE3A2: opal_datatype_add (opal_datatype_add.c:153) > ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) > ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DE124: opal_datatype_add (opal_datatype_add.c:161) > ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) > ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DE1A4: opal_datatype_add (opal_datatype_add.c:224) > ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) > ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DE1B6: opal_datatype_add (opal_datatype_add.c:237) > ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) > ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DE1E9: opal_datatype_add (opal_datatype_add.c:247) > ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) > ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DE43E: opal_datatype_add (opal_datatype_add.c:348) > ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) > ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DE758: opal_datatype_add (opal_datatype_add.c:353) > ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) > ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91E0BD1: opal_datatype_commit (opal_datatype_optimize.c:279) > ==15577== by 0x90A7F1F: ompi_coll_tuned_allgatherv_intra_neighborexchange (ompi_datatype.h:158) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91E0621: opal_datatype_optimize_short.constprop.0 (opal_datatype_optimize.c:226) > ==15577== by 0x91E0C6E: opal_datatype_commit (opal_datatype_optimize.c:308) > ==15577== by 0x90A7F1F: ompi_coll_tuned_allgatherv_intra_neighborexchange (ompi_datatype.h:158) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91E0672: opal_datatype_optimize_short.constprop.0 (opal_datatype_optimize.c:105) > ==15577== by 0x91E0C6E: opal_datatype_commit (opal_datatype_optimize.c:308) > ==15577== by 0x90A7F1F: ompi_coll_tuned_allgatherv_intra_neighborexchange (ompi_datatype.h:158) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9049798: ompi_datatype_create_indexed (ompi_datatype_create_indexed.c:52) > ==15577== by 0x90A7F84: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:470) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DBB85: opal_convertor_prepare_for_send (opal_convertor.c:559) > ==15577== by 0x91100C0: mca_pml_ob1_isend (opal_convertor.h:237) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x911A702: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:273) > ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91E0BD1: opal_datatype_commit (opal_datatype_optimize.c:279) > ==15577== by 0x90A7F99: ompi_coll_tuned_allgatherv_intra_neighborexchange (ompi_datatype.h:158) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91DB975: opal_convertor_prepare_for_recv (opal_convertor.c:523) > ==15577== by 0x9115BD7: mca_pml_ob1_recv_req_start (opal_convertor.h:258) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A0A51: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:45) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9114645: mca_pml_ob1_recv_request_progress_rndv (opal_convertor.h:284) > ==15577== by 0x9115EB4: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1026) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A0A51: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:45) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x904F223: PMPI_Allgatherv (pallgatherv.c:93) > ==15577== by 0x5112D70: AOCreate_Basic (aobasic.c:244) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5112EB4: AOCreate_Basic (aobasic.c:250) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x5112EEA: AOCreate_Basic (aobasic.c:250) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x5112EEA: AOCreate_Basic (aobasic.c:250) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5112EEA: AOCreate_Basic (aobasic.c:250) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510F92E: PetscMemcpy (petscsys.h:1656) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510F97F: PetscMemcpy (petscsys.h:1657) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510F9F8: PetscMemcpy (petscsys.h:1663) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510FA0F: PetscMemcpy (petscsys.h:1663) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) > ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3E0: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) > ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE188F: PetscSortInt (sorti.c:75) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE12EA: PetscSortInt_Private (sorti.c:33) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1362: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1385: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE13B1: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE13E1: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE15B0: PetscSortInt_Private (sorti.c:40) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE15D1: PetscSortInt_Private (sorti.c:40) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE1600: PetscSortInt_Private (sorti.c:43) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1636: PetscSortInt_Private (sorti.c:44) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE163D: PetscSortInt_Private (sorti.c:44) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE1647: PetscSortInt_Private (sorti.c:45) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16CE: PetscSortInt_Private (sorti.c:48) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16EF: PetscSortInt_Private (sorti.c:48) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE12EA: PetscSortInt_Private (sorti.c:33) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1362: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1385: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE13B1: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE13E1: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE15B0: PetscSortInt_Private (sorti.c:40) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE15D1: PetscSortInt_Private (sorti.c:40) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE1600: PetscSortInt_Private (sorti.c:43) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1636: PetscSortInt_Private (sorti.c:44) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE1647: PetscSortInt_Private (sorti.c:45) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16CE: PetscSortInt_Private (sorti.c:48) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16EF: PetscSortInt_Private (sorti.c:48) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE12F1: PetscSortInt_Private (sorti.c:34) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE12EA: PetscSortInt_Private (sorti.c:33) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1362: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1385: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE13B1: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE13E1: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1421: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1451: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1596: PetscSortInt_Private (sorti.c:40) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE15B0: PetscSortInt_Private (sorti.c:40) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE15B7: PetscSortInt_Private (sorti.c:40) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE15D1: PetscSortInt_Private (sorti.c:40) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE15D8: PetscSortInt_Private (sorti.c:41) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE1600: PetscSortInt_Private (sorti.c:43) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1615: PetscSortInt_Private (sorti.c:43) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1636: PetscSortInt_Private (sorti.c:44) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE1647: PetscSortInt_Private (sorti.c:45) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16B4: PetscSortInt_Private (sorti.c:48) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16CE: PetscSortInt_Private (sorti.c:48) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16D5: PetscSortInt_Private (sorti.c:48) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16EF: PetscSortInt_Private (sorti.c:48) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE12F1: PetscSortInt_Private (sorti.c:34) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE12FB: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1306: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1421: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1451: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1596: PetscSortInt_Private (sorti.c:40) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE15B7: PetscSortInt_Private (sorti.c:40) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE15D8: PetscSortInt_Private (sorti.c:41) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1615: PetscSortInt_Private (sorti.c:43) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16B4: PetscSortInt_Private (sorti.c:48) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16D5: PetscSortInt_Private (sorti.c:48) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE12FB: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1306: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE163D: PetscSortInt_Private (sorti.c:44) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE138B: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE161C: PetscSortInt_Private (sorti.c:43) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE138B: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE13E7: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE161C: PetscSortInt_Private (sorti.c:43) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE163D: PetscSortInt_Private (sorti.c:44) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE13E7: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE1457: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE130C: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE1457: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE130C: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51130D2: AOCreate_Basic (aobasic.c:254) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x511305D: AOCreate_Basic (aobasic.c:255) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510F92E: PetscMemcpy (petscsys.h:1656) > ==15577== by 0x51130FC: AOCreate_Basic (aobasic.c:258) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510F97F: PetscMemcpy (petscsys.h:1657) > ==15577== by 0x51130FC: AOCreate_Basic (aobasic.c:258) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510F9F8: PetscMemcpy (petscsys.h:1663) > ==15577== by 0x51130FC: AOCreate_Basic (aobasic.c:258) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510FA0F: PetscMemcpy (petscsys.h:1663) > ==15577== by 0x51130FC: AOCreate_Basic (aobasic.c:258) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4EE188F: PetscSortInt (sorti.c:75) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5113258: AOCreate_Basic (aobasic.c:260) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x51131E3: AOCreate_Basic (aobasic.c:261) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x511328F: AOCreate_Basic (aobasic.c:264) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x511328F: AOCreate_Basic (aobasic.c:264) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x511330D: AOCreate_Basic (aobasic.c:269) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x5113340: AOCreate_Basic (aobasic.c:269) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x5113340: AOCreate_Basic (aobasic.c:269) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5113340: AOCreate_Basic (aobasic.c:269) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x511335D: AOCreate_Basic (aobasic.c:269) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x5113394: AOCreate_Basic (aobasic.c:269) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x5113394: AOCreate_Basic (aobasic.c:269) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x5113394: AOCreate_Basic (aobasic.c:269) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5113415: AOCreate_Basic (aobasic.c:270) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510FB60: PetscMemzero (petscsys.h:1716) > ==15577== by 0x51134B4: AOCreate_Basic (aobasic.c:271) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) > ==15577== by 0x510FBBD: PetscMemzero (petscsys.h:1735) > ==15577== by 0x51134B4: AOCreate_Basic (aobasic.c:271) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) > ==15577== by 0x510FBBD: PetscMemzero (petscsys.h:1735) > ==15577== by 0x51134B4: AOCreate_Basic (aobasic.c:271) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x510FB60: PetscMemzero (petscsys.h:1716) > ==15577== by 0x5113528: AOCreate_Basic (aobasic.c:272) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5113721: AOCreate_Basic (aobasic.c:273) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x51135CF: AOCreate_Basic (aobasic.c:277) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x5113664: AOCreate_Basic (aobasic.c:278) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x511367A: AOCreate_Basic (aobasic.c:279) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x5113711: AOCreate_Basic (aobasic.c:280) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x511372E: AOCreate_Basic (aobasic.c:282) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x5113819: AOCreate_Basic (aobasic.c:285) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x5113819: AOCreate_Basic (aobasic.c:285) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x511385A: AOCreate_Basic (aobasic.c:285) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x511385A: AOCreate_Basic (aobasic.c:285) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x511391D: AOCreate_Basic (aobasic.c:287) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5113991: AOCreate_Basic (aobasic.c:293) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x43DA17: ReadBinary (readbinary3d.c:268) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D30A5: ISPartitioningCount (iscoloring.c:415) > ==15577== by 0x43DACB: ReadBinary (readbinary3d.c:272) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D3172: ISPartitioningCount (iscoloring.c:418) > ==15577== by 0x43DACB: ReadBinary (readbinary3d.c:272) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x50D331A: ISPartitioningCount (iscoloring.c:429) > ==15577== by 0x43DACB: ReadBinary (readbinary3d.c:272) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5110B49: AOPetscToApplication_Basic (aobasic.c:76) > ==15577== by 0x51160B0: AOPetscToApplicationIS (ao.c:134) > ==15577== by 0x43DC9A: ReadBinary (readbinary3d.c:280) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5110AC5: AOPetscToApplication_Basic (aobasic.c:77) > ==15577== by 0x51160B0: AOPetscToApplicationIS (ao.c:134) > ==15577== by 0x43DC9A: ReadBinary (readbinary3d.c:280) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5110AE1: AOPetscToApplication_Basic (aobasic.c:77) > ==15577== by 0x51160B0: AOPetscToApplicationIS (ao.c:134) > ==15577== by 0x43DC9A: ReadBinary (readbinary3d.c:280) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x5110B1A: AOPetscToApplication_Basic (aobasic.c:78) > ==15577== by 0x51160B0: AOPetscToApplicationIS (ao.c:134) > ==15577== by 0x43DC9A: ReadBinary (readbinary3d.c:280) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x511708D: AOApplicationToPetsc (ao.c:262) > ==15577== by 0x43DD1D: ReadBinary (readbinary3d.c:284) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5110E2A: AOApplicationToPetsc_Basic (aobasic.c:94) > ==15577== by 0x5117164: AOApplicationToPetsc (ao.c:263) > ==15577== by 0x43DD1D: ReadBinary (readbinary3d.c:284) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5110DA5: AOApplicationToPetsc_Basic (aobasic.c:95) > ==15577== by 0x5117164: AOApplicationToPetsc (ao.c:263) > ==15577== by 0x43DD1D: ReadBinary (readbinary3d.c:284) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x5110DC1: AOApplicationToPetsc_Basic (aobasic.c:95) > ==15577== by 0x5117164: AOApplicationToPetsc (ao.c:263) > ==15577== by 0x43DD1D: ReadBinary (readbinary3d.c:284) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x5110DFB: AOApplicationToPetsc_Basic (aobasic.c:96) > ==15577== by 0x5117164: AOApplicationToPetsc (ao.c:263) > ==15577== by 0x43DD1D: ReadBinary (readbinary3d.c:284) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x43CD6F: PetscBTCreate (petscbt.h:75) > ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x43CDB1: PetscBTCreate (petscbt.h:75) > ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x43CDB1: PetscBTCreate (petscbt.h:75) > ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x43CDB1: PetscBTCreate (petscbt.h:75) > ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x43CBCB: PetscMemzero (petscsys.h:1716) > ==15577== by 0x43CC89: PetscBTMemzero (petscbt.h:39) > ==15577== by 0x43CDD8: PetscBTCreate (petscbt.h:75) > ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) > ==15577== by 0x43CC28: PetscMemzero (petscsys.h:1735) > ==15577== by 0x43CC89: PetscBTMemzero (petscbt.h:39) > ==15577== by 0x43CDD8: PetscBTCreate (petscbt.h:75) > ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) > ==15577== by 0x43CC28: PetscMemzero (petscsys.h:1735) > ==15577== by 0x43CC89: PetscBTMemzero (petscbt.h:39) > ==15577== by 0x43CDD8: PetscBTCreate (petscbt.h:75) > ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B11D: memset (mc_replace_strmem.c:1007) > ==15577== by 0x43CC28: PetscMemzero (petscsys.h:1735) > ==15577== by 0x43CC89: PetscBTMemzero (petscbt.h:39) > ==15577== by 0x43CDD8: PetscBTCreate (petscbt.h:75) > ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2B132: memset (mc_replace_strmem.c:1007) > ==15577== by 0x43CC28: PetscMemzero (petscsys.h:1735) > ==15577== by 0x43CC89: PetscBTMemzero (petscbt.h:39) > ==15577== by 0x43CDD8: PetscBTCreate (petscbt.h:75) > ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x905D67C: PMPI_Recv (precv.c:53) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x905D7EC: PMPI_Recv (precv.c:54) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3D8: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3C3: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A3E0: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A427: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) > ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) > ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4494C1: DataPartitionVertices_Block (readbinary3d.c:1661) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x4494FA: DataPartitionVertices_Block (readbinary3d.c:1661) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x4494FA: DataPartitionVertices_Block (readbinary3d.c:1661) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x4494FA: DataPartitionVertices_Block (readbinary3d.c:1661) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x449A5C: DataPartitionVertices_Block (readbinary3d.c:1664) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44959C: DataPartitionVertices_Block (readbinary3d.c:1665) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x43CE1A: PetscBTLookupSet (petscbt.h:84) > ==15577== by 0x4495F7: DataPartitionVertices_Block (readbinary3d.c:1668) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x43CE5D: PetscBTLookupSet (petscbt.h:86) > ==15577== by 0x4495F7: DataPartitionVertices_Block (readbinary3d.c:1668) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4495FA: DataPartitionVertices_Block (readbinary3d.c:1668) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x905F748: PMPI_Send (psend.c:55) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x905F8DA: PMPI_Send (psend.c:64) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91103AA: mca_pml_ob1_send (pml_ob1_isend.c:108) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x91104AF: mca_pml_ob1_send (pml_ob1_sendreq.h:360) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2BC: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2C1: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A308: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A316: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2BC: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2C1: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A2EC: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A308: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A316: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A323: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A353: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) > ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) > ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) > ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) > ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) > ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) > ==15577== by 0x43CCC9: PetscBTDestroy (petscbt.h:44) > ==15577== by 0x449B6F: DataPartitionVertices_Block (readbinary3d.c:1696) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) > ==15577== by 0x43CCC9: PetscBTDestroy (petscbt.h:44) > ==15577== by 0x449B6F: DataPartitionVertices_Block (readbinary3d.c:1696) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x449BE1: DataPartitionVertices_Block (readbinary3d.c:1702) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) > ==15577== by 0x449C28: DataPartitionVertices_Block (readbinary3d.c:1702) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) > ==15577== by 0x449C28: DataPartitionVertices_Block (readbinary3d.c:1702) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) > ==15577== by 0x449C28: DataPartitionVertices_Block (readbinary3d.c:1702) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x43CD6F: PetscBTCreate (petscbt.h:75) > ==15577== by 0x449CA5: DataPartitionVertices_Block (readbinary3d.c:1703) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x44A129: DataPartitionVertices_Block (readbinary3d.c:1705) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x449D2D: DataPartitionVertices_Block (readbinary3d.c:1706) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x43CE1A: PetscBTLookupSet (petscbt.h:84) > ==15577== by 0x449D88: DataPartitionVertices_Block (readbinary3d.c:1708) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x43CE5D: PetscBTLookupSet (petscbt.h:86) > ==15577== by 0x449D88: DataPartitionVertices_Block (readbinary3d.c:1708) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x449DCB: DataPartitionVertices_Block (readbinary3d.c:1708) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x449D8B: DataPartitionVertices_Block (readbinary3d.c:1708) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116C0F: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:983) > ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) > ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116C40: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:998) > ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) > ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116C66: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1011) > ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) > ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116C81: mca_pml_ob1_send_request_schedule_once (opal_convertor.h:284) > ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) > ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116C8F: mca_pml_ob1_send_request_schedule_once (opal_convertor.h:294) > ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) > ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116CED: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1041) > ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) > ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x9116D84: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1079) > ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) > ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) > ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) > ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) > ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A333: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9113A26: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.c:632) > ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A0A51: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:45) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A368: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) > ==15577== by 0x9113A26: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.c:632) > ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) > ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) > ==15577== by 0x90A0A51: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:45) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x911002C: mca_pml_ob1_isend (pml_ob1_sendreq.h:372) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A435: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A43B: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A450: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A456: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A47E: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Conditional jump or move depends on uninitialised value(s) > ==15577== at 0x4C2A32D: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112D70: AOCreate_Basic (aobasic.c:244) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4C2A350: memcpy (mc_replace_strmem.c:878) > ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) > ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) > ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) > ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) > ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) > ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) > ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) > ==15577== by 0x5112D70: AOCreate_Basic (aobasic.c:244) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE14BF: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE14E2: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1522: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1552: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1689: PetscSortInt_Private (sorti.c:46) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16A6: PetscSortInt_Private (sorti.c:46) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1689: PetscSortInt_Private (sorti.c:46) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16A6: PetscSortInt_Private (sorti.c:46) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE14BF: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE14E2: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1522: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1552: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) > ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) > ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) > ==15577== by 0x405D55: main (fsi3d.c:55) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE14BF: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE14E2: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1522: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1552: PetscSortInt_Private (sorti.c:39) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE165C: PetscSortInt_Private (sorti.c:46) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1689: PetscSortInt_Private (sorti.c:46) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE168C: PetscSortInt_Private (sorti.c:46) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE16A6: PetscSortInt_Private (sorti.c:46) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1316: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1321: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1329: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1338: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE165C: PetscSortInt_Private (sorti.c:46) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE168C: PetscSortInt_Private (sorti.c:46) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1316: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1321: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1329: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== > ==15577== Use of uninitialised value of size 8 > ==15577== at 0x4EE1338: PetscSortInt_Private (sorti.c:35) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) > ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) > ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) > ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) > ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) > ==15577== -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From mailing at concert-tech.co.uk Sun Nov 3 21:15:08 2013 From: mailing at concert-tech.co.uk (Concert Tech Sound and Light) Date: Mon, 04 Nov 2013 03:15:08 +0000 Subject: [petsc-users] Concert Tech Sound and Light - a UK Based PA, Lighting, Audio Visual and Mobile Recording Services Hire Company Message-ID: <708b4498.euJ.eK6.w.5tcW7U@mailjet.com> Concert Tech Sound and Light A UK Based, PA, Lighting, Audio Visual and Mobile Recording Services Hire Company Concert Tech Sound and Light - A Brief Introduction Concert Tech Sound and Light is a UK Based, PA, Lighting, Audio Visual and Mobile Recording Services Hire Company based in Worthing, West Sussex. We are an enthusiastic, hard working company, eager to work with everyone from everywhere. With only a handful of employees but a large pool of engineers on our books we are able to maintain the 'Personal Touch' throughout your booking with us. We supply Sound Systems, Lighting Systems, LED Screens, Live Multitrack Recording Systems and Engineers to Artists, Acts, Bands and Tribute Shows throughout the UK. Productions and Events Over the past 7 years Concert Tech has been operating we have worked with a huge variety of artists, acts, bands and tributes to put on unique productions throughout the UK and Europe. With our expertise, great equipment, experienced staff and wealth of experience in the sector we are able to offer you Packages and Systems that will enable you to take your Events / Productions to the next level without breaking the bank! Large PA Hire with LED Screen & Video PA, Lighting and Video Proudction for your Event ? Flown Line Array Speaker System 2 x 2m LED Screen 4 x HD Cameras Big Lighting Check it out > ? Video of Your Event with LED Screen Your event filmed, our LED Screen Included ? LED Screen Included to Add Dynamic to your Event Full Multitrack Audio Recording 4 x HD 1080i Cameras Check it out > ? Great PA Hire Packages Unique Packages Specifically Tailored Super Low Cost Huge Variety & Tailored to Your Event Packages to Cover All Sizes of Event Indoor and Outdoor Events Covered Check it out > ? Any Occassion: Concert Tech has provided equipment for many different events, thoughout the year and throughout the country. ? Unique Productions: Concert Tech specalises in the unique. We like the challenge of taking an open space and transforming it totally! ? Recording: With our experience we have recorded everything from Quartets to Rock Bands to Orchestras and Choirs. Lighting: We believe that with careful planning we can add to every occasion and make it that something to remember. ? Live Sound: Throughout our work we have worked with some of the biggest acts out & we love to pass our skills on! ? Visual: We provide everything from Projectors to Full Video Walls for events across the spectrum. This email, document and all attachments is intended only for the person / recipient to whom it is addressed and/or otherwise authorised personnel. The information contained herein is confidential and is the sole property of Concert Tech. If for any reason you are not the intended recipient please be advised that viewing this email / document and any attachments, as well as, forwarding, printing, and disseminating any information related to this email / document / attachments is prohibited and that you are also prohibited from taking any action based on the content of this email / document and / or any attachment. Please note that the views expressed herein are solely those of the author and do not necessarily represent that of the company. Whilst all measures have been taken to ensure this email is clean and safe and that antivirus protection tools have been employed, you should check this email / document and any attachments for the presence of viruses. Concert Tech cannot be held liable for any damage caused by viruses contained within this email / document and no warranties or assurances are made in relation to the safety and content of this email / document and any attachments. Concert Tech also accepts no liability for any consequences arising from any problems caused due to this email / document and any attachments. Any reply to this document / email to this address is subject to strict monitoring for operational, security and lawful business practices. In sending this email the sender cannot be deemed to have specified authority and the contents of the email will have no contractual effect unless (in either case) it has otherwise been agreed between Concert Tech and the recipient. You email has be chosen by our team due to the fact we feel that we may be able to offer you a service that is of use to you and is not purely a random sales message. However should we have caused any offence in any way we sincerely apologise for that. If you wish to have your email removed from our list then all you need to do is click on the link below and hit send in your email browser. Remove My Email From This List -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: event_1.jpg Type: image/jpeg Size: 24105 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: event_ct_1.jpg Type: image/jpeg Size: 36278 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: event_ct_2.jpg Type: image/jpeg Size: 25099 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: event_ct_3.jpg Type: image/jpeg Size: 27203 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: event_ct_4.jpg Type: image/jpeg Size: 28033 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: event_ct_5.jpg Type: image/jpeg Size: 27757 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: event_ct_6.jpg Type: image/jpeg Size: 24499 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: header_ct.jpg Type: image/jpeg Size: 186136 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: Main Logo C.png Type: image/png Size: 35245 bytes Desc: not available URL: From bisheshkh at gmail.com Mon Nov 4 04:49:25 2013 From: bisheshkh at gmail.com (Bishesh Khanal) Date: Mon, 4 Nov 2013 11:49:25 +0100 Subject: [petsc-users] using petsc tools to solve isolated irregular domains with finite difference In-Reply-To: References: Message-ID: On Fri, Nov 1, 2013 at 1:30 PM, Matthew Knepley wrote: > On Fri, Nov 1, 2013 at 12:08 AM, Bishesh Khanal wrote: > >> On Thu, Oct 31, 2013 at 11:34 PM, Matthew Knepley wrote: >> >>> On Thu, Oct 31, 2013 at 5:02 PM, Bishesh Khanal wrote: >>> >>>> On Thu, Oct 31, 2013 at 7:43 PM, Matthew Knepley wrote: >>>> >>>>> On Thu, Oct 31, 2013 at 1:25 PM, Bishesh Khanal wrote: >>>>> >>>>>> I did not call DMCreateMatrix() before when using just dmda (and it >>>>>> is working). In any case, I added a call to DMCreateMatrix() now but still >>>>>> there are problems. Here are the code snippets and the error I get: >>>>>> >>>>> >>>>> Then you were allowing it to be called automatically by PETSc, and it >>>>> would have been this time as well. >>>>> >>>>> >>>>>> //Setting up the section and linking it to DM: >>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>> PetscInt point; >>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>> &point);CHKERRQ(ierr); >>>>>> ierr = PetscSectionSetDof(s, point, >>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>> } >>>>>> } >>>>>> } >>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>>>> >>>>>> //Set up KSP: >>>>>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>>>>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>>>>> ierr = >>>>>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>> ierr = >>>>>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>> ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); >>>>>> ierr = KSPSolve(ksp,NULL,NULL);CHKERRQ(ierr); >>>>>> >>>>>> ------------------------------------------------------ >>>>>> The computeMatrix2dSection function has: >>>>>> >>>>>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>>>>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>>>>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>>>>> for(PetscInt j = info.ys; j>>>>> for(PetscInt i = info.xs; i>>>>> if (isPosInDomain(ctx,i,j)) { >>>>>> ierr = >>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>>> ierr = PetscSectionGetDof(gs,point,&dof); >>>>>> for(PetscInt cDof = 0; cDof < dof; ++cDof) { >>>>>> num = 0; >>>>>> row+=cDof; >>>>>> col[num] = row; //(i,j) position >>>>>> v[num++] = -4; >>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>> ierr = >>>>>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>>>>> ierr = >>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>> col[num] += cDof; >>>>>> v[num++] = 1; >>>>>> } >>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>> ierr = >>>>>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>>>>> ierr = >>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>> col[num] += cDof; >>>>>> v[num++] = 1; >>>>>> } >>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>> ierr = >>>>>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>>>>> ierr = >>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>> col[num] += cDof; >>>>>> v[num++] = 1; >>>>>> } >>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>> ierr = >>>>>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>>>>> ierr = >>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>> col[num] += cDof; >>>>>> v[num++] = 1; >>>>>> } >>>>>> ierr = >>>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>> } >>>>>> } >>>>>> } >>>>>> } >>>>>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>> >>>>>> But this is not working. For e.g. for a 6X6 grid size I get the >>>>>> following error: >>>>>> >>>>> >>>>> Okay, here is how to debug this: >>>>> >>>>> 0) Go to a single scalar equations to make things easier >>>>> >>>>> 1) Use MatView(A, PETSC_VIEWER_STDOUT_WORLD), which should output a >>>>> 36 row matrix with >>>>> the preallocated nonzero pattern, filled with zeros >>>>> >>>>> 2) Make sure this is the pattern you want >>>>> >>>>> 3) Run in the debugger with -start_in_debugger >>>>> >>>>> 4) When you get the error, see >>>>> >>>>> a) If the (i, j) is one that should be setting a value >>>>> >>>>> b) Why this (i, j) was not preallocated >>>>> >>>> >>>> Up to 4 (a), it is correct. There is a problem in the way >>>> DMDAGetCellPoint and PetscSectionGetOffset works in this case scenario. I >>>> will try to explain it below for the case of 4X4 grid. >>>> First case: >>>> If I set the computational domain to be all the points of the dmda >>>> grid, (i.e. isPosInDomain(..,i,j,..) returns true for all i and j in the >>>> dmda grid), the program runs fine and does not give any error. >>>> >>>> Second case: >>>> I want the computational domain to be some part of the whole grid. >>>> There is a problem in this case. >>>> The following test is in a case where, >>>> isPosInDomain(..,i,j,..) returns true ONLY for (i,j) pairs of (2,1) >>>> (1,2) (2,2) (3,2) and (3,2). The grid with its corresponding point number >>>> in the petscsection is shown below: >>>> >>>> 12 13 14 15 >>>> 8 9 10 11 >>>> 4 5 6 7 >>>> 0 1 2 3 >>>> >>>> of which 6, 9, 10, 11 and 14 correspond to the (i,j) points that >>>> returns true for isPosInDomain(..,i,j,..) >>>> MatView gives me the 16-row matrix with the star stencil non-zero >>>> structure as expected. >>>> The error I get is: new non-zero at (0,2) caused a malloc. >>>> >>>> This error is for the value of (i,j) = (2,1), i.e. point 6. >>>> The loop to set values in the matrix starts as: >>>> for(PetscInt j = info.ys; j>>> for(PetscInt i = info.xs; i>>> if (isPosInDomain(ctx,i,j)) { >>>> ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>> >>>> Now, what I wanted from the beginning is to create the matrix >>>> containing the rows corresponding to only those points (i,j) which has true >>>> values for isPosInDomain(ctx,i,j), that means in this case 5 row matrix. In >>>> the current test, the first (i,j) that reaches DMDAGetCellPoint is (2,1), >>>> which gives sets point variable to 6. >>>> The line PetscSectionGetOffset(gs,point,&row) sets the value of row to >>>> 0. >>>> So I think this where the inconsistency lies. >>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES) expects row variable and >>>> col[] array to correspond to (i,j) based on DMDA, but PetsSectionGetOffset >>>> gives result based on how we masked away some of the points from dmda grid. >>>> Or am I missing something very basic here ? >>>> >>> >>> You are missing something basic. The row is correctly identified as 0, >>> and that means the 2 is >>> point 10. >>> >> The problem must be preallocation. First, you should send the result of >>> MatView(). >>> >> >> MatView shows that the preallocation and the matrix created is for the >> full dmda grid which is probably not what we want here, right ? Here is the >> matview result: >> > > Thats right. Did you call > > DMSetDefaultSection(da, s) > > after you made the new section and before you called DMGetMatrix()? > Do you mean before I called DMCreateMatrix() ? I called DMSetDefaultSection(da,s) as below: ... ierr = PetscSectionSetUp(s);CHKERRQ(ierr); ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); // ierr = DMGetMatrix(dm,&A);CHKERRQ(ierr); //It seems DMGetMatrix is changed to DMCreateMatrix(). ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); This MatView on A shows (result below) that it still has 16 rows for all DMDA points instead of only those petscsection points for which I set a non-zero dof. Is it what it should not be doing ? > Mat Object: 1 MPI processes >> >> type: seqaij >> >> row 0: (0, 0) (1, 0) (4, 0) >> >> row 1: (0, 0) (1, 0) (2, 0) (5, 0) >> >> row 2: (1, 0) (2, 0) (3, 0) (6, 0) >> >> row 3: (2, 0) (3, 0) (7, 0) >> >> row 4: (0, 0) (4, 0) (5, 0) (8, 0) >> >> row 5: (1, 0) (4, 0) (5, 0) (6, 0) (9, 0) >> >> row 6: (2, 0) (5, 0) (6, 0) (7, 0) (10, 0) >> >> row 7: (3, 0) (6, 0) (7, 0) (11, 0) >> >> row 8: (4, 0) (8, 0) (9, 0) (12, 0) >> >> row 9: (5, 0) (8, 0) (9, 0) (10, 0) (13, 0) >> >> row 10: (6, 0) (9, 0) (10, 0) (11, 0) (14, 0) >> >> row 11: (7, 0) (10, 0) (11, 0) (15, 0) >> >> row 12: (8, 0) (12, 0) (13, 0) >> >> row 13: (9, 0) (12, 0) (13, 0) (14, 0) >> >> row 14: (10, 0) (13, 0) (14, 0) (15, 0) >> >> row 15: (11, 0) (14, 0) (15, 0) >> >> [0]PETSC ERROR: --------------------- Error Message >> ------------------------------------ >> >> [0]PETSC ERROR: Argument out of range! >> >> [0]PETSC ERROR: New nonzero at (0,2) caused a malloc! >> >> >> For reference here is the code snippet again: ... ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { PetscInt point; if(isPosInDomain(&testPoisson,i,j,0)) { ierr = DMDAGetCellPoint(dm, i, j, 0, &point);CHKERRQ(ierr); ierr = PetscSectionSetDof(s, point, testPoisson.mDof); // No. of dofs associated with the point. } } } ierr = PetscSectionSetUp(s);CHKERRQ(ierr); ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); ierr = KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); ierr = KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); .... The function computeMatrix2dSection: #undef __FUNCT__ #define __FUNCT__ "computeMatrix2dSection" PetscErrorCode computeMatrix2dSection(KSP ksp, Mat A, Mat B, MatStructure *str, void *context) { PoissonModel *ctx = (PoissonModel*)context; PetscErrorCode ierr; DM da; DMDALocalInfo info; PetscInt row, col[5]; PetscInt dof; PetscScalar v[5]; //array to store 5 point stencil for each row PetscInt num; //non-zero position in the current row // PetscScalar dScale = 1; //to scale Dirichlet identity rows if needed PetscSection gs; //Global Section that keeps the grid info and indices PetscInt point; //Current point of the petscSection PetscFunctionBeginUser; ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); ierr = DMCreateMatrix(da,&A);CHKERRQ(ierr); // ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); for(PetscInt j = info.ys; jmDof; ++cDof) { num = 0; row+=cDof; col[num] = row; //(i,j) position v[num++] = -4; if(isPosInDomain(ctx,i+1,j)) { ierr = DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); ierr = PetscSectionGetOffset(gs,point,&col[num]); col[num] += cDof; v[num++] = 1; } if(isPosInDomain(ctx,i-1,j)) { ierr = DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); ierr = PetscSectionGetOffset(gs,point,&col[num]); col[num] += cDof; v[num++] = 1; } if(isPosInDomain(ctx,i,j+1)) { ierr = DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); ierr = PetscSectionGetOffset(gs,point,&col[num]); col[num] += cDof; v[num++] = 1; } if(isPosInDomain(ctx,i,j-1)) { ierr = DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); ierr = PetscSectionGetOffset(gs,point,&col[num]); col[num] += cDof; v[num++] = 1; } ierr = MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); } } } } ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); > >>> Matt >>> >>> >>>> >>>>> Thanks, >>>>> >>>>> Matt >>>>> >>>>> >>>>>> [0]PETSC ERROR: --------------------- Error Message >>>>>> ------------------------------------ >>>>>> [0]PETSC ERROR: Argument out of range! >>>>>> [0]PETSC ERROR: New nonzero at (0,6) caused a malloc! >>>>>> [0]PETSC ERROR: >>>>>> ------------------------------------------------------------------------ >>>>>> [0]PETSC ERROR: Petsc Development GIT revision: >>>>>> 61e5e40bb2c5bf2423e94b71a15fef47e411b0da GIT Date: 2013-10-25 21:47:45 >>>>>> -0500 >>>>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>>>> [0]PETSC ERROR: >>>>>> ------------------------------------------------------------------------ >>>>>> [0]PETSC ERROR: build/poissonIrregular on a arch-linux2-cxx-debug >>>>>> named edwards by bkhanal Thu Oct 31 19:23:58 2013 >>>>>> [0]PETSC ERROR: Libraries linked from >>>>>> /home/bkhanal/Documents/softwares/petsc/arch-linux2-cxx-debug/lib >>>>>> [0]PETSC ERROR: Configure run at Sat Oct 26 16:35:15 2013 >>>>>> [0]PETSC ERROR: Configure options --download-mpich >>>>>> -download-f-blas-lapack=1 --download-metis --download-parmetis >>>>>> --download-superlu_dist --download-scalapack --download-mumps >>>>>> --download-hypre --with-clanguage=cxx >>>>>> [0]PETSC ERROR: >>>>>> ------------------------------------------------------------------------ >>>>>> [0]PETSC ERROR: MatSetValues_SeqAIJ() line 413 in >>>>>> src/mat/impls/aij/seq/aij.c >>>>>> [0]PETSC ERROR: MatSetValues() line 1130 in src/mat/interface/matrix.c >>>>>> [0]PETSC ERROR: computeMatrix2dSection() line 318 in >>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>> [0]PETSC ERROR: KSPSetUp() line 228 in src/ksp/ksp/interface/itfunc.c >>>>>> [0]PETSC ERROR: KSPSolve() line 399 in src/ksp/ksp/interface/itfunc.c >>>>>> [0]PETSC ERROR: main() line 598 in >>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>> [unset]: aborting job: >>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>> >>>>>> >>>>>>> Matt >>>>>>> >>>>>>> >>>>>>>> However, now since I do not want the rows in the matrix >>>>>>>> corresponding to the points in the grid which do not lie in the >>>>>>>> computational domain. This means the size of the matrix will not >>>>>>>> necessarily equal the total number of cells in DMDA. So in the following >>>>>>>> code: >>>>>>>> >>>>>>>> ierr = DMDAGetLocalInfo(dm,&info);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>>> PetscInt point; >>>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>>> &point);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>>> } >>>>>>>> >>>>>>>> } >>>>>>>> } >>>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>>> >>>>>>>> should I myself compute proper nC (i.e. total no. of points for >>>>>>>> which I want the matrix to have a corresponding row) before calling >>>>>>>> PetscSectionSetChart() or is the masking out of the points inside the loop >>>>>>>> with if(isPosInDomain(&testPoisson,i,j,0)) sufficient ? >>>>>>>> And, when you write: >>>>>>>> PetscSectionGetOffset(gs, p, &ind); >>>>>>>> row = ind < 0 ? -(ind+1) : ind; >>>>>>>> >>>>>>>> it seems you allow the possibility of getting a -ve ind when using >>>>>>>> PetscSectionGetOffset. When would an offset to a particular dof and point? >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>>> num = 0; >>>>>>>>>> /*ierr = >>>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr);*/ /*Get the PetscPoint*/ >>>>>>>>>> /*But I did now understand how I would emulate >>>>>>>>>> the row.c and col.c info with PetscSection*/ >>>>>>>>>> row.i = i; row.j = j; >>>>>>>>>> >>>>>>>>> >>>>>>>>> Here you would call >>>>>>>>> >>>>>>>>> ierr = DMDAGetCellPoint(da, i, j, 0, &cell);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionGetOffset(da, cell, &row);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionGetDof(da, cell, &dof);CHKERRQ(ierr); >>>>>>>>> >>>>>>>>> This means that this cell has global rows = [row, row+dof). >>>>>>>>> >>>>>>>>> >>>>>>>>>> col[num].i = i; col[num].j = j; >>>>>>>>>> if (isPosInDomain(ctx,i,j)) { /*put the >>>>>>>>>> operator for only the values for only the points lying in the domain*/ >>>>>>>>>> v[num++] = -4; >>>>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>>>> col[num].i = i+1; col[num].j = j; >>>>>>>>>> v[num++] = 1; >>>>>>>>>> } >>>>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>>>> col[num].i = i-1; col[num].j = j; >>>>>>>>>> v[num++] = 1; >>>>>>>>>> } >>>>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>>>> col[num].i = i; col[num].j = j+1; >>>>>>>>>> v[num++] = 1; >>>>>>>>>> } >>>>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>>>> col[num].i = i; col[num].j = j-1; >>>>>>>>>> v[num++] = 1; >>>>>>>>>> } >>>>>>>>>> } else { >>>>>>>>>> v[num++] = dScale; /*set Dirichlet identity >>>>>>>>>> rows for points in the rectangle but outside the computational domain*/ >>>>>>>>>> } >>>>>>>>>> >>>>>>>>> >>>>>>>>> You do the same thing for the columns, and then call >>>>>>>>> >>>>>>>>> ierr = MatSetValues(A, dof, rows, num*dof, cols, v, >>>>>>>>> INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>> >>>>>>>>> >>>>>>>>>> ierr = >>>>>>>>>> MatSetValuesStencil(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>> } >>>>>>>>>> } >>>>>>>>>> } >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> Thanks, >>>>>>>>>>> >>>>>>>>>>> Matt >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> Matt >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> If not, just put the identity into >>>>>>>>>>>>>>>>>>>>> the rows you do not use on the full cube. It will not >>>>>>>>>>>>>>>>>>>>> hurt scalability or convergence. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> In the case of Poisson with Dirichlet condition this >>>>>>>>>>>>>>>>>>>> might be the case. But is it always true that having identity rows in the >>>>>>>>>>>>>>>>>>>> system matrix will not hurt convergence ? I thought otherwise for the >>>>>>>>>>>>>>>>>>>> following reasons: >>>>>>>>>>>>>>>>>>>> 1) Having read Jed's answer here : >>>>>>>>>>>>>>>>>>>> http://scicomp.stackexchange.com/questions/3426/why-is-pinning-a-point-to-remove-a-null-space-bad/3427#3427 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Jed is talking about a constraint on a the pressure at a >>>>>>>>>>>>>>>>>>> point. This is just decoupling these unknowns from the rest >>>>>>>>>>>>>>>>>>> of the problem. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 2) Some observation I am getting (but I am still doing >>>>>>>>>>>>>>>>>>>> more experiments to confirm) while solving my staggered-grid 3D stokes flow >>>>>>>>>>>>>>>>>>>> with schur complement and using -pc_type gamg for A00 matrix. Putting the >>>>>>>>>>>>>>>>>>>> identity rows for dirichlet boundaries and for ghost cells seemed to have >>>>>>>>>>>>>>>>>>>> effects on its convergence. I'm hoping once I know how to use PetscSection, >>>>>>>>>>>>>>>>>>>> I can get rid of using ghost cells method for the staggered grid and get >>>>>>>>>>>>>>>>>>>> rid of the identity rows too. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> It can change the exact iteration, but it does not make >>>>>>>>>>>>>>>>>>> the matrix conditioning worse. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Anyway please provide me with some pointers so that I >>>>>>>>>>>>>>>>>>>> can start trying with petscsection on top of a dmda, in the beginning for >>>>>>>>>>>>>>>>>>>> non-staggered case. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>>>> experiments lead. >>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>> experiments lead. >>>>>>>>> -- Norbert Wiener >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> What most experimenters take for granted before they begin their >>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>> experiments lead. >>>>>>> -- Norbert Wiener >>>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> What most experimenters take for granted before they begin their >>>>> experiments is infinitely more interesting than any results to which their >>>>> experiments lead. >>>>> -- Norbert Wiener >>>>> >>>> >>>> >>> >>> >>> -- >>> What most experimenters take for granted before they begin their >>> experiments is infinitely more interesting than any results to which their >>> experiments lead. >>> -- Norbert Wiener >>> >> >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From asmund.ervik at ntnu.no Mon Nov 4 07:31:05 2013 From: asmund.ervik at ntnu.no (=?UTF-8?B?w4VzbXVuZCBFcnZpaw==?=) Date: Mon, 04 Nov 2013 14:31:05 +0100 Subject: [petsc-users] Getting DMDA local coordinates In-Reply-To: References: <52651296.6030102@ntnu.no> <87sivuzvlw.fsf@mcs.anl.gov> <52651ABC.1060306@ntnu.no> <87ppqyztgc.fsf@mcs.anl.gov> <52652692.7000104@ntnu.no> <87hacazrzv.fsf@mcs.anl.gov> <52652B49.2020709@ntnu.no> Message-ID: <5277A199.1060900@ntnu.no> Hi again, On 21. okt. 2013 15:34, Jed Brown wrote: > Great. Looking forward to your example. I just submitted a pull request with my example. I added you as a reviewer. Any and all comments are appreciated. Regards, ?smund From jedbrown at mcs.anl.gov Mon Nov 4 08:12:35 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Mon, 04 Nov 2013 07:12:35 -0700 Subject: [petsc-users] Getting DMDA local coordinates In-Reply-To: <5277A199.1060900@ntnu.no> References: <52651296.6030102@ntnu.no> <87sivuzvlw.fsf@mcs.anl.gov> <52651ABC.1060306@ntnu.no> <87ppqyztgc.fsf@mcs.anl.gov> <52652692.7000104@ntnu.no> <87hacazrzv.fsf@mcs.anl.gov> <52652B49.2020709@ntnu.no> <5277A199.1060900@ntnu.no> Message-ID: <87ob602r3g.fsf@mcs.anl.gov> ?smund Ervik writes: > Hi again, > > On 21. okt. 2013 15:34, Jed Brown wrote: >> Great. Looking forward to your example. > > I just submitted a pull request with my example. I added you as a > reviewer. Any and all comments are appreciated. Better to link it here in the email so that others on the list have an opportunity to see what you're talking about. https://bitbucket.org/petsc/petsc/pull-request/118/added-dm-examples-tutorials-ex13f90 -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From rupp at mcs.anl.gov Mon Nov 4 08:51:20 2013 From: rupp at mcs.anl.gov (Karl Rupp) Date: Mon, 04 Nov 2013 15:51:20 +0100 Subject: [petsc-users] [petsc-maint] Speedup problem when using OpenMP? In-Reply-To: <5272EDAD.2070001@gmail.com> References: <5272EDAD.2070001@gmail.com> Message-ID: <5277B468.4090509@mcs.anl.gov> Hi, > I have a question on the speedup of PETSc when using OpenMP. I can get > good speedup when using MPI, but no speedup when using OpenMP. > The example is ex2f with m=100 and n=100. The number of available > processors is 16 (32 threads) and the OS is Windows Server 2012. The log > files for 4 and 8 processors are attached. > > The commands I used to run with 4 processors are as follows: > Run using MPI > mpiexec -n 4 Petsc-windows-ex2f.exe -m 100 -n 100 -log_summary > log_100x100_mpi_p4.log > > Run using OpenMP > Petsc-windows-ex2f.exe -threadcomm_type openmp -threadcomm_nthreads 4 -m > 100 -n 100 -log_summary log_100x100_openmp_p4.log > > The PETSc used for this test is PETSc for Windows > http://www.mic-tc.ch/downloads/PETScForWindows.zip, but I guess this is > not the problem because the same problem exists when I use PETSc-dev in > Cygwin. I don't know if this problem exists in Linux, would anybody help > to test? For the 100x100 case considered, the execution times per call are somewhere in the millisecond to sub-millisecond range (e.g. 1.3ms for 68 calls to VecScale with 4 processors). I'd say this is too small in order to see any reasonable performance gain when running multiple threads, consider problem sizes of about 1000x1000 instead. Moreover, keep in mind that typically you won't get a perfectly linear scaling with the number of processor cores, because ultimately the memory bandwidth is the limiting factor for standard vector operations. Best regards, Karli From lu_qin_2000 at yahoo.com Mon Nov 4 09:01:42 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Mon, 4 Nov 2013 07:01:42 -0800 (PST) Subject: [petsc-users] MatSetValues vs. MatSetValue Message-ID: <1383577302.81115.YahooMailNeo@web160204.mail.bf1.yahoo.com> Hello, ? I need to set values of a sparse matrix. Is setting coefficients row by row using MatSetValues more efficient than setting setting coefficient one by one using MatSetValue? Is MatSetValues implemented by looping over each entry of the input array calling MatSetValue? ? Thanks a lot for your help, Qin -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Mon Nov 4 09:08:43 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Mon, 04 Nov 2013 08:08:43 -0700 Subject: [petsc-users] MatSetValues vs. MatSetValue In-Reply-To: <1383577302.81115.YahooMailNeo@web160204.mail.bf1.yahoo.com> References: <1383577302.81115.YahooMailNeo@web160204.mail.bf1.yahoo.com> Message-ID: <87fvrc2ohw.fsf@mcs.anl.gov> Qin Lu writes: > Hello, > ? > I need to set values of a sparse matrix. Is setting coefficients row > by row using MatSetValues more efficient than setting setting > coefficient one by one using MatSetValue? Somewhat more efficient, yes. > Is MatSetValues implemented by looping over each entry of the input > array calling MatSetValue? Calling for a group of entries (once per row or once per "element" in finite element methods) amortizes some of the search cost. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From knepley at gmail.com Mon Nov 4 09:54:23 2013 From: knepley at gmail.com (Matthew Knepley) Date: Mon, 4 Nov 2013 09:54:23 -0600 Subject: [petsc-users] using petsc tools to solve isolated irregular domains with finite difference In-Reply-To: References: Message-ID: On Mon, Nov 4, 2013 at 4:49 AM, Bishesh Khanal wrote: > On Fri, Nov 1, 2013 at 1:30 PM, Matthew Knepley wrote: > >> On Fri, Nov 1, 2013 at 12:08 AM, Bishesh Khanal wrote: >> >>> On Thu, Oct 31, 2013 at 11:34 PM, Matthew Knepley wrote: >>> >>>> On Thu, Oct 31, 2013 at 5:02 PM, Bishesh Khanal wrote: >>>> >>>>> On Thu, Oct 31, 2013 at 7:43 PM, Matthew Knepley wrote: >>>>> >>>>>> On Thu, Oct 31, 2013 at 1:25 PM, Bishesh Khanal wrote: >>>>>> >>>>>>> I did not call DMCreateMatrix() before when using just dmda (and it >>>>>>> is working). In any case, I added a call to DMCreateMatrix() now but still >>>>>>> there are problems. Here are the code snippets and the error I get: >>>>>>> >>>>>> >>>>>> Then you were allowing it to be called automatically by PETSc, and it >>>>>> would have been this time as well. >>>>>> >>>>>> >>>>>>> //Setting up the section and linking it to DM: >>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>> PetscInt point; >>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>> &point);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>> } >>>>>>> } >>>>>>> } >>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>>>>> >>>>>>> //Set up KSP: >>>>>>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>>>>>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>>>>>> ierr = >>>>>>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>>> ierr = >>>>>>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>>> ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); >>>>>>> ierr = KSPSolve(ksp,NULL,NULL);CHKERRQ(ierr); >>>>>>> >>>>>>> ------------------------------------------------------ >>>>>>> The computeMatrix2dSection function has: >>>>>>> >>>>>>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>>>>>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>>>>>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>>>>>> for(PetscInt j = info.ys; j>>>>>> for(PetscInt i = info.xs; i>>>>>> if (isPosInDomain(ctx,i,j)) { >>>>>>> ierr = >>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>>>> ierr = PetscSectionGetDof(gs,point,&dof); >>>>>>> for(PetscInt cDof = 0; cDof < dof; ++cDof) { >>>>>>> num = 0; >>>>>>> row+=cDof; >>>>>>> col[num] = row; //(i,j) position >>>>>>> v[num++] = -4; >>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>> ierr = >>>>>>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>>>>>> ierr = >>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>> col[num] += cDof; >>>>>>> v[num++] = 1; >>>>>>> } >>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>> ierr = >>>>>>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>>>>>> ierr = >>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>> col[num] += cDof; >>>>>>> v[num++] = 1; >>>>>>> } >>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>> ierr = >>>>>>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>>>>>> ierr = >>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>> col[num] += cDof; >>>>>>> v[num++] = 1; >>>>>>> } >>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>> ierr = >>>>>>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>>>>>> ierr = >>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>> col[num] += cDof; >>>>>>> v[num++] = 1; >>>>>>> } >>>>>>> ierr = >>>>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>> } >>>>>>> } >>>>>>> } >>>>>>> } >>>>>>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>>> >>>>>>> But this is not working. For e.g. for a 6X6 grid size I get the >>>>>>> following error: >>>>>>> >>>>>> >>>>>> Okay, here is how to debug this: >>>>>> >>>>>> 0) Go to a single scalar equations to make things easier >>>>>> >>>>>> 1) Use MatView(A, PETSC_VIEWER_STDOUT_WORLD), which should output a >>>>>> 36 row matrix with >>>>>> the preallocated nonzero pattern, filled with zeros >>>>>> >>>>>> 2) Make sure this is the pattern you want >>>>>> >>>>>> 3) Run in the debugger with -start_in_debugger >>>>>> >>>>>> 4) When you get the error, see >>>>>> >>>>>> a) If the (i, j) is one that should be setting a value >>>>>> >>>>>> b) Why this (i, j) was not preallocated >>>>>> >>>>> >>>>> Up to 4 (a), it is correct. There is a problem in the way >>>>> DMDAGetCellPoint and PetscSectionGetOffset works in this case scenario. I >>>>> will try to explain it below for the case of 4X4 grid. >>>>> First case: >>>>> If I set the computational domain to be all the points of the dmda >>>>> grid, (i.e. isPosInDomain(..,i,j,..) returns true for all i and j in the >>>>> dmda grid), the program runs fine and does not give any error. >>>>> >>>>> Second case: >>>>> I want the computational domain to be some part of the whole grid. >>>>> There is a problem in this case. >>>>> The following test is in a case where, >>>>> isPosInDomain(..,i,j,..) returns true ONLY for (i,j) pairs of (2,1) >>>>> (1,2) (2,2) (3,2) and (3,2). The grid with its corresponding point number >>>>> in the petscsection is shown below: >>>>> >>>>> 12 13 14 15 >>>>> 8 9 10 11 >>>>> 4 5 6 7 >>>>> 0 1 2 3 >>>>> >>>>> of which 6, 9, 10, 11 and 14 correspond to the (i,j) points that >>>>> returns true for isPosInDomain(..,i,j,..) >>>>> MatView gives me the 16-row matrix with the star stencil non-zero >>>>> structure as expected. >>>>> The error I get is: new non-zero at (0,2) caused a malloc. >>>>> >>>>> This error is for the value of (i,j) = (2,1), i.e. point 6. >>>>> The loop to set values in the matrix starts as: >>>>> for(PetscInt j = info.ys; j>>>> for(PetscInt i = info.xs; i>>>> if (isPosInDomain(ctx,i,j)) { >>>>> ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>> >>>>> Now, what I wanted from the beginning is to create the matrix >>>>> containing the rows corresponding to only those points (i,j) which has true >>>>> values for isPosInDomain(ctx,i,j), that means in this case 5 row matrix. In >>>>> the current test, the first (i,j) that reaches DMDAGetCellPoint is (2,1), >>>>> which gives sets point variable to 6. >>>>> The line PetscSectionGetOffset(gs,point,&row) sets the value of row to >>>>> 0. >>>>> So I think this where the inconsistency lies. >>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES) expects row variable >>>>> and col[] array to correspond to (i,j) based on DMDA, but >>>>> PetsSectionGetOffset gives result based on how we masked away some of the >>>>> points from dmda grid. Or am I missing something very basic here ? >>>>> >>>> >>>> You are missing something basic. The row is correctly identified as 0, >>>> and that means the 2 is >>>> point 10. >>>> >>> The problem must be preallocation. First, you should send the result of >>>> MatView(). >>>> >>> >>> MatView shows that the preallocation and the matrix created is for the >>> full dmda grid which is probably not what we want here, right ? Here is the >>> matview result: >>> >> >> Thats right. Did you call >> >> DMSetDefaultSection(da, s) >> >> after you made the new section and before you called DMGetMatrix()? >> > Do you mean before I called DMCreateMatrix() ? I called > DMSetDefaultSection(da,s) as below: > > ... > ierr = PetscSectionSetUp(s);CHKERRQ(ierr); > ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); > ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); > // ierr = DMGetMatrix(dm,&A);CHKERRQ(ierr); //It seems DMGetMatrix is > changed to DMCreateMatrix(). > ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); > ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); > Yes, this is correct. I do exactly this in the example code and it works fine. I suspect you are not creating the section you think you are. You can send the code if you want and I will look at it. Matt > This MatView on A shows (result below) that it still has 16 rows for all > DMDA points instead of only those petscsection points for which I set a > non-zero dof. Is it what it should not be doing ? > >> Mat Object: 1 MPI processes >>> >>> type: seqaij >>> >>> row 0: (0, 0) (1, 0) (4, 0) >>> >>> row 1: (0, 0) (1, 0) (2, 0) (5, 0) >>> >>> row 2: (1, 0) (2, 0) (3, 0) (6, 0) >>> >>> row 3: (2, 0) (3, 0) (7, 0) >>> >>> row 4: (0, 0) (4, 0) (5, 0) (8, 0) >>> >>> row 5: (1, 0) (4, 0) (5, 0) (6, 0) (9, 0) >>> >>> row 6: (2, 0) (5, 0) (6, 0) (7, 0) (10, 0) >>> >>> row 7: (3, 0) (6, 0) (7, 0) (11, 0) >>> >>> row 8: (4, 0) (8, 0) (9, 0) (12, 0) >>> >>> row 9: (5, 0) (8, 0) (9, 0) (10, 0) (13, 0) >>> >>> row 10: (6, 0) (9, 0) (10, 0) (11, 0) (14, 0) >>> >>> row 11: (7, 0) (10, 0) (11, 0) (15, 0) >>> >>> row 12: (8, 0) (12, 0) (13, 0) >>> >>> row 13: (9, 0) (12, 0) (13, 0) (14, 0) >>> >>> row 14: (10, 0) (13, 0) (14, 0) (15, 0) >>> >>> row 15: (11, 0) (14, 0) (15, 0) >>> >>> [0]PETSC ERROR: --------------------- Error Message >>> ------------------------------------ >>> >>> [0]PETSC ERROR: Argument out of range! >>> >>> [0]PETSC ERROR: New nonzero at (0,2) caused a malloc! >>> >>> >>> > For reference here is the code snippet again: > ... > ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); > ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); > ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); > for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { > for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { > PetscInt point; > if(isPosInDomain(&testPoisson,i,j,0)) { > ierr = DMDAGetCellPoint(dm, i, j, 0, &point);CHKERRQ(ierr); > ierr = PetscSectionSetDof(s, point, testPoisson.mDof); // > No. of dofs associated with the point. > } > } > } > ierr = PetscSectionSetUp(s);CHKERRQ(ierr); > ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); > ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); > ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); > ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); > > ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); > ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); > > ierr = > KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); > ierr = > KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); > > .... > The function computeMatrix2dSection: > > #undef __FUNCT__ > #define __FUNCT__ "computeMatrix2dSection" > PetscErrorCode computeMatrix2dSection(KSP ksp, Mat A, Mat B, MatStructure > *str, void *context) { > PoissonModel *ctx = (PoissonModel*)context; > PetscErrorCode ierr; > DM da; > DMDALocalInfo info; > PetscInt row, col[5]; > PetscInt dof; > PetscScalar v[5]; //array to store 5 point stencil for > each row > PetscInt num; //non-zero position in the current row > // PetscScalar dScale = 1; //to scale Dirichlet > identity rows if needed > PetscSection gs; //Global Section that keeps > the grid info and indices > PetscInt point; //Current point of the > petscSection > > PetscFunctionBeginUser; > ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); > ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); > ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); > ierr = DMCreateMatrix(da,&A);CHKERRQ(ierr); > // ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); > for(PetscInt j = info.ys; j for(PetscInt i = info.xs; i if (isPosInDomain(ctx,i,j)) { > ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); > ierr = PetscSectionGetOffset(gs,point,&row); > ierr = PetscSectionGetDof(gs,point,&dof); > for(PetscInt cDof = 0; cDof < ctx->mDof; ++cDof) { > num = 0; > row+=cDof; > col[num] = row; //(i,j) position > v[num++] = -4; > if(isPosInDomain(ctx,i+1,j)) { > ierr = > DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); > ierr = PetscSectionGetOffset(gs,point,&col[num]); > col[num] += cDof; > v[num++] = 1; > } > if(isPosInDomain(ctx,i-1,j)) { > ierr = > DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); > ierr = PetscSectionGetOffset(gs,point,&col[num]); > col[num] += cDof; > v[num++] = 1; > } > if(isPosInDomain(ctx,i,j+1)) { > ierr = > DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); > ierr = PetscSectionGetOffset(gs,point,&col[num]); > col[num] += cDof; > v[num++] = 1; > } > if(isPosInDomain(ctx,i,j-1)) { > ierr = > DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); > ierr = PetscSectionGetOffset(gs,point,&col[num]); > col[num] += cDof; > v[num++] = 1; > } > ierr = > MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); > } > } > } > } > ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); > ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); > > > > >> >>>> Matt >>>> >>>> >>>>> >>>>>> Thanks, >>>>>> >>>>>> Matt >>>>>> >>>>>> >>>>>>> [0]PETSC ERROR: --------------------- Error Message >>>>>>> ------------------------------------ >>>>>>> [0]PETSC ERROR: Argument out of range! >>>>>>> [0]PETSC ERROR: New nonzero at (0,6) caused a malloc! >>>>>>> [0]PETSC ERROR: >>>>>>> ------------------------------------------------------------------------ >>>>>>> [0]PETSC ERROR: Petsc Development GIT revision: >>>>>>> 61e5e40bb2c5bf2423e94b71a15fef47e411b0da GIT Date: 2013-10-25 21:47:45 >>>>>>> -0500 >>>>>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>>>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>>>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>>>>> [0]PETSC ERROR: >>>>>>> ------------------------------------------------------------------------ >>>>>>> [0]PETSC ERROR: build/poissonIrregular on a arch-linux2-cxx-debug >>>>>>> named edwards by bkhanal Thu Oct 31 19:23:58 2013 >>>>>>> [0]PETSC ERROR: Libraries linked from >>>>>>> /home/bkhanal/Documents/softwares/petsc/arch-linux2-cxx-debug/lib >>>>>>> [0]PETSC ERROR: Configure run at Sat Oct 26 16:35:15 2013 >>>>>>> [0]PETSC ERROR: Configure options --download-mpich >>>>>>> -download-f-blas-lapack=1 --download-metis --download-parmetis >>>>>>> --download-superlu_dist --download-scalapack --download-mumps >>>>>>> --download-hypre --with-clanguage=cxx >>>>>>> [0]PETSC ERROR: >>>>>>> ------------------------------------------------------------------------ >>>>>>> [0]PETSC ERROR: MatSetValues_SeqAIJ() line 413 in >>>>>>> src/mat/impls/aij/seq/aij.c >>>>>>> [0]PETSC ERROR: MatSetValues() line 1130 in >>>>>>> src/mat/interface/matrix.c >>>>>>> [0]PETSC ERROR: computeMatrix2dSection() line 318 in >>>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>>> [0]PETSC ERROR: KSPSetUp() line 228 in src/ksp/ksp/interface/itfunc.c >>>>>>> [0]PETSC ERROR: KSPSolve() line 399 in src/ksp/ksp/interface/itfunc.c >>>>>>> [0]PETSC ERROR: main() line 598 in >>>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>>> [unset]: aborting job: >>>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>>> >>>>>>> >>>>>>>> Matt >>>>>>>> >>>>>>>> >>>>>>>>> However, now since I do not want the rows in the matrix >>>>>>>>> corresponding to the points in the grid which do not lie in the >>>>>>>>> computational domain. This means the size of the matrix will not >>>>>>>>> necessarily equal the total number of cells in DMDA. So in the following >>>>>>>>> code: >>>>>>>>> >>>>>>>>> ierr = DMDAGetLocalInfo(dm,&info);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, >>>>>>>>> &nC);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>>>> PetscInt point; >>>>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>>>> &point);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>>>> } >>>>>>>>> >>>>>>>>> } >>>>>>>>> } >>>>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>>>> >>>>>>>>> should I myself compute proper nC (i.e. total no. of points for >>>>>>>>> which I want the matrix to have a corresponding row) before calling >>>>>>>>> PetscSectionSetChart() or is the masking out of the points inside the loop >>>>>>>>> with if(isPosInDomain(&testPoisson,i,j,0)) sufficient ? >>>>>>>>> And, when you write: >>>>>>>>> PetscSectionGetOffset(gs, p, &ind); >>>>>>>>> row = ind < 0 ? -(ind+1) : ind; >>>>>>>>> >>>>>>>>> it seems you allow the possibility of getting a -ve ind when using >>>>>>>>> PetscSectionGetOffset. When would an offset to a particular dof and point? >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>>>> num = 0; >>>>>>>>>>> /*ierr = >>>>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr);*/ /*Get the PetscPoint*/ >>>>>>>>>>> /*But I did now understand how I would emulate >>>>>>>>>>> the row.c and col.c info with PetscSection*/ >>>>>>>>>>> row.i = i; row.j = j; >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Here you would call >>>>>>>>>> >>>>>>>>>> ierr = DMDAGetCellPoint(da, i, j, 0, &cell);CHKERRQ(ierr); >>>>>>>>>> ierr = PetscSectionGetOffset(da, cell, &row);CHKERRQ(ierr); >>>>>>>>>> ierr = PetscSectionGetDof(da, cell, &dof);CHKERRQ(ierr); >>>>>>>>>> >>>>>>>>>> This means that this cell has global rows = [row, row+dof). >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> col[num].i = i; col[num].j = j; >>>>>>>>>>> if (isPosInDomain(ctx,i,j)) { /*put the >>>>>>>>>>> operator for only the values for only the points lying in the domain*/ >>>>>>>>>>> v[num++] = -4; >>>>>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>>>>> col[num].i = i+1; col[num].j = j; >>>>>>>>>>> v[num++] = 1; >>>>>>>>>>> } >>>>>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>>>>> col[num].i = i-1; col[num].j = j; >>>>>>>>>>> v[num++] = 1; >>>>>>>>>>> } >>>>>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>>>>> col[num].i = i; col[num].j = j+1; >>>>>>>>>>> v[num++] = 1; >>>>>>>>>>> } >>>>>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>>>>> col[num].i = i; col[num].j = j-1; >>>>>>>>>>> v[num++] = 1; >>>>>>>>>>> } >>>>>>>>>>> } else { >>>>>>>>>>> v[num++] = dScale; /*set Dirichlet identity >>>>>>>>>>> rows for points in the rectangle but outside the computational domain*/ >>>>>>>>>>> } >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> You do the same thing for the columns, and then call >>>>>>>>>> >>>>>>>>>> ierr = MatSetValues(A, dof, rows, num*dof, cols, v, >>>>>>>>>> INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> ierr = >>>>>>>>>>> MatSetValuesStencil(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>>> } >>>>>>>>>>> } >>>>>>>>>>> } >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> Thanks, >>>>>>>>>>>> >>>>>>>>>>>> Matt >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> Matt >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> If not, just put the identity into >>>>>>>>>>>>>>>>>>>>>> the rows you do not use on the full cube. It will not >>>>>>>>>>>>>>>>>>>>>> hurt scalability or convergence. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> In the case of Poisson with Dirichlet condition this >>>>>>>>>>>>>>>>>>>>> might be the case. But is it always true that having identity rows in the >>>>>>>>>>>>>>>>>>>>> system matrix will not hurt convergence ? I thought otherwise for the >>>>>>>>>>>>>>>>>>>>> following reasons: >>>>>>>>>>>>>>>>>>>>> 1) Having read Jed's answer here : >>>>>>>>>>>>>>>>>>>>> http://scicomp.stackexchange.com/questions/3426/why-is-pinning-a-point-to-remove-a-null-space-bad/3427#3427 >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Jed is talking about a constraint on a the pressure at >>>>>>>>>>>>>>>>>>>> a point. This is just decoupling these unknowns from the rest >>>>>>>>>>>>>>>>>>>> of the problem. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> 2) Some observation I am getting (but I am still doing >>>>>>>>>>>>>>>>>>>>> more experiments to confirm) while solving my staggered-grid 3D stokes flow >>>>>>>>>>>>>>>>>>>>> with schur complement and using -pc_type gamg for A00 matrix. Putting the >>>>>>>>>>>>>>>>>>>>> identity rows for dirichlet boundaries and for ghost cells seemed to have >>>>>>>>>>>>>>>>>>>>> effects on its convergence. I'm hoping once I know how to use PetscSection, >>>>>>>>>>>>>>>>>>>>> I can get rid of using ghost cells method for the staggered grid and get >>>>>>>>>>>>>>>>>>>>> rid of the identity rows too. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> It can change the exact iteration, but it does not make >>>>>>>>>>>>>>>>>>>> the matrix conditioning worse. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Anyway please provide me with some pointers so that >>>>>>>>>>>>>>>>>>>>> I can start trying with petscsection on top of a dmda, in the beginning for >>>>>>>>>>>>>>>>>>>>> non-staggered case. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>> their experiments lead. >>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>>> experiments lead. >>>>>>>>>> -- Norbert Wiener >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> What most experimenters take for granted before they begin their >>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>> experiments lead. >>>>>>>> -- Norbert Wiener >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> What most experimenters take for granted before they begin their >>>>>> experiments is infinitely more interesting than any results to which their >>>>>> experiments lead. >>>>>> -- Norbert Wiener >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> What most experimenters take for granted before they begin their >>>> experiments is infinitely more interesting than any results to which their >>>> experiments lead. >>>> -- Norbert Wiener >>>> >>> >>> >> >> >> -- >> What most experimenters take for granted before they begin their >> experiments is infinitely more interesting than any results to which their >> experiments lead. >> -- Norbert Wiener >> > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From bisheshkh at gmail.com Mon Nov 4 10:25:45 2013 From: bisheshkh at gmail.com (Bishesh Khanal) Date: Mon, 4 Nov 2013 17:25:45 +0100 Subject: [petsc-users] using petsc tools to solve isolated irregular domains with finite difference In-Reply-To: References: Message-ID: On Mon, Nov 4, 2013 at 4:54 PM, Matthew Knepley wrote: > On Mon, Nov 4, 2013 at 4:49 AM, Bishesh Khanal wrote: > >> On Fri, Nov 1, 2013 at 1:30 PM, Matthew Knepley wrote: >> >>> On Fri, Nov 1, 2013 at 12:08 AM, Bishesh Khanal wrote: >>> >>>> On Thu, Oct 31, 2013 at 11:34 PM, Matthew Knepley wrote: >>>> >>>>> On Thu, Oct 31, 2013 at 5:02 PM, Bishesh Khanal wrote: >>>>> >>>>>> On Thu, Oct 31, 2013 at 7:43 PM, Matthew Knepley wrote: >>>>>> >>>>>>> On Thu, Oct 31, 2013 at 1:25 PM, Bishesh Khanal >>>>>> > wrote: >>>>>>> >>>>>>>> I did not call DMCreateMatrix() before when using just dmda (and it >>>>>>>> is working). In any case, I added a call to DMCreateMatrix() now but still >>>>>>>> there are problems. Here are the code snippets and the error I get: >>>>>>>> >>>>>>> >>>>>>> Then you were allowing it to be called automatically by PETSc, and >>>>>>> it would have been this time as well. >>>>>>> >>>>>>> >>>>>>>> //Setting up the section and linking it to DM: >>>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>>> PetscInt point; >>>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>>> &point);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>>> } >>>>>>>> } >>>>>>>> } >>>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>>>>>> >>>>>>>> //Set up KSP: >>>>>>>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>>>>>>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>>>>>>> ierr = >>>>>>>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>>>> ierr = >>>>>>>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>>>> ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); >>>>>>>> ierr = KSPSolve(ksp,NULL,NULL);CHKERRQ(ierr); >>>>>>>> >>>>>>>> ------------------------------------------------------ >>>>>>>> The computeMatrix2dSection function has: >>>>>>>> >>>>>>>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>>>>>>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>>>>>>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>>>>>>> for(PetscInt j = info.ys; j>>>>>>> for(PetscInt i = info.xs; i>>>>>>> if (isPosInDomain(ctx,i,j)) { >>>>>>>> ierr = >>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>>>>> ierr = PetscSectionGetDof(gs,point,&dof); >>>>>>>> for(PetscInt cDof = 0; cDof < dof; ++cDof) { >>>>>>>> num = 0; >>>>>>>> row+=cDof; >>>>>>>> col[num] = row; //(i,j) position >>>>>>>> v[num++] = -4; >>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>> ierr = >>>>>>>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>>>>>>> ierr = >>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>> col[num] += cDof; >>>>>>>> v[num++] = 1; >>>>>>>> } >>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>> ierr = >>>>>>>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>>>>>>> ierr = >>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>> col[num] += cDof; >>>>>>>> v[num++] = 1; >>>>>>>> } >>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>> ierr = >>>>>>>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>>>>>>> ierr = >>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>> col[num] += cDof; >>>>>>>> v[num++] = 1; >>>>>>>> } >>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>> ierr = >>>>>>>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>>>>>>> ierr = >>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>> col[num] += cDof; >>>>>>>> v[num++] = 1; >>>>>>>> } >>>>>>>> ierr = >>>>>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>> } >>>>>>>> } >>>>>>>> } >>>>>>>> } >>>>>>>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>>>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>>>> >>>>>>>> But this is not working. For e.g. for a 6X6 grid size I get the >>>>>>>> following error: >>>>>>>> >>>>>>> >>>>>>> Okay, here is how to debug this: >>>>>>> >>>>>>> 0) Go to a single scalar equations to make things easier >>>>>>> >>>>>>> 1) Use MatView(A, PETSC_VIEWER_STDOUT_WORLD), which should output >>>>>>> a 36 row matrix with >>>>>>> the preallocated nonzero pattern, filled with zeros >>>>>>> >>>>>>> 2) Make sure this is the pattern you want >>>>>>> >>>>>>> 3) Run in the debugger with -start_in_debugger >>>>>>> >>>>>>> 4) When you get the error, see >>>>>>> >>>>>>> a) If the (i, j) is one that should be setting a value >>>>>>> >>>>>>> b) Why this (i, j) was not preallocated >>>>>>> >>>>>> >>>>>> Up to 4 (a), it is correct. There is a problem in the way >>>>>> DMDAGetCellPoint and PetscSectionGetOffset works in this case scenario. I >>>>>> will try to explain it below for the case of 4X4 grid. >>>>>> First case: >>>>>> If I set the computational domain to be all the points of the dmda >>>>>> grid, (i.e. isPosInDomain(..,i,j,..) returns true for all i and j in the >>>>>> dmda grid), the program runs fine and does not give any error. >>>>>> >>>>>> Second case: >>>>>> I want the computational domain to be some part of the whole grid. >>>>>> There is a problem in this case. >>>>>> The following test is in a case where, >>>>>> isPosInDomain(..,i,j,..) returns true ONLY for (i,j) pairs of (2,1) >>>>>> (1,2) (2,2) (3,2) and (3,2). The grid with its corresponding point number >>>>>> in the petscsection is shown below: >>>>>> >>>>>> 12 13 14 15 >>>>>> 8 9 10 11 >>>>>> 4 5 6 7 >>>>>> 0 1 2 3 >>>>>> >>>>>> of which 6, 9, 10, 11 and 14 correspond to the (i,j) points that >>>>>> returns true for isPosInDomain(..,i,j,..) >>>>>> MatView gives me the 16-row matrix with the star stencil non-zero >>>>>> structure as expected. >>>>>> The error I get is: new non-zero at (0,2) caused a malloc. >>>>>> >>>>>> This error is for the value of (i,j) = (2,1), i.e. point 6. >>>>>> The loop to set values in the matrix starts as: >>>>>> for(PetscInt j = info.ys; j>>>>> for(PetscInt i = info.xs; i>>>>> if (isPosInDomain(ctx,i,j)) { >>>>>> ierr = >>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>>> >>>>>> Now, what I wanted from the beginning is to create the matrix >>>>>> containing the rows corresponding to only those points (i,j) which has true >>>>>> values for isPosInDomain(ctx,i,j), that means in this case 5 row matrix. In >>>>>> the current test, the first (i,j) that reaches DMDAGetCellPoint is (2,1), >>>>>> which gives sets point variable to 6. >>>>>> The line PetscSectionGetOffset(gs,point,&row) sets the value of row >>>>>> to 0. >>>>>> So I think this where the inconsistency lies. >>>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES) expects row variable >>>>>> and col[] array to correspond to (i,j) based on DMDA, but >>>>>> PetsSectionGetOffset gives result based on how we masked away some of the >>>>>> points from dmda grid. Or am I missing something very basic here ? >>>>>> >>>>> >>>>> You are missing something basic. The row is correctly identified as 0, >>>>> and that means the 2 is >>>>> point 10. >>>>> >>>> The problem must be preallocation. First, you should send the result of >>>>> MatView(). >>>>> >>>> >>>> MatView shows that the preallocation and the matrix created is for the >>>> full dmda grid which is probably not what we want here, right ? Here is the >>>> matview result: >>>> >>> >>> Thats right. Did you call >>> >>> DMSetDefaultSection(da, s) >>> >>> after you made the new section and before you called DMGetMatrix()? >>> >> Do you mean before I called DMCreateMatrix() ? I called >> DMSetDefaultSection(da,s) as below: >> >> ... >> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >> // ierr = DMGetMatrix(dm,&A);CHKERRQ(ierr); //It seems DMGetMatrix is >> changed to DMCreateMatrix(). >> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >> ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >> > > Yes, this is correct. I do exactly this in the example code and it works > fine. I suspect > you are not creating the section you think you are. You can send the code > if you want > and I will look at it. > Thanks, here is the code: https://github.com/bishesh/petscPoissonIrregular/blob/master/poissonIrregular.cxx It works for 2D and 3D without using PetscSection. (The choice of using Petsc_section is on line 68). Using PetsSection currently it implements only the 2D case with dof = 1. It works when the domain consists of all the cells of the dmda, it works (to check you can uncomment line 129 in isPosInDomain(). But it does not when I mask out some of the cells while creating the petscsection. > > Matt > > >> This MatView on A shows (result below) that it still has 16 rows for all >> DMDA points instead of only those petscsection points for which I set a >> non-zero dof. Is it what it should not be doing ? >> >>> Mat Object: 1 MPI processes >>>> >>>> type: seqaij >>>> >>>> row 0: (0, 0) (1, 0) (4, 0) >>>> >>>> row 1: (0, 0) (1, 0) (2, 0) (5, 0) >>>> >>>> row 2: (1, 0) (2, 0) (3, 0) (6, 0) >>>> >>>> row 3: (2, 0) (3, 0) (7, 0) >>>> >>>> row 4: (0, 0) (4, 0) (5, 0) (8, 0) >>>> >>>> row 5: (1, 0) (4, 0) (5, 0) (6, 0) (9, 0) >>>> >>>> row 6: (2, 0) (5, 0) (6, 0) (7, 0) (10, 0) >>>> >>>> row 7: (3, 0) (6, 0) (7, 0) (11, 0) >>>> >>>> row 8: (4, 0) (8, 0) (9, 0) (12, 0) >>>> >>>> row 9: (5, 0) (8, 0) (9, 0) (10, 0) (13, 0) >>>> >>>> row 10: (6, 0) (9, 0) (10, 0) (11, 0) (14, 0) >>>> >>>> row 11: (7, 0) (10, 0) (11, 0) (15, 0) >>>> >>>> row 12: (8, 0) (12, 0) (13, 0) >>>> >>>> row 13: (9, 0) (12, 0) (13, 0) (14, 0) >>>> >>>> row 14: (10, 0) (13, 0) (14, 0) (15, 0) >>>> >>>> row 15: (11, 0) (14, 0) (15, 0) >>>> >>>> [0]PETSC ERROR: --------------------- Error Message >>>> ------------------------------------ >>>> >>>> [0]PETSC ERROR: Argument out of range! >>>> >>>> [0]PETSC ERROR: New nonzero at (0,2) caused a malloc! >>>> >>>> >>>> >> For reference here is the code snippet again: >> ... >> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >> PetscInt point; >> if(isPosInDomain(&testPoisson,i,j,0)) { >> ierr = DMDAGetCellPoint(dm, i, j, 0, >> &point);CHKERRQ(ierr); >> ierr = PetscSectionSetDof(s, point, testPoisson.mDof); // >> No. of dofs associated with the point. >> } >> } >> } >> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >> ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >> >> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >> >> ierr = >> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >> ierr = >> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >> >> .... >> The function computeMatrix2dSection: >> >> #undef __FUNCT__ >> #define __FUNCT__ "computeMatrix2dSection" >> PetscErrorCode computeMatrix2dSection(KSP ksp, Mat A, Mat B, MatStructure >> *str, void *context) { >> PoissonModel *ctx = (PoissonModel*)context; >> PetscErrorCode ierr; >> DM da; >> DMDALocalInfo info; >> PetscInt row, col[5]; >> PetscInt dof; >> PetscScalar v[5]; //array to store 5 point stencil for >> each row >> PetscInt num; //non-zero position in the current row >> // PetscScalar dScale = 1; //to scale Dirichlet >> identity rows if needed >> PetscSection gs; //Global Section that keeps >> the grid info and indices >> PetscInt point; //Current point of the >> petscSection >> >> PetscFunctionBeginUser; >> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >> ierr = DMCreateMatrix(da,&A);CHKERRQ(ierr); >> // ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >> for(PetscInt j = info.ys; j> for(PetscInt i = info.xs; i> if (isPosInDomain(ctx,i,j)) { >> ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >> ierr = PetscSectionGetOffset(gs,point,&row); >> ierr = PetscSectionGetDof(gs,point,&dof); >> for(PetscInt cDof = 0; cDof < ctx->mDof; ++cDof) { >> num = 0; >> row+=cDof; >> col[num] = row; //(i,j) position >> v[num++] = -4; >> if(isPosInDomain(ctx,i+1,j)) { >> ierr = >> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >> ierr = PetscSectionGetOffset(gs,point,&col[num]); >> col[num] += cDof; >> v[num++] = 1; >> } >> if(isPosInDomain(ctx,i-1,j)) { >> ierr = >> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >> ierr = PetscSectionGetOffset(gs,point,&col[num]); >> col[num] += cDof; >> v[num++] = 1; >> } >> if(isPosInDomain(ctx,i,j+1)) { >> ierr = >> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >> ierr = PetscSectionGetOffset(gs,point,&col[num]); >> col[num] += cDof; >> v[num++] = 1; >> } >> if(isPosInDomain(ctx,i,j-1)) { >> ierr = >> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >> ierr = PetscSectionGetOffset(gs,point,&col[num]); >> col[num] += cDof; >> v[num++] = 1; >> } >> ierr = >> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >> } >> } >> } >> } >> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >> >> >> >> >>> >>>>> Matt >>>>> >>>>> >>>>>> >>>>>>> Thanks, >>>>>>> >>>>>>> Matt >>>>>>> >>>>>>> >>>>>>>> [0]PETSC ERROR: --------------------- Error Message >>>>>>>> ------------------------------------ >>>>>>>> [0]PETSC ERROR: Argument out of range! >>>>>>>> [0]PETSC ERROR: New nonzero at (0,6) caused a malloc! >>>>>>>> [0]PETSC ERROR: >>>>>>>> ------------------------------------------------------------------------ >>>>>>>> [0]PETSC ERROR: Petsc Development GIT revision: >>>>>>>> 61e5e40bb2c5bf2423e94b71a15fef47e411b0da GIT Date: 2013-10-25 21:47:45 >>>>>>>> -0500 >>>>>>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>>>>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>>>>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>>>>>> [0]PETSC ERROR: >>>>>>>> ------------------------------------------------------------------------ >>>>>>>> [0]PETSC ERROR: build/poissonIrregular on a arch-linux2-cxx-debug >>>>>>>> named edwards by bkhanal Thu Oct 31 19:23:58 2013 >>>>>>>> [0]PETSC ERROR: Libraries linked from >>>>>>>> /home/bkhanal/Documents/softwares/petsc/arch-linux2-cxx-debug/lib >>>>>>>> [0]PETSC ERROR: Configure run at Sat Oct 26 16:35:15 2013 >>>>>>>> [0]PETSC ERROR: Configure options --download-mpich >>>>>>>> -download-f-blas-lapack=1 --download-metis --download-parmetis >>>>>>>> --download-superlu_dist --download-scalapack --download-mumps >>>>>>>> --download-hypre --with-clanguage=cxx >>>>>>>> [0]PETSC ERROR: >>>>>>>> ------------------------------------------------------------------------ >>>>>>>> [0]PETSC ERROR: MatSetValues_SeqAIJ() line 413 in >>>>>>>> src/mat/impls/aij/seq/aij.c >>>>>>>> [0]PETSC ERROR: MatSetValues() line 1130 in >>>>>>>> src/mat/interface/matrix.c >>>>>>>> [0]PETSC ERROR: computeMatrix2dSection() line 318 in >>>>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>>>> [0]PETSC ERROR: KSPSetUp() line 228 in >>>>>>>> src/ksp/ksp/interface/itfunc.c >>>>>>>> [0]PETSC ERROR: KSPSolve() line 399 in >>>>>>>> src/ksp/ksp/interface/itfunc.c >>>>>>>> [0]PETSC ERROR: main() line 598 in >>>>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>>>> [unset]: aborting job: >>>>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>>>> >>>>>>>> >>>>>>>>> Matt >>>>>>>>> >>>>>>>>> >>>>>>>>>> However, now since I do not want the rows in the matrix >>>>>>>>>> corresponding to the points in the grid which do not lie in the >>>>>>>>>> computational domain. This means the size of the matrix will not >>>>>>>>>> necessarily equal the total number of cells in DMDA. So in the following >>>>>>>>>> code: >>>>>>>>>> >>>>>>>>>> ierr = DMDAGetLocalInfo(dm,&info);CHKERRQ(ierr); >>>>>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, >>>>>>>>>> &nC);CHKERRQ(ierr); >>>>>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>>>>> PetscInt point; >>>>>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>>>>> &point);CHKERRQ(ierr); >>>>>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>>>>> } >>>>>>>>>> >>>>>>>>>> } >>>>>>>>>> } >>>>>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>>>>> >>>>>>>>>> should I myself compute proper nC (i.e. total no. of points for >>>>>>>>>> which I want the matrix to have a corresponding row) before calling >>>>>>>>>> PetscSectionSetChart() or is the masking out of the points inside the loop >>>>>>>>>> with if(isPosInDomain(&testPoisson,i,j,0)) sufficient ? >>>>>>>>>> And, when you write: >>>>>>>>>> PetscSectionGetOffset(gs, p, &ind); >>>>>>>>>> row = ind < 0 ? -(ind+1) : ind; >>>>>>>>>> >>>>>>>>>> it seems you allow the possibility of getting a -ve ind when >>>>>>>>>> using PetscSectionGetOffset. When would an offset to a particular dof and >>>>>>>>>> point? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>>>>> num = 0; >>>>>>>>>>>> /*ierr = >>>>>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr);*/ /*Get the PetscPoint*/ >>>>>>>>>>>> /*But I did now understand how I would emulate >>>>>>>>>>>> the row.c and col.c info with PetscSection*/ >>>>>>>>>>>> row.i = i; row.j = j; >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Here you would call >>>>>>>>>>> >>>>>>>>>>> ierr = DMDAGetCellPoint(da, i, j, 0, &cell);CHKERRQ(ierr); >>>>>>>>>>> ierr = PetscSectionGetOffset(da, cell, &row);CHKERRQ(ierr); >>>>>>>>>>> ierr = PetscSectionGetDof(da, cell, &dof);CHKERRQ(ierr); >>>>>>>>>>> >>>>>>>>>>> This means that this cell has global rows = [row, row+dof). >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> col[num].i = i; col[num].j = j; >>>>>>>>>>>> if (isPosInDomain(ctx,i,j)) { /*put the >>>>>>>>>>>> operator for only the values for only the points lying in the domain*/ >>>>>>>>>>>> v[num++] = -4; >>>>>>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>>>>>> col[num].i = i+1; col[num].j = j; >>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>> } >>>>>>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>>>>>> col[num].i = i-1; col[num].j = j; >>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>> } >>>>>>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>>>>>> col[num].i = i; col[num].j = j+1; >>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>> } >>>>>>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>>>>>> col[num].i = i; col[num].j = j-1; >>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>> } >>>>>>>>>>>> } else { >>>>>>>>>>>> v[num++] = dScale; /*set Dirichlet >>>>>>>>>>>> identity rows for points in the rectangle but outside the computational >>>>>>>>>>>> domain*/ >>>>>>>>>>>> } >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> You do the same thing for the columns, and then call >>>>>>>>>>> >>>>>>>>>>> ierr = MatSetValues(A, dof, rows, num*dof, cols, v, >>>>>>>>>>> INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> ierr = >>>>>>>>>>>> MatSetValuesStencil(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>>>> } >>>>>>>>>>>> } >>>>>>>>>>>> } >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> Thanks, >>>>>>>>>>>>> >>>>>>>>>>>>> Matt >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> If not, just put the identity into >>>>>>>>>>>>>>>>>>>>>>> the rows you do not use on the full cube. It will >>>>>>>>>>>>>>>>>>>>>>> not hurt scalability or convergence. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> In the case of Poisson with Dirichlet condition this >>>>>>>>>>>>>>>>>>>>>> might be the case. But is it always true that having identity rows in the >>>>>>>>>>>>>>>>>>>>>> system matrix will not hurt convergence ? I thought otherwise for the >>>>>>>>>>>>>>>>>>>>>> following reasons: >>>>>>>>>>>>>>>>>>>>>> 1) Having read Jed's answer here : >>>>>>>>>>>>>>>>>>>>>> http://scicomp.stackexchange.com/questions/3426/why-is-pinning-a-point-to-remove-a-null-space-bad/3427#3427 >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Jed is talking about a constraint on a the pressure at >>>>>>>>>>>>>>>>>>>>> a point. This is just decoupling these unknowns from the rest >>>>>>>>>>>>>>>>>>>>> of the problem. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> 2) Some observation I am getting (but I am still >>>>>>>>>>>>>>>>>>>>>> doing more experiments to confirm) while solving my staggered-grid 3D >>>>>>>>>>>>>>>>>>>>>> stokes flow with schur complement and using -pc_type gamg for A00 matrix. >>>>>>>>>>>>>>>>>>>>>> Putting the identity rows for dirichlet boundaries and for ghost cells >>>>>>>>>>>>>>>>>>>>>> seemed to have effects on its convergence. I'm hoping once I know how to >>>>>>>>>>>>>>>>>>>>>> use PetscSection, I can get rid of using ghost cells method for the >>>>>>>>>>>>>>>>>>>>>> staggered grid and get rid of the identity rows too. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> It can change the exact iteration, but it does not >>>>>>>>>>>>>>>>>>>>> make the matrix conditioning worse. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Anyway please provide me with some pointers so that >>>>>>>>>>>>>>>>>>>>>> I can start trying with petscsection on top of a dmda, in the beginning for >>>>>>>>>>>>>>>>>>>>>> non-staggered case. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>>>> experiments lead. >>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>> experiments lead. >>>>>>>>> -- Norbert Wiener >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> What most experimenters take for granted before they begin their >>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>> experiments lead. >>>>>>> -- Norbert Wiener >>>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> What most experimenters take for granted before they begin their >>>>> experiments is infinitely more interesting than any results to which their >>>>> experiments lead. >>>>> -- Norbert Wiener >>>>> >>>> >>>> >>> >>> >>> -- >>> What most experimenters take for granted before they begin their >>> experiments is infinitely more interesting than any results to which their >>> experiments lead. >>> -- Norbert Wiener >>> >> >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From anush at bu.edu Mon Nov 4 11:09:18 2013 From: anush at bu.edu (Anush Krishnan) Date: Mon, 4 Nov 2013 12:09:18 -0500 Subject: [petsc-users] Questions about DMDA Local and Global vectors Message-ID: Hi, I have some questions about the differences between distributed local and global vectors that are created using the same DM. 1. I used a local vector for 'x' in the KSPSolve function, and I got a size mismatch error (due to extra ghost points on the local vector). So I had to use a global vector to perform the linear solve, and then later copy the values to a local vector to be able to perform finite-difference stencil-type operations. This doubles my memory usage. Is there a way to use this without the extra storage? 2. When I use VecView to print the vectors to file, the global vector gets printed in full, but only the local values on process 0 are printed for a local vector. 3. Should I assume that I always pass only global vectors to PETSc functions, and use local vectors only when I perform the stencil operations? Thank you, Anush -------------- next part -------------- An HTML attachment was scrubbed... URL: From bsmith at mcs.anl.gov Mon Nov 4 11:21:08 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Mon, 4 Nov 2013 11:21:08 -0600 Subject: [petsc-users] MatSetValues vs. MatSetValue In-Reply-To: <1383577302.81115.YahooMailNeo@web160204.mail.bf1.yahoo.com> References: <1383577302.81115.YahooMailNeo@web160204.mail.bf1.yahoo.com> Message-ID: Normally one does not ?generate? a single matrix entry at a time. For example with finite elements one generates an element stiffness which is a small dense block of values that need to be added to the global stiffness matrix. With finite differences one generates a row of the matrix entries at the time. One should then call MatSetValues() to set this collection of entries that have been computed together. Since you have all this collection it would be silly to call MatSetValue() on each. If, for some nonstandard uncommon situation one truly generates one matrix entry at a time then it would be fine to call MatSetValue() with that one entry. Barry On Nov 4, 2013, at 9:01 AM, Qin Lu wrote: > Hello, > > I need to set values of a sparse matrix. Is setting coefficients row by row using MatSetValues more efficient than setting setting coefficient one by one using MatSetValue? Is MatSetValues implemented by looping over each entry of the input array calling MatSetValue? > > Thanks a lot for your help, > Qin From danyang.su at gmail.com Mon Nov 4 13:42:45 2013 From: danyang.su at gmail.com (Danyang Su) Date: Mon, 04 Nov 2013 11:42:45 -0800 Subject: [petsc-users] [petsc-maint] Speedup problem when using OpenMP? In-Reply-To: <5277B468.4090509@mcs.anl.gov> References: <5272EDAD.2070001@gmail.com> <5277B468.4090509@mcs.anl.gov> Message-ID: <5277F8B5.9010204@gmail.com> Hi Karli, This does not make any difference. I have scaled up the matrix but the performance does not change. If I run with OpenMP, the iteration number is always the same whatever how many processors are used. This seems quite strange as the iteration number usually increase as the number of processors increased when run with MPI. I think I should move to the ubuntu system to make further test, to see if this is a windows problem. Thanks, Danyang On 04/11/2013 6:51 AM, Karl Rupp wrote: > Hi, > > > I have a question on the speedup of PETSc when using OpenMP. I can get >> good speedup when using MPI, but no speedup when using OpenMP. >> The example is ex2f with m=100 and n=100. The number of available >> processors is 16 (32 threads) and the OS is Windows Server 2012. The log >> files for 4 and 8 processors are attached. >> >> The commands I used to run with 4 processors are as follows: >> Run using MPI >> mpiexec -n 4 Petsc-windows-ex2f.exe -m 100 -n 100 -log_summary >> log_100x100_mpi_p4.log >> >> Run using OpenMP >> Petsc-windows-ex2f.exe -threadcomm_type openmp -threadcomm_nthreads 4 -m >> 100 -n 100 -log_summary log_100x100_openmp_p4.log >> >> The PETSc used for this test is PETSc for Windows >> http://www.mic-tc.ch/downloads/PETScForWindows.zip, but I guess this is >> not the problem because the same problem exists when I use PETSc-dev in >> Cygwin. I don't know if this problem exists in Linux, would anybody help >> to test? > > For the 100x100 case considered, the execution times per call are > somewhere in the millisecond to sub-millisecond range (e.g. 1.3ms for > 68 calls to VecScale with 4 processors). I'd say this is too small in > order to see any reasonable performance gain when running multiple > threads, consider problem sizes of about 1000x1000 instead. > > Moreover, keep in mind that typically you won't get a perfectly linear > scaling with the number of processor cores, because ultimately the > memory bandwidth is the limiting factor for standard vector operations. > > Best regards, > Karli > -------------- next part -------------- ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- Petsc-windows-ex2f.exe on a arch-mswin-c-opt named STARGAZER2012 with 1 processor, by danyang Mon Nov 04 09:56:14 2013 With 4 threads per MPI_Comm Using Petsc Release Version 3.4.2, Jul, 02, 2013 Max Max/Min Avg Total Time (sec): 1.851e+002 1.00000 1.851e+002 Objects: 4.500e+001 1.00000 4.500e+001 Flops: 2.203e+011 1.00000 2.203e+011 2.203e+011 Flops/sec: 1.190e+009 1.00000 1.190e+009 1.190e+009 MPI Messages: 0.000e+000 0.00000 0.000e+000 0.000e+000 MPI Message Lengths: 0.000e+000 0.00000 0.000e+000 0.000e+000 MPI Reductions: 5.236e+003 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 1.8506e+002 100.0% 2.2028e+011 100.0% 0.000e+000 0.0% 0.000e+000 0.0% 5.235e+003 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 2657 1.0 3.8941e+001 1.0 2.39e+010 1.0 0.0e+000 0.0e+000 0.0e+000 21 11 0 0 0 21 11 0 0 0 614 MatSolve 2657 1.0 5.3882e+001 1.0 2.39e+010 1.0 0.0e+000 0.0e+000 0.0e+000 29 11 0 0 0 29 11 0 0 0 443 MatLUFactorNum 1 1.0 1.0694e-001 1.0 1.10e+007 1.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 103 MatILUFactorSym 1 1.0 6.7413e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 1.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 1 1.0 5.6889e-007 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 6.2112e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 2.2756e-006 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 9.1984e-003 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 2.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecMDot 2571 1.0 3.6956e+001 1.0 7.95e+010 1.0 0.0e+000 0.0e+000 2.6e+003 20 36 0 0 49 20 36 0 0 49 2152 VecNorm 2658 1.0 9.5952e-001 1.0 5.32e+009 1.0 0.0e+000 0.0e+000 2.7e+003 1 2 0 0 51 1 2 0 0 51 5540 VecScale 2657 1.0 3.6170e+000 1.0 2.66e+009 1.0 0.0e+000 0.0e+000 0.0e+000 2 1 0 0 0 2 1 0 0 0 735 VecCopy 86 1.0 1.5124e-001 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecSet 88 1.0 9.7820e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 172 1.0 9.2293e-002 1.0 3.44e+008 1.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 3727 VecMAXPY 2657 1.0 4.9379e+001 1.0 8.47e+010 1.0 0.0e+000 0.0e+000 0.0e+000 27 38 0 0 0 27 38 0 0 0 1714 VecNormalize 2657 1.0 4.5800e+000 1.0 7.97e+009 1.0 0.0e+000 0.0e+000 2.7e+003 2 4 0 0 51 2 4 0 0 51 1740 KSPGMRESOrthog 2571 1.0 8.3362e+001 1.0 1.59e+011 1.0 0.0e+000 0.0e+000 2.6e+003 45 72 0 0 49 45 72 0 0 49 1908 KSPSetUp 1 1.0 2.0327e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 1.8441e+002 1.0 2.20e+011 1.0 0.0e+000 0.0e+000 5.2e+003100100 0 0100 100100 0 0100 1194 PCSetUp 1 1.0 1.8362e-001 1.0 1.10e+007 1.0 0.0e+000 0.0e+000 3.0e+000 0 0 0 0 0 0 0 0 0 0 60 PCApply 2657 1.0 5.3887e+001 1.0 2.39e+010 1.0 0.0e+000 0.0e+000 0.0e+000 29 11 0 0 0 29 11 0 0 0 443 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 2 2 151957324 0 Vector 37 37 296057128 0 Krylov Solver 1 1 18360 0 Preconditioner 1 1 976 0 Index Set 3 3 4002280 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 5.68889e-008 #PETSc Option Table entries: -log_summary log_1000x1000_openmp_p4.log -m 1000 -n 1000 -threadcomm_nthreads 4 -threadcomm_type openmp #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 4 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Wed Oct 2 16:35:54 2013 Configure options: --with-cc="win32fe icl" --with-cxx="win32fe icl" --with-fc="win32fe ifort" --with-blas-lapack-dir=/cygdrive/d/HardLinks/PETSc/Intel2013/mkl/lib/intel64 --with-mpi-include=/cygdrive/c/MSMPI/Inc -with-mpi-lib="[/cygdrive/C/MSMPI/Lib/amd64/msmpi.lib,/cygdrive/C/MSMPI/Lib/amd64/msmpifec.lib]" --with-openmp --with-shared-libraries --with-debugging=no --useThreads=0 ----------------------------------------- Libraries compiled on Wed Oct 2 16:35:54 2013 on NB-TT-113812 Machine characteristics: CYGWIN_NT-6.1-WOW64-1.7.25-0.270-5-3-i686-32bit Using PETSc directory: /cygdrive/d/WorkDir/petsc-3.4.2 Using PETSc arch: arch-mswin-c-opt ----------------------------------------- Using C compiler: /cygdrive/d/WorkDir/petsc-3.4.2/bin/win32fe/win32fe icl -MT -O3 -QxW -Qopenmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /cygdrive/d/WorkDir/petsc-3.4.2/bin/win32fe/win32fe ifort -MT -O3 -QxW -fpp -Qopenmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/cygdrive/d/WorkDir/petsc-3.4.2/arch-mswin-c-opt/include -I/cygdrive/d/WorkDir/petsc-3.4.2/include -I/cygdrive/d/WorkDir/petsc-3.4.2/include -I/cygdrive/d/WorkDir/petsc-3.4.2/arch-mswin-c-opt/include -I/cygdrive/c/MSMPI/Inc ----------------------------------------- Using C linker: /cygdrive/d/WorkDir/petsc-3.4.2/bin/win32fe/win32fe icl Using Fortran linker: /cygdrive/d/WorkDir/petsc-3.4.2/bin/win32fe/win32fe ifort Using libraries: -L/cygdrive/d/WorkDir/petsc-3.4.2/arch-mswin-c-opt/lib -L/cygdrive/d/WorkDir/petsc-3.4.2/arch-mswin-c-opt/lib -lpetsc /cygdrive/d/HardLinks/PETSc/Intel2013/mkl/lib/intel64/mkl_intel_lp64_dll.lib mkl_intel_thread_dll.lib mkl_core_dll.lib libiomp5md.lib /cygdrive/C/MSMPI/Lib/amd64/msmpi.lib /cygdrive/C/MSMPI/Lib/amd64/msmpifec.lib Gdi32.lib User32.lib Advapi32.lib Kernel32.lib Ws2_32.lib ----------------------------------------- -------------- next part -------------- ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- Petsc-windows-ex2f.exe on a arch-mswin-c-opt named STARGAZER2012 with 1 processor, by danyang Mon Nov 04 10:04:07 2013 With 8 threads per MPI_Comm Using Petsc Release Version 3.4.2, Jul, 02, 2013 Max Max/Min Avg Total Time (sec): 1.717e+002 1.00000 1.717e+002 Objects: 4.500e+001 1.00000 4.500e+001 Flops: 2.203e+011 1.00000 2.203e+011 2.203e+011 Flops/sec: 1.283e+009 1.00000 1.283e+009 1.283e+009 MPI Messages: 0.000e+000 0.00000 0.000e+000 0.000e+000 MPI Message Lengths: 0.000e+000 0.00000 0.000e+000 0.000e+000 MPI Reductions: 5.236e+003 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 1.7167e+002 100.0% 2.2028e+011 100.0% 0.000e+000 0.0% 0.000e+000 0.0% 5.235e+003 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 2657 1.0 3.5848e+001 1.0 2.39e+010 1.0 0.0e+000 0.0e+000 0.0e+000 21 11 0 0 0 21 11 0 0 0 666 MatSolve 2657 1.0 4.9704e+001 1.0 2.39e+010 1.0 0.0e+000 0.0e+000 0.0e+000 29 11 0 0 0 29 11 0 0 0 481 MatLUFactorNum 1 1.0 1.0581e-001 1.0 1.10e+007 1.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 104 MatILUFactorSym 1 1.0 6.6883e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 1.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 1 1.0 5.6889e-007 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 5.9349e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 1.7067e-006 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 8.4725e-003 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 2.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecMDot 2571 1.0 3.4147e+001 1.0 7.95e+010 1.0 0.0e+000 0.0e+000 2.6e+003 20 36 0 0 49 20 36 0 0 49 2328 VecNorm 2658 1.0 1.1897e+000 1.0 5.32e+009 1.0 0.0e+000 0.0e+000 2.7e+003 1 2 0 0 51 1 2 0 0 51 4468 VecScale 2657 1.0 3.6916e+000 1.0 2.66e+009 1.0 0.0e+000 0.0e+000 0.0e+000 2 1 0 0 0 2 1 0 0 0 720 VecCopy 86 1.0 1.4144e-001 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecSet 88 1.0 9.3055e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 172 1.0 1.5342e-001 1.0 3.44e+008 1.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 2242 VecMAXPY 2657 1.0 4.5747e+001 1.0 8.47e+010 1.0 0.0e+000 0.0e+000 0.0e+000 27 38 0 0 0 27 38 0 0 0 1851 VecNormalize 2657 1.0 4.8846e+000 1.0 7.97e+009 1.0 0.0e+000 0.0e+000 2.7e+003 3 4 0 0 51 3 4 0 0 51 1632 KSPGMRESOrthog 2571 1.0 7.7191e+001 1.0 1.59e+011 1.0 0.0e+000 0.0e+000 2.6e+003 45 72 0 0 49 45 72 0 0 49 2060 KSPSetUp 1 1.0 1.7942e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 1.7101e+002 1.0 2.20e+011 1.0 0.0e+000 0.0e+000 5.2e+003100100 0 0100 100100 0 0100 1288 PCSetUp 1 1.0 1.8123e-001 1.0 1.10e+007 1.0 0.0e+000 0.0e+000 3.0e+000 0 0 0 0 0 0 0 0 0 0 61 PCApply 2657 1.0 4.9709e+001 1.0 2.39e+010 1.0 0.0e+000 0.0e+000 0.0e+000 29 11 0 0 0 29 11 0 0 0 481 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 2 2 151957324 0 Vector 37 37 296057128 0 Krylov Solver 1 1 18360 0 Preconditioner 1 1 976 0 Index Set 3 3 4002280 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 5.68889e-008 #PETSc Option Table entries: -log_summary log_1000x1000_openmp_p8.log -m 1000 -n 1000 -threadcomm_nthreads 8 -threadcomm_type openmp #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 4 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Wed Oct 2 16:35:54 2013 Configure options: --with-cc="win32fe icl" --with-cxx="win32fe icl" --with-fc="win32fe ifort" --with-blas-lapack-dir=/cygdrive/d/HardLinks/PETSc/Intel2013/mkl/lib/intel64 --with-mpi-include=/cygdrive/c/MSMPI/Inc -with-mpi-lib="[/cygdrive/C/MSMPI/Lib/amd64/msmpi.lib,/cygdrive/C/MSMPI/Lib/amd64/msmpifec.lib]" --with-openmp --with-shared-libraries --with-debugging=no --useThreads=0 ----------------------------------------- Libraries compiled on Wed Oct 2 16:35:54 2013 on NB-TT-113812 Machine characteristics: CYGWIN_NT-6.1-WOW64-1.7.25-0.270-5-3-i686-32bit Using PETSc directory: /cygdrive/d/WorkDir/petsc-3.4.2 Using PETSc arch: arch-mswin-c-opt ----------------------------------------- Using C compiler: /cygdrive/d/WorkDir/petsc-3.4.2/bin/win32fe/win32fe icl -MT -O3 -QxW -Qopenmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /cygdrive/d/WorkDir/petsc-3.4.2/bin/win32fe/win32fe ifort -MT -O3 -QxW -fpp -Qopenmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/cygdrive/d/WorkDir/petsc-3.4.2/arch-mswin-c-opt/include -I/cygdrive/d/WorkDir/petsc-3.4.2/include -I/cygdrive/d/WorkDir/petsc-3.4.2/include -I/cygdrive/d/WorkDir/petsc-3.4.2/arch-mswin-c-opt/include -I/cygdrive/c/MSMPI/Inc ----------------------------------------- Using C linker: /cygdrive/d/WorkDir/petsc-3.4.2/bin/win32fe/win32fe icl Using Fortran linker: /cygdrive/d/WorkDir/petsc-3.4.2/bin/win32fe/win32fe ifort Using libraries: -L/cygdrive/d/WorkDir/petsc-3.4.2/arch-mswin-c-opt/lib -L/cygdrive/d/WorkDir/petsc-3.4.2/arch-mswin-c-opt/lib -lpetsc /cygdrive/d/HardLinks/PETSc/Intel2013/mkl/lib/intel64/mkl_intel_lp64_dll.lib mkl_intel_thread_dll.lib mkl_core_dll.lib libiomp5md.lib /cygdrive/C/MSMPI/Lib/amd64/msmpi.lib /cygdrive/C/MSMPI/Lib/amd64/msmpifec.lib Gdi32.lib User32.lib Advapi32.lib Kernel32.lib Ws2_32.lib ----------------------------------------- From mpovolot at purdue.edu Mon Nov 4 14:27:34 2013 From: mpovolot at purdue.edu (Michael Povolotskyi) Date: Mon, 04 Nov 2013 15:27:34 -0500 Subject: [petsc-users] calling lapack for LU decomposition Message-ID: <52780336.4070803@purdue.edu> Dear Petsc developers, in my work I need to measure the multi threading performance of LAPACK function dgetrf Since all matrices are of PETSc type I'd like to call Petsc functions for the LU decomposition. My code reads like this Mat A; //serial dense matrix PetscErrorCode ierr; KSP ksp; Mat F; PC pc; ierr = KSPCreate(comm, &ksp); ierr = KSPSetOperators(ksp, A, A, SAME_PRECONDITIONER); ierr = KSPSetType(ksp, KSPPREONLY); ierr = KSPGetPC(ksp, &pc); ierr = PCSetType(pc, PCLU); ierr = PCFactorSetMatSolverPackage(pc, "petsc"); ierr = KSPSetUp(ksp); ierr = PCFactorGetMatrix(pc, &F); Is this a correct way to call dgetrf ? Thank you, Michael. -- Michael Povolotskyi, PhD Research Assistant Professor Network for Computational Nanotechnology 207 S Martin Jischke Drive Purdue University, DLR, room 441-10 West Lafayette, Indiana 47907 phone: +1-765-494-9396 fax: +1-765-496-6026 From jedbrown at mcs.anl.gov Mon Nov 4 14:37:26 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Mon, 04 Nov 2013 13:37:26 -0700 Subject: [petsc-users] calling lapack for LU decomposition In-Reply-To: <52780336.4070803@purdue.edu> References: <52780336.4070803@purdue.edu> Message-ID: <878ux329a1.fsf@mcs.anl.gov> Michael Povolotskyi writes: > Dear Petsc developers, > in my work I need to measure the multi threading performance of LAPACK > function dgetrf > > Since all matrices are of PETSc type I'd like to call Petsc functions > for the LU decomposition. > My code reads like this > > Mat A; //serial dense matrix > > PetscErrorCode ierr; > KSP ksp; > Mat F; > PC pc; > > ierr = KSPCreate(comm, &ksp); > ierr = KSPSetOperators(ksp, A, A, SAME_PRECONDITIONER); > ierr = KSPSetType(ksp, KSPPREONLY); > ierr = KSPGetPC(ksp, &pc); > ierr = PCSetType(pc, PCLU); > ierr = PCFactorSetMatSolverPackage(pc, "petsc"); This is the default. > ierr = KSPSetUp(ksp); > ierr = PCFactorGetMatrix(pc, &F); This function is not needed. > Is this a correct way to call dgetrf ? Yeah, run with -log_summary and look for the time spent in the MatLUFactorNumeric event. (You _can_ call the low-level MatLUFactor() directly, but I recommend what you have written because it is more general-purpose so you can easily switch to other methods.) -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From lu_qin_2000 at yahoo.com Mon Nov 4 16:04:08 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Mon, 4 Nov 2013 14:04:08 -0800 (PST) Subject: [petsc-users] MatSetValues vs. MatSetValue In-Reply-To: References: <1383577302.81115.YahooMailNeo@web160204.mail.bf1.yahoo.com> Message-ID: <1383602648.69342.YahooMailNeo@web160202.mail.bf1.yahoo.com> Thanks for all your clarification! ? Qin On Monday, November 4, 2013 11:21 AM, Barry Smith wrote: ? Normally one does not ?generate? a single matrix entry at a time. For example with finite elements one generates an element stiffness which is a small dense block of values that need to be added to the global stiffness matrix. With finite differences one generates a row of the matrix entries at the time. ? One should then call MatSetValues() to set this collection of entries that have been computed together.? Since you have all this collection it would be silly to call MatSetValue() on each. ? If, for some nonstandard uncommon situation one truly generates one matrix entry at a time then it would be fine to call MatSetValue() with that one entry. ? Barry On Nov 4, 2013, at 9:01 AM, Qin Lu wrote: > Hello, >? > I need to set values of a sparse matrix. Is setting coefficients row by row using MatSetValues more efficient than setting setting coefficient one by one using MatSetValue? Is MatSetValues implemented by looping over each entry of the input array calling MatSetValue? >? > Thanks a lot for your help, > Qin -------------- next part -------------- An HTML attachment was scrubbed... URL: From Lukasz.Kaczmarczyk at glasgow.ac.uk Mon Nov 4 17:22:39 2013 From: Lukasz.Kaczmarczyk at glasgow.ac.uk (Lukasz Kaczmarczyk) Date: Mon, 4 Nov 2013 23:22:39 +0000 Subject: [petsc-users] MatAXPY, creations and destructions Message-ID: <08797FF1-4356-47B4-9F77-A44D39729F27@glasgow.ac.uk> Hello All, I use MatAXPY (DIFFERENT_NONZERO_PATTERN), with petsc-3.4.3. Works ok, no problem, however in -log_summary as result MatAXPY I have more destructions than creations, each time MatAXPY was used. I have tested this for MATMPIAIJ. Valgring don't show anything wrong with that. Kind regards, Lukasz From danyang.su at gmail.com Mon Nov 4 18:26:18 2013 From: danyang.su at gmail.com (Danyang Su) Date: Mon, 04 Nov 2013 16:26:18 -0800 Subject: [petsc-users] Speedup problem when using OpenMP? In-Reply-To: <5272EDAD.2070001@gmail.com> References: <5272EDAD.2070001@gmail.com> Message-ID: <52783B2A.4090909@gmail.com> Hi All, I have test the same example under Ubuntu12.04 X64. The PETSc-dev version is update to date (GIT Date: 2013-11-01 14:59:20 -0500) and the installation is smooth without any error. The speedup of MPI version is linear scalable but the speedup of OpenMP version does not change. *From the CPU usage, the program still run in one thread when use OpenMP. * The commands to run the test are as follows: openmp ./ex2f -threadcomm_type openmp -threadcomm_nthreads 4 -m 1000 -n 1000 -log_summary log_ex2f_1000x1000_ubuntu1204_omp_p4.log mpi mpiexec -n 4 ./ex2f -m 1000 -n 1000 -log_summary log_ex2f_1000x1000_ubuntu1204_mpi_p4.log This problem is so tricky to me. Can anybody confirm if KSP solver is parallelized for OpenMP version? Thanks and regards, Danyang On 31/10/2013 4:54 PM, Danyang Su wrote: > Hi All, > > I have a question on the speedup of PETSc when using OpenMP. I can get > good speedup when using MPI, but no speedup when using OpenMP. > The example is ex2f with m=100 and n=100. The number of available > processors is 16 (32 threads) and the OS is Windows Server 2012. The > log files for 4 and 8 processors are attached. > > The commands I used to run with 4 processors are as follows: > Run using MPI > mpiexec -n 4 Petsc-windows-ex2f.exe -m 100 -n 100 -log_summary > log_100x100_mpi_p4.log > > Run using OpenMP > Petsc-windows-ex2f.exe -threadcomm_type openmp -threadcomm_nthreads 4 > -m 100 -n 100 -log_summary log_100x100_openmp_p4.log > > The PETSc used for this test is PETSc for Windows > http://www.mic-tc.ch/downloads/PETScForWindows.zip, but I guess this > is not the problem because the same problem exists when I use > PETSc-dev in Cygwin. I don't know if this problem exists in Linux, > would anybody help to test? > > Thanks and regards, > > Danyang -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex2f on a linux-gnu-omp-opt named dsu-pc with 1 processor, by root Mon Nov 4 15:35:47 2013 With 4 threads per MPI_Comm Using Petsc Development GIT revision: 1beacf92f04482972e84431be0032cb960d262c7 GIT Date: 2013-11-01 14:59:20 -0500 Max Max/Min Avg Total Time (sec): 2.376e+02 1.00000 2.376e+02 Objects: 4.500e+01 1.00000 4.500e+01 Flops: 2.203e+11 1.00000 2.203e+11 2.203e+11 Flops/sec: 9.271e+08 1.00000 9.271e+08 9.271e+08 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 0.000e+00 0.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.3759e+02 100.0% 2.2028e+11 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 2657 1.0 4.1715e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 18 11 0 0 0 18 11 0 0 0 573 MatSolve 2657 1.0 6.4028e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 373 MatLUFactorNum 1 1.0 1.1149e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 99 MatILUFactorSym 1 1.0 8.2365e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 1 1.0 7.8678e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 9.1023e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 1.0014e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 1.2122e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecMDot 2571 1.0 5.1144e+01 1.0 7.95e+10 1.0 0.0e+00 0.0e+00 0.0e+00 22 36 0 0 0 22 36 0 0 0 1555 VecNorm 2658 1.0 5.4516e+00 1.0 5.32e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 2 0 0 0 2 2 0 0 0 975 VecScale 2657 1.0 3.8631e+00 1.0 2.66e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 1 0 0 0 2 1 0 0 0 688 VecCopy 86 1.0 2.2233e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 88 1.0 1.1501e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 172 1.0 4.4589e-01 1.0 3.44e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 771 VecMAXPY 2657 1.0 6.9213e+01 1.0 8.47e+10 1.0 0.0e+00 0.0e+00 0.0e+00 29 38 0 0 0 29 38 0 0 0 1223 VecNormalize 2657 1.0 9.3968e+00 1.0 7.97e+09 1.0 0.0e+00 0.0e+00 0.0e+00 4 4 0 0 0 4 4 0 0 0 848 KSPGMRESOrthog 2571 1.0 1.1630e+02 1.0 1.59e+11 1.0 0.0e+00 0.0e+00 0.0e+00 49 72 0 0 0 49 72 0 0 0 1367 KSPSetUp 1 1.0 2.8520e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 2.3699e+02 1.0 2.20e+11 1.0 0.0e+00 0.0e+00 0.0e+00100100 0 0 0 100100 0 0 0 929 PCSetUp 1 1.0 2.0609e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 53 PCApply 2657 1.0 6.4088e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 373 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 2 2 151957404 0 Vector 37 37 296057424 0 Krylov Solver 1 1 18368 0 Preconditioner 1 1 984 0 Index Set 3 3 4002304 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 6.50883e-06 #PETSc Option Table entries: -log_summary log_1000x1000_omp_p4.log -m 1000 -n 1000 -threadcomm_nthreads 4 -threadcomm_type openmp #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Mon Nov 4 15:22:12 2013 Configure options: PETSC_ARCH=linux-gnu-omp-opt --with-cc=gcc --with-fc=gfortran --download-f-blas-lapack --with-mpi=0 --with-openmp --with-debugging=0 ----------------------------------------- Libraries compiled on Mon Nov 4 15:22:12 2013 on dsu-pc Machine characteristics: Linux-3.2.0-55-generic-x86_64-with-Ubuntu-12.04-precise Using PETSc directory: /home/dsu/petsc Using PETSc arch: linux-gnu-omp-opt ----------------------------------------- Using C compiler: gcc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O -fopenmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: gfortran -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument -O -fopenmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/dsu/petsc/linux-gnu-omp-opt/include -I/home/dsu/petsc/include -I/home/dsu/petsc/include -I/home/dsu/petsc/linux-gnu-omp-opt/include -I/home/dsu/petsc/include/mpiuni ----------------------------------------- Using C linker: gcc Using Fortran linker: gfortran Using libraries: -Wl,-rpath,/home/dsu/petsc/linux-gnu-omp-opt/lib -L/home/dsu/petsc/linux-gnu-omp-opt/lib -lpetsc -Wl,-rpath,/home/dsu/petsc/linux-gnu-omp-opt/lib -L/home/dsu/petsc/linux-gnu-omp-opt/lib -lflapack -lfblas -lpthread -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lgfortran -lm -lgfortran -lm -lquadmath -lm -lstdc++ -ldl -lgcc_s -ldl ----------------------------------------- -------------- next part -------------- ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex2f on a linux-gnu-omp-opt named dsu-pc with 1 processor, by root Mon Nov 4 15:31:30 2013 Using Petsc Development GIT revision: 1beacf92f04482972e84431be0032cb960d262c7 GIT Date: 2013-11-01 14:59:20 -0500 Max Max/Min Avg Total Time (sec): 2.388e+02 1.00000 2.388e+02 Objects: 4.500e+01 1.00000 4.500e+01 Flops: 2.203e+11 1.00000 2.203e+11 2.203e+11 Flops/sec: 9.224e+08 1.00000 9.224e+08 9.224e+08 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 0.000e+00 0.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.3881e+02 100.0% 2.2028e+11 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 2657 1.0 4.0429e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 17 11 0 0 0 17 11 0 0 0 591 MatSolve 2657 1.0 6.3888e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 374 MatLUFactorNum 1 1.0 1.2874e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 85 MatILUFactorSym 1 1.0 1.3501e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 1 1.0 8.1062e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 6.8491e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 1.1921e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 1.3066e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecMDot 2571 1.0 5.2507e+01 1.0 7.95e+10 1.0 0.0e+00 0.0e+00 0.0e+00 22 36 0 0 0 22 36 0 0 0 1514 VecNorm 2658 1.0 5.4426e+00 1.0 5.32e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 2 0 0 0 2 2 0 0 0 977 VecScale 2657 1.0 3.8871e+00 1.0 2.66e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 1 0 0 0 2 1 0 0 0 684 VecCopy 86 1.0 1.9921e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 88 1.0 1.0965e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 172 1.0 4.0171e-01 1.0 3.44e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 856 VecMAXPY 2657 1.0 7.0096e+01 1.0 8.47e+10 1.0 0.0e+00 0.0e+00 0.0e+00 29 38 0 0 0 29 38 0 0 0 1208 VecNormalize 2657 1.0 9.4060e+00 1.0 7.97e+09 1.0 0.0e+00 0.0e+00 0.0e+00 4 4 0 0 0 4 4 0 0 0 847 KSPGMRESOrthog 2571 1.0 1.1847e+02 1.0 1.59e+11 1.0 0.0e+00 0.0e+00 0.0e+00 50 72 0 0 0 50 72 0 0 0 1342 KSPSetUp 1 1.0 3.7805e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 2.3820e+02 1.0 2.20e+11 1.0 0.0e+00 0.0e+00 0.0e+00100100 0 0 0 100100 0 0 0 925 PCSetUp 1 1.0 2.7698e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 40 PCApply 2657 1.0 6.3946e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 374 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 2 2 151957404 0 Vector 37 37 296057424 0 Krylov Solver 1 1 18368 0 Preconditioner 1 1 984 0 Index Set 3 3 4002304 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 8.51154e-06 #PETSc Option Table entries: -log_summary log_1000x1000_omp_p1.log -m 1000 -n 1000 -threadcomm_nthreads 1 -threadcomm_type openmp #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Mon Nov 4 15:22:12 2013 Configure options: PETSC_ARCH=linux-gnu-omp-opt --with-cc=gcc --with-fc=gfortran --download-f-blas-lapack --with-mpi=0 --with-openmp --with-debugging=0 ----------------------------------------- Libraries compiled on Mon Nov 4 15:22:12 2013 on dsu-pc Machine characteristics: Linux-3.2.0-55-generic-x86_64-with-Ubuntu-12.04-precise Using PETSc directory: /home/dsu/petsc Using PETSc arch: linux-gnu-omp-opt ----------------------------------------- Using C compiler: gcc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O -fopenmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: gfortran -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument -O -fopenmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/dsu/petsc/linux-gnu-omp-opt/include -I/home/dsu/petsc/include -I/home/dsu/petsc/include -I/home/dsu/petsc/linux-gnu-omp-opt/include -I/home/dsu/petsc/include/mpiuni ----------------------------------------- Using C linker: gcc Using Fortran linker: gfortran Using libraries: -Wl,-rpath,/home/dsu/petsc/linux-gnu-omp-opt/lib -L/home/dsu/petsc/linux-gnu-omp-opt/lib -lpetsc -Wl,-rpath,/home/dsu/petsc/linux-gnu-omp-opt/lib -L/home/dsu/petsc/linux-gnu-omp-opt/lib -lflapack -lfblas -lpthread -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lgfortran -lm -lgfortran -lm -lquadmath -lm -lstdc++ -ldl -lgcc_s -ldl ----------------------------------------- -------------- next part -------------- ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex2f on a linux-gnu-opt named dsu-pc with 4 processors, by root Mon Nov 4 16:10:24 2013 Using Petsc Development GIT revision: 1beacf92f04482972e84431be0032cb960d262c7 GIT Date: 2013-11-01 14:59:20 -0500 Max Max/Min Avg Total Time (sec): 5.364e+01 1.00045 5.362e+01 Objects: 5.600e+01 1.00000 5.600e+01 Flops: 2.837e+10 1.00010 2.837e+10 1.135e+11 Flops/sec: 5.291e+08 1.00054 5.290e+08 2.116e+09 MPI Messages: 2.744e+03 2.00000 2.058e+03 8.232e+03 MPI Message Lengths: 2.193e+07 2.00000 7.991e+03 6.578e+07 MPI Reductions: 2.720e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 5.3623e+01 100.0% 1.1347e+11 100.0% 8.232e+03 100.0% 7.991e+03 100.0% 2.719e+03 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 1370 1.0 8.7882e+00 1.0 3.08e+09 1.0 8.2e+03 8.0e+03 0.0e+00 16 11100100 0 16 11100100 0 1402 MatSolve 1370 1.0 9.0304e+00 1.0 3.08e+09 1.0 0.0e+00 0.0e+00 0.0e+00 17 11 0 0 0 17 11 0 0 0 1362 MatLUFactorNum 1 1.0 3.3336e-02 1.0 2.74e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 329 MatILUFactorSym 1 1.0 7.1875e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 1 1.0 7.2212e-0241.6 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 5.4802e-02 1.0 0.00e+00 0.0 1.2e+01 2.0e+03 9.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 1.2875e-05 2.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 4.8881e-03 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecMDot 1325 1.0 1.4754e+01 1.0 1.02e+10 1.0 0.0e+00 0.0e+00 1.3e+03 27 36 0 0 49 27 36 0 0 49 2776 VecNorm 1371 1.0 1.9989e+00 1.1 6.86e+08 1.0 0.0e+00 0.0e+00 1.4e+03 4 2 0 0 50 4 2 0 0 50 1372 VecScale 1370 1.0 4.9844e-01 1.1 3.42e+08 1.0 0.0e+00 0.0e+00 0.0e+00 1 1 0 0 0 1 1 0 0 0 2749 VecCopy 45 1.0 4.4863e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 1418 1.0 6.2273e-01 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 VecAXPY 90 1.0 1.0165e-01 1.0 4.50e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1771 VecMAXPY 1370 1.0 1.5635e+01 1.0 1.09e+10 1.0 0.0e+00 0.0e+00 0.0e+00 29 38 0 0 0 29 38 0 0 0 2789 VecScatterBegin 1370 1.0 1.6159e-01 1.8 0.00e+00 0.0 8.2e+03 8.0e+03 0.0e+00 0 0100100 0 0 0100100 0 0 VecScatterEnd 1370 1.0 9.6929e-01 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 2 0 0 0 0 2 0 0 0 0 0 VecNormalize 1370 1.0 2.5033e+00 1.1 1.03e+09 1.0 0.0e+00 0.0e+00 1.4e+03 5 4 0 0 50 5 4 0 0 50 1642 KSPGMRESOrthog 1325 1.0 2.9419e+01 1.0 2.05e+10 1.0 0.0e+00 0.0e+00 1.3e+03 54 72 0 0 49 54 72 0 0 49 2784 KSPSetUp 2 1.0 2.1291e-02 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 5.2989e+01 1.0 2.84e+10 1.0 8.2e+03 8.0e+03 2.7e+03 99100100100 99 99100100100 99 2141 PCSetUp 2 1.0 1.4600e-01 1.1 2.74e+06 1.0 0.0e+00 0.0e+00 5.0e+00 0 0 0 0 0 0 0 0 0 0 75 PCSetUpOnBlocks 1 1.0 1.1017e-01 1.0 2.74e+06 1.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 100 PCApply 1370 1.0 9.7092e+00 1.0 3.08e+09 1.0 0.0e+00 0.0e+00 0.0e+00 18 11 0 0 0 18 11 0 0 0 1267 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 4 4 56984588 0 Vector 41 41 74071440 0 Vector Scatter 1 1 1060 0 Index Set 5 5 1007832 0 Krylov Solver 2 2 19520 0 Preconditioner 2 2 1864 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 6.19888e-06 Average time for MPI_Barrier(): 0.000529623 Average time for zero size MPI_Send(): 0.000117242 #PETSc Option Table entries: -log_summary log_1000x1000_mpi_p4.log -m 1000 -n 1000 #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Mon Nov 4 14:29:26 2013 Configure options: PETSC_ARCH=linux-gnu-opt --with-cc=gcc --with-fc=gfortran --with-debugging=0 --download-f-blas-lapack --download-mpich ----------------------------------------- Libraries compiled on Mon Nov 4 14:29:26 2013 on dsu-pc Machine characteristics: Linux-3.2.0-41-generic-x86_64-with-Ubuntu-12.04-precise Using PETSc directory: /home/dsu/petsc Using PETSc arch: linux-gnu-opt ----------------------------------------- Using C compiler: /home/dsu/petsc/linux-gnu-opt/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /home/dsu/petsc/linux-gnu-opt/bin/mpif90 -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument -O ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/dsu/petsc/linux-gnu-opt/include -I/home/dsu/petsc/include -I/home/dsu/petsc/include -I/home/dsu/petsc/linux-gnu-opt/include ----------------------------------------- Using C linker: /home/dsu/petsc/linux-gnu-opt/bin/mpicc Using Fortran linker: /home/dsu/petsc/linux-gnu-opt/bin/mpif90 Using libraries: -Wl,-rpath,/home/dsu/petsc/linux-gnu-opt/lib -L/home/dsu/petsc/linux-gnu-opt/lib -lpetsc -Wl,-rpath,/home/dsu/petsc/linux-gnu-opt/lib -L/home/dsu/petsc/linux-gnu-opt/lib -lflapack -lfblas -lpthread -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lmpichf90 -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl ----------------------------------------- -------------- next part -------------- ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex2f on a linux-gnu-opt named dsu-pc with 1 processor, by root Mon Nov 4 16:14:37 2013 Using Petsc Development GIT revision: 1beacf92f04482972e84431be0032cb960d262c7 GIT Date: 2013-11-01 14:59:20 -0500 Max Max/Min Avg Total Time (sec): 2.295e+02 1.00000 2.295e+02 Objects: 4.500e+01 1.00000 4.500e+01 Flops: 2.203e+11 1.00000 2.203e+11 2.203e+11 Flops/sec: 9.597e+08 1.00000 9.597e+08 9.597e+08 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 5.236e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.2953e+02 100.0% 2.2028e+11 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 5.235e+03 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 2657 1.0 4.0388e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 18 11 0 0 0 18 11 0 0 0 592 MatSolve 2657 1.0 6.1962e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 386 MatLUFactorNum 1 1.0 1.2718e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 86 MatILUFactorSym 1 1.0 9.5901e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 1 1.0 1.2159e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 6.3241e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 7.1526e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 1.2885e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecMDot 2571 1.0 4.9771e+01 1.0 7.95e+10 1.0 0.0e+00 0.0e+00 2.6e+03 22 36 0 0 49 22 36 0 0 49 1598 VecNorm 2658 1.0 5.2489e+00 1.0 5.32e+09 1.0 0.0e+00 0.0e+00 2.7e+03 2 2 0 0 51 2 2 0 0 51 1013 VecScale 2657 1.0 3.5420e+00 1.0 2.66e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 1 0 0 0 2 1 0 0 0 750 VecCopy 86 1.0 2.0908e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 88 1.0 1.1408e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 172 1.0 4.3620e-01 1.0 3.44e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 789 VecMAXPY 2657 1.0 6.6513e+01 1.0 8.47e+10 1.0 0.0e+00 0.0e+00 0.0e+00 29 38 0 0 0 29 38 0 0 0 1273 VecNormalize 2657 1.0 8.8659e+00 1.0 7.97e+09 1.0 0.0e+00 0.0e+00 2.7e+03 4 4 0 0 51 4 4 0 0 51 899 KSPGMRESOrthog 2571 1.0 1.1234e+02 1.0 1.59e+11 1.0 0.0e+00 0.0e+00 2.6e+03 49 72 0 0 49 49 72 0 0 49 1416 KSPSetUp 1 1.0 2.9065e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 2.2896e+02 1.0 2.20e+11 1.0 0.0e+00 0.0e+00 5.2e+03100100 0 0100 100100 0 0100 962 PCSetUp 1 1.0 2.3610e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 47 PCApply 2657 1.0 6.2019e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 385 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 2 2 151957404 0 Vector 37 37 296057424 0 Krylov Solver 1 1 18368 0 Preconditioner 1 1 984 0 Index Set 3 3 4002304 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 5.81741e-06 #PETSc Option Table entries: -log_summary log_1000x1000_mpi_p1.log -m 1000 -n 1000 #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Mon Nov 4 14:29:26 2013 Configure options: PETSC_ARCH=linux-gnu-opt --with-cc=gcc --with-fc=gfortran --with-debugging=0 --download-f-blas-lapack --download-mpich ----------------------------------------- Libraries compiled on Mon Nov 4 14:29:26 2013 on dsu-pc Machine characteristics: Linux-3.2.0-41-generic-x86_64-with-Ubuntu-12.04-precise Using PETSc directory: /home/dsu/petsc Using PETSc arch: linux-gnu-opt ----------------------------------------- Using C compiler: /home/dsu/petsc/linux-gnu-opt/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /home/dsu/petsc/linux-gnu-opt/bin/mpif90 -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument -O ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/dsu/petsc/linux-gnu-opt/include -I/home/dsu/petsc/include -I/home/dsu/petsc/include -I/home/dsu/petsc/linux-gnu-opt/include ----------------------------------------- Using C linker: /home/dsu/petsc/linux-gnu-opt/bin/mpicc Using Fortran linker: /home/dsu/petsc/linux-gnu-opt/bin/mpif90 Using libraries: -Wl,-rpath,/home/dsu/petsc/linux-gnu-opt/lib -L/home/dsu/petsc/linux-gnu-opt/lib -lpetsc -Wl,-rpath,/home/dsu/petsc/linux-gnu-opt/lib -L/home/dsu/petsc/linux-gnu-opt/lib -lflapack -lfblas -lpthread -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lmpichf90 -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl ----------------------------------------- From jedbrown at mcs.anl.gov Mon Nov 4 18:32:33 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Mon, 04 Nov 2013 17:32:33 -0700 Subject: [petsc-users] Speedup problem when using OpenMP? In-Reply-To: <52783B2A.4090909@gmail.com> References: <5272EDAD.2070001@gmail.com> <52783B2A.4090909@gmail.com> Message-ID: <87vc07zo0u.fsf@mcs.anl.gov> Danyang Su writes: > Hi All, > > I have test the same example under Ubuntu12.04 X64. The PETSc-dev > version is update to date (GIT Date: 2013-11-01 14:59:20 -0500) and the > installation is smooth without any error. The speedup of MPI version is > linear scalable but the speedup of OpenMP version does not change. *From > the CPU usage, the program still run in one thread when use OpenMP. * > > The commands to run the test are as follows: > > openmp > ./ex2f -threadcomm_type openmp -threadcomm_nthreads 4 -m 1000 -n 1000 > -log_summary log_ex2f_1000x1000_ubuntu1204_omp_p4.log > > mpi > mpiexec -n 4 ./ex2f -m 1000 -n 1000 -log_summary > log_ex2f_1000x1000_ubuntu1204_mpi_p4.log > > This problem is so tricky to me. Can anybody confirm if KSP solver is > parallelized for OpenMP version? > > Thanks and regards, > > Danyang > > On 31/10/2013 4:54 PM, Danyang Su wrote: >> Hi All, >> >> I have a question on the speedup of PETSc when using OpenMP. I can get >> good speedup when using MPI, but no speedup when using OpenMP. >> The example is ex2f with m=100 and n=100. The number of available >> processors is 16 (32 threads) and the OS is Windows Server 2012. The >> log files for 4 and 8 processors are attached. >> >> The commands I used to run with 4 processors are as follows: >> Run using MPI >> mpiexec -n 4 Petsc-windows-ex2f.exe -m 100 -n 100 -log_summary >> log_100x100_mpi_p4.log >> >> Run using OpenMP >> Petsc-windows-ex2f.exe -threadcomm_type openmp -threadcomm_nthreads 4 >> -m 100 -n 100 -log_summary log_100x100_openmp_p4.log >> >> The PETSc used for this test is PETSc for Windows >> http://www.mic-tc.ch/downloads/PETScForWindows.zip, but I guess this >> is not the problem because the same problem exists when I use >> PETSc-dev in Cygwin. I don't know if this problem exists in Linux, >> would anybody help to test? >> >> Thanks and regards, >> >> Danyang > > ************************************************************************************************************************ > *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** > ************************************************************************************************************************ > > ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- > > ./ex2f on a linux-gnu-omp-opt named dsu-pc with 1 processor, by root Mon Nov 4 15:35:47 2013 > With 4 threads per MPI_Comm > Using Petsc Development GIT revision: 1beacf92f04482972e84431be0032cb960d262c7 GIT Date: 2013-11-01 14:59:20 -0500 > > Max Max/Min Avg Total > Time (sec): 2.376e+02 1.00000 2.376e+02 > Objects: 4.500e+01 1.00000 4.500e+01 > Flops: 2.203e+11 1.00000 2.203e+11 2.203e+11 > Flops/sec: 9.271e+08 1.00000 9.271e+08 9.271e+08 > MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Reductions: 0.000e+00 0.00000 > > Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) > e.g., VecAXPY() for real vectors of length N --> 2N flops > and VecAXPY() for complex vectors of length N --> 8N flops > > Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- > Avg %Total Avg %Total counts %Total Avg %Total counts %Total > 0: Main Stage: 2.3759e+02 100.0% 2.2028e+11 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% > > ------------------------------------------------------------------------------------------------------------------------ > See the 'Profiling' chapter of the users' manual for details on interpreting output. > Phase summary info: > Count: number of times phase was executed > Time and Flops: Max - maximum over all processors > Ratio - ratio of maximum to minimum over all processors > Mess: number of messages sent > Avg. len: average message length (bytes) > Reduct: number of global reductions > Global: entire computation > Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). > %T - percent time in this phase %f - percent flops in this phase > %M - percent messages in this phase %L - percent message lengths in this phase > %R - percent reductions in this phase > Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) > ------------------------------------------------------------------------------------------------------------------------ > Event Count Time (sec) Flops --- Global --- --- Stage --- Total > Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s > ------------------------------------------------------------------------------------------------------------------------ > > --- Event Stage 0: Main Stage > > MatMult 2657 1.0 4.1715e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 18 11 0 0 0 18 11 0 0 0 573 > MatSolve 2657 1.0 6.4028e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 373 > MatLUFactorNum 1 1.0 1.1149e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 99 > MatILUFactorSym 1 1.0 8.2365e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyBegin 1 1.0 7.8678e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyEnd 1 1.0 9.1023e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatGetRowIJ 1 1.0 1.0014e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatGetOrdering 1 1.0 1.2122e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecMDot 2571 1.0 5.1144e+01 1.0 7.95e+10 1.0 0.0e+00 0.0e+00 0.0e+00 22 36 0 0 0 22 36 0 0 0 1555 > VecNorm 2658 1.0 5.4516e+00 1.0 5.32e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 2 0 0 0 2 2 0 0 0 975 > VecScale 2657 1.0 3.8631e+00 1.0 2.66e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 1 0 0 0 2 1 0 0 0 688 > VecCopy 86 1.0 2.2233e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecSet 88 1.0 1.1501e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecAXPY 172 1.0 4.4589e-01 1.0 3.44e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 771 > VecMAXPY 2657 1.0 6.9213e+01 1.0 8.47e+10 1.0 0.0e+00 0.0e+00 0.0e+00 29 38 0 0 0 29 38 0 0 0 1223 > VecNormalize 2657 1.0 9.3968e+00 1.0 7.97e+09 1.0 0.0e+00 0.0e+00 0.0e+00 4 4 0 0 0 4 4 0 0 0 848 > KSPGMRESOrthog 2571 1.0 1.1630e+02 1.0 1.59e+11 1.0 0.0e+00 0.0e+00 0.0e+00 49 72 0 0 0 49 72 0 0 0 1367 > KSPSetUp 1 1.0 2.8520e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > KSPSolve 1 1.0 2.3699e+02 1.0 2.20e+11 1.0 0.0e+00 0.0e+00 0.0e+00100100 0 0 0 100100 0 0 0 929 > PCSetUp 1 1.0 2.0609e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 53 > PCApply 2657 1.0 6.4088e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 373 > ------------------------------------------------------------------------------------------------------------------------ > > Memory usage is given in bytes: > > Object Type Creations Destructions Memory Descendants' Mem. > Reports information only for process 0. > > --- Event Stage 0: Main Stage > > Matrix 2 2 151957404 0 > Vector 37 37 296057424 0 > Krylov Solver 1 1 18368 0 > Preconditioner 1 1 984 0 > Index Set 3 3 4002304 0 > Viewer 1 0 0 0 > ======================================================================================================================== > Average time to get PetscTime(): 6.50883e-06 > #PETSc Option Table entries: > -log_summary log_1000x1000_omp_p4.log > -m 1000 > -n 1000 > -threadcomm_nthreads 4 > -threadcomm_type openmp > #End of PETSc Option Table entries > Compiled without FORTRAN kernels > Compiled with full precision matrices (default) > sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 > Configure run at: Mon Nov 4 15:22:12 2013 > Configure options: PETSC_ARCH=linux-gnu-omp-opt --with-cc=gcc --with-fc=gfortran --download-f-blas-lapack --with-mpi=0 --with-openmp --with-debugging=0 Add --with-threadcomm --with-pthreadclasses to the configuration. Using --with-openmp on its own doesn't turn on threadcomm. Use all three flags for now and compare -threadcomm_type openmp to -threadcomm_type pthread. http://www.mcs.anl.gov/petsc/documentation/installation.html#threads > ----------------------------------------- > Libraries compiled on Mon Nov 4 15:22:12 2013 on dsu-pc > Machine characteristics: Linux-3.2.0-55-generic-x86_64-with-Ubuntu-12.04-precise > Using PETSc directory: /home/dsu/petsc > Using PETSc arch: linux-gnu-omp-opt > ----------------------------------------- > > Using C compiler: gcc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O -fopenmp ${COPTFLAGS} ${CFLAGS} > Using Fortran compiler: gfortran -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument -O -fopenmp ${FOPTFLAGS} ${FFLAGS} > ----------------------------------------- > > Using include paths: -I/home/dsu/petsc/linux-gnu-omp-opt/include -I/home/dsu/petsc/include -I/home/dsu/petsc/include -I/home/dsu/petsc/linux-gnu-omp-opt/include -I/home/dsu/petsc/include/mpiuni > ----------------------------------------- > > Using C linker: gcc > Using Fortran linker: gfortran > Using libraries: -Wl,-rpath,/home/dsu/petsc/linux-gnu-omp-opt/lib -L/home/dsu/petsc/linux-gnu-omp-opt/lib -lpetsc -Wl,-rpath,/home/dsu/petsc/linux-gnu-omp-opt/lib -L/home/dsu/petsc/linux-gnu-omp-opt/lib -lflapack -lfblas -lpthread -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lgfortran -lm -lgfortran -lm -lquadmath -lm -lstdc++ -ldl -lgcc_s -ldl > ----------------------------------------- > > ************************************************************************************************************************ > *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** > ************************************************************************************************************************ > > ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- > > ./ex2f on a linux-gnu-omp-opt named dsu-pc with 1 processor, by root Mon Nov 4 15:31:30 2013 > Using Petsc Development GIT revision: 1beacf92f04482972e84431be0032cb960d262c7 GIT Date: 2013-11-01 14:59:20 -0500 > > Max Max/Min Avg Total > Time (sec): 2.388e+02 1.00000 2.388e+02 > Objects: 4.500e+01 1.00000 4.500e+01 > Flops: 2.203e+11 1.00000 2.203e+11 2.203e+11 > Flops/sec: 9.224e+08 1.00000 9.224e+08 9.224e+08 > MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Reductions: 0.000e+00 0.00000 > > Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) > e.g., VecAXPY() for real vectors of length N --> 2N flops > and VecAXPY() for complex vectors of length N --> 8N flops > > Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- > Avg %Total Avg %Total counts %Total Avg %Total counts %Total > 0: Main Stage: 2.3881e+02 100.0% 2.2028e+11 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% > > ------------------------------------------------------------------------------------------------------------------------ > See the 'Profiling' chapter of the users' manual for details on interpreting output. > Phase summary info: > Count: number of times phase was executed > Time and Flops: Max - maximum over all processors > Ratio - ratio of maximum to minimum over all processors > Mess: number of messages sent > Avg. len: average message length (bytes) > Reduct: number of global reductions > Global: entire computation > Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). > %T - percent time in this phase %f - percent flops in this phase > %M - percent messages in this phase %L - percent message lengths in this phase > %R - percent reductions in this phase > Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) > ------------------------------------------------------------------------------------------------------------------------ > Event Count Time (sec) Flops --- Global --- --- Stage --- Total > Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s > ------------------------------------------------------------------------------------------------------------------------ > > --- Event Stage 0: Main Stage > > MatMult 2657 1.0 4.0429e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 17 11 0 0 0 17 11 0 0 0 591 > MatSolve 2657 1.0 6.3888e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 374 > MatLUFactorNum 1 1.0 1.2874e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 85 > MatILUFactorSym 1 1.0 1.3501e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyBegin 1 1.0 8.1062e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyEnd 1 1.0 6.8491e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatGetRowIJ 1 1.0 1.1921e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatGetOrdering 1 1.0 1.3066e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecMDot 2571 1.0 5.2507e+01 1.0 7.95e+10 1.0 0.0e+00 0.0e+00 0.0e+00 22 36 0 0 0 22 36 0 0 0 1514 > VecNorm 2658 1.0 5.4426e+00 1.0 5.32e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 2 0 0 0 2 2 0 0 0 977 > VecScale 2657 1.0 3.8871e+00 1.0 2.66e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 1 0 0 0 2 1 0 0 0 684 > VecCopy 86 1.0 1.9921e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecSet 88 1.0 1.0965e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecAXPY 172 1.0 4.0171e-01 1.0 3.44e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 856 > VecMAXPY 2657 1.0 7.0096e+01 1.0 8.47e+10 1.0 0.0e+00 0.0e+00 0.0e+00 29 38 0 0 0 29 38 0 0 0 1208 > VecNormalize 2657 1.0 9.4060e+00 1.0 7.97e+09 1.0 0.0e+00 0.0e+00 0.0e+00 4 4 0 0 0 4 4 0 0 0 847 > KSPGMRESOrthog 2571 1.0 1.1847e+02 1.0 1.59e+11 1.0 0.0e+00 0.0e+00 0.0e+00 50 72 0 0 0 50 72 0 0 0 1342 > KSPSetUp 1 1.0 3.7805e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > KSPSolve 1 1.0 2.3820e+02 1.0 2.20e+11 1.0 0.0e+00 0.0e+00 0.0e+00100100 0 0 0 100100 0 0 0 925 > PCSetUp 1 1.0 2.7698e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 40 > PCApply 2657 1.0 6.3946e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 374 > ------------------------------------------------------------------------------------------------------------------------ > > Memory usage is given in bytes: > > Object Type Creations Destructions Memory Descendants' Mem. > Reports information only for process 0. > > --- Event Stage 0: Main Stage > > Matrix 2 2 151957404 0 > Vector 37 37 296057424 0 > Krylov Solver 1 1 18368 0 > Preconditioner 1 1 984 0 > Index Set 3 3 4002304 0 > Viewer 1 0 0 0 > ======================================================================================================================== > Average time to get PetscTime(): 8.51154e-06 > #PETSc Option Table entries: > -log_summary log_1000x1000_omp_p1.log > -m 1000 > -n 1000 > -threadcomm_nthreads 1 > -threadcomm_type openmp > #End of PETSc Option Table entries > Compiled without FORTRAN kernels > Compiled with full precision matrices (default) > sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 > Configure run at: Mon Nov 4 15:22:12 2013 > Configure options: PETSC_ARCH=linux-gnu-omp-opt --with-cc=gcc --with-fc=gfortran --download-f-blas-lapack --with-mpi=0 --with-openmp --with-debugging=0 > ----------------------------------------- > Libraries compiled on Mon Nov 4 15:22:12 2013 on dsu-pc > Machine characteristics: Linux-3.2.0-55-generic-x86_64-with-Ubuntu-12.04-precise > Using PETSc directory: /home/dsu/petsc > Using PETSc arch: linux-gnu-omp-opt > ----------------------------------------- > > Using C compiler: gcc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O -fopenmp ${COPTFLAGS} ${CFLAGS} > Using Fortran compiler: gfortran -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument -O -fopenmp ${FOPTFLAGS} ${FFLAGS} > ----------------------------------------- > > Using include paths: -I/home/dsu/petsc/linux-gnu-omp-opt/include -I/home/dsu/petsc/include -I/home/dsu/petsc/include -I/home/dsu/petsc/linux-gnu-omp-opt/include -I/home/dsu/petsc/include/mpiuni > ----------------------------------------- > > Using C linker: gcc > Using Fortran linker: gfortran > Using libraries: -Wl,-rpath,/home/dsu/petsc/linux-gnu-omp-opt/lib -L/home/dsu/petsc/linux-gnu-omp-opt/lib -lpetsc -Wl,-rpath,/home/dsu/petsc/linux-gnu-omp-opt/lib -L/home/dsu/petsc/linux-gnu-omp-opt/lib -lflapack -lfblas -lpthread -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lgfortran -lm -lgfortran -lm -lquadmath -lm -lstdc++ -ldl -lgcc_s -ldl > ----------------------------------------- > > ************************************************************************************************************************ > *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** > ************************************************************************************************************************ > > ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- > > ./ex2f on a linux-gnu-opt named dsu-pc with 4 processors, by root Mon Nov 4 16:10:24 2013 > Using Petsc Development GIT revision: 1beacf92f04482972e84431be0032cb960d262c7 GIT Date: 2013-11-01 14:59:20 -0500 > > Max Max/Min Avg Total > Time (sec): 5.364e+01 1.00045 5.362e+01 > Objects: 5.600e+01 1.00000 5.600e+01 > Flops: 2.837e+10 1.00010 2.837e+10 1.135e+11 > Flops/sec: 5.291e+08 1.00054 5.290e+08 2.116e+09 > MPI Messages: 2.744e+03 2.00000 2.058e+03 8.232e+03 > MPI Message Lengths: 2.193e+07 2.00000 7.991e+03 6.578e+07 > MPI Reductions: 2.720e+03 1.00000 > > Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) > e.g., VecAXPY() for real vectors of length N --> 2N flops > and VecAXPY() for complex vectors of length N --> 8N flops > > Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- > Avg %Total Avg %Total counts %Total Avg %Total counts %Total > 0: Main Stage: 5.3623e+01 100.0% 1.1347e+11 100.0% 8.232e+03 100.0% 7.991e+03 100.0% 2.719e+03 100.0% > > ------------------------------------------------------------------------------------------------------------------------ > See the 'Profiling' chapter of the users' manual for details on interpreting output. > Phase summary info: > Count: number of times phase was executed > Time and Flops: Max - maximum over all processors > Ratio - ratio of maximum to minimum over all processors > Mess: number of messages sent > Avg. len: average message length (bytes) > Reduct: number of global reductions > Global: entire computation > Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). > %T - percent time in this phase %f - percent flops in this phase > %M - percent messages in this phase %L - percent message lengths in this phase > %R - percent reductions in this phase > Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) > ------------------------------------------------------------------------------------------------------------------------ > Event Count Time (sec) Flops --- Global --- --- Stage --- Total > Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s > ------------------------------------------------------------------------------------------------------------------------ > > --- Event Stage 0: Main Stage > > MatMult 1370 1.0 8.7882e+00 1.0 3.08e+09 1.0 8.2e+03 8.0e+03 0.0e+00 16 11100100 0 16 11100100 0 1402 > MatSolve 1370 1.0 9.0304e+00 1.0 3.08e+09 1.0 0.0e+00 0.0e+00 0.0e+00 17 11 0 0 0 17 11 0 0 0 1362 > MatLUFactorNum 1 1.0 3.3336e-02 1.0 2.74e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 329 > MatILUFactorSym 1 1.0 7.1875e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyBegin 1 1.0 7.2212e-0241.6 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyEnd 1 1.0 5.4802e-02 1.0 0.00e+00 0.0 1.2e+01 2.0e+03 9.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatGetRowIJ 1 1.0 1.2875e-05 2.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatGetOrdering 1 1.0 4.8881e-03 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecMDot 1325 1.0 1.4754e+01 1.0 1.02e+10 1.0 0.0e+00 0.0e+00 1.3e+03 27 36 0 0 49 27 36 0 0 49 2776 > VecNorm 1371 1.0 1.9989e+00 1.1 6.86e+08 1.0 0.0e+00 0.0e+00 1.4e+03 4 2 0 0 50 4 2 0 0 50 1372 > VecScale 1370 1.0 4.9844e-01 1.1 3.42e+08 1.0 0.0e+00 0.0e+00 0.0e+00 1 1 0 0 0 1 1 0 0 0 2749 > VecCopy 45 1.0 4.4863e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecSet 1418 1.0 6.2273e-01 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 > VecAXPY 90 1.0 1.0165e-01 1.0 4.50e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1771 > VecMAXPY 1370 1.0 1.5635e+01 1.0 1.09e+10 1.0 0.0e+00 0.0e+00 0.0e+00 29 38 0 0 0 29 38 0 0 0 2789 > VecScatterBegin 1370 1.0 1.6159e-01 1.8 0.00e+00 0.0 8.2e+03 8.0e+03 0.0e+00 0 0100100 0 0 0100100 0 0 > VecScatterEnd 1370 1.0 9.6929e-01 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 2 0 0 0 0 2 0 0 0 0 0 > VecNormalize 1370 1.0 2.5033e+00 1.1 1.03e+09 1.0 0.0e+00 0.0e+00 1.4e+03 5 4 0 0 50 5 4 0 0 50 1642 > KSPGMRESOrthog 1325 1.0 2.9419e+01 1.0 2.05e+10 1.0 0.0e+00 0.0e+00 1.3e+03 54 72 0 0 49 54 72 0 0 49 2784 > KSPSetUp 2 1.0 2.1291e-02 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > KSPSolve 1 1.0 5.2989e+01 1.0 2.84e+10 1.0 8.2e+03 8.0e+03 2.7e+03 99100100100 99 99100100100 99 2141 > PCSetUp 2 1.0 1.4600e-01 1.1 2.74e+06 1.0 0.0e+00 0.0e+00 5.0e+00 0 0 0 0 0 0 0 0 0 0 75 > PCSetUpOnBlocks 1 1.0 1.1017e-01 1.0 2.74e+06 1.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 100 > PCApply 1370 1.0 9.7092e+00 1.0 3.08e+09 1.0 0.0e+00 0.0e+00 0.0e+00 18 11 0 0 0 18 11 0 0 0 1267 > ------------------------------------------------------------------------------------------------------------------------ > > Memory usage is given in bytes: > > Object Type Creations Destructions Memory Descendants' Mem. > Reports information only for process 0. > > --- Event Stage 0: Main Stage > > Matrix 4 4 56984588 0 > Vector 41 41 74071440 0 > Vector Scatter 1 1 1060 0 > Index Set 5 5 1007832 0 > Krylov Solver 2 2 19520 0 > Preconditioner 2 2 1864 0 > Viewer 1 0 0 0 > ======================================================================================================================== > Average time to get PetscTime(): 6.19888e-06 > Average time for MPI_Barrier(): 0.000529623 > Average time for zero size MPI_Send(): 0.000117242 > #PETSc Option Table entries: > -log_summary log_1000x1000_mpi_p4.log > -m 1000 > -n 1000 > #End of PETSc Option Table entries > Compiled without FORTRAN kernels > Compiled with full precision matrices (default) > sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 > Configure run at: Mon Nov 4 14:29:26 2013 > Configure options: PETSC_ARCH=linux-gnu-opt --with-cc=gcc --with-fc=gfortran --with-debugging=0 --download-f-blas-lapack --download-mpich > ----------------------------------------- > Libraries compiled on Mon Nov 4 14:29:26 2013 on dsu-pc > Machine characteristics: Linux-3.2.0-41-generic-x86_64-with-Ubuntu-12.04-precise > Using PETSc directory: /home/dsu/petsc > Using PETSc arch: linux-gnu-opt > ----------------------------------------- > > Using C compiler: /home/dsu/petsc/linux-gnu-opt/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} > Using Fortran compiler: /home/dsu/petsc/linux-gnu-opt/bin/mpif90 -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument -O ${FOPTFLAGS} ${FFLAGS} > ----------------------------------------- > > Using include paths: -I/home/dsu/petsc/linux-gnu-opt/include -I/home/dsu/petsc/include -I/home/dsu/petsc/include -I/home/dsu/petsc/linux-gnu-opt/include > ----------------------------------------- > > Using C linker: /home/dsu/petsc/linux-gnu-opt/bin/mpicc > Using Fortran linker: /home/dsu/petsc/linux-gnu-opt/bin/mpif90 > Using libraries: -Wl,-rpath,/home/dsu/petsc/linux-gnu-opt/lib -L/home/dsu/petsc/linux-gnu-opt/lib -lpetsc -Wl,-rpath,/home/dsu/petsc/linux-gnu-opt/lib -L/home/dsu/petsc/linux-gnu-opt/lib -lflapack -lfblas -lpthread -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lmpichf90 -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl > ----------------------------------------- > > ************************************************************************************************************************ > *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** > ************************************************************************************************************************ > > ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- > > ./ex2f on a linux-gnu-opt named dsu-pc with 1 processor, by root Mon Nov 4 16:14:37 2013 > Using Petsc Development GIT revision: 1beacf92f04482972e84431be0032cb960d262c7 GIT Date: 2013-11-01 14:59:20 -0500 > > Max Max/Min Avg Total > Time (sec): 2.295e+02 1.00000 2.295e+02 > Objects: 4.500e+01 1.00000 4.500e+01 > Flops: 2.203e+11 1.00000 2.203e+11 2.203e+11 > Flops/sec: 9.597e+08 1.00000 9.597e+08 9.597e+08 > MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Reductions: 5.236e+03 1.00000 > > Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) > e.g., VecAXPY() for real vectors of length N --> 2N flops > and VecAXPY() for complex vectors of length N --> 8N flops > > Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- > Avg %Total Avg %Total counts %Total Avg %Total counts %Total > 0: Main Stage: 2.2953e+02 100.0% 2.2028e+11 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 5.235e+03 100.0% > > ------------------------------------------------------------------------------------------------------------------------ > See the 'Profiling' chapter of the users' manual for details on interpreting output. > Phase summary info: > Count: number of times phase was executed > Time and Flops: Max - maximum over all processors > Ratio - ratio of maximum to minimum over all processors > Mess: number of messages sent > Avg. len: average message length (bytes) > Reduct: number of global reductions > Global: entire computation > Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). > %T - percent time in this phase %f - percent flops in this phase > %M - percent messages in this phase %L - percent message lengths in this phase > %R - percent reductions in this phase > Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) > ------------------------------------------------------------------------------------------------------------------------ > Event Count Time (sec) Flops --- Global --- --- Stage --- Total > Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s > ------------------------------------------------------------------------------------------------------------------------ > > --- Event Stage 0: Main Stage > > MatMult 2657 1.0 4.0388e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 18 11 0 0 0 18 11 0 0 0 592 > MatSolve 2657 1.0 6.1962e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 386 > MatLUFactorNum 1 1.0 1.2718e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 86 > MatILUFactorSym 1 1.0 9.5901e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyBegin 1 1.0 1.2159e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyEnd 1 1.0 6.3241e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatGetRowIJ 1 1.0 7.1526e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatGetOrdering 1 1.0 1.2885e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecMDot 2571 1.0 4.9771e+01 1.0 7.95e+10 1.0 0.0e+00 0.0e+00 2.6e+03 22 36 0 0 49 22 36 0 0 49 1598 > VecNorm 2658 1.0 5.2489e+00 1.0 5.32e+09 1.0 0.0e+00 0.0e+00 2.7e+03 2 2 0 0 51 2 2 0 0 51 1013 > VecScale 2657 1.0 3.5420e+00 1.0 2.66e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 1 0 0 0 2 1 0 0 0 750 > VecCopy 86 1.0 2.0908e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecSet 88 1.0 1.1408e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecAXPY 172 1.0 4.3620e-01 1.0 3.44e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 789 > VecMAXPY 2657 1.0 6.6513e+01 1.0 8.47e+10 1.0 0.0e+00 0.0e+00 0.0e+00 29 38 0 0 0 29 38 0 0 0 1273 > VecNormalize 2657 1.0 8.8659e+00 1.0 7.97e+09 1.0 0.0e+00 0.0e+00 2.7e+03 4 4 0 0 51 4 4 0 0 51 899 > KSPGMRESOrthog 2571 1.0 1.1234e+02 1.0 1.59e+11 1.0 0.0e+00 0.0e+00 2.6e+03 49 72 0 0 49 49 72 0 0 49 1416 > KSPSetUp 1 1.0 2.9065e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > KSPSolve 1 1.0 2.2896e+02 1.0 2.20e+11 1.0 0.0e+00 0.0e+00 5.2e+03100100 0 0100 100100 0 0100 962 > PCSetUp 1 1.0 2.3610e-01 1.0 1.10e+07 1.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 47 > PCApply 2657 1.0 6.2019e+01 1.0 2.39e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 11 0 0 0 27 11 0 0 0 385 > ------------------------------------------------------------------------------------------------------------------------ > > Memory usage is given in bytes: > > Object Type Creations Destructions Memory Descendants' Mem. > Reports information only for process 0. > > --- Event Stage 0: Main Stage > > Matrix 2 2 151957404 0 > Vector 37 37 296057424 0 > Krylov Solver 1 1 18368 0 > Preconditioner 1 1 984 0 > Index Set 3 3 4002304 0 > Viewer 1 0 0 0 > ======================================================================================================================== > Average time to get PetscTime(): 5.81741e-06 > #PETSc Option Table entries: > -log_summary log_1000x1000_mpi_p1.log > -m 1000 > -n 1000 > #End of PETSc Option Table entries > Compiled without FORTRAN kernels > Compiled with full precision matrices (default) > sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 > Configure run at: Mon Nov 4 14:29:26 2013 > Configure options: PETSC_ARCH=linux-gnu-opt --with-cc=gcc --with-fc=gfortran --with-debugging=0 --download-f-blas-lapack --download-mpich > ----------------------------------------- > Libraries compiled on Mon Nov 4 14:29:26 2013 on dsu-pc > Machine characteristics: Linux-3.2.0-41-generic-x86_64-with-Ubuntu-12.04-precise > Using PETSc directory: /home/dsu/petsc > Using PETSc arch: linux-gnu-opt > ----------------------------------------- > > Using C compiler: /home/dsu/petsc/linux-gnu-opt/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} > Using Fortran compiler: /home/dsu/petsc/linux-gnu-opt/bin/mpif90 -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument -O ${FOPTFLAGS} ${FFLAGS} > ----------------------------------------- > > Using include paths: -I/home/dsu/petsc/linux-gnu-opt/include -I/home/dsu/petsc/include -I/home/dsu/petsc/include -I/home/dsu/petsc/linux-gnu-opt/include > ----------------------------------------- > > Using C linker: /home/dsu/petsc/linux-gnu-opt/bin/mpicc > Using Fortran linker: /home/dsu/petsc/linux-gnu-opt/bin/mpif90 > Using libraries: -Wl,-rpath,/home/dsu/petsc/linux-gnu-opt/lib -L/home/dsu/petsc/linux-gnu-opt/lib -lpetsc -Wl,-rpath,/home/dsu/petsc/linux-gnu-opt/lib -L/home/dsu/petsc/linux-gnu-opt/lib -lflapack -lfblas -lpthread -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lmpichf90 -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl > ----------------------------------------- -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From mpovolot at purdue.edu Mon Nov 4 21:37:02 2013 From: mpovolot at purdue.edu (Michael Povolotskyi) Date: Mon, 04 Nov 2013 22:37:02 -0500 Subject: [petsc-users] calling lapack for LU decomposition In-Reply-To: <878ux329a1.fsf@mcs.anl.gov> References: <52780336.4070803@purdue.edu> <878ux329a1.fsf@mcs.anl.gov> Message-ID: <527867DE.3030703@purdue.edu> On 11/4/2013 3:37 PM, Jed Brown wrote: > Michael Povolotskyi writes: > >> Dear Petsc developers, >> in my work I need to measure the multi threading performance of LAPACK >> function dgetrf >> >> Since all matrices are of PETSc type I'd like to call Petsc functions >> for the LU decomposition. >> My code reads like this >> >> Mat A; //serial dense matrix >> >> PetscErrorCode ierr; >> KSP ksp; >> Mat F; >> PC pc; >> >> ierr = KSPCreate(comm, &ksp); >> ierr = KSPSetOperators(ksp, A, A, SAME_PRECONDITIONER); >> ierr = KSPSetType(ksp, KSPPREONLY); >> ierr = KSPGetPC(ksp, &pc); >> ierr = PCSetType(pc, PCLU); >> ierr = PCFactorSetMatSolverPackage(pc, "petsc"); > This is the default. > >> ierr = KSPSetUp(ksp); >> ierr = PCFactorGetMatrix(pc, &F); > This function is not needed. > >> Is this a correct way to call dgetrf ? > Yeah, run with -log_summary and look for the time spent in the > MatLUFactorNumeric event. > > (You _can_ call the low-level MatLUFactor() directly, but I recommend > what you have written because it is more general-purpose so you can > easily switch to other methods.) Thank you, Michael. From mpovolot at purdue.edu Mon Nov 4 21:43:10 2013 From: mpovolot at purdue.edu (Michael Povolotskyi) Date: Mon, 04 Nov 2013 22:43:10 -0500 Subject: [petsc-users] question about alternative to MatMatSolve Message-ID: <5278694E.10708@purdue.edu> Dear Petsc developers, I'm using the MatMatSolve function to solve a linear system with a right hand side containing many columns. In my case each right hand side column is an almost empty vector with a couple of non zero entries. Is there any way to use the fact that the right hand side is very sparse and make the calculations faster? Thank you, Michael. From jedbrown at mcs.anl.gov Mon Nov 4 21:49:50 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Mon, 04 Nov 2013 20:49:50 -0700 Subject: [petsc-users] question about alternative to MatMatSolve In-Reply-To: <5278694E.10708@purdue.edu> References: <5278694E.10708@purdue.edu> Message-ID: <87ppqfzew1.fsf@mcs.anl.gov> Michael Povolotskyi writes: > In my case each right hand side column is an almost empty vector with a > couple of non zero entries. > Is there any way to use the fact that the right hand side is very sparse > and make the calculations faster? Optimizing for this case would only give a small benefit because the vector becomes dense during the forward-solve and the entire back-solve has the dense vector. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From mpovolot at purdue.edu Mon Nov 4 22:02:20 2013 From: mpovolot at purdue.edu (Michael Povolotskyi) Date: Mon, 04 Nov 2013 23:02:20 -0500 Subject: [petsc-users] question about alternative to MatMatSolve In-Reply-To: <87ppqfzew1.fsf@mcs.anl.gov> References: <5278694E.10708@purdue.edu> <87ppqfzew1.fsf@mcs.anl.gov> Message-ID: <52786DCC.1090601@purdue.edu> On 11/4/2013 10:49 PM, Jed Brown wrote: > Michael Povolotskyi writes: >> In my case each right hand side column is an almost empty vector with a >> couple of non zero entries. >> Is there any way to use the fact that the right hand side is very sparse >> and make the calculations faster? > Optimizing for this case would only give a small benefit because the > vector becomes dense during the forward-solve and the entire back-solve > has the dense vector. Thank you. Michael From rupp at mcs.anl.gov Tue Nov 5 03:22:01 2013 From: rupp at mcs.anl.gov (Karl Rupp) Date: Tue, 05 Nov 2013 10:22:01 +0100 Subject: [petsc-users] [petsc-maint] Speedup problem when using OpenMP? In-Reply-To: <5277F8B5.9010204@gmail.com> References: <5272EDAD.2070001@gmail.com> <5277B468.4090509@mcs.anl.gov> <5277F8B5.9010204@gmail.com> Message-ID: <5278B8B9.4070703@mcs.anl.gov> Hi Danyang, > This does not make any difference. I have scaled up the matrix but the > performance does not change. If I run with OpenMP, the iteration number > is always the same whatever how many processors are used. This seems > quite strange as the iteration number usually increase as the number of > processors increased when run with MPI. I think I should move to the > ubuntu system to make further test, to see if this is a windows problem. OpenMP and MPI are two different parallelization approaches: - With MPI, we split up the system matrix into different strips, where each of the strips is assigned to one MPI process. This then leads (among others) to block-Jacobi preconditioner techniques, where you usually see an increase in iteration counts. In the ex2 case, however, this even leads to a reduction of iteration counts. - With OpenMP, the system matrix is contiguous in memory, so one still computes preconditioners for the full matrix (as is for example the case with ILU). Thus, the use of OpenMP is transparent with respect to the algorithms employed, so you don't see any change in iteration counts. The typical vector operations like VecScale() (should) make use of OpenMP, but apparently this is not the case. I'm double-checking on my machine (Linux Mint Maya, based on Ubuntu 12.04 LTS) and let you know. Best regards, Karli > On 04/11/2013 6:51 AM, Karl Rupp wrote: >> Hi, >> >> > I have a question on the speedup of PETSc when using OpenMP. I can get >>> good speedup when using MPI, but no speedup when using OpenMP. >>> The example is ex2f with m=100 and n=100. The number of available >>> processors is 16 (32 threads) and the OS is Windows Server 2012. The log >>> files for 4 and 8 processors are attached. >>> >>> The commands I used to run with 4 processors are as follows: >>> Run using MPI >>> mpiexec -n 4 Petsc-windows-ex2f.exe -m 100 -n 100 -log_summary >>> log_100x100_mpi_p4.log >>> >>> Run using OpenMP >>> Petsc-windows-ex2f.exe -threadcomm_type openmp -threadcomm_nthreads 4 -m >>> 100 -n 100 -log_summary log_100x100_openmp_p4.log >>> >>> The PETSc used for this test is PETSc for Windows >>> http://www.mic-tc.ch/downloads/PETScForWindows.zip, but I guess this is >>> not the problem because the same problem exists when I use PETSc-dev in >>> Cygwin. I don't know if this problem exists in Linux, would anybody help >>> to test? >> >> For the 100x100 case considered, the execution times per call are >> somewhere in the millisecond to sub-millisecond range (e.g. 1.3ms for >> 68 calls to VecScale with 4 processors). I'd say this is too small in >> order to see any reasonable performance gain when running multiple >> threads, consider problem sizes of about 1000x1000 instead. >> >> Moreover, keep in mind that typically you won't get a perfectly linear >> scaling with the number of processor cores, because ultimately the >> memory bandwidth is the limiting factor for standard vector operations. >> >> Best regards, >> Karli >> > From rupp at mcs.anl.gov Tue Nov 5 03:59:41 2013 From: rupp at mcs.anl.gov (Karl Rupp) Date: Tue, 05 Nov 2013 10:59:41 +0100 Subject: [petsc-users] [petsc-maint] Speedup problem when using OpenMP? In-Reply-To: <52783B2A.4090909@gmail.com> References: <5272EDAD.2070001@gmail.com> <52783B2A.4090909@gmail.com> Message-ID: <5278C18D.5060107@mcs.anl.gov> Hi again, as Jed noted, adding --with-threadcomm to configure finally gives you the OpenMP parallelization (--with-pthreadclasses also gives you pthread parallelization, but this may not be available on your Windows machine). With this I get the following execution times on my laptop for e.g. VecScale() for n=m=500: Single-threaded: 9.3448e-01 sec OpenMP: 4.8728e-01 sec PThreads: 4.3991e-01 sec This factor of two is fairly common for VecScale() and indicates a saturation of the memory bus. Other operations may saturate later. Best regards, Karli On 11/05/2013 01:26 AM, Danyang Su wrote: > Hi All, > > I have test the same example under Ubuntu12.04 X64. The PETSc-dev > version is update to date (GIT Date: 2013-11-01 14:59:20 -0500) and the > installation is smooth without any error. The speedup of MPI version is > linear scalable but the speedup of OpenMP version does not change. *From > the CPU usage, the program still run in one thread when use OpenMP. * > > The commands to run the test are as follows: > > openmp > ./ex2f -threadcomm_type openmp -threadcomm_nthreads 4 -m 1000 -n 1000 > -log_summary log_ex2f_1000x1000_ubuntu1204_omp_p4.log > > mpi > mpiexec -n 4 ./ex2f -m 1000 -n 1000 -log_summary > log_ex2f_1000x1000_ubuntu1204_mpi_p4.log > > This problem is so tricky to me. Can anybody confirm if KSP solver is > parallelized for OpenMP version? > > Thanks and regards, > > Danyang > > On 31/10/2013 4:54 PM, Danyang Su wrote: >> Hi All, >> >> I have a question on the speedup of PETSc when using OpenMP. I can get >> good speedup when using MPI, but no speedup when using OpenMP. >> The example is ex2f with m=100 and n=100. The number of available >> processors is 16 (32 threads) and the OS is Windows Server 2012. The >> log files for 4 and 8 processors are attached. >> >> The commands I used to run with 4 processors are as follows: >> Run using MPI >> mpiexec -n 4 Petsc-windows-ex2f.exe -m 100 -n 100 -log_summary >> log_100x100_mpi_p4.log >> >> Run using OpenMP >> Petsc-windows-ex2f.exe -threadcomm_type openmp -threadcomm_nthreads 4 >> -m 100 -n 100 -log_summary log_100x100_openmp_p4.log >> >> The PETSc used for this test is PETSc for Windows >> http://www.mic-tc.ch/downloads/PETScForWindows.zip, but I guess this >> is not the problem because the same problem exists when I use >> PETSc-dev in Cygwin. I don't know if this problem exists in Linux, >> would anybody help to test? >> >> Thanks and regards, >> >> Danyang > From lu_qin_2000 at yahoo.com Tue Nov 5 09:41:47 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Tue, 5 Nov 2013 07:41:47 -0800 (PST) Subject: [petsc-users] Directory name with space used by hypre configure Message-ID: <1383666107.29402.YahooMailNeo@web160204.mail.bf1.yahoo.com> Hello, I am trying to build hypre in Windows 7 (using cygwin) since it will be included in my PETSc lib. The configure options of hypre contains directory such as: ./configure --with-MPI-include="/cygdrive/c/Program Files/mpich2x64/include" Somehow hypre's configure took this directory as two separate directories ("/cygdrive/c/Program" and "Files/mpich2x64/include") since there is a space in the middle. (Strangely, PETSc's configure read this directory correctly). I tried to replacing the space with "\ " but it still did not work. How can I make hypre's configure read this directory correctly? Thanks for your suggestions, Qin -------------- next part -------------- An HTML attachment was scrubbed... URL: From hsahasra at purdue.edu Tue Nov 5 09:52:23 2013 From: hsahasra at purdue.edu (Harshad Sahasrabudhe) Date: Tue, 5 Nov 2013 10:52:23 -0500 (EST) Subject: [petsc-users] MatConvert Message-ID: <544930282.313771.1383666743415.JavaMail.root@mailhub027.itcs.purdue.edu> Hi, Can we use the original Mat in Matconvert to store the converted matrix? E.g. MatConvert(A, MATDENSE, MAT_INITIAL_MATRIX, &A); Or do I need MAT_REUSE_MATRIX flag instead of MAT_INITIAL_MATRIX? A is initially MATAIJ. Thanks, Harshad From jedbrown at mcs.anl.gov Tue Nov 5 10:00:36 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Tue, 05 Nov 2013 09:00:36 -0700 Subject: [petsc-users] MatConvert In-Reply-To: <544930282.313771.1383666743415.JavaMail.root@mailhub027.itcs.purdue.edu> References: <544930282.313771.1383666743415.JavaMail.root@mailhub027.itcs.purdue.edu> Message-ID: <87d2mezvmj.fsf@mcs.anl.gov> Harshad Sahasrabudhe writes: > Hi, > > Can we use the original Mat in Matconvert to store the converted matrix? > > E.g. MatConvert(A, MATDENSE, MAT_INITIAL_MATRIX, &A); > Or do I need MAT_REUSE_MATRIX flag instead of MAT_INITIAL_MATRIX? Do this: MatConvert(A, MATDENSE, MAT_REUSE_MATRIX, &A); Note that conversion is not really "in place", it creates a new matrix and destroys the old one (all except the header). This is potentially dangerous if other components of your code hold references to the old matrix and do not expect its type to change. It does not use more memory to create a new matrix and eventually destroy the old one. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From bsmith at mcs.anl.gov Tue Nov 5 10:03:40 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Tue, 5 Nov 2013 10:03:40 -0600 Subject: [petsc-users] Directory name with space used by hypre configure In-Reply-To: <1383666107.29402.YahooMailNeo@web160204.mail.bf1.yahoo.com> References: <1383666107.29402.YahooMailNeo@web160204.mail.bf1.yahoo.com> Message-ID: <6BC19C6F-1B33-4217-BD3D-9F21DD41A35B@mcs.anl.gov> Reinstall the MPI in a location without spaces in the directory name. On Nov 5, 2013, at 9:41 AM, Qin Lu wrote: > Hello, > I am trying to build hypre in Windows 7 (using cygwin) since it will be included in my PETSc lib. The configure options of hypre contains directory such as: > ./configure --with-MPI-include="/cygdrive/c/Program Files/mpich2x64/include" > Somehow hypre's configure took this directory as two separate directories ("/cygdrive/c/Program" and "Files/mpich2x64/include") since there is a space in the middle. (Strangely, PETSc's configure read this directory correctly). I tried to replacing the space with "\ " but it still did not work. > How can I make hypre's configure read this directory correctly? > Thanks for your suggestions, > Qin From balay at mcs.anl.gov Tue Nov 5 10:05:52 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Tue, 5 Nov 2013 10:05:52 -0600 (CST) Subject: [petsc-users] Directory name with space used by hypre configure In-Reply-To: <6BC19C6F-1B33-4217-BD3D-9F21DD41A35B@mcs.anl.gov> References: <1383666107.29402.YahooMailNeo@web160204.mail.bf1.yahoo.com> <6BC19C6F-1B33-4217-BD3D-9F21DD41A35B@mcs.anl.gov> Message-ID: Alctually you'll have to look at hypre docs for installation instructions for windows. [download-hypre does not work] Satish On Tue, 5 Nov 2013, Barry Smith wrote: > > Reinstall the MPI in a location without spaces in the directory name. > > On Nov 5, 2013, at 9:41 AM, Qin Lu wrote: > > > Hello, > > I am trying to build hypre in Windows 7 (using cygwin) since it will be included in my PETSc lib. The configure options of hypre contains directory such as: > > ./configure --with-MPI-include="/cygdrive/c/Program Files/mpich2x64/include" > > Somehow hypre's configure took this directory as two separate directories ("/cygdrive/c/Program" and "Files/mpich2x64/include") since there is a space in the middle. (Strangely, PETSc's configure read this directory correctly). I tried to replacing the space with "\ " but it still did not work. > > How can I make hypre's configure read this directory correctly? > > Thanks for your suggestions, > > Qin > > From gaetank at gmail.com Tue Nov 5 10:29:26 2013 From: gaetank at gmail.com (Gaetan Kenway) Date: Tue, 5 Nov 2013 11:29:26 -0500 Subject: [petsc-users] Parallel Graph Coloring Message-ID: Hi I have a quick question regarding the MatGetColoring() function. According to the documentation, "For parallel matrices currently converts to sequential matrix and uses the sequential coloring on that." I am wondering does this give actually give a valid parallel coloring? My application is a cell centered multi-block finite volume code, (with block-based decomposition) and the communications are done using a two level exchange of halo cells. I would like to use FD (actually forward mode AD to compute the jacbian matrix). So, what I'm wondering is, if a cell with color 0 is perturbed on processor 0, is it guaranteed that a residual on processor 1, that is influenced by the 0-colored cell on proc zero is *only* influenced by the color 0 from proc 0 and not a 0-colored cell on proc 1? I hope that is clear Thank you, Gaetan Kenway -------------- next part -------------- An HTML attachment was scrubbed... URL: From brune at mcs.anl.gov Tue Nov 5 10:39:29 2013 From: brune at mcs.anl.gov (Peter Brune) Date: Tue, 5 Nov 2013 10:39:29 -0600 Subject: [petsc-users] Parallel Graph Coloring In-Reply-To: References: Message-ID: On Tue, Nov 5, 2013 at 10:29 AM, Gaetan Kenway wrote: > Hi > > I have a quick question regarding the MatGetColoring() function. According > to the documentation, > > "For parallel matrices currently converts to sequential matrix and uses > the sequential coloring on that." > > The meaning of this is that it transfers the parallel matrix to a single processor and does the coloring there, not that it only considers on-processor entries. What comes out is therefore a valid column coloring of the whole matrix. Parallel matrix coloring is in development but is not quite ready for public consumption. Anecdotal evidence suggests that the serial coloring does not become a bottleneck on small parallel runs. A good option is to provide the coloring through your discretization. This will often be more efficient than the matrix colorings, as you know the structure of your problem. - Peter > I am wondering does this give actually give a valid parallel coloring? My > application is a cell centered multi-block finite volume code, (with > block-based decomposition) and the communications are done using a two > level exchange of halo cells. I would like to use FD (actually forward mode > AD to compute the jacbian matrix). So, what I'm wondering is, if a cell > with color 0 is perturbed on processor 0, is it guaranteed that a residual > on processor 1, that is influenced by the 0-colored cell on proc zero is > *only* influenced by the color 0 from proc 0 and not a 0-colored cell on > proc 1? > > I hope that is clear > > Thank you, > > Gaetan Kenway > > > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From gaetank at gmail.com Tue Nov 5 11:07:32 2013 From: gaetank at gmail.com (Gaetan Kenway) Date: Tue, 5 Nov 2013 12:07:32 -0500 Subject: [petsc-users] Parallel Graph Coloring In-Reply-To: References: Message-ID: Hi Peter Thanks for the info. I actually already have a coloring that is done through the discretrization that is near optimal. The issue is that I have a funny interprocessor dependence that is tricky to do my own sequential coloring. So I was looking at doing it in a global, parallel sense. For interest, I tried using the matGetColor() on my assembled matrix just to see how many colors it would take. My discretrization coloring has 35 colors (for a stencil with 33 cells) and the matGetColor() resulted in approximately 78 colors or about twice as many. Thanks, Gaetan On Tue, Nov 5, 2013 at 11:39 AM, Peter Brune wrote: > > > > On Tue, Nov 5, 2013 at 10:29 AM, Gaetan Kenway wrote: > >> Hi >> >> I have a quick question regarding the MatGetColoring() function. >> According to the documentation, >> >> "For parallel matrices currently converts to sequential matrix and uses >> the sequential coloring on that." >> >> > The meaning of this is that it transfers the parallel matrix to a single > processor and does the coloring there, not that it only considers > on-processor entries. What comes out is therefore a valid column coloring > of the whole matrix. Parallel matrix coloring is in development but is not > quite ready for public consumption. Anecdotal evidence suggests that the > serial coloring does not become a bottleneck on small parallel runs. > > A good option is to provide the coloring through your discretization. > This will often be more efficient than the matrix colorings, as you know > the structure of your problem. > > - Peter > > >> I am wondering does this give actually give a valid parallel coloring? My >> application is a cell centered multi-block finite volume code, (with >> block-based decomposition) and the communications are done using a two >> level exchange of halo cells. I would like to use FD (actually forward mode >> AD to compute the jacbian matrix). So, what I'm wondering is, if a cell >> with color 0 is perturbed on processor 0, is it guaranteed that a residual >> on processor 1, that is influenced by the 0-colored cell on proc zero is >> *only* influenced by the color 0 from proc 0 and not a 0-colored cell on >> proc 1? >> >> I hope that is clear >> >> Thank you, >> >> Gaetan Kenway >> >> >> >> >> >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From prbrune at gmail.com Tue Nov 5 15:08:48 2013 From: prbrune at gmail.com (Peter Brune) Date: Tue, 5 Nov 2013 15:08:48 -0600 Subject: [petsc-users] Parallel Graph Coloring In-Reply-To: References: Message-ID: On Tue, Nov 5, 2013 at 11:07 AM, Gaetan Kenway wrote: > Hi Peter > > Thanks for the info. I actually already have a coloring that is done > through the discretrization that is near optimal. The issue is that I have > a funny interprocessor dependence that is tricky to do my own sequential > coloring. So I was looking at doing it in a global, parallel sense. For > interest, I tried using the matGetColor() on my assembled matrix just to > see how many colors it would take. My discretrization coloring has 35 > colors (for a stencil with 33 cells) and the matGetColor() resulted in > approximately 78 colors or about twice as many. > Jacobian coloring requires a column, or distance-2 coloring of the matrix. The distance-2 colorings we get from the serial algorithms aren't that bad, so I suspect that you might be doing a 1-coloring from your mesh; I assume that you mean that your residual/Jacobian stencil has 33 entries rather than some distance-2 thing? How did the problems with your coloring manifest themselves? - Peter > > Thanks, > > Gaetan > > > On Tue, Nov 5, 2013 at 11:39 AM, Peter Brune wrote: > >> >> >> >> On Tue, Nov 5, 2013 at 10:29 AM, Gaetan Kenway wrote: >> >>> Hi >>> >>> I have a quick question regarding the MatGetColoring() function. >>> According to the documentation, >>> >>> "For parallel matrices currently converts to sequential matrix and uses >>> the sequential coloring on that." >>> >>> >> The meaning of this is that it transfers the parallel matrix to a single >> processor and does the coloring there, not that it only considers >> on-processor entries. What comes out is therefore a valid column coloring >> of the whole matrix. Parallel matrix coloring is in development but is not >> quite ready for public consumption. Anecdotal evidence suggests that the >> serial coloring does not become a bottleneck on small parallel runs. >> >> A good option is to provide the coloring through your discretization. >> This will often be more efficient than the matrix colorings, as you know >> the structure of your problem. >> >> - Peter >> >> >>> I am wondering does this give actually give a valid parallel coloring? >>> My application is a cell centered multi-block finite volume code, (with >>> block-based decomposition) and the communications are done using a two >>> level exchange of halo cells. I would like to use FD (actually forward mode >>> AD to compute the jacbian matrix). So, what I'm wondering is, if a cell >>> with color 0 is perturbed on processor 0, is it guaranteed that a residual >>> on processor 1, that is influenced by the 0-colored cell on proc zero is >>> *only* influenced by the color 0 from proc 0 and not a 0-colored cell on >>> proc 1? >>> >>> I hope that is clear >>> >>> Thank you, >>> >>> Gaetan Kenway >>> >>> >>> >>> >>> >>> >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From gaetank at gmail.com Tue Nov 5 15:24:28 2013 From: gaetank at gmail.com (Gaetan Kenway) Date: Tue, 5 Nov 2013 16:24:28 -0500 Subject: [petsc-users] Parallel Graph Coloring In-Reply-To: References: Message-ID: I'm not aware of the difference between a distance-2 coloring and a 1-coloring. What is the difference? Yes, I mean that my Jacibain stencil has 33 entires; it is a 3x3x3 cube with 6 extra cells on the middle of each of the 6 faces. For that stencil you can find an analytic coloring that requires 35 evaluations. The issue I have arises from the application of boundary conditions; The normal residual evaluation extrapolates values to halo cells and then these halo values are communicated. I am assembling the matrix using forward mode AD and in a fashion that doesn't require communication; the idea is to peturb "halo" cells in addition to the "real" cells and this eliminates the communication for the halo exchange. However, I didn't have a way of relating the extrapolated values from a different processor with the real cell on the other processor. It's a little confusing and fairly specific to the CFD solver I'm using. If I could get a global coloring that was somewhat close to what I can get analytically, I'd use that since it is simpler and requires less modification. Gaetan On Tue, Nov 5, 2013 at 4:08 PM, Peter Brune wrote: > > > > On Tue, Nov 5, 2013 at 11:07 AM, Gaetan Kenway wrote: > >> Hi Peter >> >> Thanks for the info. I actually already have a coloring that is done >> through the discretrization that is near optimal. The issue is that I have >> a funny interprocessor dependence that is tricky to do my own sequential >> coloring. So I was looking at doing it in a global, parallel sense. For >> interest, I tried using the matGetColor() on my assembled matrix just to >> see how many colors it would take. My discretrization coloring has 35 >> colors (for a stencil with 33 cells) and the matGetColor() resulted in >> approximately 78 colors or about twice as many. >> > > Jacobian coloring requires a column, or distance-2 coloring of the matrix. > The distance-2 colorings we get from the serial algorithms aren't that > bad, so I suspect that you might be doing a 1-coloring from your mesh; I > assume that you mean that your residual/Jacobian stencil has 33 entries > rather than some distance-2 thing? How did the problems with your coloring > manifest themselves? > > - Peter > > >> >> Thanks, >> >> Gaetan >> >> >> On Tue, Nov 5, 2013 at 11:39 AM, Peter Brune wrote: >> >>> >>> >>> >>> On Tue, Nov 5, 2013 at 10:29 AM, Gaetan Kenway wrote: >>> >>>> Hi >>>> >>>> I have a quick question regarding the MatGetColoring() function. >>>> According to the documentation, >>>> >>>> "For parallel matrices currently converts to sequential matrix and >>>> uses the sequential coloring on that." >>>> >>>> >>> The meaning of this is that it transfers the parallel matrix to a single >>> processor and does the coloring there, not that it only considers >>> on-processor entries. What comes out is therefore a valid column coloring >>> of the whole matrix. Parallel matrix coloring is in development but is not >>> quite ready for public consumption. Anecdotal evidence suggests that the >>> serial coloring does not become a bottleneck on small parallel runs. >>> >>> A good option is to provide the coloring through your discretization. >>> This will often be more efficient than the matrix colorings, as you know >>> the structure of your problem. >>> >>> - Peter >>> >>> >>>> I am wondering does this give actually give a valid parallel coloring? >>>> My application is a cell centered multi-block finite volume code, (with >>>> block-based decomposition) and the communications are done using a two >>>> level exchange of halo cells. I would like to use FD (actually forward mode >>>> AD to compute the jacbian matrix). So, what I'm wondering is, if a cell >>>> with color 0 is perturbed on processor 0, is it guaranteed that a residual >>>> on processor 1, that is influenced by the 0-colored cell on proc zero is >>>> *only* influenced by the color 0 from proc 0 and not a 0-colored cell on >>>> proc 1? >>>> >>>> I hope that is clear >>>> >>>> Thank you, >>>> >>>> Gaetan Kenway >>>> >>>> >>>> >>>> >>>> >>>> >>> >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From zhenglun.wei at gmail.com Tue Nov 5 21:21:01 2013 From: zhenglun.wei at gmail.com (Alan) Date: Tue, 05 Nov 2013 21:21:01 -0600 Subject: [petsc-users] Memory and Speed Issue of using MG as preconditioner Message-ID: <5279B59D.9030804@gmail.com> Dear all, I hope you're having a nice day. Recently, I came across a problem on using MG as preconditioner. Basically, to achieve the same finest mesh with pc_type = mg, the memory usage for -da_refine 2 is much more than that for -da_refine 5. To my limited knowledge, more refinement should consume more memory, which is contradict to the behavior of pc_type = mg in PETsc. Here, I provide two output files. They are all from /src/ksp/ksp/example/tutorial/ex45.c with 32 processes. The execute file for out-level2 is mpiexec -np 32 ./ex45 -pc_type mg -ksp_type cg -da_refine 2 -pc_mg_galerkin -ksp_rtol 1.0e-7 -mg_levels_pc_type jacobi -mg_levels_ksp_type chebyshev -dm_view -log_summary -pc_mg_log -pc_mg_monitor -ksp_view -ksp_monitor > out & and in ex45.c, KSPCreate is changed as: ierr = DMDACreate3d(PETSC_COMM_WORLD,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_STENCIL_STAR,-65,-33,-33,PETSC_DECIDE,PETSC_DECIDE,PETSC_DECIDE,1,1,0,0,0,&da);CHKERRQ(ierr); On the other hand, the execute file for out-level5 is mpiexec -np 32 ./ex45 -pc_type mg -ksp_type cg -da_refine 5 -pc_mg_galerkin -ksp_rtol 1.0e-7 -mg_levels_pc_type jacobi -mg_levels_ksp_type chebyshev -dm_view -log_summary -pc_mg_log -pc_mg_monitor -ksp_view -ksp_monitor > out & and in ex45.c, KSPCreate is changed as: ierr = DMDACreate3d(PETSC_COMM_WORLD,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_STENCIL_STAR,-9,-5,-5,PETSC_DECIDE,PETSC_DECIDE,PETSC_DECIDE,1,1,0,0,0,&da);CHKERRQ(ierr); In summary, the final finest meshes obtained for both cases are 257*129*129 as documented in both files. However, the out-level2 shows that the Matrix requested 822871308 memory while out-level5 only need 36052932. Furthermore, although the total iterations for KSP solver are shown as 5 times in both files. the wall time elapsed for out-level2 is around 150s, while out-level5 is about 4.7s. At last, there is a minor question. In both files, under 'Down solver (pre-smoother) on level 1' and 'Down solver (pre-smoother) on level 2', the type of "KSP Object: (mg_levels_1_est_)" and "KSP Object: (mg_levels_2_est_)" are all 'gmres'. Since I'm using uniformly Cartesian mesh, would it be helpful to speed up the solver if the 'gmres' is replaced by 'cg' here? If so, which PETSc option can change the type of KSP object. sincerely appreciate, Alan -------------- next part -------------- Processor [0] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 0 33 Processor [1] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 0 33 Processor [2] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 0 33 Processor [3] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 0 33 Processor [4] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 0 33 Processor [5] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 0 33 Processor [6] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 0 33 Processor [7] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 0 33 Processor [8] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 33 65 Processor [9] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 33 65 Processor [10] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 33 65 Processor [11] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 33 65 Processor [12] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 33 65 Processor [13] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 33 65 Processor [14] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 33 65 Processor [15] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 33 65 Processor [16] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 65 97 Processor [17] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 65 97 Processor [18] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 65 97 Processor [19] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 65 97 Processor [20] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 65 97 Processor [21] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 65 97 Processor [22] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 65 97 Processor [23] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 65 97 Processor [24] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 97 129 Processor [25] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 97 129 Processor [26] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 97 129 Processor [27] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 97 129 Processor [28] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 97 129 Processor [29] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 97 129 Processor [30] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 97 129 Processor [31] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 97 129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 Processor [0] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 0 17 Processor [1] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 0 17 Processor [2] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 0 17 Processor [3] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 0 17 Processor [4] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 0 17 Processor [5] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 0 17 Processor [6] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 0 17 Processor [7] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 0 17 Processor [8] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 17 33 Processor [9] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 17 33 Processor [10] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 17 33 Processor [11] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 17 33 Processor [12] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 17 33 Processor [13] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 17 33 Processor [14] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 17 33 Processor [15] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 17 33 Processor [16] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 33 49 Processor [17] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 33 49 Processor [18] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 33 49 Processor [19] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 33 49 Processor [20] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 33 49 Processor [21] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 33 49 Processor [22] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 33 49 Processor [23] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 33 49 Processor [24] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 49 65 Processor [25] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 49 65 Processor [26] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 49 65 Processor [27] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 49 65 Processor [28] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 49 65 Processor [29] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 49 65 Processor [30] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 49 65 Processor [31] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 49 65 Processor [0] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 0 9 Processor [1] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 0 9 Processor [2] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 0 9 Processor [3] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 0 9 Processor [4] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 0 9 Processor [5] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 0 9 Processor [6] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 0 9 Processor [7] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 0 9 Processor [8] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 9 17 Processor [9] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 9 17 Processor [10] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 9 17 Processor [11] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 9 17 Processor [12] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 9 17 Processor [13] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 9 17 Processor [14] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 9 17 Processor [15] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 9 17 Processor [16] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 17 25 Processor [17] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 17 25 Processor [18] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 17 25 Processor [19] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 17 25 Processor [20] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 17 25 Processor [21] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 17 25 Processor [22] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 17 25 Processor [23] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 17 25 Processor [24] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 25 33 Processor [25] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 25 33 Processor [26] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 25 33 Processor [27] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 25 33 Processor [28] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 25 33 Processor [29] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 25 33 Processor [30] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 25 33 Processor [31] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 25 33 0 KSP Residual norm 2.036594349596e+03 1 KSP Residual norm 8.756270777762e+01 2 KSP Residual norm 3.092374574522e+00 3 KSP Residual norm 1.220382147945e-01 4 KSP Residual norm 2.871729837203e-02 KSP Object: 32 MPI processes type: cg maximum iterations=10000 tolerances: relative=1e-07, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using PRECONDITIONED norm type for convergence test PC Object: 32 MPI processes type: mg MG: type is MULTIPLICATIVE, levels=3 cycles=v Cycles per PCApply=1 Using Galerkin computed coarse grid matrices Coarse grid solver -- level ------------------------------- KSP Object: (mg_coarse_) 32 MPI processes type: preonly maximum iterations=1, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_) 32 MPI processes type: redundant Redundant preconditioner: First (color=0) of 32 PCs follows KSP Object: (mg_coarse_redundant_) 1 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_redundant_) 1 MPI processes type: lu LU: out-of-place factorization tolerance for zero pivot 2.22045e-14 using diagonal shift on blocks to prevent zero pivot matrix ordering: nd factor fill ratio given 5, needed 35.0339 Factored matrix follows: Matrix Object: 1 MPI processes type: seqaij rows=70785, cols=70785 package used to perform factorization: petsc total: nonzeros=63619375, allocated nonzeros=63619375 total number of mallocs used during MatSetValues calls =0 not using I-node routines linear system matrix = precond matrix: Matrix Object: 1 MPI processes type: seqaij rows=70785, cols=70785 total: nonzeros=1815937, allocated nonzeros=1911195 total number of mallocs used during MatSetValues calls =0 not using I-node routines linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=70785, cols=70785 total: nonzeros=1815937, allocated nonzeros=1815937 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Down solver (pre-smoother) on level 1 ------------------------------- KSP Object: (mg_levels_1_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.231542, max = 2.54696 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_1_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_1_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=545025, cols=545025 total: nonzeros=14340865, allocated nonzeros=14340865 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_1_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=545025, cols=545025 total: nonzeros=14340865, allocated nonzeros=14340865 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 2 ------------------------------- KSP Object: (mg_levels_2_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.155706, max = 1.71277 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_2_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_2_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_2_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 Up solver (post-smoother) same as down solver (pre-smoother) linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 Residual norm 0.000941649 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex45 on a linux-gnu-c-nodebug named n042 with 32 processors, by zlwei Tue Nov 5 12:21:51 2013 Using Petsc Development GIT revision: d696997672013bb4513d3ff57c61cc10e09b71f6 GIT Date: 2013-06-13 10:28:37 -0500 Max Max/Min Avg Total Time (sec): 5.699e+02 1.00005 5.699e+02 Objects: 1.750e+02 1.00000 1.750e+02 Flops: 7.470e+10 1.00022 7.468e+10 2.390e+12 Flops/sec: 1.311e+08 1.00020 1.310e+08 4.193e+09 MPI Messages: 1.594e+03 1.77506 1.234e+03 3.948e+04 MPI Message Lengths: 3.375e+07 1.17496 2.581e+04 1.019e+09 MPI Reductions: 3.500e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.7734e+02 48.7% 2.3636e+12 98.9% 1.188e+04 30.1% 1.963e+04 76.0% 2.430e+02 69.4% 1: MG Apply: 2.9257e+02 51.3% 2.6346e+10 1.1% 2.760e+04 69.9% 6.185e+03 24.0% 1.060e+02 30.3% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage KSPSetUp 5 1.0 4.6016e-02 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 1.8e+01 0 0 0 0 5 0 0 0 0 7 0 Warning -- total time of even greater than time of entire stage -- something is wrong with the timer KSPSolve 1 1.0 5.6976e+02 1.0 7.47e+10 1.0 3.9e+04 2.6e+04 3.2e+02100100 99100 93 205101327131134 4194 VecTDot 9 1.0 1.4363e-01 3.3 2.51e+06 1.1 0.0e+00 0.0e+00 9.0e+00 0 0 0 0 3 0 0 0 0 4 536 VecNorm 6 1.0 3.8997e-0114.8 1.67e+06 1.1 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 2 0 0 0 0 2 132 VecCopy 2 1.0 6.8030e-0312.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 12 1.0 2.0368e-03 6.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 9 1.0 4.1642e-0215.4 2.51e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1849 VecAYPX 3 1.0 2.5070e-0233.2 8.37e+05 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1024 VecScatterBegin 7 1.0 9.5501e-0321.5 0.00e+00 0.0 8.7e+02 1.7e+04 0.0e+00 0 0 2 1 0 0 0 7 2 0 0 VecScatterEnd 7 1.0 3.0083e-0255.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 5 1.0 1.0018e-01 7.1 8.98e+06 1.1 6.4e+02 2.3e+04 0.0e+00 0 0 2 1 0 0 0 5 2 0 2758 MatMultTranspose 2 1.0 2.0180e-0212.1 1.04e+06 1.0 2.3e+02 2.1e+03 0.0e+00 0 0 1 0 0 0 0 2 0 0 1601 MatLUFactorSym 1 1.0 4.2020e+00 2.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 1 0 0 0 1 1 0 0 0 1 0 MatLUFactorNum 1 1.0 5.5597e+02 4.6 7.38e+10 1.0 0.0e+00 0.0e+00 0.0e+00 47 99 0 0 0 97100 0 0 0 4249 MatAssemblyBegin 11 1.0 4.3128e-01 2.9 0.00e+00 0.0 0.0e+00 0.0e+00 1.2e+01 0 0 0 0 3 0 0 0 0 5 0 MatAssemblyEnd 11 1.0 2.1238e-01 1.8 0.00e+00 0.0 2.2e+03 1.0e+03 4.0e+01 0 0 6 0 11 0 0 18 0 16 0 MatGetRowIJ 1 1.0 4.3992e-02 3.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 3.3257e-01 2.6 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 1 0 0 0 0 1 0 MatView 8 1.3 4.6694e-03 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 2 0 0 0 0 2 0 MatPtAP 2 1.0 8.1194e-01 1.0 2.06e+07 1.1 3.9e+03 8.8e+03 5.0e+01 0 0 10 3 14 0 0 33 4 21 781 MatPtAPSymbolic 2 1.0 5.0288e-01 1.1 0.00e+00 0.0 2.2e+03 1.2e+04 3.0e+01 0 0 6 2 9 0 0 18 3 12 0 MatPtAPNumeric 2 1.0 3.5799e-01 1.2 2.06e+07 1.1 1.7e+03 5.2e+03 2.0e+01 0 0 4 1 6 0 0 14 1 8 1771 MatGetRedundant 1 1.0 2.1102e+00 1.2 0.00e+00 0.0 3.0e+03 2.3e+05 4.0e+00 0 0 8 68 1 1 0 25 89 2 0 MatGetLocalMat 2 1.0 4.9906e-02 6.4 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 0 0 0 0 1 0 0 0 0 2 0 MatGetBrAoCol 2 1.0 5.5355e-02 2.8 0.00e+00 0.0 1.5e+03 1.4e+04 4.0e+00 0 0 4 2 1 0 0 13 3 2 0 MatGetSymTrans 4 1.0 2.7369e-02 8.7 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCSetUp 1 1.0 5.6158e+02 4.5 7.38e+10 1.0 1.1e+04 7.2e+04 1.8e+02 48 99 27 74 51 100100 89 98 73 4208 Warning -- total time of even greater than time of entire stage -- something is wrong with the timer PCApply 5 1.0 4.4313e+0259.4 8.32e+08 1.0 2.8e+04 8.8e+03 1.1e+02 51 1 70 24 30 105 1232 32 44 59 MGSetup Level 0 1 1.0 5.6044e+02 4.5 7.38e+10 1.0 5.1e+03 1.4e+05 2.7e+01 48 99 13 71 8 99100 43 93 11 4215 MGSetup Level 1 1 1.0 9.1751e-03 2.0 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 2 0 0 0 0 2 0 MGSetup Level 2 1 1.0 1.5626e-02 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 2 0 0 0 0 2 0 --- Event Stage 1: MG Apply KSPGMRESOrthog 20 1.0 4.6161e-01 2.0 3.47e+07 1.1 0.0e+00 0.0e+00 2.0e+01 0 0 0 0 6 0 4 0 0 19 2298 KSPSetUp 2 1.0 4.3606e+029821.1 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+01 50 0 0 0 6 98 0 0 0 19 0 KSPSolve 25 1.0 4.4278e+0262.2 8.08e+08 1.0 2.3e+04 9.7e+03 1.1e+02 51 1 58 22 30 100 97 83 90100 58 VecMDot 20 1.0 4.4832e-01 3.1 1.74e+07 1.1 0.0e+00 0.0e+00 2.0e+01 0 0 0 0 6 0 2 0 0 19 1183 VecNorm 22 1.0 2.1354e-01 2.9 3.47e+06 1.1 0.0e+00 0.0e+00 2.2e+01 0 0 0 0 6 0 0 0 0 21 497 VecScale 62 1.0 6.3914e-0213.6 4.90e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 2339 VecCopy 12 1.0 2.0800e-0211.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 51 1.0 2.0875e-0213.7 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 84 1.0 8.1026e-02 8.7 1.33e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 2 0 0 0 4999 VecAYPX 80 1.0 8.6726e-02 7.8 7.90e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 2780 VecMAXPY 22 1.0 1.3574e-01 9.2 2.05e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 2 0 0 0 4618 VecPointwiseMult 82 1.0 1.3558e-01 8.2 6.48e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 1458 VecScatterBegin 112 1.0 8.0228e-02 8.3 0.00e+00 0.0 2.8e+04 8.8e+03 0.0e+00 0 0 70 24 0 0 0100100 0 0 VecScatterEnd 112 1.0 5.4348e+0012.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 VecNormalize 22 1.0 2.1602e-01 2.4 5.21e+06 1.1 0.0e+00 0.0e+00 2.2e+01 0 0 0 0 6 0 1 0 0 21 737 MatMult 82 1.0 4.7078e+00 4.2 1.12e+08 1.1 2.0e+04 7.5e+03 0.0e+00 0 0 52 15 0 1 13 74 62 0 726 MatMultAdd 10 1.0 2.7111e+0069.2 5.21e+06 1.0 1.2e+03 2.1e+03 0.0e+00 0 0 3 0 0 0 1 4 1 0 60 MatMultTranspose 10 1.0 1.4186e-0115.2 5.21e+06 1.0 1.2e+03 2.1e+03 0.0e+00 0 0 3 0 0 0 1 4 1 0 1139 MatSolve 5 1.0 4.9743e+00 6.4 6.36e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 1 77 0 0 0 4090 PCApply 87 1.0 5.2570e+00 5.0 6.42e+08 1.0 5.0e+03 1.8e+04 4.0e+00 0 1 13 9 1 1 78 18 36 4 3908 MGSmooth Level 0 5 1.0 5.0933e+00 5.1 6.36e+08 1.0 5.0e+03 1.8e+04 0.0e+00 0 1 13 9 0 1 77 18 36 0 3995 MGSmooth Level 1 10 1.0 4.4056e+00 9.2 4.25e+07 1.1 1.3e+04 2.1e+03 5.3e+01 0 0 34 3 15 1 5 48 11 50 287 MGResid Level 1 5 1.0 1.0928e-01 2.7 4.80e+06 1.1 1.8e+03 2.1e+03 0.0e+00 0 0 5 0 0 0 1 7 2 0 1312 MGInterp Level 1 10 1.0 2.6571e+00620.1 1.20e+06 1.1 1.2e+03 8.7e+02 0.0e+00 0 0 3 0 0 0 0 4 0 0 14 MGSmooth Level 2 10 1.0 4.3692e+02296.7 1.29e+08 1.1 4.6e+03 2.3e+04 5.3e+01 50 0 12 10 15 98 15 17 43 50 9 MGResid Level 2 5 1.0 1.4266e-01 2.4 9.67e+06 1.1 6.4e+02 2.3e+04 0.0e+00 0 0 2 1 0 0 1 2 6 0 2087 MGInterp Level 2 10 1.0 1.2601e-01 3.0 9.22e+06 1.0 1.2e+03 3.3e+03 0.0e+00 0 0 3 0 0 0 1 4 2 0 2276 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 1 1 564 0 Krylov Solver 7 7 66456 0 DMKSP interface 2 2 1296 0 Vector 46 78 36381912 0 Vector Scatter 13 13 13676 0 Matrix 21 21 822871308 0 Distributed Mesh 3 3 1391808 0 Bipartite Graph 6 6 4752 0 Viewer 2 1 728 0 Index Set 32 32 1989612 0 IS L to G Mapping 3 3 690348 0 Preconditioner 7 7 6432 0 --- Event Stage 1: MG Apply Vector 32 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 5.00679e-07 Average time for MPI_Barrier(): 0.000607204 Average time for zero size MPI_Send(): 0.000628747 #PETSc Option Table entries: -da_refine 2 -dm_view -ksp_monitor -ksp_rtol 1.0e-7 -ksp_type cg -ksp_view -log_summary -mg_levels_ksp_type chebyshev -mg_levels_pc_type jacobi -pc_mg_galerkin -pc_mg_log -pc_mg_monitor -pc_type mg #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Thu Jun 13 15:51:55 2013 Configure options: --download-f-blas-lapack --download-hypre --download-mpich --with-cc=gcc --with-debugging=no --with-fc=gfortran PETSC_ARCH=linux-gnu-c-nodebug ----------------------------------------- Libraries compiled on Thu Jun 13 15:51:55 2013 on login1.ittc.ku.edu Machine characteristics: Linux-2.6.32-220.13.1.el6.x86_64-x86_64-with-redhat-6.2-Santiago Using PETSc directory: /bio/work1/zlwei/PETSc/petsc-dev Using PETSc arch: linux-gnu-c-nodebug ----------------------------------------- Using C compiler: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpif90 -fPIC -Wall -Wno-unused-variable -O ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/include -I/bio/work1/zlwei/PETSc/petsc-dev/include -I/bio/work1/zlwei/PETSc/petsc-dev/include -I/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/include ----------------------------------------- Using C linker: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpicc Using Fortran linker: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpif90 Using libraries: -Wl,-rpath,/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -L/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -lpetsc -Wl,-rpath,/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -L/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -lHYPRE -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lmpichcxx -lstdc++ -lflapack -lfblas -lX11 -lpthread -lmpichf90 -lgfortran -lm -lm -lmpichcxx -lstdc++ -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl ----------------------------------------- -------------- next part -------------- Processor [0] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 0 33 Processor [1] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 0 33 Processor [2] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 0 33 Processor [3] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 0 33 Processor [4] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 0 33 Processor [5] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 0 33 Processor [6] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 0 33 Processor [7] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 0 33 Processor [8] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 33 65 Processor [9] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 33 65 Processor [10] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 33 65 Processor [11] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 33 65 Processor [12] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 33 65 Processor [13] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 33 65 Processor [14] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 33 65 Processor [15] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 33 65 Processor [16] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 65 97 Processor [17] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 65 97 Processor [18] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 65 97 Processor [19] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 65 97 Processor [20] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 65 97 Processor [21] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 65 97 Processor [22] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 65 97 Processor [23] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 65 97 Processor [24] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 97 129 Processor [25] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 97 129 Processor [26] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 97 129 Processor [27] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 97 129 Processor [28] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 97 129 Processor [29] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 97 129 Processor [30] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 97 129 Processor [31] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 97 129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 Processor [0] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 0 17 Processor [1] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 0 17 Processor [2] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 0 17 Processor [3] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 0 17 Processor [4] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 0 17 Processor [5] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 0 17 Processor [6] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 0 17 Processor [7] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 0 17 Processor [8] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 17 33 Processor [9] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 17 33 Processor [10] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 17 33 Processor [11] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 17 33 Processor [12] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 17 33 Processor [13] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 17 33 Processor [14] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 17 33 Processor [15] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 17 33 Processor [16] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 33 49 Processor [17] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 33 49 Processor [18] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 33 49 Processor [19] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 33 49 Processor [20] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 33 49 Processor [21] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 33 49 Processor [22] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 33 49 Processor [23] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 33 49 Processor [24] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 49 65 Processor [25] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 49 65 Processor [26] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 49 65 Processor [27] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 49 65 Processor [28] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 49 65 Processor [29] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 49 65 Processor [30] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 49 65 Processor [31] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 49 65 Processor [0] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 0 9 Processor [1] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 0 9 Processor [2] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 0 9 Processor [3] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 0 9 Processor [4] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 0 9 Processor [5] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 0 9 Processor [6] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 0 9 Processor [7] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 0 9 Processor [8] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 9 17 Processor [9] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 9 17 Processor [10] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 9 17 Processor [11] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 9 17 Processor [12] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 9 17 Processor [13] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 9 17 Processor [14] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 9 17 Processor [15] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 9 17 Processor [16] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 17 25 Processor [17] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 17 25 Processor [18] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 17 25 Processor [19] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 17 25 Processor [20] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 17 25 Processor [21] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 17 25 Processor [22] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 17 25 Processor [23] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 17 25 Processor [24] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 25 33 Processor [25] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 25 33 Processor [26] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 25 33 Processor [27] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 25 33 Processor [28] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 25 33 Processor [29] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 25 33 Processor [30] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 25 33 Processor [31] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 25 33 Processor [0] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 0 9, Z range of indices: 0 5 Processor [1] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 0 9, Z range of indices: 0 5 Processor [2] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 0 9, Z range of indices: 0 5 Processor [3] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 0 9, Z range of indices: 0 5 Processor [4] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 9 17, Z range of indices: 0 5 Processor [5] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 9 17, Z range of indices: 0 5 Processor [6] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 9 17, Z range of indices: 0 5 Processor [7] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 9 17, Z range of indices: 0 5 Processor [8] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 0 9, Z range of indices: 5 9 Processor [9] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 0 9, Z range of indices: 5 9 Processor [10] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 0 9, Z range of indices: 5 9 Processor [11] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 0 9, Z range of indices: 5 9 Processor [12] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 9 17, Z range of indices: 5 9 Processor [13] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 9 17, Z range of indices: 5 9 Processor [14] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 9 17, Z range of indices: 5 9 Processor [15] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 9 17, Z range of indices: 5 9 Processor [16] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 0 9, Z range of indices: 9 13 Processor [17] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 0 9, Z range of indices: 9 13 Processor [18] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 0 9, Z range of indices: 9 13 Processor [19] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 0 9, Z range of indices: 9 13 Processor [20] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 9 17, Z range of indices: 9 13 Processor [21] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 9 17, Z range of indices: 9 13 Processor [22] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 9 17, Z range of indices: 9 13 Processor [23] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 9 17, Z range of indices: 9 13 Processor [24] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 0 9, Z range of indices: 13 17 Processor [25] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 0 9, Z range of indices: 13 17 Processor [26] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 0 9, Z range of indices: 13 17 Processor [27] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 0 9, Z range of indices: 13 17 Processor [28] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 9 17, Z range of indices: 13 17 Processor [29] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 9 17, Z range of indices: 13 17 Processor [30] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 9 17, Z range of indices: 13 17 Processor [31] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 9 17, Z range of indices: 13 17 Processor [0] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 0 5, Z range of indices: 0 3 Processor [1] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 0 5, Z range of indices: 0 3 Processor [2] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 0 5, Z range of indices: 0 3 Processor [3] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 0 5, Z range of indices: 0 3 Processor [4] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 5 9, Z range of indices: 0 3 Processor [5] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 5 9, Z range of indices: 0 3 Processor [6] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 5 9, Z range of indices: 0 3 Processor [7] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 5 9, Z range of indices: 0 3 Processor [8] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 0 5, Z range of indices: 3 5 Processor [9] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 0 5, Z range of indices: 3 5 Processor [10] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 0 5, Z range of indices: 3 5 Processor [11] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 0 5, Z range of indices: 3 5 Processor [12] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 5 9, Z range of indices: 3 5 Processor [13] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 5 9, Z range of indices: 3 5 Processor [14] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 5 9, Z range of indices: 3 5 Processor [15] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 5 9, Z range of indices: 3 5 Processor [16] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 0 5, Z range of indices: 5 7 Processor [17] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 0 5, Z range of indices: 5 7 Processor [18] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 0 5, Z range of indices: 5 7 Processor [19] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 0 5, Z range of indices: 5 7 Processor [20] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 5 9, Z range of indices: 5 7 Processor [21] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 5 9, Z range of indices: 5 7 Processor [22] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 5 9, Z range of indices: 5 7 Processor [23] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 5 9, Z range of indices: 5 7 Processor [24] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 0 5, Z range of indices: 7 9 Processor [25] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 0 5, Z range of indices: 7 9 Processor [26] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 0 5, Z range of indices: 7 9 Processor [27] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 0 5, Z range of indices: 7 9 Processor [28] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 5 9, Z range of indices: 7 9 Processor [29] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 5 9, Z range of indices: 7 9 Processor [30] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 5 9, Z range of indices: 7 9 Processor [31] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 5 9, Z range of indices: 7 9 Processor [0] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 0 3, Z range of indices: 0 2 Processor [1] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 0 3, Z range of indices: 0 2 Processor [2] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 0 3, Z range of indices: 0 2 Processor [3] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 0 3, Z range of indices: 0 2 Processor [4] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 3 5, Z range of indices: 0 2 Processor [5] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 3 5, Z range of indices: 0 2 Processor [6] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 3 5, Z range of indices: 0 2 Processor [7] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 3 5, Z range of indices: 0 2 Processor [8] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 0 3, Z range of indices: 2 3 Processor [9] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 0 3, Z range of indices: 2 3 Processor [10] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 0 3, Z range of indices: 2 3 Processor [11] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 0 3, Z range of indices: 2 3 Processor [12] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 3 5, Z range of indices: 2 3 Processor [13] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 3 5, Z range of indices: 2 3 Processor [14] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 3 5, Z range of indices: 2 3 Processor [15] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 3 5, Z range of indices: 2 3 Processor [16] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 0 3, Z range of indices: 3 4 Processor [17] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 0 3, Z range of indices: 3 4 Processor [18] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 0 3, Z range of indices: 3 4 Processor [19] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 0 3, Z range of indices: 3 4 Processor [20] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 3 5, Z range of indices: 3 4 Processor [21] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 3 5, Z range of indices: 3 4 Processor [22] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 3 5, Z range of indices: 3 4 Processor [23] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 3 5, Z range of indices: 3 4 Processor [24] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 0 3, Z range of indices: 4 5 Processor [25] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 0 3, Z range of indices: 4 5 Processor [26] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 0 3, Z range of indices: 4 5 Processor [27] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 0 3, Z range of indices: 4 5 Processor [28] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 3 5, Z range of indices: 4 5 Processor [29] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 3 5, Z range of indices: 4 5 Processor [30] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 3 5, Z range of indices: 4 5 Processor [31] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 3 5, Z range of indices: 4 5 0 KSP Residual norm 1.990474015208e+03 1 KSP Residual norm 1.163078153200e+02 2 KSP Residual norm 2.809444096980e+00 3 KSP Residual norm 2.139770554363e-01 4 KSP Residual norm 4.835908670272e-02 KSP Object: 32 MPI processes type: cg maximum iterations=10000 tolerances: relative=1e-07, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using PRECONDITIONED norm type for convergence test PC Object: 32 MPI processes type: mg MG: type is MULTIPLICATIVE, levels=6 cycles=v Cycles per PCApply=1 Using Galerkin computed coarse grid matrices Coarse grid solver -- level ------------------------------- KSP Object: (mg_coarse_) 32 MPI processes type: preonly maximum iterations=1, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_) 32 MPI processes type: redundant Redundant preconditioner: First (color=0) of 32 PCs follows KSP Object: (mg_coarse_redundant_) 1 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_redundant_) 1 MPI processes type: lu LU: out-of-place factorization tolerance for zero pivot 2.22045e-14 using diagonal shift on blocks to prevent zero pivot matrix ordering: nd factor fill ratio given 5, needed 3.15101 Factored matrix follows: Matrix Object: 1 MPI processes type: seqaij rows=225, cols=225 package used to perform factorization: petsc total: nonzeros=13313, allocated nonzeros=13313 total number of mallocs used during MatSetValues calls =0 using I-node routines: found 174 nodes, limit used is 5 linear system matrix = precond matrix: Matrix Object: 1 MPI processes type: seqaij rows=225, cols=225 total: nonzeros=4225, allocated nonzeros=6075 total number of mallocs used during MatSetValues calls =0 not using I-node routines linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=225, cols=225 total: nonzeros=4225, allocated nonzeros=4225 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Down solver (pre-smoother) on level 1 ------------------------------- KSP Object: (mg_levels_1_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.283115, max = 3.11426 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_1_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_1_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=1377, cols=1377 total: nonzeros=30625, allocated nonzeros=30625 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_1_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=1377, cols=1377 total: nonzeros=30625, allocated nonzeros=30625 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 2 ------------------------------- KSP Object: (mg_levels_2_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.285627, max = 3.1419 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_2_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_2_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=9537, cols=9537 total: nonzeros=232897, allocated nonzeros=232897 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_2_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=9537, cols=9537 total: nonzeros=232897, allocated nonzeros=232897 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 3 ------------------------------- KSP Object: (mg_levels_3_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.275571, max = 3.03128 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_3_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_3_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=70785, cols=70785 total: nonzeros=1815937, allocated nonzeros=1815937 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_3_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=70785, cols=70785 total: nonzeros=1815937, allocated nonzeros=1815937 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 4 ------------------------------- KSP Object: (mg_levels_4_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.231542, max = 2.54696 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_4_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_4_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=545025, cols=545025 total: nonzeros=14340865, allocated nonzeros=14340865 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_4_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=545025, cols=545025 total: nonzeros=14340865, allocated nonzeros=14340865 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 5 ------------------------------- KSP Object: (mg_levels_5_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.155706, max = 1.71277 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_5_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_5_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_5_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 Up solver (post-smoother) same as down solver (pre-smoother) linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 Residual norm 0.000947651 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex45 on a linux-gnu-c-nodebug named n042 with 32 processors, by zlwei Tue Nov 5 15:22:44 2013 Using Petsc Development GIT revision: d696997672013bb4513d3ff57c61cc10e09b71f6 GIT Date: 2013-06-13 10:28:37 -0500 Max Max/Min Avg Total Time (sec): 4.761e+00 1.00178 4.756e+00 Objects: 3.570e+02 1.00000 3.570e+02 Flops: 2.438e+08 1.08372 2.320e+08 7.424e+09 Flops/sec: 5.127e+07 1.08375 4.879e+07 1.561e+09 MPI Messages: 4.234e+03 2.12444 3.045e+03 9.745e+04 MPI Message Lengths: 9.073e+06 1.78864 2.350e+03 2.291e+08 MPI Reductions: 7.370e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.3328e+00 49.1% 1.2102e+09 16.3% 2.114e+04 21.7% 6.174e+02 26.3% 4.710e+02 63.9% 1: MG Apply: 2.4230e+00 50.9% 6.2142e+09 83.7% 7.631e+04 78.3% 1.733e+03 73.7% 2.650e+02 36.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage KSPSetUp 8 1.0 6.4316e-02 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 3.6e+01 1 0 0 0 5 2 0 0 0 8 0 Warning -- total time of even greater than time of entire stage -- something is wrong with the timer KSPSolve 1 1.0 4.6004e+00 1.0 2.41e+08 1.1 9.7e+04 2.3e+03 7.1e+02 97 99 99 98 96 197607458373150 1598 VecTDot 9 1.0 1.2305e-01 2.1 2.51e+06 1.1 0.0e+00 0.0e+00 9.0e+00 2 1 0 0 1 4 6 0 0 2 626 VecNorm 6 1.0 4.0866e-0115.1 1.67e+06 1.1 0.0e+00 0.0e+00 6.0e+00 4 1 0 0 1 9 4 0 0 1 126 VecCopy 2 1.0 9.8240e-03 7.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 27 1.0 1.7531e-03 3.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 9 1.0 3.7209e-0213.7 2.51e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 1 6 0 0 0 2069 VecAYPX 3 1.0 1.4415e-0219.0 8.37e+05 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 2 0 0 0 1780 VecScatterBegin 10 1.0 1.0550e-0217.5 0.00e+00 0.0 1.2e+03 1.2e+04 0.0e+00 0 0 1 7 0 0 0 6 25 0 0 VecScatterEnd 10 1.0 3.9624e-0252.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 5 1.0 9.6178e-02 4.1 8.98e+06 1.1 6.4e+02 2.3e+04 0.0e+00 1 4 1 6 0 2 23 3 24 0 2873 MatMultTranspose 5 1.0 2.2073e-0210.4 1.06e+06 1.0 5.8e+02 9.0e+02 0.0e+00 0 0 1 0 0 0 3 3 1 0 1488 MatLUFactorSym 1 1.0 9.3551e-0325.4 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 1 0 MatLUFactorNum 1 1.0 3.3901e-03 6.0 4.79e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 4518 MatAssemblyBegin 23 1.0 5.2334e-01 5.1 0.00e+00 0.0 0.0e+00 0.0e+00 2.4e+01 7 0 0 0 3 14 0 0 0 5 0 MatAssemblyEnd 23 1.0 1.9211e-01 1.4 0.00e+00 0.0 5.1e+03 4.5e+02 8.8e+01 3 0 5 1 12 7 0 24 4 19 0 MatGetRowIJ 1 1.0 1.4806e-04 4.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 3.9680e-0327.4 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 14 1.2 8.8232e-03 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 1.2e+01 0 0 0 0 2 0 0 0 0 3 0 MatPtAP 5 1.0 8.5644e-01 1.0 2.14e+07 1.1 1.1e+04 3.4e+03 1.2e+02 18 9 11 16 17 37 54 51 61 27 765 MatPtAPSymbolic 5 1.0 5.1170e-01 1.1 0.00e+00 0.0 6.5e+03 4.1e+03 7.5e+01 10 0 7 12 10 21 0 31 45 16 0 MatPtAPNumeric 5 1.0 3.8337e-01 1.1 2.14e+07 1.1 4.3e+03 2.2e+03 5.0e+01 8 9 4 4 7 15 54 20 16 11 1708 MatGetRedundant 1 1.0 1.8767e-02 2.7 0.00e+00 0.0 3.0e+03 5.5e+02 4.0e+00 0 0 3 1 1 0 0 14 3 1 0 MatGetLocalMat 5 1.0 6.9065e-02 8.1 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 1 1 0 0 0 2 0 MatGetBrAoCol 5 1.0 9.8569e-02 2.5 0.00e+00 0.0 4.8e+03 4.6e+03 1.0e+01 1 0 5 10 1 3 0 23 37 2 0 MatGetSymTrans 10 1.0 1.9155e-02 5.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCSetUp 1 1.0 1.3558e+00 1.0 2.29e+07 1.1 2.0e+04 2.1e+03 4.0e+02 28 9 20 18 54 58 58 94 70 85 518 Warning -- total time of even greater than time of entire stage -- something is wrong with the timer PCApply 5 1.0 2.6004e+00 1.2 2.04e+08 1.1 7.6e+04 2.2e+03 2.6e+02 51 84 78 74 36 104513361281 56 2390 MGSetup Level 0 1 1.0 5.4679e-02 1.3 4.79e+05 1.0 5.1e+03 3.4e+02 2.7e+01 1 0 5 1 4 2 1 24 3 6 280 MGSetup Level 1 1 1.0 9.1040e-03 2.1 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 0 0 0 0 1 0 MGSetup Level 2 1 1.0 5.1131e-03 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 0 0 0 0 1 0 MGSetup Level 3 1 1.0 4.5691e-03 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 0 0 0 0 1 0 MGSetup Level 4 1 1.0 6.5250e-03 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 0 0 0 0 1 0 MGSetup Level 5 1 1.0 2.0500e-02 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 1 0 0 0 1 0 --- Event Stage 1: MG Apply KSPGMRESOrthog 50 1.0 4.9737e-01 1.9 3.54e+07 1.1 0.0e+00 0.0e+00 5.0e+01 8 15 0 0 7 16 17 0 0 19 2169 KSPSetUp 5 1.0 1.6936e-01 2.5 0.00e+00 0.0 0.0e+00 0.0e+00 5.0e+01 2 0 0 0 7 5 0 0 0 19 0 KSPSolve 55 1.0 2.2558e+00 1.2 1.79e+08 1.1 6.3e+04 2.3e+03 2.6e+02 44 73 64 63 36 87 87 82 85100 2404 VecMDot 50 1.0 4.7352e-01 2.6 1.77e+07 1.1 0.0e+00 0.0e+00 5.0e+01 7 7 0 0 7 14 9 0 0 19 1139 VecNorm 55 1.0 2.4168e-01 2.1 3.54e+06 1.1 0.0e+00 0.0e+00 5.5e+01 4 1 0 0 7 8 2 0 0 21 446 VecScale 155 1.0 7.0602e-0214.9 4.99e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 2 0 0 0 1 2 0 0 0 2153 VecCopy 30 1.0 2.8271e-0214.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 105 1.0 1.3875e-0210.7 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 210 1.0 7.9889e-02 8.2 1.35e+07 1.1 0.0e+00 0.0e+00 0.0e+00 1 6 0 0 0 2 7 0 0 0 5156 VecAYPX 200 1.0 8.6185e-02 7.8 8.05e+06 1.1 0.0e+00 0.0e+00 0.0e+00 1 3 0 0 0 2 4 0 0 0 2845 VecMAXPY 55 1.0 1.3215e-0110.3 2.09e+07 1.1 0.0e+00 0.0e+00 0.0e+00 1 9 0 0 0 2 10 0 0 0 4824 VecPointwiseMult 205 1.0 1.3009e-01 7.2 6.60e+06 1.1 0.0e+00 0.0e+00 0.0e+00 1 3 0 0 0 3 3 0 0 0 1545 VecScatterBegin 265 1.0 9.1651e-02 5.2 0.00e+00 0.0 7.6e+04 2.2e+03 0.0e+00 1 0 78 74 0 2 0100100 0 0 VecScatterEnd 265 1.0 1.2055e+00 3.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 16 0 0 0 0 31 0 0 0 0 0 VecNormalize 55 1.0 2.4665e-01 2.0 5.31e+06 1.1 0.0e+00 0.0e+00 5.5e+01 4 2 0 0 7 8 3 0 0 21 656 MatMult 205 1.0 1.3930e+00 1.4 1.18e+08 1.1 6.6e+04 2.5e+03 0.0e+00 26 48 67 71 0 51 58 86 97 0 2575 MatMultAdd 25 1.0 1.4686e-01 4.6 5.31e+06 1.0 2.9e+03 9.0e+02 0.0e+00 2 2 3 1 0 3 3 4 2 0 1118 MatMultTranspose 25 1.0 1.0048e-01 9.3 5.31e+06 1.0 2.9e+03 9.0e+02 0.0e+00 1 2 3 1 0 2 3 4 2 0 1635 MatSolve 5 1.0 5.9199e-04 2.9 1.32e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 7135 PCApply 210 1.0 1.8144e-01 2.6 6.73e+06 1.1 5.0e+03 5.6e+01 1.0e+01 3 3 5 0 1 5 3 6 0 4 1131 MGSmooth Level 0 5 1.0 1.7867e-02 4.3 1.32e+05 1.0 5.0e+03 5.6e+01 0.0e+00 0 0 5 0 0 1 0 6 0 0 236 MGSmooth Level 1 10 1.0 7.2614e-02 1.1 1.45e+05 2.5 1.3e+04 5.3e+01 5.3e+01 1 0 14 0 7 3 0 17 0 20 38 MGResid Level 1 5 1.0 9.3150e-03 2.7 1.57e+04 2.6 1.8e+03 5.3e+01 0.0e+00 0 0 2 0 0 0 0 2 0 0 33 MGInterp Level 1 10 1.0 1.1652e-02 7.1 3.92e+03 1.8 1.2e+03 2.4e+01 0.0e+00 0 0 1 0 0 0 0 2 0 0 7 MGSmooth Level 2 10 1.0 8.3309e-02 1.1 8.54e+05 1.6 1.3e+04 1.6e+02 5.3e+01 2 0 14 1 7 3 0 17 1 20 250 MGResid Level 2 5 1.0 6.0241e-03 1.9 9.46e+04 1.6 1.8e+03 1.6e+02 0.0e+00 0 0 2 0 0 0 0 2 0 0 387 MGInterp Level 2 10 1.0 6.3021e-03 2.4 2.37e+04 1.4 1.2e+03 7.1e+01 0.0e+00 0 0 1 0 0 0 0 2 0 0 97 MGSmooth Level 3 10 1.0 1.6508e-01 1.5 5.79e+06 1.3 1.3e+04 5.7e+02 5.3e+01 3 2 14 3 7 6 3 17 4 20 975 MGResid Level 3 5 1.0 1.8781e-02 3.5 6.50e+05 1.3 1.8e+03 5.7e+02 0.0e+00 0 0 2 0 0 0 0 2 1 0 967 MGInterp Level 3 10 1.0 1.3515e-0212.8 1.62e+05 1.2 1.2e+03 2.4e+02 0.0e+00 0 0 1 0 0 0 0 2 0 0 345 MGSmooth Level 4 10 1.0 6.3244e-01 1.5 4.25e+07 1.1 1.3e+04 2.1e+03 5.3e+01 11 17 14 12 7 22 20 17 17 20 2001 MGResid Level 4 5 1.0 1.0258e-01 3.5 4.80e+06 1.1 1.8e+03 2.1e+03 0.0e+00 1 2 2 2 0 2 2 2 2 0 1398 MGInterp Level 4 10 1.0 3.4274e-02 3.4 1.20e+06 1.1 1.2e+03 8.7e+02 0.0e+00 0 0 1 0 0 1 1 2 1 0 1060 MGSmooth Level 5 10 1.0 1.5438e+00 1.6 1.29e+08 1.1 4.6e+03 2.3e+04 5.3e+01 27 53 5 46 7 53 64 6 62 20 2571 MGResid Level 5 5 1.0 1.3734e-01 1.7 9.67e+06 1.1 6.4e+02 2.3e+04 0.0e+00 2 4 1 6 0 4 5 1 9 0 2168 MGInterp Level 5 10 1.0 1.3824e-01 3.1 9.22e+06 1.0 1.2e+03 3.3e+03 0.0e+00 2 4 1 2 0 4 5 2 2 0 2075 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 1 1 564 0 Krylov Solver 13 13 160752 0 DMKSP interface 4 4 2592 0 Vector 91 171 35911200 0 Vector Scatter 25 25 26300 0 Matrix 45 45 36052932 0 Distributed Mesh 6 6 1412736 0 Bipartite Graph 12 12 9504 0 Viewer 2 1 728 0 Index Set 59 59 788488 0 IS L to G Mapping 6 6 695256 0 Preconditioner 13 13 11736 0 --- Event Stage 1: MG Apply Vector 80 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 5.00679e-07 Average time for MPI_Barrier(): 0.000676203 Average time for zero size MPI_Send(): 0.000420213 #PETSc Option Table entries: -da_refine 5 -dm_view -ksp_monitor -ksp_rtol 1.0e-7 -ksp_type cg -ksp_view -log_summary -mg_levels_ksp_type chebyshev -mg_levels_pc_type jacobi -pc_mg_galerkin -pc_mg_log -pc_mg_monitor -pc_type mg #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Thu Jun 13 15:51:55 2013 Configure options: --download-f-blas-lapack --download-hypre --download-mpich --with-cc=gcc --with-debugging=no --with-fc=gfortran PETSC_ARCH=linux-gnu-c-nodebug ----------------------------------------- Libraries compiled on Thu Jun 13 15:51:55 2013 on login1.ittc.ku.edu Machine characteristics: Linux-2.6.32-220.13.1.el6.x86_64-x86_64-with-redhat-6.2-Santiago Using PETSc directory: /bio/work1/zlwei/PETSc/petsc-dev Using PETSc arch: linux-gnu-c-nodebug ----------------------------------------- Using C compiler: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpif90 -fPIC -Wall -Wno-unused-variable -O ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/include -I/bio/work1/zlwei/PETSc/petsc-dev/include -I/bio/work1/zlwei/PETSc/petsc-dev/include -I/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/include ----------------------------------------- Using C linker: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpicc Using Fortran linker: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpif90 Using libraries: -Wl,-rpath,/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -L/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -lpetsc -Wl,-rpath,/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -L/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -lHYPRE -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lmpichcxx -lstdc++ -lflapack -lfblas -lX11 -lpthread -lmpichf90 -lgfortran -lm -lm -lmpichcxx -lstdc++ -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl ----------------------------------------- From rongliang.chan at gmail.com Tue Nov 5 23:18:02 2013 From: rongliang.chan at gmail.com (Rongliang Chen) Date: Wed, 06 Nov 2013 13:18:02 +0800 Subject: [petsc-users] Debug AOCreateBasic In-Reply-To: <8761sbpfq0.fsf@mcs.anl.gov> References: <52733BB5.5030203@gmail.com> <8761scheg2.fsf@mcs.anl.gov> <52734F7B.2060002@gmail.com> <8761sbpfq0.fsf@mcs.anl.gov> Message-ID: <5279D10A.8080204@gmail.com> Hi Jed, I tested my code on a locally 16-cores workstation. Both the MPICH and OPENMPI version work well for the 30 million unstructured meshes case (this workstation has 256G memory). My code is also valgrind-clean on this workstation. Best, Rongliang On 11/02/2013 12:52 PM, Jed Brown wrote: > Please keep replies on the list. > > Rongliang Chen writes: > >> Hi Jed, >> >> Thank you for your suggestions. >> >> The valgrind results are attached. Can you help me to check it? (I do >> not understand the result). > There is a ton of noise from Open MPI. You can use suppression files to > hide it. And then you have a number of places that are not inside the > MPI stack. I would install locally (on your laptop/workstation) using > MPICH and check that it is valgrind-clean for small sizes. > >> I am using a supercomputer where they do not have mpich (Are there other >> ways to check if the bug comes from openmpi?). > Easiest is to try a different implementation. > >> For your third >> suggestion, I have no idea how to do that. > Check the docs or ask the staff at your computing center. > >> Best, >> Rongliang >> >> On 11/01/2013 01:35 PM, Jed Brown wrote: >>> Rongliang Chen writes: >>> >>>> Hi there, >>>> >>>> My code died in the AOCreateBasic and the error messages are followed. >>>> Do you have any suggestions to debug this? >>> 1. Make sure your code is valgrind-clean for small sizes (to provide >>> more evidence that it is getting the right answer for the right reason). >>> >>> 2. Try MPICH instead to see if you error in the same place. >>> >>> 3. Set up your system to dump core on selected ranks. >>> >>>> Notes: >>>> 1. In this case, it has about 30,000,000 unstructured mesh and use 96 >>>> processors (I also tried 1024 processors and it has the same problem). >>>> 2. My code works well for a smaller case (about 25,000,000 unstructured >>>> meshes) . >>>> 3. I also check the memory usage of this case and it is very small >>>> because the solution stage does not start yet. >>>> >>>> Best, >>>> Rongliang >>>> >>>> -------------------------------------------------------------------------- >>>> MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD >>>> with errorcode 59. >>>> >>>> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. >>>> You may or may not see output from other processes, depending on >>>> exactly when Open MPI kills them. >>>> -------------------------------------------------------------------------- >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the >>>> batch system) has told this process to end >>>> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger >>>> [0]PETSC ERROR: or see >>>> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSC >>>> ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to >>>> find memory corruption errors >>>> [0]PETSC ERROR: likely location of problem given in stack below >>>> [0]PETSC ERROR: --------------------- Stack Frames >>>> ------------------------------------ >>>> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available, >>>> [0]PETSC ERROR: INSTEAD the line number of the start of the function >>>> [0]PETSC ERROR: is given. >>>> [0]PETSC ERROR: [0] AOCreate_Basic line 203 >>>> src/vec/is/ao/impls/basic/aobasic.c >>>> [0]PETSC ERROR: [0] AOSetType line 35 src/vec/is/ao/interface/aoreg.c >>>> [0]PETSC ERROR: [0] AOCreateBasicIS line 380 >>>> src/vec/is/ao/impls/basic/aobasic.c >>>> [0]PETSC ERROR: [0] AOCreateBasic line 335 >>>> src/vec/is/ao/impls/basic/aobasic.c >>>> [0]PETSC ERROR: [0] DataPartitionVertices_Block line 1634 >>>> /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/readbinary3d.c >>>> [0]PETSC ERROR: [0] ReadBinary line 184 >>>> /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/readbinary3d.c >>>> [0]PETSC ERROR: [0] LoadGrid line 720 >>>> /projects/ronglian/soft/3Dfluid_new/3DWindturbine/WindturbineFor3.4/codes/loadgrid3d.c >>>> [0]PETSC ERROR: --------------------- Error Message >>>> ------------------------------------ >>>> [0]PETSC ERROR: Signal received! >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: Petsc Development GIT revision: >>>> ee17fca9fd6ac48e6579ef235144daafbb22b801 GIT Date: 2013-10-23 14:21:20 >>>> -0500 >>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: ./fsi3d on a Janus-debug-64bit named node0880 by >>>> ronglian Thu Oct 31 21:58:20 2013 >>>> [0]PETSC ERROR: Libraries linked from >>>> /projects/ronglian/soft/petsc-dev-latest/Janus-debug-64bit/lib >>>> [0]PETSC ERROR: Configure run at Thu Oct 24 21:24:31 2013 >>>> [0]PETSC ERROR: Configure options --known-level1-dcache-size=32768 >>>> --known-level1-dcache-linesize=64 --known-level1-dcache-assoc=8 >>>> --known-memcmp-ok=1 --known-sizeof-char=1 --known-sizeof-void-p=8 >>>> --known-sizeof-short=2 --known-sizeof-int=4 --known-sizeof-long=8 >>>> --known-sizeof-long-long=8 --known-sizeof-float=4 >>>> --known-sizeof-double=8 --known-sizeof-size_t=8 --known-bits-per-byte=8 >>>> --known-sizeof-MPI_Comm=8 --known-sizeof-MPI_Fint=4 >>>> --known-mpi-long-double=1 --known-mpi-c-double-complex=0 >>>> --download-blacs=1 --download-f-blas-lapack=1 --download-metis=1 >>>> --download-parmetis=1 --download-scalapack=1 --download-superlu_dist=1 >>>> --known-mpi-shared-libraries=0 --with-64-bit-indices --with-batch=1 >>>> --with-mpi-shared=1 --download-exodusii=1 --download-hdf5=1 >>>> --download-netcdf=1 --known-64-bit-blas-indices --with-debugging=1 >>>> COPTFLAGS="-O0 -g" >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: User provided function() line 0 in unknown file >>>> [1]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [1]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the >>>> batch system) has told this process to end >>>> [1]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger >>>> [1]PETSC ERROR: or see >>>> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[1]PETSC >>>> ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to >>>> find memory corruption errors >>>> [1]PETSC ERROR: likely location of problem given in stack below >> --15577-- WARNING: Serious error when reading debug info >> --15577-- When reading debug info from /projects/ronglian/soft/petsc-dev-latest/Janus-debug-64bit/lib/libnetcdf.so.7.2.0: >> --15577-- DWARF line info appears to be corrupt - the section is too small >> --15577-- Warning: DWARF2 CFI reader: unhandled DW_OP_ opcode 0x8 >> --15577-- Warning: DWARF2 CFI reader: unhandled DW_OP_ opcode 0x8 >> --15577-- Warning: DWARF2 CFI reader: unhandled DW_OP_ opcode 0x8 >> --15577-- Warning: DWARF2 CFI reader: unhandled DW_OP_ opcode 0x8 >> --15577-- Warning: DWARF2 CFI reader: unhandled DW_OP_ opcode 0x8 >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0278: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9755825: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9755967: rdma_create_event_channel (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9091663: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:313) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xAF13D2B: _itoa_word (in /lib64/libc-2.12.so) >> ==15577== by 0xAF150F5: vfprintf (in /lib64/libc-2.12.so) >> ==15577== by 0xAFCF2BF: __vsnprintf_chk (in /lib64/libc-2.12.so) >> ==15577== by 0xAFCF1F9: __snprintf_chk (in /lib64/libc-2.12.so) >> ==15577== by 0xF1C2F06: mlx4_query_device (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C02CE: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9755825: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9755967: rdma_create_event_channel (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9091663: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:313) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xAF13D35: _itoa_word (in /lib64/libc-2.12.so) >> ==15577== by 0xAF150F5: vfprintf (in /lib64/libc-2.12.so) >> ==15577== by 0xAFCF2BF: __vsnprintf_chk (in /lib64/libc-2.12.so) >> ==15577== by 0xAFCF1F9: __snprintf_chk (in /lib64/libc-2.12.so) >> ==15577== by 0xF1C2F06: mlx4_query_device (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C02CE: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9755825: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9755967: rdma_create_event_channel (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9091663: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:313) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0306: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9755825: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9755967: rdma_create_event_channel (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9091663: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:313) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x975656A: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x90916AE: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:330) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffdfac is on thread 1's stack >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x9755C0E: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9755C5D: rdma_destroy_id (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9091766: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:370) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffdf80 is on thread 1's stack >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9755C62: rdma_destroy_id (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9091766: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:370) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9755C8A: rdma_destroy_id (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9091766: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:370) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x9756207: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x90916AE: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:330) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffde70 is on thread 1's stack >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x975621C: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x90916AE: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:330) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9756368: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x90916AE: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:330) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x975637F: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x90916AE: mca_btl_openib_build_rdma_addr_list (btl_openib_ip.c:330) >> ==15577== by 0x90961A9: rdmacm_component_init (btl_openib_connect_rdmacm.c:2009) >> ==15577== by 0x90923D3: ompi_btl_openib_connect_base_init (btl_openib_connect_base.c:204) >> ==15577== by 0x9080A84: btl_openib_component_init (btl_openib_component.c:2535) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0278: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x90813A7: btl_openib_component_init (btl_openib_component.c:1663) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0306: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x90813A7: btl_openib_component_init (btl_openib_component.c:1663) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90815CC: btl_openib_component_init (btl_openib_component.c:1248) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9081829: btl_openib_component_init (btl_openib_component.c:1685) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x908E768: ompi_btl_openib_ini_query (btl_openib_ini.c:176) >> ==15577== by 0x9081854: btl_openib_component_init (btl_openib_component.c:1694) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x908E76E: ompi_btl_openib_ini_query (btl_openib_ini.c:176) >> ==15577== by 0x9081854: btl_openib_component_init (btl_openib_component.c:1694) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9081AC8: btl_openib_component_init (btl_openib_component.c:1896) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9081AB8: btl_openib_component_init (btl_openib_component.c:1905) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9081B01: btl_openib_component_init (btl_openib_component.c:1910) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x902DA00: init_one_port (btl_openib_component.c:681) >> ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x902DB0E: init_one_port (btl_openib_component.c:712) >> ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x902DBBF: init_one_port (btl_openib_component.c:758) >> ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x902E062: init_one_port (btl_openib_component.c:778) >> ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xAF13D2B: _itoa_word (in /lib64/libc-2.12.so) >> ==15577== by 0xAF150F5: vfprintf (in /lib64/libc-2.12.so) >> ==15577== by 0xAF391C8: vsprintf (in /lib64/libc-2.12.so) >> ==15577== by 0xAF1F167: sprintf (in /lib64/libc-2.12.so) >> ==15577== by 0x902DE57: init_one_port (btl_openib_component.c:824) >> ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xAF13D35: _itoa_word (in /lib64/libc-2.12.so) >> ==15577== by 0xAF150F5: vfprintf (in /lib64/libc-2.12.so) >> ==15577== by 0xAF391C8: vsprintf (in /lib64/libc-2.12.so) >> ==15577== by 0xAF1F167: sprintf (in /lib64/libc-2.12.so) >> ==15577== by 0x902DE57: init_one_port (btl_openib_component.c:824) >> ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x902DF1D: init_one_port (btl_openib_component.c:856) >> ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x902DF80: init_one_port (btl_openib_component.c:888) >> ==15577== by 0x9081C3D: btl_openib_component_init (btl_openib_component.c:1918) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9092BC4: oob_component_query (btl_openib_connect_oob.c:131) >> ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) >> ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x975656A: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9098B22: rdmacm_component_query (btl_openib_connect_rdmacm.c:1859) >> ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) >> ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffdf7c is on thread 1's stack >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x9756207: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9098B22: rdmacm_component_query (btl_openib_connect_rdmacm.c:1859) >> ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) >> ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffde40 is on thread 1's stack >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x975621C: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9098B22: rdmacm_component_query (btl_openib_connect_rdmacm.c:1859) >> ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) >> ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9756368: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9756579: rdma_bind_addr (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9098B22: rdmacm_component_query (btl_openib_connect_rdmacm.c:1859) >> ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) >> ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9091545: mca_btl_openib_rdma_get_ipv4addr (btl_openib_ip.c:138) >> ==15577== by 0x9098B5E: rdmacm_component_query (btl_openib_connect_rdmacm.c:1702) >> ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) >> ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x9756470: rdma_listen (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9098D41: rdmacm_component_query (btl_openib_connect_rdmacm.c:1881) >> ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) >> ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffdf98 is on thread 1's stack >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x9756207: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x975648C: rdma_listen (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9098D41: rdmacm_component_query (btl_openib_connect_rdmacm.c:1881) >> ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) >> ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffde80 is on thread 1's stack >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x975621C: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x975648C: rdma_listen (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9098D41: rdmacm_component_query (btl_openib_connect_rdmacm.c:1881) >> ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) >> ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9756368: ??? (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x975648C: rdma_listen (in /usr/lib64/librdmacm.so.1.0.0) >> ==15577== by 0x9098D41: rdmacm_component_query (btl_openib_connect_rdmacm.c:1881) >> ==15577== by 0x90925C2: ompi_btl_openib_connect_base_select_for_local_port (btl_openib_connect_base.c:255) >> ==15577== by 0x9082878: btl_openib_component_init (btl_openib_component.c:2860) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x9082D96: btl_openib_component_init (btl_openib_component.c:1085) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0xda015cc is 268 bytes inside a block of size 8,784 alloc'd >> ==15577== at 0x4C2694F: calloc (vg_replace_malloc.c:593) >> ==15577== by 0xF1C0126: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99623AA: ibv_open_device (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x90813A7: btl_openib_component_init (btl_openib_component.c:1663) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x908EA9B: btl_openib_async_command_done (btl_openib_async.c:483) >> ==15577== by 0x9082DAE: btl_openib_component_init (btl_openib_component.c:1092) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9082DBF: btl_openib_component_init (btl_openib_component.c:1110) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x995EEB6: ibv_cmd_reg_mr (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1C2D79: mlx4_reg_mr (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965782: ibv_reg_mr (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9080016: openib_reg_mr (btl_openib_component.c:602) >> ==15577== by 0x90ED046: register_cache_bypass (mpool_rdma_module.c:172) >> ==15577== by 0x90ED728: mca_mpool_rdma_alloc (mpool_rdma_module.c:115) >> ==15577== by 0x902FAA7: ompi_free_list_grow (ompi_free_list.c:203) >> ==15577== by 0x9082EA4: btl_openib_component_init (btl_openib_component.c:1150) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffdf18 is on thread 1's stack >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x907E6A4: mca_btl_openib_get_transport_type (btl_openib.c:436) >> ==15577== by 0x9080188: btl_openib_modex_send (btl_openib_component.c:386) >> ==15577== by 0x90829AD: btl_openib_component_init (btl_openib_component.c:2925) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x907E6CA: mca_btl_openib_get_transport_type (btl_openib.c:459) >> ==15577== by 0x9080188: btl_openib_modex_send (btl_openib_component.c:386) >> ==15577== by 0x90829AD: btl_openib_component_init (btl_openib_component.c:2925) >> ==15577== by 0x9072AB3: mca_btl_base_select (btl_base_select.c:111) >> ==15577== by 0x9071D90: mca_bml_r2_component_init (bml_r2_component.c:85) >> ==15577== by 0x9070655: mca_bml_base_init (bml_base_init.c:71) >> ==15577== by 0x910E5BD: mca_pml_ob1_component_init (pml_ob1_component.c:181) >> ==15577== by 0x91077A2: mca_pml_base_select (pml_base_select.c:132) >> ==15577== by 0x9044897: ompi_mpi_init (ompi_mpi_init.c:521) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param writev(vector[...]) points to uninitialised byte(s) >> ==15577== at 0xAFB0377: writev (in /lib64/libc-2.12.so) >> ==15577== by 0x918AFE2: mca_oob_tcp_msg_send_handler (oob_tcp_msg.c:249) >> ==15577== by 0x918BD4C: mca_oob_tcp_peer_send (oob_tcp_peer.c:204) >> ==15577== by 0x918EEFB: mca_oob_tcp_send_nb (oob_tcp_send.c:167) >> ==15577== by 0x91A8EEE: orte_rml_oob_send (rml_oob_send.c:136) >> ==15577== by 0x91A9468: orte_rml_oob_send_buffer (rml_oob_send.c:270) >> ==15577== by 0x917549F: modex (grpcomm_bad_module.c:573) >> ==15577== by 0x9044943: ompi_mpi_init (ompi_mpi_init.c:541) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0xdb504c1 is 161 bytes inside a block of size 512 alloc'd >> ==15577== at 0x4C2884E: realloc (vg_replace_malloc.c:662) >> ==15577== by 0x91C0B17: opal_dss_buffer_extend (dss_internal_functions.c:63) >> ==15577== by 0x91C0E4D: opal_dss_copy_payload (dss_load_unload.c:164) >> ==15577== by 0x916F255: orte_grpcomm_base_pack_modex_entries (grpcomm_base_modex.c:861) >> ==15577== by 0x917537E: modex (grpcomm_bad_module.c:563) >> ==15577== by 0x9044943: ompi_mpi_init (ompi_mpi_init.c:541) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x907E6A4: mca_btl_openib_get_transport_type (btl_openib.c:436) >> ==15577== by 0x907EE11: mca_btl_openib_add_procs (btl_openib.c:475) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x907E6CA: mca_btl_openib_get_transport_type (btl_openib.c:459) >> ==15577== by 0x907EE11: mca_btl_openib_add_procs (btl_openib.c:475) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C2B16: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C2B9C: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C2BA6: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param mmap(length) contains uninitialised byte(s) >> ==15577== at 0xAFB4C6A: mmap (in /lib64/libc-2.12.so) >> ==15577== by 0xF1BF01A: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x996412E: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9964218: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9964224: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x99639A3: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99642FA: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x99639B1: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99642FA: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param madvise(length) contains uninitialised byte(s) >> ==15577== at 0xAFB4D57: madvise (in /lib64/libc-2.12.so) >> ==15577== by 0x9964449: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x996426D: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BF216: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) >> ==15577== by 0xF1BF1CC: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) >> ==15577== by 0xF1BF1CC: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B11D: memset (mc_replace_strmem.c:1007) >> ==15577== by 0xF1BF1CC: mlx4_alloc_cq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2BC8: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907ED89: mca_btl_openib_add_procs (btl_openib.c:405) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C2B16: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907EDBA: mca_btl_openib_add_procs (btl_openib.c:410) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C2B9C: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907EDBA: mca_btl_openib_add_procs (btl_openib.c:410) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C2BA6: mlx4_create_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x996544E: ibv_create_cq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907E63F: adjust_cq (btl_openib.c:163) >> ==15577== by 0x907EDBA: mca_btl_openib_add_procs (btl_openib.c:410) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x9960C21: ibv_cmd_create_srq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1C2869: mlx4_create_srq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99652BB: ibv_create_srq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907F359: mca_btl_openib_add_procs (btl_openib.c:255) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffe138 is on thread 1's stack >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x9960766: ibv_cmd_modify_srq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1C2731: mlx4_modify_srq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x907F92A: mca_btl_openib_add_procs (btl_openib.c:279) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffe1a8 is on thread 1's stack >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x995FDD1: ibv_cmd_destroy_srq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1C266D: mlx4_destroy_srq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x907F3FC: mca_btl_openib_add_procs (btl_openib.c:285) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== Address 0x7feffe160 is on thread 1's stack >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x995FE0B: ibv_cmd_destroy_srq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1C266D: mlx4_destroy_srq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x907F3FC: mca_btl_openib_add_procs (btl_openib.c:285) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x99641A4: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C1A79: mlx4_alloc_srq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2807: mlx4_create_srq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99652BB: ibv_create_srq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907F54E: mca_btl_openib_add_procs (btl_openib.c:334) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x99641BB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C1A79: mlx4_alloc_srq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2807: mlx4_create_srq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99652BB: ibv_create_srq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907F54E: mca_btl_openib_add_procs (btl_openib.c:334) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x99641CD: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1BF02E: mlx4_alloc_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C1A79: mlx4_alloc_srq_buf (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C2807: mlx4_create_srq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x99652BB: ibv_create_srq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x907F54E: mca_btl_openib_add_procs (btl_openib.c:334) >> ==15577== by 0x9070D2A: mca_bml_r2_add_procs (bml_r2.c:208) >> ==15577== by 0x910CB45: mca_pml_ob1_add_procs (pml_ob1.c:326) >> ==15577== by 0x9044F93: ompi_mpi_init (ompi_mpi_init.c:741) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1BF64A: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9174254: barrier (grpcomm_bad_module.c:277) >> ==15577== by 0x90450C1: ompi_mpi_init (ompi_mpi_init.c:783) >> ==15577== by 0x905A56D: PMPI_Init_thread (pinit_thread.c:84) >> ==15577== by 0x4F807D6: PetscInitialize (pinit.c:689) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C2F5B: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C2F60: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C2FCA: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C2FCF: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x9960AA8: ibv_cmd_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1C31A5: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== Address 0x7feffaec8 is on thread 1's stack >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C089F: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C26906: calloc (vg_replace_malloc.c:593) >> ==15577== by 0xF1C0902: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C0906: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C091A: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C08B3: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C08BE: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C08C3: mlx4_store_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C31BD: mlx4_create_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9965161: ibv_create_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092F18: qp_create_all (btl_openib_connect_oob.c:486) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x995F923: ibv_cmd_modify_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1C24BF: mlx4_modify_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9964C73: ibv_modify_qp (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9092FA7: qp_create_all (btl_openib_connect_oob.c:514) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== Address 0x7feffaef8 is on thread 1's stack >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0B91: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C0C4E: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9087EE4: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0C62: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9087EE4: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0CD4: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9087EE4: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C0CAF: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9087EE4: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C0D0D: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9087EE4: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0B91: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C0C4E: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x908824E: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0C62: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x908824E: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C0CAF: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x908824E: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0CD4: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x908824E: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C0D0D: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x908824E: mca_btl_openib_endpoint_post_recvs (verbs.h:1120) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Syscall param write(buf) points to uninitialised byte(s) >> ==15577== at 0x6CC24ED: ??? (in /lib64/libpthread-2.12.so) >> ==15577== by 0x9960766: ibv_cmd_modify_srq (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0xF1C2731: mlx4_modify_srq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x908446D: mca_btl_openib_post_srr (btl_openib_component.c:3760) >> ==15577== by 0x9087E5F: mca_btl_openib_endpoint_post_recvs (btl_openib_endpoint.c:493) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== Address 0x7feffadf8 is on thread 1's stack >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x99641A4: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9965746: ibv_reg_mr (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9080016: openib_reg_mr (btl_openib_component.c:602) >> ==15577== by 0x90ED046: register_cache_bypass (mpool_rdma_module.c:172) >> ==15577== by 0x90ED728: mca_mpool_rdma_alloc (mpool_rdma_module.c:115) >> ==15577== by 0x902FAA7: ompi_free_list_grow (ompi_free_list.c:203) >> ==15577== by 0x9084273: mca_btl_openib_post_srr (ompi_free_list.h:253) >> ==15577== by 0x9087E5F: mca_btl_openib_endpoint_post_recvs (btl_openib_endpoint.c:493) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x99641BB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9965746: ibv_reg_mr (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9080016: openib_reg_mr (btl_openib_component.c:602) >> ==15577== by 0x90ED046: register_cache_bypass (mpool_rdma_module.c:172) >> ==15577== by 0x90ED728: mca_mpool_rdma_alloc (mpool_rdma_module.c:115) >> ==15577== by 0x902FAA7: ompi_free_list_grow (ompi_free_list.c:203) >> ==15577== by 0x9084273: mca_btl_openib_post_srr (ompi_free_list.h:253) >> ==15577== by 0x9087E5F: mca_btl_openib_endpoint_post_recvs (btl_openib_endpoint.c:493) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x99641CD: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x99645CB: ??? (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9965746: ibv_reg_mr (in /usr/lib64/libibverbs.so.1.0.0) >> ==15577== by 0x9080016: openib_reg_mr (btl_openib_component.c:602) >> ==15577== by 0x90ED046: register_cache_bypass (mpool_rdma_module.c:172) >> ==15577== by 0x90ED728: mca_mpool_rdma_alloc (mpool_rdma_module.c:115) >> ==15577== by 0x902FAA7: ompi_free_list_grow (ompi_free_list.c:203) >> ==15577== by 0x9084273: mca_btl_openib_post_srr (ompi_free_list.h:253) >> ==15577== by 0x9087E5F: mca_btl_openib_endpoint_post_recvs (btl_openib_endpoint.c:493) >> ==15577== by 0x909311F: qp_create_all (btl_openib_connect_oob.c:434) >> ==15577== by 0x9093EB4: rml_recv_cb (btl_openib_connect_oob.c:246) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== >> ==15577== >> ==15577== More than 100 errors detected. Subsequent errors >> ==15577== will still be recorded, but in less detail than before. >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9093D3C: rml_recv_cb (btl_openib_connect_oob.c:784) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9093D42: rml_recv_cb (btl_openib_connect_oob.c:785) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1BF6BB: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1BF6CE: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C07EF: mlx4_find_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BFA2F: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C07FD: mlx4_find_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BFA2F: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C0802: mlx4_find_qp (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1BFA2F: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1BF8D0: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1BF909: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1BF9C8: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1BF659: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9084584: btl_openib_handle_incoming (btl_openib_component.c:1374) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x90845A2: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9111AF0: mca_pml_ob1_recv_frag_callback_match (opal_pointer_array.h:130) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111B34: mca_pml_ob1_recv_frag_callback_match (opal_pointer_array.h:134) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111B5B: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:157) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9111B61: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:157) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111B67: mca_pml_ob1_recv_frag_callback_match (opal_list.h:322) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111B78: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:167) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91115EF: match_incomming.isra.2 (opal_list.h:197) >> ==15577== by 0x9111B93: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:470) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91116A5: match_incomming.isra.2 (pml_ob1_recvfrag.c:445) >> ==15577== by 0x9111B93: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:470) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91116F0: match_incomming.isra.2 (opal_list.h:378) >> ==15577== by 0x9111B93: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:470) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111BA7: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:474) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90845D2: btl_openib_handle_incoming (btl_openib_component.c:3104) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90845DE: btl_openib_handle_incoming (btl_openib_component.c:3119) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9084614: btl_openib_handle_incoming (btl_openib_component.c:1382) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90848DF: btl_openib_handle_incoming (btl_openib_component.c:3170) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1BFC00: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1BF710: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x902E678: append_frag_to_list (opal_list.h:431) >> ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66986: PetscOptionsInsertFile (options.c:471) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66A34: PetscOptionsInsertFile (options.c:473) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66A3E: PetscOptionsInsertFile (options.c:473) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66B04: PetscOptionsInsertFile (options.c:479) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66DB3: PetscOptionsInsertFile (options.c:493) >> ==15577== by 0x4F6805E: PetscOptionsInsert (options.c:624) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115B7B: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:896) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115C39: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1020) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115C41: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1020) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115C49: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1020) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F668F7: PetscOptionsInsertFile (options.c:467) >> ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66986: PetscOptionsInsertFile (options.c:471) >> ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66A34: PetscOptionsInsertFile (options.c:473) >> ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66A3E: PetscOptionsInsertFile (options.c:473) >> ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66B04: PetscOptionsInsertFile (options.c:479) >> ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66DB3: PetscOptionsInsertFile (options.c:493) >> ==15577== by 0x4F680D1: PetscOptionsInsert (options.c:625) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66986: PetscOptionsInsertFile (options.c:471) >> ==15577== by 0x4F68144: PetscOptionsInsert (options.c:626) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66A34: PetscOptionsInsertFile (options.c:473) >> ==15577== by 0x4F68144: PetscOptionsInsert (options.c:626) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66A3E: PetscOptionsInsertFile (options.c:473) >> ==15577== by 0x4F68144: PetscOptionsInsert (options.c:626) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66B04: PetscOptionsInsertFile (options.c:479) >> ==15577== by 0x4F68144: PetscOptionsInsert (options.c:626) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F66DB3: PetscOptionsInsertFile (options.c:493) >> ==15577== by 0x4F68144: PetscOptionsInsert (options.c:626) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9115C1C: mca_pml_ob1_recv_req_start (opal_list.h:376) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4F682EE: PetscOptionsInsert (options.c:638) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F6834E: PetscOptionsInsert (options.c:639) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F68415: PetscOptionsInsert (options.c:643) >> ==15577== by 0x4F8114A: PetscInitialize (pinit.c:778) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1BF814: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4EF1345: PetscWorldIsSingleHost (pdisplay.c:94) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1BFA5F: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x4EF1345: PetscWorldIsSingleHost (pdisplay.c:94) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C29BDC: strcmp (mc_replace_strmem.c:725) >> ==15577== by 0x4EE8F19: PetscStrcmp (str.c:440) >> ==15577== by 0x4EF13BA: PetscWorldIsSingleHost (pdisplay.c:95) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C29BF0: strcmp (mc_replace_strmem.c:725) >> ==15577== by 0x4EE8F19: PetscStrcmp (str.c:440) >> ==15577== by 0x4EF13BA: PetscWorldIsSingleHost (pdisplay.c:95) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE8F21: PetscStrcmp (str.c:441) >> ==15577== by 0x4EF13BA: PetscWorldIsSingleHost (pdisplay.c:95) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0B91: ??? (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0xF1C0E28: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x907D977: mca_btl_openib_sendi (verbs.h:1111) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C13DC: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x907D977: mca_btl_openib_sendi (verbs.h:1111) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C13DC: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x908701E: mca_btl_openib_endpoint_post_send (verbs.h:1111) >> ==15577== by 0x9088458: mca_btl_openib_endpoint_connected (btl_openib_endpoint.c:689) >> ==15577== by 0x9093E0C: rml_recv_cb (btl_openib_connect_oob.c:914) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C144B: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x908701E: mca_btl_openib_endpoint_post_send (verbs.h:1111) >> ==15577== by 0x9088458: mca_btl_openib_endpoint_connected (btl_openib_endpoint.c:689) >> ==15577== by 0x9093E0C: rml_recv_cb (btl_openib_connect_oob.c:914) >> ==15577== by 0x91A826B: orte_rml_recv_msg_callback (rml_oob_recv.c:62) >> ==15577== by 0x918B7DE: mca_oob_tcp_msg_recv_complete (oob_tcp_msg.c:530) >> ==15577== by 0x918CC8C: mca_oob_tcp_peer_recv_handler (oob_tcp_peer.c:925) >> ==15577== by 0x91C70D3: opal_event_base_loop (event.c:686) >> ==15577== by 0x91E6C88: opal_progress (opal_progress.c:189) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1BF728: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1BF7DE: mlx4_poll_cq (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x90852FB: poll_device (verbs.h:976) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x4EF1479: PetscWorldIsSingleHost (pdisplay.c:99) >> ==15577== by 0x4EF193F: PetscSetDisplay (pdisplay.c:123) >> ==15577== by 0x4F787C1: PetscOptionsCheckInitial_Private (init.c:323) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C144B: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x907D977: mca_btl_openib_sendi (verbs.h:1111) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4FA3DC7: PetscLogBegin_Private (plog.c:248) >> ==15577== by 0x4FA4103: PetscLogBegin (plog.c:292) >> ==15577== by 0x4F7A5FF: PetscOptionsCheckInitial_Private (init.c:491) >> ==15577== by 0x4F81298: PetscInitialize (pinit.c:788) >> ==15577== by 0x405831: main (fsi3d.c:27) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x441A29: DataRead (readbinary3d.c:678) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x441B80: DataRead (readbinary3d.c:696) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x441B98: DataRead (readbinary3d.c:697) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x441BD9: DataRead (readbinary3d.c:697) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x441BD9: DataRead (readbinary3d.c:697) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x441BD9: DataRead (readbinary3d.c:697) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x441C05: DataRead (readbinary3d.c:698) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x441C46: DataRead (readbinary3d.c:698) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F231A2: PetscMallocValidate (mtr.c:139) >> ==15577== by 0x4F235E5: PetscTrMallocDefault (mtr.c:182) >> ==15577== by 0x441C46: DataRead (readbinary3d.c:698) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x441C46: DataRead (readbinary3d.c:698) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x441C46: DataRead (readbinary3d.c:698) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x905061F: PMPI_Bcast (pbcast.c:75) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90506A3: PMPI_Bcast (pbcast.c:101) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90A0E8D: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:251) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115BA2: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.h:200) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DB968: opal_convertor_prepare_for_recv (opal_convertor.c:523) >> ==15577== by 0x9115BD7: mca_pml_ob1_recv_req_start (opal_convertor.h:258) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90A92FA: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:224) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9111BE8: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:194) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DB703: opal_convertor_unpack (opal_convertor.c:282) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3B4: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A435: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A43B: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A450: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A456: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A47E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9111C64: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvreq.h:168) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x904365C: ompi_request_default_wait (condition.h:99) >> ==15577== by 0x90A97BB: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:238) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441F6B: DataRead (readbinary3d.c:712) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x905061F: PMPI_Bcast (pbcast.c:75) >> ==15577== by 0x441FF9: DataRead (readbinary3d.c:713) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90506A3: PMPI_Bcast (pbcast.c:101) >> ==15577== by 0x441FF9: DataRead (readbinary3d.c:713) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91139FC: mca_pml_ob1_recv_request_progress_match (opal_convertor.h:284) >> ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441FF9: DataRead (readbinary3d.c:713) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DB703: opal_convertor_unpack (opal_convertor.c:282) >> ==15577== by 0x9113A26: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.c:632) >> ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441FF9: DataRead (readbinary3d.c:713) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9113BE8: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.h:168) >> ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A92EC: ompi_coll_tuned_bcast_intra_generic (coll_tuned_bcast.c:219) >> ==15577== by 0x90A9C08: ompi_coll_tuned_bcast_intra_binomial (coll_tuned_bcast.c:369) >> ==15577== by 0x90A0F3B: ompi_coll_tuned_bcast_intra_dec_fixed (coll_tuned_decision_fixed.c:254) >> ==15577== by 0x90B0DF8: mca_coll_sync_bcast (coll_sync_bcast.c:44) >> ==15577== by 0x90506D7: PMPI_Bcast (pbcast.c:110) >> ==15577== by 0x441FF9: DataRead (readbinary3d.c:713) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44205D: DataRead (readbinary3d.c:718) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x442208: DataRead (readbinary3d.c:719) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44208D: DataRead (readbinary3d.c:731) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4420E8: DataRead (readbinary3d.c:733) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x442143: DataRead (readbinary3d.c:735) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x445163: DataReadAndSplitGeneric (readbinary3d.c:1068) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x445163: DataReadAndSplitGeneric (readbinary3d.c:1068) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x445163: DataReadAndSplitGeneric (readbinary3d.c:1068) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x905D67C: PMPI_Recv (precv.c:53) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x905D7EC: PMPI_Recv (precv.c:54) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115BA2: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.h:200) >> ==15577== by 0x910F5C1: mca_pml_ob1_recv (pml_ob1_irecv.c:104) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9111793: mca_pml_ob1_recv_frag_match (opal_pointer_array.h:130) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91117EC: mca_pml_ob1_recv_frag_match (opal_pointer_array.h:134) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x911180E: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:606) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9111816: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:607) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111836: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:618) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91115E0: match_incomming.isra.2 (opal_list.h:322) >> ==15577== by 0x9111854: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:470) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91115EF: match_incomming.isra.2 (opal_list.h:197) >> ==15577== by 0x9111854: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:470) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91116A5: match_incomming.isra.2 (pml_ob1_recvfrag.c:445) >> ==15577== by 0x9111854: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:470) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91116F0: match_incomming.isra.2 (opal_list.h:378) >> ==15577== by 0x9111854: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:470) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x911186A: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:474) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9112D10: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:258) >> ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9112DEB: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:265) >> ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9112E16: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:271) >> ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9112E2E: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:282) >> ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9112D3C: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:302) >> ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91145B8: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9114645: mca_pml_ob1_recv_request_progress_rndv (opal_convertor.h:284) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DB703: opal_convertor_unpack (opal_convertor.c:282) >> ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9114526: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.h:185) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91118A2: mca_pml_ob1_recv_frag_match (opal_list.h:322) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9115644: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115655: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9115678: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:284) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115689: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:284) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x911568F: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:294) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115696: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:294) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91DB694: opal_convertor_unpack (opal_convertor.c:266) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91DB6B0: opal_convertor_unpack (opal_convertor.c:266) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91DB6CC: opal_convertor_unpack (opal_convertor.c:276) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DB703: opal_convertor_unpack (opal_convertor.c:282) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3A0: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3B4: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A435: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A456: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91DB754: opal_convertor_unpack (opal_convertor.c:290) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91156B3: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:465) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91156CA: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:185) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91156D1: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:185) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91156EB: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:467) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3A0: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91DB785: opal_convertor_unpack (opal_convertor.c:296) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91156D3: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:58) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91157D8: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:152) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x911581F: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:161) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9115834: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:166) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115854: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:168) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9115861: mca_pml_ob1_recv_request_progress_frag (request.h:395) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9115876: mca_pml_ob1_recv_request_progress_frag (request.h:399) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44223B: DataRead (readbinary3d.c:756) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) >> ==15577== by 0x4422BE: DataRead (readbinary3d.c:757) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x445106: DataReadAndSplitGeneric (readbinary3d.c:1062) >> ==15577== by 0x4422BE: DataRead (readbinary3d.c:757) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) >> ==15577== by 0x4422BE: DataRead (readbinary3d.c:757) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) >> ==15577== by 0x442341: DataRead (readbinary3d.c:758) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) >> ==15577== by 0x442341: DataRead (readbinary3d.c:758) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) >> ==15577== by 0x4423C4: DataRead (readbinary3d.c:759) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) >> ==15577== by 0x4423C4: DataRead (readbinary3d.c:759) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4426CC: DataRead (readbinary3d.c:785) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) >> ==15577== by 0x442702: DataRead (readbinary3d.c:786) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) >> ==15577== by 0x442702: DataRead (readbinary3d.c:786) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) >> ==15577== by 0x442788: DataRead (readbinary3d.c:787) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) >> ==15577== by 0x442788: DataRead (readbinary3d.c:787) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4450E4: DataReadAndSplitGeneric (readbinary3d.c:1059) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x445106: DataReadAndSplitGeneric (readbinary3d.c:1062) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44512A: DataReadAndSplitGeneric (readbinary3d.c:1068) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084581: btl_openib_handle_incoming (btl_openib_component.c:1374) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9084584: btl_openib_handle_incoming (btl_openib_component.c:1374) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x908458A: btl_openib_handle_incoming (btl_openib_component.c:3096) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x90845A2: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9111F8B: mca_pml_ob1_recv_frag_callback_rndv (pml_ob1_recvfrag.c:254) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x911177D: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:568) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91117F5: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:585) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111841: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:470) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91144C0: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:564) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9112D09: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:258) >> ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9112DE0: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:266) >> ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9112E26: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:282) >> ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9112D52: mca_pml_ob1_recv_request_ack (pml_ob1_recvreq.c:307) >> ==15577== by 0x91144FE: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:568) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9114502: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:574) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91145F7: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A43B: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A450: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A47E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x911454E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:597) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084E80: btl_openib_handle_incoming (btl_openib_component.c:3099) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90845D2: btl_openib_handle_incoming (btl_openib_component.c:3104) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x90845D8: btl_openib_handle_incoming (btl_openib_component.c:3119) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90845DE: btl_openib_handle_incoming (btl_openib_component.c:3119) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90848DF: btl_openib_handle_incoming (btl_openib_component.c:3170) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084A3F: btl_openib_handle_incoming (btl_openib_endpoint.h:464) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x902E6AE: opal_convertor_set_position.part.8 (opal_convertor.h:297) >> ==15577== by 0x9115A4E: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:156) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x902E6D8: opal_convertor_set_position.part.8 (opal_convertor.h:303) >> ==15577== by 0x9115A4E: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:156) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91156F9: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:467) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x44524E: DataReadAndSplitGeneric (readbinary3d.c:1069) >> ==15577== by 0x44280E: DataRead (readbinary3d.c:791) >> ==15577== by 0x43D22A: ReadBinary (readbinary3d.c:206) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x43D2C0: ReadBinary (readbinary3d.c:214) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44631F: DataPartitionElements (readbinary3d.c:1190) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x446355: DataPartitionElements (readbinary3d.c:1190) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x446355: DataPartitionElements (readbinary3d.c:1190) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x446355: DataPartitionElements (readbinary3d.c:1190) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4463C6: DataPartitionElements (readbinary3d.c:1191) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x4463FC: DataPartitionElements (readbinary3d.c:1191) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x4463FC: DataPartitionElements (readbinary3d.c:1191) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4463FC: DataPartitionElements (readbinary3d.c:1191) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x446597: DataPartitionElements (readbinary3d.c:1196) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4464ED: DataPartitionElements (readbinary3d.c:1202) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44650F: DataPartitionElements (readbinary3d.c:1203) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x446531: DataPartitionElements (readbinary3d.c:1204) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x446553: DataPartitionElements (readbinary3d.c:1205) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4EE790A: PetscStrallocpy (str.c:188) >> ==15577== by 0x4FBF60A: PetscClassRegLogRegister (classlog.c:253) >> ==15577== by 0x4FBBCCE: PetscClassIdRegister (plog.c:2570) >> ==15577== by 0x5481173: MatMFFDInitializePackage (mffd.c:57) >> ==15577== by 0x52A42A0: MatInitializePackage (dlregismat.c:88) >> ==15577== by 0x577B3AF: MatCreate (gcreate.c:60) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F0997C: PetscFunctionListAdd_Private (reg.c:180) >> ==15577== by 0x548274A: MatMFFDRegister (mffd.c:205) >> ==15577== by 0x548EBF7: MatMFFDRegisterAll (mfregis.c:28) >> ==15577== by 0x54811CD: MatMFFDInitializePackage (mffd.c:59) >> ==15577== by 0x52A42A0: MatInitializePackage (dlregismat.c:88) >> ==15577== by 0x577B3AF: MatCreate (gcreate.c:60) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F09C8C: PetscFunctionListAdd_Private (reg.c:210) >> ==15577== by 0x548274A: MatMFFDRegister (mffd.c:205) >> ==15577== by 0x548EC62: MatMFFDRegisterAll (mfregis.c:29) >> ==15577== by 0x54811CD: MatMFFDInitializePackage (mffd.c:59) >> ==15577== by 0x52A42A0: MatInitializePackage (dlregismat.c:88) >> ==15577== by 0x577B3AF: MatCreate (gcreate.c:60) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x529D004: MatRegisterBaseName (matreg.c:170) >> ==15577== by 0x529D970: MatRegisterAll (matregis.c:97) >> ==15577== by 0x52A457C: MatInitializePackage (dlregismat.c:97) >> ==15577== by 0x577B3AF: MatCreate (gcreate.c:60) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x577B42F: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x577B47F: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x577B4D6: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9037693: ompi_comm_nextcid (comm_cid.c:236) >> ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) >> ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) >> ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90376EA: ompi_comm_nextcid (comm_cid.c:255) >> ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) >> ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) >> ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91E6821: opal_pointer_array_set_item (opal_pointer_array.c:172) >> ==15577== by 0x903778C: ompi_comm_nextcid (comm_cid.c:272) >> ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) >> ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) >> ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91E6837: opal_pointer_array_set_item (opal_pointer_array.c:189) >> ==15577== by 0x903778C: ompi_comm_nextcid (comm_cid.c:272) >> ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) >> ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) >> ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91E6844: opal_pointer_array_set_item (opal_pointer_array.c:193) >> ==15577== by 0x903778C: ompi_comm_nextcid (comm_cid.c:272) >> ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) >> ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) >> ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91E6846: opal_pointer_array_set_item (opal_pointer_array.c:205) >> ==15577== by 0x903778C: ompi_comm_nextcid (comm_cid.c:272) >> ==15577== by 0x903452D: ompi_comm_dup (comm.c:675) >> ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) >> ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xAF16CBD: vfprintf (in /lib64/libc-2.12.so) >> ==15577== by 0xAF3F201: vsnprintf (in /lib64/libc-2.12.so) >> ==15577== by 0xAF1F0D2: snprintf (in /lib64/libc-2.12.so) >> ==15577== by 0x9034558: ompi_comm_dup (comm.c:687) >> ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) >> ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xAF151DC: vfprintf (in /lib64/libc-2.12.so) >> ==15577== by 0xAF3F201: vsnprintf (in /lib64/libc-2.12.so) >> ==15577== by 0xAF1F0D2: snprintf (in /lib64/libc-2.12.so) >> ==15577== by 0x9034558: ompi_comm_dup (comm.c:687) >> ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) >> ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x910CD57: mca_pml_ob1_add_comm (pml_ob1.c:191) >> ==15577== by 0x9037A6F: ompi_comm_activate (comm_cid.c:427) >> ==15577== by 0x9034576: ompi_comm_dup (comm.c:691) >> ==15577== by 0x905300F: PMPI_Comm_dup (pcomm_dup.c:62) >> ==15577== by 0x4F518F4: PetscCommDuplicate (tagm.c:148) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F51A35: PetscCommDuplicate (tagm.c:151) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F2A82D: PetscThreadCommCreate (threadcomm.c:146) >> ==15577== by 0x4F32F1D: PetscThreadCommWorldInitialize (threadcomm.c:1228) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F2A8F9: PetscThreadCommCreate (threadcomm.c:150) >> ==15577== by 0x4F32F1D: PetscThreadCommWorldInitialize (threadcomm.c:1228) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F231A2: PetscMallocValidate (mtr.c:139) >> ==15577== by 0x4F23DA4: PetscTrFreeDefault (mtr.c:266) >> ==15577== by 0x4F8CD1D: PetscOptionsEnd_Private (aoptions.c:486) >> ==15577== by 0x4F2C03B: PetscThreadCommSetNThreads (threadcomm.c:340) >> ==15577== by 0x4F32F93: PetscThreadCommWorldInitialize (threadcomm.c:1230) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F2C6E9: PetscThreadCommSetAffinities (threadcomm.c:423) >> ==15577== by 0x4F32FF9: PetscThreadCommWorldInitialize (threadcomm.c:1231) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F3307C: PetscThreadCommWorldInitialize (threadcomm.c:1232) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F333DC: PetscThreadCommWorldInitialize (threadcomm.c:1240) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F334C0: PetscThreadCommWorldInitialize (threadcomm.c:1241) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F3668F: PetscThreadCommReductionCreate (threadcommred.c:432) >> ==15577== by 0x4F336BD: PetscThreadCommWorldInitialize (threadcomm.c:1251) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F3674F: PetscThreadCommReductionCreate (threadcommred.c:435) >> ==15577== by 0x4F336BD: PetscThreadCommWorldInitialize (threadcomm.c:1251) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F367F2: PetscThreadCommReductionCreate (threadcommred.c:436) >> ==15577== by 0x4F336BD: PetscThreadCommWorldInitialize (threadcomm.c:1251) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F368AC: PetscThreadCommReductionCreate (threadcommred.c:440) >> ==15577== by 0x4F336BD: PetscThreadCommWorldInitialize (threadcomm.c:1251) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F36A3A: PetscThreadCommReductionCreate (threadcommred.c:448) >> ==15577== by 0x4F336BD: PetscThreadCommWorldInitialize (threadcomm.c:1251) >> ==15577== by 0x4F2A1E2: PetscGetThreadCommWorld (threadcomm.c:82) >> ==15577== by 0x4F2A4DB: PetscCommGetThreadComm (threadcomm.c:117) >> ==15577== by 0x4F5206F: PetscCommDuplicate (tagm.c:195) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F579F1: PetscHeaderCreate_Private (inherit.c:68) >> ==15577== by 0x577B546: MatCreate (gcreate.c:62) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x509452A: PetscLayoutCreate (pmap.c:53) >> ==15577== by 0x577B60E: MatCreate (gcreate.c:63) >> ==15577== by 0x54986E3: MatCreateMPIAdj (mpiadj.c:656) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x577BBC1: MatSetSizes (gcreate.c:110) >> ==15577== by 0x5498760: MatCreateMPIAdj (mpiadj.c:657) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x577BCB3: MatSetSizes (gcreate.c:110) >> ==15577== by 0x5498760: MatCreateMPIAdj (mpiadj.c:657) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x577BD73: MatSetSizes (gcreate.c:112) >> ==15577== by 0x5498760: MatCreateMPIAdj (mpiadj.c:657) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x577BD80: MatSetSizes (gcreate.c:112) >> ==15577== by 0x5498760: MatCreateMPIAdj (mpiadj.c:657) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5497EEB: MatCreate_MPIAdj (mpiadj.c:576) >> ==15577== by 0x529C6B9: MatSetType (matreg.c:74) >> ==15577== by 0x54987CB: MatCreateMPIAdj (mpiadj.c:658) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5094D7F: PetscLayoutSetUp (pmap.c:146) >> ==15577== by 0x549681D: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:450) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5094F50: PetscLayoutSetUp (pmap.c:150) >> ==15577== by 0x549681D: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:450) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF6740: PetscSplitOwnership (psplit.c:81) >> ==15577== by 0x5094FD4: PetscLayoutSetUp (pmap.c:152) >> ==15577== by 0x549681D: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:450) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50950BE: PetscLayoutSetUp (pmap.c:156) >> ==15577== by 0x549681D: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:450) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5094F87: PetscLayoutSetUp (pmap.c:151) >> ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF6733: PetscSplitOwnership (psplit.c:81) >> ==15577== by 0x5094FD4: PetscLayoutSetUp (pmap.c:152) >> ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF6790: PetscSplitOwnership (psplit.c:83) >> ==15577== by 0x5094FD4: PetscLayoutSetUp (pmap.c:152) >> ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9111ADC: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:114) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111ADE: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:120) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111B3D: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:136) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111B85: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:470) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9111BD3: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:686) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x5496885: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:451) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5496A42: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:455) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x5496B24: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:458) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5496A6E: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:459) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5496A98: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:459) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x5496B81: MatMPIAdjSetPreallocation_MPIAdj (mpiadj.c:468) >> ==15577== by 0x549840E: MatMPIAdjSetPreallocation (mpiadj.c:610) >> ==15577== by 0x549883B: MatCreateMPIAdj (mpiadj.c:659) >> ==15577== by 0x4465CC: DataPartitionElements (readbinary3d.c:1214) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x520E2D2: MatPartitioningCreate (partition.c:403) >> ==15577== by 0x44663A: DataPartitionElements (readbinary3d.c:1220) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x520E322: MatPartitioningCreate (partition.c:403) >> ==15577== by 0x44663A: DataPartitionElements (readbinary3d.c:1220) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x520E379: MatPartitioningCreate (partition.c:403) >> ==15577== by 0x44663A: DataPartitionElements (readbinary3d.c:1220) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084A50: btl_openib_handle_incoming (btl_openib_endpoint.h:465) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x520E3E9: MatPartitioningCreate (partition.c:403) >> ==15577== by 0x44663A: DataPartitionElements (readbinary3d.c:1220) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084A5A: btl_openib_handle_incoming (btl_openib_endpoint.h:465) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x520E3E9: MatPartitioningCreate (partition.c:403) >> ==15577== by 0x44663A: DataPartitionElements (readbinary3d.c:1220) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50CB3A1: ISCreate (isreg.c:41) >> ==15577== by 0x50B113E: ISCreateStride (stride.c:417) >> ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) >> ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) >> ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50CB3F7: ISCreate (isreg.c:41) >> ==15577== by 0x50B113E: ISCreateStride (stride.c:417) >> ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) >> ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) >> ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50CB454: ISCreate (isreg.c:41) >> ==15577== by 0x50B113E: ISCreateStride (stride.c:417) >> ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) >> ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) >> ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50B1517: ISCreate_Stride (stride.c:432) >> ==15577== by 0x50CBC08: ISSetType (isreg.c:87) >> ==15577== by 0x50B11A9: ISCreateStride (stride.c:418) >> ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) >> ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) >> ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B093F: ISStrideSetStride (stride.c:352) >> ==15577== by 0x50B1219: ISCreateStride (stride.c:419) >> ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) >> ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) >> ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B0E84: ISStrideSetStride_Stride (stride.c:373) >> ==15577== by 0x50B0A23: ISStrideSetStride (stride.c:353) >> ==15577== by 0x50B1219: ISCreateStride (stride.c:419) >> ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) >> ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) >> ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B0EA5: ISStrideSetStride_Stride (stride.c:374) >> ==15577== by 0x50B0A23: ISStrideSetStride (stride.c:353) >> ==15577== by 0x50B1219: ISCreateStride (stride.c:419) >> ==15577== by 0x520ABE5: MatPartitioningApply_Current (partition.c:28) >> ==15577== by 0x520C74A: MatPartitioningApply (partition.c:225) >> ==15577== by 0x44676E: DataPartitionElements (readbinary3d.c:1228) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50AE960: ISGetIndices_Stride (stride.c:160) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x50AE997: ISGetIndices_Stride (stride.c:160) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x50AE997: ISGetIndices_Stride (stride.c:160) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50AE997: ISGetIndices_Stride (stride.c:160) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50AEA0A: ISGetIndices_Stride (stride.c:161) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50AEA72: ISGetIndices_Stride (stride.c:163) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x50D23CD: ISPartitioningToNumbering (iscoloring.c:328) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D2463: ISPartitioningToNumbering (iscoloring.c:330) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D2533: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x50D2566: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x50D2566: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50D2566: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D2584: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x50D25B7: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x50D25B7: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50D25B7: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D25D1: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x50D2604: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x50D2604: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50D2604: ISPartitioningToNumbering (iscoloring.c:339) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50CEF3C: PetscMemzero (petscsys.h:1716) >> ==15577== by 0x50D2690: ISPartitioningToNumbering (iscoloring.c:340) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x50CEF99: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x50D2690: ISPartitioningToNumbering (iscoloring.c:340) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x50CEF99: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x50D2690: ISPartitioningToNumbering (iscoloring.c:340) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D2724: ISPartitioningToNumbering (iscoloring.c:341) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x904F6AD: PMPI_Allreduce (pallreduce.c:87) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x904F6F2: PMPI_Allreduce (pallreduce.c:96) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90A0D42: ompi_coll_tuned_allreduce_intra_dec_fixed (coll_tuned_decision_fixed.c:60) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DF202: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:87) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91DF240: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:104) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DF07B: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3B4: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3BA: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3E0: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A427: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DF0A4: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) >> ==15577== by 0x90A3163: ompi_coll_tuned_allreduce_intra_recursivedoubling (ompi_datatype.h:290) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x910FCA8: mca_pml_ob1_isend (pml_ob1_isend.c:76) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DBB78: opal_convertor_prepare_for_send (opal_convertor.c:559) >> ==15577== by 0x91100C0: mca_pml_ob1_isend (opal_convertor.h:237) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x910FDAA: mca_pml_ob1_isend (pml_ob1_sendreq.h:360) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x911002C: mca_pml_ob1_isend (pml_ob1_sendreq.h:372) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90797D8: mca_btl_sm_sendi (btl_sm.c:815) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x907984E: mca_btl_sm_sendi (btl_sm.c:840) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DB575: opal_convertor_pack (opal_convertor.c:232) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A333: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A35E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A368: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9111BE8: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:194) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9111C64: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvreq.h:168) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116204: mca_pml_ob1_send_request_free (pml_ob1_sendreq.c:117) >> ==15577== by 0x9043BC9: ompi_request_default_wait_all (request.h:350) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90F04C8: ompi_op_base_sum_int64_t (op_base_functions.c:264) >> ==15577== by 0x90A3533: ompi_coll_tuned_allreduce_intra_recursivedoubling (op.h:498) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90F04E3: ompi_op_base_sum_int64_t (op_base_functions.c:264) >> ==15577== by 0x90A3533: ompi_coll_tuned_allreduce_intra_recursivedoubling (op.h:498) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90F04C8: ompi_op_base_sum_int64_t (op_base_functions.c:264) >> ==15577== by 0x90A3573: ompi_coll_tuned_allreduce_intra_recursivedoubling (op.h:498) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90F04E3: ompi_op_base_sum_int64_t (op_base_functions.c:264) >> ==15577== by 0x90A3573: ompi_coll_tuned_allreduce_intra_recursivedoubling (op.h:498) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x907D557: mca_btl_openib_sendi (btl_openib_frag.h:396) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x907D568: mca_btl_openib_sendi (btl_openib_frag.h:396) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x907D5E5: mca_btl_openib_sendi (btl_openib.c:1533) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x907D6AB: mca_btl_openib_sendi (btl_openib.c:1573) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DB575: opal_convertor_pack (opal_convertor.c:232) >> ==15577== by 0x907DB50: mca_btl_openib_sendi (btl_openib.c:1582) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A37E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907DB50: mca_btl_openib_sendi (btl_openib.c:1582) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x907D777: mca_btl_openib_sendi (btl_openib_endpoint.h:305) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0F5A: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x907D977: mca_btl_openib_sendi (verbs.h:1111) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D2775: ISPartitioningToNumbering (iscoloring.c:342) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x905EDD2: PMPI_Scan (pscan.c:83) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x905EC62: PMPI_Scan (pscan.c:92) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DF202: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:87) >> ==15577== by 0x90BDBED: mca_coll_basic_scan_intra (ompi_datatype.h:290) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91DF240: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:104) >> ==15577== by 0x90BDBED: mca_coll_basic_scan_intra (ompi_datatype.h:290) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DF07B: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) >> ==15577== by 0x90BDBED: mca_coll_basic_scan_intra (ompi_datatype.h:290) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DF0A4: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) >> ==15577== by 0x90BDBED: mca_coll_basic_scan_intra (ompi_datatype.h:290) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90F04C8: ompi_op_base_sum_int64_t (op_base_functions.c:264) >> ==15577== by 0x90BDCFB: mca_coll_basic_scan_intra (op.h:498) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90F04E3: ompi_op_base_sum_int64_t (op_base_functions.c:264) >> ==15577== by 0x90BDCFB: mca_coll_basic_scan_intra (op.h:498) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91103AA: mca_pml_ob1_send (pml_ob1_isend.c:108) >> ==15577== by 0x90BDB8E: mca_coll_basic_scan_intra (coll_basic_scan.c:122) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DBB78: opal_convertor_prepare_for_send (opal_convertor.c:559) >> ==15577== by 0x9110A69: mca_pml_ob1_send (opal_convertor.h:237) >> ==15577== by 0x90BDB8E: mca_coll_basic_scan_intra (coll_basic_scan.c:122) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91104AF: mca_pml_ob1_send (pml_ob1_sendreq.h:360) >> ==15577== by 0x90BDB8E: mca_coll_basic_scan_intra (coll_basic_scan.c:122) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9110840: mca_pml_ob1_send (pml_ob1_sendreq.h:372) >> ==15577== by 0x90BDB8E: mca_coll_basic_scan_intra (coll_basic_scan.c:122) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116204: mca_pml_ob1_send_request_free (pml_ob1_sendreq.c:117) >> ==15577== by 0x91109F0: mca_pml_ob1_send (request.h:350) >> ==15577== by 0x90BDB8E: mca_coll_basic_scan_intra (coll_basic_scan.c:122) >> ==15577== by 0x90B16DA: mca_coll_sync_scan (coll_sync_scan.c:43) >> ==15577== by 0x905ECB5: PMPI_Scan (pscan.c:101) >> ==15577== by 0x50D2808: ISPartitioningToNumbering (iscoloring.c:343) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D28B4: ISPartitioningToNumbering (iscoloring.c:344) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D2947: ISPartitioningToNumbering (iscoloring.c:345) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D2958: ISPartitioningToNumbering (iscoloring.c:353) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x50D298B: ISPartitioningToNumbering (iscoloring.c:353) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x50D298B: ISPartitioningToNumbering (iscoloring.c:353) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50D298B: ISPartitioningToNumbering (iscoloring.c:353) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D2A41: ISPartitioningToNumbering (iscoloring.c:354) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x50D2A6E: ISPartitioningToNumbering (iscoloring.c:355) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x50D2A6E: ISPartitioningToNumbering (iscoloring.c:355) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F22C01: PetscMemzero (petscsys.h:1716) >> ==15577== by 0x4F2417A: PetscTrFreeDefault (mtr.c:308) >> ==15577== by 0x50D2A6E: ISPartitioningToNumbering (iscoloring.c:355) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x4F22C5E: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x4F2417A: PetscTrFreeDefault (mtr.c:308) >> ==15577== by 0x50D2A6E: ISPartitioningToNumbering (iscoloring.c:355) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x4F22C5E: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x4F2417A: PetscTrFreeDefault (mtr.c:308) >> ==15577== by 0x50D2A6E: ISPartitioningToNumbering (iscoloring.c:355) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x50D2AA6: ISPartitioningToNumbering (iscoloring.c:355) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x50D2AA6: ISPartitioningToNumbering (iscoloring.c:355) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x50D2ADE: ISPartitioningToNumbering (iscoloring.c:355) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x50D2ADE: ISPartitioningToNumbering (iscoloring.c:355) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x50AECB3: ISRestoreIndices_Stride (stride.c:175) >> ==15577== by 0x50C47D1: ISRestoreIndices (index.c:452) >> ==15577== by 0x50D2B63: ISPartitioningToNumbering (iscoloring.c:357) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x50AECB3: ISRestoreIndices_Stride (stride.c:175) >> ==15577== by 0x50C47D1: ISRestoreIndices (index.c:452) >> ==15577== by 0x50D2B63: ISPartitioningToNumbering (iscoloring.c:357) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50BFC1B: ISCreate_General (general.c:510) >> ==15577== by 0x50CBC08: ISSetType (isreg.c:87) >> ==15577== by 0x50BEEC1: ISCreateGeneral (general.c:438) >> ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50BF554: ISGeneralSetIndices_General (general.c:480) >> ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) >> ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) >> ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50BF59E: ISGeneralSetIndices_General (general.c:481) >> ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) >> ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) >> ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50BE9DD: ISCreateGeneral_Private (general.c:378) >> ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) >> ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) >> ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) >> ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50BE9C5: ISCreateGeneral_Private (general.c:379) >> ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) >> ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) >> ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) >> ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50BE9E6: ISCreateGeneral_Private (general.c:381) >> ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) >> ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) >> ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) >> ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50BEA90: ISCreateGeneral_Private (general.c:383) >> ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) >> ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) >> ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) >> ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50BEA31: ISCreateGeneral_Private (general.c:384) >> ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) >> ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) >> ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) >> ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50BEA67: ISCreateGeneral_Private (general.c:385) >> ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) >> ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) >> ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) >> ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x5493F54: MatDestroy_MPIAdj (mpiadj.c:73) >> ==15577== by 0x523BDF2: MatDestroy (matrix.c:1029) >> ==15577== by 0x4468C1: DataPartitionElements (readbinary3d.c:1246) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x5493F54: MatDestroy_MPIAdj (mpiadj.c:73) >> ==15577== by 0x523BDF2: MatDestroy (matrix.c:1029) >> ==15577== by 0x4468C1: DataPartitionElements (readbinary3d.c:1246) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x5493FFB: MatDestroy_MPIAdj (mpiadj.c:74) >> ==15577== by 0x523BDF2: MatDestroy (matrix.c:1029) >> ==15577== by 0x4468C1: DataPartitionElements (readbinary3d.c:1246) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x5493FFB: MatDestroy_MPIAdj (mpiadj.c:74) >> ==15577== by 0x523BDF2: MatDestroy (matrix.c:1029) >> ==15577== by 0x4468C1: DataPartitionElements (readbinary3d.c:1246) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x446C58: DataMoveGeneric (readbinary3d.c:1297) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D30A5: ISPartitioningCount (iscoloring.c:415) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2BC: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A32D: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A350: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A353: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A35E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D30F8: ISPartitioningCount (iscoloring.c:416) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D3172: ISPartitioningCount (iscoloring.c:418) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50D3207: ISPartitioningCount (iscoloring.c:427) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D331A: ISPartitioningCount (iscoloring.c:429) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x902E5D0: append_frag_to_list (pml_ob1_recvfrag.c:72) >> ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x902E608: append_frag_to_list (pml_ob1_recvfrag.c:72) >> ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x906F575: mca_allocator_bucket_alloc (allocator_bucket_alloc.c:91) >> ==15577== by 0x902E638: append_frag_to_list (pml_ob1_recvfrag.c:72) >> ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x906F591: mca_allocator_bucket_alloc (allocator_bucket_alloc.c:91) >> ==15577== by 0x902E638: append_frag_to_list (pml_ob1_recvfrag.c:72) >> ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x902E664: append_frag_to_list (pml_ob1_recvfrag.c:72) >> ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x902E664: append_frag_to_list (pml_ob1_recvfrag.c:72) >> ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9113982: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.c:632) >> ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A3243: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:216) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91139B7: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.c:632) >> ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A3243: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:216) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115C7B: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1037) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A3243: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:216) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50D3438: ISPartitioningCount (iscoloring.c:432) >> ==15577== by 0x446D0F: DataMoveGeneric (readbinary3d.c:1301) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4FC852D: EventRegLogRegister (eventlog.c:310) >> ==15577== by 0x4FA6657: PetscLogEventRegister (plog.c:753) >> ==15577== by 0x517D85C: VecInitializePackage (dlregisvec.c:127) >> ==15577== by 0x517AB60: VecCreate (veccreate.c:35) >> ==15577== by 0x446D7D: DataMoveGeneric (readbinary3d.c:1306) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4FC7EE7: EventPerfLogEnsureSize (eventlog.c:211) >> ==15577== by 0x4FA66E5: PetscLogEventRegister (plog.c:755) >> ==15577== by 0x517D85C: VecInitializePackage (dlregisvec.c:127) >> ==15577== by 0x517AB60: VecCreate (veccreate.c:35) >> ==15577== by 0x446D7D: DataMoveGeneric (readbinary3d.c:1306) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x517ABE0: VecCreate (veccreate.c:37) >> ==15577== by 0x446D7D: DataMoveGeneric (readbinary3d.c:1306) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x517AC30: VecCreate (veccreate.c:37) >> ==15577== by 0x446D7D: DataMoveGeneric (readbinary3d.c:1306) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x517AC87: VecCreate (veccreate.c:37) >> ==15577== by 0x446D7D: DataMoveGeneric (readbinary3d.c:1306) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x517B71C: VecSetType (vecreg.c:49) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x514FDC7: VecCreate_MPI_Private (pbvec.c:185) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5094D7F: PetscLayoutSetUp (pmap.c:146) >> ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5094F50: PetscLayoutSetUp (pmap.c:150) >> ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084E1A: btl_openib_handle_incoming (btl_openib_component.c:1374) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9084E1D: btl_openib_handle_incoming (btl_openib_component.c:1374) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9083BD8: btl_openib_control (btl_openib_component.c:478) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9083BDD: btl_openib_control (btl_openib_component.c:478) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9083DD0: btl_openib_control (btl_openib_component.c:507) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x514FFC2: VecCreate_MPI_Private (pbvec.c:197) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x514FFF5: VecCreate_MPI_Private (pbvec.c:197) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x514FFF5: VecCreate_MPI_Private (pbvec.c:197) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x514FFF5: VecCreate_MPI_Private (pbvec.c:197) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5150068: VecCreate_MPI_Private (pbvec.c:198) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x514E9F4: PetscMemzero (petscsys.h:1716) >> ==15577== by 0x515010F: VecCreate_MPI_Private (pbvec.c:199) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x514EA51: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x515010F: VecCreate_MPI_Private (pbvec.c:199) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x514EA51: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x515010F: VecCreate_MPI_Private (pbvec.c:199) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084E65: btl_openib_handle_incoming (btl_openib_component.c:3106) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) >> ==15577== by 0x51F6263: VecStashCreate_Private (vecstash.c:31) >> ==15577== by 0x51501C0: VecCreate_MPI_Private (pbvec.c:211) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90848C9: btl_openib_handle_incoming (btl_openib_component.c:3167) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) >> ==15577== by 0x51F6263: VecStashCreate_Private (vecstash.c:31) >> ==15577== by 0x51501C0: VecCreate_MPI_Private (pbvec.c:211) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9084903: btl_openib_handle_incoming (btl_openib_component.c:3173) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) >> ==15577== by 0x51F6263: VecStashCreate_Private (vecstash.c:31) >> ==15577== by 0x51501C0: VecCreate_MPI_Private (pbvec.c:211) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51F6454: VecStashCreate_Private (vecstash.c:37) >> ==15577== by 0x51501C0: VecCreate_MPI_Private (pbvec.c:211) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x446FE3: DataMoveGeneric (readbinary3d.c:1319) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x446FE3: DataMoveGeneric (readbinary3d.c:1319) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x446FE3: DataMoveGeneric (readbinary3d.c:1319) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B5E51: ISCreateBlock (block.c:376) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B5E9B: ISCreateBlock (block.c:377) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084A66: btl_openib_handle_incoming (btl_openib_endpoint.h:475) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084A76: btl_openib_handle_incoming (btl_openib_endpoint.h:475) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088709: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:768) >> ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088C4F: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:773) >> ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x908872E: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.h:526) >> ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088747: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:789) >> ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088769: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:799) >> ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088773: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:799) >> ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088792: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.c:802) >> ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088822: mca_btl_openib_endpoint_send_credits (btl_openib_endpoint.h:322) >> ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C13DC: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9088991: mca_btl_openib_endpoint_send_credits (verbs.h:1111) >> ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C144B: mlx4_post_send (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9088991: mca_btl_openib_endpoint_send_credits (verbs.h:1111) >> ==15577== by 0x9084A8B: btl_openib_handle_incoming (btl_openib_endpoint.h:476) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088EC2: mca_btl_openib_endpoint_credits (btl_openib_endpoint.h:455) >> ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088ECF: mca_btl_openib_endpoint_credits (btl_openib_endpoint.c:745) >> ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088EDA: mca_btl_openib_endpoint_credits (btl_openib_endpoint.c:745) >> ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088F04: mca_btl_openib_endpoint_credits (btl_openib_endpoint.c:748) >> ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088F80: mca_btl_openib_endpoint_credits (btl_openib_endpoint.c:748) >> ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088F95: mca_btl_openib_endpoint_credits (btl_openib_endpoint.c:748) >> ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088F19: mca_btl_openib_endpoint_credits (btl_openib_endpoint.h:465) >> ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9088F1D: mca_btl_openib_endpoint_credits (btl_openib_endpoint.h:465) >> ==15577== by 0x908595E: poll_device (btl_openib_component.c:3373) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50B7D85: ISCreate_Block (block.c:568) >> ==15577== by 0x50CBC08: ISSetType (isreg.c:87) >> ==15577== by 0x50B5FD0: ISCreateBlock (block.c:380) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x907D5F4: mca_btl_openib_sendi (btl_openib_endpoint.h:526) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A3282: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:219) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50B5718: ISBlockSetIndices_Block (block.c:307) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B57D3: ISBlockSetIndices_Block (block.c:308) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B57BB: ISBlockSetIndices_Block (block.c:309) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B57DC: ISBlockSetIndices_Block (block.c:311) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B5886: ISBlockSetIndices_Block (block.c:313) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B5827: ISBlockSetIndices_Block (block.c:314) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B585D: ISBlockSetIndices_Block (block.c:315) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B589D: ISBlockSetIndices_Block (block.c:318) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x50B58D4: ISBlockSetIndices_Block (block.c:318) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x50B58D4: ISBlockSetIndices_Block (block.c:318) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50B58D4: ISBlockSetIndices_Block (block.c:318) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B5948: ISBlockSetIndices_Block (block.c:319) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B1DDE: PetscMemcpy (petscsys.h:1656) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B1E2F: PetscMemcpy (petscsys.h:1657) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B1EA8: PetscMemcpy (petscsys.h:1663) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B1EBF: PetscMemcpy (petscsys.h:1663) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) >> ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3E0: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x50B1F35: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x50B59E9: ISBlockSetIndices_Block (block.c:320) >> ==15577== by 0x50B52D6: ISBlockSetIndices (block.c:291) >> ==15577== by 0x50B6046: ISCreateBlock (block.c:381) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x447225: DataMoveGeneric (readbinary3d.c:1329) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x447225: DataMoveGeneric (readbinary3d.c:1329) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5174EA5: VecSetSizes (vector.c:1430) >> ==15577== by 0x514CF9A: VecCreateSeq (vseqcr.c:38) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5174F91: VecSetSizes (vector.c:1430) >> ==15577== by 0x514CF9A: VecCreateSeq (vseqcr.c:38) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5174FE8: VecSetSizes (vector.c:1431) >> ==15577== by 0x514CF9A: VecCreateSeq (vseqcr.c:38) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5174FF2: VecSetSizes (vector.c:1431) >> ==15577== by 0x514CF9A: VecCreateSeq (vseqcr.c:38) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x517B71C: VecSetType (vecreg.c:49) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x514D594: VecCreate_Seq (bvec3.c:38) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x514D5C7: VecCreate_Seq (bvec3.c:38) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x514D5C7: VecCreate_Seq (bvec3.c:38) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x514D5C7: VecCreate_Seq (bvec3.c:38) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x514D637: VecCreate_Seq (bvec3.c:39) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5141A03: VecCreate_Seq_Private (bvec2.c:1245) >> ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5094D7F: PetscLayoutSetUp (pmap.c:146) >> ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) >> ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5094D90: PetscLayoutSetUp (pmap.c:146) >> ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) >> ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5094F50: PetscLayoutSetUp (pmap.c:150) >> ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) >> ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5094F87: PetscLayoutSetUp (pmap.c:151) >> ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) >> ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF685C: PetscSplitOwnership (psplit.c:85) >> ==15577== by 0x5094FD4: PetscLayoutSetUp (pmap.c:152) >> ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) >> ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF6A4B: PetscSplitOwnership (psplit.c:93) >> ==15577== by 0x5094FD4: PetscLayoutSetUp (pmap.c:152) >> ==15577== by 0x5141B62: VecCreate_Seq_Private (bvec2.c:1253) >> ==15577== by 0x514D6C8: VecCreate_Seq (bvec3.c:40) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5145D6F: PetscMemzero (petscsys.h:1716) >> ==15577== by 0x5149E73: VecSet_Seq (dvec2.c:729) >> ==15577== by 0x5185F76: VecSet (rvector.c:571) >> ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x5145DCC: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x5149E73: VecSet_Seq (dvec2.c:729) >> ==15577== by 0x5185F76: VecSet (rvector.c:571) >> ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x5145DCC: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x5149E73: VecSet_Seq (dvec2.c:729) >> ==15577== by 0x5185F76: VecSet (rvector.c:571) >> ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F8811A: PetscObjectComposedDataIncreaseReal (state.c:169) >> ==15577== by 0x518619F: VecSet (rvector.c:577) >> ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4F8822D: PetscObjectComposedDataIncreaseReal (state.c:171) >> ==15577== by 0x518619F: VecSet (rvector.c:577) >> ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x83FA903: sqrt (in /lib64/libm-2.12.so) >> ==15577== by 0x518635A: VecSet (rvector.c:579) >> ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x83FA90D: sqrt (in /lib64/libm-2.12.so) >> ==15577== by 0x518635A: VecSet (rvector.c:579) >> ==15577== by 0x514D748: VecCreate_Seq (bvec3.c:45) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x514D005: VecCreateSeq (vseqcr.c:39) >> ==15577== by 0x4472BD: DataMoveGeneric (readbinary3d.c:1334) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51B46C3: VecScatterCreate (vscat.c:1028) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51B4713: VecScatterCreate (vscat.c:1028) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51B476A: VecScatterCreate (vscat.c:1028) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3BA: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x51BAC0B: VecScatterCreate (vscat.c:1548) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3D1: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x51BAC0B: VecScatterCreate (vscat.c:1548) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3DA: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9111EC4: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:217) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x51BAC0B: VecScatterCreate (vscat.c:1548) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B27A4: ISGetIndices_Block (block.c:45) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x51BB45D: VecScatterCreate (vscat.c:1583) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x50B27DC: ISGetIndices_Block (block.c:45) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x51BB45D: VecScatterCreate (vscat.c:1583) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x50B27DC: ISGetIndices_Block (block.c:45) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x51BB45D: VecScatterCreate (vscat.c:1583) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x50B27DC: ISGetIndices_Block (block.c:45) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x51BB45D: VecScatterCreate (vscat.c:1583) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B28CC: ISGetIndices_Block (block.c:49) >> ==15577== by 0x50C3F71: ISGetIndices (index.c:374) >> ==15577== by 0x51BB45D: VecScatterCreate (vscat.c:1583) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51BB4C4: VecScatterCreate (vscat.c:1584) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9084E1D: btl_openib_handle_incoming (btl_openib_component.c:1374) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) >> ==15577== by 0x4F51037: PetscObjectGetNewTag (tagm.c:47) >> ==15577== by 0x51E1339: VecScatterCreate_PtoS (vpscat.c:1736) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9084F20: btl_openib_handle_incoming (btl_openib_component.c:1382) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) >> ==15577== by 0x4F51037: PetscObjectGetNewTag (tagm.c:47) >> ==15577== by 0x51E1339: VecScatterCreate_PtoS (vpscat.c:1736) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Invalid read of size 4 >> ==15577== at 0x902E5D0: append_frag_to_list (pml_ob1_recvfrag.c:72) >> ==15577== by 0x9111F45: mca_pml_ob1_recv_frag_callback_match (pml_ob1_recvfrag.c:490) >> ==15577== by 0x90845A7: btl_openib_handle_incoming (btl_openib_component.c:3097) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) >> ==15577== by 0x4F51037: PetscObjectGetNewTag (tagm.c:47) >> ==15577== by 0x51E1339: VecScatterCreate_PtoS (vpscat.c:1736) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== Address 0x145854be is 209,982 bytes inside a block of size 209,984 alloc'd >> ==15577== at 0x4C266CA: memalign (vg_replace_malloc.c:727) >> ==15577== by 0x4C26765: posix_memalign (vg_replace_malloc.c:876) >> ==15577== by 0x90ED708: mca_mpool_rdma_alloc (mpool_rdma_module.c:103) >> ==15577== by 0x908903C: mca_btl_openib_endpoint_connect_eager_rdma (btl_openib_endpoint.c:947) >> ==15577== by 0x908592B: poll_device (btl_openib_component.c:3424) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A329E: ompi_coll_tuned_allreduce_intra_recursivedoubling (coll_tuned_allreduce.c:223) >> ==15577== by 0x904F745: PMPI_Allreduce (pallreduce.c:105) >> ==15577== by 0x50BE922: ISCreateGeneral_Private (general.c:377) >> ==15577== by 0x50BF91D: ISGeneralSetIndices_General (general.c:497) >> ==15577== by 0x50BF22B: ISGeneralSetIndices (general.c:468) >> ==15577== by 0x50BEF30: ISCreateGeneral (general.c:439) >> ==15577== by 0x50D2BDB: ISPartitioningToNumbering (iscoloring.c:358) >> ==15577== by 0x446860: DataPartitionElements (readbinary3d.c:1241) >> ==15577== by 0x43D42D: ReadBinary (readbinary3d.c:229) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E1615: VecScatterCreate_PtoS (vpscat.c:1745) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E1635: VecScatterCreate_PtoS (vpscat.c:1745) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x51E166E: VecScatterCreate_PtoS (vpscat.c:1745) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x51E166E: VecScatterCreate_PtoS (vpscat.c:1745) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E166E: VecScatterCreate_PtoS (vpscat.c:1745) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E18C9: VecScatterCreate_PtoS (vpscat.c:1750) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E17B3: VecScatterCreate_PtoS (vpscat.c:1752) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E17DD: VecScatterCreate_PtoS (vpscat.c:1754) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4EF806D: PetscGatherNumberOfMessages (mpimesg.c:45) >> ==15577== by 0x51E1938: VecScatterCreate_PtoS (vpscat.c:1767) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4EF80BD: PetscGatherNumberOfMessages (mpimesg.c:45) >> ==15577== by 0x51E1938: VecScatterCreate_PtoS (vpscat.c:1767) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF86B1: PetscGatherMessageLengths (mpimesg.c:112) >> ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF8715: PetscGatherMessageLengths (mpimesg.c:112) >> ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF87EB: PetscGatherMessageLengths (mpimesg.c:116) >> ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF8998: PetscGatherMessageLengths (mpimesg.c:117) >> ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF8AEA: PetscGatherMessageLengths (mpimesg.c:130) >> ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF8BC3: PetscGatherMessageLengths (mpimesg.c:133) >> ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EF8CA6: PetscGatherMessageLengths (mpimesg.c:134) >> ==15577== by 0x51E19C3: VecScatterCreate_PtoS (vpscat.c:1768) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE4C2B: PetscSortMPIIntWithArray (sorti.c:476) >> ==15577== by 0x51E1A36: VecScatterCreate_PtoS (vpscat.c:1769) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE4D72: PetscSortMPIIntWithArray (sorti.c:477) >> ==15577== by 0x51E1A36: VecScatterCreate_PtoS (vpscat.c:1769) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E1AC9: VecScatterCreate_PtoS (vpscat.c:1770) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E1B31: VecScatterCreate_PtoS (vpscat.c:1773) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E1B8D: VecScatterCreate_PtoS (vpscat.c:1773) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E1DA8: VecScatterCreate_PtoS (vpscat.c:1775) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E1DBC: VecScatterCreate_PtoS (vpscat.c:1784) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x51E1DF5: VecScatterCreate_PtoS (vpscat.c:1784) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x51E1DF5: VecScatterCreate_PtoS (vpscat.c:1784) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E1DF5: VecScatterCreate_PtoS (vpscat.c:1784) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E1EAE: VecScatterCreate_PtoS (vpscat.c:1784) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E2038: VecScatterCreate_PtoS (vpscat.c:1788) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E23A6: VecScatterCreate_PtoS (vpscat.c:1803) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E2438: VecScatterCreate_PtoS (vpscat.c:1814) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E2509: VecScatterCreate_PtoS (vpscat.c:1817) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E268E: VecScatterCreate_PtoS (vpscat.c:1818) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x51E26D0: VecScatterCreate_PtoS (vpscat.c:1818) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x51E26D0: VecScatterCreate_PtoS (vpscat.c:1818) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E26D0: VecScatterCreate_PtoS (vpscat.c:1818) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E26F5: VecScatterCreate_PtoS (vpscat.c:1818) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E27D6: VecScatterCreate_PtoS (vpscat.c:1819) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E2877: VecScatterCreate_PtoS (vpscat.c:1819) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E297A: VecScatterCreate_PtoS (vpscat.c:1824) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E2D8D: VecScatterCreate_PtoS (vpscat.c:1841) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E2F1B: VecScatterCreate_PtoS (vpscat.c:1845) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E2F9F: VecScatterCreate_PtoS (vpscat.c:1845) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E304F: VecScatterCreate_PtoS (vpscat.c:1845) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E318D: VecScatterCreate_PtoS (vpscat.c:1849) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E31E7: VecScatterCreate_PtoS (vpscat.c:1849) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E3488: VecScatterCreate_PtoS (vpscat.c:1859) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x51E35A4: VecScatterCreate_PtoS (vpscat.c:1866) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x51E35A4: VecScatterCreate_PtoS (vpscat.c:1866) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x51E393C: VecScatterCreate_PtoS (vpscat.c:1874) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x51E393C: VecScatterCreate_PtoS (vpscat.c:1874) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E3A25: VecScatterCreate_PtoS (vpscat.c:1879) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51E3AD6: VecScatterCreate_PtoS (vpscat.c:1880) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E3CC2: VecScatterCreate_PtoS (vpscat.c:1882) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E3B97: VecScatterCreate_PtoS (vpscat.c:1884) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E3BBF: VecScatterCreate_PtoS (vpscat.c:1884) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E3C6D: VecScatterCreate_PtoS (vpscat.c:1887) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E5BAA: VecScatterCreateCommon_PtoS (vpscat.c:2070) >> ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E66EB: VecScatterCreateCommon_PtoS (vpscat.c:2100) >> ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E69AA: VecScatterCreateCommon_PtoS (vpscat.c:2113) >> ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E6E80: VecScatterCreateCommon_PtoS (vpscat.c:2129) >> ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51E6E93: VecScatterCreateCommon_PtoS (vpscat.c:2129) >> ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51C32DC: VecScatterLocalOptimizeCopy_Private (vpscat.c:263) >> ==15577== by 0x51E716C: VecScatterCreateCommon_PtoS (vpscat.c:2175) >> ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x50B2B31: ISRestoreIndices_Block (block.c:65) >> ==15577== by 0x50C47D1: ISRestoreIndices (index.c:452) >> ==15577== by 0x51BB627: VecScatterCreate (vscat.c:1587) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x50B2B31: ISRestoreIndices_Block (block.c:65) >> ==15577== by 0x50C47D1: ISRestoreIndices (index.c:452) >> ==15577== by 0x51BB627: VecScatterCreate (vscat.c:1587) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x50B21FC: ISDestroy_Block (block.c:24) >> ==15577== by 0x50C29BD: ISDestroy (index.c:220) >> ==15577== by 0x4474FD: DataMoveGeneric (readbinary3d.c:1349) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x50B21FC: ISDestroy_Block (block.c:24) >> ==15577== by 0x50C29BD: ISDestroy (index.c:220) >> ==15577== by 0x4474FD: DataMoveGeneric (readbinary3d.c:1349) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51BC86C: VecScatterBegin (vscat.c:1718) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51BC880: VecScatterBegin (vscat.c:1718) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51BCA5A: VecScatterBegin (vscat.c:1725) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51BCACA: VecScatterBegin (vscat.c:1726) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51D1012: VecScatterBegin_1 (vpscat.h:81) >> ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51BFE32: PetscMemcpy (petscsys.h:1657) >> ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) >> ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51BFE7F: PetscMemcpy (petscsys.h:1661) >> ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) >> ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51BFE8D: PetscMemcpy (petscsys.h:1663) >> ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) >> ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51BFEBB: PetscMemcpy (petscsys.h:1663) >> ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) >> ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) >> ==15577== by 0x51BFF31: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) >> ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) >> ==15577== by 0x51BFF31: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) >> ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x51BFF31: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) >> ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x51BFF31: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) >> ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x51BFF31: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x51D1881: VecScatterBegin_1 (vpscat.h:124) >> ==15577== by 0x51BCEC4: VecScatterBegin (vscat.c:1733) >> ==15577== by 0x447571: DataMoveGeneric (readbinary3d.c:1350) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51D242A: VecScatterEnd_1 (vpscat.h:189) >> ==15577== by 0x51BD974: VecScatterEnd (vscat.c:1780) >> ==15577== by 0x4475E5: DataMoveGeneric (readbinary3d.c:1351) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x51C2E20: VecScatterDestroy_PtoP (vpscat.c:234) >> ==15577== by 0x51BE0FD: VecScatterDestroy (vscat.c:1813) >> ==15577== by 0x447646: DataMoveGeneric (readbinary3d.c:1352) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x51C2E20: VecScatterDestroy_PtoP (vpscat.c:234) >> ==15577== by 0x51BE0FD: VecScatterDestroy (vscat.c:1813) >> ==15577== by 0x447646: DataMoveGeneric (readbinary3d.c:1352) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x5140F02: VecDestroy_Seq (bvec2.c:1132) >> ==15577== by 0x516CA2A: VecDestroy (vector.c:547) >> ==15577== by 0x4476A7: DataMoveGeneric (readbinary3d.c:1353) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x5140F02: VecDestroy_Seq (bvec2.c:1132) >> ==15577== by 0x516CA2A: VecDestroy (vector.c:547) >> ==15577== by 0x4476A7: DataMoveGeneric (readbinary3d.c:1353) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x44774D: DataMoveGeneric (readbinary3d.c:1355) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x44774D: DataMoveGeneric (readbinary3d.c:1355) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x447841: DataMoveGeneric (readbinary3d.c:1364) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x447841: DataMoveGeneric (readbinary3d.c:1364) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x447841: DataMoveGeneric (readbinary3d.c:1364) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x51554A7: VecDestroy_MPI (pdvec.c:20) >> ==15577== by 0x516CA2A: VecDestroy (vector.c:547) >> ==15577== by 0x447A21: DataMoveGeneric (readbinary3d.c:1376) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x51554A7: VecDestroy_MPI (pdvec.c:20) >> ==15577== by 0x516CA2A: VecDestroy (vector.c:547) >> ==15577== by 0x447A21: DataMoveGeneric (readbinary3d.c:1376) >> ==15577== by 0x43D4A0: ReadBinary (readbinary3d.c:233) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90848C9: btl_openib_handle_incoming (btl_openib_component.c:3167) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9084903: btl_openib_handle_incoming (btl_openib_component.c:3173) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A6990: ompi_coll_tuned_allgather_intra_bruck (coll_tuned_util.h:57) >> ==15577== by 0x904F084: PMPI_Allgather (pallgather.c:117) >> ==15577== by 0x509518E: PetscLayoutSetUp (pmap.c:158) >> ==15577== by 0x514FF1B: VecCreate_MPI_Private (pbvec.c:191) >> ==15577== by 0x5150532: VecCreate_MPI (pbvec.c:240) >> ==15577== by 0x517B76F: VecSetType (vecreg.c:53) >> ==15577== by 0x5174355: VecSetTypeFromOptions_Private (vector.c:1354) >> ==15577== by 0x517480E: VecSetFromOptions (vector.c:1389) >> ==15577== by 0x446ECF: DataMoveGeneric (readbinary3d.c:1309) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9085177: btl_openib_handle_incoming (btl_openib_component.c:3150) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x90846B5: btl_openib_handle_incoming (btl_openib_endpoint.h:391) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x90846C9: btl_openib_handle_incoming (btl_openib_endpoint.h:398) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084C3B: btl_openib_handle_incoming (btl_openib_endpoint.h:409) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084C50: btl_openib_handle_incoming (btl_openib_endpoint.h:410) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0C62: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9084F8D: btl_openib_handle_incoming (verbs.h:1120) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C0CAF: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9084F8D: btl_openib_handle_incoming (verbs.h:1120) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0xF1C0CD4: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9084F8D: btl_openib_handle_incoming (verbs.h:1120) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0xF1C0D0D: mlx4_post_recv (in /usr/lib64/libmlx4-m-rdmav2.so) >> ==15577== by 0x9084F8D: btl_openib_handle_incoming (verbs.h:1120) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084FA8: btl_openib_handle_incoming (btl_openib_endpoint.h:416) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084FBA: btl_openib_handle_incoming (btl_openib_endpoint.h:417) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084A3F: btl_openib_handle_incoming (btl_openib_endpoint.h:464) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084A50: btl_openib_handle_incoming (btl_openib_endpoint.h:465) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084A5A: btl_openib_handle_incoming (btl_openib_endpoint.h:465) >> ==15577== by 0x90858B8: poll_device (btl_openib_component.c:3410) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) >> ==15577== by 0x43D513: ReadBinary (readbinary3d.c:234) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) >> ==15577== by 0x43D586: ReadBinary (readbinary3d.c:235) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) >> ==15577== by 0x43D586: ReadBinary (readbinary3d.c:235) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) >> ==15577== by 0x43D586: ReadBinary (readbinary3d.c:235) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) >> ==15577== by 0x43D586: ReadBinary (readbinary3d.c:235) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) >> ==15577== by 0x43D586: ReadBinary (readbinary3d.c:235) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) >> ==15577== by 0x43D5F9: ReadBinary (readbinary3d.c:236) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) >> ==15577== by 0x43D5F9: ReadBinary (readbinary3d.c:236) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) >> ==15577== by 0x43D5F9: ReadBinary (readbinary3d.c:236) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) >> ==15577== by 0x43D5F9: ReadBinary (readbinary3d.c:236) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) >> ==15577== by 0x43D5F9: ReadBinary (readbinary3d.c:236) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x43D75A: ReadBinary (readbinary3d.c:255) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) >> ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) >> ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) >> ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9087052: mca_btl_openib_endpoint_post_send (btl_openib_endpoint.h:526) >> ==15577== by 0x90886AC: mca_btl_openib_endpoint_send (btl_openib_endpoint.c:714) >> ==15577== by 0x9089275: mca_btl_openib_endpoint_connect_eager_rdma (btl_openib_endpoint.c:901) >> ==15577== by 0x908592B: poll_device (btl_openib_component.c:3424) >> ==15577== by 0x9085E58: btl_openib_component_progress (btl_openib_component.c:3660) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51580: PetscCommGetNewTag (tagm.c:99) >> ==15577== by 0x4F51037: PetscObjectGetNewTag (tagm.c:47) >> ==15577== by 0x51E410A: VecScatterCreateCommon_PtoS (vpscat.c:1929) >> ==15577== by 0x51E3DB8: VecScatterCreate_PtoS (vpscat.c:1909) >> ==15577== by 0x51E7418: VecScatterCreate_StoP (vpscat.c:2197) >> ==15577== by 0x51BB54B: VecScatterCreate (vscat.c:1585) >> ==15577== by 0x44749C: DataMoveGeneric (readbinary3d.c:1348) >> ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) >> ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) >> ==15577== by 0x43D780: ReadBinary (readbinary3d.c:256) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x446FA9: DataMoveGeneric (readbinary3d.c:1319) >> ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4470FC: DataMoveGeneric (readbinary3d.c:1322) >> ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084E32: btl_openib_handle_incoming (btl_openib_component.c:3111) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084E40: btl_openib_handle_incoming (btl_openib_component.c:3113) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x9084E52: btl_openib_handle_incoming (btl_openib_component.c:3113) >> ==15577== by 0x9085D70: btl_openib_component_progress (btl_openib_component.c:3645) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8E4A: ompi_coll_tuned_barrier_intra_bruck (coll_tuned_barrier.c:225) >> ==15577== by 0x90504F5: PMPI_Barrier (pbarrier.c:70) >> ==15577== by 0x4F51E11: PetscCommDuplicate (tagm.c:180) >> ==15577== by 0x4F577F0: PetscHeaderCreate_Private (inherit.c:55) >> ==15577== by 0x50CB4CA: ISCreate (isreg.c:41) >> ==15577== by 0x50B5F65: ISCreateBlock (block.c:379) >> ==15577== by 0x447132: DataMoveGeneric (readbinary3d.c:1326) >> ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44740F: DataMoveGeneric (readbinary3d.c:1336) >> ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4477FC: DataMoveGeneric (readbinary3d.c:1364) >> ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4479A8: DataMoveGeneric (readbinary3d.c:1366) >> ==15577== by 0x43D7F6: ReadBinary (readbinary3d.c:257) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B0ED3: ISStrideSetStride_Stride (stride.c:377) >> ==15577== by 0x50B0A23: ISStrideSetStride (stride.c:353) >> ==15577== by 0x50B1219: ISCreateStride (stride.c:419) >> ==15577== by 0x43D898: ReadBinary (readbinary3d.c:263) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50B0EE4: ISStrideSetStride_Stride (stride.c:377) >> ==15577== by 0x50B0A23: ISStrideSetStride (stride.c:353) >> ==15577== by 0x50B1219: ISCreateStride (stride.c:419) >> ==15577== by 0x43D898: ReadBinary (readbinary3d.c:263) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5119745: AOCreate (ao.c:577) >> ==15577== by 0x5114195: AOCreateBasicIS (aobasic.c:382) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5119795: AOCreate (ao.c:577) >> ==15577== by 0x5114195: AOCreateBasicIS (aobasic.c:382) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51197EC: AOCreate (ao.c:577) >> ==15577== by 0x5114195: AOCreateBasicIS (aobasic.c:382) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5118EAF: AOSetIS (ao.c:498) >> ==15577== by 0x5114201: AOCreateBasicIS (aobasic.c:383) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x51122DE: AOCreate_Basic (aobasic.c:205) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510FD16: PetscMPIIntCast (petscsys.h:1983) >> ==15577== by 0x5112568: AOCreate_Basic (aobasic.c:213) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5112740: AOCreate_Basic (aobasic.c:219) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5112790: AOCreate_Basic (aobasic.c:219) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510FD16: PetscMPIIntCast (petscsys.h:1983) >> ==15577== by 0x51128F5: AOCreate_Basic (aobasic.c:223) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51129A1: AOCreate_Basic (aobasic.c:230) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5112B3A: AOCreate_Basic (aobasic.c:242) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x5112B70: AOCreate_Basic (aobasic.c:242) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x5112B70: AOCreate_Basic (aobasic.c:242) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5112B70: AOCreate_Basic (aobasic.c:242) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5112B8D: AOCreate_Basic (aobasic.c:242) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x5112BC3: AOCreate_Basic (aobasic.c:242) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x5112BC3: AOCreate_Basic (aobasic.c:242) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5112BC3: AOCreate_Basic (aobasic.c:242) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x904F223: PMPI_Allgatherv (pallgatherv.c:93) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x90A1442: ompi_coll_tuned_allgatherv_intra_dec_fixed (coll_tuned_decision_fixed.c:643) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x904DFD9: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:54) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DF202: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:87) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DF223: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:98) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x91DF225: opal_datatype_copy_content_same_ddt (opal_datatype_copy.c:99) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DF07B: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2BC: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2EC: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A316: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A323: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DF0A0: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:154) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DF0A4: non_overlap_copy_content_same_ddt (opal_datatype_copy.h:146) >> ==15577== by 0x904E48D: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:61) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x904E491: ompi_datatype_sndrcv (ompi_datatype_sndrcv.c:62) >> ==15577== by 0x90A7D56: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:398) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x910FCA8: mca_pml_ob1_isend (pml_ob1_isend.c:76) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x910FDAA: mca_pml_ob1_isend (pml_ob1_sendreq.h:360) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DB575: opal_convertor_pack (opal_convertor.c:232) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) >> ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) >> ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) >> ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) >> ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3A0: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) >> ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3B4: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) >> ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3BA: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) >> ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) >> ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x911989F: mca_pml_ob1_send_request_start_rndv (bml.h:339) >> ==15577== by 0x91101A4: mca_pml_ob1_isend (pml_ob1_sendreq.h:401) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2BC: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x911466E: mca_pml_ob1_recv_request_progress_rndv (pml_ob1_recvreq.c:581) >> ==15577== by 0x911196C: mca_pml_ob1_recv_frag_match (pml_ob1_recvfrag.c:647) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x911668C: mca_pml_ob1_send_request_copy_in_out (pml_ob1_sendreq.c:880) >> ==15577== by 0x9112002: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_recvfrag.c:304) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x911204E: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:273) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116C0F: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:983) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116C40: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:998) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116C66: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1011) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116C81: mca_pml_ob1_send_request_schedule_once (opal_convertor.h:284) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116C8F: mca_pml_ob1_send_request_schedule_once (opal_convertor.h:294) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9079B0B: mca_btl_sm_prepare_src (btl_sm.c:696) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9079B45: mca_btl_sm_prepare_src (btl_sm.c:705) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A249: memcpy (mc_replace_strmem.c:124) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2AB: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3A0: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3B4: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3BA: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3E0: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A427: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116CED: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1041) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116D84: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1079) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9112290: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:273) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115689: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:284) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115696: mca_pml_ob1_recv_request_progress_frag (opal_convertor.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2EC: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A316: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A323: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9115854: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.h:168) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x911A65F: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:273) >> ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116204: mca_pml_ob1_send_request_free (pml_ob1_sendreq.c:117) >> ==15577== by 0x9043B1D: ompi_request_default_wait_all (request.h:350) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A7E44: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9049798: ompi_datatype_create_indexed (ompi_datatype_create_indexed.c:52) >> ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DE0F7: opal_datatype_add (opal_datatype_add.c:153) >> ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) >> ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DE3A2: opal_datatype_add (opal_datatype_add.c:153) >> ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) >> ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DE124: opal_datatype_add (opal_datatype_add.c:161) >> ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) >> ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DE1A4: opal_datatype_add (opal_datatype_add.c:224) >> ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) >> ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DE1B6: opal_datatype_add (opal_datatype_add.c:237) >> ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) >> ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DE1E9: opal_datatype_add (opal_datatype_add.c:247) >> ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) >> ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DE43E: opal_datatype_add (opal_datatype_add.c:348) >> ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) >> ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DE758: opal_datatype_add (opal_datatype_add.c:353) >> ==15577== by 0x90497DE: ompi_datatype_create_indexed (ompi_datatype.h:183) >> ==15577== by 0x90A7F0A: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:460) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91E0BD1: opal_datatype_commit (opal_datatype_optimize.c:279) >> ==15577== by 0x90A7F1F: ompi_coll_tuned_allgatherv_intra_neighborexchange (ompi_datatype.h:158) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91E0621: opal_datatype_optimize_short.constprop.0 (opal_datatype_optimize.c:226) >> ==15577== by 0x91E0C6E: opal_datatype_commit (opal_datatype_optimize.c:308) >> ==15577== by 0x90A7F1F: ompi_coll_tuned_allgatherv_intra_neighborexchange (ompi_datatype.h:158) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91E0672: opal_datatype_optimize_short.constprop.0 (opal_datatype_optimize.c:105) >> ==15577== by 0x91E0C6E: opal_datatype_commit (opal_datatype_optimize.c:308) >> ==15577== by 0x90A7F1F: ompi_coll_tuned_allgatherv_intra_neighborexchange (ompi_datatype.h:158) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9049798: ompi_datatype_create_indexed (ompi_datatype_create_indexed.c:52) >> ==15577== by 0x90A7F84: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_allgatherv.c:470) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DBB85: opal_convertor_prepare_for_send (opal_convertor.c:559) >> ==15577== by 0x91100C0: mca_pml_ob1_isend (opal_convertor.h:237) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x911A702: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:273) >> ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91E0BD1: opal_datatype_commit (opal_datatype_optimize.c:279) >> ==15577== by 0x90A7F99: ompi_coll_tuned_allgatherv_intra_neighborexchange (ompi_datatype.h:158) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91DB975: opal_convertor_prepare_for_recv (opal_convertor.c:523) >> ==15577== by 0x9115BD7: mca_pml_ob1_recv_req_start (opal_convertor.h:258) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A0A51: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:45) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9114645: mca_pml_ob1_recv_request_progress_rndv (opal_convertor.h:284) >> ==15577== by 0x9115EB4: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1026) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A0A51: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:45) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x904F223: PMPI_Allgatherv (pallgatherv.c:93) >> ==15577== by 0x5112D70: AOCreate_Basic (aobasic.c:244) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5112EB4: AOCreate_Basic (aobasic.c:250) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x5112EEA: AOCreate_Basic (aobasic.c:250) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x5112EEA: AOCreate_Basic (aobasic.c:250) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5112EEA: AOCreate_Basic (aobasic.c:250) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510F92E: PetscMemcpy (petscsys.h:1656) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510F97F: PetscMemcpy (petscsys.h:1657) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510F9F8: PetscMemcpy (petscsys.h:1663) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510FA0F: PetscMemcpy (petscsys.h:1663) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A236: memcpy (mc_replace_strmem.c:115) >> ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A25E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3E0: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x510FA85: PetscMemcpy (petscsys.h:1686) >> ==15577== by 0x5112F76: AOCreate_Basic (aobasic.c:252) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE188F: PetscSortInt (sorti.c:75) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE12EA: PetscSortInt_Private (sorti.c:33) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1362: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1385: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE13B1: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE13E1: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE15B0: PetscSortInt_Private (sorti.c:40) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE15D1: PetscSortInt_Private (sorti.c:40) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE1600: PetscSortInt_Private (sorti.c:43) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1636: PetscSortInt_Private (sorti.c:44) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE163D: PetscSortInt_Private (sorti.c:44) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE1647: PetscSortInt_Private (sorti.c:45) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16CE: PetscSortInt_Private (sorti.c:48) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16EF: PetscSortInt_Private (sorti.c:48) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE12EA: PetscSortInt_Private (sorti.c:33) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1362: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1385: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE13B1: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE13E1: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE15B0: PetscSortInt_Private (sorti.c:40) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE15D1: PetscSortInt_Private (sorti.c:40) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE1600: PetscSortInt_Private (sorti.c:43) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1636: PetscSortInt_Private (sorti.c:44) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE1647: PetscSortInt_Private (sorti.c:45) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16CE: PetscSortInt_Private (sorti.c:48) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16EF: PetscSortInt_Private (sorti.c:48) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE12F1: PetscSortInt_Private (sorti.c:34) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE12EA: PetscSortInt_Private (sorti.c:33) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1362: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1385: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE13B1: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE13E1: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1421: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1451: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1596: PetscSortInt_Private (sorti.c:40) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE15B0: PetscSortInt_Private (sorti.c:40) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE15B7: PetscSortInt_Private (sorti.c:40) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE15D1: PetscSortInt_Private (sorti.c:40) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE15D8: PetscSortInt_Private (sorti.c:41) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE1600: PetscSortInt_Private (sorti.c:43) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1615: PetscSortInt_Private (sorti.c:43) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1636: PetscSortInt_Private (sorti.c:44) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE1647: PetscSortInt_Private (sorti.c:45) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16B4: PetscSortInt_Private (sorti.c:48) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16CE: PetscSortInt_Private (sorti.c:48) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16D5: PetscSortInt_Private (sorti.c:48) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16EF: PetscSortInt_Private (sorti.c:48) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE12F1: PetscSortInt_Private (sorti.c:34) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE12FB: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1306: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1421: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1451: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1596: PetscSortInt_Private (sorti.c:40) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE15B7: PetscSortInt_Private (sorti.c:40) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE15D8: PetscSortInt_Private (sorti.c:41) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1615: PetscSortInt_Private (sorti.c:43) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16B4: PetscSortInt_Private (sorti.c:48) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16D5: PetscSortInt_Private (sorti.c:48) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE12FB: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1306: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE163D: PetscSortInt_Private (sorti.c:44) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE138B: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE161C: PetscSortInt_Private (sorti.c:43) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE138B: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE13E7: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE161C: PetscSortInt_Private (sorti.c:43) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE163D: PetscSortInt_Private (sorti.c:44) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE13E7: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE1457: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE130C: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE1457: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE130C: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5112FE1: AOCreate_Basic (aobasic.c:253) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51130D2: AOCreate_Basic (aobasic.c:254) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x511305D: AOCreate_Basic (aobasic.c:255) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510F92E: PetscMemcpy (petscsys.h:1656) >> ==15577== by 0x51130FC: AOCreate_Basic (aobasic.c:258) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510F97F: PetscMemcpy (petscsys.h:1657) >> ==15577== by 0x51130FC: AOCreate_Basic (aobasic.c:258) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510F9F8: PetscMemcpy (petscsys.h:1663) >> ==15577== by 0x51130FC: AOCreate_Basic (aobasic.c:258) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510FA0F: PetscMemcpy (petscsys.h:1663) >> ==15577== by 0x51130FC: AOCreate_Basic (aobasic.c:258) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4EE188F: PetscSortInt (sorti.c:75) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5113258: AOCreate_Basic (aobasic.c:260) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x51131E3: AOCreate_Basic (aobasic.c:261) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x511328F: AOCreate_Basic (aobasic.c:264) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x511328F: AOCreate_Basic (aobasic.c:264) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x511330D: AOCreate_Basic (aobasic.c:269) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x5113340: AOCreate_Basic (aobasic.c:269) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x5113340: AOCreate_Basic (aobasic.c:269) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5113340: AOCreate_Basic (aobasic.c:269) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x511335D: AOCreate_Basic (aobasic.c:269) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x5113394: AOCreate_Basic (aobasic.c:269) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x5113394: AOCreate_Basic (aobasic.c:269) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x5113394: AOCreate_Basic (aobasic.c:269) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5113415: AOCreate_Basic (aobasic.c:270) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510FB60: PetscMemzero (petscsys.h:1716) >> ==15577== by 0x51134B4: AOCreate_Basic (aobasic.c:271) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x510FBBD: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x51134B4: AOCreate_Basic (aobasic.c:271) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x510FBBD: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x51134B4: AOCreate_Basic (aobasic.c:271) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x510FB60: PetscMemzero (petscsys.h:1716) >> ==15577== by 0x5113528: AOCreate_Basic (aobasic.c:272) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5113721: AOCreate_Basic (aobasic.c:273) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x51135CF: AOCreate_Basic (aobasic.c:277) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x5113664: AOCreate_Basic (aobasic.c:278) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x511367A: AOCreate_Basic (aobasic.c:279) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x5113711: AOCreate_Basic (aobasic.c:280) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x511372E: AOCreate_Basic (aobasic.c:282) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x5113819: AOCreate_Basic (aobasic.c:285) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x5113819: AOCreate_Basic (aobasic.c:285) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x511385A: AOCreate_Basic (aobasic.c:285) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x511385A: AOCreate_Basic (aobasic.c:285) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x511391D: AOCreate_Basic (aobasic.c:287) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5113991: AOCreate_Basic (aobasic.c:293) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x43D919: ReadBinary (readbinary3d.c:266) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x43DA17: ReadBinary (readbinary3d.c:268) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D30A5: ISPartitioningCount (iscoloring.c:415) >> ==15577== by 0x43DACB: ReadBinary (readbinary3d.c:272) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D3172: ISPartitioningCount (iscoloring.c:418) >> ==15577== by 0x43DACB: ReadBinary (readbinary3d.c:272) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x50D331A: ISPartitioningCount (iscoloring.c:429) >> ==15577== by 0x43DACB: ReadBinary (readbinary3d.c:272) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5110B49: AOPetscToApplication_Basic (aobasic.c:76) >> ==15577== by 0x51160B0: AOPetscToApplicationIS (ao.c:134) >> ==15577== by 0x43DC9A: ReadBinary (readbinary3d.c:280) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5110AC5: AOPetscToApplication_Basic (aobasic.c:77) >> ==15577== by 0x51160B0: AOPetscToApplicationIS (ao.c:134) >> ==15577== by 0x43DC9A: ReadBinary (readbinary3d.c:280) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5110AE1: AOPetscToApplication_Basic (aobasic.c:77) >> ==15577== by 0x51160B0: AOPetscToApplicationIS (ao.c:134) >> ==15577== by 0x43DC9A: ReadBinary (readbinary3d.c:280) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x5110B1A: AOPetscToApplication_Basic (aobasic.c:78) >> ==15577== by 0x51160B0: AOPetscToApplicationIS (ao.c:134) >> ==15577== by 0x43DC9A: ReadBinary (readbinary3d.c:280) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x511708D: AOApplicationToPetsc (ao.c:262) >> ==15577== by 0x43DD1D: ReadBinary (readbinary3d.c:284) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5110E2A: AOApplicationToPetsc_Basic (aobasic.c:94) >> ==15577== by 0x5117164: AOApplicationToPetsc (ao.c:263) >> ==15577== by 0x43DD1D: ReadBinary (readbinary3d.c:284) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5110DA5: AOApplicationToPetsc_Basic (aobasic.c:95) >> ==15577== by 0x5117164: AOApplicationToPetsc (ao.c:263) >> ==15577== by 0x43DD1D: ReadBinary (readbinary3d.c:284) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x5110DC1: AOApplicationToPetsc_Basic (aobasic.c:95) >> ==15577== by 0x5117164: AOApplicationToPetsc (ao.c:263) >> ==15577== by 0x43DD1D: ReadBinary (readbinary3d.c:284) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x5110DFB: AOApplicationToPetsc_Basic (aobasic.c:96) >> ==15577== by 0x5117164: AOApplicationToPetsc (ao.c:263) >> ==15577== by 0x43DD1D: ReadBinary (readbinary3d.c:284) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x43CD6F: PetscBTCreate (petscbt.h:75) >> ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x43CDB1: PetscBTCreate (petscbt.h:75) >> ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x43CDB1: PetscBTCreate (petscbt.h:75) >> ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x43CDB1: PetscBTCreate (petscbt.h:75) >> ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x43CBCB: PetscMemzero (petscsys.h:1716) >> ==15577== by 0x43CC89: PetscBTMemzero (petscbt.h:39) >> ==15577== by 0x43CDD8: PetscBTCreate (petscbt.h:75) >> ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B0D9: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x43CC28: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x43CC89: PetscBTMemzero (petscbt.h:39) >> ==15577== by 0x43CDD8: PetscBTCreate (petscbt.h:75) >> ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B115: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x43CC28: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x43CC89: PetscBTMemzero (petscbt.h:39) >> ==15577== by 0x43CDD8: PetscBTCreate (petscbt.h:75) >> ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B11D: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x43CC28: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x43CC89: PetscBTMemzero (petscbt.h:39) >> ==15577== by 0x43CDD8: PetscBTCreate (petscbt.h:75) >> ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2B132: memset (mc_replace_strmem.c:1007) >> ==15577== by 0x43CC28: PetscMemzero (petscsys.h:1735) >> ==15577== by 0x43CC89: PetscBTMemzero (petscbt.h:39) >> ==15577== by 0x43CDD8: PetscBTCreate (petscbt.h:75) >> ==15577== by 0x4492E6: DataPartitionVertices_Block (readbinary3d.c:1639) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x905D67C: PMPI_Recv (precv.c:53) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x905D7EC: PMPI_Recv (precv.c:54) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB730: opal_convertor_unpack (opal_convertor.c:285) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3D8: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3C3: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A3E0: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3F8: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A3FD: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A40A: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A427: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x91156B2: mca_pml_ob1_recv_request_progress_frag (pml_ob1_recvreq.c:448) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x910F6B4: mca_pml_ob1_recv (condition.h:99) >> ==15577== by 0x905D73E: PMPI_Recv (precv.c:78) >> ==15577== by 0x4493DA: DataPartitionVertices_Block (readbinary3d.c:1643) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4494C1: DataPartitionVertices_Block (readbinary3d.c:1661) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x4494FA: DataPartitionVertices_Block (readbinary3d.c:1661) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x4494FA: DataPartitionVertices_Block (readbinary3d.c:1661) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x4494FA: DataPartitionVertices_Block (readbinary3d.c:1661) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x449A5C: DataPartitionVertices_Block (readbinary3d.c:1664) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44959C: DataPartitionVertices_Block (readbinary3d.c:1665) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x43CE1A: PetscBTLookupSet (petscbt.h:84) >> ==15577== by 0x4495F7: DataPartitionVertices_Block (readbinary3d.c:1668) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x43CE5D: PetscBTLookupSet (petscbt.h:86) >> ==15577== by 0x4495F7: DataPartitionVertices_Block (readbinary3d.c:1668) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4495FA: DataPartitionVertices_Block (readbinary3d.c:1668) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x905F748: PMPI_Send (psend.c:55) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x905F8DA: PMPI_Send (psend.c:64) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91103AA: mca_pml_ob1_send (pml_ob1_isend.c:108) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x91104AF: mca_pml_ob1_send (pml_ob1_sendreq.h:360) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2BC: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2C1: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A308: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A316: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB628: opal_convertor_pack (opal_convertor.c:238) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A24F: memcpy (mc_replace_strmem.c:127) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2BC: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2C1: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A2EC: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A308: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A30B: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A316: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A323: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A353: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x9079B72: mca_btl_sm_prepare_src (btl_sm.c:713) >> ==15577== by 0x9116CD4: mca_pml_ob1_send_request_schedule_once (bml.h:339) >> ==15577== by 0x9112257: mca_pml_ob1_recv_frag_callback_ack (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BA7C: mca_btl_sm_component_progress (btl_sm_component.c:629) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x9110A04: mca_pml_ob1_send (condition.h:99) >> ==15577== by 0x905F7D1: PMPI_Send (psend.c:75) >> ==15577== by 0x449AFB: DataPartitionVertices_Block (readbinary3d.c:1694) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F23EC3: PetscTrFreeDefault (mtr.c:279) >> ==15577== by 0x43CCC9: PetscBTDestroy (petscbt.h:44) >> ==15577== by 0x449B6F: DataPartitionVertices_Block (readbinary3d.c:1696) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F2411A: PetscTrFreeDefault (mtr.c:298) >> ==15577== by 0x43CCC9: PetscBTDestroy (petscbt.h:44) >> ==15577== by 0x449B6F: DataPartitionVertices_Block (readbinary3d.c:1696) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x449BE1: DataPartitionVertices_Block (readbinary3d.c:1702) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2357E: PetscTrMallocDefault (mtr.c:179) >> ==15577== by 0x449C28: DataPartitionVertices_Block (readbinary3d.c:1702) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4F237F2: PetscTrMallocDefault (mtr.c:202) >> ==15577== by 0x449C28: DataPartitionVertices_Block (readbinary3d.c:1702) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4F2381E: PetscTrMallocDefault (mtr.c:205) >> ==15577== by 0x449C28: DataPartitionVertices_Block (readbinary3d.c:1702) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x43CD6F: PetscBTCreate (petscbt.h:75) >> ==15577== by 0x449CA5: DataPartitionVertices_Block (readbinary3d.c:1703) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x44A129: DataPartitionVertices_Block (readbinary3d.c:1705) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x449D2D: DataPartitionVertices_Block (readbinary3d.c:1706) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x43CE1A: PetscBTLookupSet (petscbt.h:84) >> ==15577== by 0x449D88: DataPartitionVertices_Block (readbinary3d.c:1708) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x43CE5D: PetscBTLookupSet (petscbt.h:86) >> ==15577== by 0x449D88: DataPartitionVertices_Block (readbinary3d.c:1708) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x449DCB: DataPartitionVertices_Block (readbinary3d.c:1708) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x449D8B: DataPartitionVertices_Block (readbinary3d.c:1708) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116C0F: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:983) >> ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116C40: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:998) >> ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116C66: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1011) >> ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116C81: mca_pml_ob1_send_request_schedule_once (opal_convertor.h:284) >> ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116C8F: mca_pml_ob1_send_request_schedule_once (opal_convertor.h:294) >> ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116CED: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1041) >> ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x9116D84: mca_pml_ob1_send_request_schedule_once (pml_ob1_sendreq.c:1079) >> ==15577== by 0x911A6D7: mca_pml_ob1_frag_completion (pml_ob1_sendreq.h:294) >> ==15577== by 0x907BCDC: mca_btl_sm_component_progress (btl_sm_component.c:651) >> ==15577== by 0x91E6C39: opal_progress (opal_progress.c:207) >> ==15577== by 0x90439BC: ompi_request_default_wait_all (condition.h:99) >> ==15577== by 0x90A0AF3: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:54) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A333: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9113A26: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.c:632) >> ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A0A51: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:45) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A368: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB784: opal_convertor_unpack (opal_convertor.c:295) >> ==15577== by 0x9113A26: mca_pml_ob1_recv_request_progress_match (pml_ob1_recvreq.c:632) >> ==15577== by 0x9115C67: mca_pml_ob1_recv_req_start (pml_ob1_recvreq.c:1022) >> ==15577== by 0x910F416: mca_pml_ob1_irecv (pml_ob1_irecv.c:76) >> ==15577== by 0x90A0A51: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:45) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x911002C: mca_pml_ob1_isend (pml_ob1_sendreq.h:372) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A435: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A43B: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A450: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A456: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A464: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A47E: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112CA1: AOCreate_Basic (aobasic.c:243) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Conditional jump or move depends on uninitialised value(s) >> ==15577== at 0x4C2A32D: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112D70: AOCreate_Basic (aobasic.c:244) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4C2A350: memcpy (mc_replace_strmem.c:878) >> ==15577== by 0x91DB674: opal_convertor_pack (opal_convertor.c:251) >> ==15577== by 0x907999E: mca_btl_sm_sendi (btl_sm.c:849) >> ==15577== by 0x91186A2: mca_pml_ob1_send_request_start_copy (bml.h:304) >> ==15577== by 0x911003D: mca_pml_ob1_isend (pml_ob1_sendreq.h:375) >> ==15577== by 0x90A0AD4: ompi_coll_tuned_sendrecv_actual (coll_tuned_util.c:50) >> ==15577== by 0x90A8006: ompi_coll_tuned_allgatherv_intra_neighborexchange (coll_tuned_util.h:57) >> ==15577== by 0x904F2F3: PMPI_Allgatherv (pallgatherv.c:139) >> ==15577== by 0x5112D70: AOCreate_Basic (aobasic.c:244) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE14BF: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE14E2: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1522: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1552: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1689: PetscSortInt_Private (sorti.c:46) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16A6: PetscSortInt_Private (sorti.c:46) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1689: PetscSortInt_Private (sorti.c:46) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16A6: PetscSortInt_Private (sorti.c:46) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE14BF: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE14E2: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1522: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1552: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== by 0x44A2F9: DataPartitionVertices_Block (readbinary3d.c:1751) >> ==15577== by 0x43DE25: ReadBinary (readbinary3d.c:307) >> ==15577== by 0x42F5BC: LoadGrid (loadgrid3d.c:726) >> ==15577== by 0x405D55: main (fsi3d.c:55) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE14BF: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE14E2: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1522: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1552: PetscSortInt_Private (sorti.c:39) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE165C: PetscSortInt_Private (sorti.c:46) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1689: PetscSortInt_Private (sorti.c:46) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE168C: PetscSortInt_Private (sorti.c:46) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE16A6: PetscSortInt_Private (sorti.c:46) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== by 0x5113DEA: AOCreateBasic (aobasic.c:342) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1316: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1321: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1329: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1338: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE165C: PetscSortInt_Private (sorti.c:46) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE168C: PetscSortInt_Private (sorti.c:46) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1316: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1321: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1329: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== >> ==15577== Use of uninitialised value of size 8 >> ==15577== at 0x4EE1338: PetscSortInt_Private (sorti.c:35) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE173B: PetscSortInt_Private (sorti.c:50) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE1708: PetscSortInt_Private (sorti.c:49) >> ==15577== by 0x4EE19A1: PetscSortInt (sorti.c:85) >> ==15577== by 0x5113167: AOCreate_Basic (aobasic.c:259) >> ==15577== by 0x511AEE4: AOSetType (aoreg.c:48) >> ==15577== by 0x5114269: AOCreateBasicIS (aobasic.c:384) >> ==15577== From jedbrown at mcs.anl.gov Tue Nov 5 23:22:56 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Tue, 05 Nov 2013 22:22:56 -0700 Subject: [petsc-users] Debug AOCreateBasic In-Reply-To: <5279D10A.8080204@gmail.com> References: <52733BB5.5030203@gmail.com> <8761scheg2.fsf@mcs.anl.gov> <52734F7B.2060002@gmail.com> <8761sbpfq0.fsf@mcs.anl.gov> <5279D10A.8080204@gmail.com> Message-ID: <877gcmxfwv.fsf@mcs.anl.gov> Rongliang Chen writes: > Hi Jed, > > I tested my code on a locally 16-cores workstation. Both the MPICH and > OPENMPI version work well for the 30 million unstructured meshes case > (this workstation has 256G memory). Then I would start thinking about an MPI problem and see if you can configure to get a core dump. > My code is also valgrind-clean on this workstation. Good. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From dave.mayhem23 at gmail.com Wed Nov 6 02:07:43 2013 From: dave.mayhem23 at gmail.com (Dave May) Date: Wed, 6 Nov 2013 09:07:43 +0100 Subject: [petsc-users] Memory and Speed Issue of using MG as preconditioner In-Reply-To: <5279B59D.9030804@gmail.com> References: <5279B59D.9030804@gmail.com> Message-ID: Hey Alan, 1/ One difference in the memory footprint is likely coming from your coarse grid solver which is redundant LU. The 2 level case has a coarse grid problem with 70785 unknowns whilst the 5 level case has a coarse grid problem with 225 unknowns. 2/ The solve time difference will be affected by your coarse grid size. Add the command line argument -pc_mg_log to profile the setup time spent on the coarse grid and all other levels. See http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/PC/PCMG.html 3/ You can change the smoother on all levels by using the command line argument with the appropriate prefix, eg -mg_levels_ksp_type cg Note the prefix is displayed in the result of -ksp_view Also, your mesh size can be altered at run time using arguments like -da_grid_x 5 You shouldn't have to modify the source code each time See http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/DM/DMDACreate3d.html Cheers, Dave On 6 November 2013 04:21, Alan wrote: > Dear all, > I hope you're having a nice day. > Recently, I came across a problem on using MG as preconditioner. > Basically, to achieve the same finest mesh with pc_type = mg, the memory > usage for -da_refine 2 is much more than that for -da_refine 5. To my > limited knowledge, more refinement should consume more memory, which is > contradict to the behavior of pc_type = mg in PETsc. > Here, I provide two output files. They are all from > /src/ksp/ksp/example/tutorial/ex45.c with 32 processes. > The execute file for out-level2 is > mpiexec -np 32 ./ex45 -pc_type mg -ksp_type cg -da_refine 2 > -pc_mg_galerkin -ksp_rtol 1.0e-7 -mg_levels_pc_type jacobi > -mg_levels_ksp_type chebyshev -dm_view -log_summary -pc_mg_log > -pc_mg_monitor -ksp_view -ksp_monitor > out & > and in ex45.c, KSPCreate is changed as: > ierr = > > DMDACreate3d(PETSC_COMM_WORLD,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_STENCIL_STAR,-65,-33,-33,PETSC_DECIDE,PETSC_DECIDE,PETSC_DECIDE,1,1,0,0,0,&da);CHKERRQ(ierr); > On the other hand, the execute file for out-level5 is > mpiexec -np 32 ./ex45 -pc_type mg -ksp_type cg -da_refine 5 > -pc_mg_galerkin -ksp_rtol 1.0e-7 -mg_levels_pc_type jacobi > -mg_levels_ksp_type chebyshev -dm_view -log_summary -pc_mg_log > -pc_mg_monitor -ksp_view -ksp_monitor > out & > and in ex45.c, KSPCreate is changed as: > ierr = > > DMDACreate3d(PETSC_COMM_WORLD,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_STENCIL_STAR,-9,-5,-5,PETSC_DECIDE,PETSC_DECIDE,PETSC_DECIDE,1,1,0,0,0,&da);CHKERRQ(ierr); > In summary, the final finest meshes obtained for both cases are > 257*129*129 as documented in both files. However, the out-level2 shows > that the Matrix requested 822871308 memory while out-level5 only need > 36052932. > Furthermore, although the total iterations for KSP solver are shown as 5 > times in both files. the wall time elapsed for out-level2 is around > 150s, while out-level5 is about 4.7s. > At last, there is a minor question. In both files, under 'Down solver > (pre-smoother) on level 1' and 'Down solver (pre-smoother) on level 2', > the type of "KSP Object: (mg_levels_1_est_)" and "KSP Object: > (mg_levels_2_est_)" are all 'gmres'. Since I'm using uniformly Cartesian > mesh, would it be helpful to speed up the solver if the 'gmres' is > replaced by 'cg' here? If so, which PETSc option can change the type of > KSP object. > > sincerely appreciate, > Alan > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From lu_qin_2000 at yahoo.com Wed Nov 6 09:57:56 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Wed, 6 Nov 2013 07:57:56 -0800 (PST) Subject: [petsc-users] Directory name with space used by hypre configure In-Reply-To: References: <1383666107.29402.YahooMailNeo@web160204.mail.bf1.yahoo.com> <6BC19C6F-1B33-4217-BD3D-9F21DD41A35B@mcs.anl.gov> Message-ID: <1383753476.10521.YahooMailNeo@web160205.mail.bf1.yahoo.com> I tried to put the MPICH2 under a directory without space, but the Intel C++ compiler icl could not find the mpi.h file even if I set the path for the compiler (-I/cygdrive/c/cygwin_cache/mpich2x64/include/). See the error message: ? =================== Making struct_mv ... make[1]: Entering directory '/cygdrive/c/cygwin_cache/hypre-2.8.0b/src/struct_mv' icl -O2 -I/cygdrive/c/cygwin_cache/mpich2x64/include/ -DHAVE_CONFIG_H -I.. -I. -I./.. -I./../utilities?? -c assumed_part.c Intel(R) C++ Intel(R) 64 Compiler Professional for applications running on Intel(R) 64, Version 11.1????Build 20091012 Package ID: w_cproc_p_11.1.051 Copyright (C) 1985-2009 Intel Corporation.??All rights reserved. assumed_part.c C:\cygwin_cache\hypre-2.8.0b\src\utilities\HYPRE_utilities.h(26): catastrophic error: could not open source file "mpi.h" #include "mpi.h" ==================== ? Do you have any idea what can be wrong? ? Thanks, Qin On Tuesday, November 5, 2013 10:05 AM, Satish Balay wrote: Alctually you'll have to look at hypre docs for installation instructions for windows. [download-hypre does not work] Satish On Tue, 5 Nov 2013, Barry Smith wrote: > >? ? Reinstall the MPI in a location without spaces in the directory name. > > On Nov 5, 2013, at 9:41 AM, Qin Lu wrote: > > > Hello, > > I am trying to build hypre in Windows 7 (using cygwin) since it will be included in my PETSc lib. The configure options of hypre contains directory such as: > > ./configure --with-MPI-include="/cygdrive/c/Program Files/mpich2x64/include" > > Somehow hypre's configure took this directory as two separate directories ("/cygdrive/c/Program" and "Files/mpich2x64/include") since there is a space in the middle. (Strangely, PETSc's configure read this directory correctly). I tried to replacing the space with "\ " but it still did not work. > > How can I make hypre's configure read this directory correctly? > > Thanks for your suggestions, > > Qin > > ??? From balay at mcs.anl.gov Wed Nov 6 10:34:46 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Wed, 6 Nov 2013 10:34:46 -0600 (CST) Subject: [petsc-users] Directory name with space used by hypre configure In-Reply-To: <1383753476.10521.YahooMailNeo@web160205.mail.bf1.yahoo.com> References: <1383666107.29402.YahooMailNeo@web160204.mail.bf1.yahoo.com> <6BC19C6F-1B33-4217-BD3D-9F21DD41A35B@mcs.anl.gov> <1383753476.10521.YahooMailNeo@web160205.mail.bf1.yahoo.com> Message-ID: icl does not understand cygwin paths [like /cygdrive/c./....] As mentioned before - you should check hypre installation instructions on how to install it on windows. Satish On Wed, 6 Nov 2013, Qin Lu wrote: > I tried to put the MPICH2 under a directory without space, but the Intel C++ compiler icl could not find the mpi.h file even if I set the path for the compiler (-I/cygdrive/c/cygwin_cache/mpich2x64/include/). See the error message: > ? > =================== > Making struct_mv ... > make[1]: Entering directory '/cygdrive/c/cygwin_cache/hypre-2.8.0b/src/struct_mv' > icl -O2 -I/cygdrive/c/cygwin_cache/mpich2x64/include/ -DHAVE_CONFIG_H -I.. -I. -I./.. -I./../utilities?? -c assumed_part.c > Intel(R) C++ Intel(R) 64 Compiler Professional for applications running on Intel(R) 64, Version 11.1????Build 20091012 Package ID: w_cproc_p_11.1.051 > Copyright (C) 1985-2009 Intel Corporation.??All rights reserved. > assumed_part.c > C:\cygwin_cache\hypre-2.8.0b\src\utilities\HYPRE_utilities.h(26): catastrophic error: could not open source file "mpi.h" > #include "mpi.h" > ==================== > ? > Do you have any idea what can be wrong? > ? > Thanks, > Qin > > > On Tuesday, November 5, 2013 10:05 AM, Satish Balay wrote: > > Alctually you'll have to look at hypre docs for installation > instructions for windows. [download-hypre does not work] > > Satish > > > > On Tue, 5 Nov 2013, Barry Smith wrote: > > > > >? ? Reinstall the MPI in a location without spaces in the directory name. > > > > On Nov 5, 2013, at 9:41 AM, Qin Lu wrote: > > > > > Hello, > > > I am trying to build hypre in Windows 7 (using cygwin) since it will be included in my PETSc lib. The configure options of hypre contains directory such as: > > > ./configure --with-MPI-include="/cygdrive/c/Program Files/mpich2x64/include" > > > Somehow hypre's configure took this directory as two separate directories ("/cygdrive/c/Program" and "Files/mpich2x64/include") since there is a space in the middle. (Strangely, PETSc's configure read this directory correctly). I tried to replacing the space with "\ " but it still did not work. > > > How can I make hypre's configure read this directory correctly? > > > Thanks for your suggestions, > > > Qin > > > > ??? > From lu_qin_2000 at yahoo.com Wed Nov 6 11:25:41 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Wed, 6 Nov 2013 09:25:41 -0800 (PST) Subject: [petsc-users] Directory name with space used by hypre configure In-Reply-To: References: <1383666107.29402.YahooMailNeo@web160204.mail.bf1.yahoo.com> <6BC19C6F-1B33-4217-BD3D-9F21DD41A35B@mcs.anl.gov> <1383753476.10521.YahooMailNeo@web160205.mail.bf1.yahoo.com> Message-ID: <1383758741.88660.YahooMailNeo@web160201.mail.bf1.yahoo.com> I did read hypre's instructions (no offense, it is not as clear as PETSc's documents), which?gives a configure option example for gcc and g++ using the cygwin path, such as ./configure CC=gcc CXX=g++ CXXFLAGS="O2 -I/cygdrive/c/MPICH2/include/" Does this mean that gcc/g++ understand cygwin paths, while icl does not? But icl worked for?me building PETSc using cygwin paths. Thanks, Qin ??? On Wednesday, November 6, 2013 10:34 AM, Satish Balay wrote: icl does not understand cygwin paths [like /cygdrive/c./....] As mentioned before - you should check hypre installation instructions on how to install it on windows. Satish On Wed, 6 Nov 2013, Qin Lu wrote: > I tried to put the MPICH2 under a directory without space, but the Intel C++ compiler icl could not find the mpi.h file even if I set the path for the compiler (-I/cygdrive/c/cygwin_cache/mpich2x64/include/). See the error message: > ? > =================== > Making struct_mv ... > make[1]: Entering directory '/cygdrive/c/cygwin_cache/hypre-2.8.0b/src/struct_mv' > icl -O2 -I/cygdrive/c/cygwin_cache/mpich2x64/include/ -DHAVE_CONFIG_H -I.. -I. -I./.. -I./../utilities?? -c assumed_part.c > Intel(R) C++ Intel(R) 64 Compiler Professional for applications running on Intel(R) 64, Version 11.1????Build 20091012 Package ID: w_cproc_p_11.1.051 > Copyright (C) 1985-2009 Intel Corporation.??All rights reserved. > assumed_part.c > C:\cygwin_cache\hypre-2.8.0b\src\utilities\HYPRE_utilities.h(26): catastrophic error: could not open source file "mpi.h" > #include "mpi.h" > ==================== > ? > Do you have any idea what can be wrong? > ? > Thanks, > Qin > > > On Tuesday, November 5, 2013 10:05 AM, Satish Balay wrote: > > Alctually you'll have to look at hypre docs for installation > instructions for windows. [download-hypre does not work] > > Satish > > > > On Tue, 5 Nov 2013, Barry Smith wrote: > > > > >? ? Reinstall the MPI in a location without spaces in the directory name. > > > > On Nov 5, 2013, at 9:41 AM, Qin Lu wrote: > > > > > Hello, > > > I am trying to build hypre in Windows 7 (using cygwin) since it will be included in my PETSc lib. The configure options of hypre contains directory such as: > > > ./configure --with-MPI-include="/cygdrive/c/Program Files/mpich2x64/include" > > > Somehow hypre's configure took this directory as two separate directories ("/cygdrive/c/Program" and "Files/mpich2x64/include") since there is a space in the middle. (Strangely, PETSc's configure read this directory correctly). I tried to replacing the space with "\ " but it still did not work. > > > How can I make hypre's configure read this directory correctly? > > > Thanks for your suggestions, > > > Qin > > > > ??? > From balay at mcs.anl.gov Wed Nov 6 11:41:44 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Wed, 6 Nov 2013 11:41:44 -0600 (CST) Subject: [petsc-users] Directory name with space used by hypre configure In-Reply-To: <1383758741.88660.YahooMailNeo@web160201.mail.bf1.yahoo.com> References: <1383666107.29402.YahooMailNeo@web160204.mail.bf1.yahoo.com> <6BC19C6F-1B33-4217-BD3D-9F21DD41A35B@mcs.anl.gov> <1383753476.10521.YahooMailNeo@web160205.mail.bf1.yahoo.com> <1383758741.88660.YahooMailNeo@web160201.mail.bf1.yahoo.com> Message-ID: On Wed, 6 Nov 2013, Qin Lu wrote: > I did read hypre's instructions (no offense, it is not as clear as PETSc's documents), which?gives a configure option example for gcc and g++ using the cygwin path, such as hypre/README has: ========================================== HYPRE Installation Information using CMake ========================================== CMake is another way to build HYPRE that is particularly useful for building the code on Windows machines. CMake provides a uniform interface for setting configuration options on different platforms. It does not actually build the code, but generates input for other "native" build systems such as Make (Unix platforms) or Visual Studio (Windows). Here are the basic steps: 1. First, ensure that CMake version 2.8.8 or later is installed on the system. 2. To build the library, run CMake on the top-level HYPRE source directory to generate files appropriate for the native build system. To prevent writing over the Makefiles in HYPRE's configure/make system above, only out-of-source builds are currently allowed with CMake. Directories 'src/cmbuild' and 'src/test/cmbuild' are provided in the release for convenience, but alternative build directories may be created by the user: - Unix: From the 'src/cmbuild' directory, type 'cmake ..'. - Windows: Set the source and build directories to 'src' and 'src/cmbuild', then click on 'Configure' following by 'Generate'. 3. To complete the build of the library, compile with the native build system: - Unix: From the 'src/cmbuild' directory, type 'make install'. - Windows Visual Studio: Open the 'hypre' VS solution file generated by CMake and build the 'ALL_BUILD' target, then the 'INSTALL' target. Optional Features: Various configuration options can be set from within CMake. Change these as appropriate, then reconfigure/generate: - Unix: From the 'src/cmbuild' directory, type 'ccmake ..', change options, then type 'c' then 'q'. - Windows VS: Change options, then click on 'Configure' then 'Generate'. ===================== > > ./configure CC=gcc CXX=g++ CXXFLAGS="O2 -I/cygdrive/c/MPICH2/include/" > > Does this mean that gcc/g++ understand cygwin paths, while icl does not? yes. > But icl worked for?me building PETSc using cygwin paths. Because we provide a wrapper 'win32fe' that translates the paths and other options to icc notation. However hypre's configure does not work with win32fe. Satish From zhenglun.wei at gmail.com Wed Nov 6 12:35:12 2013 From: zhenglun.wei at gmail.com (Alan Z. Wei) Date: Wed, 06 Nov 2013 12:35:12 -0600 Subject: [petsc-users] Memory and Speed Issue of using MG as preconditioner In-Reply-To: References: <5279B59D.9030804@gmail.com> Message-ID: <527A8BE0.9020101@gmail.com> Thanks Dave, I further simulated the problem with -pc_mg_log and output these files in the attachments. I found that the smoothing process of the last level always consumes the most time, i.e. 'MGSmooth Level 5' in out-level5 and "MGSmooth Level 2' in out-level2. However, as I tested several other -mg_levels_pc_type such as 'bjacobi', 'asm' etc. The default one, which is 'jacobi', actually works the best. Therefore, I decide to keep using it. However, do you have any suggestions to speed up this smoothing process other than changing -mg_levels_pc_type? Also, as you suggested to change -mg_levels_ksp_type, it does not influence much if replacing 'chebyshev' by 'cg'. However, this part never change while I modify '-mg_levels_ksp_type': PC Object: (mg_coarse_) 32 MPI processes type: redundant Redundant preconditioner: First (color=0) of 32 PCs follows KSP Object: (mg_coarse_redundant_) 1 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_redundant_) 1 MPI processes type: lu LU: out-of-place factorization As you mentioned that the redundant LU for the coarse grid solver primarily cause the large memory request for the 2-level case. How could I change the coarse grid solver to reduce the memory requirement or speed up the solver. thanks again, Alan > Hey Alan, > > 1/ One difference in the memory footprint is likely coming from your > coarse grid solver which is redundant LU. > The 2 level case has a coarse grid problem with 70785 unknowns whilst > the 5 level case has a coarse grid problem with 225 unknowns. > > 2/ The solve time difference will be affected by your coarse grid > size. Add the command line argument > -pc_mg_log > to profile the setup time spent on the coarse grid and all other levels. > See > http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/PC/PCMG.html > > 3/ You can change the smoother on all levels by using the command line > argument with the appropriate prefix, eg > -mg_levels_ksp_type cg > Note the prefix is displayed in the result of -ksp_view > > Also, your mesh size can be altered at run time using arguments like > -da_grid_x 5 > You shouldn't have to modify the source code each time > See > http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/DM/DMDACreate3d.html > > > Cheers, > Dave > > > On 6 November 2013 04:21, Alan > wrote: > > Dear all, > I hope you're having a nice day. > Recently, I came across a problem on using MG as preconditioner. > Basically, to achieve the same finest mesh with pc_type = mg, the > memory > usage for -da_refine 2 is much more than that for -da_refine 5. To my > limited knowledge, more refinement should consume more memory, > which is > contradict to the behavior of pc_type = mg in PETsc. > Here, I provide two output files. They are all from > /src/ksp/ksp/example/tutorial/ex45.c with 32 processes. > The execute file for out-level2 is > mpiexec -np 32 ./ex45 -pc_type mg -ksp_type cg -da_refine 2 > -pc_mg_galerkin -ksp_rtol 1.0e-7 -mg_levels_pc_type jacobi > -mg_levels_ksp_type chebyshev -dm_view -log_summary -pc_mg_log > -pc_mg_monitor -ksp_view -ksp_monitor > out & > and in ex45.c, KSPCreate is changed as: > ierr = > DMDACreate3d(PETSC_COMM_WORLD,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_STENCIL_STAR,-65,-33,-33,PETSC_DECIDE,PETSC_DECIDE,PETSC_DECIDE,1,1,0,0,0,&da);CHKERRQ(ierr); > On the other hand, the execute file for out-level5 is > mpiexec -np 32 ./ex45 -pc_type mg -ksp_type cg -da_refine 5 > -pc_mg_galerkin -ksp_rtol 1.0e-7 -mg_levels_pc_type jacobi > -mg_levels_ksp_type chebyshev -dm_view -log_summary -pc_mg_log > -pc_mg_monitor -ksp_view -ksp_monitor > out & > and in ex45.c, KSPCreate is changed as: > ierr = > DMDACreate3d(PETSC_COMM_WORLD,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_STENCIL_STAR,-9,-5,-5,PETSC_DECIDE,PETSC_DECIDE,PETSC_DECIDE,1,1,0,0,0,&da);CHKERRQ(ierr); > In summary, the final finest meshes obtained for both cases are > 257*129*129 as documented in both files. However, the out-level2 shows > that the Matrix requested 822871308 memory while out-level5 only need > 36052932. > Furthermore, although the total iterations for KSP solver are > shown as 5 > times in both files. the wall time elapsed for out-level2 is around > 150s, while out-level5 is about 4.7s. > At last, there is a minor question. In both files, under 'Down solver > (pre-smoother) on level 1' and 'Down solver (pre-smoother) on > level 2', > the type of "KSP Object: (mg_levels_1_est_)" and "KSP Object: > (mg_levels_2_est_)" are all 'gmres'. Since I'm using uniformly > Cartesian > mesh, would it be helpful to speed up the solver if the 'gmres' is > replaced by 'cg' here? If so, which PETSc option can change the > type of > KSP object. > > sincerely appreciate, > Alan > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- Processor [0] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 0 33 Processor [1] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 0 33 Processor [2] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 0 33 Processor [3] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 0 33 Processor [4] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 0 33 Processor [5] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 0 33 Processor [6] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 0 33 Processor [7] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 0 33 Processor [8] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 33 65 Processor [9] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 33 65 Processor [10] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 33 65 Processor [11] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 33 65 Processor [12] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 33 65 Processor [13] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 33 65 Processor [14] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 33 65 Processor [15] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 33 65 Processor [16] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 65 97 Processor [17] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 65 97 Processor [18] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 65 97 Processor [19] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 65 97 Processor [20] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 65 97 Processor [21] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 65 97 Processor [22] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 65 97 Processor [23] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 65 97 Processor [24] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 97 129 Processor [25] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 97 129 Processor [26] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 97 129 Processor [27] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 97 129 Processor [28] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 97 129 Processor [29] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 97 129 Processor [30] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 97 129 Processor [31] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 97 129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 Processor [0] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 0 17 Processor [1] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 0 17 Processor [2] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 0 17 Processor [3] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 0 17 Processor [4] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 0 17 Processor [5] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 0 17 Processor [6] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 0 17 Processor [7] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 0 17 Processor [8] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 17 33 Processor [9] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 17 33 Processor [10] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 17 33 Processor [11] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 17 33 Processor [12] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 17 33 Processor [13] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 17 33 Processor [14] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 17 33 Processor [15] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 17 33 Processor [16] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 33 49 Processor [17] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 33 49 Processor [18] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 33 49 Processor [19] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 33 49 Processor [20] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 33 49 Processor [21] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 33 49 Processor [22] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 33 49 Processor [23] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 33 49 Processor [24] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 49 65 Processor [25] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 49 65 Processor [26] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 49 65 Processor [27] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 49 65 Processor [28] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 49 65 Processor [29] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 49 65 Processor [30] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 49 65 Processor [31] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 49 65 Processor [0] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 0 9 Processor [1] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 0 9 Processor [2] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 0 9 Processor [3] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 0 9 Processor [4] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 0 9 Processor [5] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 0 9 Processor [6] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 0 9 Processor [7] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 0 9 Processor [8] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 9 17 Processor [9] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 9 17 Processor [10] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 9 17 Processor [11] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 9 17 Processor [12] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 9 17 Processor [13] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 9 17 Processor [14] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 9 17 Processor [15] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 9 17 Processor [16] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 17 25 Processor [17] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 17 25 Processor [18] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 17 25 Processor [19] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 17 25 Processor [20] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 17 25 Processor [21] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 17 25 Processor [22] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 17 25 Processor [23] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 17 25 Processor [24] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 25 33 Processor [25] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 25 33 Processor [26] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 25 33 Processor [27] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 25 33 Processor [28] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 25 33 Processor [29] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 25 33 Processor [30] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 25 33 Processor [31] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 25 33 0 KSP Residual norm 2.036594349596e+03 1 KSP Residual norm 8.756270777762e+01 2 KSP Residual norm 3.092374574522e+00 3 KSP Residual norm 1.220382147945e-01 4 KSP Residual norm 2.871729837207e-02 KSP Object: 32 MPI processes type: cg maximum iterations=10000 tolerances: relative=1e-07, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using PRECONDITIONED norm type for convergence test PC Object: 32 MPI processes type: mg MG: type is MULTIPLICATIVE, levels=3 cycles=v Cycles per PCApply=1 Using Galerkin computed coarse grid matrices Coarse grid solver -- level ------------------------------- KSP Object: (mg_coarse_) 32 MPI processes type: preonly maximum iterations=1, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_) 32 MPI processes type: redundant Redundant preconditioner: First (color=0) of 32 PCs follows KSP Object: (mg_coarse_redundant_) 1 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_redundant_) 1 MPI processes type: lu LU: out-of-place factorization tolerance for zero pivot 2.22045e-14 using diagonal shift on blocks to prevent zero pivot matrix ordering: nd factor fill ratio given 5, needed 35.0339 Factored matrix follows: Matrix Object: 1 MPI processes type: seqaij rows=70785, cols=70785 package used to perform factorization: petsc total: nonzeros=63619375, allocated nonzeros=63619375 total number of mallocs used during MatSetValues calls =0 not using I-node routines linear system matrix = precond matrix: Matrix Object: 1 MPI processes type: seqaij rows=70785, cols=70785 total: nonzeros=1815937, allocated nonzeros=1911195 total number of mallocs used during MatSetValues calls =0 not using I-node routines linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=70785, cols=70785 total: nonzeros=1815937, allocated nonzeros=1815937 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Down solver (pre-smoother) on level 1 ------------------------------- KSP Object: (mg_levels_1_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.231542, max = 2.54696 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_1_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_1_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=545025, cols=545025 total: nonzeros=14340865, allocated nonzeros=14340865 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_1_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=545025, cols=545025 total: nonzeros=14340865, allocated nonzeros=14340865 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 2 ------------------------------- KSP Object: (mg_levels_2_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.155706, max = 1.71277 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_2_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_2_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_2_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 Up solver (post-smoother) same as down solver (pre-smoother) linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 Residual norm 0.000941649 Total Time Elapsed: 570.948138 Total Time Elapsed: 571.226165 Total Time Elapsed: 571.226142 Total Time Elapsed: 571.227743 Total Time Elapsed: 571.155077 Total Time Elapsed: 571.230839 Total Time Elapsed: 571.230841 Total Time Elapsed: 571.379246 Total Time Elapsed: 571.231696 Total Time Elapsed: 571.162397 Total Time Elapsed: 570.963429 Total Time Elapsed: 571.234322 Total Time Elapsed: 570.952416 Total Time Elapsed: 571.235487 Total Time Elapsed: 570.955526 Total Time Elapsed: 570.952019 Total Time Elapsed: 570.956828 Total Time Elapsed: 571.384060 Total Time Elapsed: 571.168546 Total Time Elapsed: 570.955564 Total Time Elapsed: 571.386214 Total Time Elapsed: 571.167004 Total Time Elapsed: 570.958710 Total Time Elapsed: 571.389300 Total Time Elapsed: 571.173462 Total Time Elapsed: 571.394695 Total Time Elapsed: 571.169628 Total Time Elapsed: 571.397931 Total Time Elapsed: 571.401814 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- Total Time Elapsed: 571.401327 ./ex45 on a linux-gnu-c-nodebug named n042 with 32 processors, by zlwei Wed Nov 6 11:48:06 2013 Using Petsc Development GIT revision: d696997672013bb4513d3ff57c61cc10e09b71f6 GIT Date: 2013-06-13 10:28:37 -0500 Max Max/Min Avg Total Total Time Elapsed: 571.186741 Total Time Elapsed: 571.188217 Time (sec): 5.710e+02 1.00007 5.709e+02 Objects: 1.750e+02 1.00000 1.750e+02 Flops: 7.470e+10 1.00022 7.468e+10 2.390e+12 Flops/sec: 1.308e+08 1.00018 1.308e+08 4.186e+09 MPI Messages: 1.594e+03 1.77506 1.234e+03 3.948e+04 MPI Message Lengths: 3.375e+07 1.17496 2.581e+04 1.019e+09 MPI Reductions: 3.500e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.8451e+02 49.8% 2.3636e+12 98.9% 1.188e+04 30.1% 1.963e+04 76.0% 2.430e+02 69.4% 1: MG Apply: 2.8643e+02 50.2% 2.6346e+10 1.1% 2.760e+04 69.9% 6.185e+03 24.0% 1.060e+02 30.3% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage KSPSetUp 5 1.0 4.9497e-02 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 1.8e+01 0 0 0 0 5 0 0 0 0 7 0 Warning -- total time of even greater than time of entire stage -- something is wrong with the timer KSPSolve 1 1.0 5.7077e+02 1.0 7.47e+10 1.0 3.9e+04 2.6e+04 3.2e+02100100 99100 93 201101327131134 4187 VecTDot 9 1.0 1.3759e-01 2.2 2.51e+06 1.1 0.0e+00 0.0e+00 9.0e+00 0 0 0 0 3 0 0 0 0 4 560 VecNorm 6 1.0 3.9218e-0117.3 1.67e+06 1.1 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 2 0 0 0 0 2 131 VecCopy 2 1.0 6.6240e-03 9.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 12 1.0 1.3879e-0242.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 9 1.0 3.7683e-0214.0 2.51e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2043 VecAYPX 3 1.0 7.1573e-03 7.7 8.37e+05 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 3585 VecScatterBegin 7 1.0 5.3980e-03 9.0 0.00e+00 0.0 8.7e+02 1.7e+04 0.0e+00 0 0 2 1 0 0 0 7 2 0 0 VecScatterEnd 7 1.0 3.0406e-0248.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 5 1.0 1.0779e-01 4.6 8.98e+06 1.1 6.4e+02 2.3e+04 0.0e+00 0 0 2 1 0 0 0 5 2 0 2564 MatMultTranspose 2 1.0 1.7044e-02 6.5 1.04e+06 1.0 2.3e+02 2.1e+03 0.0e+00 0 0 1 0 0 0 0 2 0 0 1896 MatLUFactorSym 1 1.0 3.1107e+00 2.5 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 1 1 0 0 0 1 0 MatLUFactorNum 1 1.0 5.5701e+02 7.3 7.38e+10 1.0 0.0e+00 0.0e+00 0.0e+00 49 99 0 0 0 98100 0 0 0 4241 MatAssemblyBegin 11 1.0 5.8176e-01 4.1 0.00e+00 0.0 0.0e+00 0.0e+00 1.2e+01 0 0 0 0 3 0 0 0 0 5 0 MatAssemblyEnd 11 1.0 2.9766e-01 2.0 0.00e+00 0.0 2.2e+03 1.0e+03 4.0e+01 0 0 6 0 11 0 0 18 0 16 0 MatGetRowIJ 1 1.0 6.0652e-02 7.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 3.3588e-01 3.4 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 1 0 0 0 0 1 0 MatView 8 1.3 4.5514e-03 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 2 0 0 0 0 2 0 MatPtAP 2 1.0 7.9281e-01 1.0 2.06e+07 1.1 3.9e+03 8.8e+03 5.0e+01 0 0 10 3 14 0 0 33 4 21 800 MatPtAPSymbolic 2 1.0 4.4065e-01 1.1 0.00e+00 0.0 2.2e+03 1.2e+04 3.0e+01 0 0 6 2 9 0 0 18 3 12 0 MatPtAPNumeric 2 1.0 3.9084e-01 1.1 2.06e+07 1.1 1.7e+03 5.2e+03 2.0e+01 0 0 4 1 6 0 0 14 1 8 1623 MatGetRedundant 1 1.0 1.8018e+00 1.3 0.00e+00 0.0 3.0e+03 2.3e+05 4.0e+00 0 0 8 68 1 1 0 25 89 2 0 MatGetLocalMat 2 1.0 5.1957e-02 6.7 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 0 0 0 0 1 0 0 0 0 2 0 MatGetBrAoCol 2 1.0 6.7300e-02 2.9 0.00e+00 0.0 1.5e+03 1.4e+04 4.0e+00 0 0 4 2 1 0 0 13 3 2 0 MatGetSymTrans 4 1.0 2.9506e-02 8.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCSetUp 1 1.0 5.6251e+02 7.0 7.38e+10 1.0 1.1e+04 7.2e+04 1.8e+02 50 99 27 74 51 100100 89 98 73 4201 Warning -- total time of even greater than time of entire stage -- something is wrong with the timer PCApply 5 1.0 4.8945e+0265.0 8.32e+08 1.0 2.8e+04 8.8e+03 1.1e+02 50 1 70 24 30 101 1232 32 44 54 MGSetup Level 0 1 1.0 5.6137e+02 7.1 7.38e+10 1.0 5.1e+03 1.4e+05 2.7e+01 49 99 13 71 8 99100 43 93 11 4208 MGSetup Level 1 1 1.0 8.1902e-03 2.0 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 2 0 0 0 0 2 0 MGSetup Level 2 1 1.0 1.7457e-02 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 2 0 0 0 0 2 0 --- Event Stage 1: MG Apply KSPGMRESOrthog 20 1.0 4.4386e-01 1.9 3.47e+07 1.1 0.0e+00 0.0e+00 2.0e+01 0 0 0 0 6 0 4 0 0 19 2390 KSPSetUp 2 1.0 4.8228e+029246.6 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+01 49 0 0 0 6 97 0 0 0 19 0 KSPSolve 25 1.0 4.8849e+0270.1 8.08e+08 1.0 2.3e+04 9.7e+03 1.1e+02 50 1 58 22 30 99 97 83 90100 52 VecMDot 20 1.0 4.1889e-01 3.0 1.74e+07 1.1 0.0e+00 0.0e+00 2.0e+01 0 0 0 0 6 0 2 0 0 19 1266 VecNorm 22 1.0 2.4155e-01 2.4 3.47e+06 1.1 0.0e+00 0.0e+00 2.2e+01 0 0 0 0 6 0 0 0 0 21 439 VecScale 62 1.0 6.6617e-0213.4 4.90e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 2244 VecCopy 12 1.0 1.9117e-0210.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 51 1.0 3.4841e-0229.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 84 1.0 9.3068e-02 9.6 1.33e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 2 0 0 0 4352 VecAYPX 80 1.0 8.7920e-02 8.3 7.90e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 2742 VecMAXPY 22 1.0 1.4221e-01 7.5 2.05e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 2 0 0 0 4408 VecPointwiseMult 82 1.0 1.4117e-01 9.0 6.48e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 1400 VecScatterBegin 112 1.0 1.0815e-01 8.5 0.00e+00 0.0 2.8e+04 8.8e+03 0.0e+00 0 0 70 24 0 0 0100100 0 0 VecScatterEnd 112 1.0 5.3921e+0012.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 VecNormalize 22 1.0 2.4458e-01 1.9 5.21e+06 1.1 0.0e+00 0.0e+00 2.2e+01 0 0 0 0 6 0 1 0 0 21 651 MatMult 82 1.0 4.5575e+00 4.0 1.12e+08 1.1 2.0e+04 7.5e+03 0.0e+00 0 0 52 15 0 1 13 74 62 0 750 MatMultAdd 10 1.0 2.4382e+0030.3 5.21e+06 1.0 1.2e+03 2.1e+03 0.0e+00 0 0 3 0 0 0 1 4 1 0 66 MatMultTranspose 10 1.0 1.6895e-0117.1 5.21e+06 1.0 1.2e+03 2.1e+03 0.0e+00 0 0 3 0 0 0 1 4 1 0 956 MatSolve 5 1.0 5.0062e+00 7.7 6.36e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 1 77 0 0 0 4064 PCApply 87 1.0 5.3308e+00 5.4 6.42e+08 1.0 5.0e+03 1.8e+04 4.0e+00 0 1 13 9 1 1 78 18 36 4 3854 MGSmooth Level 0 5 1.0 5.1727e+00 5.6 6.36e+08 1.0 5.0e+03 1.8e+04 0.0e+00 0 1 13 9 0 1 77 18 36 0 3934 MGSmooth Level 1 10 1.0 4.2581e+00 9.2 4.25e+07 1.1 1.3e+04 2.1e+03 5.3e+01 0 0 34 3 15 1 5 48 11 50 297 MGResid Level 1 5 1.0 1.1527e-01 3.0 4.80e+06 1.1 1.8e+03 2.1e+03 0.0e+00 0 0 5 0 0 0 1 7 2 0 1244 MGInterp Level 1 10 1.0 2.3899e+0026.1 1.20e+06 1.1 1.2e+03 8.7e+02 0.0e+00 0 0 3 0 0 0 0 4 0 0 15 MGSmooth Level 2 10 1.0 4.8321e+02325.1 1.29e+08 1.1 4.6e+03 2.3e+04 5.3e+01 49 0 12 10 15 98 15 17 43 50 8 MGResid Level 2 5 1.0 1.5715e-01 2.6 9.67e+06 1.1 6.4e+02 2.3e+04 0.0e+00 0 0 2 1 0 0 1 2 6 0 1894 MGInterp Level 2 10 1.0 1.4948e-01 3.1 9.22e+06 1.0 1.2e+03 3.3e+03 0.0e+00 0 0 3 0 0 0 1 4 2 0 1919 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 1 1 564 0 Krylov Solver 7 7 66456 0 DMKSP interface 2 2 1296 0 Vector 46 78 36381912 0 Vector Scatter 13 13 13676 0 Matrix 21 21 822871308 0 Distributed Mesh 3 3 1391808 0 Bipartite Graph 6 6 4752 0 Viewer 2 1 728 0 Index Set 32 32 1989612 0 IS L to G Mapping 3 3 690348 0 Preconditioner 7 7 6432 0 --- Event Stage 1: MG Apply Vector 32 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 5.00679e-07 Average time for MPI_Barrier(): 0.000592613 Average time for zero size MPI_Send(): 0.000596561 #PETSc Option Table entries: -da_refine 2 -dm_view -ksp_monitor -ksp_rtol 1.0e-7 -ksp_type cg -ksp_view -log_summary -mg_levels_ksp_type chebyshev -mg_levels_pc_type jacobi -pc_mg_galerkin -pc_mg_log -pc_mg_monitor -pc_type mg #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Thu Jun 13 15:51:55 2013 Configure options: --download-f-blas-lapack --download-hypre --download-mpich --with-cc=gcc --with-debugging=no --with-fc=gfortran PETSC_ARCH=linux-gnu-c-nodebug ----------------------------------------- Libraries compiled on Thu Jun 13 15:51:55 2013 on login1.ittc.ku.edu Machine characteristics: Linux-2.6.32-220.13.1.el6.x86_64-x86_64-with-redhat-6.2-Santiago Using PETSc directory: /bio/work1/zlwei/PETSc/petsc-dev Using PETSc arch: linux-gnu-c-nodebug ----------------------------------------- Using C compiler: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpif90 -fPIC -Wall -Wno-unused-variable -O ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/include -I/bio/work1/zlwei/PETSc/petsc-dev/include -I/bio/work1/zlwei/PETSc/petsc-dev/include -I/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/include ----------------------------------------- Using C linker: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpicc Using Fortran linker: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpif90 Using libraries: -Wl,-rpath,/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -L/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -lpetsc -Wl,-rpath,/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -L/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -lHYPRE -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lmpichcxx -lstdc++ -lflapack -lfblas -lX11 -lpthread -lmpichf90 -lgfortran -lm -lm -lmpichcxx -lstdc++ -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl ----------------------------------------- -------------- next part -------------- Processor [0] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 0 33 Processor [1] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 0 33 Processor [2] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 0 33 Processor [3] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 0 33 Processor [4] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 0 33 Processor [5] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 0 33 Processor [6] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 0 33 Processor [7] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 0 33 Processor [8] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 33 65 Processor [9] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 33 65 Processor [10] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 33 65 Processor [11] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 33 65 Processor [12] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 33 65 Processor [13] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 33 65 Processor [14] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 33 65 Processor [15] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 33 65 Processor [16] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 65 97 Processor [17] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 65 97 Processor [18] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 65 97 Processor [19] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 65 97 Processor [20] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 65 97 Processor [21] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 65 97 Processor [22] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 65 97 Processor [23] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 65 97 Processor [24] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 65, Z range of indices: 97 129 Processor [25] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 0 65, Z range of indices: 97 129 Processor [26] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 0 65, Z range of indices: 97 129 Processor [27] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 0 65, Z range of indices: 97 129 Processor [28] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 65, Y range of indices: 65 129, Z range of indices: 97 129 Processor [29] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 129, Y range of indices: 65 129, Z range of indices: 97 129 Processor [30] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 129 193, Y range of indices: 65 129, Z range of indices: 97 129 Processor [31] M 257 N 129 P 129 m 4 n 2 p 4 w 1 s 1 X range of indices: 193 257, Y range of indices: 65 129, Z range of indices: 97 129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 mx = 257, my = 129, mz =129 Processor [0] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 0 17 Processor [1] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 0 17 Processor [2] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 0 17 Processor [3] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 0 17 Processor [4] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 0 17 Processor [5] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 0 17 Processor [6] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 0 17 Processor [7] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 0 17 Processor [8] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 17 33 Processor [9] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 17 33 Processor [10] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 17 33 Processor [11] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 17 33 Processor [12] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 17 33 Processor [13] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 17 33 Processor [14] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 17 33 Processor [15] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 17 33 Processor [16] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 33 49 Processor [17] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 33 49 Processor [18] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 33 49 Processor [19] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 33 49 Processor [20] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 33 49 Processor [21] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 33 49 Processor [22] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 33 49 Processor [23] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 33 49 Processor [24] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 33, Z range of indices: 49 65 Processor [25] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 0 33, Z range of indices: 49 65 Processor [26] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 0 33, Z range of indices: 49 65 Processor [27] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 0 33, Z range of indices: 49 65 Processor [28] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 33, Y range of indices: 33 65, Z range of indices: 49 65 Processor [29] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 65, Y range of indices: 33 65, Z range of indices: 49 65 Processor [30] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 65 97, Y range of indices: 33 65, Z range of indices: 49 65 Processor [31] M 129 N 65 P 65 m 4 n 2 p 4 w 1 s 1 X range of indices: 97 129, Y range of indices: 33 65, Z range of indices: 49 65 Processor [0] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 0 9 Processor [1] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 0 9 Processor [2] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 0 9 Processor [3] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 0 9 Processor [4] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 0 9 Processor [5] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 0 9 Processor [6] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 0 9 Processor [7] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 0 9 Processor [8] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 9 17 Processor [9] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 9 17 Processor [10] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 9 17 Processor [11] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 9 17 Processor [12] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 9 17 Processor [13] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 9 17 Processor [14] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 9 17 Processor [15] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 9 17 Processor [16] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 17 25 Processor [17] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 17 25 Processor [18] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 17 25 Processor [19] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 17 25 Processor [20] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 17 25 Processor [21] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 17 25 Processor [22] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 17 25 Processor [23] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 17 25 Processor [24] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 17, Z range of indices: 25 33 Processor [25] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 0 17, Z range of indices: 25 33 Processor [26] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 0 17, Z range of indices: 25 33 Processor [27] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 0 17, Z range of indices: 25 33 Processor [28] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 17, Y range of indices: 17 33, Z range of indices: 25 33 Processor [29] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 33, Y range of indices: 17 33, Z range of indices: 25 33 Processor [30] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 33 49, Y range of indices: 17 33, Z range of indices: 25 33 Processor [31] M 65 N 33 P 33 m 4 n 2 p 4 w 1 s 1 X range of indices: 49 65, Y range of indices: 17 33, Z range of indices: 25 33 Processor [0] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 0 9, Z range of indices: 0 5 Processor [1] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 0 9, Z range of indices: 0 5 Processor [2] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 0 9, Z range of indices: 0 5 Processor [3] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 0 9, Z range of indices: 0 5 Processor [4] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 9 17, Z range of indices: 0 5 Processor [5] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 9 17, Z range of indices: 0 5 Processor [6] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 9 17, Z range of indices: 0 5 Processor [7] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 9 17, Z range of indices: 0 5 Processor [8] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 0 9, Z range of indices: 5 9 Processor [9] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 0 9, Z range of indices: 5 9 Processor [10] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 0 9, Z range of indices: 5 9 Processor [11] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 0 9, Z range of indices: 5 9 Processor [12] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 9 17, Z range of indices: 5 9 Processor [13] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 9 17, Z range of indices: 5 9 Processor [14] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 9 17, Z range of indices: 5 9 Processor [15] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 9 17, Z range of indices: 5 9 Processor [16] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 0 9, Z range of indices: 9 13 Processor [17] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 0 9, Z range of indices: 9 13 Processor [18] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 0 9, Z range of indices: 9 13 Processor [19] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 0 9, Z range of indices: 9 13 Processor [20] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 9 17, Z range of indices: 9 13 Processor [21] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 9 17, Z range of indices: 9 13 Processor [22] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 9 17, Z range of indices: 9 13 Processor [23] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 9 17, Z range of indices: 9 13 Processor [24] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 0 9, Z range of indices: 13 17 Processor [25] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 0 9, Z range of indices: 13 17 Processor [26] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 0 9, Z range of indices: 13 17 Processor [27] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 0 9, Z range of indices: 13 17 Processor [28] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 9, Y range of indices: 9 17, Z range of indices: 13 17 Processor [29] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 17, Y range of indices: 9 17, Z range of indices: 13 17 Processor [30] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 17 25, Y range of indices: 9 17, Z range of indices: 13 17 Processor [31] M 33 N 17 P 17 m 4 n 2 p 4 w 1 s 1 X range of indices: 25 33, Y range of indices: 9 17, Z range of indices: 13 17 Processor [0] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 0 5, Z range of indices: 0 3 Processor [1] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 0 5, Z range of indices: 0 3 Processor [2] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 0 5, Z range of indices: 0 3 Processor [3] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 0 5, Z range of indices: 0 3 Processor [4] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 5 9, Z range of indices: 0 3 Processor [5] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 5 9, Z range of indices: 0 3 Processor [6] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 5 9, Z range of indices: 0 3 Processor [7] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 5 9, Z range of indices: 0 3 Processor [8] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 0 5, Z range of indices: 3 5 Processor [9] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 0 5, Z range of indices: 3 5 Processor [10] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 0 5, Z range of indices: 3 5 Processor [11] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 0 5, Z range of indices: 3 5 Processor [12] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 5 9, Z range of indices: 3 5 Processor [13] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 5 9, Z range of indices: 3 5 Processor [14] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 5 9, Z range of indices: 3 5 Processor [15] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 5 9, Z range of indices: 3 5 Processor [16] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 0 5, Z range of indices: 5 7 Processor [17] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 0 5, Z range of indices: 5 7 Processor [18] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 0 5, Z range of indices: 5 7 Processor [19] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 0 5, Z range of indices: 5 7 Processor [20] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 5 9, Z range of indices: 5 7 Processor [21] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 5 9, Z range of indices: 5 7 Processor [22] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 5 9, Z range of indices: 5 7 Processor [23] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 5 9, Z range of indices: 5 7 Processor [24] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 0 5, Z range of indices: 7 9 Processor [25] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 0 5, Z range of indices: 7 9 Processor [26] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 0 5, Z range of indices: 7 9 Processor [27] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 0 5, Z range of indices: 7 9 Processor [28] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 5, Y range of indices: 5 9, Z range of indices: 7 9 Processor [29] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 9, Y range of indices: 5 9, Z range of indices: 7 9 Processor [30] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 9 13, Y range of indices: 5 9, Z range of indices: 7 9 Processor [31] M 17 N 9 P 9 m 4 n 2 p 4 w 1 s 1 X range of indices: 13 17, Y range of indices: 5 9, Z range of indices: 7 9 Processor [0] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 0 3, Z range of indices: 0 2 Processor [1] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 0 3, Z range of indices: 0 2 Processor [2] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 0 3, Z range of indices: 0 2 Processor [3] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 0 3, Z range of indices: 0 2 Processor [4] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 3 5, Z range of indices: 0 2 Processor [5] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 3 5, Z range of indices: 0 2 Processor [6] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 3 5, Z range of indices: 0 2 Processor [7] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 3 5, Z range of indices: 0 2 Processor [8] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 0 3, Z range of indices: 2 3 Processor [9] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 0 3, Z range of indices: 2 3 Processor [10] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 0 3, Z range of indices: 2 3 Processor [11] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 0 3, Z range of indices: 2 3 Processor [12] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 3 5, Z range of indices: 2 3 Processor [13] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 3 5, Z range of indices: 2 3 Processor [14] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 3 5, Z range of indices: 2 3 Processor [15] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 3 5, Z range of indices: 2 3 Processor [16] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 0 3, Z range of indices: 3 4 Processor [17] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 0 3, Z range of indices: 3 4 Processor [18] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 0 3, Z range of indices: 3 4 Processor [19] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 0 3, Z range of indices: 3 4 Processor [20] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 3 5, Z range of indices: 3 4 Processor [21] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 3 5, Z range of indices: 3 4 Processor [22] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 3 5, Z range of indices: 3 4 Processor [23] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 3 5, Z range of indices: 3 4 Processor [24] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 0 3, Z range of indices: 4 5 Processor [25] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 0 3, Z range of indices: 4 5 Processor [26] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 0 3, Z range of indices: 4 5 Processor [27] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 0 3, Z range of indices: 4 5 Processor [28] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 0 3, Y range of indices: 3 5, Z range of indices: 4 5 Processor [29] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 3 5, Y range of indices: 3 5, Z range of indices: 4 5 Processor [30] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 5 7, Y range of indices: 3 5, Z range of indices: 4 5 Processor [31] M 9 N 5 P 5 m 4 n 2 p 4 w 1 s 1 X range of indices: 7 9, Y range of indices: 3 5, Z range of indices: 4 5 0 KSP Residual norm 1.990474015208e+03 1 KSP Residual norm 1.163078153200e+02 2 KSP Residual norm 2.809444096980e+00 3 KSP Residual norm 2.139770554363e-01 4 KSP Residual norm 4.835908670273e-02 KSP Object: 32 MPI processes type: cg maximum iterations=10000 tolerances: relative=1e-07, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using PRECONDITIONED norm type for convergence test PC Object: 32 MPI processes type: mg MG: type is MULTIPLICATIVE, levels=6 cycles=v Cycles per PCApply=1 Using Galerkin computed coarse grid matrices Coarse grid solver -- level ------------------------------- KSP Object: (mg_coarse_) 32 MPI processes type: preonly maximum iterations=1, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_) 32 MPI processes type: redundant Redundant preconditioner: First (color=0) of 32 PCs follows KSP Object: (mg_coarse_redundant_) 1 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_redundant_) 1 MPI processes type: lu LU: out-of-place factorization tolerance for zero pivot 2.22045e-14 using diagonal shift on blocks to prevent zero pivot matrix ordering: nd factor fill ratio given 5, needed 3.15101 Factored matrix follows: Matrix Object: 1 MPI processes type: seqaij rows=225, cols=225 package used to perform factorization: petsc total: nonzeros=13313, allocated nonzeros=13313 total number of mallocs used during MatSetValues calls =0 using I-node routines: found 174 nodes, limit used is 5 linear system matrix = precond matrix: Matrix Object: 1 MPI processes type: seqaij rows=225, cols=225 total: nonzeros=4225, allocated nonzeros=6075 total number of mallocs used during MatSetValues calls =0 not using I-node routines linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=225, cols=225 total: nonzeros=4225, allocated nonzeros=4225 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Down solver (pre-smoother) on level 1 ------------------------------- KSP Object: (mg_levels_1_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.283115, max = 3.11426 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_1_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_1_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=1377, cols=1377 total: nonzeros=30625, allocated nonzeros=30625 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_1_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=1377, cols=1377 total: nonzeros=30625, allocated nonzeros=30625 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 2 ------------------------------- KSP Object: (mg_levels_2_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.285627, max = 3.1419 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_2_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_2_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=9537, cols=9537 total: nonzeros=232897, allocated nonzeros=232897 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_2_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=9537, cols=9537 total: nonzeros=232897, allocated nonzeros=232897 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 3 ------------------------------- KSP Object: (mg_levels_3_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.275571, max = 3.03128 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_3_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_3_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=70785, cols=70785 total: nonzeros=1815937, allocated nonzeros=1815937 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_3_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=70785, cols=70785 total: nonzeros=1815937, allocated nonzeros=1815937 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 4 ------------------------------- KSP Object: (mg_levels_4_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.231542, max = 2.54696 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_4_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_4_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=545025, cols=545025 total: nonzeros=14340865, allocated nonzeros=14340865 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_4_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=545025, cols=545025 total: nonzeros=14340865, allocated nonzeros=14340865 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 5 ------------------------------- KSP Object: (mg_levels_5_) 32 MPI processes type: chebyshev Chebyshev: eigenvalue estimates: min = 0.155706, max = 1.71277 Chebyshev: estimated using: [0 0.1; 0 1.1] KSP Object: (mg_levels_5_est_) 32 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_5_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (mg_levels_5_) 32 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 Up solver (post-smoother) same as down solver (pre-smoother) linear system matrix = precond matrix: Matrix Object: 32 MPI processes type: mpiaij rows=4276737, cols=4276737 total: nonzeros=29771265, allocated nonzeros=29771265 total number of mallocs used during MatSetValues calls =0 Residual norm 0.000947651 Total Time Elapsed: 5.583765 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex45 on a linux-gnu-c-nodebug named n042 with 32 processors, by zlwei Wed Nov 6 11:35:09 2013 Total Time Elapsed: 5.584491 Using Petsc Development GIT revision: d696997672013bb4513d3ff57c61cc10e09b71f6 GIT Date: 2013-06-13 10:28:37 -0500 Max Max/Min Avg Total Total Time Elapsed: 5.245622 Total Time Elapsed: 5.245764 Total Time Elapsed: 5.585519 Total Time Elapsed: 5.245926 Total Time Elapsed: 5.585674 Total Time Elapsed: 4.979911 Total Time Elapsed: 4.983441 Total Time Elapsed: 4.980154 Total Time Elapsed: 4.983256 Total Time Elapsed: 4.979283 Total Time Elapsed: 4.983043 Total Time Elapsed: 4.982123 Total Time Elapsed: 5.247401 Total Time Elapsed: 5.247372 Total Time Elapsed: 5.247512 Total Time Elapsed: 5.248704 Total Time Elapsed: 5.588453 Total Time Elapsed: 5.248888 Total Time Elapsed: 4.987055 Total Time Elapsed: 5.589496 Total Time Elapsed: 5.113992 Total Time Elapsed: 5.117246 Total Time Elapsed: 5.591055 Total Time Elapsed: 5.591275 Total Time Elapsed: 5.116060 Total Time Elapsed: 5.119928 Total Time Elapsed: 5.114534 Total Time Elapsed: 5.118732 Total Time Elapsed: 5.123729 Total Time Elapsed: 5.125026 Time (sec): 4.978e+00 1.00287 4.969e+00 Objects: 3.570e+02 1.00000 3.570e+02 Flops: 2.438e+08 1.08372 2.320e+08 7.424e+09 Flops/sec: 4.911e+07 1.08434 4.669e+07 1.494e+09 MPI Messages: 4.234e+03 2.12444 3.045e+03 9.745e+04 MPI Message Lengths: 9.073e+06 1.78864 2.350e+03 2.291e+08 MPI Reductions: 7.370e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.5054e+00 50.4% 1.2102e+09 16.3% 2.114e+04 21.7% 6.174e+02 26.3% 4.710e+02 63.9% 1: MG Apply: 2.4639e+00 49.6% 6.2142e+09 83.7% 7.631e+04 78.3% 1.733e+03 73.7% 2.650e+02 36.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage KSPSetUp 8 1.0 6.8008e-02 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 3.6e+01 1 0 0 0 5 2 0 0 0 8 0 Warning -- total time of even greater than time of entire stage -- something is wrong with the timer KSPSolve 1 1.0 4.7911e+00 1.0 2.41e+08 1.1 9.7e+04 2.3e+03 7.1e+02 96 99 99 98 96 191607458373150 1535 VecTDot 9 1.0 1.3888e-01 4.1 2.51e+06 1.1 0.0e+00 0.0e+00 9.0e+00 2 1 0 0 1 4 6 0 0 2 554 VecNorm 6 1.0 3.6860e-0110.9 1.67e+06 1.1 0.0e+00 0.0e+00 6.0e+00 4 1 0 0 1 8 4 0 0 1 139 VecCopy 2 1.0 3.1111e-03 4.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 27 1.0 1.3712e-0241.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 9 1.0 3.5091e-0213.5 2.51e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 6 0 0 0 2194 VecAYPX 3 1.0 2.2887e-0224.0 8.37e+05 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 2 0 0 0 1121 VecScatterBegin 10 1.0 3.3240e-03 5.2 0.00e+00 0.0 1.2e+03 1.2e+04 0.0e+00 0 0 1 7 0 0 0 6 25 0 0 VecScatterEnd 10 1.0 3.6685e-0251.7 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 1 0 0 0 0 0 MatMult 5 1.0 9.9055e-02 5.5 8.98e+06 1.1 6.4e+02 2.3e+04 0.0e+00 1 4 1 6 0 2 23 3 24 0 2790 MatMultTranspose 5 1.0 2.3519e-02 8.9 1.06e+06 1.0 5.8e+02 9.0e+02 0.0e+00 0 0 1 0 0 0 3 3 1 0 1397 MatLUFactorSym 1 1.0 3.5100e-0310.2 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 1 0 MatLUFactorNum 1 1.0 3.3560e-03 6.0 4.79e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 4564 MatAssemblyBegin 23 1.0 5.5205e-01 4.4 0.00e+00 0.0 0.0e+00 0.0e+00 2.4e+01 6 0 0 0 3 13 0 0 0 5 0 MatAssemblyEnd 23 1.0 2.5726e-01 1.6 0.00e+00 0.0 5.1e+03 4.5e+02 8.8e+01 4 0 5 1 12 8 0 24 4 19 0 MatGetRowIJ 1 1.0 1.5712e-04 6.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 5.7721e-04 4.2 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 14 1.2 7.8599e-03 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 1.2e+01 0 0 0 0 2 0 0 0 0 3 0 MatPtAP 5 1.0 8.9238e-01 1.0 2.14e+07 1.1 1.1e+04 3.4e+03 1.2e+02 18 9 11 16 17 36 54 51 61 27 734 MatPtAPSymbolic 5 1.0 5.3509e-01 1.1 0.00e+00 0.0 6.5e+03 4.1e+03 7.5e+01 10 0 7 12 10 21 0 31 45 16 0 MatPtAPNumeric 5 1.0 3.8300e-01 1.1 2.14e+07 1.1 4.3e+03 2.2e+03 5.0e+01 7 9 4 4 7 15 54 20 16 11 1710 MatGetRedundant 1 1.0 2.0813e-02 3.6 0.00e+00 0.0 3.0e+03 5.5e+02 4.0e+00 0 0 3 1 1 0 0 14 3 1 0 MatGetLocalMat 5 1.0 7.4365e-02 8.8 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 1 0 0 0 1 1 0 0 0 2 0 MatGetBrAoCol 5 1.0 9.1380e-02 2.3 0.00e+00 0.0 4.8e+03 4.6e+03 1.0e+01 1 0 5 10 1 3 0 23 37 2 0 MatGetSymTrans 10 1.0 2.0689e-02 5.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCSetUp 1 1.0 1.4868e+00 1.0 2.29e+07 1.1 2.0e+04 2.1e+03 4.0e+02 30 9 20 18 54 59 58 94 70 85 473 PCApply 5 1.0 2.6097e+00 1.1 2.04e+08 1.1 7.6e+04 2.2e+03 2.6e+02 50 84 78 74 36 98513361281 56 2381 MGSetup Level 0 1 1.0 5.2823e-02 1.4 4.79e+05 1.0 5.1e+03 3.4e+02 2.7e+01 1 0 5 1 4 2 1 24 3 6 290 MGSetup Level 1 1 1.0 9.4671e-03 2.1 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 0 0 0 0 1 0 MGSetup Level 2 1 1.0 4.6070e-03 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 0 0 0 0 1 0 MGSetup Level 3 1 1.0 4.6048e-03 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 0 0 0 0 1 0 MGSetup Level 4 1 1.0 6.0458e-03 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 0 0 0 0 1 0 MGSetup Level 5 1 1.0 2.0175e-02 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 1 0 0 0 1 0 --- Event Stage 1: MG Apply KSPGMRESOrthog 50 1.0 5.0590e-01 1.9 3.54e+07 1.1 0.0e+00 0.0e+00 5.0e+01 8 15 0 0 7 16 17 0 0 19 2132 KSPSetUp 5 1.0 1.7280e-01 3.1 0.00e+00 0.0 0.0e+00 0.0e+00 5.0e+01 2 0 0 0 7 4 0 0 0 19 0 KSPSolve 55 1.0 2.2699e+00 1.2 1.79e+08 1.1 6.3e+04 2.3e+03 2.6e+02 43 73 64 63 36 86 87 82 85100 2389 VecMDot 50 1.0 4.8221e-01 2.4 1.77e+07 1.1 0.0e+00 0.0e+00 5.0e+01 7 7 0 0 7 14 9 0 0 19 1119 VecNorm 55 1.0 2.5947e-01 2.1 3.54e+06 1.1 0.0e+00 0.0e+00 5.5e+01 4 1 0 0 7 8 2 0 0 21 416 VecScale 155 1.0 5.0950e-02 9.8 4.99e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 2 0 0 0 1 2 0 0 0 2983 VecCopy 30 1.0 1.3113e-02 7.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 105 1.0 1.9075e-0219.7 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 210 1.0 8.1555e-02 8.1 1.35e+07 1.1 0.0e+00 0.0e+00 0.0e+00 1 6 0 0 0 2 7 0 0 0 5050 VecAYPX 200 1.0 8.0734e-02 7.4 8.05e+06 1.1 0.0e+00 0.0e+00 0.0e+00 1 3 0 0 0 2 4 0 0 0 3037 VecMAXPY 55 1.0 1.2952e-01 9.1 2.09e+07 1.1 0.0e+00 0.0e+00 0.0e+00 1 9 0 0 0 2 10 0 0 0 4921 VecPointwiseMult 205 1.0 1.2314e-01 8.1 6.60e+06 1.1 0.0e+00 0.0e+00 0.0e+00 1 3 0 0 0 3 3 0 0 0 1633 VecScatterBegin 265 1.0 1.2376e-01 6.9 0.00e+00 0.0 7.6e+04 2.2e+03 0.0e+00 1 0 78 74 0 2 0100100 0 0 VecScatterEnd 265 1.0 1.2838e+00 3.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 16 0 0 0 0 33 0 0 0 0 0 VecNormalize 55 1.0 2.6221e-01 1.8 5.31e+06 1.1 0.0e+00 0.0e+00 5.5e+01 4 2 0 0 7 8 3 0 0 21 617 MatMult 205 1.0 1.4328e+00 1.4 1.18e+08 1.1 6.6e+04 2.5e+03 0.0e+00 25 48 67 71 0 51 58 86 97 0 2503 MatMultAdd 25 1.0 1.4013e-01 2.4 5.31e+06 1.0 2.9e+03 9.0e+02 0.0e+00 2 2 3 1 0 3 3 4 2 0 1172 MatMultTranspose 25 1.0 1.2134e-0111.7 5.31e+06 1.0 2.9e+03 9.0e+02 0.0e+00 1 2 3 1 0 3 3 4 2 0 1354 MatSolve 5 1.0 6.1536e-04 3.8 1.32e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 6865 PCApply 210 1.0 1.7584e-01 2.9 6.73e+06 1.1 5.0e+03 5.6e+01 1.0e+01 2 3 5 0 1 5 3 6 0 4 1167 MGSmooth Level 0 5 1.0 1.6812e-02 3.7 1.32e+05 1.0 5.0e+03 5.6e+01 0.0e+00 0 0 5 0 0 0 0 6 0 0 251 MGSmooth Level 1 10 1.0 7.2483e-02 1.1 1.45e+05 2.5 1.3e+04 5.3e+01 5.3e+01 1 0 14 0 7 3 0 17 0 20 39 MGResid Level 1 5 1.0 8.1933e-03 2.0 1.57e+04 2.6 1.8e+03 5.3e+01 0.0e+00 0 0 2 0 0 0 0 2 0 0 37 MGInterp Level 1 10 1.0 1.0109e-0210.1 3.92e+03 1.8 1.2e+03 2.4e+01 0.0e+00 0 0 1 0 0 0 0 2 0 0 8 MGSmooth Level 2 10 1.0 7.6197e-02 1.1 8.54e+05 1.6 1.3e+04 1.6e+02 5.3e+01 1 0 14 1 7 3 0 17 1 20 273 MGResid Level 2 5 1.0 6.6566e-03 2.0 9.46e+04 1.6 1.8e+03 1.6e+02 0.0e+00 0 0 2 0 0 0 0 2 0 0 350 MGInterp Level 2 10 1.0 7.4329e-03 4.0 2.37e+04 1.4 1.2e+03 7.1e+01 0.0e+00 0 0 1 0 0 0 0 2 0 0 82 MGSmooth Level 3 10 1.0 1.4607e-01 1.4 5.79e+06 1.3 1.3e+04 5.7e+02 5.3e+01 3 2 14 3 7 5 3 17 4 20 1102 MGResid Level 3 5 1.0 1.9478e-02 3.1 6.50e+05 1.3 1.8e+03 5.7e+02 0.0e+00 0 0 2 0 0 0 0 2 1 0 932 MGInterp Level 3 10 1.0 1.0718e-02 3.6 1.62e+05 1.2 1.2e+03 2.4e+02 0.0e+00 0 0 1 0 0 0 0 2 0 0 435 MGSmooth Level 4 10 1.0 6.5122e-01 1.4 4.25e+07 1.1 1.3e+04 2.1e+03 5.3e+01 12 17 14 12 7 23 20 17 17 20 1943 MGResid Level 4 5 1.0 1.2598e-01 5.1 4.80e+06 1.1 1.8e+03 2.1e+03 0.0e+00 1 2 2 2 0 2 2 2 2 0 1138 MGInterp Level 4 10 1.0 4.8600e-02 5.7 1.20e+06 1.1 1.2e+03 8.7e+02 0.0e+00 0 0 1 0 0 1 1 2 1 0 747 MGSmooth Level 5 10 1.0 1.5219e+00 1.5 1.29e+08 1.1 4.6e+03 2.3e+04 5.3e+01 26 53 5 46 7 52 64 6 62 20 2608 MGResid Level 5 5 1.0 1.6775e-01 2.4 9.67e+06 1.1 6.4e+02 2.3e+04 0.0e+00 2 4 1 6 0 4 5 1 9 0 1775 MGInterp Level 5 10 1.0 1.9435e-01 3.1 9.22e+06 1.0 1.2e+03 3.3e+03 0.0e+00 2 4 1 2 0 4 5 2 2 0 1476 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 1 1 564 0 Krylov Solver 13 13 160752 0 DMKSP interface 4 4 2592 0 Vector 91 171 35911200 0 Vector Scatter 25 25 26300 0 Matrix 45 45 36052932 0 Distributed Mesh 6 6 1412736 0 Bipartite Graph 12 12 9504 0 Viewer 2 1 728 0 Index Set 59 59 788488 0 IS L to G Mapping 6 6 695256 0 Preconditioner 13 13 11736 0 --- Event Stage 1: MG Apply Vector 80 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 5.00679e-07 Average time for MPI_Barrier(): 0.000635576 Average time for zero size MPI_Send(): 0.000971369 #PETSc Option Table entries: -da_refine 5 -dm_view -ksp_monitor -ksp_rtol 1.0e-7 -ksp_type cg -ksp_view -log_summary -mg_levels_ksp_type chebyshev -mg_levels_pc_type jacobi -pc_mg_galerkin -pc_mg_log -pc_mg_monitor -pc_type mg #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Thu Jun 13 15:51:55 2013 Configure options: --download-f-blas-lapack --download-hypre --download-mpich --with-cc=gcc --with-debugging=no --with-fc=gfortran PETSC_ARCH=linux-gnu-c-nodebug ----------------------------------------- Libraries compiled on Thu Jun 13 15:51:55 2013 on login1.ittc.ku.edu Machine characteristics: Linux-2.6.32-220.13.1.el6.x86_64-x86_64-with-redhat-6.2-Santiago Using PETSc directory: /bio/work1/zlwei/PETSc/petsc-dev Using PETSc arch: linux-gnu-c-nodebug ----------------------------------------- Using C compiler: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpif90 -fPIC -Wall -Wno-unused-variable -O ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/include -I/bio/work1/zlwei/PETSc/petsc-dev/include -I/bio/work1/zlwei/PETSc/petsc-dev/include -I/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/include ----------------------------------------- Using C linker: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpicc Using Fortran linker: /bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/bin/mpif90 Using libraries: -Wl,-rpath,/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -L/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -lpetsc -Wl,-rpath,/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -L/bio/work1/zlwei/PETSc/petsc-dev/linux-gnu-c-nodebug/lib -lHYPRE -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lmpichcxx -lstdc++ -lflapack -lfblas -lX11 -lpthread -lmpichf90 -lgfortran -lm -lm -lmpichcxx -lstdc++ -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl ----------------------------------------- From dave.mayhem23 at gmail.com Wed Nov 6 12:40:16 2013 From: dave.mayhem23 at gmail.com (Dave May) Date: Wed, 6 Nov 2013 19:40:16 +0100 Subject: [petsc-users] Memory and Speed Issue of using MG as preconditioner In-Reply-To: <527A8BE0.9020101@gmail.com> References: <5279B59D.9030804@gmail.com> <527A8BE0.9020101@gmail.com> Message-ID: To speed up the smoother you could: 1) do less smoothing, but this might incur needing to do more KSP iterations 2) run on more cores 3) if you are happy with cheby/jacobi, you could write a matrix free implementation for your discretisation which gets used on the fine level. Depending on your discretisation, this might improve the speed and scalability. As I said, the prefix for the solver is mentioned in the result of -ksp_view e.g. PC Object: (*mg_coarse_*) 32 MPI processes type: redundant The string in brackets (bolded) is the solver prefix. So if you want to change the coarse grid solver, do this -mg_coarse_ksp_type XXXX -mg_coarse_pc_type YYYY Cheers, Dave On 6 November 2013 19:35, Alan Z. Wei wrote: > Thanks Dave, > I further simulated the problem with -pc_mg_log and output these files > in the attachments. > I found that the smoothing process of the last level always consumes > the most time, i.e. 'MGSmooth Level 5' in out-level5 and "MGSmooth Level 2' > in out-level2. However, as I tested several other -mg_levels_pc_type such > as 'bjacobi', 'asm' etc. The default one, which is 'jacobi', actually works > the best. Therefore, I decide to keep using it. However, do you have any > suggestions to speed up this smoothing process other than changing > -mg_levels_pc_type? > Also, as you suggested to change -mg_levels_ksp_type, it does not > influence much if replacing 'chebyshev' by 'cg'. However, this part never > change while I modify '-mg_levels_ksp_type': > PC Object: (mg_coarse_) 32 MPI processes > type: redundant > Redundant preconditioner: First (color=0) of 32 PCs follows > KSP Object: (mg_coarse_redundant_) 1 MPI processes > type: preonly > maximum iterations=10000, initial guess is zero > tolerances: relative=1e-05, absolute=1e-50, divergence=10000 > left preconditioning > using NONE norm type for convergence test > PC Object: (mg_coarse_redundant_) 1 MPI processes > type: lu > LU: out-of-place factorization > As you mentioned that the redundant LU for the coarse grid solver > primarily cause the large memory request for the 2-level case. How could I > change the coarse grid solver to reduce the memory requirement or speed up > the solver. > > thanks again, > Alan > > Hey Alan, > > 1/ One difference in the memory footprint is likely coming from your > coarse grid solver which is redundant LU. > The 2 level case has a coarse grid problem with 70785 unknowns whilst the > 5 level case has a coarse grid problem with 225 unknowns. > > 2/ The solve time difference will be affected by your coarse grid size. > Add the command line argument > -pc_mg_log > to profile the setup time spent on the coarse grid and all other levels. > See > http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/PC/PCMG.html > > 3/ You can change the smoother on all levels by using the command line > argument with the appropriate prefix, eg > -mg_levels_ksp_type cg > Note the prefix is displayed in the result of -ksp_view > > Also, your mesh size can be altered at run time using arguments like > -da_grid_x 5 > You shouldn't have to modify the source code each time > See > > http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/DM/DMDACreate3d.html > > > Cheers, > Dave > > > On 6 November 2013 04:21, Alan wrote: > >> Dear all, >> I hope you're having a nice day. >> Recently, I came across a problem on using MG as preconditioner. >> Basically, to achieve the same finest mesh with pc_type = mg, the memory >> usage for -da_refine 2 is much more than that for -da_refine 5. To my >> limited knowledge, more refinement should consume more memory, which is >> contradict to the behavior of pc_type = mg in PETsc. >> Here, I provide two output files. They are all from >> /src/ksp/ksp/example/tutorial/ex45.c with 32 processes. >> The execute file for out-level2 is >> mpiexec -np 32 ./ex45 -pc_type mg -ksp_type cg -da_refine 2 >> -pc_mg_galerkin -ksp_rtol 1.0e-7 -mg_levels_pc_type jacobi >> -mg_levels_ksp_type chebyshev -dm_view -log_summary -pc_mg_log >> -pc_mg_monitor -ksp_view -ksp_monitor > out & >> and in ex45.c, KSPCreate is changed as: >> ierr = >> >> DMDACreate3d(PETSC_COMM_WORLD,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_STENCIL_STAR,-65,-33,-33,PETSC_DECIDE,PETSC_DECIDE,PETSC_DECIDE,1,1,0,0,0,&da);CHKERRQ(ierr); >> On the other hand, the execute file for out-level5 is >> mpiexec -np 32 ./ex45 -pc_type mg -ksp_type cg -da_refine 5 >> -pc_mg_galerkin -ksp_rtol 1.0e-7 -mg_levels_pc_type jacobi >> -mg_levels_ksp_type chebyshev -dm_view -log_summary -pc_mg_log >> -pc_mg_monitor -ksp_view -ksp_monitor > out & >> and in ex45.c, KSPCreate is changed as: >> ierr = >> >> DMDACreate3d(PETSC_COMM_WORLD,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_BOUNDARY_NONE,DMDA_STENCIL_STAR,-9,-5,-5,PETSC_DECIDE,PETSC_DECIDE,PETSC_DECIDE,1,1,0,0,0,&da);CHKERRQ(ierr); >> In summary, the final finest meshes obtained for both cases are >> 257*129*129 as documented in both files. However, the out-level2 shows >> that the Matrix requested 822871308 memory while out-level5 only need >> 36052932. >> Furthermore, although the total iterations for KSP solver are shown as 5 >> times in both files. the wall time elapsed for out-level2 is around >> 150s, while out-level5 is about 4.7s. >> At last, there is a minor question. In both files, under 'Down solver >> (pre-smoother) on level 1' and 'Down solver (pre-smoother) on level 2', >> the type of "KSP Object: (mg_levels_1_est_)" and "KSP Object: >> (mg_levels_2_est_)" are all 'gmres'. Since I'm using uniformly Cartesian >> mesh, would it be helpful to speed up the solver if the 'gmres' is >> replaced by 'cg' here? If so, which PETSc option can change the type of >> KSP object. >> >> sincerely appreciate, >> Alan >> >> >> >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Wed Nov 6 12:47:55 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 06 Nov 2013 11:47:55 -0700 Subject: [petsc-users] Memory and Speed Issue of using MG as preconditioner In-Reply-To: <527A8BE0.9020101@gmail.com> References: <5279B59D.9030804@gmail.com> <527A8BE0.9020101@gmail.com> Message-ID: <87y551v02s.fsf@mcs.anl.gov> "Alan Z. Wei" writes: > As you mentioned that the redundant LU for the coarse grid solver > primarily cause the large memory request for the 2-level case. How could > I change the coarse grid solver to reduce the memory requirement or > speed up the solver. You should use more levels or coarsen faster. You cannot have a scalable method with a huge coarse level. You could try switch to algebraic multigrid on the coarse levels if geometric is not converging well. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From lu_qin_2000 at yahoo.com Wed Nov 6 16:02:42 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Wed, 6 Nov 2013 14:02:42 -0800 (PST) Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 Message-ID: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> Hello, ? I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I?want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: ? --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work ? The configure.log is attached. ? Many thanks for any suggestions, ? Qin? -------------- next part -------------- A non-text attachment was scrubbed... Name: configure.log Type: application/octet-stream Size: 3383137 bytes Desc: not available URL: From balay at mcs.anl.gov Wed Nov 6 16:11:32 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Wed, 6 Nov 2013 16:11:32 -0600 (CST) Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 In-Reply-To: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> References: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> Message-ID: 1. use latest hypre version - i.e https://computation.llnl.gov/casc/hypre/download/hypre-2.9.1a.tar.gz 2. Make sure its buit with -MT compiler option 3. specify hypre with --with-hypre-include and --with-hypre-lib options 4. send us the following [so we know whats there - and why configure is not finding it] ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release Satish On Wed, 6 Nov 2013, Qin Lu wrote: > Hello, > ? > I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I?want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: > ? > --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work > ? > The configure.log is attached. > ? > Many thanks for any suggestions, > ? > Qin? From lu_qin_2000 at yahoo.com Wed Nov 6 18:10:45 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Wed, 6 Nov 2013 16:10:45 -0800 (PST) Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 In-Reply-To: References: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> Message-ID: <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> Satis, 1. hypre-2.9 did not build in Win-7 for some reason, so I still use hypre-2.8.0b. 2. I?used CXXFLAGS="-O2 -MT" for hypre configure. 3. Tried --with-hypre-include and --with-hypre-lib but still got error: ?===============================================================================???????????????????????????????????????????????????????????????????????????????????????????????? TESTING: check from config.libraries(config/BuildSystem/config/libraries.py:145)????????????????????????????????????????????????????????????????????????????????????????????????******************************************************************************* UNABLE to CONFIGURE with GIVEN OPTIONS????(see configure.log for details): ------------------------------------------------------------------------------- --with-hypre-lib=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib'] and --with-hypre-include=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/include'] did not work ******************************************************************************* 4. Content of hypre installation: ? $ ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release /cygdrive/c/Lib/hypre-2.8.0b-win64-release: include??lib /cygdrive/c/Lib/hypre-2.8.0b-win64-release/include: Data.h?????????????????????????????????????? HYPRE_krylov.h????????????????HYPRE_struct_mv.h?? _hypre_IJ_mv.h?????? csr_block_matrix.h????par_csr_block_matrix.h HYPRE.h??????????????????????????????????????HYPRE_lobpcg.h????????????????HYPREf.h????????????_hypre_parcsr_ls.h?? csr_matmultivec.h???? par_csr_matmultivec.h HYPRE_DistributedMatrixPilutSolver_protos.h??HYPRE_matrix_matrix_protos.h??LLNL_FEI_Fei.h??????_hypre_parcsr_mv.h?? distributed_matrix.h??par_csr_pmvcomm.h HYPRE_DistributedMatrixPilutSolver_types.h?? HYPRE_parcsr_ls.h???????????? LLNL_FEI_Impl.h???? _hypre_sstruct_ls.h??fei_defs.h????????????par_multivector.h HYPRE_FEI_includes.h???????????????????????? HYPRE_parcsr_mv.h???????????? LLNL_FEI_LSCore.h?? _hypre_sstruct_mv.h??hypre_cfei.h??????????seq_multivector.h HYPRE_IJ_mv.h????????????????????????????????HYPRE_seq_mv.h????????????????LLNL_FEI_Matrix.h?? _hypre_struct_ls.h?? interpreter.h???????? seq_mv.h HYPRE_LinSysCore.h?????????????????????????? HYPRE_sstruct_ls.h????????????LLNL_FEI_Solver.h?? _hypre_struct_mv.h?? krylov.h??????????????temp_multivector.h HYPRE_MatvecFunctions.h??????????????????????HYPRE_sstruct_mv.h????????????LinearSystemCore.h??cfei-hypre.h???????? lobpcg.h HYPRE_config.h?????????????????????????????? HYPRE_struct_ls.h???????????? Lookup.h????????????cfei_hypre.h???????? multivector.h /cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib: libHYPRE.a 5. I used compilation option -O3 for PETSc, but -O2 for hypre, can this be a problem? ? Thanks, Qin On Wednesday, November 6, 2013 4:11 PM, Satish Balay wrote: 1. use latest hypre version - i.e https://computation.llnl.gov/casc/hypre/download/hypre-2.9.1a.tar.gz 2. Make sure its buit with -MT compiler option 3. specify hypre with --with-hypre-include and --with-hypre-lib options 4. send us the following [so we know whats there - and why configure is not finding it] ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release Satish On Wed, 6 Nov 2013, Qin Lu wrote: > Hello, > ? > I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I?want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: > ? > --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work > ? > The configure.log is attached. > ? > Many thanks for any suggestions, > ? > Qin???? From bsmith at mcs.anl.gov Wed Nov 6 18:33:30 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Wed, 6 Nov 2013 18:33:30 -0600 Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 In-Reply-To: <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> References: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> Message-ID: <09FDD89A-2138-4AE5-A117-D5894824667E@mcs.anl.gov> You need to send configure.log so we can see WHY configure thought it could not use hypre. On Nov 6, 2013, at 6:10 PM, Qin Lu wrote: > Satis, > > 1. hypre-2.9 did not build in Win-7 for some reason, so I still use hypre-2.8.0b. > 2. I used CXXFLAGS="-O2 -MT" for hypre configure. > 3. Tried --with-hypre-include and --with-hypre-lib but still got error: > > =============================================================================== TESTING: check from config.libraries(config/BuildSystem/config/libraries.py:145) ******************************************************************************* > UNABLE to CONFIGURE with GIVEN OPTIONS (see configure.log for details): > ------------------------------------------------------------------------------- > --with-hypre-lib=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib'] and > --with-hypre-include=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/include'] did not work > ******************************************************************************* > > 4. Content of hypre installation: > > $ ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > /cygdrive/c/Lib/hypre-2.8.0b-win64-release: > include lib > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/include: > Data.h HYPRE_krylov.h HYPRE_struct_mv.h _hypre_IJ_mv.h csr_block_matrix.h par_csr_block_matrix.h > HYPRE.h HYPRE_lobpcg.h HYPREf.h _hypre_parcsr_ls.h csr_matmultivec.h par_csr_matmultivec.h > HYPRE_DistributedMatrixPilutSolver_protos.h HYPRE_matrix_matrix_protos.h LLNL_FEI_Fei.h _hypre_parcsr_mv.h distributed_matrix.h par_csr_pmvcomm.h > HYPRE_DistributedMatrixPilutSolver_types.h HYPRE_parcsr_ls.h LLNL_FEI_Impl.h _hypre_sstruct_ls.h fei_defs.h par_multivector.h > HYPRE_FEI_includes.h HYPRE_parcsr_mv.h LLNL_FEI_LSCore.h _hypre_sstruct_mv.h hypre_cfei.h seq_multivector.h > HYPRE_IJ_mv.h HYPRE_seq_mv.h LLNL_FEI_Matrix.h _hypre_struct_ls.h interpreter.h seq_mv.h > HYPRE_LinSysCore.h HYPRE_sstruct_ls.h LLNL_FEI_Solver.h _hypre_struct_mv.h krylov.h temp_multivector.h > HYPRE_MatvecFunctions.h HYPRE_sstruct_mv.h LinearSystemCore.h cfei-hypre.h lobpcg.h > HYPRE_config.h HYPRE_struct_ls.h Lookup.h cfei_hypre.h multivector.h > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib: > libHYPRE.a > > 5. I used compilation option -O3 for PETSc, but -O2 for hypre, can this be a problem? > > Thanks, > Qin > > On Wednesday, November 6, 2013 4:11 PM, Satish Balay wrote: > 1. use latest hypre version - i.e > https://computation.llnl.gov/casc/hypre/download/hypre-2.9.1a.tar.gz > > 2. Make sure its buit with -MT compiler option > > 3. specify hypre with --with-hypre-include and --with-hypre-lib options > > 4. send us the following [so we know whats there - and why configure is not finding it] > > ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > Satish > > > On Wed, 6 Nov 2013, Qin Lu wrote: > >> Hello, >> >> I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: >> >> --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work >> >> The configure.log is attached. >> >> Many thanks for any suggestions, >> >> Qin From balay at mcs.anl.gov Wed Nov 6 18:37:37 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Wed, 6 Nov 2013 18:37:37 -0600 (CST) Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 In-Reply-To: <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> References: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> Message-ID: For one - the --with-hypre-lib would be: --with-hypre-lib=/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib/libHYPRE.a However - it appears you built hypre with gcc - not icc [which I thought was your goal] If gcc/gfortran would work - you might as well install hypre via petsc configure [--download-hypre] Satish > Satis, > > 1. hypre-2.9 did not build in Win-7 for some reason, so I still use hypre-2.8.0b. > 2. I?used CXXFLAGS="-O2 -MT" for hypre configure. > 3. Tried --with-hypre-include and --with-hypre-lib but still got error: > > ?===============================================================================???????????????????????????????????????????????????????????????????????????????????????????????? TESTING: check from config.libraries(config/BuildSystem/config/libraries.py:145)????????????????????????????????????????????????????????????????????????????????????????????????******************************************************************************* > UNABLE to CONFIGURE with GIVEN OPTIONS????(see configure.log for details): > ------------------------------------------------------------------------------- > --with-hypre-lib=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib'] and > --with-hypre-include=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/include'] did not work > ******************************************************************************* > > 4. Content of hypre installation: > ? > $ ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > /cygdrive/c/Lib/hypre-2.8.0b-win64-release: > include??lib > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/include: > Data.h?????????????????????????????????????? HYPRE_krylov.h????????????????HYPRE_struct_mv.h?? _hypre_IJ_mv.h?????? csr_block_matrix.h????par_csr_block_matrix.h > HYPRE.h??????????????????????????????????????HYPRE_lobpcg.h????????????????HYPREf.h????????????_hypre_parcsr_ls.h?? csr_matmultivec.h???? par_csr_matmultivec.h > HYPRE_DistributedMatrixPilutSolver_protos.h??HYPRE_matrix_matrix_protos.h??LLNL_FEI_Fei.h??????_hypre_parcsr_mv.h?? distributed_matrix.h??par_csr_pmvcomm.h > HYPRE_DistributedMatrixPilutSolver_types.h?? HYPRE_parcsr_ls.h???????????? LLNL_FEI_Impl.h???? _hypre_sstruct_ls.h??fei_defs.h????????????par_multivector.h > HYPRE_FEI_includes.h???????????????????????? HYPRE_parcsr_mv.h???????????? LLNL_FEI_LSCore.h?? _hypre_sstruct_mv.h??hypre_cfei.h??????????seq_multivector.h > HYPRE_IJ_mv.h????????????????????????????????HYPRE_seq_mv.h????????????????LLNL_FEI_Matrix.h?? _hypre_struct_ls.h?? interpreter.h???????? seq_mv.h > HYPRE_LinSysCore.h?????????????????????????? HYPRE_sstruct_ls.h????????????LLNL_FEI_Solver.h?? _hypre_struct_mv.h?? krylov.h??????????????temp_multivector.h > HYPRE_MatvecFunctions.h??????????????????????HYPRE_sstruct_mv.h????????????LinearSystemCore.h??cfei-hypre.h???????? lobpcg.h > HYPRE_config.h?????????????????????????????? HYPRE_struct_ls.h???????????? Lookup.h????????????cfei_hypre.h???????? multivector.h > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib: > libHYPRE.a > > 5. I used compilation option -O3 for PETSc, but -O2 for hypre, can this be a problem? > ? > Thanks, > Qin > > On Wednesday, November 6, 2013 4:11 PM, Satish Balay wrote: > 1. use latest hypre version - i.e > https://computation.llnl.gov/casc/hypre/download/hypre-2.9.1a.tar.gz > > 2. Make sure its buit with -MT compiler option > > 3. specify hypre with --with-hypre-include and --with-hypre-lib options > > 4. send us the following [so we know whats there - and why configure is not finding it] > > ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > Satish > > > On Wed, 6 Nov 2013, Qin Lu wrote: > > > Hello, > > ? > > I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I?want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: > > ? > > --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work > > ? > > The configure.log is attached. > > ? > > Many thanks for any suggestions, > > ? > > Qin???? > From rongliang.chan at gmail.com Wed Nov 6 20:32:13 2013 From: rongliang.chan at gmail.com (Rongliang Chen) Date: Thu, 07 Nov 2013 10:32:13 +0800 Subject: [petsc-users] Debug AOCreateBasic In-Reply-To: <877gcmxfwv.fsf@mcs.anl.gov> References: <52733BB5.5030203@gmail.com> <8761scheg2.fsf@mcs.anl.gov> <52734F7B.2060002@gmail.com> <8761sbpfq0.fsf@mcs.anl.gov> <5279D10A.8080204@gmail.com> <877gcmxfwv.fsf@mcs.anl.gov> Message-ID: <527AFBAD.6030904@gmail.com> Hi Jed, I have not find a way to "dump core on selected ranks" yet and I will continue to do that. I run my code with the option "-on_error_attach_debugger" and got the following message: -------------------------------------------------------------------------- An MPI process has executed an operation involving a call to the "fork()" system call to create a child process. Open MPI is currently operating in a condition that could result in memory corruption or other system errors; your MPI job may hang, crash, or produce silent data corruption. The use of fork() (or system() or other calls that create child processes) is strongly discouraged. The process that invoked fork was: Local host: node1529 (PID 3701) MPI_COMM_WORLD rank: 0 If you are *absolutely sure* that your application will successfully and correctly survive a call to fork(), you may disable this warning by setting the mpi_warn_on_fork MCA parameter to 0. -------------------------------------------------------------------------- [node1529:03700] 13 more processes have sent help message help-mpi-runtime.txt / mpi_init:warn-fork [node1529:03700] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages -------------------------------------------------------------------------- Is this message useful for the debugging? Best, Rongliang On 11/06/2013 01:22 PM, Jed Brown wrote: > Rongliang Chen writes: > >> Hi Jed, >> >> I tested my code on a locally 16-cores workstation. Both the MPICH and >> OPENMPI version work well for the 30 million unstructured meshes case >> (this workstation has 256G memory). > Then I would start thinking about an MPI problem and see if you can > configure to get a core dump. > >> My code is also valgrind-clean on this workstation. > Good. From jedbrown at mcs.anl.gov Wed Nov 6 20:38:38 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 06 Nov 2013 19:38:38 -0700 Subject: [petsc-users] Debug AOCreateBasic In-Reply-To: <527AFBAD.6030904@gmail.com> References: <52733BB5.5030203@gmail.com> <8761scheg2.fsf@mcs.anl.gov> <52734F7B.2060002@gmail.com> <8761sbpfq0.fsf@mcs.anl.gov> <5279D10A.8080204@gmail.com> <877gcmxfwv.fsf@mcs.anl.gov> <527AFBAD.6030904@gmail.com> Message-ID: <87siv9szpt.fsf@mcs.anl.gov> Rongliang Chen writes: > Hi Jed, > > I have not find a way to "dump core on selected ranks" yet and I will > continue to do that. Ask the administrators at your facility. There are a few common ways, but I'm not going to play a guessing game on the mailing list. > I run my code with the option "-on_error_attach_debugger" and got the > following message: > > -------------------------------------------------------------------------- > An MPI process has executed an operation involving a call to the > "fork()" system call to create a child process. Open MPI is currently > operating in a condition that could result in memory corruption or > other system errors; your MPI job may hang, crash, or produce silent > data corruption. The use of fork() (or system() or other calls that > create child processes) is strongly discouraged. > > The process that invoked fork was: > > Local host: node1529 (PID 3701) > MPI_COMM_WORLD rank: 0 > > If you are *absolutely sure* that your application will successfully > and correctly survive a call to fork(), you may disable this warning > by setting the mpi_warn_on_fork MCA parameter to 0. > -------------------------------------------------------------------------- > [node1529:03700] 13 more processes have sent help message > help-mpi-runtime.txt / mpi_init:warn-fork > [node1529:03700] Set MCA parameter "orte_base_help_aggregate" to 0 to > see all help / error messages > -------------------------------------------------------------------------- > > Is this message useful for the debugging? This is just a possibly technical problem attaching a debugger in your environment, but you have to actually attach the debugger and poke around (stack trace, etc). Can you create an interactive session and run your job from there? -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From charlesj at purdue.edu Wed Nov 6 23:27:33 2013 From: charlesj at purdue.edu (James A Charles) Date: Thu, 7 Nov 2013 00:27:33 -0500 (EST) Subject: [petsc-users] QR factorization of Parallel Dense Matrix Message-ID: <897242198.367610.1383802053952.JavaMail.root@mailhub050.itcs.purdue.edu> Hello, I have a parallel dense matrix, how can I perform a QR factorization? Thanks, James From jedbrown at mcs.anl.gov Wed Nov 6 23:53:25 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 06 Nov 2013 22:53:25 -0700 Subject: [petsc-users] QR factorization of Parallel Dense Matrix In-Reply-To: <897242198.367610.1383802053952.JavaMail.root@mailhub050.itcs.purdue.edu> References: <897242198.367610.1383802053952.JavaMail.root@mailhub050.itcs.purdue.edu> Message-ID: <871u2su59m.fsf@mcs.anl.gov> James A Charles writes: > Hello, > > I have a parallel dense matrix, how can I perform a QR factorization? What dimension is the matrix and how do you want the factors distributed? -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From gryllida at fastmail.fm Thu Nov 7 00:39:24 2013 From: gryllida at fastmail.fm (Gryllida) Date: Thu, 07 Nov 2013 17:09:24 +1030 Subject: [petsc-users] stensils Message-ID: <1383806364.13625.44181269.7618017B@webmail.messagingengine.com> http://courses.engr.illinois.edu/cs554/fa2011/notes/petscpde.pdf has a "this is a simple but inefficient way to set up the matrix" at slide 18. Later they set it as a stensil instead. What is a recommended example of doing the same with OPENMP in FORTRAN, if I have matrices A_central, A_top, A_bottom, A_left, A_right, containing the stensil items? Svetlana From lu_qin_2000 at yahoo.com Thu Nov 7 09:13:34 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Thu, 7 Nov 2013 07:13:34 -0800 (PST) Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 In-Reply-To: References: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> Message-ID: <1383837214.32247.YahooMailNeo@web160202.mail.bf1.yahoo.com> Satish, ? I tried what you suggested for petsc configure but it still did not work. See the attached configure.log. ? How did you see I was using gcc for hypre? I think I used Intel compiler icl since I specified "CC=icl CXX=icl" for hypre configure. I double-checked hypre's Makefile.config and CC and CXX are icl. ? Thanks, Qin On Wednesday, November 6, 2013 6:37 PM, Satish Balay wrote: For one - the? --with-hypre-lib would be: --with-hypre-lib=/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib/libHYPRE.a However - it appears you built hypre with gcc - not icc [which I thought was your goal] If gcc/gfortran would work - you might as well install hypre via petsc configure [--download-hypre] Satish > Satis, > > 1. hypre-2.9 did not build in Win-7 for some reason, so I still use hypre-2.8.0b. > 2. I?used CXXFLAGS="-O2 -MT" for hypre configure. > 3. Tried --with-hypre-include and --with-hypre-lib but still got error: > > ?===============================================================================???????????????????????????????????????????????????????????????????????????????????????????????? TESTING: check from config.libraries(config/BuildSystem/config/libraries.py:145)????????????????????????????????????????????????????????????????????????????????????????????????******************************************************************************* > UNABLE to CONFIGURE with GIVEN OPTIONS????(see configure.log for details): > ------------------------------------------------------------------------------- > --with-hypre-lib=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib'] and > --with-hypre-include=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/include'] did not work > ******************************************************************************* > > 4. Content of hypre installation: > ? > $ ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > /cygdrive/c/Lib/hypre-2.8.0b-win64-release: > include??lib > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/include: > Data.h?????????????????????????????????????? HYPRE_krylov.h????????????????HYPRE_struct_mv.h?? _hypre_IJ_mv.h?????? csr_block_matrix.h????par_csr_block_matrix.h > HYPRE.h??????????????????????????????????????HYPRE_lobpcg.h????????????????HYPREf.h????????????_hypre_parcsr_ls.h?? csr_matmultivec.h???? par_csr_matmultivec.h > HYPRE_DistributedMatrixPilutSolver_protos.h??HYPRE_matrix_matrix_protos.h??LLNL_FEI_Fei.h??????_hypre_parcsr_mv.h?? distributed_matrix.h??par_csr_pmvcomm.h > HYPRE_DistributedMatrixPilutSolver_types.h?? HYPRE_parcsr_ls.h???????????? LLNL_FEI_Impl.h???? _hypre_sstruct_ls.h??fei_defs.h????????????par_multivector.h > HYPRE_FEI_includes.h???????????????????????? HYPRE_parcsr_mv.h???????????? LLNL_FEI_LSCore.h?? _hypre_sstruct_mv.h??hypre_cfei.h??????????seq_multivector.h > HYPRE_IJ_mv.h????????????????????????????????HYPRE_seq_mv.h????????????????LLNL_FEI_Matrix.h?? _hypre_struct_ls.h?? interpreter.h???????? seq_mv.h > HYPRE_LinSysCore.h?????????????????????????? HYPRE_sstruct_ls.h????????????LLNL_FEI_Solver.h?? _hypre_struct_mv.h?? krylov.h??????????????temp_multivector.h > HYPRE_MatvecFunctions.h??????????????????????HYPRE_sstruct_mv.h????????????LinearSystemCore.h??cfei-hypre.h???????? lobpcg.h > HYPRE_config.h?????????????????????????????? HYPRE_struct_ls.h???????????? Lookup.h????????????cfei_hypre.h???????? multivector.h > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib: > libHYPRE.a > > 5. I used compilation option -O3 for PETSc, but -O2 for hypre, can this be a problem? > ? > Thanks, > Qin > > On Wednesday, November 6, 2013 4:11 PM, Satish Balay wrote: > 1. use latest hypre version - i.e > https://computation.llnl.gov/casc/hypre/download/hypre-2.9.1a.tar.gz > > 2. Make sure its buit with -MT compiler option > > 3. specify hypre with --with-hypre-include and --with-hypre-lib options > > 4. send us the following [so we know whats there - and why configure is not finding it] > > ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > Satish > > > On Wed, 6 Nov 2013, Qin Lu wrote: > > > Hello, > > ? > > I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I?want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: > > ? > > --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work > > ? > > The configure.log is attached. > > ? > > Many thanks for any suggestions, > > ? > > Qin???? > -------------- next part -------------- A non-text attachment was scrubbed... Name: configure.log Type: application/octet-stream Size: 3377246 bytes Desc: not available URL: From balay at mcs.anl.gov Thu Nov 7 09:20:16 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Thu, 7 Nov 2013 09:20:16 -0600 (CST) Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 In-Reply-To: <1383837214.32247.YahooMailNeo@web160202.mail.bf1.yahoo.com> References: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> <1383837214.32247.YahooMailNeo@web160202.mail.bf1.yahoo.com> Message-ID: On Thu, 7 Nov 2013, Qin Lu wrote: > Satish, > ? > I tried what you suggested for petsc configure but it still did not work. See the attached configure.log. > ? > How did you see I was using gcc for hypre? For cl/icl the library should be libhypre.lib or hypre.lib [not libhypre.a] > I think I used Intel compiler icl since I specified "CC=icl CXX=icl" for hypre configure. I double-checked hypre's Makefile.config and CC and CXX are icl. If you follow the instructions I copy/pasted in my previous e-mail regarding 'windows' installation - there is no mention of using configure - but of using cmake. [I think you have to install/use native windows cmake - not the one from cygwin] Satish > ? > Thanks, > Qin > > > > On Wednesday, November 6, 2013 6:37 PM, Satish Balay wrote: > For one - the? --with-hypre-lib would be: > > --with-hypre-lib=/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib/libHYPRE.a > > However - it appears you built hypre with gcc - not icc [which I thought was your goal] > > If gcc/gfortran would work - you might as well install hypre via petsc configure [--download-hypre] > > Satish > > > > Satis, > > > > 1. hypre-2.9 did not build in Win-7 for some reason, so I still use hypre-2.8.0b. > > 2. I?used CXXFLAGS="-O2 -MT" for hypre configure. > > 3. Tried --with-hypre-include and --with-hypre-lib but still got error: > > > > ?===============================================================================???????????????????????????????????????????????????????????????????????????????????????????????? TESTING: check from config.libraries(config/BuildSystem/config/libraries.py:145)????????????????????????????????????????????????????????????????????????????????????????????????******************************************************************************* > > UNABLE to CONFIGURE with GIVEN OPTIONS????(see configure.log for details): > > ------------------------------------------------------------------------------- > > --with-hypre-lib=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib'] and > > --with-hypre-include=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/include'] did not work > > ******************************************************************************* > > > > 4. Content of hypre installation: > > ? > > $ ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release: > > include??lib > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/include: > > Data.h?????????????????????????????????????? HYPRE_krylov.h????????????????HYPRE_struct_mv.h?? _hypre_IJ_mv.h?????? csr_block_matrix.h????par_csr_block_matrix.h > > HYPRE.h??????????????????????????????????????HYPRE_lobpcg.h????????????????HYPREf.h????????????_hypre_parcsr_ls.h?? csr_matmultivec.h???? par_csr_matmultivec.h > > HYPRE_DistributedMatrixPilutSolver_protos.h??HYPRE_matrix_matrix_protos.h??LLNL_FEI_Fei.h??????_hypre_parcsr_mv.h?? distributed_matrix.h??par_csr_pmvcomm.h > > HYPRE_DistributedMatrixPilutSolver_types.h?? HYPRE_parcsr_ls.h???????????? LLNL_FEI_Impl.h???? _hypre_sstruct_ls.h??fei_defs.h????????????par_multivector.h > > HYPRE_FEI_includes.h???????????????????????? HYPRE_parcsr_mv.h???????????? LLNL_FEI_LSCore.h?? _hypre_sstruct_mv.h??hypre_cfei.h??????????seq_multivector.h > > HYPRE_IJ_mv.h????????????????????????????????HYPRE_seq_mv.h????????????????LLNL_FEI_Matrix.h?? _hypre_struct_ls.h?? interpreter.h???????? seq_mv.h > > HYPRE_LinSysCore.h?????????????????????????? HYPRE_sstruct_ls.h????????????LLNL_FEI_Solver.h?? _hypre_struct_mv.h?? krylov.h??????????????temp_multivector.h > > HYPRE_MatvecFunctions.h??????????????????????HYPRE_sstruct_mv.h????????????LinearSystemCore.h??cfei-hypre.h???????? lobpcg.h > > HYPRE_config.h?????????????????????????????? HYPRE_struct_ls.h???????????? Lookup.h????????????cfei_hypre.h???????? multivector.h > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib: > > libHYPRE.a > > > > 5. I used compilation option -O3 for PETSc, but -O2 for hypre, can this be a problem? > > ? > > Thanks, > > Qin > > > > On Wednesday, November 6, 2013 4:11 PM, Satish Balay wrote: > > 1. use latest hypre version - i.e > > https://computation.llnl.gov/casc/hypre/download/hypre-2.9.1a.tar.gz > > > > 2. Make sure its buit with -MT compiler option > > > > 3. specify hypre with --with-hypre-include and --with-hypre-lib options > > > > 4. send us the following [so we know whats there - and why configure is not finding it] > > > > ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > > > Satish > > > > > > On Wed, 6 Nov 2013, Qin Lu wrote: > > > > > Hello, > > > ? > > > I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I?want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: > > > ? > > > --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work > > > ? > > > The configure.log is attached. > > > ? > > > Many thanks for any suggestions, > > > ? > > > Qin???? > > From lu_qin_2000 at yahoo.com Thu Nov 7 09:56:20 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Thu, 7 Nov 2013 07:56:20 -0800 (PST) Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 In-Reply-To: References: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> <1383837214.32247.YahooMailNeo@web160202.mail.bf1.yahoo.com> Message-ID: <1383839780.27927.YahooMailNeo@web160204.mail.bf1.yahoo.com> I?managed to build hypre-2.8.0b with cygwin (instead of cmake) and it worked, but have no idea why it created libhypre.a rather than libhypre.lib. From hypre build's screen output, it did use icl. Please let me know if there is a way to force to create .lib file. Anyway, PETSc's configure should not care how?hypre was built, right? I tried to rename libhypre.a to libhypre.lib (for testing purpose) but got the same error. Thank, Qin? On Thursday, November 7, 2013 9:20 AM, Satish Balay wrote: On Thu, 7 Nov 2013, Qin Lu wrote: > Satish, > ? > I tried what you suggested for petsc configure but it still did not work. See the attached configure.log. > ? > How did you see I was using gcc for hypre? For cl/icl the library should be libhypre.lib or hypre.lib [not libhypre.a] > I think I used Intel compiler icl since I specified "CC=icl CXX=icl" for hypre configure. I double-checked hypre's Makefile.config and CC and CXX are icl. If you follow the instructions I copy/pasted in my previous e-mail regarding 'windows' installation - there is no mention of using configure - but of using cmake. [I think you have to install/use native windows cmake - not the one from cygwin] Satish > ? > Thanks, > Qin > > > > On Wednesday, November 6, 2013 6:37 PM, Satish Balay wrote: > For one - the? --with-hypre-lib would be: > > --with-hypre-lib=/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib/libHYPRE.a > > However - it appears you built hypre with gcc - not icc [which I thought was your goal] > > If gcc/gfortran would work - you might as well install hypre via petsc configure [--download-hypre] > > Satish > > > > Satis, > > > > 1. hypre-2.9 did not build in Win-7 for some reason, so I still use hypre-2.8.0b. > > 2. I?used CXXFLAGS="-O2 -MT" for hypre configure. > > 3. Tried --with-hypre-include and --with-hypre-lib but still got error: > > > > ?===============================================================================???????????????????????????????????????????????????????????????????????????????????????????????? TESTING: check from config.libraries(config/BuildSystem/config/libraries.py:145)????????????????????????????????????????????????????????????????????????????????????????????????******************************************************************************* > > UNABLE to CONFIGURE with GIVEN OPTIONS????(see configure.log for details): > > ------------------------------------------------------------------------------- > > --with-hypre-lib=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib'] and > > --with-hypre-include=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/include'] did not work > > ******************************************************************************* > > > > 4. Content of hypre installation: > > ? > > $ ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release: > > include??lib > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/include: > > Data.h?????????????????????????????????????? HYPRE_krylov.h????????????????HYPRE_struct_mv.h?? _hypre_IJ_mv.h?????? csr_block_matrix.h????par_csr_block_matrix.h > > HYPRE.h??????????????????????????????????????HYPRE_lobpcg.h????????????????HYPREf.h????????????_hypre_parcsr_ls.h?? csr_matmultivec.h???? par_csr_matmultivec.h > > HYPRE_DistributedMatrixPilutSolver_protos.h??HYPRE_matrix_matrix_protos.h??LLNL_FEI_Fei.h??????_hypre_parcsr_mv.h?? distributed_matrix.h??par_csr_pmvcomm.h > > HYPRE_DistributedMatrixPilutSolver_types.h?? HYPRE_parcsr_ls.h???????????? LLNL_FEI_Impl.h???? _hypre_sstruct_ls.h??fei_defs.h????????????par_multivector.h > > HYPRE_FEI_includes.h???????????????????????? HYPRE_parcsr_mv.h???????????? LLNL_FEI_LSCore.h?? _hypre_sstruct_mv.h??hypre_cfei.h??????????seq_multivector.h > > HYPRE_IJ_mv.h????????????????????????????????HYPRE_seq_mv.h????????????????LLNL_FEI_Matrix.h?? _hypre_struct_ls.h?? interpreter.h???????? seq_mv.h > > HYPRE_LinSysCore.h?????????????????????????? HYPRE_sstruct_ls.h????????????LLNL_FEI_Solver.h?? _hypre_struct_mv.h?? krylov.h??????????????temp_multivector.h > > HYPRE_MatvecFunctions.h??????????????????????HYPRE_sstruct_mv.h????????????LinearSystemCore.h??cfei-hypre.h???????? lobpcg.h > > HYPRE_config.h?????????????????????????????? HYPRE_struct_ls.h???????????? Lookup.h????????????cfei_hypre.h???????? multivector.h > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib: > > libHYPRE.a > > > > 5. I used compilation option -O3 for PETSc, but -O2 for hypre, can this be a problem? > > ? > > Thanks, > > Qin > > > > On Wednesday, November 6, 2013 4:11 PM, Satish Balay wrote: > > 1. use latest hypre version - i.e > > https://computation.llnl.gov/casc/hypre/download/hypre-2.9.1a.tar.gz > > > > 2. Make sure its buit with -MT compiler option > > > > 3. specify hypre with --with-hypre-include and --with-hypre-lib options > > > > 4. send us the following [so we know whats there - and why configure is not finding it] > > > > ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > > > Satish > > > > > > On Wed, 6 Nov 2013, Qin Lu wrote: > > > > > Hello, > > > ? > > > I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I?want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: > > > ? > > > --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work > > > ? > > > The configure.log is attached. > > > ? > > > Many thanks for any suggestions, > > > ? > > > Qin???? > > From balay at mcs.anl.gov Thu Nov 7 10:06:47 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Thu, 7 Nov 2013 10:06:47 -0600 (CST) Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 In-Reply-To: <1383839780.27927.YahooMailNeo@web160204.mail.bf1.yahoo.com> References: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> <1383837214.32247.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383839780.27927.YahooMailNeo@web160204.mail.bf1.yahoo.com> Message-ID: Sorry - looks like you are ignoring things I say. [I pointed you many times to the hypre installation instructions which you are ingoring. And then complaining that things don't work] And there is no consistancy in your requirements. [you say you want to build hypre with icc - but then you say you've built it with cygwin - i.e gcc? And then expect it to work with icc via petsc configure? I see no point in replying any further. Its a waste of time.. Satish On Thu, 7 Nov 2013, Qin Lu wrote: > I?managed to build hypre-2.8.0b with cygwin (instead of cmake) and it worked, but have no idea why it created libhypre.a rather than libhypre.lib. From hypre build's screen output, it did use icl. Please let me know if there is a way to force to create .lib file. > > Anyway, PETSc's configure should not care how?hypre was built, right? I tried to rename libhypre.a to libhypre.lib (for testing purpose) but got the same error. > > Thank, > Qin? > > On Thursday, November 7, 2013 9:20 AM, Satish Balay wrote: > > On Thu, 7 Nov 2013, Qin Lu wrote: > > > Satish, > > ? > > I tried what you suggested for petsc configure but it still did not work. See the attached configure.log. > > ? > > How did you see I was using gcc for hypre? > > For cl/icl the library should be libhypre.lib or hypre.lib [not libhypre.a] > > > I think I used Intel compiler icl since I specified "CC=icl CXX=icl" for hypre configure. I double-checked hypre's Makefile.config and CC and CXX are icl. > > If you follow the instructions I copy/pasted in my previous e-mail regarding 'windows' installation > - there is no mention of using configure - but of using cmake. [I think you have to install/use native > windows cmake - not the one from cygwin] > > > Satish > > > ? > > Thanks, > > Qin > > > > > > > > On Wednesday, November 6, 2013 6:37 PM, Satish Balay wrote: > > For one - the? --with-hypre-lib would be: > > > > --with-hypre-lib=/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib/libHYPRE.a > > > > However - it appears you built hypre with gcc - not icc [which I thought was your goal] > > > > If gcc/gfortran would work - you might as well install hypre via petsc configure [--download-hypre] > > > > Satish > > > > > > > Satis, > > > > > > 1. hypre-2.9 did not build in Win-7 for some reason, so I still use hypre-2.8.0b. > > > 2. I?used CXXFLAGS="-O2 -MT" for hypre configure. > > > 3. Tried --with-hypre-include and --with-hypre-lib but still got error: > > > > > > ?===============================================================================???????????????????????????????????????????????????????????????????????????????????????????????? TESTING: check from config.libraries(config/BuildSystem/config/libraries.py:145)????????????????????????????????????????????????????????????????????????????????????????????????******************************************************************************* > > > UNABLE to CONFIGURE with GIVEN OPTIONS????(see configure.log for details): > > > ------------------------------------------------------------------------------- > > > --with-hypre-lib=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib'] and > > > --with-hypre-include=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/include'] did not work > > > ******************************************************************************* > > > > > > 4. Content of hypre installation: > > > ? > > > $ ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release: > > > include??lib > > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/include: > > > Data.h?????????????????????????????????????? HYPRE_krylov.h????????????????HYPRE_struct_mv.h?? _hypre_IJ_mv.h?????? csr_block_matrix.h????par_csr_block_matrix.h > > > HYPRE.h??????????????????????????????????????HYPRE_lobpcg.h????????????????HYPREf.h????????????_hypre_parcsr_ls.h?? csr_matmultivec.h???? par_csr_matmultivec.h > > > HYPRE_DistributedMatrixPilutSolver_protos.h??HYPRE_matrix_matrix_protos.h??LLNL_FEI_Fei.h??????_hypre_parcsr_mv.h?? distributed_matrix.h??par_csr_pmvcomm.h > > > HYPRE_DistributedMatrixPilutSolver_types.h?? HYPRE_parcsr_ls.h???????????? LLNL_FEI_Impl.h???? _hypre_sstruct_ls.h??fei_defs.h????????????par_multivector.h > > > HYPRE_FEI_includes.h???????????????????????? HYPRE_parcsr_mv.h???????????? LLNL_FEI_LSCore.h?? _hypre_sstruct_mv.h??hypre_cfei.h??????????seq_multivector.h > > > HYPRE_IJ_mv.h????????????????????????????????HYPRE_seq_mv.h????????????????LLNL_FEI_Matrix.h?? _hypre_struct_ls.h?? interpreter.h???????? seq_mv.h > > > HYPRE_LinSysCore.h?????????????????????????? HYPRE_sstruct_ls.h????????????LLNL_FEI_Solver.h?? _hypre_struct_mv.h?? krylov.h??????????????temp_multivector.h > > > HYPRE_MatvecFunctions.h??????????????????????HYPRE_sstruct_mv.h????????????LinearSystemCore.h??cfei-hypre.h???????? lobpcg.h > > > HYPRE_config.h?????????????????????????????? HYPRE_struct_ls.h???????????? Lookup.h????????????cfei_hypre.h???????? multivector.h > > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib: > > > libHYPRE.a > > > > > > 5. I used compilation option -O3 for PETSc, but -O2 for hypre, can this be a problem? > > > ? > > > Thanks, > > > Qin > > > > > > On Wednesday, November 6, 2013 4:11 PM, Satish Balay wrote: > > > 1. use latest hypre version - i.e > > > https://computation.llnl.gov/casc/hypre/download/hypre-2.9.1a.tar.gz > > > > > > 2. Make sure its buit with -MT compiler option > > > > > > 3. specify hypre with --with-hypre-include and --with-hypre-lib options > > > > > > 4. send us the following [so we know whats there - and why configure is not finding it] > > > > > > ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > > > > > Satish > > > > > > > > > On Wed, 6 Nov 2013, Qin Lu wrote: > > > > > > > Hello, > > > > ? > > > > I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I?want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: > > > > ? > > > > --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work > > > > ? > > > > The configure.log is attached. > > > > ? > > > > Many thanks for any suggestions, > > > > ? > > > > Qin???? > > > > From lu_qin_2000 at yahoo.com Thu Nov 7 10:25:46 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Thu, 7 Nov 2013 08:25:46 -0800 (PST) Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 In-Reply-To: References: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> <1383837214.32247.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383839780.27927.YahooMailNeo@web160204.mail.bf1.yahoo.com> Message-ID: <1383841546.84925.YahooMailNeo@web160203.mail.bf1.yahoo.com> Satish, ? Sorry if I caused any confusions. Let me clarify: ? 1. I built hypre-2.8.0b using cygwin with Intel compiler icl. It created libHYPRE.a rather than libHYPRE.lib, which I don't know why. I have never said I used gcc, since I mentioned several times that I used "CC=icl CXX=icl" in hypre's configure options. ? 2. I am configuring PETSc-3.4.2 using cygwin with Intel comiler icl and with hypre-2.8.0b, but --with-hypre-lib, etc., did not work. ? 3. If it is known that?hypre lib built by cygwin (with icl) does not work for PETSc, please let me know. I will then figure out how to build it with cmake; if this is not the case, at least we should understand why --with-hypre-lib, etc., did not work. Thanks, Qin ? On Thursday, November 7, 2013 10:06 AM, Satish Balay wrote: Sorry - looks like you are ignoring things I say. [I pointed you many times to the hypre installation instructions which you are ingoring.? And then complaining that things don't work] And there is no consistancy in your requirements. [you say you want to build hypre with icc - but then you say you've built it with cygwin - i.e gcc? And then expect it to work with icc via petsc configure? I see no point in replying any further. Its a waste of time.. Satish On Thu, 7 Nov 2013, Qin Lu wrote: > I?managed to build hypre-2.8.0b with cygwin (instead of cmake) and it worked, but have no idea why it created libhypre.a rather than libhypre.lib. From hypre build's screen output, it did use icl. Please let me know if there is a way to force to create .lib file. > > Anyway, PETSc's configure should not care how?hypre was built, right? I tried to rename libhypre.a to libhypre.lib (for testing purpose) but got the same error. > > Thank, > Qin? > > On Thursday, November 7, 2013 9:20 AM, Satish Balay wrote: > > On Thu, 7 Nov 2013, Qin Lu wrote: > > > Satish, > > ? > > I tried what you suggested for petsc configure but it still did not work. See the attached configure.log. > > ? > > How did you see I was using gcc for hypre? > > For cl/icl the library should be libhypre.lib or hypre.lib [not libhypre.a] > > > I think I used Intel compiler icl since I specified "CC=icl CXX=icl" for hypre configure. I double-checked hypre's Makefile.config and CC and CXX are icl. > > If you follow the instructions I copy/pasted in my previous e-mail regarding 'windows' installation > - there is no mention of using configure - but of using cmake. [I think you have to install/use native > windows cmake - not the one from cygwin] > > > Satish > > > ? > > Thanks, > > Qin > > > > > > > > On Wednesday, November 6, 2013 6:37 PM, Satish Balay wrote: > > For one - the? --with-hypre-lib would be: > > > > --with-hypre-lib=/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib/libHYPRE.a > > > > However - it appears you built hypre with gcc - not icc [which I thought was your goal] > > > > If gcc/gfortran would work - you might as well install hypre via petsc configure [--download-hypre] > > > > Satish > > > > > > > Satis, > > > > > > 1. hypre-2.9 did not build in Win-7 for some reason, so I still use hypre-2.8.0b. > > > 2. I?used CXXFLAGS="-O2 -MT" for hypre configure. > > > 3. Tried --with-hypre-include and --with-hypre-lib but still got error: > > > > > > ?===============================================================================???????????????????????????????????????????????????????????????????????????????????????????????? TESTING: check from config.libraries(config/BuildSystem/config/libraries.py:145)????????????????????????????????????????????????????????????????????????????????????????????????******************************************************************************* > > > UNABLE to CONFIGURE with GIVEN OPTIONS????(see configure.log for details): > > > ------------------------------------------------------------------------------- > > > --with-hypre-lib=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib'] and > > > --with-hypre-include=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/include'] did not work > > > ******************************************************************************* > > > > > > 4. Content of hypre installation: > > > ? > > > $ ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release: > > > include??lib > > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/include: > > > Data.h?????????????????????????????????????? HYPRE_krylov.h????????????????HYPRE_struct_mv.h?? _hypre_IJ_mv.h?????? csr_block_matrix.h????par_csr_block_matrix.h > > > HYPRE.h??????????????????????????????????????HYPRE_lobpcg.h????????????????HYPREf.h????????????_hypre_parcsr_ls.h?? csr_matmultivec.h???? par_csr_matmultivec.h > > > HYPRE_DistributedMatrixPilutSolver_protos.h??HYPRE_matrix_matrix_protos.h??LLNL_FEI_Fei.h??????_hypre_parcsr_mv.h?? distributed_matrix.h??par_csr_pmvcomm.h > > > HYPRE_DistributedMatrixPilutSolver_types.h?? HYPRE_parcsr_ls.h???????????? LLNL_FEI_Impl.h???? _hypre_sstruct_ls.h??fei_defs.h????????????par_multivector.h > > > HYPRE_FEI_includes.h???????????????????????? HYPRE_parcsr_mv.h???????????? LLNL_FEI_LSCore.h?? _hypre_sstruct_mv.h??hypre_cfei.h??????????seq_multivector.h > > > HYPRE_IJ_mv.h????????????????????????????????HYPRE_seq_mv.h????????????????LLNL_FEI_Matrix.h?? _hypre_struct_ls.h?? interpreter.h???????? seq_mv.h > > > HYPRE_LinSysCore.h?????????????????????????? HYPRE_sstruct_ls.h????????????LLNL_FEI_Solver.h?? _hypre_struct_mv.h?? krylov.h??????????????temp_multivector.h > > > HYPRE_MatvecFunctions.h??????????????????????HYPRE_sstruct_mv.h????????????LinearSystemCore.h??cfei-hypre.h???????? lobpcg.h > > > HYPRE_config.h?????????????????????????????? HYPRE_struct_ls.h???????????? Lookup.h????????????cfei_hypre.h???????? multivector.h > > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib: > > > libHYPRE.a > > > > > > 5. I used compilation option -O3 for PETSc, but -O2 for hypre, can this be a problem? > > > ? > > > Thanks, > > > Qin > > > > > > On Wednesday, November 6, 2013 4:11 PM, Satish Balay wrote: > > > 1. use latest hypre version - i.e > > > https://computation.llnl.gov/casc/hypre/download/hypre-2.9.1a.tar.gz > > > > > > 2. Make sure its buit with -MT compiler option > > > > > > 3. specify hypre with --with-hypre-include and --with-hypre-lib options > > > > > > 4. send us the following [so we know whats there - and why configure is not finding it] > > > > > > ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > > > > > Satish > > > > > > > > > On Wed, 6 Nov 2013, Qin Lu wrote: > > > > > > > Hello, > > > > ? > > > > I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I?want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: > > > > ? > > > > --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work > > > > ? > > > > The configure.log is attached. > > > > ? > > > > Many thanks for any suggestions, > > > > ? > > > > Qin???? > > > > From Wadud.Miah at awe.co.uk Thu Nov 7 11:10:42 2013 From: Wadud.Miah at awe.co.uk (Wadud.Miah at awe.co.uk) Date: Thu, 7 Nov 2013 17:10:42 +0000 Subject: [petsc-users] PETSc types in Fortran Message-ID: <201311071710.rA7HAjS5013916@msw2.awe.co.uk> Hello, I want to use some PETSc types in Fortran but am running into some problems. When include the PETSc header file: #include "finclude/petsc.h" which defines the data types, it starts to complain about my MPI header file. If I remove the above include statement, then it complains that the data type is not recognised. Any help will be greatly appreciated. Many thanks, Wadud. ___________________________________________________ ____________________________ The information in this email and in any attachment(s) is commercial in confidence. If you are not the named addressee(s) or if you receive this email in error then any distribution, copying or use of this communication or the information in it is strictly prohibited. Please notify us immediately by email at admin.internet(at)awe.co.uk, and then delete this message from your computer. While attachments are virus checked, AWE plc does not accept any liability in respect of any virus which is not detected. AWE Plc Registered in England and Wales Registration No 02763902 AWE, Aldermaston, Reading, RG7 4PR -------------- next part -------------- An HTML attachment was scrubbed... URL: From bsmith at mcs.anl.gov Thu Nov 7 11:17:39 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Thu, 7 Nov 2013 11:17:39 -0600 Subject: [petsc-users] PETSc types in Fortran In-Reply-To: <201311071710.rA7HAjS5013916@msw2.awe.co.uk> References: <201311071710.rA7HAjS5013916@msw2.awe.co.uk> Message-ID: <1E75A36D-0EDD-4213-8091-22D21300BFB9@mcs.anl.gov> Please send the entire error message when it complains about the MPI header file Barry On Nov 7, 2013, at 11:10 AM, wrote: > Hello, > > I want to use some PETSc types in Fortran but am running into some problems. When include the PETSc header file: > > #include ?finclude/petsc.h? > > which defines the data types, it starts to complain about my MPI header file. If I remove the above include statement, then it complains that the data type is not recognised. Any help will be greatly appreciated. > > Many thanks, > Wadud. > ___________________________________________________ ____________________________ The information in this email and in any attachment(s) is commercial in confidence. If you are not the named addressee(s) or if you receive this email in error then any distribution, copying or use of this communication or the information in it is strictly prohibited. Please notify us immediately by email at admin.internet(at)awe.co.uk, and then delete this message from your computer. While attachments are virus checked, AWE plc does not accept any liability in respect of any virus which is not detected. AWE Plc Registered in England and Wales Registration No 02763902 AWE, Aldermaston, Reading, RG7 4PR > From balay at mcs.anl.gov Thu Nov 7 11:22:41 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Thu, 7 Nov 2013 11:22:41 -0600 (CST) Subject: [petsc-users] --with-hypre-dir did not work for petsc configure in Win-7 In-Reply-To: <1383841546.84925.YahooMailNeo@web160203.mail.bf1.yahoo.com> References: <1383775362.41849.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383783045.43878.YahooMailNeo@web160201.mail.bf1.yahoo.com> <1383837214.32247.YahooMailNeo@web160202.mail.bf1.yahoo.com> <1383839780.27927.YahooMailNeo@web160204.mail.bf1.yahoo.com> <1383841546.84925.YahooMailNeo@web160203.mail.bf1.yahoo.com> Message-ID: On Thu, 7 Nov 2013, Qin Lu wrote: > Satish, > ? > Sorry if I caused any confusions. there is no confusion on my side. > Let me clarify: > ? > 1. I built hypre-2.8.0b using cygwin with Intel compiler icl. It created libHYPRE.a rather than libHYPRE.lib, which I don't know why. I have never said I used gcc, since I mentioned several times that I used "CC=icl CXX=icl" in hypre's configure options. And I've responded saying : that can't be true [because icl would not create libhypre.a - but hypre.lib]. Then you play arround 'with renaming files' insteading of following up on what I say. In a normal conversation - one would have responded by either asserting their statemet by providing proof [by including the build logs for hypre build] - or followed suggested recourse. And I've repatedly pointed out hypre installation instructions [use native-cmake for windows]. And you repatedly try something else and expect it to work [cygwin/configure with CC=icl] In a normal conversation - one would try the suggestions and repsond if it worked or not [and elaborate on why the suggestion was not appropriate - or if/why it didn't work] Satish > ? > 2. I am configuring PETSc-3.4.2 using cygwin with Intel comiler icl and with hypre-2.8.0b, but --with-hypre-lib, etc., did not work. > ? > 3. If it is known that?hypre lib built by cygwin (with icl) does not work for PETSc, please let me know. I will then figure out how to build it with cmake; if this is not the case, at least we should understand why --with-hypre-lib, etc., did not work. > > Thanks, > Qin > > ? > On Thursday, November 7, 2013 10:06 AM, Satish Balay wrote: > Sorry - looks like you are ignoring things I say. > > [I pointed you many times to the hypre installation instructions which > you are ingoring.? And then complaining that things don't work] > > And there is no consistancy in your requirements. > > [you say you want to build hypre with icc - but then you say you've > built it with cygwin - i.e gcc? And then expect it to work with icc > via petsc configure? > > I see no point in replying any further. Its a waste of time.. > > Satish > > > On Thu, 7 Nov 2013, Qin Lu wrote: > > > I?managed to build hypre-2.8.0b with cygwin (instead of cmake) and it worked, but have no idea why it created libhypre.a rather than libhypre.lib. From hypre build's screen output, it did use icl. Please let me know if there is a way to force to create .lib file. > > > > Anyway, PETSc's configure should not care how?hypre was built, right? I tried to rename libhypre.a to libhypre.lib (for testing purpose) but got the same error. > > > > Thank, > > Qin? > > > > On Thursday, November 7, 2013 9:20 AM, Satish Balay wrote: > > > > On Thu, 7 Nov 2013, Qin Lu wrote: > > > > > Satish, > > > ? > > > I tried what you suggested for petsc configure but it still did not work. See the attached configure.log. > > > ? > > > How did you see I was using gcc for hypre? > > > > For cl/icl the library should be libhypre.lib or hypre.lib [not libhypre.a] > > > > > I think I used Intel compiler icl since I specified "CC=icl CXX=icl" for hypre configure. I double-checked hypre's Makefile.config and CC and CXX are icl. > > > > If you follow the instructions I copy/pasted in my previous e-mail regarding 'windows' installation > > - there is no mention of using configure - but of using cmake. [I think you have to install/use native > > windows cmake - not the one from cygwin] > > > > > > Satish > > > > > ? > > > Thanks, > > > Qin > > > > > > > > > > > > On Wednesday, November 6, 2013 6:37 PM, Satish Balay wrote: > > > For one - the? --with-hypre-lib would be: > > > > > > --with-hypre-lib=/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib/libHYPRE.a > > > > > > However - it appears you built hypre with gcc - not icc [which I thought was your goal] > > > > > > If gcc/gfortran would work - you might as well install hypre via petsc configure [--download-hypre] > > > > > > Satish > > > > > > > > > > Satis, > > > > > > > > 1. hypre-2.9 did not build in Win-7 for some reason, so I still use hypre-2.8.0b. > > > > 2. I?used CXXFLAGS="-O2 -MT" for hypre configure. > > > > 3. Tried --with-hypre-include and --with-hypre-lib but still got error: > > > > > > > > ?===============================================================================???????????????????????????????????????????????????????????????????????????????????????????????? TESTING: check from config.libraries(config/BuildSystem/config/libraries.py:145)????????????????????????????????????????????????????????????????????????????????????????????????******************************************************************************* > > > > UNABLE to CONFIGURE with GIVEN OPTIONS????(see configure.log for details): > > > > ------------------------------------------------------------------------------- > > > > --with-hypre-lib=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib'] and > > > > --with-hypre-include=['/cygdrive/c/Lib/hypre-2.8.0b-win64-release/include'] did not work > > > > ******************************************************************************* > > > > > > > > 4. Content of hypre installation: > > > > ? > > > > $ ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release: > > > > include??lib > > > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/include: > > > > Data.h?????????????????????????????????????? HYPRE_krylov.h????????????????HYPRE_struct_mv.h?? _hypre_IJ_mv.h?????? csr_block_matrix.h????par_csr_block_matrix.h > > > > HYPRE.h??????????????????????????????????????HYPRE_lobpcg.h????????????????HYPREf.h????????????_hypre_parcsr_ls.h?? csr_matmultivec.h???? par_csr_matmultivec.h > > > > HYPRE_DistributedMatrixPilutSolver_protos.h??HYPRE_matrix_matrix_protos.h??LLNL_FEI_Fei.h??????_hypre_parcsr_mv.h?? distributed_matrix.h??par_csr_pmvcomm.h > > > > HYPRE_DistributedMatrixPilutSolver_types.h?? HYPRE_parcsr_ls.h???????????? LLNL_FEI_Impl.h???? _hypre_sstruct_ls.h??fei_defs.h????????????par_multivector.h > > > > HYPRE_FEI_includes.h???????????????????????? HYPRE_parcsr_mv.h???????????? LLNL_FEI_LSCore.h?? _hypre_sstruct_mv.h??hypre_cfei.h??????????seq_multivector.h > > > > HYPRE_IJ_mv.h????????????????????????????????HYPRE_seq_mv.h????????????????LLNL_FEI_Matrix.h?? _hypre_struct_ls.h?? interpreter.h???????? seq_mv.h > > > > HYPRE_LinSysCore.h?????????????????????????? HYPRE_sstruct_ls.h????????????LLNL_FEI_Solver.h?? _hypre_struct_mv.h?? krylov.h??????????????temp_multivector.h > > > > HYPRE_MatvecFunctions.h??????????????????????HYPRE_sstruct_mv.h????????????LinearSystemCore.h??cfei-hypre.h???????? lobpcg.h > > > > HYPRE_config.h?????????????????????????????? HYPRE_struct_ls.h???????????? Lookup.h????????????cfei_hypre.h???????? multivector.h > > > > /cygdrive/c/Lib/hypre-2.8.0b-win64-release/lib: > > > > libHYPRE.a > > > > > > > > 5. I used compilation option -O3 for PETSc, but -O2 for hypre, can this be a problem? > > > > ? > > > > Thanks, > > > > Qin > > > > > > > > On Wednesday, November 6, 2013 4:11 PM, Satish Balay wrote: > > > > 1. use latest hypre version - i.e > > > > https://computation.llnl.gov/casc/hypre/download/hypre-2.9.1a.tar.gz > > > > > > > > 2. Make sure its buit with -MT compiler option > > > > > > > > 3. specify hypre with --with-hypre-include and --with-hypre-lib options > > > > > > > > 4. send us the following [so we know whats there - and why configure is not finding it] > > > > > > > > ls -R /cygdrive/c/Lib/hypre-2.8.0b-win64-release > > > > > > > > Satish > > > > > > > > > > > > On Wed, 6 Nov 2013, Qin Lu wrote: > > > > > > > > > Hello, > > > > > ? > > > > > I have built hypre-2.8.0b in Win-7 and have put the installlation under C:\Lib\hypre-2.8.0b-win64-release. Now I?want to include it in my PETSc build but PETSc configure (within cygwin) gave an error message: > > > > > ? > > > > > --with-hypre-dir=/cygdrive/c/Lib/hypre-2.8.0b-win64-release did not work > > > > > ? > > > > > The configure.log is attached. > > > > > ? > > > > > Many thanks for any suggestions, > > > > > ? > > > > > Qin???? > > > > > > > From Lukasz.Kaczmarczyk at glasgow.ac.uk Thu Nov 7 13:01:58 2013 From: Lukasz.Kaczmarczyk at glasgow.ac.uk (Lukasz Kaczmarczyk) Date: Thu, 7 Nov 2013 19:01:58 +0000 Subject: [petsc-users] MatAXPY, creations and destructions In-Reply-To: <08797FF1-4356-47B4-9F77-A44D39729F27@glasgow.ac.uk> References: <08797FF1-4356-47B4-9F77-A44D39729F27@glasgow.ac.uk> Message-ID: <5CE9FBF9-BE45-48AA-83F2-A8D5AD98098A@glasgow.ac.uk> Hello, Just to confirm. More destructions than creations after use MatAXPY it is known bug in petsc? Could you confirm that you have the same problem, or it is something unique to my implementation. Regards, Lukasz On 4 Nov 2013, at 23:22, Lukasz Kaczmarczyk wrote: > Hello All, > > I use MatAXPY (DIFFERENT_NONZERO_PATTERN), with petsc-3.4.3. Works ok, no problem, however in -log_summary as result MatAXPY I have more destructions than creations, each time MatAXPY was used. I have tested this for MATMPIAIJ. > > Valgring don't show anything wrong with that. > > > Kind regards, > Lukasz From juhaj at iki.fi Thu Nov 7 15:35:06 2013 From: juhaj at iki.fi (Juha =?ISO-8859-1?Q?J=E4ykk=E4?=) Date: Thu, 07 Nov 2013 22:35:06 +0100 Subject: [petsc-users] several DAs with different dofs Message-ID: <1846736.A6xhnzay5e@rigel> Hi list! I am thinking of pre-computing and storing some values that are used often during a single TS iteration. I thought I'd create another vector from my DA stick those values there: that would be the cleanest solution. However, the dof of the DA is not big enough to hold all my pre-computed values! Any suggestions how to do this? Unless I miss something obvious, here are my "requirements". I do not need the data across iterations. but I'd rather not realloc it every time either. The data need no ghosts (the data are just local values of some functions of my variables, so any ghosts needed are the ghosts of the variables). I would value a data[][][]-style indexing for simplicity (of code). And finally, I would value the possibility of saving these for inspection under some circumstances in the same format as the main data (i.e. VecView). Cheers, Juha -- ----------------------------------------------- | Juha J?ykk?, juhaj at iki.fi | | http://koti.kapsi.fi/~juhaj/ | ----------------------------------------------- From charlesj at purdue.edu Thu Nov 7 15:38:11 2013 From: charlesj at purdue.edu (James A Charles) Date: Thu, 7 Nov 2013 16:38:11 -0500 (EST) Subject: [petsc-users] QR factorization of Parallel Dense Matrix In-Reply-To: <871u2su59m.fsf@mcs.anl.gov> Message-ID: <1445057878.370962.1383860291715.JavaMail.root@mailhub050.itcs.purdue.edu> I did some more investigation into this problem. Do you think using Scalapack for the QR factorization would be my best bet? If so, How do I interface Petsc to Scalapack? Are the parallelizations of the matrices in Petsc compatible with Scalapack? Thanks, James ----- Original Message ----- From: "Jed Brown" To: "James A Charles" , petsc-users at mcs.anl.gov Sent: Thursday, November 7, 2013 12:53:25 AM Subject: Re: [petsc-users] QR factorization of Parallel Dense Matrix James A Charles writes: > Hello, > > I have a parallel dense matrix, how can I perform a QR factorization? What dimension is the matrix and how do you want the factors distributed? From bsmith at mcs.anl.gov Thu Nov 7 15:59:17 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Thu, 7 Nov 2013 15:59:17 -0600 Subject: [petsc-users] several DAs with different dofs In-Reply-To: <1846736.A6xhnzay5e@rigel> References: <1846736.A6xhnzay5e@rigel> Message-ID: Just create another DA with a different dof argument and then use the DMGetGlobalVector() to get an empty vector and fill it up as needed with the data[][][] style Barry On Nov 7, 2013, at 3:35 PM, Juha J?ykk? wrote: > Hi list! > > I am thinking of pre-computing and storing some values that are used often > during a single TS iteration. I thought I'd create another vector from my DA > stick those values there: that would be the cleanest solution. However, the > dof of the DA is not big enough to hold all my pre-computed values! > > Any suggestions how to do this? > > Unless I miss something obvious, here are my "requirements". I do not need the > data across iterations. but I'd rather not realloc it every time either. The > data need no ghosts (the data are just local values of some functions of my > variables, so any ghosts needed are the ghosts of the variables). I would > value a data[][][]-style indexing for simplicity (of code). And finally, I > would value the possibility of saving these for inspection under some > circumstances in the same format as the main data (i.e. VecView). > > Cheers, > Juha > > -- > ----------------------------------------------- > | Juha J?ykk?, juhaj at iki.fi | > | http://koti.kapsi.fi/~juhaj/ | > ----------------------------------------------- > From juhaj at iki.fi Thu Nov 7 16:10:42 2013 From: juhaj at iki.fi (Juha =?ISO-8859-1?Q?J=E4ykk=E4?=) Date: Thu, 07 Nov 2013 23:10:42 +0100 Subject: [petsc-users] several DAs with different dofs In-Reply-To: References: <1846736.A6xhnzay5e@rigel> Message-ID: <3333088.Qujan1rfMj@rigel> > Just create another DA with a different dof argument and then use the > DMGetGlobalVector() to get an empty vector and fill it up as needed with > the data[][][] style Thanks for the idea! Will that be guaranteed to have the same layout? I would not want to end up with one DA split to 4 MPI ranks along some axis and the other DA split to 3: then I would not have the right data. This also has another disadvantage: I end up using a LOT more memory than I actually need: all the DA structures that get allocated before I create any vectors, would get allocated mostly for nothing because I will never create a local vector out of it and never do any MPI calls with it. Bottom line: if the answer to my question above is "yes", I would still prefer a way which consumes less memory if there is an easy one. Cheers, Juha > > Barry > > On Nov 7, 2013, at 3:35 PM, Juha J?ykk? wrote: > > Hi list! > > > > I am thinking of pre-computing and storing some values that are used often > > during a single TS iteration. I thought I'd create another vector from my > > DA stick those values there: that would be the cleanest solution. > > However, the dof of the DA is not big enough to hold all my pre-computed > > values! > > > > Any suggestions how to do this? > > > > Unless I miss something obvious, here are my "requirements". I do not need > > the data across iterations. but I'd rather not realloc it every time > > either. The data need no ghosts (the data are just local values of some > > functions of my variables, so any ghosts needed are the ghosts of the > > variables). I would value a data[][][]-style indexing for simplicity (of > > code). And finally, I would value the possibility of saving these for > > inspection under some circumstances in the same format as the main data > > (i.e. VecView). > > > > Cheers, > > Juha -- ----------------------------------------------- | Juha J?ykk?, juhaj at iki.fi | | http://koti.kapsi.fi/~juhaj/ | ----------------------------------------------- From jedbrown at mcs.anl.gov Thu Nov 7 16:14:13 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 07 Nov 2013 15:14:13 -0700 Subject: [petsc-users] QR factorization of Parallel Dense Matrix In-Reply-To: <1445057878.370962.1383860291715.JavaMail.root@mailhub050.itcs.purdue.edu> References: <1445057878.370962.1383860291715.JavaMail.root@mailhub050.itcs.purdue.edu> Message-ID: <87zjpfq2q2.fsf@mcs.anl.gov> James A Charles writes: > I did some more investigation into this problem. Do you think using > Scalapack for the QR factorization would be my best bet? If so, How do > I interface Petsc to Scalapack? Are the parallelizations of the > matrices in Petsc compatible with Scalapack? The algorithm you really want here is "TSQR". It works by performing a local QR factorization and then sending the R factors up a tree. An alternative that only needs one MPI_Allreduce and no other communication is "Cholesky QR", which looks like: L * L^T = chol(A^T * A) Q = A * L^{-T} Q and R=L^T are factors of A This is just a few lines of code, though it is less numerically stable than TSQR. TSQR would be a good contribution to PETSc, if you have a little bit of time to implement (not more than a few hours if you are familiar with PETSc). We can provide some guidance, otherwise one of us will get to it eventually. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From bsmith at mcs.anl.gov Thu Nov 7 16:28:18 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Thu, 7 Nov 2013 16:28:18 -0600 Subject: [petsc-users] several DAs with different dofs In-Reply-To: <3333088.Qujan1rfMj@rigel> References: <1846736.A6xhnzay5e@rigel> <3333088.Qujan1rfMj@rigel> Message-ID: On Nov 7, 2013, at 4:10 PM, Juha J?ykk? wrote: >> Just create another DA with a different dof argument and then use the >> DMGetGlobalVector() to get an empty vector and fill it up as needed with >> the data[][][] style > > Thanks for the idea! Will that be guaranteed to have the same layout? Yes, unless you manually control the layout of one of them they will all have the same layout > I would > not want to end up with one DA split to 4 MPI ranks along some axis and the > other DA split to 3: then I would not have the right data. > > This also has another disadvantage: I end up using a LOT more memory than I > actually need: all the DA structures that get allocated before I create any > vectors, would get allocated mostly for nothing because I will never create a > local vector out of it and never do any MPI calls with it. > > Bottom line: if the answer to my question above is "yes", I would still prefer > a way which consumes less memory if there is an easy one. There is no easier way. Just live with the memory usage. > > Cheers, > Juha > >> >> Barry >> >> On Nov 7, 2013, at 3:35 PM, Juha J?ykk? wrote: >>> Hi list! >>> >>> I am thinking of pre-computing and storing some values that are used often >>> during a single TS iteration. I thought I'd create another vector from my >>> DA stick those values there: that would be the cleanest solution. >>> However, the dof of the DA is not big enough to hold all my pre-computed >>> values! >>> >>> Any suggestions how to do this? >>> >>> Unless I miss something obvious, here are my "requirements". I do not need >>> the data across iterations. but I'd rather not realloc it every time >>> either. The data need no ghosts (the data are just local values of some >>> functions of my variables, so any ghosts needed are the ghosts of the >>> variables). I would value a data[][][]-style indexing for simplicity (of >>> code). And finally, I would value the possibility of saving these for >>> inspection under some circumstances in the same format as the main data >>> (i.e. VecView). >>> >>> Cheers, >>> Juha > -- > ----------------------------------------------- > | Juha J?ykk?, juhaj at iki.fi | > | http://koti.kapsi.fi/~juhaj/ | > ----------------------------------------------- > From jedbrown at mcs.anl.gov Thu Nov 7 16:31:46 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 07 Nov 2013 15:31:46 -0700 Subject: [petsc-users] several DAs with different dofs In-Reply-To: References: <1846736.A6xhnzay5e@rigel> <3333088.Qujan1rfMj@rigel> Message-ID: <87r4arq1wt.fsf@mcs.anl.gov> Barry Smith writes: > There is no easier way. Just live with the memory usage. DMDA with stencil width of 0 could use contiguous index sets (ISStride) so that the scatters take up no memory, right? Why should stencil width 0 use any problem-sized memory? -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From gryllida at fastmail.fm Thu Nov 7 17:40:00 2013 From: gryllida at fastmail.fm (Gryllida) Date: Fri, 08 Nov 2013 10:10:00 +1030 Subject: [petsc-users] stensils In-Reply-To: <1383806364.13625.44181269.7618017B@webmail.messagingengine.com> References: <1383806364.13625.44181269.7618017B@webmail.messagingengine.com> Message-ID: <1383867600.8121.44547781.40F3E5A8@webmail.messagingengine.com> On Thu, 7 Nov 2013, at 17:09, Gryllida wrote: > http://courses.engr.illinois.edu/cs554/fa2011/notes/petscpde.pdf has a "this is a simple but inefficient way to set up the matrix" at slide 18. Later they set it as a stensil instead. What is a recommended example of doing the same with OPENMP in FORTRAN, if I have matrices A_central, A_top, A_bottom, A_left, A_right, containing the stensil items? > > Svetlana Sent that to petsc-dev as the stable version doesn't support openmp. Apologies for sending to the wrong list initially. From bsmith at mcs.anl.gov Thu Nov 7 20:46:21 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Thu, 7 Nov 2013 20:46:21 -0600 Subject: [petsc-users] several DAs with different dofs In-Reply-To: <87r4arq1wt.fsf@mcs.anl.gov> References: <1846736.A6xhnzay5e@rigel> <3333088.Qujan1rfMj@rigel> <87r4arq1wt.fsf@mcs.anl.gov> Message-ID: <5FE08A6D-4C7D-4DAA-8971-577044B29261@mcs.anl.gov> On Nov 7, 2013, at 4:31 PM, Jed Brown wrote: > Barry Smith writes: >> There is no easier way. Just live with the memory usage. > > DMDA with stencil width of 0 could use contiguous index sets (ISStride) > so that the scatters take up no memory, right? Why should stencil width > 0 use any problem-sized memory? Maybe it doesn?t. Just give it a try. From anush at bu.edu Fri Nov 8 19:54:19 2013 From: anush at bu.edu (Anush Krishnan) Date: Fri, 8 Nov 2013 20:54:19 -0500 Subject: [petsc-users] Creating a larger matrix from submatrices Message-ID: Hi all, I'm trying to solve the Navier-Stokes equations on a staggered grid. I've created three DA vectors U, V and P to store the values of the x- and y-components of velocity and the pressure. Each of these has different global dimensions. I have assembled matrices Gx and Gy to calculate each component of the gradient of the pressure field: Gx * P = Rx Gy * P = Ry Rx and Ry have the same lengths and distributed structures of U and V. I would like to create matrix G and vector R such that G*P = R where G = /Gx\ \Gy/ and R = /Rx\ \Ry/ and vector R is created from the DMComposite of the DAs of U and V. What would be the best way to go about doing this? Thank you, Anush -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Fri Nov 8 21:22:49 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 08 Nov 2013 20:22:49 -0700 Subject: [petsc-users] Creating a larger matrix from submatrices In-Reply-To: References: Message-ID: <87vc02l0mu.fsf@mcs.anl.gov> Anush Krishnan writes: > Hi all, > > I'm trying to solve the Navier-Stokes equations on a staggered grid. I've > created three DA vectors U, V and P to store the values of the x- and > y-components of velocity and the pressure. Each of these has different > global dimensions. I have assembled matrices Gx and Gy to calculate each > component of the gradient of the pressure field: > > Gx * P = Rx > Gy * P = Ry > > Rx and Ry have the same lengths and distributed structures of U and V. > > I would like to create matrix G and vector R such that > > G*P = R > > where > > G = /Gx\ > \Gy/ > and > > R = /Rx\ > \Ry/ > > and vector R is created from the DMComposite of the DAs of U and V. DMCreateGlobalVector() gives you the vector R. Then use MatCreate() and MatSetSizes() where the local and global row sizes match R and the column sizes match P. For your staggered grid, you should be able to preallocate simply by using MatSeqAIJSetPreallocation() and MatMPIAIJSetPreallocation() with a constant row size of 2 for the "diagonal block" and 1 in the off-diagonal block. (You can also preallocate exactly to be more accurate with memory, but may as well know you are assembling the correct thing first.) You can call DMCompositeGetGlobalISs() to get index sets holding the global indices for the U and V rows of your matrix, or just compute them From the ownership range and the sizes. The column indices will come From the DMDA that holds P. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From jedbrown at mcs.anl.gov Sat Nov 9 01:11:58 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Sat, 09 Nov 2013 00:11:58 -0700 Subject: [petsc-users] MatAXPY, creations and destructions In-Reply-To: <5CE9FBF9-BE45-48AA-83F2-A8D5AD98098A@glasgow.ac.uk> References: <08797FF1-4356-47B4-9F77-A44D39729F27@glasgow.ac.uk> <5CE9FBF9-BE45-48AA-83F2-A8D5AD98098A@glasgow.ac.uk> Message-ID: <87mwlekq0x.fsf@mcs.anl.gov> Lukasz Kaczmarczyk writes: > Hello, > > Just to confirm. More destructions than creations after use MatAXPY it > is known bug in petsc? Could you confirm that you have the same > problem, or it is something unique to my implementation. Thanks for the report. I have fixed the problem in 'next' and will merge it into 'maint' in a day or two. commit 03c4025afd7bfddd354430f069f83de621f6646b (jed/fix-mat-header-replace-log) Author: Jed Brown Date: Sat Nov 9 00:03:28 2013 -0700 MatHeaderReplace: fix duplicate call to PetscLogObjectDestroy PetscLogObjectDestroy includes tallying of creations and destructions, but MatHeaderReplace already calls PetscHeaderDestroy_Private, which calls PetscLogObjectDestroy itself. This led to tallies with more destructions than creations for operations when calling functions that use MatHeaderReplace (like MatAXPY and MatAYPX). Reported-by: Lukasz Kaczmarczyk diff --git a/src/mat/utils/gcreate.c b/src/mat/utils/gcreate.c index b00fcaa..cbe1f72 100644 --- a/src/mat/utils/gcreate.c +++ b/src/mat/utils/gcreate.c @@ -363,7 +363,6 @@ PETSC_EXTERN PetscErrorCode MatHeaderReplace(Mat A,Mat C) ((PetscObject)A)->refct = refct; - ierr = PetscLogObjectDestroy((PetscObject)C);CHKERRQ(ierr); ierr = PetscFree(C);CHKERRQ(ierr); PetscFunctionReturn(0); } -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From anush at bu.edu Sat Nov 9 20:02:05 2013 From: anush at bu.edu (Anush Krishnan) Date: Sat, 9 Nov 2013 21:02:05 -0500 Subject: [petsc-users] Creating a larger matrix from submatrices In-Reply-To: <87vc02l0mu.fsf@mcs.anl.gov> References: <87vc02l0mu.fsf@mcs.anl.gov> Message-ID: On 8 November 2013 22:22, Jed Brown wrote: > > I would like to create matrix G and vector R such that > > > > G*P = R > > > > where > > > > G = /Gx\ > > \Gy/ > > and > > > > R = /Rx\ > > \Ry/ > > > > and vector R is created from the DMComposite of the DAs of U and V. > > DMCreateGlobalVector() gives you the vector R. Then use MatCreate() and > MatSetSizes() where the local and global row sizes match R and the > column sizes match P. For your staggered grid, you should be able to > preallocate simply by using MatSeqAIJSetPreallocation() and > MatMPIAIJSetPreallocation() with a constant row size of 2 for the > "diagonal block" and 1 in the off-diagonal block. (You can also > preallocate exactly to be more accurate with memory, but may as well > know you are assembling the correct thing first.) > > You can call DMCompositeGetGlobalISs() to get index sets holding the > global indices for the U and V rows of your matrix, or just compute them > From the ownership range and the sizes. The column indices will come > From the DMDA that holds P. > Hi Jed, Thank you very much! That worked. I also have a couple of general questions regarding this: 1. In this case, I followed your advice and created a single matrix G of type MPIAIJ and assembled it. How do I decide if I should rather use Block Matrices (MPIBAIJ) or MatGetLocalSubMatrix to set it up? 2. Various functions exist to help us access the correct indices when required (e.g. DMDAGetAO, DMCompositeGetGlobalISs, DMDAGetCorners, VecGetOwnershipRange, etc.) And sometimes, there are multiple ways by which we can obtain the same indices. In such cases, are some of these functions preferable to others (either because they are faster or need to allocate less memory)? Thank you, Anush -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Sat Nov 9 20:15:12 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Sat, 09 Nov 2013 19:15:12 -0700 Subject: [petsc-users] Creating a larger matrix from submatrices In-Reply-To: References: <87vc02l0mu.fsf@mcs.anl.gov> Message-ID: <87k3ghj93j.fsf@jedbrown.org> Anush Krishnan writes: > Thank you very much! That worked. Great! > I also have a couple of general questions regarding this: > > 1. In this case, I followed your advice and created a single matrix G of > type MPIAIJ and assembled it. How do I decide if I should rather use Block > Matrices (MPIBAIJ) or MatGetLocalSubMatrix to set it up? BAIJ only does square blocks and you don't have block structurex to begin with (because you don't interlace the x- and y-components of the fields). MatGetLocalSubMatrix is a convenience when it makes sense to assembled blocks on their own. That probably doesn't apply here because you don't have a reason to assemble, say, the gradient of pressure in the y-direction only. > 2. Various functions exist to help us access the correct indices when > required (e.g. DMDAGetAO, DMCompositeGetGlobalISs, DMDAGetCorners, > VecGetOwnershipRange, etc.) And sometimes, there are multiple ways by which > we can obtain the same indices. In such cases, are some of these functions > preferable to others (either because they are faster or need to allocate > less memory)? Many uses of AO are fairly heavy-weight because any global-to-local operation either needs non-scalable memory or a lookup structure that is slower than an array. DMDAGetCorners() is extremely cheap and doesn't use much memory, so if that is all you need, use that. DMCompositeGetGlobalISs() is cheap on its own, but accessing the resulting ISs often boils down to an array of indices, which would mildly increase your memory bandwidth demands if you do it in a performance-sensitive traversal. Usually these things are only needed during setup, at which point they are cheap as long as they are scalable. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From anush at bu.edu Sat Nov 9 20:27:54 2013 From: anush at bu.edu (Anush Krishnan) Date: Sat, 9 Nov 2013 21:27:54 -0500 Subject: [petsc-users] Creating a larger matrix from submatrices In-Reply-To: <87k3ghj93j.fsf@jedbrown.org> References: <87vc02l0mu.fsf@mcs.anl.gov> <87k3ghj93j.fsf@jedbrown.org> Message-ID: Thanks, Jed. That was helpful. On 9 November 2013 21:15, Jed Brown wrote: > Anush Krishnan writes: > > Thank you very much! That worked. > > Great! > > > I also have a couple of general questions regarding this: > > > > 1. In this case, I followed your advice and created a single matrix G of > > type MPIAIJ and assembled it. How do I decide if I should rather use > Block > > Matrices (MPIBAIJ) or MatGetLocalSubMatrix to set it up? > > BAIJ only does square blocks and you don't have block structurex to > begin with (because you don't interlace the x- and y-components of the > fields). > > MatGetLocalSubMatrix is a convenience when it makes sense to assembled > blocks on their own. That probably doesn't apply here because you don't > have a reason to assemble, say, the gradient of pressure in the > y-direction only. > > > 2. Various functions exist to help us access the correct indices when > > required (e.g. DMDAGetAO, DMCompositeGetGlobalISs, DMDAGetCorners, > > VecGetOwnershipRange, etc.) And sometimes, there are multiple ways by > which > > we can obtain the same indices. In such cases, are some of these > functions > > preferable to others (either because they are faster or need to allocate > > less memory)? > > Many uses of AO are fairly heavy-weight because any global-to-local > operation either needs non-scalable memory or a lookup structure that is > slower than an array. > > DMDAGetCorners() is extremely cheap and doesn't use much memory, so if > that is all you need, use that. DMCompositeGetGlobalISs() is cheap on > its own, but accessing the resulting ISs often boils down to an array of > indices, which would mildly increase your memory bandwidth demands if > you do it in a performance-sensitive traversal. Usually these things > are only needed during setup, at which point they are cheap as long as > they are scalable. > -------------- next part -------------- An HTML attachment was scrubbed... URL: From pengxwang at hotmail.com Sun Nov 10 12:34:29 2013 From: pengxwang at hotmail.com (Roc Wang) Date: Sun, 10 Nov 2013 12:34:29 -0600 Subject: [petsc-users] approaches to reduce computing time Message-ID: Hi all, I am trying to minimize the computing time to solve a large sparse matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to reduce the computing time from two directions: 1 finding a Pre-conditioner to reduce the number of iterations which reduces the time numerically, 2 requesting more cores. ----For the first method, I tried several methods: 1 default KSP and PC, 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type jacobi, 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm (PCASM) The iterations and timing is like the following with 128 cores requested: case# iter timing (s) 1 1436 816 2 3 12658 3 1069 669.64 4 872 768.12 5 927 513.14 It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment can help to reduce the iterations but not the timing (comparing case 3 and 4). Second, the PCASM helps a lot. Although the second option is able to reduce iterations, the timing increases very much. Is it because more operations are needed in the PC? My questions here are: 1. Which direction should I take to select -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger restart with large augment is better or larger restart with smaller augment is better? ----For the second method, I tried with -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different number of cores. I found the speedup ratio increases slowly when more than 32 to 64 cores are requested. I searched the milling list archives and found that I am very likely running into the memory bandwidth bottleneck. http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: # of cores iter timing 1 923 19541.83 4 929 5897.06 8 932 4854.72 16 924 1494.33 32 924 1480.88 64 928 686.89 128 927 627.33 256 926 552.93 My question here is: Is there any other PC can help on both reducing iterations and increasing scalability? Thanks. -------------- next part -------------- An HTML attachment was scrubbed... URL: From ahmeed.salawu at syria-web.biz Sat Nov 9 08:51:36 2013 From: ahmeed.salawu at syria-web.biz (Ahmed Salawu) Date: Sat, 9 Nov 2013 14:51:36 +0000 Subject: [petsc-users] Proposal Message-ID: <2532235928952123127520@MAIL7> Please read the attached business proposal and get back to me at your earliest convenience with your thoughts and reaction. Email me on: ahmed.salawu at syria-web.biz so that we can go over the details together. Regards. Ahmed. -------------- next part -------------- A non-text attachment was scrubbed... Name: Proposal.jpg Type: image/jpeg Size: 653273 bytes Desc: not available URL: From jedbrown at mcs.anl.gov Sun Nov 10 13:20:18 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Sun, 10 Nov 2013 12:20:18 -0700 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: References: Message-ID: <87zjpcgj2l.fsf@jedbrown.org> Roc Wang writes: > Hi all, > > I am trying to minimize the computing time to solve a large sparse matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to reduce the computing time from two directions: 1 finding a Pre-conditioner to reduce the number of iterations which reduces the time numerically, 2 requesting more cores. > > ----For the first method, I tried several methods: > 1 default KSP and PC, > 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type jacobi, > 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, > 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, > 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm (PCASM) > > The iterations and timing is like the following with 128 cores requested: > case# iter timing (s) > 1 1436 816 > 2 3 12658 > 3 1069 669.64 > 4 872 768.12 > 5 927 513.14 > > It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment can help to reduce the iterations but not the timing (comparing case 3 and 4). Second, the PCASM helps a lot. Although the second option is able to reduce iterations, the timing increases very much. Is it because more operations are needed in the PC? > > My questions here are: 1. Which direction should I take to select > -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger > restart with large augment is better or larger restart with smaller > augment is better? Look at the -log_summary. By increasing the restart, the work in KSPGMRESOrthog will increase linearly, but the number of iterations might decrease enough to compensate. There is no general rule here since it depends on the relative expense of operations for your problem on your machine. > ----For the second method, I tried with -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different number of cores. I found the speedup ratio increases slowly when more than 32 to 64 cores are requested. I searched the milling list archives and found that I am very likely running into the memory bandwidth bottleneck. http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: > > # of cores iter timing > 1 923 19541.83 > 4 929 5897.06 > 8 932 4854.72 > 16 924 1494.33 > 32 924 1480.88 > 64 928 686.89 > 128 927 627.33 > 256 926 552.93 The bandwidth issue has more to do with using multiple cores within a node rather than between nodes. Likely the above is a load balancing problem or bad communication. > My question here is: Is there any other PC can help on both reducing iterations and increasing scalability? Thanks. Always send -log_summary with questions like this, but algebraic multigrid is a good place to start. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From s.kramer at imperial.ac.uk Mon Nov 11 05:33:53 2013 From: s.kramer at imperial.ac.uk (Stephan Kramer) Date: Mon, 11 Nov 2013 22:33:53 +1100 Subject: [petsc-users] Fwd: MatZeroRowsColumns for mpibaij In-Reply-To: References: <5271BD65.1010507@imperial.ac.uk> Message-ID: <5280C0A1.9090304@imperial.ac.uk> On 01/11/13 06:01, Matthew Knepley wrote: > On Wed, Oct 30, 2013 at 9:19 PM, Rhodri Davies > wrote: > > Dear all, > > Would you mind giving us an update on your thoughts re: this question? We're very keen to be able to use such functionally and it would be good to know if you're looking into this / have any > suggestions? > > > I looked at the code. It does not look hard to make a MPIBAIJ version. Since the scatter context for BAIJ is already exploded, most > of the MPIAIJ code goes straight through, it just a matter of handling the blocking in a few places. > > Here is what would be really helpful for us. Write a nice test. The code would probably only take an hour of being careful. However, > the test assembly would be more. If you give us a nice test example for it, writing it could happen pretty quickly. > > Thanks, > > Matt That would be great. I had a look at the existing tests and found src/mat/examples/tests/ex140.c which already claims to test MatZeroRowsColumns with MPIBAIJ - except it isn't being run in parallel. Would running this test in parallel be a sufficient test to work with? It currently fails with the expected message "No support for this operation for this object type!". I'd be happy to write another test but it would basically be creating another matrix and testing it does the right thing with the matrix and the rhs. Let me know if there's anything else we could help out with Cheers Stephan >> From: Stephan Kramer >> Subject: Fwd: [petsc-users] MatZeroRowsColumns for mpibaij >> Date: 31 October 2013 13:16:05 GMT+11:00 >> To: Rhodri Davies >> >> >> >> >> -------- Original Message -------- >> Subject: [petsc-users] MatZeroRowsColumns for mpibaij >> Date: Wed, 19 Jun 2013 13:22:31 +0100 >> From: Stephan Kramer >> To: PETSc users list >> >> Dear all, >> >> We have found the MatZeroRowsColumns() routine to be very useful for lifting boundary conditions. Unfortunately we also found that it isn't implemented for matrix type mpibaij and we're quite keen on >> keeping the baij structure as it seems to give us significant savings for DG problems (in particular the blocked sor). I had a quick look at the implementation of MatZeroRowsColums_MPIAIJ but couldn't >> really estimate how hard it would be to do something similar for mpibaij (an implementation for seqbaij exists already). Is this something that's likely to be looked at in the future? >> >> Stephan From a.vergottis at ucl.ac.uk Mon Nov 11 06:10:59 2013 From: a.vergottis at ucl.ac.uk (Anthony Vergottis) Date: Mon, 11 Nov 2013 12:10:59 +0000 Subject: [petsc-users] Parallel DM example for FEM Message-ID: Dear All, I am trying to find a good example within PETSc that examples with some detail the use of DM objects within a parallel and FEM framework. I have looked at most of the example and have not found anything like this. Is there any other material available on the subject from any other source or maybe I missed something within the PETSc full download file? Thanks in advance. Regards, Adonis -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Mon Nov 11 06:19:41 2013 From: knepley at gmail.com (Matthew Knepley) Date: Mon, 11 Nov 2013 06:19:41 -0600 Subject: [petsc-users] Parallel DM example for FEM In-Reply-To: References: Message-ID: On Mon, Nov 11, 2013 at 6:10 AM, Anthony Vergottis wrote: > Dear All, > > I am trying to find a good example within PETSc that examples with some > detail the use of DM objects within a parallel and FEM framework. I have > looked at most of the example and have not found anything like this. Is > there any other material available on the subject from any other source or > maybe I missed something within the PETSc full download file? > Did you look at SNES ex12? Is it not understandable? Thanks, Matt > Thanks in advance. > > Regards, > Adonis > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From a.vergottis at ucl.ac.uk Mon Nov 11 06:29:21 2013 From: a.vergottis at ucl.ac.uk (Anthony Vergottis) Date: Mon, 11 Nov 2013 12:29:21 +0000 Subject: [petsc-users] Parallel DM example for FEM In-Reply-To: References: Message-ID: I compile ex12 but when I run the program I get this in the terminal. adoni at Adoni:~/Desktop/petsc-lat/src/snes/examples/tutorials$ make ex12 /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/bin/mpicxx -o ex12.o -c -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g -fPIC -I/home/adoni/Desktop/petsc-lat/include -I/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/include /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/bin/mpicxx -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g -o ex12 ex12.o -Wl,-rpath,/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib -L/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib -lpetsc -Wl,-rpath,/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib -lflapack -lfblas -ltriangle -lX11 -lparmetis -lmetis -lpthread -lchaco -lctetgen -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lmpichf90 -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl /bin/rm -f ex12.o adoni at Adoni:~/Desktop/petsc-lat/src/snes/examples/tutorials$ mpiexec -np 2 ./ex12 [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Invalid argument! [0]PETSC ERROR: The section cell closure size 0 != dual space dimension 1! [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Petsc Development GIT revision: unknown GIT Date: unknown [0]PETSC ERROR: See docs/changes/index.html for recent updates. [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [0]PETSC ERROR: See docs/index.html for manual pages. [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: ./ex12 on a arch-linux2-c-debug named Adoni by adoni Mon Nov 11 12:21:49 2013 [0]PETSC ERROR: Libraries linked from /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib [0]PETSC ERROR: Configure run at Fri Nov 8 20:03:26 2013 [0]PETSC ERROR: Configure options --with-clanguage=c++ --with-shared-libraries=1 --download-f-blas-lapack --download-mpich --download-boost --download-triangle --download-ctetgen --download-chaco --download-metis --download-parmetis [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: DMPlexProjectFunctionLocal() line 207 in src/dm/impls/plex/plexfem.c [0]PETSC ERROR: DMPlexProjectFunction() line 258 in src/dm/impls/plex/plexfem.c [0]PETSC ERROR: main() line 686 in /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Invalid argument! [0]PETSC ERROR: The section cell closure size 0 != dual space dimension 1! [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Petsc Development GIT revision: unknown GIT Date: unknown [0]PETSC ERROR: See docs/changes/index.html for recent updates. [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [0]PETSC ERROR: See docs/index.html for manual pages. [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: ./ex12 on a arch-linux2-c-debug named Adoni by adoni Mon Nov 11 12:21:49 2013 [0]PETSC ERROR: Libraries linked from /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib [0]PETSC ERROR: Configure run at Fri Nov 8 20:03:26 2013 [0]PETSC ERROR: Configure options --with-clanguage=c++ --with-shared-libraries=1 --download-f-blas-lapack --download-mpich --download-boost --download-triangle --download-ctetgen --download-chaco --download-metis --download-parmetis [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: DMPlexProjectFunctionLocal() line 207 in src/dm/impls/plex/plexfem.c [0]PETSC ERROR: DMPlexProjectFunction() line 258 in src/dm/impls/plex/plexfem.c [0]PETSC ERROR: main() line 686 in /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 [unset]: aborting job: application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 [unset]: aborting job: application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 This occurs for all process counts. Thanks. Adoni On 11 November 2013 12:19, Matthew Knepley wrote: > On Mon, Nov 11, 2013 at 6:10 AM, Anthony Vergottis wrote: > >> Dear All, >> >> I am trying to find a good example within PETSc that examples with some >> detail the use of DM objects within a parallel and FEM framework. I have >> looked at most of the example and have not found anything like this. Is >> there any other material available on the subject from any other source or >> maybe I missed something within the PETSc full download file? >> > > Did you look at SNES ex12? Is it not understandable? > > Thanks, > > Matt > > >> Thanks in advance. >> >> Regards, >> Adonis >> > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Mon Nov 11 06:31:14 2013 From: knepley at gmail.com (Matthew Knepley) Date: Mon, 11 Nov 2013 06:31:14 -0600 Subject: [petsc-users] Fwd: MatZeroRowsColumns for mpibaij In-Reply-To: <5280C0A1.9090304@imperial.ac.uk> References: <5271BD65.1010507@imperial.ac.uk> <5280C0A1.9090304@imperial.ac.uk> Message-ID: On Mon, Nov 11, 2013 at 5:33 AM, Stephan Kramer wrote: > On 01/11/13 06:01, Matthew Knepley wrote: > > On Wed, Oct 30, 2013 at 9:19 PM, Rhodri Davies > rhodri.davies at anu.edu.au>> wrote: >> >> Dear all, >> >> Would you mind giving us an update on your thoughts re: this >> question? We're very keen to be able to use such functionally and it would >> be good to know if you're looking into this / have any >> suggestions? >> >> >> I looked at the code. It does not look hard to make a MPIBAIJ version. >> Since the scatter context for BAIJ is already exploded, most >> of the MPIAIJ code goes straight through, it just a matter of handling >> the blocking in a few places. >> >> Here is what would be really helpful for us. Write a nice test. The code >> would probably only take an hour of being careful. However, >> the test assembly would be more. If you give us a nice test example for >> it, writing it could happen pretty quickly. >> >> Thanks, >> >> Matt >> > > That would be great. I had a look at the existing tests and found > src/mat/examples/tests/ex140.c which already claims to test > MatZeroRowsColumns with MPIBAIJ - except it isn't being run in parallel. > Would running this test in parallel be a sufficient test to work with? It > currently fails with the expected message "No support for this operation > for this object type!". I'd be happy to write another test but it would > basically be creating another matrix and testing it does the right thing > with the matrix and the rhs. Let me know if there's anything else we could > help out with > As far as I can tell ex140 does not really work. I am using ex12 to check my rewrite of MatZeroRows(). I would like a companion to that which uses RowsColumns(). And the work here is to provide the output to check against. Even for ex140 I would have to do that, which means more time here. Matt > Cheers > Stephan > > > From: Stephan Kramer >>> Subject: Fwd: [petsc-users] MatZeroRowsColumns for mpibaij >>> Date: 31 October 2013 13:16:05 GMT+11:00 >>> To: Rhodri Davies >>> >>> >>> >>> >>> -------- Original Message -------- >>> Subject: [petsc-users] MatZeroRowsColumns for mpibaij >>> Date: Wed, 19 Jun 2013 13:22:31 +0100 >>> From: Stephan Kramer >>> To: PETSc users list >>> >>> Dear all, >>> >>> We have found the MatZeroRowsColumns() routine to be very useful for >>> lifting boundary conditions. Unfortunately we also found that it isn't >>> implemented for matrix type mpibaij and we're quite keen on >>> keeping the baij structure as it seems to give us significant savings >>> for DG problems (in particular the blocked sor). I had a quick look at the >>> implementation of MatZeroRowsColums_MPIAIJ but couldn't >>> really estimate how hard it would be to do something similar for mpibaij >>> (an implementation for seqbaij exists already). Is this something that's >>> likely to be looked at in the future? >>> >>> Stephan >>> >> -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From a.vergottis at ucl.ac.uk Mon Nov 11 06:47:38 2013 From: a.vergottis at ucl.ac.uk (Anthony Vergottis) Date: Mon, 11 Nov 2013 12:47:38 +0000 Subject: [petsc-users] Parallel DM example for FEM In-Reply-To: References: Message-ID: I would also like to ask as I have some confusion. Are the DM objects capable of handling all aspects of the parallel computations within a given problem? Let me elaborate - for example if I am passing a mesh structure (triangular 3 node elements) as the following: Elem 0: N1 N5 N6 etc..... where N# is the node tag. into a DM object and then perform a decomposition for as many processes I required. When the mesh partitions are distributed to their processes do I then have the ability by using DM objects or something else built into PETSc to compute specific calculations on each process without the need to hardcode any MPI? I hope I have explained this well. Thanks again. Adoni On 11 November 2013 12:33, Anthony Vergottis wrote: > Thanks mate that worked! > > Thanks. > Adoni > > > > On 11 November 2013 12:29, Matthew Knepley wrote: > >> On Mon, Nov 11, 2013 at 6:25 AM, Anthony Vergottis > > wrote: >> >>> I compile ex12 but when I run the program I get this in the terminal. >>> >> >> I guess I should make better defaults. Here are the runs I use for ex12: >> >> >> https://bitbucket.org/petsc/petsc/src/a965ca046084fa53248a41da989a0a62cb6266ea/config/builder.py?at=master#cl-180 >> >> You can see that the problem here is that the -petscspace_order has not >> been set. So to test P_1 I use >> >> -run_type test -refinement_limit 0.0 -bc_type dirichlet -interpolate 0 >> -petscspace_order 1 -show_initial -dm_plex_print_fem 1 >> >> and then take away run_type (or use full) to solve the problem. SNES ex62 >> is the Stokes problem, and run parameters are >> in the same place. >> >> Thanks, >> >> Matt >> >> >>> adoni at Adoni:~/Desktop/petsc-lat/src/snes/examples/tutorials$ make ex12 >>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/bin/mpicxx -o ex12.o >>> -c -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g >>> -fPIC -I/home/adoni/Desktop/petsc-lat/include >>> -I/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/include >>> /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c >>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/bin/mpicxx -Wall >>> -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g -o ex12 >>> ex12.o -Wl,-rpath,/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib >>> -L/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib -lpetsc >>> -Wl,-rpath,/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib -lflapack >>> -lfblas -ltriangle -lX11 -lparmetis -lmetis -lpthread -lchaco -lctetgen -lm >>> -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 >>> -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu >>> -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu >>> -L/lib/x86_64-linux-gnu -lmpichf90 -lgfortran -lm -lgfortran -lm -lquadmath >>> -lm -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s >>> -ldl >>> /bin/rm -f ex12.o >>> adoni at Adoni:~/Desktop/petsc-lat/src/snes/examples/tutorials$ mpiexec >>> -np 2 ./ex12 >>> [0]PETSC ERROR: --------------------- Error Message >>> ------------------------------------ >>> [0]PETSC ERROR: Invalid argument! >>> [0]PETSC ERROR: The section cell closure size 0 != dual space dimension >>> 1! >>> [0]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [0]PETSC ERROR: Petsc Development GIT revision: unknown GIT Date: >>> unknown >>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>> [0]PETSC ERROR: See docs/index.html for manual pages. >>> [0]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [0]PETSC ERROR: ./ex12 on a arch-linux2-c-debug named Adoni by adoni Mon >>> Nov 11 12:21:49 2013 >>> [0]PETSC ERROR: Libraries linked from >>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib >>> [0]PETSC ERROR: Configure run at Fri Nov 8 20:03:26 2013 >>> [0]PETSC ERROR: Configure options --with-clanguage=c++ >>> --with-shared-libraries=1 --download-f-blas-lapack --download-mpich >>> --download-boost --download-triangle --download-ctetgen --download-chaco >>> --download-metis --download-parmetis >>> [0]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [0]PETSC ERROR: DMPlexProjectFunctionLocal() line 207 in >>> src/dm/impls/plex/plexfem.c >>> [0]PETSC ERROR: DMPlexProjectFunction() line 258 in >>> src/dm/impls/plex/plexfem.c >>> [0]PETSC ERROR: main() line 686 in >>> /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c >>> [0]PETSC ERROR: --------------------- Error Message >>> ------------------------------------ >>> [0]PETSC ERROR: Invalid argument! >>> [0]PETSC ERROR: The section cell closure size 0 != dual space dimension >>> 1! >>> [0]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [0]PETSC ERROR: Petsc Development GIT revision: unknown GIT Date: >>> unknown >>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>> [0]PETSC ERROR: See docs/index.html for manual pages. >>> [0]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [0]PETSC ERROR: ./ex12 on a arch-linux2-c-debug named Adoni by adoni Mon >>> Nov 11 12:21:49 2013 >>> [0]PETSC ERROR: Libraries linked from >>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib >>> [0]PETSC ERROR: Configure run at Fri Nov 8 20:03:26 2013 >>> [0]PETSC ERROR: Configure options --with-clanguage=c++ >>> --with-shared-libraries=1 --download-f-blas-lapack --download-mpich >>> --download-boost --download-triangle --download-ctetgen --download-chaco >>> --download-metis --download-parmetis >>> [0]PETSC ERROR: >>> ------------------------------------------------------------------------ >>> [0]PETSC ERROR: DMPlexProjectFunctionLocal() line 207 in >>> src/dm/impls/plex/plexfem.c >>> [0]PETSC ERROR: DMPlexProjectFunction() line 258 in >>> src/dm/impls/plex/plexfem.c >>> [0]PETSC ERROR: main() line 686 in >>> /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c >>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>> [unset]: aborting job: >>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>> [unset]: aborting job: >>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>> >>> This occurs for all process counts. >>> >>> Thanks. >>> >>> Adoni >>> >>> >>> On 11 November 2013 12:19, Matthew Knepley wrote: >>> >>>> On Mon, Nov 11, 2013 at 6:10 AM, Anthony Vergottis < >>>> a.vergottis at ucl.ac.uk> wrote: >>>> >>>>> Dear All, >>>>> >>>>> I am trying to find a good example within PETSc that examples with >>>>> some detail the use of DM objects within a parallel and FEM framework. I >>>>> have looked at most of the example and have not found anything like this. >>>>> Is there any other material available on the subject from any other source >>>>> or maybe I missed something within the PETSc full download file? >>>>> >>>> >>>> Did you look at SNES ex12? Is it not understandable? >>>> >>>> Thanks, >>>> >>>> Matt >>>> >>>> >>>>> Thanks in advance. >>>>> >>>>> Regards, >>>>> Adonis >>>>> >>>> >>>> >>>> >>>> -- >>>> What most experimenters take for granted before they begin their >>>> experiments is infinitely more interesting than any results to which their >>>> experiments lead. >>>> -- Norbert Wiener >>>> >>> >>> >> >> >> -- >> What most experimenters take for granted before they begin their >> experiments is infinitely more interesting than any results to which their >> experiments lead. >> -- Norbert Wiener >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Mon Nov 11 06:49:54 2013 From: knepley at gmail.com (Matthew Knepley) Date: Mon, 11 Nov 2013 06:49:54 -0600 Subject: [petsc-users] Parallel DM example for FEM In-Reply-To: References: Message-ID: On Mon, Nov 11, 2013 at 6:47 AM, Anthony Vergottis wrote: > I would also like to ask as I have some confusion. Are the DM objects > capable of handling all aspects of the parallel computations within a given > problem? > > Let me elaborate - for example if I am passing a mesh structure > (triangular 3 node elements) as the following: > > Elem 0: N1 N5 N6 > etc..... > where N# is the node tag. > > into a DM object and then perform a decomposition for as many processes I > required. When the mesh partitions are distributed to their processes do I > then have the ability by using DM objects or something else built into > PETSc to compute specific calculations on each process without the need to > hardcode any MPI? > I believe the answer is yes. I could be more specific with a concrete example. For instance, the DM can calculate a parallel residual without the user ever seeing MPI. Matt > I hope I have explained this well. > > Thanks again. > Adoni > > > On 11 November 2013 12:33, Anthony Vergottis wrote: > >> Thanks mate that worked! >> >> Thanks. >> Adoni >> >> >> >> On 11 November 2013 12:29, Matthew Knepley wrote: >> >>> On Mon, Nov 11, 2013 at 6:25 AM, Anthony Vergottis < >>> a.vergottis at ucl.ac.uk> wrote: >>> >>>> I compile ex12 but when I run the program I get this in the terminal. >>>> >>> >>> I guess I should make better defaults. Here are the runs I use for ex12: >>> >>> >>> https://bitbucket.org/petsc/petsc/src/a965ca046084fa53248a41da989a0a62cb6266ea/config/builder.py?at=master#cl-180 >>> >>> You can see that the problem here is that the -petscspace_order has not >>> been set. So to test P_1 I use >>> >>> -run_type test -refinement_limit 0.0 -bc_type dirichlet -interpolate >>> 0 -petscspace_order 1 -show_initial -dm_plex_print_fem 1 >>> >>> and then take away run_type (or use full) to solve the problem. SNES >>> ex62 is the Stokes problem, and run parameters are >>> in the same place. >>> >>> Thanks, >>> >>> Matt >>> >>> >>>> adoni at Adoni:~/Desktop/petsc-lat/src/snes/examples/tutorials$ make ex12 >>>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/bin/mpicxx -o ex12.o >>>> -c -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g >>>> -fPIC -I/home/adoni/Desktop/petsc-lat/include >>>> -I/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/include >>>> /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c >>>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/bin/mpicxx -Wall >>>> -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g -o ex12 >>>> ex12.o -Wl,-rpath,/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib >>>> -L/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib -lpetsc >>>> -Wl,-rpath,/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib -lflapack >>>> -lfblas -ltriangle -lX11 -lparmetis -lmetis -lpthread -lchaco -lctetgen -lm >>>> -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 >>>> -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu >>>> -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu >>>> -L/lib/x86_64-linux-gnu -lmpichf90 -lgfortran -lm -lgfortran -lm -lquadmath >>>> -lm -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s >>>> -ldl >>>> /bin/rm -f ex12.o >>>> adoni at Adoni:~/Desktop/petsc-lat/src/snes/examples/tutorials$ mpiexec >>>> -np 2 ./ex12 >>>> [0]PETSC ERROR: --------------------- Error Message >>>> ------------------------------------ >>>> [0]PETSC ERROR: Invalid argument! >>>> [0]PETSC ERROR: The section cell closure size 0 != dual space dimension >>>> 1! >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: Petsc Development GIT revision: unknown GIT Date: >>>> unknown >>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: ./ex12 on a arch-linux2-c-debug named Adoni by adoni >>>> Mon Nov 11 12:21:49 2013 >>>> [0]PETSC ERROR: Libraries linked from >>>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib >>>> [0]PETSC ERROR: Configure run at Fri Nov 8 20:03:26 2013 >>>> [0]PETSC ERROR: Configure options --with-clanguage=c++ >>>> --with-shared-libraries=1 --download-f-blas-lapack --download-mpich >>>> --download-boost --download-triangle --download-ctetgen --download-chaco >>>> --download-metis --download-parmetis >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: DMPlexProjectFunctionLocal() line 207 in >>>> src/dm/impls/plex/plexfem.c >>>> [0]PETSC ERROR: DMPlexProjectFunction() line 258 in >>>> src/dm/impls/plex/plexfem.c >>>> [0]PETSC ERROR: main() line 686 in >>>> /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c >>>> [0]PETSC ERROR: --------------------- Error Message >>>> ------------------------------------ >>>> [0]PETSC ERROR: Invalid argument! >>>> [0]PETSC ERROR: The section cell closure size 0 != dual space dimension >>>> 1! >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: Petsc Development GIT revision: unknown GIT Date: >>>> unknown >>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: ./ex12 on a arch-linux2-c-debug named Adoni by adoni >>>> Mon Nov 11 12:21:49 2013 >>>> [0]PETSC ERROR: Libraries linked from >>>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib >>>> [0]PETSC ERROR: Configure run at Fri Nov 8 20:03:26 2013 >>>> [0]PETSC ERROR: Configure options --with-clanguage=c++ >>>> --with-shared-libraries=1 --download-f-blas-lapack --download-mpich >>>> --download-boost --download-triangle --download-ctetgen --download-chaco >>>> --download-metis --download-parmetis >>>> [0]PETSC ERROR: >>>> ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: DMPlexProjectFunctionLocal() line 207 in >>>> src/dm/impls/plex/plexfem.c >>>> [0]PETSC ERROR: DMPlexProjectFunction() line 258 in >>>> src/dm/impls/plex/plexfem.c >>>> [0]PETSC ERROR: main() line 686 in >>>> /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c >>>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>>> [unset]: aborting job: >>>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>>> [unset]: aborting job: >>>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>>> >>>> This occurs for all process counts. >>>> >>>> Thanks. >>>> >>>> Adoni >>>> >>>> >>>> On 11 November 2013 12:19, Matthew Knepley wrote: >>>> >>>>> On Mon, Nov 11, 2013 at 6:10 AM, Anthony Vergottis < >>>>> a.vergottis at ucl.ac.uk> wrote: >>>>> >>>>>> Dear All, >>>>>> >>>>>> I am trying to find a good example within PETSc that examples with >>>>>> some detail the use of DM objects within a parallel and FEM framework. I >>>>>> have looked at most of the example and have not found anything like this. >>>>>> Is there any other material available on the subject from any other source >>>>>> or maybe I missed something within the PETSc full download file? >>>>>> >>>>> >>>>> Did you look at SNES ex12? Is it not understandable? >>>>> >>>>> Thanks, >>>>> >>>>> Matt >>>>> >>>>> >>>>>> Thanks in advance. >>>>>> >>>>>> Regards, >>>>>> Adonis >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> What most experimenters take for granted before they begin their >>>>> experiments is infinitely more interesting than any results to which their >>>>> experiments lead. >>>>> -- Norbert Wiener >>>>> >>>> >>>> >>> >>> >>> -- >>> What most experimenters take for granted before they begin their >>> experiments is infinitely more interesting than any results to which their >>> experiments lead. >>> -- Norbert Wiener >>> >> >> > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From a.vergottis at ucl.ac.uk Mon Nov 11 06:52:17 2013 From: a.vergottis at ucl.ac.uk (Anthony Vergottis) Date: Mon, 11 Nov 2013 12:52:17 +0000 Subject: [petsc-users] Parallel DM example for FEM In-Reply-To: References: Message-ID: Ok thank you very much yet again! Adoni On 11 November 2013 12:49, Matthew Knepley wrote: > On Mon, Nov 11, 2013 at 6:47 AM, Anthony Vergottis wrote: > >> I would also like to ask as I have some confusion. Are the DM objects >> capable of handling all aspects of the parallel computations within a given >> problem? >> >> Let me elaborate - for example if I am passing a mesh structure >> (triangular 3 node elements) as the following: >> >> Elem 0: N1 N5 N6 >> etc..... >> where N# is the node tag. >> >> into a DM object and then perform a decomposition for as many processes I >> required. When the mesh partitions are distributed to their processes do I >> then have the ability by using DM objects or something else built into >> PETSc to compute specific calculations on each process without the need to >> hardcode any MPI? >> > > I believe the answer is yes. I could be more specific with a concrete > example. For instance, the DM can > calculate a parallel residual without the user ever seeing MPI. > > Matt > > >> I hope I have explained this well. >> >> Thanks again. >> Adoni >> >> >> On 11 November 2013 12:33, Anthony Vergottis wrote: >> >>> Thanks mate that worked! >>> >>> Thanks. >>> Adoni >>> >>> >>> >>> On 11 November 2013 12:29, Matthew Knepley wrote: >>> >>>> On Mon, Nov 11, 2013 at 6:25 AM, Anthony Vergottis < >>>> a.vergottis at ucl.ac.uk> wrote: >>>> >>>>> I compile ex12 but when I run the program I get this in the terminal. >>>>> >>>> >>>> I guess I should make better defaults. Here are the runs I use for ex12: >>>> >>>> >>>> https://bitbucket.org/petsc/petsc/src/a965ca046084fa53248a41da989a0a62cb6266ea/config/builder.py?at=master#cl-180 >>>> >>>> You can see that the problem here is that the -petscspace_order has not >>>> been set. So to test P_1 I use >>>> >>>> -run_type test -refinement_limit 0.0 -bc_type dirichlet -interpolate >>>> 0 -petscspace_order 1 -show_initial -dm_plex_print_fem 1 >>>> >>>> and then take away run_type (or use full) to solve the problem. SNES >>>> ex62 is the Stokes problem, and run parameters are >>>> in the same place. >>>> >>>> Thanks, >>>> >>>> Matt >>>> >>>> >>>>> adoni at Adoni:~/Desktop/petsc-lat/src/snes/examples/tutorials$ make ex12 >>>>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/bin/mpicxx -o ex12.o >>>>> -c -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g >>>>> -fPIC -I/home/adoni/Desktop/petsc-lat/include >>>>> -I/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/include >>>>> /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c >>>>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/bin/mpicxx -Wall >>>>> -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g -o ex12 >>>>> ex12.o -Wl,-rpath,/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib >>>>> -L/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib -lpetsc >>>>> -Wl,-rpath,/home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib -lflapack >>>>> -lfblas -ltriangle -lX11 -lparmetis -lmetis -lpthread -lchaco -lctetgen -lm >>>>> -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 >>>>> -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu >>>>> -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu >>>>> -L/lib/x86_64-linux-gnu -lmpichf90 -lgfortran -lm -lgfortran -lm -lquadmath >>>>> -lm -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s >>>>> -ldl >>>>> /bin/rm -f ex12.o >>>>> adoni at Adoni:~/Desktop/petsc-lat/src/snes/examples/tutorials$ mpiexec >>>>> -np 2 ./ex12 >>>>> [0]PETSC ERROR: --------------------- Error Message >>>>> ------------------------------------ >>>>> [0]PETSC ERROR: Invalid argument! >>>>> [0]PETSC ERROR: The section cell closure size 0 != dual space >>>>> dimension 1! >>>>> [0]PETSC ERROR: >>>>> ------------------------------------------------------------------------ >>>>> [0]PETSC ERROR: Petsc Development GIT revision: unknown GIT Date: >>>>> unknown >>>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>>> [0]PETSC ERROR: >>>>> ------------------------------------------------------------------------ >>>>> [0]PETSC ERROR: ./ex12 on a arch-linux2-c-debug named Adoni by adoni >>>>> Mon Nov 11 12:21:49 2013 >>>>> [0]PETSC ERROR: Libraries linked from >>>>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib >>>>> [0]PETSC ERROR: Configure run at Fri Nov 8 20:03:26 2013 >>>>> [0]PETSC ERROR: Configure options --with-clanguage=c++ >>>>> --with-shared-libraries=1 --download-f-blas-lapack --download-mpich >>>>> --download-boost --download-triangle --download-ctetgen --download-chaco >>>>> --download-metis --download-parmetis >>>>> [0]PETSC ERROR: >>>>> ------------------------------------------------------------------------ >>>>> [0]PETSC ERROR: DMPlexProjectFunctionLocal() line 207 in >>>>> src/dm/impls/plex/plexfem.c >>>>> [0]PETSC ERROR: DMPlexProjectFunction() line 258 in >>>>> src/dm/impls/plex/plexfem.c >>>>> [0]PETSC ERROR: main() line 686 in >>>>> /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c >>>>> [0]PETSC ERROR: --------------------- Error Message >>>>> ------------------------------------ >>>>> [0]PETSC ERROR: Invalid argument! >>>>> [0]PETSC ERROR: The section cell closure size 0 != dual space >>>>> dimension 1! >>>>> [0]PETSC ERROR: >>>>> ------------------------------------------------------------------------ >>>>> [0]PETSC ERROR: Petsc Development GIT revision: unknown GIT Date: >>>>> unknown >>>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>>> [0]PETSC ERROR: >>>>> ------------------------------------------------------------------------ >>>>> [0]PETSC ERROR: ./ex12 on a arch-linux2-c-debug named Adoni by adoni >>>>> Mon Nov 11 12:21:49 2013 >>>>> [0]PETSC ERROR: Libraries linked from >>>>> /home/adoni/Desktop/petsc-lat/arch-linux2-c-debug/lib >>>>> [0]PETSC ERROR: Configure run at Fri Nov 8 20:03:26 2013 >>>>> [0]PETSC ERROR: Configure options --with-clanguage=c++ >>>>> --with-shared-libraries=1 --download-f-blas-lapack --download-mpich >>>>> --download-boost --download-triangle --download-ctetgen --download-chaco >>>>> --download-metis --download-parmetis >>>>> [0]PETSC ERROR: >>>>> ------------------------------------------------------------------------ >>>>> [0]PETSC ERROR: DMPlexProjectFunctionLocal() line 207 in >>>>> src/dm/impls/plex/plexfem.c >>>>> [0]PETSC ERROR: DMPlexProjectFunction() line 258 in >>>>> src/dm/impls/plex/plexfem.c >>>>> [0]PETSC ERROR: main() line 686 in >>>>> /home/adoni/Desktop/petsc-lat/src/snes/examples/tutorials/ex12.c >>>>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>>>> [unset]: aborting job: >>>>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>>>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>>>> [unset]: aborting job: >>>>> application called MPI_Abort(MPI_COMM_WORLD, 62) - process 0 >>>>> >>>>> This occurs for all process counts. >>>>> >>>>> Thanks. >>>>> >>>>> Adoni >>>>> >>>>> >>>>> On 11 November 2013 12:19, Matthew Knepley wrote: >>>>> >>>>>> On Mon, Nov 11, 2013 at 6:10 AM, Anthony Vergottis < >>>>>> a.vergottis at ucl.ac.uk> wrote: >>>>>> >>>>>>> Dear All, >>>>>>> >>>>>>> I am trying to find a good example within PETSc that examples with >>>>>>> some detail the use of DM objects within a parallel and FEM framework. I >>>>>>> have looked at most of the example and have not found anything like this. >>>>>>> Is there any other material available on the subject from any other source >>>>>>> or maybe I missed something within the PETSc full download file? >>>>>>> >>>>>> >>>>>> Did you look at SNES ex12? Is it not understandable? >>>>>> >>>>>> Thanks, >>>>>> >>>>>> Matt >>>>>> >>>>>> >>>>>>> Thanks in advance. >>>>>>> >>>>>>> Regards, >>>>>>> Adonis >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> What most experimenters take for granted before they begin their >>>>>> experiments is infinitely more interesting than any results to which their >>>>>> experiments lead. >>>>>> -- Norbert Wiener >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> What most experimenters take for granted before they begin their >>>> experiments is infinitely more interesting than any results to which their >>>> experiments lead. >>>> -- Norbert Wiener >>>> >>> >>> >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From davidblogueur at gmail.com Mon Nov 11 11:22:26 2013 From: davidblogueur at gmail.com (David Laroche) Date: Mon, 11 Nov 2013 18:22:26 +0100 Subject: [petsc-users] [DL] Mon guide sur la confiance en soi GRATUIT aujourd'hui Message-ID: <4389289489271e3509a12cee6bd8102e@smtp3.ymlpserver.net> -------------------------------------------------------------------------------- Cette newsletter vous a ?t? envoy?e au format graphique HTML. Si vous lisez cette version, alors votre logiciel de messagerie pr?f?re les e-mails au format texte. Vous pouvez lire la version originale en ligne: http://ymlp229.net/z7QJaO -------------------------------------------------------------------------------- Bonjour, Je viens de terminer ?le petit guide de la confiance en soi? et j?aimerais vous le partager aujourd'hui et vous en faire profiter. Ce sont 40 pages de conseils et strat?gies pour acc?l?rer les r?sultats, gagner confiance en soi, ?tre plus ? l'aise avec les autres et lib?rer son potentiel. Ce livre se base sur mes 9 derni?res ann?es de travail sur le domaine de la confiance en soi pendant lesquelles j'ai pu rencontrer personnellement John Gray (Les hommes viennent de Mars, les femmes de V?nus), Tony Robbins, Brian Tracy, Darren Hardy, Seth Godin... de grands entrepreneurs Fran?ais et Am?ricains, sportifs olympiques, com?diens... ps : je suis un ancien timide, aujourd'hui coach et conf?rencier sur le domaine de la confiance Cliquez ici pour t?l?charger le guide ( http://lefacteurcle.com/guideconfiancegratuit ) Je suis ? votre disposition si vous souhaitez de nouveaux ?l?ments. David _____________________________ Unsubscribe / Change Profile: http://ymlp229.net/ugbhewuugsgeshwygeggjeuhhb Powered par YourMailingListProvider -------------- next part -------------- An HTML attachment was scrubbed... URL: From a.vergottis at ucl.ac.uk Mon Nov 11 11:46:46 2013 From: a.vergottis at ucl.ac.uk (Anthony Vergottis) Date: Mon, 11 Nov 2013 17:46:46 +0000 Subject: [petsc-users] Triangle geometry data as input to DM object to generate a mesh Message-ID: Dear All, I would like some assistance on how a mesh is created using Triangle with the DM objects. Specifically how does one input the geometry of the computational domain into PETSc/DM objects to then generate a 2D triangular mesh? Is there a function that can do such a thing? Currently I have hard coded all mesh I/O but I would like to make things a bit cleaner if its possible. Thanks in advance for any help. Best regards, Anthony -------------- next part -------------- An HTML attachment was scrubbed... URL: From qiyuelu1 at gmail.com Mon Nov 11 11:55:24 2013 From: qiyuelu1 at gmail.com (Qiyue Lu) Date: Mon, 11 Nov 2013 11:55:24 -0600 Subject: [petsc-users] -ksp_rtol problem Message-ID: Dear All: I am working on a linear A*x=b system. When I use *-ksp_rtol 1.0e-5 *option in the command line, I am not very sure how the relative tolerance is defined. In my understanding: At the beginning, r= b-A*x_i, so rtol = ||r|| / ||b|| with 2-norm has a value. with the calculation, rtol gets smaller, when the rtol_new / rtol_original < 1.0e-5, the calculation will stop. Is my understanding correct? If so, how to decide the rtol at the first iteration? If initial solution values are all zeros, then rtol should be 1.0. In my output file, it's a value related to the matrix and can be at 1.0e-4 scale. Is there any initial guess solution from preconditioner used there? Could you help to clarify this definition? Thanks Qiyue Lu -------------- next part -------------- An HTML attachment was scrubbed... URL: From qiyuelu1 at gmail.com Mon Nov 11 12:03:24 2013 From: qiyuelu1 at gmail.com (Qiyue Lu) Date: Mon, 11 Nov 2013 12:03:24 -0600 Subject: [petsc-users] -ksp_rtol problem In-Reply-To: References: Message-ID: Sorry, I made a mistake in privous email. in my output file, the *KSP residual norm* not the rtol is 1.0e-4 at the 1st iteration. My question is how this residual norm calculated? Is there any initial guess there? thanks On Mon, Nov 11, 2013 at 11:55 AM, Qiyue Lu wrote: > Dear All: > > I am working on a linear A*x=b system. > When I use *-ksp_rtol 1.0e-5 *option in the command line, I am not very > sure how the relative tolerance is defined. > > In my understanding: > > At the beginning, r= b-A*x_i, so rtol = ||r|| / ||b|| with 2-norm has a > value. with the calculation, rtol gets smaller, when the rtol_new / > rtol_original < 1.0e-5, the calculation will stop. > > Is my understanding correct? > > If so, how to decide the rtol at the first iteration? If initial solution > values are all zeros, then rtol should be 1.0. In my output file, it's a > value related to the matrix and can be at 1.0e-4 scale. Is there any > initial guess solution from preconditioner used there? > > Could you help to clarify this definition? Thanks > > > Qiyue Lu > -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Mon Nov 11 12:21:47 2013 From: knepley at gmail.com (Matthew Knepley) Date: Mon, 11 Nov 2013 12:21:47 -0600 Subject: [petsc-users] Triangle geometry data as input to DM object to generate a mesh In-Reply-To: References: Message-ID: On Mon, Nov 11, 2013 at 11:46 AM, Anthony Vergottis wrote: > Dear All, > > I would like some assistance on how a mesh is created using Triangle with > the DM objects. Specifically how does one input the geometry of the > computational domain into PETSc/DM objects to then generate a 2D triangular > mesh? Is there a function that can do such a thing? Currently I have hard > coded all mesh I/O but I would like to make things a bit cleaner if its > possible. > The default 2D mesh generator in PETSc is Triangle. If you have a 1D boundary, you can generate the 2D mesh using DMPlexGenerate(boundaryDM, NULL, PETSC_FALSE, &dm) or PETSC_TRUE if you want edges included. I suggest using DMPlexCreateFromDAG() if you have programmatically constructed your boundary, or DMPlexCreateFromCellList() if you get your boundary from a meshing tool. Thanks, Matt > Thanks in advance for any help. > > Best regards, > Anthony > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Mon Nov 11 12:23:27 2013 From: knepley at gmail.com (Matthew Knepley) Date: Mon, 11 Nov 2013 12:23:27 -0600 Subject: [petsc-users] -ksp_rtol problem In-Reply-To: References: Message-ID: On Mon, Nov 11, 2013 at 12:03 PM, Qiyue Lu wrote: > Sorry, I made a mistake in privous email. > in my output file, the *KSP residual norm* not the rtol is 1.0e-4 at the > 1st iteration. My question is how this residual norm calculated? Is there > any initial guess there? thanks > The residual r for some approximate solution x is r = b - Ax The initial residual uses the initial guess, often 0. Matt > > > On Mon, Nov 11, 2013 at 11:55 AM, Qiyue Lu wrote: > >> Dear All: >> >> I am working on a linear A*x=b system. >> When I use *-ksp_rtol 1.0e-5 *option in the command line, I am not very >> sure how the relative tolerance is defined. >> >> In my understanding: >> >> At the beginning, r= b-A*x_i, so rtol = ||r|| / ||b|| with 2-norm has a >> value. with the calculation, rtol gets smaller, when the rtol_new / >> rtol_original < 1.0e-5, the calculation will stop. >> >> Is my understanding correct? >> >> If so, how to decide the rtol at the first iteration? If initial solution >> values are all zeros, then rtol should be 1.0. In my output file, it's a >> value related to the matrix and can be at 1.0e-4 scale. Is there any >> initial guess solution from preconditioner used there? >> >> Could you help to clarify this definition? Thanks >> >> >> Qiyue Lu >> > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From a.vergottis at ucl.ac.uk Mon Nov 11 12:35:40 2013 From: a.vergottis at ucl.ac.uk (Anthony Vergottis) Date: Mon, 11 Nov 2013 18:35:40 +0000 Subject: [petsc-users] Triangle geometry data as input to DM object to generate a mesh In-Reply-To: References: Message-ID: Thank you very much. Best regards, Anthony On 11 November 2013 18:21, Matthew Knepley wrote: > On Mon, Nov 11, 2013 at 11:46 AM, Anthony Vergottis > wrote: > >> Dear All, >> >> I would like some assistance on how a mesh is created using Triangle with >> the DM objects. Specifically how does one input the geometry of the >> computational domain into PETSc/DM objects to then generate a 2D triangular >> mesh? Is there a function that can do such a thing? Currently I have hard >> coded all mesh I/O but I would like to make things a bit cleaner if its >> possible. >> > > The default 2D mesh generator in PETSc is Triangle. If you have a 1D > boundary, you can generate the 2D mesh using > > DMPlexGenerate(boundaryDM, NULL, PETSC_FALSE, &dm) > > or PETSC_TRUE if you want edges included. I suggest using > DMPlexCreateFromDAG() if you have programmatically > constructed your boundary, or DMPlexCreateFromCellList() if you get your > boundary from a meshing tool. > > Thanks, > > Matt > > >> Thanks in advance for any help. >> >> Best regards, >> Anthony >> > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From bsmith at mcs.anl.gov Mon Nov 11 12:41:08 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Mon, 11 Nov 2013 12:41:08 -0600 Subject: [petsc-users] -ksp_rtol problem In-Reply-To: References: Message-ID: <528AE660-5878-4843-A079-1C9F3AD9DC95@mcs.anl.gov> On Nov 11, 2013, at 12:23 PM, Matthew Knepley wrote: > On Mon, Nov 11, 2013 at 12:03 PM, Qiyue Lu wrote: > Sorry, I made a mistake in privous email. > in my output file, the KSP residual norm not the rtol is 1.0e-4 at the 1st iteration. My question is how this residual norm calculated? Is there any initial guess there? thanks > > The residual r for some approximate solution x is > > r = b - Ax > > The initial residual uses the initial guess, often 0. Note that for left preconditioning, like the default GMRES in PETSc, the residual norm that is used is the preconditioned residual norm Br = B(b - Ax) which can be scaled differently from the unpreconditioned residual norm. See KPSSetPCSide(). You can run with -ksp_monitor_true_residual to see also the unpreconditioned residual. Barry > > Matt > > > > On Mon, Nov 11, 2013 at 11:55 AM, Qiyue Lu wrote: > Dear All: > > I am working on a linear A*x=b system. > When I use -ksp_rtol 1.0e-5 option in the command line, I am not very sure how the relative tolerance is defined. > > In my understanding: > > At the beginning, r= b-A*x_i, so rtol = ||r|| / ||b|| with 2-norm has a value. with the calculation, rtol gets smaller, when the rtol_new / rtol_original < 1.0e-5, the calculation will stop. > > Is my understanding correct? > > If so, how to decide the rtol at the first iteration? If initial solution values are all zeros, then rtol should be 1.0. In my output file, it's a value related to the matrix and can be at 1.0e-4 scale. Is there any initial guess solution from preconditioner used there? > > Could you help to clarify this definition? Thanks > > > Qiyue Lu > > > > > -- > What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. > -- Norbert Wiener From bisheshkh at gmail.com Mon Nov 11 13:08:45 2013 From: bisheshkh at gmail.com (Bishesh Khanal) Date: Mon, 11 Nov 2013 20:08:45 +0100 Subject: [petsc-users] using petsc tools to solve isolated irregular domains with finite difference In-Reply-To: References: Message-ID: On Mon, Nov 4, 2013 at 5:25 PM, Bishesh Khanal wrote: > > > > On Mon, Nov 4, 2013 at 4:54 PM, Matthew Knepley wrote: > >> On Mon, Nov 4, 2013 at 4:49 AM, Bishesh Khanal wrote: >> >>> On Fri, Nov 1, 2013 at 1:30 PM, Matthew Knepley wrote: >>> >>>> On Fri, Nov 1, 2013 at 12:08 AM, Bishesh Khanal wrote: >>>> >>>>> On Thu, Oct 31, 2013 at 11:34 PM, Matthew Knepley wrote: >>>>> >>>>>> On Thu, Oct 31, 2013 at 5:02 PM, Bishesh Khanal wrote: >>>>>> >>>>>>> On Thu, Oct 31, 2013 at 7:43 PM, Matthew Knepley wrote: >>>>>>> >>>>>>>> On Thu, Oct 31, 2013 at 1:25 PM, Bishesh Khanal < >>>>>>>> bisheshkh at gmail.com> wrote: >>>>>>>> >>>>>>>>> I did not call DMCreateMatrix() before when using just dmda (and >>>>>>>>> it is working). In any case, I added a call to DMCreateMatrix() now but >>>>>>>>> still there are problems. Here are the code snippets and the error I get: >>>>>>>>> >>>>>>>> >>>>>>>> Then you were allowing it to be called automatically by PETSc, and >>>>>>>> it would have been this time as well. >>>>>>>> >>>>>>>> >>>>>>>>> //Setting up the section and linking it to DM: >>>>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, >>>>>>>>> &nC);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>>>> PetscInt point; >>>>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>>>> &point);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>>>> } >>>>>>>>> } >>>>>>>>> } >>>>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>>>>>>> >>>>>>>>> //Set up KSP: >>>>>>>>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>>>>>>>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>>>>>>>> ierr = >>>>>>>>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>>>>> ierr = >>>>>>>>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>>>>> ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); >>>>>>>>> ierr = KSPSolve(ksp,NULL,NULL);CHKERRQ(ierr); >>>>>>>>> >>>>>>>>> ------------------------------------------------------ >>>>>>>>> The computeMatrix2dSection function has: >>>>>>>>> >>>>>>>>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>>>>>>>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>>>>>>>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>> if (isPosInDomain(ctx,i,j)) { >>>>>>>>> ierr = >>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>>>>>> ierr = PetscSectionGetDof(gs,point,&dof); >>>>>>>>> for(PetscInt cDof = 0; cDof < dof; ++cDof) { >>>>>>>>> num = 0; >>>>>>>>> row+=cDof; >>>>>>>>> col[num] = row; //(i,j) position >>>>>>>>> v[num++] = -4; >>>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>>> ierr = >>>>>>>>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>>>>>>>> ierr = >>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>> col[num] += cDof; >>>>>>>>> v[num++] = 1; >>>>>>>>> } >>>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>>> ierr = >>>>>>>>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>>>>>>>> ierr = >>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>> col[num] += cDof; >>>>>>>>> v[num++] = 1; >>>>>>>>> } >>>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>>> ierr = >>>>>>>>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>>>>>>>> ierr = >>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>> col[num] += cDof; >>>>>>>>> v[num++] = 1; >>>>>>>>> } >>>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>>> ierr = >>>>>>>>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>>>>>>>> ierr = >>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>> col[num] += cDof; >>>>>>>>> v[num++] = 1; >>>>>>>>> } >>>>>>>>> ierr = >>>>>>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>> } >>>>>>>>> } >>>>>>>>> } >>>>>>>>> } >>>>>>>>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>>>>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>>>>> >>>>>>>>> But this is not working. For e.g. for a 6X6 grid size I get the >>>>>>>>> following error: >>>>>>>>> >>>>>>>> >>>>>>>> Okay, here is how to debug this: >>>>>>>> >>>>>>>> 0) Go to a single scalar equations to make things easier >>>>>>>> >>>>>>>> 1) Use MatView(A, PETSC_VIEWER_STDOUT_WORLD), which should output >>>>>>>> a 36 row matrix with >>>>>>>> the preallocated nonzero pattern, filled with zeros >>>>>>>> >>>>>>>> 2) Make sure this is the pattern you want >>>>>>>> >>>>>>>> 3) Run in the debugger with -start_in_debugger >>>>>>>> >>>>>>>> 4) When you get the error, see >>>>>>>> >>>>>>>> a) If the (i, j) is one that should be setting a value >>>>>>>> >>>>>>>> b) Why this (i, j) was not preallocated >>>>>>>> >>>>>>> >>>>>>> Up to 4 (a), it is correct. There is a problem in the way >>>>>>> DMDAGetCellPoint and PetscSectionGetOffset works in this case scenario. I >>>>>>> will try to explain it below for the case of 4X4 grid. >>>>>>> First case: >>>>>>> If I set the computational domain to be all the points of the dmda >>>>>>> grid, (i.e. isPosInDomain(..,i,j,..) returns true for all i and j in the >>>>>>> dmda grid), the program runs fine and does not give any error. >>>>>>> >>>>>>> Second case: >>>>>>> I want the computational domain to be some part of the whole grid. >>>>>>> There is a problem in this case. >>>>>>> The following test is in a case where, >>>>>>> isPosInDomain(..,i,j,..) returns true ONLY for (i,j) pairs of (2,1) >>>>>>> (1,2) (2,2) (3,2) and (3,2). The grid with its corresponding point number >>>>>>> in the petscsection is shown below: >>>>>>> >>>>>>> 12 13 14 15 >>>>>>> 8 9 10 11 >>>>>>> 4 5 6 7 >>>>>>> 0 1 2 3 >>>>>>> >>>>>>> of which 6, 9, 10, 11 and 14 correspond to the (i,j) points that >>>>>>> returns true for isPosInDomain(..,i,j,..) >>>>>>> MatView gives me the 16-row matrix with the star stencil non-zero >>>>>>> structure as expected. >>>>>>> The error I get is: new non-zero at (0,2) caused a malloc. >>>>>>> >>>>>>> This error is for the value of (i,j) = (2,1), i.e. point 6. >>>>>>> The loop to set values in the matrix starts as: >>>>>>> for(PetscInt j = info.ys; j>>>>>> for(PetscInt i = info.xs; i>>>>>> if (isPosInDomain(ctx,i,j)) { >>>>>>> ierr = >>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>>>> >>>>>>> Now, what I wanted from the beginning is to create the matrix >>>>>>> containing the rows corresponding to only those points (i,j) which has true >>>>>>> values for isPosInDomain(ctx,i,j), that means in this case 5 row matrix. In >>>>>>> the current test, the first (i,j) that reaches DMDAGetCellPoint is (2,1), >>>>>>> which gives sets point variable to 6. >>>>>>> The line PetscSectionGetOffset(gs,point,&row) sets the value of row >>>>>>> to 0. >>>>>>> So I think this where the inconsistency lies. >>>>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES) expects row variable >>>>>>> and col[] array to correspond to (i,j) based on DMDA, but >>>>>>> PetsSectionGetOffset gives result based on how we masked away some of the >>>>>>> points from dmda grid. Or am I missing something very basic here ? >>>>>>> >>>>>> >>>>>> You are missing something basic. The row is correctly identified as >>>>>> 0, and that means the 2 is >>>>>> point 10. >>>>>> >>>>> The problem must be preallocation. First, you should send the result >>>>>> of MatView(). >>>>>> >>>>> >>>>> MatView shows that the preallocation and the matrix created is for the >>>>> full dmda grid which is probably not what we want here, right ? Here is the >>>>> matview result: >>>>> >>>> >>>> Thats right. Did you call >>>> >>>> DMSetDefaultSection(da, s) >>>> >>>> after you made the new section and before you called DMGetMatrix()? >>>> >>> Do you mean before I called DMCreateMatrix() ? I called >>> DMSetDefaultSection(da,s) as below: >>> >>> ... >>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>> // ierr = DMGetMatrix(dm,&A);CHKERRQ(ierr); //It seems DMGetMatrix >>> is changed to DMCreateMatrix(). >>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>> ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >>> >> >> Yes, this is correct. I do exactly this in the example code and it works >> fine. I suspect >> you are not creating the section you think you are. You can send the >> code if you want >> and I will look at it. >> > > Thanks, here is the code: > https://github.com/bishesh/petscPoissonIrregular/blob/master/poissonIrregular.cxx > It works for 2D and 3D without using PetscSection. (The choice of using > Petsc_section is on line 68). Using PetsSection currently it implements > only the 2D case with dof = 1. It works when the domain consists of all the > cells of the dmda, it works (to check you can uncomment line 129 in > isPosInDomain(). But it does not when I mask out some of the cells while > creating the petscsection. > Hi Matt, I'm afraid if my last email somehow escaped you. Could you please let me know what could be possibly going wrong with my code linked above ? > > >> >> Matt >> >> >>> This MatView on A shows (result below) that it still has 16 rows for all >>> DMDA points instead of only those petscsection points for which I set a >>> non-zero dof. Is it what it should not be doing ? >>> >>>> Mat Object: 1 MPI processes >>>>> >>>>> type: seqaij >>>>> >>>>> row 0: (0, 0) (1, 0) (4, 0) >>>>> >>>>> row 1: (0, 0) (1, 0) (2, 0) (5, 0) >>>>> >>>>> row 2: (1, 0) (2, 0) (3, 0) (6, 0) >>>>> >>>>> row 3: (2, 0) (3, 0) (7, 0) >>>>> >>>>> row 4: (0, 0) (4, 0) (5, 0) (8, 0) >>>>> >>>>> row 5: (1, 0) (4, 0) (5, 0) (6, 0) (9, 0) >>>>> >>>>> row 6: (2, 0) (5, 0) (6, 0) (7, 0) (10, 0) >>>>> >>>>> row 7: (3, 0) (6, 0) (7, 0) (11, 0) >>>>> >>>>> row 8: (4, 0) (8, 0) (9, 0) (12, 0) >>>>> >>>>> row 9: (5, 0) (8, 0) (9, 0) (10, 0) (13, 0) >>>>> >>>>> row 10: (6, 0) (9, 0) (10, 0) (11, 0) (14, 0) >>>>> >>>>> row 11: (7, 0) (10, 0) (11, 0) (15, 0) >>>>> >>>>> row 12: (8, 0) (12, 0) (13, 0) >>>>> >>>>> row 13: (9, 0) (12, 0) (13, 0) (14, 0) >>>>> >>>>> row 14: (10, 0) (13, 0) (14, 0) (15, 0) >>>>> >>>>> row 15: (11, 0) (14, 0) (15, 0) >>>>> >>>>> [0]PETSC ERROR: --------------------- Error Message >>>>> ------------------------------------ >>>>> >>>>> [0]PETSC ERROR: Argument out of range! >>>>> >>>>> [0]PETSC ERROR: New nonzero at (0,2) caused a malloc! >>>>> >>>>> >>>>> >>> For reference here is the code snippet again: >>> ... >>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>> PetscInt point; >>> if(isPosInDomain(&testPoisson,i,j,0)) { >>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>> &point);CHKERRQ(ierr); >>> ierr = PetscSectionSetDof(s, point, testPoisson.mDof); >>> // No. of dofs associated with the point. >>> } >>> } >>> } >>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>> ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >>> >>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>> >>> ierr = >>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>> ierr = >>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>> >>> .... >>> The function computeMatrix2dSection: >>> >>> #undef __FUNCT__ >>> #define __FUNCT__ "computeMatrix2dSection" >>> PetscErrorCode computeMatrix2dSection(KSP ksp, Mat A, Mat B, >>> MatStructure *str, void *context) { >>> PoissonModel *ctx = (PoissonModel*)context; >>> PetscErrorCode ierr; >>> DM da; >>> DMDALocalInfo info; >>> PetscInt row, col[5]; >>> PetscInt dof; >>> PetscScalar v[5]; //array to store 5 point stencil for >>> each row >>> PetscInt num; //non-zero position in the current row >>> // PetscScalar dScale = 1; //to scale Dirichlet >>> identity rows if needed >>> PetscSection gs; //Global Section that keeps >>> the grid info and indices >>> PetscInt point; //Current point of the >>> petscSection >>> >>> PetscFunctionBeginUser; >>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>> ierr = DMCreateMatrix(da,&A);CHKERRQ(ierr); >>> // ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >>> for(PetscInt j = info.ys; j>> for(PetscInt i = info.xs; i>> if (isPosInDomain(ctx,i,j)) { >>> ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>> ierr = PetscSectionGetOffset(gs,point,&row); >>> ierr = PetscSectionGetDof(gs,point,&dof); >>> for(PetscInt cDof = 0; cDof < ctx->mDof; ++cDof) { >>> num = 0; >>> row+=cDof; >>> col[num] = row; //(i,j) position >>> v[num++] = -4; >>> if(isPosInDomain(ctx,i+1,j)) { >>> ierr = >>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>> ierr = PetscSectionGetOffset(gs,point,&col[num]); >>> col[num] += cDof; >>> v[num++] = 1; >>> } >>> if(isPosInDomain(ctx,i-1,j)) { >>> ierr = >>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>> ierr = PetscSectionGetOffset(gs,point,&col[num]); >>> col[num] += cDof; >>> v[num++] = 1; >>> } >>> if(isPosInDomain(ctx,i,j+1)) { >>> ierr = >>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>> ierr = PetscSectionGetOffset(gs,point,&col[num]); >>> col[num] += cDof; >>> v[num++] = 1; >>> } >>> if(isPosInDomain(ctx,i,j-1)) { >>> ierr = >>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>> ierr = PetscSectionGetOffset(gs,point,&col[num]); >>> col[num] += cDof; >>> v[num++] = 1; >>> } >>> ierr = >>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>> } >>> } >>> } >>> } >>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>> >>> >>> >>> >>>> >>>>>> Matt >>>>>> >>>>>> >>>>>>> >>>>>>>> Thanks, >>>>>>>> >>>>>>>> Matt >>>>>>>> >>>>>>>> >>>>>>>>> [0]PETSC ERROR: --------------------- Error Message >>>>>>>>> ------------------------------------ >>>>>>>>> [0]PETSC ERROR: Argument out of range! >>>>>>>>> [0]PETSC ERROR: New nonzero at (0,6) caused a malloc! >>>>>>>>> [0]PETSC ERROR: >>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>> [0]PETSC ERROR: Petsc Development GIT revision: >>>>>>>>> 61e5e40bb2c5bf2423e94b71a15fef47e411b0da GIT Date: 2013-10-25 21:47:45 >>>>>>>>> -0500 >>>>>>>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>>>>>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>>>>>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>>>>>>> [0]PETSC ERROR: >>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>> [0]PETSC ERROR: build/poissonIrregular on a arch-linux2-cxx-debug >>>>>>>>> named edwards by bkhanal Thu Oct 31 19:23:58 2013 >>>>>>>>> [0]PETSC ERROR: Libraries linked from >>>>>>>>> /home/bkhanal/Documents/softwares/petsc/arch-linux2-cxx-debug/lib >>>>>>>>> [0]PETSC ERROR: Configure run at Sat Oct 26 16:35:15 2013 >>>>>>>>> [0]PETSC ERROR: Configure options --download-mpich >>>>>>>>> -download-f-blas-lapack=1 --download-metis --download-parmetis >>>>>>>>> --download-superlu_dist --download-scalapack --download-mumps >>>>>>>>> --download-hypre --with-clanguage=cxx >>>>>>>>> [0]PETSC ERROR: >>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>> [0]PETSC ERROR: MatSetValues_SeqAIJ() line 413 in >>>>>>>>> src/mat/impls/aij/seq/aij.c >>>>>>>>> [0]PETSC ERROR: MatSetValues() line 1130 in >>>>>>>>> src/mat/interface/matrix.c >>>>>>>>> [0]PETSC ERROR: computeMatrix2dSection() line 318 in >>>>>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>>>>> [0]PETSC ERROR: KSPSetUp() line 228 in >>>>>>>>> src/ksp/ksp/interface/itfunc.c >>>>>>>>> [0]PETSC ERROR: KSPSolve() line 399 in >>>>>>>>> src/ksp/ksp/interface/itfunc.c >>>>>>>>> [0]PETSC ERROR: main() line 598 in >>>>>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>>>>> [unset]: aborting job: >>>>>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>>>>> >>>>>>>>> >>>>>>>>>> Matt >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> However, now since I do not want the rows in the matrix >>>>>>>>>>> corresponding to the points in the grid which do not lie in the >>>>>>>>>>> computational domain. This means the size of the matrix will not >>>>>>>>>>> necessarily equal the total number of cells in DMDA. So in the following >>>>>>>>>>> code: >>>>>>>>>>> >>>>>>>>>>> ierr = DMDAGetLocalInfo(dm,&info);CHKERRQ(ierr); >>>>>>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, >>>>>>>>>>> &s);CHKERRQ(ierr); >>>>>>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, >>>>>>>>>>> &nC);CHKERRQ(ierr); >>>>>>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>>>>>> PetscInt point; >>>>>>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>>>>>> &point);CHKERRQ(ierr); >>>>>>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>>>>>> } >>>>>>>>>>> >>>>>>>>>>> } >>>>>>>>>>> } >>>>>>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>>>>>> >>>>>>>>>>> should I myself compute proper nC (i.e. total no. of points for >>>>>>>>>>> which I want the matrix to have a corresponding row) before calling >>>>>>>>>>> PetscSectionSetChart() or is the masking out of the points inside the loop >>>>>>>>>>> with if(isPosInDomain(&testPoisson,i,j,0)) sufficient ? >>>>>>>>>>> And, when you write: >>>>>>>>>>> PetscSectionGetOffset(gs, p, &ind); >>>>>>>>>>> row = ind < 0 ? -(ind+1) : ind; >>>>>>>>>>> >>>>>>>>>>> it seems you allow the possibility of getting a -ve ind when >>>>>>>>>>> using PetscSectionGetOffset. When would an offset to a particular dof and >>>>>>>>>>> point? >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>>>>>> num = 0; >>>>>>>>>>>>> /*ierr = >>>>>>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr);*/ /*Get the PetscPoint*/ >>>>>>>>>>>>> /*But I did now understand how I would emulate >>>>>>>>>>>>> the row.c and col.c info with PetscSection*/ >>>>>>>>>>>>> row.i = i; row.j = j; >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Here you would call >>>>>>>>>>>> >>>>>>>>>>>> ierr = DMDAGetCellPoint(da, i, j, 0, &cell);CHKERRQ(ierr); >>>>>>>>>>>> ierr = PetscSectionGetOffset(da, cell, &row);CHKERRQ(ierr); >>>>>>>>>>>> ierr = PetscSectionGetDof(da, cell, &dof);CHKERRQ(ierr); >>>>>>>>>>>> >>>>>>>>>>>> This means that this cell has global rows = [row, row+dof). >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> col[num].i = i; col[num].j = j; >>>>>>>>>>>>> if (isPosInDomain(ctx,i,j)) { /*put the >>>>>>>>>>>>> operator for only the values for only the points lying in the domain*/ >>>>>>>>>>>>> v[num++] = -4; >>>>>>>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>>>>>>> col[num].i = i+1; col[num].j = j; >>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>> } >>>>>>>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>>>>>>> col[num].i = i-1; col[num].j = j; >>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>> } >>>>>>>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>>>>>>> col[num].i = i; col[num].j = j+1; >>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>> } >>>>>>>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>>>>>>> col[num].i = i; col[num].j = j-1; >>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>> } >>>>>>>>>>>>> } else { >>>>>>>>>>>>> v[num++] = dScale; /*set Dirichlet >>>>>>>>>>>>> identity rows for points in the rectangle but outside the computational >>>>>>>>>>>>> domain*/ >>>>>>>>>>>>> } >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> You do the same thing for the columns, and then call >>>>>>>>>>>> >>>>>>>>>>>> ierr = MatSetValues(A, dof, rows, num*dof, cols, v, >>>>>>>>>>>> INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> ierr = >>>>>>>>>>>>> MatSetValuesStencil(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>>>>> } >>>>>>>>>>>>> } >>>>>>>>>>>>> } >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>> >>>>>>>>>>>>>> Matt >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> If not, just put the identity into >>>>>>>>>>>>>>>>>>>>>>>> the rows you do not use on the full cube. It will >>>>>>>>>>>>>>>>>>>>>>>> not hurt scalability or convergence. >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> In the case of Poisson with Dirichlet condition this >>>>>>>>>>>>>>>>>>>>>>> might be the case. But is it always true that having identity rows in the >>>>>>>>>>>>>>>>>>>>>>> system matrix will not hurt convergence ? I thought otherwise for the >>>>>>>>>>>>>>>>>>>>>>> following reasons: >>>>>>>>>>>>>>>>>>>>>>> 1) Having read Jed's answer here : >>>>>>>>>>>>>>>>>>>>>>> http://scicomp.stackexchange.com/questions/3426/why-is-pinning-a-point-to-remove-a-null-space-bad/3427#3427 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Jed is talking about a constraint on a the pressure >>>>>>>>>>>>>>>>>>>>>> at a point. This is just decoupling these unknowns from the rest >>>>>>>>>>>>>>>>>>>>>> of the problem. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2) Some observation I am getting (but I am still >>>>>>>>>>>>>>>>>>>>>>> doing more experiments to confirm) while solving my staggered-grid 3D >>>>>>>>>>>>>>>>>>>>>>> stokes flow with schur complement and using -pc_type gamg for A00 matrix. >>>>>>>>>>>>>>>>>>>>>>> Putting the identity rows for dirichlet boundaries and for ghost cells >>>>>>>>>>>>>>>>>>>>>>> seemed to have effects on its convergence. I'm hoping once I know how to >>>>>>>>>>>>>>>>>>>>>>> use PetscSection, I can get rid of using ghost cells method for the >>>>>>>>>>>>>>>>>>>>>>> staggered grid and get rid of the identity rows too. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> It can change the exact iteration, but it does not >>>>>>>>>>>>>>>>>>>>>> make the matrix conditioning worse. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Anyway please provide me with some pointers so >>>>>>>>>>>>>>>>>>>>>>> that I can start trying with petscsection on top of a dmda, in the >>>>>>>>>>>>>>>>>>>>>>> beginning for non-staggered case. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before >>>>>>>>>>>>>>>>>>>>>>>> they begin their experiments is infinitely more interesting than any >>>>>>>>>>>>>>>>>>>>>>>> results to which their experiments lead. >>>>>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>> their experiments lead. >>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>>> experiments lead. >>>>>>>>>> -- Norbert Wiener >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> What most experimenters take for granted before they begin their >>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>> experiments lead. >>>>>>>> -- Norbert Wiener >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> What most experimenters take for granted before they begin their >>>>>> experiments is infinitely more interesting than any results to which their >>>>>> experiments lead. >>>>>> -- Norbert Wiener >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> What most experimenters take for granted before they begin their >>>> experiments is infinitely more interesting than any results to which their >>>> experiments lead. >>>> -- Norbert Wiener >>>> >>> >>> >> >> >> -- >> What most experimenters take for granted before they begin their >> experiments is infinitely more interesting than any results to which their >> experiments lead. >> -- Norbert Wiener >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Mon Nov 11 13:46:57 2013 From: irving at naml.us (Geoffrey Irving) Date: Mon, 11 Nov 2013 11:46:57 -0800 Subject: [petsc-users] petsc4py and mixed python/c++ code Message-ID: I'm revamping Otherlab's use of petsc, and taking another look at petsc4py as part of that. Previously we rolled our own Python bindings, but it'd be great to not have to do that. However, our usage would involve writing new C/C++ functions which take petsc types as arguments, and then exposing these new functions to python. Is there a way to do this if the petsc4py bindings are written in Cython? Geoffrey From knepley at gmail.com Mon Nov 11 14:24:10 2013 From: knepley at gmail.com (Matthew Knepley) Date: Mon, 11 Nov 2013 14:24:10 -0600 Subject: [petsc-users] petsc4py and mixed python/c++ code In-Reply-To: References: Message-ID: On Mon, Nov 11, 2013 at 1:46 PM, Geoffrey Irving wrote: > I'm revamping Otherlab's use of petsc, and taking another look at > petsc4py as part of that. Previously we rolled our own Python > bindings, but it'd be great to not have to do that. > > However, our usage would involve writing new C/C++ functions which > take petsc types as arguments, and then exposing these new functions > to python. Is there a way to do this if the petsc4py bindings are > written in Cython? 1) Lisandro is the expert 2) C is fairly trivial using Cython 3) If you use a bunch of C++ stuff, at least when I evaluated things last time, Cython was fairly hard to manage (esp. templates). However, SWIG worked alright. I have never tried mixing SWIG and Cython, but I would think its possible, since all you have to do is pull out a wrapped pointer. I have never seen mixed language programming be more efficient for a developing library. Wrapping stable, debugged libraries makes a lot of sense, and can really speed up development. Wrapping libraries you are still playing with makes interfaces changes harder, and debugging a nightmare. Matt > > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Mon Nov 11 16:39:20 2013 From: knepley at gmail.com (Matthew Knepley) Date: Mon, 11 Nov 2013 16:39:20 -0600 Subject: [petsc-users] using petsc tools to solve isolated irregular domains with finite difference In-Reply-To: References: Message-ID: On Mon, Nov 11, 2013 at 1:08 PM, Bishesh Khanal wrote: > On Mon, Nov 4, 2013 at 5:25 PM, Bishesh Khanal wrote: >> >> On Mon, Nov 4, 2013 at 4:54 PM, Matthew Knepley wrote: >> >>> On Mon, Nov 4, 2013 at 4:49 AM, Bishesh Khanal wrote: >>> >>>> On Fri, Nov 1, 2013 at 1:30 PM, Matthew Knepley wrote: >>>> >>>>> On Fri, Nov 1, 2013 at 12:08 AM, Bishesh Khanal wrote: >>>>> >>>>>> On Thu, Oct 31, 2013 at 11:34 PM, Matthew Knepley wrote: >>>>>> >>>>>>> On Thu, Oct 31, 2013 at 5:02 PM, Bishesh Khanal >>>>>> > wrote: >>>>>>> >>>>>>>> On Thu, Oct 31, 2013 at 7:43 PM, Matthew Knepley >>>>>>> > wrote: >>>>>>>> >>>>>>>>> On Thu, Oct 31, 2013 at 1:25 PM, Bishesh Khanal < >>>>>>>>> bisheshkh at gmail.com> wrote: >>>>>>>>> >>>>>>>>>> I did not call DMCreateMatrix() before when using just dmda (and >>>>>>>>>> it is working). In any case, I added a call to DMCreateMatrix() now but >>>>>>>>>> still there are problems. Here are the code snippets and the error I get: >>>>>>>>>> >>>>>>>>> >>>>>>>>> Then you were allowing it to be called automatically by PETSc, and >>>>>>>>> it would have been this time as well. >>>>>>>>> >>>>>>>>> >>>>>>>>>> //Setting up the section and linking it to DM: >>>>>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, >>>>>>>>>> &nC);CHKERRQ(ierr); >>>>>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>>>>> PetscInt point; >>>>>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>>>>> &point);CHKERRQ(ierr); >>>>>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>>>>> } >>>>>>>>>> } >>>>>>>>>> } >>>>>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>>>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>>>>>>>> >>>>>>>>>> //Set up KSP: >>>>>>>>>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>>>>>>>>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>>>>>>>>> ierr = >>>>>>>>>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>>>>>> ierr = >>>>>>>>>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>>>>>> ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); >>>>>>>>>> ierr = KSPSolve(ksp,NULL,NULL);CHKERRQ(ierr); >>>>>>>>>> >>>>>>>>>> ------------------------------------------------------ >>>>>>>>>> The computeMatrix2dSection function has: >>>>>>>>>> >>>>>>>>>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>>>>>>>>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>>>>>>>>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>>> if (isPosInDomain(ctx,i,j)) { >>>>>>>>>> ierr = >>>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>>>>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>>>>>>> ierr = PetscSectionGetDof(gs,point,&dof); >>>>>>>>>> for(PetscInt cDof = 0; cDof < dof; ++cDof) { >>>>>>>>>> num = 0; >>>>>>>>>> row+=cDof; >>>>>>>>>> col[num] = row; //(i,j) position >>>>>>>>>> v[num++] = -4; >>>>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>>>> ierr = >>>>>>>>>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>>>>>>>>> ierr = >>>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>>> col[num] += cDof; >>>>>>>>>> v[num++] = 1; >>>>>>>>>> } >>>>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>>>> ierr = >>>>>>>>>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>>>>>>>>> ierr = >>>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>>> col[num] += cDof; >>>>>>>>>> v[num++] = 1; >>>>>>>>>> } >>>>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>>>> ierr = >>>>>>>>>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>>>>>>>>> ierr = >>>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>>> col[num] += cDof; >>>>>>>>>> v[num++] = 1; >>>>>>>>>> } >>>>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>>>> ierr = >>>>>>>>>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>>>>>>>>> ierr = >>>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>>> col[num] += cDof; >>>>>>>>>> v[num++] = 1; >>>>>>>>>> } >>>>>>>>>> ierr = >>>>>>>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>> } >>>>>>>>>> } >>>>>>>>>> } >>>>>>>>>> } >>>>>>>>>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>>>>>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>>>>>> >>>>>>>>>> But this is not working. For e.g. for a 6X6 grid size I get the >>>>>>>>>> following error: >>>>>>>>>> >>>>>>>>> >>>>>>>>> Okay, here is how to debug this: >>>>>>>>> >>>>>>>>> 0) Go to a single scalar equations to make things easier >>>>>>>>> >>>>>>>>> 1) Use MatView(A, PETSC_VIEWER_STDOUT_WORLD), which should >>>>>>>>> output a 36 row matrix with >>>>>>>>> the preallocated nonzero pattern, filled with zeros >>>>>>>>> >>>>>>>>> 2) Make sure this is the pattern you want >>>>>>>>> >>>>>>>>> 3) Run in the debugger with -start_in_debugger >>>>>>>>> >>>>>>>>> 4) When you get the error, see >>>>>>>>> >>>>>>>>> a) If the (i, j) is one that should be setting a value >>>>>>>>> >>>>>>>>> b) Why this (i, j) was not preallocated >>>>>>>>> >>>>>>>> >>>>>>>> Up to 4 (a), it is correct. There is a problem in the way >>>>>>>> DMDAGetCellPoint and PetscSectionGetOffset works in this case scenario. I >>>>>>>> will try to explain it below for the case of 4X4 grid. >>>>>>>> First case: >>>>>>>> If I set the computational domain to be all the points of the dmda >>>>>>>> grid, (i.e. isPosInDomain(..,i,j,..) returns true for all i and j in the >>>>>>>> dmda grid), the program runs fine and does not give any error. >>>>>>>> >>>>>>>> Second case: >>>>>>>> I want the computational domain to be some part of the whole grid. >>>>>>>> There is a problem in this case. >>>>>>>> The following test is in a case where, >>>>>>>> isPosInDomain(..,i,j,..) returns true ONLY for (i,j) pairs of (2,1) >>>>>>>> (1,2) (2,2) (3,2) and (3,2). The grid with its corresponding point number >>>>>>>> in the petscsection is shown below: >>>>>>>> >>>>>>>> 12 13 14 15 >>>>>>>> 8 9 10 11 >>>>>>>> 4 5 6 7 >>>>>>>> 0 1 2 3 >>>>>>>> >>>>>>>> of which 6, 9, 10, 11 and 14 correspond to the (i,j) points that >>>>>>>> returns true for isPosInDomain(..,i,j,..) >>>>>>>> MatView gives me the 16-row matrix with the star stencil non-zero >>>>>>>> structure as expected. >>>>>>>> The error I get is: new non-zero at (0,2) caused a malloc. >>>>>>>> >>>>>>>> This error is for the value of (i,j) = (2,1), i.e. point 6. >>>>>>>> The loop to set values in the matrix starts as: >>>>>>>> for(PetscInt j = info.ys; j>>>>>>> for(PetscInt i = info.xs; i>>>>>>> if (isPosInDomain(ctx,i,j)) { >>>>>>>> ierr = >>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>>>>> >>>>>>>> Now, what I wanted from the beginning is to create the matrix >>>>>>>> containing the rows corresponding to only those points (i,j) which has true >>>>>>>> values for isPosInDomain(ctx,i,j), that means in this case 5 row matrix. In >>>>>>>> the current test, the first (i,j) that reaches DMDAGetCellPoint is (2,1), >>>>>>>> which gives sets point variable to 6. >>>>>>>> The line PetscSectionGetOffset(gs,point,&row) sets the value of row >>>>>>>> to 0. >>>>>>>> So I think this where the inconsistency lies. >>>>>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES) expects row variable >>>>>>>> and col[] array to correspond to (i,j) based on DMDA, but >>>>>>>> PetsSectionGetOffset gives result based on how we masked away some of the >>>>>>>> points from dmda grid. Or am I missing something very basic here ? >>>>>>>> >>>>>>> >>>>>>> You are missing something basic. The row is correctly identified as >>>>>>> 0, and that means the 2 is >>>>>>> point 10. >>>>>>> >>>>>> The problem must be preallocation. First, you should send the result >>>>>>> of MatView(). >>>>>>> >>>>>> >>>>>> MatView shows that the preallocation and the matrix created is for >>>>>> the full dmda grid which is probably not what we want here, right ? Here is >>>>>> the matview result: >>>>>> >>>>> >>>>> Thats right. Did you call >>>>> >>>>> DMSetDefaultSection(da, s) >>>>> >>>>> after you made the new section and before you called DMGetMatrix()? >>>>> >>>> Do you mean before I called DMCreateMatrix() ? I called >>>> DMSetDefaultSection(da,s) as below: >>>> >>>> ... >>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>> // ierr = DMGetMatrix(dm,&A);CHKERRQ(ierr); //It seems DMGetMatrix >>>> is changed to DMCreateMatrix(). >>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>> ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >>>> >>> >>> Yes, this is correct. I do exactly this in the example code and it works >>> fine. I suspect >>> you are not creating the section you think you are. You can send the >>> code if you want >>> and I will look at it. >>> >> >> Thanks, here is the code: >> https://github.com/bishesh/petscPoissonIrregular/blob/master/poissonIrregular.cxx >> It works for 2D and 3D without using PetscSection. (The choice of using >> Petsc_section is on line 68). Using PetsSection currently it implements >> only the 2D case with dof = 1. It works when the domain consists of all the >> cells of the dmda, it works (to check you can uncomment line 129 in >> isPosInDomain(). But it does not when I mask out some of the cells while >> creating the petscsection. >> > > Hi Matt, I'm afraid if my last email somehow escaped you. Could you please > let me know what could be possibly going wrong with my code linked above ? > There is a bug in DMCreateMatrix() for DMDA when using a PetscSection. I am fixing it. I will mail when its done. Thanks, Matt > > >> >>> Matt >>> >>> >>>> This MatView on A shows (result below) that it still has 16 rows for >>>> all DMDA points instead of only those petscsection points for which I set a >>>> non-zero dof. Is it what it should not be doing ? >>>> >>>>> Mat Object: 1 MPI processes >>>>>> >>>>>> type: seqaij >>>>>> >>>>>> row 0: (0, 0) (1, 0) (4, 0) >>>>>> >>>>>> row 1: (0, 0) (1, 0) (2, 0) (5, 0) >>>>>> >>>>>> row 2: (1, 0) (2, 0) (3, 0) (6, 0) >>>>>> >>>>>> row 3: (2, 0) (3, 0) (7, 0) >>>>>> >>>>>> row 4: (0, 0) (4, 0) (5, 0) (8, 0) >>>>>> >>>>>> row 5: (1, 0) (4, 0) (5, 0) (6, 0) (9, 0) >>>>>> >>>>>> row 6: (2, 0) (5, 0) (6, 0) (7, 0) (10, 0) >>>>>> >>>>>> row 7: (3, 0) (6, 0) (7, 0) (11, 0) >>>>>> >>>>>> row 8: (4, 0) (8, 0) (9, 0) (12, 0) >>>>>> >>>>>> row 9: (5, 0) (8, 0) (9, 0) (10, 0) (13, 0) >>>>>> >>>>>> row 10: (6, 0) (9, 0) (10, 0) (11, 0) (14, 0) >>>>>> >>>>>> row 11: (7, 0) (10, 0) (11, 0) (15, 0) >>>>>> >>>>>> row 12: (8, 0) (12, 0) (13, 0) >>>>>> >>>>>> row 13: (9, 0) (12, 0) (13, 0) (14, 0) >>>>>> >>>>>> row 14: (10, 0) (13, 0) (14, 0) (15, 0) >>>>>> >>>>>> row 15: (11, 0) (14, 0) (15, 0) >>>>>> >>>>>> [0]PETSC ERROR: --------------------- Error Message >>>>>> ------------------------------------ >>>>>> >>>>>> [0]PETSC ERROR: Argument out of range! >>>>>> >>>>>> [0]PETSC ERROR: New nonzero at (0,2) caused a malloc! >>>>>> >>>>>> >>>>>> >>>> For reference here is the code snippet again: >>>> ... >>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>> PetscInt point; >>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>> &point);CHKERRQ(ierr); >>>> ierr = PetscSectionSetDof(s, point, testPoisson.mDof); >>>> // No. of dofs associated with the point. >>>> } >>>> } >>>> } >>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>> ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >>>> >>>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>>> >>>> ierr = >>>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>> ierr = >>>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>> >>>> .... >>>> The function computeMatrix2dSection: >>>> >>>> #undef __FUNCT__ >>>> #define __FUNCT__ "computeMatrix2dSection" >>>> PetscErrorCode computeMatrix2dSection(KSP ksp, Mat A, Mat B, >>>> MatStructure *str, void *context) { >>>> PoissonModel *ctx = (PoissonModel*)context; >>>> PetscErrorCode ierr; >>>> DM da; >>>> DMDALocalInfo info; >>>> PetscInt row, col[5]; >>>> PetscInt dof; >>>> PetscScalar v[5]; //array to store 5 point stencil for >>>> each row >>>> PetscInt num; //non-zero position in the current row >>>> // PetscScalar dScale = 1; //to scale Dirichlet >>>> identity rows if needed >>>> PetscSection gs; //Global Section that keeps >>>> the grid info and indices >>>> PetscInt point; //Current point of the >>>> petscSection >>>> >>>> PetscFunctionBeginUser; >>>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>>> ierr = DMCreateMatrix(da,&A);CHKERRQ(ierr); >>>> // ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >>>> for(PetscInt j = info.ys; j>>> for(PetscInt i = info.xs; i>>> if (isPosInDomain(ctx,i,j)) { >>>> ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>> ierr = PetscSectionGetDof(gs,point,&dof); >>>> for(PetscInt cDof = 0; cDof < ctx->mDof; ++cDof) { >>>> num = 0; >>>> row+=cDof; >>>> col[num] = row; //(i,j) position >>>> v[num++] = -4; >>>> if(isPosInDomain(ctx,i+1,j)) { >>>> ierr = >>>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>>> ierr = >>>> PetscSectionGetOffset(gs,point,&col[num]); >>>> col[num] += cDof; >>>> v[num++] = 1; >>>> } >>>> if(isPosInDomain(ctx,i-1,j)) { >>>> ierr = >>>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>>> ierr = >>>> PetscSectionGetOffset(gs,point,&col[num]); >>>> col[num] += cDof; >>>> v[num++] = 1; >>>> } >>>> if(isPosInDomain(ctx,i,j+1)) { >>>> ierr = >>>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>>> ierr = >>>> PetscSectionGetOffset(gs,point,&col[num]); >>>> col[num] += cDof; >>>> v[num++] = 1; >>>> } >>>> if(isPosInDomain(ctx,i,j-1)) { >>>> ierr = >>>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>>> ierr = >>>> PetscSectionGetOffset(gs,point,&col[num]); >>>> col[num] += cDof; >>>> v[num++] = 1; >>>> } >>>> ierr = >>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>> } >>>> } >>>> } >>>> } >>>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>> >>>> >>>> >>>> >>>>> >>>>>>> Matt >>>>>>> >>>>>>> >>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> >>>>>>>>> Matt >>>>>>>>> >>>>>>>>> >>>>>>>>>> [0]PETSC ERROR: --------------------- Error Message >>>>>>>>>> ------------------------------------ >>>>>>>>>> [0]PETSC ERROR: Argument out of range! >>>>>>>>>> [0]PETSC ERROR: New nonzero at (0,6) caused a malloc! >>>>>>>>>> [0]PETSC ERROR: >>>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>>> [0]PETSC ERROR: Petsc Development GIT revision: >>>>>>>>>> 61e5e40bb2c5bf2423e94b71a15fef47e411b0da GIT Date: 2013-10-25 21:47:45 >>>>>>>>>> -0500 >>>>>>>>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>>>>>>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble >>>>>>>>>> shooting. >>>>>>>>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>>>>>>>> [0]PETSC ERROR: >>>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>>> [0]PETSC ERROR: build/poissonIrregular on a arch-linux2-cxx-debug >>>>>>>>>> named edwards by bkhanal Thu Oct 31 19:23:58 2013 >>>>>>>>>> [0]PETSC ERROR: Libraries linked from >>>>>>>>>> /home/bkhanal/Documents/softwares/petsc/arch-linux2-cxx-debug/lib >>>>>>>>>> [0]PETSC ERROR: Configure run at Sat Oct 26 16:35:15 2013 >>>>>>>>>> [0]PETSC ERROR: Configure options --download-mpich >>>>>>>>>> -download-f-blas-lapack=1 --download-metis --download-parmetis >>>>>>>>>> --download-superlu_dist --download-scalapack --download-mumps >>>>>>>>>> --download-hypre --with-clanguage=cxx >>>>>>>>>> [0]PETSC ERROR: >>>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>>> [0]PETSC ERROR: MatSetValues_SeqAIJ() line 413 in >>>>>>>>>> src/mat/impls/aij/seq/aij.c >>>>>>>>>> [0]PETSC ERROR: MatSetValues() line 1130 in >>>>>>>>>> src/mat/interface/matrix.c >>>>>>>>>> [0]PETSC ERROR: computeMatrix2dSection() line 318 in >>>>>>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>>>>>> [0]PETSC ERROR: KSPSetUp() line 228 in >>>>>>>>>> src/ksp/ksp/interface/itfunc.c >>>>>>>>>> [0]PETSC ERROR: KSPSolve() line 399 in >>>>>>>>>> src/ksp/ksp/interface/itfunc.c >>>>>>>>>> [0]PETSC ERROR: main() line 598 in >>>>>>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>>>>>> [unset]: aborting job: >>>>>>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> Matt >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> However, now since I do not want the rows in the matrix >>>>>>>>>>>> corresponding to the points in the grid which do not lie in the >>>>>>>>>>>> computational domain. This means the size of the matrix will not >>>>>>>>>>>> necessarily equal the total number of cells in DMDA. So in the following >>>>>>>>>>>> code: >>>>>>>>>>>> >>>>>>>>>>>> ierr = DMDAGetLocalInfo(dm,&info);CHKERRQ(ierr); >>>>>>>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, >>>>>>>>>>>> &s);CHKERRQ(ierr); >>>>>>>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, >>>>>>>>>>>> &nC);CHKERRQ(ierr); >>>>>>>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>>>>>>> PetscInt point; >>>>>>>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>>>>>>> &point);CHKERRQ(ierr); >>>>>>>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>>>>>>> } >>>>>>>>>>>> >>>>>>>>>>>> } >>>>>>>>>>>> } >>>>>>>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>>>>>>> >>>>>>>>>>>> should I myself compute proper nC (i.e. total no. of points for >>>>>>>>>>>> which I want the matrix to have a corresponding row) before calling >>>>>>>>>>>> PetscSectionSetChart() or is the masking out of the points inside the loop >>>>>>>>>>>> with if(isPosInDomain(&testPoisson,i,j,0)) sufficient ? >>>>>>>>>>>> And, when you write: >>>>>>>>>>>> PetscSectionGetOffset(gs, p, &ind); >>>>>>>>>>>> row = ind < 0 ? -(ind+1) : ind; >>>>>>>>>>>> >>>>>>>>>>>> it seems you allow the possibility of getting a -ve ind when >>>>>>>>>>>> using PetscSectionGetOffset. When would an offset to a particular dof and >>>>>>>>>>>> point? >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>>>>>>> { >>>>>>>>>>>>>> num = 0; >>>>>>>>>>>>>> /*ierr = >>>>>>>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr);*/ /*Get the PetscPoint*/ >>>>>>>>>>>>>> /*But I did now understand how I would >>>>>>>>>>>>>> emulate the row.c and col.c info with PetscSection*/ >>>>>>>>>>>>>> row.i = i; row.j = j; >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Here you would call >>>>>>>>>>>>> >>>>>>>>>>>>> ierr = DMDAGetCellPoint(da, i, j, 0, &cell);CHKERRQ(ierr); >>>>>>>>>>>>> ierr = PetscSectionGetOffset(da, cell, &row);CHKERRQ(ierr); >>>>>>>>>>>>> ierr = PetscSectionGetDof(da, cell, &dof);CHKERRQ(ierr); >>>>>>>>>>>>> >>>>>>>>>>>>> This means that this cell has global rows = [row, row+dof). >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> col[num].i = i; col[num].j = j; >>>>>>>>>>>>>> if (isPosInDomain(ctx,i,j)) { /*put the >>>>>>>>>>>>>> operator for only the values for only the points lying in the domain*/ >>>>>>>>>>>>>> v[num++] = -4; >>>>>>>>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>>>>>>>> col[num].i = i+1; col[num].j = j; >>>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>>> } >>>>>>>>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>>>>>>>> col[num].i = i-1; col[num].j = j; >>>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>>> } >>>>>>>>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>>>>>>>> col[num].i = i; col[num].j = j+1; >>>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>>> } >>>>>>>>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>>>>>>>> col[num].i = i; col[num].j = j-1; >>>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>>> } >>>>>>>>>>>>>> } else { >>>>>>>>>>>>>> v[num++] = dScale; /*set Dirichlet >>>>>>>>>>>>>> identity rows for points in the rectangle but outside the computational >>>>>>>>>>>>>> domain*/ >>>>>>>>>>>>>> } >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> You do the same thing for the columns, and then call >>>>>>>>>>>>> >>>>>>>>>>>>> ierr = MatSetValues(A, dof, rows, num*dof, cols, v, >>>>>>>>>>>>> INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> ierr = >>>>>>>>>>>>>> MatSetValuesStencil(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>>>>>> } >>>>>>>>>>>>>> } >>>>>>>>>>>>>> } >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> If not, just put the identity into >>>>>>>>>>>>>>>>>>>>>>>>> the rows you do not use on the full cube. It will >>>>>>>>>>>>>>>>>>>>>>>>> not hurt scalability or convergence. >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> In the case of Poisson with Dirichlet condition >>>>>>>>>>>>>>>>>>>>>>>> this might be the case. But is it always true that having identity rows in >>>>>>>>>>>>>>>>>>>>>>>> the system matrix will not hurt convergence ? I thought otherwise for the >>>>>>>>>>>>>>>>>>>>>>>> following reasons: >>>>>>>>>>>>>>>>>>>>>>>> 1) Having read Jed's answer here : >>>>>>>>>>>>>>>>>>>>>>>> http://scicomp.stackexchange.com/questions/3426/why-is-pinning-a-point-to-remove-a-null-space-bad/3427#3427 >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Jed is talking about a constraint on a the pressure >>>>>>>>>>>>>>>>>>>>>>> at a point. This is just decoupling these unknowns from the rest >>>>>>>>>>>>>>>>>>>>>>> of the problem. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> 2) Some observation I am getting (but I am still >>>>>>>>>>>>>>>>>>>>>>>> doing more experiments to confirm) while solving my staggered-grid 3D >>>>>>>>>>>>>>>>>>>>>>>> stokes flow with schur complement and using -pc_type gamg for A00 matrix. >>>>>>>>>>>>>>>>>>>>>>>> Putting the identity rows for dirichlet boundaries and for ghost cells >>>>>>>>>>>>>>>>>>>>>>>> seemed to have effects on its convergence. I'm hoping once I know how to >>>>>>>>>>>>>>>>>>>>>>>> use PetscSection, I can get rid of using ghost cells method for the >>>>>>>>>>>>>>>>>>>>>>>> staggered grid and get rid of the identity rows too. >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> It can change the exact iteration, but it does not >>>>>>>>>>>>>>>>>>>>>>> make the matrix conditioning worse. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Anyway please provide me with some pointers so >>>>>>>>>>>>>>>>>>>>>>>> that I can start trying with petscsection on top of a dmda, in the >>>>>>>>>>>>>>>>>>>>>>>> beginning for non-staggered case. >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before >>>>>>>>>>>>>>>>>>>>>>>>> they begin their experiments is infinitely more interesting than any >>>>>>>>>>>>>>>>>>>>>>>>> results to which their experiments lead. >>>>>>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>>>> experiments lead. >>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>> experiments lead. >>>>>>>>> -- Norbert Wiener >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> What most experimenters take for granted before they begin their >>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>> experiments lead. >>>>>>> -- Norbert Wiener >>>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> What most experimenters take for granted before they begin their >>>>> experiments is infinitely more interesting than any results to which their >>>>> experiments lead. >>>>> -- Norbert Wiener >>>>> >>>> >>>> >>> >>> >>> -- >>> What most experimenters take for granted before they begin their >>> experiments is infinitely more interesting than any results to which their >>> experiments lead. >>> -- Norbert Wiener >>> >> >> > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From anush at bu.edu Mon Nov 11 18:38:21 2013 From: anush at bu.edu (Anush Krishnan) Date: Mon, 11 Nov 2013 19:38:21 -0500 Subject: [petsc-users] Calculating the PETSc index of a DMComposite vector Message-ID: Hi all, I have two 2D distributed arrays U and V, and a vector created from the DMComposite of both. If I know just the Natural Ordering index of an element of U or V, can I calculate the PETSc Ordering index of that element in the composite vector? Thank you, Anush -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Mon Nov 11 18:59:58 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Mon, 11 Nov 2013 17:59:58 -0700 Subject: [petsc-users] Calculating the PETSc index of a DMComposite vector In-Reply-To: References: Message-ID: <87habicu41.fsf@jedbrown.org> Anush Krishnan writes: > Hi all, > > I have two 2D distributed arrays U and V, and a vector created from the > DMComposite of both. If I know just the Natural Ordering index of an > element of U or V, can I calculate the PETSc Ordering index of that > element in the composite vector? That direction is fragile and/or expensive unless it is within the subdomain overlap/ghost region. Where are you getting this index from? -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From anush at bu.edu Mon Nov 11 19:57:28 2013 From: anush at bu.edu (Anush Krishnan) Date: Mon, 11 Nov 2013 20:57:28 -0500 Subject: [petsc-users] Calculating the PETSc index of a DMComposite vector In-Reply-To: <87habicu41.fsf@jedbrown.org> References: <87habicu41.fsf@jedbrown.org> Message-ID: On 11 November 2013 19:59, Jed Brown wrote: > That direction is fragile and/or expensive unless it is within the > subdomain overlap/ghost region. Where are you getting this index from? > Let me give a brief description of the problem I'm working on: I use a structured staggered grid to store a fluid velocity field, and I need to interpolate this field on to a set of points that represent an immersed body present in the flow. These points can be arbitrarily placed in the domain and need not conform to the grid. Around each body point, I choose a small rectangular region of velocity grid points and perform an interpolation with these values to obtain the velocity at the body point. This can be represented as: [C][u] = [b] where [C] contains the interpolation coefficients, [u] contains velocity values from the small rectangular region and [b] is the velocity at the body point. These matrices can be expanded to include the entire velocity field and all the body points. Currently, I'm storing a copy of the body on every process. And since my grid is structured and cartesian, it's easy to calculate the indices of the velocity grid points around each body point. This is how I obtain the Natural Ordering index. I created the vector [b] using VecCreate, and PETSC_DECIDE for the size on each process. The length of [b] is 2*nb, where nb is the number of body points, and the first half of the vector stores the x-component and the second stores the y-component of velocity. [u] is a DMComposite of the U and V components of velocity. Currently, my fix is to loop over all the velocity grid points, check if they come under the influence of any of the body points, and correspondingly set the interpolation coefficient value in the matrix. But this procedure requires a double loop. I haven't thought about parallelising the body storage yet. -------------- next part -------------- An HTML attachment was scrubbed... URL: From rongliang.chan at gmail.com Mon Nov 11 23:06:23 2013 From: rongliang.chan at gmail.com (Rongliang Chen) Date: Tue, 12 Nov 2013 13:06:23 +0800 Subject: [petsc-users] Debug AOCreateBasic In-Reply-To: <87siv9szpt.fsf@mcs.anl.gov> References: <52733BB5.5030203@gmail.com> <8761scheg2.fsf@mcs.anl.gov> <52734F7B.2060002@gmail.com> <8761sbpfq0.fsf@mcs.anl.gov> <5279D10A.8080204@gmail.com> <877gcmxfwv.fsf@mcs.anl.gov> <527AFBAD.6030904@gmail.com> <87siv9szpt.fsf@mcs.anl.gov> Message-ID: <5281B74F.4010100@gmail.com> Hi Jed, I tried the mpich version petsc on Janus (configured with option --download-mpich) and my code stopped at another place. The error message is followed. Do you have any suggestions? For the core dump, I emailed the administrators of the Janus for help about a week ago but have not get any reply yet. Best, Rongliang ---------------------------- [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the batch system) has told this process to end [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger [0]PETSC ERROR: or see http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSC ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory corruption errors [0]PETSC ERROR: likely location of problem given in stack below [0]PETSC ERROR: --------------------- Stack Frames ------------------------------------ [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available, [0]PETSC ERROR: INSTEAD the line number of the start of the function [0]PETSC ERROR: is given. [0]PETSC ERROR: [0] PetscBinaryRead line 234 /projects/ronglian/soft/petsc-3.4.3/src/sys/fileio/sysio.c [0]PETSC ERROR: [0] GetPieceData line 1096 readbinary3d.c [0]PETSC ERROR: [0] DataReadAndSplitGeneric line 962 readbinary3d.c [0]PETSC ERROR: [0] DataRead line 621 readbinary3d.c [0]PETSC ERROR: [0] ReadBinary line 184 readbinary3d.c [0]PETSC ERROR: [0] LoadGrid line 720 loadgrid3d.c [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Signal received! [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 [0]PETSC ERROR: See docs/changes/index.html for recent updates. [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [0]PETSC ERROR: See docs/index.html for manual pages. [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: ./fsi3d on a Janus-debug-64bit-mpich named node0718 by ronglian Mon Nov 11 20:54:09 2013 [0]PETSC ERROR: Libraries linked from /projects/ronglian/soft/petsc-3.4.3/Janus-debug-64bit-mpich/lib [0]PETSC ERROR: Configure run at Mon Nov 11 20:49:25 2013 [0]PETSC ERROR: Configure options --known-level1-dcache-size=32768 --known-level1-dcache-linesize=64 --known-level1-dcache-assoc=8 --known-memcmp-ok=1 --known-sizeof-char=1 --known-sizeof-void-p=8 --known-sizeof-short=2 --known-sizeof-int=4 --known-sizeof-long=8 --known-sizeof-long-long=8 --known-sizeof-float=4 --known-sizeof-double=8 --known-sizeof-size_t=8 --known-bits-per-byte=8 --known-sizeof-MPI_Comm=4 --known-sizeof-MPI_Fint=4 --known-mpi-long-double=1 --known-mpi-c-double-complex=1 --download-mpich=1 --download-blacs=1 --download-f-blas-lapack=1 --download-metis=1 --download-parmetis=1 --download-scalapack=1 --download-superlu_dist=1 --known-mpi-shared-libraries=0 --with-64-bit-indices --with-batch=1 --download-exodusii=1 --download-hdf5=1 --download-netcdf=1 --known-64-bit-blas-indices --with-debugging=1 [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: User provided function() line 0 in unknown directory unknown file application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0 [unset]: aborting job: application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0 [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the batch system) has told this process to end [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger [0]PETSC ERROR: or see http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSC ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory corruption errors [0]PETSC ERROR: likely location of problem given in stack below On 11/07/2013 10:38 AM, Jed Brown wrote: > Rongliang Chen writes: > >> Hi Jed, >> >> I have not find a way to "dump core on selected ranks" yet and I will >> continue to do that. > Ask the administrators at your facility. There are a few common ways, > but I'm not going to play a guessing game on the mailing list. > >> I run my code with the option "-on_error_attach_debugger" and got the >> following message: >> >> -------------------------------------------------------------------------- >> An MPI process has executed an operation involving a call to the >> "fork()" system call to create a child process. Open MPI is currently >> operating in a condition that could result in memory corruption or >> other system errors; your MPI job may hang, crash, or produce silent >> data corruption. The use of fork() (or system() or other calls that >> create child processes) is strongly discouraged. >> >> The process that invoked fork was: >> >> Local host: node1529 (PID 3701) >> MPI_COMM_WORLD rank: 0 >> >> If you are *absolutely sure* that your application will successfully >> and correctly survive a call to fork(), you may disable this warning >> by setting the mpi_warn_on_fork MCA parameter to 0. >> -------------------------------------------------------------------------- >> [node1529:03700] 13 more processes have sent help message >> help-mpi-runtime.txt / mpi_init:warn-fork >> [node1529:03700] Set MCA parameter "orte_base_help_aggregate" to 0 to >> see all help / error messages >> -------------------------------------------------------------------------- >> >> Is this message useful for the debugging? > This is just a possibly technical problem attaching a debugger in your > environment, but you have to actually attach the debugger and poke > around (stack trace, etc). > > Can you create an interactive session and run your job from there? From zonexo at gmail.com Tue Nov 12 02:31:10 2013 From: zonexo at gmail.com (TAY wee-beng) Date: Tue, 12 Nov 2013 16:31:10 +0800 Subject: [petsc-users] Error linking with DMSETFUNCTION and DMSETJACOBIAN Message-ID: <5281E74E.10401@gmail.com> Hi, I am running PETSc 3.4 on windows 7 VS2008, using Intel Fortran. When I tried to build my code, during linking, it gives the error: 1>PETSc_solvers.obj : error LNK2019: unresolved external symbol *DMSETFUNCTION* referenced in function PETSC_SOLVERS_mp_P_MATRIX_SOLV_PETSC_MULTIGRID 1>PETSc_solvers.obj : error LNK2019: unresolved external symbol *DMSETJACOBIAN* referenced in function PETSC_SOLVERS_mp_P_MATRIX_SOLV_PETSC_MULTIGRID My subroutine is : #include "finclude/petsc.h90" PetscErrorCode ierr ... call DMSetFunction(da,ComputeRHS,ierr) ... call DMSetJacobian(da,ComputeMatrix,ierr) ... I also tried to use #include #include #include #include #include #include #include #include instead of #include "finclude/petsc.h90" but it still can't work. I remember it used to work, maybe when I'm using petsc-dev. Please help. Thank you! -- Yours sincerely, TAY wee-beng -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Tue Nov 12 02:34:23 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Tue, 12 Nov 2013 01:34:23 -0700 Subject: [petsc-users] Error linking with DMSETFUNCTION and DMSETJACOBIAN In-Reply-To: <5281E74E.10401@gmail.com> References: <5281E74E.10401@gmail.com> Message-ID: <87zjpaaui8.fsf@jedbrown.org> TAY wee-beng writes: > Hi, > > I am running PETSc 3.4 on windows 7 VS2008, using Intel Fortran. > > When I tried to build my code, during linking, it gives the error: > > 1>PETSc_solvers.obj : error LNK2019: unresolved external symbol > *DMSETFUNCTION* referenced in function * DMSetFunction() and DMSetJacobian() have been removed use SNESSetFunction() and SNESSetJacobian() instead, note the calling sequences are slightly different http://www.mcs.anl.gov/petsc/documentation/changes/34.html -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From zonexo at gmail.com Tue Nov 12 02:53:11 2013 From: zonexo at gmail.com (TAY wee-beng) Date: Tue, 12 Nov 2013 16:53:11 +0800 Subject: [petsc-users] Error linking with DMSETFUNCTION and DMSETJACOBIAN In-Reply-To: <87zjpaaui8.fsf@jedbrown.org> References: <5281E74E.10401@gmail.com> <87zjpaaui8.fsf@jedbrown.org> Message-ID: <5281EC77.5080101@gmail.com> Hi Jed, Is there any tutorial to show the new calling sequence? I looked at some examples and it seems that I need to create a new SNES using SNESCreate, and use SNESSetDM. However, I am not sure what are the arguments for SNESSetFunction and SNESSetJacobian in Fortran? Thanks. Yours sincerely, TAY wee-beng On 12/11/2013 16:34, Jed Brown wrote: > TAY wee-beng writes: > >> Hi, >> >> I am running PETSc 3.4 on windows 7 VS2008, using Intel Fortran. >> >> When I tried to build my code, during linking, it gives the error: >> >> 1>PETSc_solvers.obj : error LNK2019: unresolved external symbol >> *DMSETFUNCTION* referenced in function > * DMSetFunction() and DMSetJacobian() have been removed use > SNESSetFunction() and SNESSetJacobian() instead, note the calling > sequences are slightly different > > http://www.mcs.anl.gov/petsc/documentation/changes/34.html From jedbrown at mcs.anl.gov Tue Nov 12 03:12:49 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Tue, 12 Nov 2013 02:12:49 -0700 Subject: [petsc-users] Error linking with DMSETFUNCTION and DMSETJACOBIAN In-Reply-To: <5281EC77.5080101@gmail.com> References: <5281E74E.10401@gmail.com> <87zjpaaui8.fsf@jedbrown.org> <5281EC77.5080101@gmail.com> Message-ID: <87wqkeasq6.fsf@jedbrown.org> TAY wee-beng writes: > Hi Jed, > > Is there any tutorial to show the new calling sequence? > > I looked at some examples and it seems that I need to create a new SNES > using SNESCreate, and use SNESSetDM. > > However, I am not sure what are the arguments for SNESSetFunction and > SNESSetJacobian in Fortran? http://www.mcs.anl.gov/petsc/petsc-current/src/snes/examples/tutorials/ex5f90.F.html -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From juhaj at iki.fi Tue Nov 12 04:37:13 2013 From: juhaj at iki.fi (Juha =?ISO-8859-1?Q?J=E4ykk=E4?=) Date: Tue, 12 Nov 2013 11:37:13 +0100 Subject: [petsc-users] Unable to create >4GB sized HDF5 files on Cray XC30 In-Reply-To: <87li1kjh54.fsf@mcs.anl.gov> References: <2569883.TWAWmgZMUo@rigel> <5862566.rSSLWvQenN@rigel> <87li1kjh54.fsf@mcs.anl.gov> Message-ID: <6408526.5lg4qfbu4O@dhcp071> Hi Jed, Let's see what comes of this... I forked petsc/petsc and created a pull request. There are probably loads of unneeded casts and at least one "+1" just to make sure I'm not rounding a float down into an integer under any circumstances. > Thanks, the logic looks reasonable. Unfortunately, I don't think we can > have comprehensive tests because the large file sizes won't work in many > test environments. That said, we could have a configurable parameter > instead of hard-coding 4 GiB, in which case we could set it to 1 MiB (or > less) for testing. I'll leave the parameter for others to implement: I don't have a good idea of how to do that. I believe I can provide a testing environment (but no access to it, I'm afraid) for those big files, just let me know what kind of test suite I should create. Not that the old chunking seems to have had much testing considering how seriously broken it is. ;) > > + // just in case some libraries do not empty memory for local variables: > We have to use C89-style comments because of Microsoft. What is this, Microsoft's compiler cannot do C99? Is it too new a standard, at 14 years old? > > + > > + // some requirements for the chunks (assuming no split along time-step) > > + int ranks; > Declarations cannot be mixed with statements in C89. Microsoft again? I've reverted these *standard* pieces of code into an obsolete, old standard. I really had no idea there are STILL compilers around which cannot do C99. > PetscObjectComm((PetscObject)xin). And please name the variable "size" > or "comm_size". Done. > One reasonable thing is to write > > hsize_t Ki = 1024; > > and then > > Gi = Ki*Ki*Ki; Done. > I would prefer to move all the logic below into a function > VecGetHDF5ChunkSize(). Note that almost the same logic appears in Done. > Please use PetscMax() and PetscMin(). You can define the 3-way versions > at the top of the file if need be. Done. I also removed the 3-way versions. > You mixed tabs and spaces here. Please use all spaces. Probably forgot. =) I mostly write python, so I'm very surprised I managed to mix tabs and spaces, considering how nice side-effects that often has in python. I hope this works and can get to the next release. Cheers, Juha -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 230 bytes Desc: This is a digitally signed message part. URL: From knepley at gmail.com Tue Nov 12 05:45:32 2013 From: knepley at gmail.com (Matthew Knepley) Date: Tue, 12 Nov 2013 05:45:32 -0600 Subject: [petsc-users] Debug AOCreateBasic In-Reply-To: <5281B74F.4010100@gmail.com> References: <52733BB5.5030203@gmail.com> <8761scheg2.fsf@mcs.anl.gov> <52734F7B.2060002@gmail.com> <8761sbpfq0.fsf@mcs.anl.gov> <5279D10A.8080204@gmail.com> <877gcmxfwv.fsf@mcs.anl.gov> <527AFBAD.6030904@gmail.com> <87siv9szpt.fsf@mcs.anl.gov> <5281B74F.4010100@gmail.com> Message-ID: On Mon, Nov 11, 2013 at 11:06 PM, Rongliang Chen wrote: > Hi Jed, > > I tried the mpich version petsc on Janus (configured with option > --download-mpich) and my code stopped at another place. The error message > is followed. Do you have any suggestions? > 1) I believe you said that you ran under valgrind without errors, so we guess that GetPieceData() is fine. 2) I think it is quite unlikely that there is an error in PetscBinaryRead() 3) Wrong file size or permission should not cause an SEGV 4) Thus to me it clearly looks like a driver issue here with the filesystem. If this is reproducible, it should be easy for the administrators of the machine to look at, and this is definitely there job. Move up the hierarchy now. Matt > For the core dump, I emailed the administrators of the Janus for help > about a week ago but have not get any reply yet. > > Best, > Rongliang > > ---------------------------- > [0]PETSC ERROR: ------------------------------ > ------------------------------------------ > [0]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the > batch system) has told this process to end > [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger > [0]PETSC ERROR: or see http://www.mcs.anl.gov/petsc/ > documentation/faq.html#valgrind[0]PETSC ERROR: or try http://valgrind.orgon GNU/linux and Apple Mac OS X to find memory corruption errors > [0]PETSC ERROR: likely location of problem given in stack below > [0]PETSC ERROR: --------------------- Stack Frames > ------------------------------------ > [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not > available, > [0]PETSC ERROR: INSTEAD the line number of the start of the function > [0]PETSC ERROR: is given. > [0]PETSC ERROR: [0] PetscBinaryRead line 234 /projects/ronglian/soft/petsc- > 3.4.3/src/sys/fileio/sysio.c > [0]PETSC ERROR: [0] GetPieceData line 1096 readbinary3d.c > [0]PETSC ERROR: [0] DataReadAndSplitGeneric line 962 readbinary3d.c > [0]PETSC ERROR: [0] DataRead line 621 readbinary3d.c > [0]PETSC ERROR: [0] ReadBinary line 184 readbinary3d.c > [0]PETSC ERROR: [0] LoadGrid line 720 loadgrid3d.c > [0]PETSC ERROR: --------------------- Error Message > ------------------------------------ > [0]PETSC ERROR: Signal received! > [0]PETSC ERROR: ------------------------------ > ------------------------------------------ > [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > [0]PETSC ERROR: See docs/changes/index.html for recent updates. > [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > [0]PETSC ERROR: See docs/index.html for manual pages. > [0]PETSC ERROR: ------------------------------ > ------------------------------------------ > [0]PETSC ERROR: ./fsi3d on a Janus-debug-64bit-mpich named node0718 by > ronglian Mon Nov 11 20:54:09 2013 > [0]PETSC ERROR: Libraries linked from /projects/ronglian/soft/petsc- > 3.4.3/Janus-debug-64bit-mpich/lib > [0]PETSC ERROR: Configure run at Mon Nov 11 20:49:25 2013 > [0]PETSC ERROR: Configure options --known-level1-dcache-size=32768 > --known-level1-dcache-linesize=64 --known-level1-dcache-assoc=8 > --known-memcmp-ok=1 --known-sizeof-char=1 --known-sizeof-void-p=8 > --known-sizeof-short=2 --known-sizeof-int=4 --known-sizeof-long=8 > --known-sizeof-long-long=8 --known-sizeof-float=4 --known-sizeof-double=8 > --known-sizeof-size_t=8 --known-bits-per-byte=8 --known-sizeof-MPI_Comm=4 > --known-sizeof-MPI_Fint=4 --known-mpi-long-double=1 > --known-mpi-c-double-complex=1 --download-mpich=1 --download-blacs=1 > --download-f-blas-lapack=1 --download-metis=1 --download-parmetis=1 > --download-scalapack=1 --download-superlu_dist=1 > --known-mpi-shared-libraries=0 --with-64-bit-indices --with-batch=1 > --download-exodusii=1 --download-hdf5=1 --download-netcdf=1 > --known-64-bit-blas-indices --with-debugging=1 > [0]PETSC ERROR: ------------------------------ > ------------------------------------------ > [0]PETSC ERROR: User provided function() line 0 in unknown directory > unknown file > application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0 > [unset]: aborting job: > application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0 > [0]PETSC ERROR: ------------------------------ > ------------------------------------------ > [0]PETSC ERROR: Caught signal number 15 Terminate: Somet process (or the > batch system) has told this process to end > [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger > [0]PETSC ERROR: or see http://www.mcs.anl.gov/petsc/ > documentation/faq.html#valgrind[0]PETSC ERROR: or try http://valgrind.orgon GNU/linux and Apple Mac OS X to find memory corruption errors > [0]PETSC ERROR: likely location of problem given in stack below > > > On 11/07/2013 10:38 AM, Jed Brown wrote: > >> Rongliang Chen writes: >> >> Hi Jed, >>> >>> I have not find a way to "dump core on selected ranks" yet and I will >>> continue to do that. >>> >> Ask the administrators at your facility. There are a few common ways, >> but I'm not going to play a guessing game on the mailing list. >> >> I run my code with the option "-on_error_attach_debugger" and got the >>> following message: >>> >>> ------------------------------------------------------------ >>> -------------- >>> An MPI process has executed an operation involving a call to the >>> "fork()" system call to create a child process. Open MPI is currently >>> operating in a condition that could result in memory corruption or >>> other system errors; your MPI job may hang, crash, or produce silent >>> data corruption. The use of fork() (or system() or other calls that >>> create child processes) is strongly discouraged. >>> >>> The process that invoked fork was: >>> >>> Local host: node1529 (PID 3701) >>> MPI_COMM_WORLD rank: 0 >>> >>> If you are *absolutely sure* that your application will successfully >>> and correctly survive a call to fork(), you may disable this warning >>> by setting the mpi_warn_on_fork MCA parameter to 0. >>> ------------------------------------------------------------ >>> -------------- >>> [node1529:03700] 13 more processes have sent help message >>> help-mpi-runtime.txt / mpi_init:warn-fork >>> [node1529:03700] Set MCA parameter "orte_base_help_aggregate" to 0 to >>> see all help / error messages >>> ------------------------------------------------------------ >>> -------------- >>> >>> Is this message useful for the debugging? >>> >> This is just a possibly technical problem attaching a debugger in your >> environment, but you have to actually attach the debugger and poke >> around (stack trace, etc). >> >> Can you create an interactive session and run your job from there? >> > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From bisheshkh at gmail.com Tue Nov 12 08:27:03 2013 From: bisheshkh at gmail.com (Bishesh Khanal) Date: Tue, 12 Nov 2013 15:27:03 +0100 Subject: [petsc-users] using petsc tools to solve isolated irregular domains with finite difference In-Reply-To: References: Message-ID: On Mon, Nov 11, 2013 at 11:39 PM, Matthew Knepley wrote: > On Mon, Nov 11, 2013 at 1:08 PM, Bishesh Khanal wrote: > >> On Mon, Nov 4, 2013 at 5:25 PM, Bishesh Khanal wrote: >>> >>> On Mon, Nov 4, 2013 at 4:54 PM, Matthew Knepley wrote: >>> >>>> On Mon, Nov 4, 2013 at 4:49 AM, Bishesh Khanal wrote: >>>> >>>>> On Fri, Nov 1, 2013 at 1:30 PM, Matthew Knepley wrote: >>>>> >>>>>> On Fri, Nov 1, 2013 at 12:08 AM, Bishesh Khanal wrote: >>>>>> >>>>>>> On Thu, Oct 31, 2013 at 11:34 PM, Matthew Knepley >>>>>> > wrote: >>>>>>> >>>>>>>> On Thu, Oct 31, 2013 at 5:02 PM, Bishesh Khanal < >>>>>>>> bisheshkh at gmail.com> wrote: >>>>>>>> >>>>>>>>> On Thu, Oct 31, 2013 at 7:43 PM, Matthew Knepley < >>>>>>>>> knepley at gmail.com> wrote: >>>>>>>>> >>>>>>>>>> On Thu, Oct 31, 2013 at 1:25 PM, Bishesh Khanal < >>>>>>>>>> bisheshkh at gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> I did not call DMCreateMatrix() before when using just dmda (and >>>>>>>>>>> it is working). In any case, I added a call to DMCreateMatrix() now but >>>>>>>>>>> still there are problems. Here are the code snippets and the error I get: >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Then you were allowing it to be called automatically by PETSc, >>>>>>>>>> and it would have been this time as well. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> //Setting up the section and linking it to DM: >>>>>>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>>>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, >>>>>>>>>>> &nC);CHKERRQ(ierr); >>>>>>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>>>>>> PetscInt point; >>>>>>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>>>>>> &point);CHKERRQ(ierr); >>>>>>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>>>>>> } >>>>>>>>>>> } >>>>>>>>>>> } >>>>>>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>>>>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>>>>>>>>> >>>>>>>>>>> //Set up KSP: >>>>>>>>>>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>>>>>>>>>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>>>>>>>>>> ierr = >>>>>>>>>>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>>>>>>> ierr = >>>>>>>>>>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>>>>>>>> ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); >>>>>>>>>>> ierr = KSPSolve(ksp,NULL,NULL);CHKERRQ(ierr); >>>>>>>>>>> >>>>>>>>>>> ------------------------------------------------------ >>>>>>>>>>> The computeMatrix2dSection function has: >>>>>>>>>>> >>>>>>>>>>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>>>>>>>>>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>>>>>>>>>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>>>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>>>> if (isPosInDomain(ctx,i,j)) { >>>>>>>>>>> ierr = >>>>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>>>>>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>>>>>>>> ierr = PetscSectionGetDof(gs,point,&dof); >>>>>>>>>>> for(PetscInt cDof = 0; cDof < dof; ++cDof) { >>>>>>>>>>> num = 0; >>>>>>>>>>> row+=cDof; >>>>>>>>>>> col[num] = row; //(i,j) position >>>>>>>>>>> v[num++] = -4; >>>>>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>>>>> ierr = >>>>>>>>>>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>>>>>>>>>> ierr = >>>>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>>>> col[num] += cDof; >>>>>>>>>>> v[num++] = 1; >>>>>>>>>>> } >>>>>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>>>>> ierr = >>>>>>>>>>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>>>>>>>>>> ierr = >>>>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>>>> col[num] += cDof; >>>>>>>>>>> v[num++] = 1; >>>>>>>>>>> } >>>>>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>>>>> ierr = >>>>>>>>>>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>>>>>>>>>> ierr = >>>>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>>>> col[num] += cDof; >>>>>>>>>>> v[num++] = 1; >>>>>>>>>>> } >>>>>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>>>>> ierr = >>>>>>>>>>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>>>>>>>>>> ierr = >>>>>>>>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>>>>>>>> col[num] += cDof; >>>>>>>>>>> v[num++] = 1; >>>>>>>>>>> } >>>>>>>>>>> ierr = >>>>>>>>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>>> } >>>>>>>>>>> } >>>>>>>>>>> } >>>>>>>>>>> } >>>>>>>>>>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>>>>>>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>>>>>>>> >>>>>>>>>>> But this is not working. For e.g. for a 6X6 grid size I get the >>>>>>>>>>> following error: >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Okay, here is how to debug this: >>>>>>>>>> >>>>>>>>>> 0) Go to a single scalar equations to make things easier >>>>>>>>>> >>>>>>>>>> 1) Use MatView(A, PETSC_VIEWER_STDOUT_WORLD), which should >>>>>>>>>> output a 36 row matrix with >>>>>>>>>> the preallocated nonzero pattern, filled with zeros >>>>>>>>>> >>>>>>>>>> 2) Make sure this is the pattern you want >>>>>>>>>> >>>>>>>>>> 3) Run in the debugger with -start_in_debugger >>>>>>>>>> >>>>>>>>>> 4) When you get the error, see >>>>>>>>>> >>>>>>>>>> a) If the (i, j) is one that should be setting a value >>>>>>>>>> >>>>>>>>>> b) Why this (i, j) was not preallocated >>>>>>>>>> >>>>>>>>> >>>>>>>>> Up to 4 (a), it is correct. There is a problem in the way >>>>>>>>> DMDAGetCellPoint and PetscSectionGetOffset works in this case scenario. I >>>>>>>>> will try to explain it below for the case of 4X4 grid. >>>>>>>>> First case: >>>>>>>>> If I set the computational domain to be all the points of the dmda >>>>>>>>> grid, (i.e. isPosInDomain(..,i,j,..) returns true for all i and j in the >>>>>>>>> dmda grid), the program runs fine and does not give any error. >>>>>>>>> >>>>>>>>> Second case: >>>>>>>>> I want the computational domain to be some part of the whole grid. >>>>>>>>> There is a problem in this case. >>>>>>>>> The following test is in a case where, >>>>>>>>> isPosInDomain(..,i,j,..) returns true ONLY for (i,j) pairs of >>>>>>>>> (2,1) (1,2) (2,2) (3,2) and (3,2). The grid with its corresponding point >>>>>>>>> number in the petscsection is shown below: >>>>>>>>> >>>>>>>>> 12 13 14 15 >>>>>>>>> 8 9 10 11 >>>>>>>>> 4 5 6 7 >>>>>>>>> 0 1 2 3 >>>>>>>>> >>>>>>>>> of which 6, 9, 10, 11 and 14 correspond to the (i,j) points that >>>>>>>>> returns true for isPosInDomain(..,i,j,..) >>>>>>>>> MatView gives me the 16-row matrix with the star stencil non-zero >>>>>>>>> structure as expected. >>>>>>>>> The error I get is: new non-zero at (0,2) caused a malloc. >>>>>>>>> >>>>>>>>> This error is for the value of (i,j) = (2,1), i.e. point 6. >>>>>>>>> The loop to set values in the matrix starts as: >>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>> if (isPosInDomain(ctx,i,j)) { >>>>>>>>> ierr = >>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>>>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>>>>>> >>>>>>>>> Now, what I wanted from the beginning is to create the matrix >>>>>>>>> containing the rows corresponding to only those points (i,j) which has true >>>>>>>>> values for isPosInDomain(ctx,i,j), that means in this case 5 row matrix. In >>>>>>>>> the current test, the first (i,j) that reaches DMDAGetCellPoint is (2,1), >>>>>>>>> which gives sets point variable to 6. >>>>>>>>> The line PetscSectionGetOffset(gs,point,&row) sets the value of >>>>>>>>> row to 0. >>>>>>>>> So I think this where the inconsistency lies. >>>>>>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES) expects row >>>>>>>>> variable and col[] array to correspond to (i,j) based on DMDA, but >>>>>>>>> PetsSectionGetOffset gives result based on how we masked away some of the >>>>>>>>> points from dmda grid. Or am I missing something very basic here ? >>>>>>>>> >>>>>>>> >>>>>>>> You are missing something basic. The row is correctly identified as >>>>>>>> 0, and that means the 2 is >>>>>>>> point 10. >>>>>>>> >>>>>>> The problem must be preallocation. First, you should send the result >>>>>>>> of MatView(). >>>>>>>> >>>>>>> >>>>>>> MatView shows that the preallocation and the matrix created is for >>>>>>> the full dmda grid which is probably not what we want here, right ? Here is >>>>>>> the matview result: >>>>>>> >>>>>> >>>>>> Thats right. Did you call >>>>>> >>>>>> DMSetDefaultSection(da, s) >>>>>> >>>>>> after you made the new section and before you called DMGetMatrix()? >>>>>> >>>>> Do you mean before I called DMCreateMatrix() ? I called >>>>> DMSetDefaultSection(da,s) as below: >>>>> >>>>> ... >>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>> // ierr = DMGetMatrix(dm,&A);CHKERRQ(ierr); //It seems DMGetMatrix >>>>> is changed to DMCreateMatrix(). >>>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>>> ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >>>>> >>>> >>>> Yes, this is correct. I do exactly this in the example code and it >>>> works fine. I suspect >>>> you are not creating the section you think you are. You can send the >>>> code if you want >>>> and I will look at it. >>>> >>> >>> Thanks, here is the code: >>> https://github.com/bishesh/petscPoissonIrregular/blob/master/poissonIrregular.cxx >>> It works for 2D and 3D without using PetscSection. (The choice of using >>> Petsc_section is on line 68). Using PetsSection currently it implements >>> only the 2D case with dof = 1. It works when the domain consists of all the >>> cells of the dmda, it works (to check you can uncomment line 129 in >>> isPosInDomain(). But it does not when I mask out some of the cells while >>> creating the petscsection. >>> >> >> Hi Matt, I'm afraid if my last email somehow escaped you. Could you >> please let me know what could be possibly going wrong with my code linked >> above ? >> > > There is a bug in DMCreateMatrix() for DMDA when using a PetscSection. I > am fixing it. I will mail when its done. > Thanks Matt, I've one more application where I think PetscSection might be useful in stokes like equation with staggered grid; I will start another thread for that while waiting for the bug to get fixed in this thread. > > Thanks, > > Matt > > >> >> >>> >>>> Matt >>>> >>>> >>>>> This MatView on A shows (result below) that it still has 16 rows for >>>>> all DMDA points instead of only those petscsection points for which I set a >>>>> non-zero dof. Is it what it should not be doing ? >>>>> >>>>>> Mat Object: 1 MPI processes >>>>>>> >>>>>>> type: seqaij >>>>>>> >>>>>>> row 0: (0, 0) (1, 0) (4, 0) >>>>>>> >>>>>>> row 1: (0, 0) (1, 0) (2, 0) (5, 0) >>>>>>> >>>>>>> row 2: (1, 0) (2, 0) (3, 0) (6, 0) >>>>>>> >>>>>>> row 3: (2, 0) (3, 0) (7, 0) >>>>>>> >>>>>>> row 4: (0, 0) (4, 0) (5, 0) (8, 0) >>>>>>> >>>>>>> row 5: (1, 0) (4, 0) (5, 0) (6, 0) (9, 0) >>>>>>> >>>>>>> row 6: (2, 0) (5, 0) (6, 0) (7, 0) (10, 0) >>>>>>> >>>>>>> row 7: (3, 0) (6, 0) (7, 0) (11, 0) >>>>>>> >>>>>>> row 8: (4, 0) (8, 0) (9, 0) (12, 0) >>>>>>> >>>>>>> row 9: (5, 0) (8, 0) (9, 0) (10, 0) (13, 0) >>>>>>> >>>>>>> row 10: (6, 0) (9, 0) (10, 0) (11, 0) (14, 0) >>>>>>> >>>>>>> row 11: (7, 0) (10, 0) (11, 0) (15, 0) >>>>>>> >>>>>>> row 12: (8, 0) (12, 0) (13, 0) >>>>>>> >>>>>>> row 13: (9, 0) (12, 0) (13, 0) (14, 0) >>>>>>> >>>>>>> row 14: (10, 0) (13, 0) (14, 0) (15, 0) >>>>>>> >>>>>>> row 15: (11, 0) (14, 0) (15, 0) >>>>>>> >>>>>>> [0]PETSC ERROR: --------------------- Error Message >>>>>>> ------------------------------------ >>>>>>> >>>>>>> [0]PETSC ERROR: Argument out of range! >>>>>>> >>>>>>> [0]PETSC ERROR: New nonzero at (0,2) caused a malloc! >>>>>>> >>>>>>> >>>>>>> >>>>> For reference here is the code snippet again: >>>>> ... >>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, &s);CHKERRQ(ierr); >>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, &nC);CHKERRQ(ierr); >>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>> PetscInt point; >>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>> &point);CHKERRQ(ierr); >>>>> ierr = PetscSectionSetDof(s, point, testPoisson.mDof); >>>>> // No. of dofs associated with the point. >>>>> } >>>>> } >>>>> } >>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>> ierr = DMCreateMatrix(dm,&A);CHKERRQ(ierr); >>>>> ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >>>>> >>>>> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >>>>> ierr = KSPSetDM(ksp,dm);CHKERRQ(ierr); >>>>> >>>>> ierr = >>>>> KSPSetComputeOperators(ksp,computeMatrix2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>> ierr = >>>>> KSPSetComputeRHS(ksp,computeRhs2dSection,(void*)&testPoisson);CHKERRQ(ierr); >>>>> >>>>> .... >>>>> The function computeMatrix2dSection: >>>>> >>>>> #undef __FUNCT__ >>>>> #define __FUNCT__ "computeMatrix2dSection" >>>>> PetscErrorCode computeMatrix2dSection(KSP ksp, Mat A, Mat B, >>>>> MatStructure *str, void *context) { >>>>> PoissonModel *ctx = (PoissonModel*)context; >>>>> PetscErrorCode ierr; >>>>> DM da; >>>>> DMDALocalInfo info; >>>>> PetscInt row, col[5]; >>>>> PetscInt dof; >>>>> PetscScalar v[5]; //array to store 5 point stencil for >>>>> each row >>>>> PetscInt num; //non-zero position in the current row >>>>> // PetscScalar dScale = 1; //to scale Dirichlet >>>>> identity rows if needed >>>>> PetscSection gs; //Global Section that >>>>> keeps the grid info and indices >>>>> PetscInt point; //Current point of the >>>>> petscSection >>>>> >>>>> PetscFunctionBeginUser; >>>>> ierr = KSPGetDM(ksp,&da);CHKERRQ(ierr); >>>>> ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); >>>>> ierr = DMGetDefaultGlobalSection(da,&gs);CHKERRQ(ierr); >>>>> ierr = DMCreateMatrix(da,&A);CHKERRQ(ierr); >>>>> // ierr = MatView(A,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); >>>>> for(PetscInt j = info.ys; j>>>> for(PetscInt i = info.xs; i>>>> if (isPosInDomain(ctx,i,j)) { >>>>> ierr = DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr); >>>>> ierr = PetscSectionGetOffset(gs,point,&row); >>>>> ierr = PetscSectionGetDof(gs,point,&dof); >>>>> for(PetscInt cDof = 0; cDof < ctx->mDof; ++cDof) { >>>>> num = 0; >>>>> row+=cDof; >>>>> col[num] = row; //(i,j) position >>>>> v[num++] = -4; >>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>> ierr = >>>>> DMDAGetCellPoint(da,i+1,j,0,&point);CHKERRQ(ierr); >>>>> ierr = >>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>> col[num] += cDof; >>>>> v[num++] = 1; >>>>> } >>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>> ierr = >>>>> DMDAGetCellPoint(da,i-1,j,0,&point);CHKERRQ(ierr); >>>>> ierr = >>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>> col[num] += cDof; >>>>> v[num++] = 1; >>>>> } >>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>> ierr = >>>>> DMDAGetCellPoint(da,i,j+1,0,&point);CHKERRQ(ierr); >>>>> ierr = >>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>> col[num] += cDof; >>>>> v[num++] = 1; >>>>> } >>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>> ierr = >>>>> DMDAGetCellPoint(da,i,j-1,0,&point);CHKERRQ(ierr); >>>>> ierr = >>>>> PetscSectionGetOffset(gs,point,&col[num]); >>>>> col[num] += cDof; >>>>> v[num++] = 1; >>>>> } >>>>> ierr = >>>>> MatSetValues(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>> } >>>>> } >>>>> } >>>>> } >>>>> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >>>>> >>>>> >>>>> >>>>> >>>>>> >>>>>>>> Matt >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> >>>>>>>>>> Matt >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> [0]PETSC ERROR: --------------------- Error Message >>>>>>>>>>> ------------------------------------ >>>>>>>>>>> [0]PETSC ERROR: Argument out of range! >>>>>>>>>>> [0]PETSC ERROR: New nonzero at (0,6) caused a malloc! >>>>>>>>>>> [0]PETSC ERROR: >>>>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>>>> [0]PETSC ERROR: Petsc Development GIT revision: >>>>>>>>>>> 61e5e40bb2c5bf2423e94b71a15fef47e411b0da GIT Date: 2013-10-25 21:47:45 >>>>>>>>>>> -0500 >>>>>>>>>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>>>>>>>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble >>>>>>>>>>> shooting. >>>>>>>>>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>>>>>>>>> [0]PETSC ERROR: >>>>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>>>> [0]PETSC ERROR: build/poissonIrregular on a >>>>>>>>>>> arch-linux2-cxx-debug named edwards by bkhanal Thu Oct 31 19:23:58 2013 >>>>>>>>>>> [0]PETSC ERROR: Libraries linked from >>>>>>>>>>> /home/bkhanal/Documents/softwares/petsc/arch-linux2-cxx-debug/lib >>>>>>>>>>> [0]PETSC ERROR: Configure run at Sat Oct 26 16:35:15 2013 >>>>>>>>>>> [0]PETSC ERROR: Configure options --download-mpich >>>>>>>>>>> -download-f-blas-lapack=1 --download-metis --download-parmetis >>>>>>>>>>> --download-superlu_dist --download-scalapack --download-mumps >>>>>>>>>>> --download-hypre --with-clanguage=cxx >>>>>>>>>>> [0]PETSC ERROR: >>>>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>>>> [0]PETSC ERROR: MatSetValues_SeqAIJ() line 413 in >>>>>>>>>>> src/mat/impls/aij/seq/aij.c >>>>>>>>>>> [0]PETSC ERROR: MatSetValues() line 1130 in >>>>>>>>>>> src/mat/interface/matrix.c >>>>>>>>>>> [0]PETSC ERROR: computeMatrix2dSection() line 318 in >>>>>>>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>>>>>>> [0]PETSC ERROR: KSPSetUp() line 228 in >>>>>>>>>>> src/ksp/ksp/interface/itfunc.c >>>>>>>>>>> [0]PETSC ERROR: KSPSolve() line 399 in >>>>>>>>>>> src/ksp/ksp/interface/itfunc.c >>>>>>>>>>> [0]PETSC ERROR: main() line 598 in >>>>>>>>>>> /user/bkhanal/home/works/poissonIrregular/poissonIrregular.cxx >>>>>>>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>>>>>>> [unset]: aborting job: >>>>>>>>>>> application called MPI_Abort(MPI_COMM_WORLD, 63) - process 0 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> Matt >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> However, now since I do not want the rows in the matrix >>>>>>>>>>>>> corresponding to the points in the grid which do not lie in the >>>>>>>>>>>>> computational domain. This means the size of the matrix will not >>>>>>>>>>>>> necessarily equal the total number of cells in DMDA. So in the following >>>>>>>>>>>>> code: >>>>>>>>>>>>> >>>>>>>>>>>>> ierr = DMDAGetLocalInfo(dm,&info);CHKERRQ(ierr); >>>>>>>>>>>>> ierr = PetscSectionCreate(PETSC_COMM_WORLD, >>>>>>>>>>>>> &s);CHKERRQ(ierr); >>>>>>>>>>>>> ierr = DMDAGetNumCells(dm, NULL, NULL, NULL, >>>>>>>>>>>>> &nC);CHKERRQ(ierr); >>>>>>>>>>>>> ierr = PetscSectionSetChart(s, 0, nC);CHKERRQ(ierr); >>>>>>>>>>>>> for (PetscInt j = info.ys; j < info.ys+info.ym; ++j) { >>>>>>>>>>>>> for (PetscInt i = info.xs; i < info.xs+info.xm; ++i) { >>>>>>>>>>>>> PetscInt point; >>>>>>>>>>>>> if(isPosInDomain(&testPoisson,i,j,0)) { >>>>>>>>>>>>> ierr = DMDAGetCellPoint(dm, i, j, 0, >>>>>>>>>>>>> &point);CHKERRQ(ierr); >>>>>>>>>>>>> ierr = PetscSectionSetDof(s, point, >>>>>>>>>>>>> testPoisson.mDof); // No. of dofs associated with the point. >>>>>>>>>>>>> } >>>>>>>>>>>>> >>>>>>>>>>>>> } >>>>>>>>>>>>> } >>>>>>>>>>>>> ierr = PetscSectionSetUp(s);CHKERRQ(ierr); >>>>>>>>>>>>> ierr = DMSetDefaultSection(dm, s);CHKERRQ(ierr); >>>>>>>>>>>>> ierr = PetscSectionDestroy(&s);CHKERRQ(ierr); >>>>>>>>>>>>> >>>>>>>>>>>>> should I myself compute proper nC (i.e. total no. of points >>>>>>>>>>>>> for which I want the matrix to have a corresponding row) before calling >>>>>>>>>>>>> PetscSectionSetChart() or is the masking out of the points inside the loop >>>>>>>>>>>>> with if(isPosInDomain(&testPoisson,i,j,0)) sufficient ? >>>>>>>>>>>>> And, when you write: >>>>>>>>>>>>> PetscSectionGetOffset(gs, p, &ind); >>>>>>>>>>>>> row = ind < 0 ? -(ind+1) : ind; >>>>>>>>>>>>> >>>>>>>>>>>>> it seems you allow the possibility of getting a -ve ind when >>>>>>>>>>>>> using PetscSectionGetOffset. When would an offset to a particular dof and >>>>>>>>>>>>> point? >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> for(PetscInt j = info.ys; j>>>>>>>>>>>>>> for(PetscInt i = info.xs; i>>>>>>>>>>>>>> ++i) { >>>>>>>>>>>>>>> num = 0; >>>>>>>>>>>>>>> /*ierr = >>>>>>>>>>>>>>> DMDAGetCellPoint(da,i,j,0,&point);CHKERRQ(ierr);*/ /*Get the PetscPoint*/ >>>>>>>>>>>>>>> /*But I did now understand how I would >>>>>>>>>>>>>>> emulate the row.c and col.c info with PetscSection*/ >>>>>>>>>>>>>>> row.i = i; row.j = j; >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Here you would call >>>>>>>>>>>>>> >>>>>>>>>>>>>> ierr = DMDAGetCellPoint(da, i, j, 0, &cell);CHKERRQ(ierr); >>>>>>>>>>>>>> ierr = PetscSectionGetOffset(da, cell, &row);CHKERRQ(ierr); >>>>>>>>>>>>>> ierr = PetscSectionGetDof(da, cell, &dof);CHKERRQ(ierr); >>>>>>>>>>>>>> >>>>>>>>>>>>>> This means that this cell has global rows = [row, row+dof). >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> col[num].i = i; col[num].j = j; >>>>>>>>>>>>>>> if (isPosInDomain(ctx,i,j)) { /*put the >>>>>>>>>>>>>>> operator for only the values for only the points lying in the domain*/ >>>>>>>>>>>>>>> v[num++] = -4; >>>>>>>>>>>>>>> if(isPosInDomain(ctx,i+1,j)) { >>>>>>>>>>>>>>> col[num].i = i+1; col[num].j = j; >>>>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> if(isPosInDomain(ctx,i-1,j)) { >>>>>>>>>>>>>>> col[num].i = i-1; col[num].j = j; >>>>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> if(isPosInDomain(ctx,i,j+1)) { >>>>>>>>>>>>>>> col[num].i = i; col[num].j = j+1; >>>>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> if(isPosInDomain(ctx,i,j-1)) { >>>>>>>>>>>>>>> col[num].i = i; col[num].j = j-1; >>>>>>>>>>>>>>> v[num++] = 1; >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> } else { >>>>>>>>>>>>>>> v[num++] = dScale; /*set Dirichlet >>>>>>>>>>>>>>> identity rows for points in the rectangle but outside the computational >>>>>>>>>>>>>>> domain*/ >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> You do the same thing for the columns, and then call >>>>>>>>>>>>>> >>>>>>>>>>>>>> ierr = MatSetValues(A, dof, rows, num*dof, cols, v, >>>>>>>>>>>>>> INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> ierr = >>>>>>>>>>>>>>> MatSetValuesStencil(A,1,&row,num,col,v,INSERT_VALUES);CHKERRQ(ierr); >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> If not, just put the identity into >>>>>>>>>>>>>>>>>>>>>>>>>> the rows you do not use on the full cube. It will >>>>>>>>>>>>>>>>>>>>>>>>>> not hurt scalability or convergence. >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> In the case of Poisson with Dirichlet condition >>>>>>>>>>>>>>>>>>>>>>>>> this might be the case. But is it always true that having identity rows in >>>>>>>>>>>>>>>>>>>>>>>>> the system matrix will not hurt convergence ? I thought otherwise for the >>>>>>>>>>>>>>>>>>>>>>>>> following reasons: >>>>>>>>>>>>>>>>>>>>>>>>> 1) Having read Jed's answer here : >>>>>>>>>>>>>>>>>>>>>>>>> http://scicomp.stackexchange.com/questions/3426/why-is-pinning-a-point-to-remove-a-null-space-bad/3427#3427 >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Jed is talking about a constraint on a the pressure >>>>>>>>>>>>>>>>>>>>>>>> at a point. This is just decoupling these unknowns from the rest >>>>>>>>>>>>>>>>>>>>>>>> of the problem. >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> 2) Some observation I am getting (but I am still >>>>>>>>>>>>>>>>>>>>>>>>> doing more experiments to confirm) while solving my staggered-grid 3D >>>>>>>>>>>>>>>>>>>>>>>>> stokes flow with schur complement and using -pc_type gamg for A00 matrix. >>>>>>>>>>>>>>>>>>>>>>>>> Putting the identity rows for dirichlet boundaries and for ghost cells >>>>>>>>>>>>>>>>>>>>>>>>> seemed to have effects on its convergence. I'm hoping once I know how to >>>>>>>>>>>>>>>>>>>>>>>>> use PetscSection, I can get rid of using ghost cells method for the >>>>>>>>>>>>>>>>>>>>>>>>> staggered grid and get rid of the identity rows too. >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> It can change the exact iteration, but it does not >>>>>>>>>>>>>>>>>>>>>>>> make the matrix conditioning worse. >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> Anyway please provide me with some pointers so >>>>>>>>>>>>>>>>>>>>>>>>> that I can start trying with petscsection on top of a dmda, in the >>>>>>>>>>>>>>>>>>>>>>>>> beginning for non-staggered case. >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> Matt >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>>>>>>>> Bishesh >>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before >>>>>>>>>>>>>>>>>>>>>>>>>> they begin their experiments is infinitely more interesting than any >>>>>>>>>>>>>>>>>>>>>>>>>> results to which their experiments lead. >>>>>>>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before >>>>>>>>>>>>>>>>>>>>>>>> they begin their experiments is infinitely more interesting than any >>>>>>>>>>>>>>>>>>>>>>>> results to which their experiments lead. >>>>>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> What most experimenters take for granted before they >>>>>>>>>>>>>>>>>> begin their experiments is infinitely more interesting than any results to >>>>>>>>>>>>>>>>>> which their experiments lead. >>>>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>>>> their experiments lead. >>>>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> What most experimenters take for granted before they begin >>>>>>>>>>>> their experiments is infinitely more interesting than any results to which >>>>>>>>>>>> their experiments lead. >>>>>>>>>>>> -- Norbert Wiener >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> What most experimenters take for granted before they begin their >>>>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>>>> experiments lead. >>>>>>>>>> -- Norbert Wiener >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> What most experimenters take for granted before they begin their >>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>> experiments lead. >>>>>>>> -- Norbert Wiener >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> What most experimenters take for granted before they begin their >>>>>> experiments is infinitely more interesting than any results to which their >>>>>> experiments lead. >>>>>> -- Norbert Wiener >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> What most experimenters take for granted before they begin their >>>> experiments is infinitely more interesting than any results to which their >>>> experiments lead. >>>> -- Norbert Wiener >>>> >>> >>> >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From bisheshkh at gmail.com Tue Nov 12 09:14:30 2013 From: bisheshkh at gmail.com (Bishesh Khanal) Date: Tue, 12 Nov 2013 16:14:30 +0100 Subject: [petsc-users] solving stokes-like equation in 3D staggered grid in irregular domain with petscsection Message-ID: Dear all, I have an implementation of stokes flow equation in cuboid using a staggered grid with DMDA. Now I want to solve a problem in a slightly different (but more difficult) situation where I think PetscSection might be useful although I have not used PetscSection that much and I understand that some of its aspects is still under development in petsc. I'm ok to work with the dev version of petsc if what I want to do using petscsection is plausible. Here is my problem description: I have a 3D binary image that partitions a cuboid into domain A and domain B by tagging each cell as 0 or 1. The shape of A and therefore B can be fairly arbitrary. I want to solve the following sets of equations: A. In A: div(mu(grad(u))) - grad(p) = f1 , where f1 depends on the position x. div(u) = f2 , where f2 is non-zero and depends on the position x. --------------------------------------- B. In B: div(mu(grad(u)) = f3, where f3 depends on the position x. ----------------------------------------- u = constant on the cuboid boundaries. (Dirichlet boundaries) p = constant on domain B (if needed as boundary condition for domain A) ------------------------------------------- Now, I was thinking of creating a single matrix M that contains the operators for both A and B domains, and solve a linear system MX = R, where X and R are the solution and Rhs vectors of roughly the sizes 4*nA + 3*nB where, nA => no. of cells in A, nB => no. of cells in B. Some questions: 1. Would PetscSection be useful in this case due to the arbitrary shape of domain A and because of different number of dofs in domain A and domain B ? Is there a simple example that uses petscsection for the staggered grid case ? I would like to be able to implement staggered grid without the use of ghost/fictitious cells. 2. Is it a good idea to create a single matrix corresponding to both the domains A and B (particularly in view of the difficulty it might bring in solving with some preconditioners using schur fieldsplit) ? 3. I'm trying to put both domains in a single matrix to avoid the difficulty I would have if I want to consider only the domain A. In this case I would need a traction free boundary condition on the irregular boundary of domain A, and it seems a bit too challenging for me to incorporate it with the staggered grid. If there is an idea to implement this and if you think this could be more suitable than the approach in 2 above, I would like to learn about that too! Thanks, Bishesh -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Tue Nov 12 12:57:52 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Tue, 12 Nov 2013 11:57:52 -0700 Subject: [petsc-users] Unable to create >4GB sized HDF5 files on Cray XC30 In-Reply-To: <6408526.5lg4qfbu4O@dhcp071> References: <2569883.TWAWmgZMUo@rigel> <5862566.rSSLWvQenN@rigel> <87li1kjh54.fsf@mcs.anl.gov> <6408526.5lg4qfbu4O@dhcp071> Message-ID: <87fvr18n2n.fsf@jedbrown.org> Juha J?ykk? writes: > I'll leave the parameter for others to implement: I don't have a good idea of > how to do that. I believe I can provide a testing environment (but no access > to it, I'm afraid) for those big files, just let me know what kind of test > suite I should create. Not that the old chunking seems to have had much > testing considering how seriously broken it is. ;) Yes, but we like to improve test coverage when we make changes. The new code is more lines doing more complicated (hopefully less wrong) things, so it ought to be tested. The parameter could be controlled by: PetscViewerHDF5SetMaxChunkSize(PetscViewer viewer,size_t max_chunk); >> > + // just in case some libraries do not empty memory for local variables: >> We have to use C89-style comments because of Microsoft. > > What is this, Microsoft's compiler cannot do C99? Is it too new a standard, at > 14 years old? Sadly, Microsoft does not intend to ever implement C99. http://blog.reverberate.org/2012/03/microsoft-please-support-at-least-tiny.html That said, I personally would not have a problem with using those C99 features that Microsoft's C++ compiler supports. That even includes variadic macros (well, a bastardized version, because when they "cherry-picked" it from C99, they implemented the semantics wrong, though that mistake is only triggered in peculiar cases). But unless PETSc as a whole make a decision to allow C99 features, we have to stick to C89. (Public headers should use C89 syntax forever because we don't want to impose a dialect on users.) >> > + >> > + // some requirements for the chunks (assuming no split along time-step) >> > + int ranks; >> Declarations cannot be mixed with statements in C89. > > Microsoft again? I've reverted these *standard* pieces of code into an > obsolete, old standard. I really had no idea there are STILL compilers around > which cannot do C99. Shocking, I know. If only they didn't have so much market share... -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From torquil at gmail.com Tue Nov 12 13:16:15 2013 From: torquil at gmail.com (=?UTF-8?Q?Torquil_Macdonald_S=C3=B8rensen?=) Date: Tue, 12 Nov 2013 20:16:15 +0100 Subject: [petsc-users] Multigrid and interpolation matrices Message-ID: Hi! I'm having some problems getting PCMG to work when manually populating the interpolation matrix, i.e. using MatSetValue to set its values, instead of the use of DMCreateInterpolation, as I have seen in an example. I have to use MatSetValue to populate my PETSc matrices, because I'm obtaining them in a different format from a FEM library. I have reduced my program to a minimal test case (without involving the FEM library), where the system matrix is simply an identity matrix, with two multigrid levels of 8 and 4 degrees of freedom, and a very simple interpolation matrix that can be seen by looking at the code I'm attaching. I'm not including KSPSolve because the problem occurs prior to that, on line 56, which corresponds to where I'm calling KSPSetUp(). I'm actually not certain how to create the matrix that will be passed to PCMGSetInterpolation(). At the moment I'm only running sequentially, but how should this non-square matrix be distributed across the MPI nodes when running in parallel, i.e. what values should I pass to MatSetSizes() for that matrix? In any case, the error message I'm getting (running it with no command line options) is: [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Object is in wrong state! [0]PETSC ERROR: Mat object's type is not set: Argument # 1! [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 [0]PETSC ERROR: See docs/changes/index.html for recent updates. [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [0]PETSC ERROR: See docs/index.html for manual pages. [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: ./multigrid2.elf on a arch-linux2-cxx-debug named enlil.uio.no by tmac Tue Nov 12 19:57:59 2013 [0]PETSC ERROR: Libraries linked from /mn/anatu/cma-u3/tmac/usr/stow/petsc_complex_dbg/lib [0]PETSC ERROR: Configure run at Tue Nov 12 16:27:32 2013 [0]PETSC ERROR: Configure options CPPFLAGS=-I/mn/anatu/cma-u3/tmac/usr/include LDFLAGS=-L/mn/anatu/cma-u3/tmac/usr/lib -L/mn/anatu/cma-u3/tmac/usr/lib64 --prefix=/mn/anatu/cma-u3/tmac/usr/stow/petsc_complex_dbg --with-clanguage=C++ --with-scalar-type=complex --with-shared-libraries=1 --with-debugging=1 --with-superlu=1 --with-superlu-lib=/mn/anatu/cma-u3/tmac/usr/lib/libsuperlu.so --with-superlu-include=/mn/anatu/cma-u3/tmac/usr/include/superlu/ [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: MatGetVecs() line 8131 in /tmp/petsc-3.4.3/src/mat/interface/matrix.c [0]PETSC ERROR: KSPGetVecs() line 929 in /tmp/petsc-3.4.3/src/ksp/ksp/interface/iterativ.c [0]PETSC ERROR: PCSetUp_MG() line 691 in /tmp/petsc-3.4.3/src/ksp/pc/impls/mg/mg.c [0]PETSC ERROR: PCSetUp() line 890 in /tmp/petsc-3.4.3/src/ksp/pc/interface/precon.c [0]PETSC ERROR: KSPSetUp() line 278 in /tmp/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c [0]PETSC ERROR: main() line 56 in "unknowndirectory/"/mn/anatu/cma-u3/tmac/programming/fem/getfem/source/multigrid2.cpp -------------------------------------------------------------------------- MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD with errorcode 73. NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. You may or may not see output from other processes, depending on exactly when Open MPI kills them. -------------------------------------------------------------------------- The full code is: #include "petscksp.h" int main(int argc, char ** argv) { PetscErrorCode ierr = PetscInitialize(&argc, &argv, 0, 0); CHKERRQ(ierr); Vec x; Mat A; PetscInt rstart, rend, nlocal; PetscInt const n = 8; ierr = VecCreate(PETSC_COMM_WORLD, &x); CHKERRQ(ierr); ierr = VecSetSizes(x, PETSC_DECIDE, n); CHKERRQ(ierr); ierr = VecSetFromOptions(x); CHKERRQ(ierr); ierr = VecGetOwnershipRange(x, &rstart, &rend); CHKERRQ(ierr); ierr = VecGetLocalSize(x, &nlocal); CHKERRQ(ierr); ierr = MatCreate(PETSC_COMM_WORLD, &A); CHKERRQ(ierr); ierr = MatSetSizes(A, nlocal, nlocal, n, n); CHKERRQ(ierr); ierr = MatSetFromOptions(A); CHKERRQ(ierr); ierr = MatSetUp(A); CHKERRQ(ierr); for(PetscInt i = rstart; i < rend; ++i) { ierr = MatSetValue(A, i, i, 1.0, INSERT_VALUES); CHKERRQ(ierr); } ierr = MatAssemblyBegin(A, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); ierr = MatAssemblyEnd(A, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); KSP ksp; ierr = KSPCreate(PETSC_COMM_WORLD, &ksp); CHKERRQ(ierr); ierr = KSPSetOperators(ksp, A, A, SAME_NONZERO_PATTERN); CHKERRQ(ierr); ierr = KSPSetType(ksp, KSPFGMRES); CHKERRQ(ierr); PC pc; ierr = KSPGetPC(ksp, &pc); CHKERRQ(ierr); ierr = PCSetType(pc, PCMG); CHKERRQ(ierr); ierr = PCMGSetLevels(pc, 2, 0); CHKERRQ(ierr); Mat interp; ierr = MatCreate(PETSC_COMM_WORLD, &interp); CHKERRQ(ierr); ierr = MatSetSizes(interp, PETSC_DECIDE, PETSC_DECIDE, n, n/2); CHKERRQ(ierr); ierr = MatSetFromOptions(interp); CHKERRQ(ierr); ierr = MatSetUp(interp); CHKERRQ(ierr); ierr = MatGetOwnershipRange(interp, &rstart, &rend); CHKERRQ(ierr); for(PetscInt i = rstart; i < rend; ++i) { ierr = MatSetValue(interp, i, i/2, 1.0, INSERT_VALUES); CHKERRQ(ierr); } ierr = MatAssemblyBegin(interp, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); ierr = MatAssemblyEnd(interp, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); ierr = PCMGSetInterpolation(pc, 1, interp); CHKERRQ(ierr); ierr = MatDestroy(&interp); CHKERRQ(ierr); ierr = KSPSetFromOptions(ksp); CHKERRQ(ierr); ierr = KSPSetUp(ksp); CHKERRQ(ierr); ierr = VecDestroy(&x); CHKERRQ(ierr); ierr = MatDestroy(&A); CHKERRQ(ierr); ierr = KSPDestroy(&ksp); CHKERRQ(ierr); ierr = PetscFinalize(); CHKERRQ(ierr); return 0; } -------------- next part -------------- An HTML attachment was scrubbed... URL: From bsmith at mcs.anl.gov Tue Nov 12 13:27:12 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Tue, 12 Nov 2013 13:27:12 -0600 Subject: [petsc-users] Multigrid and interpolation matrices In-Reply-To: References: Message-ID: <240E25F2-7485-4C22-9D6C-E310945B51AA@mcs.anl.gov> y = A *x The number of local rows in A is the number of local rows in y. The number of local columns in A is the number of local rows in x. To keep it from crashing you need to either 1) provide the coarse grid operator or 2) run with -pc_mg_galerkin to have the coarse grid operator computed from the interpolation you provided and the fine grid operator you provided. Barry On Nov 12, 2013, at 1:16 PM, Torquil Macdonald S?rensen wrote: > Hi! > > I'm having some problems getting PCMG to work when manually populating the interpolation matrix, i.e. using MatSetValue to set its values, instead of the use of DMCreateInterpolation, as I have seen in an example. I have to use MatSetValue to populate my PETSc matrices, because I'm obtaining them in a different format from a FEM library. > > I have reduced my program to a minimal test case (without involving the FEM library), where the system matrix is simply an identity matrix, with two multigrid levels of 8 and 4 degrees of freedom, and a very simple interpolation matrix that can be seen by looking at the code I'm attaching. I'm not including KSPSolve because the problem occurs prior to that, on line 56, which corresponds to where I'm calling KSPSetUp(). > > I'm actually not certain how to create the matrix that will be passed to PCMGSetInterpolation(). At the moment I'm only running sequentially, but how should this non-square matrix be distributed across the MPI nodes when running in parallel, i.e. what values should I pass to MatSetSizes() for that matrix? > > In any case, the error message I'm getting (running it with no command line options) is: > > [0]PETSC ERROR: --------------------- Error Message ------------------------------------ > [0]PETSC ERROR: Object is in wrong state! > [0]PETSC ERROR: Mat object's type is not set: Argument # 1! > [0]PETSC ERROR: ------------------------------------------------------------------------ > [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > [0]PETSC ERROR: See docs/changes/index.html for recent updates. > [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > [0]PETSC ERROR: See docs/index.html for manual pages. > [0]PETSC ERROR: ------------------------------------------------------------------------ > [0]PETSC ERROR: ./multigrid2.elf on a arch-linux2-cxx-debug named enlil.uio.no by tmac Tue Nov 12 19:57:59 2013 > [0]PETSC ERROR: Libraries linked from /mn/anatu/cma-u3/tmac/usr/stow/petsc_complex_dbg/lib > [0]PETSC ERROR: Configure run at Tue Nov 12 16:27:32 2013 > [0]PETSC ERROR: Configure options CPPFLAGS=-I/mn/anatu/cma-u3/tmac/usr/include LDFLAGS=-L/mn/anatu/cma-u3/tmac/usr/lib -L/mn/anatu/cma-u3/tmac/usr/lib64 --prefix=/mn/anatu/cma-u3/tmac/usr/stow/petsc_complex_dbg --with-clanguage=C++ --with-scalar-type=complex --with-shared-libraries=1 --with-debugging=1 --with-superlu=1 --with-superlu-lib=/mn/anatu/cma-u3/tmac/usr/lib/libsuperlu.so --with-superlu-include=/mn/anatu/cma-u3/tmac/usr/include/superlu/ > [0]PETSC ERROR: ------------------------------------------------------------------------ > [0]PETSC ERROR: MatGetVecs() line 8131 in /tmp/petsc-3.4.3/src/mat/interface/matrix.c > [0]PETSC ERROR: KSPGetVecs() line 929 in /tmp/petsc-3.4.3/src/ksp/ksp/interface/iterativ.c > [0]PETSC ERROR: PCSetUp_MG() line 691 in /tmp/petsc-3.4.3/src/ksp/pc/impls/mg/mg.c > [0]PETSC ERROR: PCSetUp() line 890 in /tmp/petsc-3.4.3/src/ksp/pc/interface/precon.c > [0]PETSC ERROR: KSPSetUp() line 278 in /tmp/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c > [0]PETSC ERROR: main() line 56 in "unknowndirectory/"/mn/anatu/cma-u3/tmac/programming/fem/getfem/source/multigrid2.cpp > -------------------------------------------------------------------------- > MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD > with errorcode 73. > > NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. > You may or may not see output from other processes, depending on > exactly when Open MPI kills them. > -------------------------------------------------------------------------- > > The full code is: > > #include "petscksp.h" > > int main(int argc, char ** argv) > { > PetscErrorCode ierr = PetscInitialize(&argc, &argv, 0, 0); CHKERRQ(ierr); > > Vec x; > Mat A; > PetscInt rstart, rend, nlocal; > > PetscInt const n = 8; > > ierr = VecCreate(PETSC_COMM_WORLD, &x); CHKERRQ(ierr); > ierr = VecSetSizes(x, PETSC_DECIDE, n); CHKERRQ(ierr); > ierr = VecSetFromOptions(x); CHKERRQ(ierr); > ierr = VecGetOwnershipRange(x, &rstart, &rend); CHKERRQ(ierr); > ierr = VecGetLocalSize(x, &nlocal); CHKERRQ(ierr); > > ierr = MatCreate(PETSC_COMM_WORLD, &A); CHKERRQ(ierr); > ierr = MatSetSizes(A, nlocal, nlocal, n, n); CHKERRQ(ierr); > ierr = MatSetFromOptions(A); CHKERRQ(ierr); > ierr = MatSetUp(A); CHKERRQ(ierr); > for(PetscInt i = rstart; i < rend; ++i) { > ierr = MatSetValue(A, i, i, 1.0, INSERT_VALUES); CHKERRQ(ierr); > } > ierr = MatAssemblyBegin(A, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); > ierr = MatAssemblyEnd(A, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); > > KSP ksp; > ierr = KSPCreate(PETSC_COMM_WORLD, &ksp); CHKERRQ(ierr); > ierr = KSPSetOperators(ksp, A, A, SAME_NONZERO_PATTERN); CHKERRQ(ierr); > ierr = KSPSetType(ksp, KSPFGMRES); CHKERRQ(ierr); > > PC pc; > ierr = KSPGetPC(ksp, &pc); CHKERRQ(ierr); > ierr = PCSetType(pc, PCMG); CHKERRQ(ierr); > ierr = PCMGSetLevels(pc, 2, 0); CHKERRQ(ierr); > > Mat interp; > ierr = MatCreate(PETSC_COMM_WORLD, &interp); CHKERRQ(ierr); > ierr = MatSetSizes(interp, PETSC_DECIDE, PETSC_DECIDE, n, n/2); CHKERRQ(ierr); > ierr = MatSetFromOptions(interp); CHKERRQ(ierr); > ierr = MatSetUp(interp); CHKERRQ(ierr); > ierr = MatGetOwnershipRange(interp, &rstart, &rend); CHKERRQ(ierr); > for(PetscInt i = rstart; i < rend; ++i) { > ierr = MatSetValue(interp, i, i/2, 1.0, INSERT_VALUES); CHKERRQ(ierr); > } > ierr = MatAssemblyBegin(interp, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); > ierr = MatAssemblyEnd(interp, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); > > ierr = PCMGSetInterpolation(pc, 1, interp); CHKERRQ(ierr); > ierr = MatDestroy(&interp); CHKERRQ(ierr); > > ierr = KSPSetFromOptions(ksp); CHKERRQ(ierr); > ierr = KSPSetUp(ksp); CHKERRQ(ierr); > > ierr = VecDestroy(&x); CHKERRQ(ierr); > ierr = MatDestroy(&A); CHKERRQ(ierr); > ierr = KSPDestroy(&ksp); CHKERRQ(ierr); > ierr = PetscFinalize(); CHKERRQ(ierr); > > return 0; > } From jedbrown at mcs.anl.gov Tue Nov 12 13:43:54 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Tue, 12 Nov 2013 12:43:54 -0700 Subject: [petsc-users] Calculating the PETSc index of a DMComposite vector In-Reply-To: References: <87habicu41.fsf@jedbrown.org> Message-ID: <877gcd8kxx.fsf@jedbrown.org> Anush Krishnan writes: > Around each body point, I choose a small rectangular region of velocity > grid points and perform an interpolation with these values to obtain the > velocity at the body point. How large are these rectangular regions? > This can be represented as: [C][u] = [b] > > where [C] contains the interpolation coefficients, [u] contains velocity > values from the small rectangular region and [b] is the velocity at the > body point. These matrices can be expanded to include the entire velocity > field and all the body points. What are you doing with the C once it is built? Depending on the distribution and access pattern, it could make sense to assemble C^T instead, but it likely doesn't matter. > Currently, my fix is to loop over all the velocity grid points, check if > they come under the influence of any of the body points, and > correspondingly set the interpolation coefficient value in the matrix. But > this procedure requires a double loop. I would loop over the body points and classify them as being "within range" of the grid points that you own. Ignore the points that are out of range. From the interpolation patch size, you know analytically which owned grid points need to contribute values. Call MatSetValues() or MatSetValuesLocal() for that partial row of C. (Don't worry about whether you own the body point or not, and only set entries for velocity points that you own---because the other process will be setting the entries for the points they own.) -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From m.guterres at gmail.com Tue Nov 12 13:56:53 2013 From: m.guterres at gmail.com (Marcelo Xavier Guterres) Date: Tue, 12 Nov 2013 17:56:53 -0200 Subject: [petsc-users] how to sort a vector petsc In-Reply-To: References: Message-ID: <52828805.9090605@gmail.com> Dears, how to sort a vector petsc ? Marcelo Guterres Brazil From knepley at gmail.com Tue Nov 12 13:58:29 2013 From: knepley at gmail.com (Matthew Knepley) Date: Tue, 12 Nov 2013 13:58:29 -0600 Subject: [petsc-users] how to sort a vector petsc In-Reply-To: <52828805.9090605@gmail.com> References: <52828805.9090605@gmail.com> Message-ID: On Tue, Nov 12, 2013 at 1:56 PM, Marcelo Xavier Guterres < m.guterres at gmail.com> wrote: > Dears, > > how to sort a vector petsc ? > We do not have parallel sort, but we do have PetscSortInt/Real. Matt > Marcelo Guterres > Brazil > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From pengxwang at hotmail.com Tue Nov 12 14:14:47 2013 From: pengxwang at hotmail.com (Roc Wang) Date: Tue, 12 Nov 2013 14:14:47 -0600 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: <87zjpcgj2l.fsf@jedbrown.org> References: , <87zjpcgj2l.fsf@jedbrown.org> Message-ID: Thanks Jed, I have questions about load balance and PC type below. > From: jedbrown at mcs.anl.gov > To: pengxwang at hotmail.com; petsc-users at mcs.anl.gov > Subject: Re: [petsc-users] approaches to reduce computing time > Date: Sun, 10 Nov 2013 12:20:18 -0700 > > Roc Wang writes: > > > Hi all, > > > > I am trying to minimize the computing time to solve a large sparse matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to reduce the computing time from two directions: 1 finding a Pre-conditioner to reduce the number of iterations which reduces the time numerically, 2 requesting more cores. > > > > ----For the first method, I tried several methods: > > 1 default KSP and PC, > > 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type jacobi, > > 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, > > 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, > > 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm (PCASM) > > > > The iterations and timing is like the following with 128 cores requested: > > case# iter timing (s) > > 1 1436 816 > > 2 3 12658 > > 3 1069 669.64 > > 4 872 768.12 > > 5 927 513.14 > > > > It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment can help to reduce the iterations but not the timing (comparing case 3 and 4). Second, the PCASM helps a lot. Although the second option is able to reduce iterations, the timing increases very much. Is it because more operations are needed in the PC? > > > > My questions here are: 1. Which direction should I take to select > > -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger > > restart with large augment is better or larger restart with smaller > > augment is better? > > Look at the -log_summary. By increasing the restart, the work in > KSPGMRESOrthog will increase linearly, but the number of iterations > might decrease enough to compensate. There is no general rule here > since it depends on the relative expense of operations for your problem > on your machine. > > > ----For the second method, I tried with -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different number of cores. I found the speedup ratio increases slowly when more than 32 to 64 cores are requested. I searched the milling list archives and found that I am very likely running into the memory bandwidth bottleneck. http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: > > > > # of cores iter timing > > 1 923 19541.83 > > 4 929 5897.06 > > 8 932 4854.72 > > 16 924 1494.33 > > 32 924 1480.88 > > 64 928 686.89 > > 128 927 627.33 > > 256 926 552.93 > > The bandwidth issue has more to do with using multiple cores within a > node rather than between nodes. Likely the above is a load balancing > problem or bad communication. I use DM to manage the distributed data. The DM was created by calling DMDACreate3d() and let PETSc decide the local number of nodes in each direction. To my understand the load of each core is determined at this stage. If the load balance is done when DMDACreate3d() is called and use PETSC_DECIDE option? Or how should make the load balanced after DM is created? > > > My question here is: Is there any other PC can help on both reducing iterations and increasing scalability? Thanks. > > Always send -log_summary with questions like this, but algebraic multigrid is a good place to start. Please take a look at the attached log file, they are for 128 cores and 256 cores, respectively. Based on the log files, what should be done to increase the scalability? Thanks. -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- An embedded and charset-unspecified text was scrubbed... Name: PCASM_m40-k10_p256_5294.log URL: -------------- next part -------------- An embedded and charset-unspecified text was scrubbed... Name: PCASM_m40-k10_p128_5295.log URL: From knepley at gmail.com Tue Nov 12 14:22:35 2013 From: knepley at gmail.com (Matthew Knepley) Date: Tue, 12 Nov 2013 14:22:35 -0600 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: References: <87zjpcgj2l.fsf@jedbrown.org> Message-ID: On Tue, Nov 12, 2013 at 2:14 PM, Roc Wang wrote: > Thanks Jed, > > I have questions about load balance and PC type below. > > > From: jedbrown at mcs.anl.gov > > To: pengxwang at hotmail.com; petsc-users at mcs.anl.gov > > Subject: Re: [petsc-users] approaches to reduce computing time > > Date: Sun, 10 Nov 2013 12:20:18 -0700 > > > > Roc Wang writes: > > > > > Hi all, > > > > > > I am trying to minimize the computing time to solve a large sparse > matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to > reduce the computing time from two directions: 1 finding a Pre-conditioner > to reduce the number of iterations which reduces the time numerically, 2 > requesting more cores. > > > > > > ----For the first method, I tried several methods: > > > 1 default KSP and PC, > > > 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type > jacobi, > > > 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, > > > 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, > > > 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 > -pc_type asm (PCASM) > > > > > > The iterations and timing is like the following with 128 cores > requested: > > > case# iter timing (s) > > > 1 1436 816 > > > 2 3 12658 > > > 3 1069 669.64 > > > 4 872 768.12 > > > 5 927 513.14 > > > > > > It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment > can help to reduce the iterations but not the timing (comparing case 3 and > 4). Second, the PCASM helps a lot. Although the second option is able to > reduce iterations, the timing increases very much. Is it because more > operations are needed in the PC? > > > > > > My questions here are: 1. Which direction should I take to select > > > -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger > > > restart with large augment is better or larger restart with smaller > > > augment is better? > > > > Look at the -log_summary. By increasing the restart, the work in > > KSPGMRESOrthog will increase linearly, but the number of iterations > > might decrease enough to compensate. There is no general rule here > > since it depends on the relative expense of operations for your problem > > on your machine. > > > > > ----For the second method, I tried with -ksp_type lgmres > -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different > number of cores. I found the speedup ratio increases slowly when more than > 32 to 64 cores are requested. I searched the milling list archives and > found that I am very likely running into the memory bandwidth bottleneck. > http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: > > > > > > # of cores iter timing > > > 1 923 19541.83 > > > 4 929 5897.06 > > > 8 932 4854.72 > > > 16 924 1494.33 > > > 32 924 1480.88 > > > 64 928 686.89 > > > 128 927 627.33 > > > 256 926 552.93 > > > > The bandwidth issue has more to do with using multiple cores within a > > node rather than between nodes. Likely the above is a load balancing > > problem or bad communication. > > I use DM to manage the distributed data. The DM was created by calling > DMDACreate3d() and let PETSc decide the local number of nodes in each > direction. To my understand the load of each core is determined at this > stage. If the load balance is done when DMDACreate3d() is called and use > PETSC_DECIDE option? Or how should make the load balanced after DM is > created? > We do not have a way to do fine-grained load balancing for the DMDA since it is intended for very simple topologies. You can see if it is load imbalance from the division by running with a cube that is evenly divisible with a cube number of processes. Matt > > > > > My question here is: Is there any other PC can help on both reducing > iterations and increasing scalability? Thanks. > > > > Always send -log_summary with questions like this, but algebraic > multigrid is a good place to start. > > Please take a look at the attached log file, they are for 128 cores and > 256 cores, respectively. Based on the log files, what should be done to > increase the scalability? Thanks. > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From torquil at gmail.com Tue Nov 12 14:32:32 2013 From: torquil at gmail.com (=?ISO-8859-1?Q?Torquil_Macdonald_S=F8rensen?=) Date: Tue, 12 Nov 2013 21:32:32 +0100 Subject: [petsc-users] Multigrid and interpolation matrices In-Reply-To: <240E25F2-7485-4C22-9D6C-E310945B51AA@mcs.anl.gov> References: <240E25F2-7485-4C22-9D6C-E310945B51AA@mcs.anl.gov> Message-ID: <52829060.6080706@gmail.com> Thanks! For some reason I was thinking that -pc_mg_galerkin was a default, which it is not as can be seen from the -help output. Now I have hardcoded it, and it works. I'll try to apply that rows/cols logic to the distributed interpolation matrices for the parallel case, after I have verified that the problem is solved correctly in the sequential case. Right now I only know that it runs without errors. Best regards Torquil S?rensen On 12/11/13 20:27, Barry Smith wrote: > y = A *x > > The number of local rows in A is the number of local rows in y. The number of local columns in A is the number of local rows in x. > > To keep it from crashing you need to either > > 1) provide the coarse grid operator or > > 2) run with -pc_mg_galerkin to have the coarse grid operator computed from the interpolation you provided and the fine grid operator you provided. > > Barry > > On Nov 12, 2013, at 1:16 PM, Torquil Macdonald S?rensen wrote: > >> Hi! >> >> I'm having some problems getting PCMG to work when manually populating the interpolation matrix, i.e. using MatSetValue to set its values, instead of the use of DMCreateInterpolation, as I have seen in an example. I have to use MatSetValue to populate my PETSc matrices, because I'm obtaining them in a different format from a FEM library. >> >> I have reduced my program to a minimal test case (without involving the FEM library), where the system matrix is simply an identity matrix, with two multigrid levels of 8 and 4 degrees of freedom, and a very simple interpolation matrix that can be seen by looking at the code I'm attaching. I'm not including KSPSolve because the problem occurs prior to that, on line 56, which corresponds to where I'm calling KSPSetUp(). >> >> I'm actually not certain how to create the matrix that will be passed to PCMGSetInterpolation(). At the moment I'm only running sequentially, but how should this non-square matrix be distributed across the MPI nodes when running in parallel, i.e. what values should I pass to MatSetSizes() for that matrix? >> >> In any case, the error message I'm getting (running it with no command line options) is: >> >> [0]PETSC ERROR: --------------------- Error Message ------------------------------------ >> [0]PETSC ERROR: Object is in wrong state! >> [0]PETSC ERROR: Mat object's type is not set: Argument # 1! >> [0]PETSC ERROR: ------------------------------------------------------------------------ >> [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 >> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >> [0]PETSC ERROR: See docs/index.html for manual pages. >> [0]PETSC ERROR: ------------------------------------------------------------------------ >> [0]PETSC ERROR: ./multigrid2.elf on a arch-linux2-cxx-debug named enlil.uio.no by tmac Tue Nov 12 19:57:59 2013 >> [0]PETSC ERROR: Libraries linked from /mn/anatu/cma-u3/tmac/usr/stow/petsc_complex_dbg/lib >> [0]PETSC ERROR: Configure run at Tue Nov 12 16:27:32 2013 >> [0]PETSC ERROR: Configure options CPPFLAGS=-I/mn/anatu/cma-u3/tmac/usr/include LDFLAGS=-L/mn/anatu/cma-u3/tmac/usr/lib -L/mn/anatu/cma-u3/tmac/usr/lib64 --prefix=/mn/anatu/cma-u3/tmac/usr/stow/petsc_complex_dbg --with-clanguage=C++ --with-scalar-type=complex --with-shared-libraries=1 --with-debugging=1 --with-superlu=1 --with-superlu-lib=/mn/anatu/cma-u3/tmac/usr/lib/libsuperlu.so --with-superlu-include=/mn/anatu/cma-u3/tmac/usr/include/superlu/ >> [0]PETSC ERROR: ------------------------------------------------------------------------ >> [0]PETSC ERROR: MatGetVecs() line 8131 in /tmp/petsc-3.4.3/src/mat/interface/matrix.c >> [0]PETSC ERROR: KSPGetVecs() line 929 in /tmp/petsc-3.4.3/src/ksp/ksp/interface/iterativ.c >> [0]PETSC ERROR: PCSetUp_MG() line 691 in /tmp/petsc-3.4.3/src/ksp/pc/impls/mg/mg.c >> [0]PETSC ERROR: PCSetUp() line 890 in /tmp/petsc-3.4.3/src/ksp/pc/interface/precon.c >> [0]PETSC ERROR: KSPSetUp() line 278 in /tmp/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >> [0]PETSC ERROR: main() line 56 in "unknowndirectory/"/mn/anatu/cma-u3/tmac/programming/fem/getfem/source/multigrid2.cpp >> -------------------------------------------------------------------------- >> MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD >> with errorcode 73. >> >> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. >> You may or may not see output from other processes, depending on >> exactly when Open MPI kills them. >> -------------------------------------------------------------------------- >> >> The full code is: >> >> #include "petscksp.h" >> >> int main(int argc, char ** argv) >> { >> PetscErrorCode ierr = PetscInitialize(&argc, &argv, 0, 0); CHKERRQ(ierr); >> >> Vec x; >> Mat A; >> PetscInt rstart, rend, nlocal; >> >> PetscInt const n = 8; >> >> ierr = VecCreate(PETSC_COMM_WORLD, &x); CHKERRQ(ierr); >> ierr = VecSetSizes(x, PETSC_DECIDE, n); CHKERRQ(ierr); >> ierr = VecSetFromOptions(x); CHKERRQ(ierr); >> ierr = VecGetOwnershipRange(x, &rstart, &rend); CHKERRQ(ierr); >> ierr = VecGetLocalSize(x, &nlocal); CHKERRQ(ierr); >> >> ierr = MatCreate(PETSC_COMM_WORLD, &A); CHKERRQ(ierr); >> ierr = MatSetSizes(A, nlocal, nlocal, n, n); CHKERRQ(ierr); >> ierr = MatSetFromOptions(A); CHKERRQ(ierr); >> ierr = MatSetUp(A); CHKERRQ(ierr); >> for(PetscInt i = rstart; i < rend; ++i) { >> ierr = MatSetValue(A, i, i, 1.0, INSERT_VALUES); CHKERRQ(ierr); >> } >> ierr = MatAssemblyBegin(A, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); >> ierr = MatAssemblyEnd(A, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); >> >> KSP ksp; >> ierr = KSPCreate(PETSC_COMM_WORLD, &ksp); CHKERRQ(ierr); >> ierr = KSPSetOperators(ksp, A, A, SAME_NONZERO_PATTERN); CHKERRQ(ierr); >> ierr = KSPSetType(ksp, KSPFGMRES); CHKERRQ(ierr); >> >> PC pc; >> ierr = KSPGetPC(ksp, &pc); CHKERRQ(ierr); >> ierr = PCSetType(pc, PCMG); CHKERRQ(ierr); >> ierr = PCMGSetLevels(pc, 2, 0); CHKERRQ(ierr); >> >> Mat interp; >> ierr = MatCreate(PETSC_COMM_WORLD, &interp); CHKERRQ(ierr); >> ierr = MatSetSizes(interp, PETSC_DECIDE, PETSC_DECIDE, n, n/2); CHKERRQ(ierr); >> ierr = MatSetFromOptions(interp); CHKERRQ(ierr); >> ierr = MatSetUp(interp); CHKERRQ(ierr); >> ierr = MatGetOwnershipRange(interp, &rstart, &rend); CHKERRQ(ierr); >> for(PetscInt i = rstart; i < rend; ++i) { >> ierr = MatSetValue(interp, i, i/2, 1.0, INSERT_VALUES); CHKERRQ(ierr); >> } >> ierr = MatAssemblyBegin(interp, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); >> ierr = MatAssemblyEnd(interp, MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); >> >> ierr = PCMGSetInterpolation(pc, 1, interp); CHKERRQ(ierr); >> ierr = MatDestroy(&interp); CHKERRQ(ierr); >> >> ierr = KSPSetFromOptions(ksp); CHKERRQ(ierr); >> ierr = KSPSetUp(ksp); CHKERRQ(ierr); >> >> ierr = VecDestroy(&x); CHKERRQ(ierr); >> ierr = MatDestroy(&A); CHKERRQ(ierr); >> ierr = KSPDestroy(&ksp); CHKERRQ(ierr); >> ierr = PetscFinalize(); CHKERRQ(ierr); >> >> return 0; >> } From pengxwang at hotmail.com Tue Nov 12 14:48:46 2013 From: pengxwang at hotmail.com (Roc Wang) Date: Tue, 12 Nov 2013 14:48:46 -0600 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: References: , <87zjpcgj2l.fsf@jedbrown.org>, , Message-ID: Date: Tue, 12 Nov 2013 14:22:35 -0600 Subject: Re: [petsc-users] approaches to reduce computing time From: knepley at gmail.com To: pengxwang at hotmail.com CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov On Tue, Nov 12, 2013 at 2:14 PM, Roc Wang wrote: Thanks Jed, I have questions about load balance and PC type below. > From: jedbrown at mcs.anl.gov > To: pengxwang at hotmail.com; petsc-users at mcs.anl.gov > Subject: Re: [petsc-users] approaches to reduce computing time > Date: Sun, 10 Nov 2013 12:20:18 -0700 > > Roc Wang writes: > > > Hi all, > > > > I am trying to minimize the computing time to solve a large sparse matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to reduce the computing time from two directions: 1 finding a Pre-conditioner to reduce the number of iterations which reduces the time numerically, 2 requesting more cores. > > > > ----For the first method, I tried several methods: > > 1 default KSP and PC, > > 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type jacobi, > > 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, > > 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, > > 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm (PCASM) > > > > The iterations and timing is like the following with 128 cores requested: > > case# iter timing (s) > > 1 1436 816 > > 2 3 12658 > > 3 1069 669.64 > > 4 872 768.12 > > 5 927 513.14 > > > > It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment can help to reduce the iterations but not the timing (comparing case 3 and 4). Second, the PCASM helps a lot. Although the second option is able to reduce iterations, the timing increases very much. Is it because more operations are needed in the PC? > > > > My questions here are: 1. Which direction should I take to select > > -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger > > restart with large augment is better or larger restart with smaller > > augment is better? > > Look at the -log_summary. By increasing the restart, the work in > KSPGMRESOrthog will increase linearly, but the number of iterations > might decrease enough to compensate. There is no general rule here > since it depends on the relative expense of operations for your problem > on your machine. > > > ----For the second method, I tried with -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different number of cores. I found the speedup ratio increases slowly when more than 32 to 64 cores are requested. I searched the milling list archives and found that I am very likely running into the memory bandwidth bottleneck. http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: > > > > # of cores iter timing > > 1 923 19541.83 > > 4 929 5897.06 > > 8 932 4854.72 > > 16 924 1494.33 > > 32 924 1480.88 > > 64 928 686.89 > > 128 927 627.33 > > 256 926 552.93 > > The bandwidth issue has more to do with using multiple cores within a > node rather than between nodes. Likely the above is a load balancing > problem or bad communication. I use DM to manage the distributed data. The DM was created by calling DMDACreate3d() and let PETSc decide the local number of nodes in each direction. To my understand the load of each core is determined at this stage. If the load balance is done when DMDACreate3d() is called and use PETSC_DECIDE option? Or how should make the load balanced after DM is created? We do not have a way to do fine-grained load balancing for the DMDA since it is intended for very simple topologies. You can seeif it is load imbalance from the division by running with a cube that is evenly divisible with a cube number of processes. Matt So, I have nothing to do to make the load balanced if I use DMDA? Would you please take a look at the attached log summary files and give me some suggestions on how to improve the speedup ratio? Thanks. > > > My question here is: Is there any other PC can help on both reducing iterations and increasing scalability? Thanks. > > Always send -log_summary with questions like this, but algebraic multigrid is a good place to start. Please take a look at the attached log file, they are for 128 cores and 256 cores, respectively. Based on the log files, what should be done to increase the scalability? Thanks. -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- An embedded and charset-unspecified text was scrubbed... Name: PCASM_m40-k10_p128_5295.log URL: -------------- next part -------------- An embedded and charset-unspecified text was scrubbed... Name: PCASM_m40-k10_p256_5294.log URL: From knepley at gmail.com Tue Nov 12 14:59:30 2013 From: knepley at gmail.com (Matthew Knepley) Date: Tue, 12 Nov 2013 14:59:30 -0600 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: References: <87zjpcgj2l.fsf@jedbrown.org> Message-ID: On Tue, Nov 12, 2013 at 2:48 PM, Roc Wang wrote: > > > ------------------------------ > Date: Tue, 12 Nov 2013 14:22:35 -0600 > Subject: Re: [petsc-users] approaches to reduce computing time > From: knepley at gmail.com > To: pengxwang at hotmail.com > CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov > > On Tue, Nov 12, 2013 at 2:14 PM, Roc Wang wrote: > > Thanks Jed, > > I have questions about load balance and PC type below. > > > From: jedbrown at mcs.anl.gov > > To: pengxwang at hotmail.com; petsc-users at mcs.anl.gov > > Subject: Re: [petsc-users] approaches to reduce computing time > > Date: Sun, 10 Nov 2013 12:20:18 -0700 > > > > Roc Wang writes: > > > > > Hi all, > > > > > > I am trying to minimize the computing time to solve a large sparse > matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to > reduce the computing time from two directions: 1 finding a Pre-conditioner > to reduce the number of iterations which reduces the time numerically, 2 > requesting more cores. > > > > > > ----For the first method, I tried several methods: > > > 1 default KSP and PC, > > > 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type > jacobi, > > > 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, > > > 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, > > > 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 > -pc_type asm (PCASM) > > > > > > The iterations and timing is like the following with 128 cores > requested: > > > case# iter timing (s) > > > 1 1436 816 > > > 2 3 12658 > > > 3 1069 669.64 > > > 4 872 768.12 > > > 5 927 513.14 > > > > > > It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment > can help to reduce the iterations but not the timing (comparing case 3 and > 4). Second, the PCASM helps a lot. Although the second option is able to > reduce iterations, the timing increases very much. Is it because more > operations are needed in the PC? > > > > > > My questions here are: 1. Which direction should I take to select > > > -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger > > > restart with large augment is better or larger restart with smaller > > > augment is better? > > > > Look at the -log_summary. By increasing the restart, the work in > > KSPGMRESOrthog will increase linearly, but the number of iterations > > might decrease enough to compensate. There is no general rule here > > since it depends on the relative expense of operations for your problem > > on your machine. > > > > > ----For the second method, I tried with -ksp_type lgmres > -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different > number of cores. I found the speedup ratio increases slowly when more than > 32 to 64 cores are requested. I searched the milling list archives and > found that I am very likely running into the memory bandwidth bottleneck. > http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: > > > > > > # of cores iter timing > > > 1 923 19541.83 > > > 4 929 5897.06 > > > 8 932 4854.72 > > > 16 924 1494.33 > > > 32 924 1480.88 > > > 64 928 686.89 > > > 128 927 627.33 > > > 256 926 552.93 > > > > The bandwidth issue has more to do with using multiple cores within a > > node rather than between nodes. Likely the above is a load balancing > > problem or bad communication. > > I use DM to manage the distributed data. The DM was created by calling > DMDACreate3d() and let PETSc decide the local number of nodes in each > direction. To my understand the load of each core is determined at this > stage. If the load balance is done when DMDACreate3d() is called and use > PETSC_DECIDE option? Or how should make the load balanced after DM is > created? > > > We do not have a way to do fine-grained load balancing for the DMDA since > it is intended for very simple topologies. You can see > if it is load imbalance from the division by running with a cube that is > evenly divisible with a cube number of processes. > > Matt > > So, I have nothing to do to make the load balanced if I use DMDA? Would > you please take a look at the attached log summary files and give me some > suggestions on how to improve the speedup ratio? Thanks. > Please try what I suggested above. And it looks like there is a little load imbalance VecAXPY 234 1.0 1.0124e+00 3.4 1.26e+08 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 15290 VecAXPY 234 1.0 4.2862e-01 3.6 6.37e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 36115 although it is not limiting the speedup. The time imbalance is really strange. I am guessing other jobs are running on this machine. Matt > > > > > My question here is: Is there any other PC can help on both reducing > iterations and increasing scalability? Thanks. > > > > Always send -log_summary with questions like this, but algebraic > multigrid is a good place to start. > > Please take a look at the attached log file, they are for 128 cores and > 256 cores, respectively. Based on the log files, what should be done to > increase the scalability? Thanks. > > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From pengxwang at hotmail.com Tue Nov 12 15:22:38 2013 From: pengxwang at hotmail.com (Roc Wang) Date: Tue, 12 Nov 2013 15:22:38 -0600 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: References: , <87zjpcgj2l.fsf@jedbrown.org>, , , , Message-ID: Date: Tue, 12 Nov 2013 14:59:30 -0600 Subject: Re: [petsc-users] approaches to reduce computing time From: knepley at gmail.com To: pengxwang at hotmail.com CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov On Tue, Nov 12, 2013 at 2:48 PM, Roc Wang wrote: Date: Tue, 12 Nov 2013 14:22:35 -0600 Subject: Re: [petsc-users] approaches to reduce computing time From: knepley at gmail.com To: pengxwang at hotmail.com CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov On Tue, Nov 12, 2013 at 2:14 PM, Roc Wang wrote: Thanks Jed, I have questions about load balance and PC type below. > From: jedbrown at mcs.anl.gov > To: pengxwang at hotmail.com; petsc-users at mcs.anl.gov > Subject: Re: [petsc-users] approaches to reduce computing time > Date: Sun, 10 Nov 2013 12:20:18 -0700 > > Roc Wang writes: > > > Hi all, > > > > I am trying to minimize the computing time to solve a large sparse matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to reduce the computing time from two directions: 1 finding a Pre-conditioner to reduce the number of iterations which reduces the time numerically, 2 requesting more cores. > > > > ----For the first method, I tried several methods: > > 1 default KSP and PC, > > 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type jacobi, > > 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, > > 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, > > 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm (PCASM) > > > > The iterations and timing is like the following with 128 cores requested: > > case# iter timing (s) > > 1 1436 816 > > 2 3 12658 > > 3 1069 669.64 > > 4 872 768.12 > > 5 927 513.14 > > > > It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment can help to reduce the iterations but not the timing (comparing case 3 and 4). Second, the PCASM helps a lot. Although the second option is able to reduce iterations, the timing increases very much. Is it because more operations are needed in the PC? > > > > My questions here are: 1. Which direction should I take to select > > -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger > > restart with large augment is better or larger restart with smaller > > augment is better? > > Look at the -log_summary. By increasing the restart, the work in > KSPGMRESOrthog will increase linearly, but the number of iterations > might decrease enough to compensate. There is no general rule here > since it depends on the relative expense of operations for your problem > on your machine. > > > ----For the second method, I tried with -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different number of cores. I found the speedup ratio increases slowly when more than 32 to 64 cores are requested. I searched the milling list archives and found that I am very likely running into the memory bandwidth bottleneck. http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: > > > > # of cores iter timing > > 1 923 19541.83 > > 4 929 5897.06 > > 8 932 4854.72 > > 16 924 1494.33 > > 32 924 1480.88 > > 64 928 686.89 > > 128 927 627.33 > > 256 926 552.93 > > The bandwidth issue has more to do with using multiple cores within a > node rather than between nodes. Likely the above is a load balancing > problem or bad communication. I use DM to manage the distributed data. The DM was created by calling DMDACreate3d() and let PETSc decide the local number of nodes in each direction. To my understand the load of each core is determined at this stage. If the load balance is done when DMDACreate3d() is called and use PETSC_DECIDE option? Or how should make the load balanced after DM is created? We do not have a way to do fine-grained load balancing for the DMDA since it is intended for very simple topologies. You can seeif it is load imbalance from the division by running with a cube that is evenly divisible with a cube number of processes. Matt So, I have nothing to do to make the load balanced if I use DMDA? Would you please take a look at the attached log summary files and give me some suggestions on how to improve the speedup ratio? Thanks. Please try what I suggested above. And it looks like there is a little load imbalance Roc----So if the domain is a cube, then the number of the processors is better to be like 2^3=8, 3^3=9, 4^4 =16, and so on, right? I am also wondering whether the physical boundary type effects the load balance? Since freed node, Dirichlet node and Neumann node has different number of neighbors? VecAXPY 234 1.0 1.0124e+00 3.4 1.26e+08 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 15290 VecAXPY 234 1.0 4.2862e-01 3.6 6.37e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 36115 although it is not limiting the speedup. The time imbalance is really strange. I am guessing other jobs are running on this machine. Roc----The code was run a cluster. There should be other jobs were running. Do you mean those jobs affect the load balance of my job or speed of the cluster? I am just trying to improve the scalability of the code, but really don't know what's the reason that the speedup ratio decreases so quickly? Thanks. Matt > > > My question here is: Is there any other PC can help on both reducing iterations and increasing scalability? Thanks. > > Always send -log_summary with questions like this, but algebraic multigrid is a good place to start. Please take a look at the attached log file, they are for 128 cores and 256 cores, respectively. Based on the log files, what should be done to increase the scalability? Thanks. -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Tue Nov 12 15:34:16 2013 From: knepley at gmail.com (Matthew Knepley) Date: Tue, 12 Nov 2013 15:34:16 -0600 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: References: <87zjpcgj2l.fsf@jedbrown.org> Message-ID: On Tue, Nov 12, 2013 at 3:22 PM, Roc Wang wrote: > > > ------------------------------ > Date: Tue, 12 Nov 2013 14:59:30 -0600 > Subject: Re: [petsc-users] approaches to reduce computing time > From: knepley at gmail.com > To: pengxwang at hotmail.com > CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov > > On Tue, Nov 12, 2013 at 2:48 PM, Roc Wang wrote: > > > > ------------------------------ > Date: Tue, 12 Nov 2013 14:22:35 -0600 > Subject: Re: [petsc-users] approaches to reduce computing time > From: knepley at gmail.com > To: pengxwang at hotmail.com > CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov > > On Tue, Nov 12, 2013 at 2:14 PM, Roc Wang wrote: > > Thanks Jed, > > I have questions about load balance and PC type below. > > > From: jedbrown at mcs.anl.gov > > To: pengxwang at hotmail.com; petsc-users at mcs.anl.gov > > Subject: Re: [petsc-users] approaches to reduce computing time > > Date: Sun, 10 Nov 2013 12:20:18 -0700 > > > > Roc Wang writes: > > > > > Hi all, > > > > > > I am trying to minimize the computing time to solve a large sparse > matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to > reduce the computing time from two directions: 1 finding a Pre-conditioner > to reduce the number of iterations which reduces the time numerically, 2 > requesting more cores. > > > > > > ----For the first method, I tried several methods: > > > 1 default KSP and PC, > > > 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type > jacobi, > > > 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, > > > 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, > > > 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 > -pc_type asm (PCASM) > > > > > > The iterations and timing is like the following with 128 cores > requested: > > > case# iter timing (s) > > > 1 1436 816 > > > 2 3 12658 > > > 3 1069 669.64 > > > 4 872 768.12 > > > 5 927 513.14 > > > > > > It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment > can help to reduce the iterations but not the timing (comparing case 3 and > 4). Second, the PCASM helps a lot. Although the second option is able to > reduce iterations, the timing increases very much. Is it because more > operations are needed in the PC? > > > > > > My questions here are: 1. Which direction should I take to select > > > -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger > > > restart with large augment is better or larger restart with smaller > > > augment is better? > > > > Look at the -log_summary. By increasing the restart, the work in > > KSPGMRESOrthog will increase linearly, but the number of iterations > > might decrease enough to compensate. There is no general rule here > > since it depends on the relative expense of operations for your problem > > on your machine. > > > > > ----For the second method, I tried with -ksp_type lgmres > -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different > number of cores. I found the speedup ratio increases slowly when more than > 32 to 64 cores are requested. I searched the milling list archives and > found that I am very likely running into the memory bandwidth bottleneck. > http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: > > > > > > # of cores iter timing > > > 1 923 19541.83 > > > 4 929 5897.06 > > > 8 932 4854.72 > > > 16 924 1494.33 > > > 32 924 1480.88 > > > 64 928 686.89 > > > 128 927 627.33 > > > 256 926 552.93 > > > > The bandwidth issue has more to do with using multiple cores within a > > node rather than between nodes. Likely the above is a load balancing > > problem or bad communication. > > I use DM to manage the distributed data. The DM was created by calling > DMDACreate3d() and let PETSc decide the local number of nodes in each > direction. To my understand the load of each core is determined at this > stage. If the load balance is done when DMDACreate3d() is called and use > PETSC_DECIDE option? Or how should make the load balanced after DM is > created? > > > We do not have a way to do fine-grained load balancing for the DMDA since > it is intended for very simple topologies. You can see > if it is load imbalance from the division by running with a cube that is > evenly divisible with a cube number of processes. > > Matt > > So, I have nothing to do to make the load balanced if I use DMDA? Would > you please take a look at the attached log summary files and give me some > suggestions on how to improve the speedup ratio? Thanks. > > > Please try what I suggested above. And it looks like there is a little > load imbalance > > Roc----So if the domain is a cube, then the number of the processors is > better to be like 2^3=8, 3^3=9, 4^4 =16, and so on, right? > I want you to try this to eliminate load imbalance as a reason for poor speedup. I don't think it is, but we will see. > I am also wondering whether the physical boundary type effects the load > balance? Since freed node, Dirichlet node and Neumann node has different > number of neighbors? > > VecAXPY 234 1.0 1.0124e+00 3.4 1.26e+08 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 15290 > > VecAXPY 234 1.0 4.2862e-01 3.6 6.37e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 36115 > > > although it is not limiting the speedup. The time imbalance is really > strange. I am guessing other jobs are running on this machine. > > > Roc----The code was run a cluster. There should be other jobs were > running. Do you mean those jobs affect the load balance of my job or speed > of the cluster? I am just trying to improve the scalability of the code, > but really don't know what's the reason that the speedup ratio decreases > so quickly? Thanks. > Yes, other people running can definitely screw up speedup and cause imbalance. Usually timing runs are made with dedicated time. Your VecAXPY and MatMult are speeding up just fine. It is reductions which are killing your computation. You should switch to a more effective preconditioner, so you can avoid all those dot products. Also, you might try something like BiCG with fewer dot products. Matt > Matt > > > > > > > My question here is: Is there any other PC can help on both reducing > iterations and increasing scalability? Thanks. > > > > Always send -log_summary with questions like this, but algebraic > multigrid is a good place to start. > > Please take a look at the attached log file, they are for 128 cores and > 256 cores, respectively. Based on the log files, what should be done to > increase the scalability? Thanks. > > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From a.vergottis at ucl.ac.uk Tue Nov 12 16:10:34 2013 From: a.vergottis at ucl.ac.uk (Anthony Vergottis) Date: Tue, 12 Nov 2013 22:10:34 +0000 Subject: [petsc-users] Local and Global numberings after DMPlexDistribute Message-ID: Dear All, How does one get the local and global numbers of nodes/elements on each process after the DMPlexDistribute function has been used to partition the mesh (DM object)? Are there any built in PETSc functions that offer such functionality? Thanks in advance. Regards, Anthony -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Tue Nov 12 15:55:28 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Tue, 12 Nov 2013 14:55:28 -0700 Subject: [petsc-users] solving stokes-like equation in 3D staggered grid in irregular domain with petscsection In-Reply-To: References: Message-ID: <874n7hb7zj.fsf@jedbrown.org> Bishesh Khanal writes: > Dear all, > I have an implementation of stokes flow equation in cuboid using a > staggered grid with DMDA. Now I want to solve a problem in a slightly > different (but more difficult) situation where I think PetscSection might > be useful although I have not used PetscSection that much and I understand > that some of its aspects is still under development in petsc. > I'm ok to work with the dev version of petsc if what I want to do using > petscsection is plausible. > > Here is my problem description: > > I have a 3D binary image that partitions a cuboid into domain A and domain > B by tagging each cell as 0 or 1. The shape of A and therefore B can be > fairly arbitrary. I want to solve the following sets of equations: > > A. In A: > div(mu(grad(u))) - grad(p) = f1 , where f1 depends on the position x. Is mu constant or variable? Should this be mu (grad(u) + grad(u)^T)/2 ? > div(u) = f2 , where f2 is non-zero and > depends on the position x. Does it really only depend on position or does it depend on state variables (u,p)? > --------------------------------------- > B. In B: > div(mu(grad(u)) = f3, where f3 depends on the position x. What physics does this represent? > ----------------------------------------- > u = constant on the cuboid boundaries. (Dirichlet boundaries) > p = constant on domain B (if needed as boundary condition for domain A) > ------------------------------------------- > > Now, I was thinking of creating a single matrix M that contains the > operators for both A and B domains, and solve a linear system MX = R, where > X and R are the solution and Rhs vectors of roughly the sizes 4*nA + 3*nB > where, nA => no. of cells in A, nB => no. of cells in B. > > Some questions: > > 1. Would PetscSection be useful in this case due to the arbitrary shape of > domain A and because of different number of dofs in domain A and domain B ? > Is there a simple example that uses petscsection for the staggered grid > case ? I would like to be able to implement staggered grid without the use > of ghost/fictitious cells. Just put degrees of freedom on faces and cell centers when creating the section. > 2. Is it a good idea to create a single matrix corresponding to both the > domains A and B (particularly in view of the difficulty it might bring in > solving with some preconditioners using schur fieldsplit) ? I would start with a single matrix. Use MatSetValuesLocal() in assembly so that you can easily switch to a different format. > 3. I'm trying to put both domains in a single matrix to avoid the > difficulty I would have if I want to consider only the domain A. In this > case I would need a traction free boundary condition on the irregular > boundary of domain A, and it seems a bit too challenging for me to > incorporate it with the staggered grid. If there is an idea to implement > this and if you think this could be more suitable than the approach in 2 > above, I would like to learn about that too! Complexity of implementing boundary conditions on staggered grids is one reason some people turn to other discretization technology, such as finite elements. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From dave.mayhem23 at gmail.com Tue Nov 12 17:00:32 2013 From: dave.mayhem23 at gmail.com (Dave May) Date: Wed, 13 Nov 2013 00:00:32 +0100 Subject: [petsc-users] solving stokes-like equation in 3D staggered grid in irregular domain with petscsection In-Reply-To: <874n7hb7zj.fsf@jedbrown.org> References: <874n7hb7zj.fsf@jedbrown.org> Message-ID: Since your using staggered grids, the physical boundary of your domain will be approximate by a "stair-case" type of boundary. (Correct me if this is not what you were thinking to do) Thus, imposing traction boundary conditions on the stair-case boundary should be no more complex than it was in your standard cube domain with staggered grids. The only exception is that you have to implement, in a cell-wise manner, the imposition of the traction condition. The functionality should already exist in your original cube staggered grid implementation, but possible the implementation of this boundary condition was done "wall-wise" rather than cell-wise. > 3. I'm trying to put both domains in a single matrix to avoid the > > difficulty I would have if I want to consider only the domain A. In this > > case I would need a traction free boundary condition on the irregular > > boundary of domain A, and it seems a bit too challenging for me to > > incorporate it with the staggered grid. If there is an idea to implement > > this and if you think this could be more suitable than the approach in 2 > > above, I would like to learn about that too! > > Complexity of implementing boundary conditions on staggered grids is one > reason some people turn to other discretization technology, such as > finite elements. > -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Tue Nov 12 17:25:21 2013 From: knepley at gmail.com (Matthew Knepley) Date: Tue, 12 Nov 2013 17:25:21 -0600 Subject: [petsc-users] Local and Global numberings after DMPlexDistribute In-Reply-To: References: Message-ID: On Tue, Nov 12, 2013 at 4:10 PM, Anthony Vergottis wrote: > Dear All, > > How does one get the local and global numbers of nodes/elements on each > process after the DMPlexDistribute function has been used to partition the > mesh (DM object)? > Can you be more explicit? Vertices and cells are renumbered after distribution to be contiguous on each process: cell: [0, numCells) vertices: [numCells, numCells+numVertices) Matt > Are there any built in PETSc functions that offer such functionality? > > Thanks in advance. > > Regards, > Anthony > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From a.vergottis at ucl.ac.uk Tue Nov 12 17:32:28 2013 From: a.vergottis at ucl.ac.uk (Anthony Vergottis) Date: Tue, 12 Nov 2013 23:32:28 +0000 Subject: [petsc-users] Local and Global numberings after DMPlexDistribute In-Reply-To: References: Message-ID: I would like to calculate the area of all the elements in my unstructured mesh. This is the process that I have envisaged: 1) Partition the mesh with DMPlexDistribute. 2) Obtain which cell/vertices are owned by each process. 3) Have each process calculate the area of the cells it own, as I would like to do this in parallel. I hope this explains things a bit better. Thanks. Regards, Anthony On 12 November 2013 23:25, Matthew Knepley wrote: > On Tue, Nov 12, 2013 at 4:10 PM, Anthony Vergottis wrote: > >> Dear All, >> >> How does one get the local and global numbers of nodes/elements on each >> process after the DMPlexDistribute function has been used to partition the >> mesh (DM object)? >> > > Can you be more explicit? Vertices and cells are renumbered after > distribution to be contiguous on > each process: > > cell: [0, numCells) > vertices: [numCells, numCells+numVertices) > > Matt > > >> Are there any built in PETSc functions that offer such functionality? >> >> Thanks in advance. >> >> Regards, >> Anthony >> > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Tue Nov 12 18:10:21 2013 From: knepley at gmail.com (Matthew Knepley) Date: Tue, 12 Nov 2013 18:10:21 -0600 Subject: [petsc-users] Local and Global numberings after DMPlexDistribute In-Reply-To: References: Message-ID: On Tue, Nov 12, 2013 at 5:32 PM, Anthony Vergottis wrote: > I would like to calculate the area of all the elements in my unstructured > mesh. This is the process that I have envisaged: > > 1) Partition the mesh with DMPlexDistribute. > > 2) Obtain which cell/vertices are owned by each process. > > 3) Have each process calculate the area of the cells it own, as I would > like to do this in parallel. > > I hope this explains things a bit better. > ierr = DMPlexGetHeightStratum(dm, 0, &cStart, &cEnd);CHKERRQ(ierr); for (c = cStart; c < cEnd; ++c) { PetscReal vol; ierr = DMPlexComputeCellGeometryFVM(dm, c, &vol, NULL, NULL);CHKERRQ(ierr); } Matt > Thanks. > > Regards, > Anthony > > > > On 12 November 2013 23:25, Matthew Knepley wrote: > >> On Tue, Nov 12, 2013 at 4:10 PM, Anthony Vergottis > > wrote: >> >>> Dear All, >>> >>> How does one get the local and global numbers of nodes/elements on each >>> process after the DMPlexDistribute function has been used to partition the >>> mesh (DM object)? >>> >> >> Can you be more explicit? Vertices and cells are renumbered after >> distribution to be contiguous on >> each process: >> >> cell: [0, numCells) >> vertices: [numCells, numCells+numVertices) >> >> Matt >> >> >>> Are there any built in PETSc functions that offer such functionality? >>> >>> Thanks in advance. >>> >>> Regards, >>> Anthony >>> >> >> >> >> -- >> What most experimenters take for granted before they begin their >> experiments is infinitely more interesting than any results to which their >> experiments lead. >> -- Norbert Wiener >> > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From a.vergottis at ucl.ac.uk Tue Nov 12 18:30:57 2013 From: a.vergottis at ucl.ac.uk (Anthony Vergottis) Date: Wed, 13 Nov 2013 00:30:57 +0000 Subject: [petsc-users] Local and Global numberings after DMPlexDistribute In-Reply-To: References: Message-ID: Thanks a lot Matt. That helped. Regards, Anthony On 13 November 2013 00:10, Matthew Knepley wrote: > On Tue, Nov 12, 2013 at 5:32 PM, Anthony Vergottis wrote: > >> I would like to calculate the area of all the elements in my unstructured >> mesh. This is the process that I have envisaged: >> >> 1) Partition the mesh with DMPlexDistribute. >> >> 2) Obtain which cell/vertices are owned by each process. >> >> 3) Have each process calculate the area of the cells it own, as I would >> like to do this in parallel. >> >> I hope this explains things a bit better. >> > > ierr = DMPlexGetHeightStratum(dm, 0, &cStart, &cEnd);CHKERRQ(ierr); > for (c = cStart; c < cEnd; ++c) { > PetscReal vol; > ierr = DMPlexComputeCellGeometryFVM(dm, c, &vol, NULL, > NULL);CHKERRQ(ierr); > > } > > Matt > > >> Thanks. >> >> Regards, >> Anthony >> >> >> >> On 12 November 2013 23:25, Matthew Knepley wrote: >> >>> On Tue, Nov 12, 2013 at 4:10 PM, Anthony Vergottis < >>> a.vergottis at ucl.ac.uk> wrote: >>> >>>> Dear All, >>>> >>>> How does one get the local and global numbers of nodes/elements on each >>>> process after the DMPlexDistribute function has been used to partition the >>>> mesh (DM object)? >>>> >>> >>> Can you be more explicit? Vertices and cells are renumbered after >>> distribution to be contiguous on >>> each process: >>> >>> cell: [0, numCells) >>> vertices: [numCells, numCells+numVertices) >>> >>> Matt >>> >>> >>>> Are there any built in PETSc functions that offer such functionality? >>>> >>>> Thanks in advance. >>>> >>>> Regards, >>>> Anthony >>>> >>> >>> >>> >>> -- >>> What most experimenters take for granted before they begin their >>> experiments is infinitely more interesting than any results to which their >>> experiments lead. >>> -- Norbert Wiener >>> >> >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From anush at bu.edu Tue Nov 12 18:42:20 2013 From: anush at bu.edu (Anush Krishnan) Date: Tue, 12 Nov 2013 19:42:20 -0500 Subject: [petsc-users] Calculating the PETSc index of a DMComposite vector In-Reply-To: <877gcd8kxx.fsf@jedbrown.org> References: <87habicu41.fsf@jedbrown.org> <877gcd8kxx.fsf@jedbrown.org> Message-ID: Hi Jed, > > I would loop over the body points and classify them as being "within > range" of the grid points that you own. Ignore the points that are out > of range. From the interpolation patch size, you know analytically > which owned grid points need to contribute values. Call MatSetValues() > or MatSetValuesLocal() for that partial row of C. (Don't worry about > whether you own the body point or not, and only set entries for velocity > points that you own---because the other process will be setting the > entries for the points they own.) > That's exactly what I ended up doing. Thanks! How large are these rectangular regions? > They encompass 3x4 grid points. The velocity grid could potentially have hundreds of nodes in each direction. So they're small compared to the rest of the grid. Depending on the distribution and access pattern, it could make sense to > assemble C^T instead, but it likely doesn't matter. Could you explain why it might be better to assemble C^T? As far as the code is concerned, I don't see any difference between the two, except interchanging the row and column when I call MatSetValues. For memory allocation, I know that every row of C contains exactly 12 non-zeros (since I interpolate from a 3x4 grid). But I have no idea if they are in the diagonal or the off-diagonal region of the matrix. For C^T, I can't predict beforehand how many non-zeros will be present in each row. Most rows will in fact be empty. But if I assume a "nice" solid boundary, I would imagine each velocity grid point should not be influenced by more than ~4 body points. Just to give you an idea of the values of d_nz and o_nz that I ended up using: The test problem consisted of a 20x20 grid in a square domain, interpolating on to a circle in the center of the domain with diameter equal to half of the domain edge. The circle is represented by 32 boundary points. For C^T, I used d_nz=5 and o_nz=5 (A smaller number for either would produce "Argument out of range!" errors) For C, I needed d_nz=12 and o_nz=12. (I should clarify: The test problem actually created the matrix [Cx Cy] to map from [u v]^T to [bx by]^T. u was of size 19x20, and v was 20x19.) What are you doing with the C once it is built? > I need to compute the matrix C * C^T, and use that as the left-hand side matrix of a linear system. I will also have to include C^T as part of a bigger matrix Q, and calculate Q^T*Q, but I should probably ask questions about that on a different thread. Thanks, Anush -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Tue Nov 12 19:21:14 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Tue, 12 Nov 2013 18:21:14 -0700 Subject: [petsc-users] Calculating the PETSc index of a DMComposite vector In-Reply-To: References: <87habicu41.fsf@jedbrown.org> <877gcd8kxx.fsf@jedbrown.org> Message-ID: <87bo1p9jw5.fsf@jedbrown.org> Anush Krishnan writes: > Could you explain why it might be better to assemble C^T? As far as the > code is concerned, I don't see any difference between the two, except > interchanging the row and column when I call MatSetValues. Correct, it's a trivial change for you, but it would mean that all the entries would be local, rather than being communicated to the owner of the body point. > For memory allocation, I know that every row of C contains exactly 12 > non-zeros (since I interpolate from a 3x4 grid). But I have no idea if they > are in the diagonal or the off-diagonal region of the matrix. For C^T, I > can't predict beforehand how many non-zeros will be present in each row. > Most rows will in fact be empty. But if I assume a "nice" solid boundary, I > would imagine each velocity grid point should not be influenced by more > than ~4 body points. Can you just make one pass to count and another to insert? > Just to give you an idea of the values of d_nz and o_nz that I ended up > using: > The test problem consisted of a 20x20 grid in a square domain, > interpolating on to a circle in the center of the domain with diameter > equal to half of the domain edge. The circle is represented by 32 boundary > points. > For C^T, I used d_nz=5 and o_nz=5 (A smaller number for either would > produce "Argument out of range!" errors) > For C, I needed d_nz=12 and o_nz=12. > (I should clarify: The test problem actually created the matrix [Cx Cy] to > map from [u v]^T to [bx by]^T. u was of size 19x20, and v was 20x19.) > > What are you doing with the C once it is built? >> > > I need to compute the matrix C * C^T, and use that as the left-hand side > matrix of a linear system. I will also have to include C^T as part of a > bigger matrix Q, and calculate Q^T*Q, but I should probably ask questions > about that on a different thread. C * C^T is the influence of particles on each other? Why can't you compute that analytically (without referencing the grid directly)? -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From bisheshkh at gmail.com Wed Nov 13 04:45:45 2013 From: bisheshkh at gmail.com (Bishesh Khanal) Date: Wed, 13 Nov 2013 11:45:45 +0100 Subject: [petsc-users] solving stokes-like equation in 3D staggered grid in irregular domain with petscsection In-Reply-To: <874n7hb7zj.fsf@jedbrown.org> References: <874n7hb7zj.fsf@jedbrown.org> Message-ID: On Tue, Nov 12, 2013 at 10:55 PM, Jed Brown wrote: > Bishesh Khanal writes: > > > Dear all, > > I have an implementation of stokes flow equation in cuboid using a > > staggered grid with DMDA. Now I want to solve a problem in a slightly > > different (but more difficult) situation where I think PetscSection might > > be useful although I have not used PetscSection that much and I > understand > > that some of its aspects is still under development in petsc. > > I'm ok to work with the dev version of petsc if what I want to do using > > petscsection is plausible. > > > > Here is my problem description: > > > > I have a 3D binary image that partitions a cuboid into domain A and > domain > > B by tagging each cell as 0 or 1. The shape of A and therefore B can be > > fairly arbitrary. I want to solve the following sets of equations: > > > > A. In A: > > div(mu(grad(u))) - grad(p) = f1 , where f1 depends on the position x. > > Is mu constant or variable? Should this be mu (grad(u) + grad(u)^T)/2 ? > Within A, for now, I can consider mu to be constant, although later if possible it can be a variable even a tensor to describe anisotropy. But to start with I want put this a constant. The original equations start with mu (grad(u) + grad(u)^T) but then simplifications occur due to div(u) = f2 where f2 is known and is an input to the system. Some terms involving f2 go to the right hand side to form f1. > > > div(u) = f2 , where f2 is non-zero > and > > depends on the position x. > > Does it really only depend on position or does it depend on state > variables (u,p)? > f2 comes from my other models, it is an output of an image processing pipeline, thus is a 3D image. So it does vary depending on position, and is independent of (u,p) for this time independent sets of equations. > > > --------------------------------------- > > B. In B: > > div(mu(grad(u)) = f3, where f3 depends on the position x. > > What physics does this represent? > I'm mostly interested in the phenomenon in A with my model, here B is the extension of the very irregular domain of A to get a cuboid. Here, in B I release the div(u) = f2 constraint and just put a regularisation to penalize large deformation. What is of importance here is to compensate the net volume expansion in domain A by corresponding contraction in domain B so that the boundaries of the cuboid do not move. It does not particularly represent any physics except probably that it gives me a velocity field having a certain divergence field that penalizes big deformations. > > ----------------------------------------- > > u = constant on the cuboid boundaries. (Dirichlet boundaries) > > p = constant on domain B (if needed as boundary condition for domain A) > > ------------------------------------------- > > > > Now, I was thinking of creating a single matrix M that contains the > > operators for both A and B domains, and solve a linear system MX = R, > where > > X and R are the solution and Rhs vectors of roughly the sizes 4*nA + 3*nB > > where, nA => no. of cells in A, nB => no. of cells in B. > > > > Some questions: > > > > 1. Would PetscSection be useful in this case due to the arbitrary shape > of > > domain A and because of different number of dofs in domain A and domain > B ? > > Is there a simple example that uses petscsection for the staggered grid > > case ? I would like to be able to implement staggered grid without the > use > > of ghost/fictitious cells. > > Just put degrees of freedom on faces and cell centers when creating the > section. > > I've not understood petscsection well. I tried reading up in the manual but it does not look easy to follow with all the terms I'm unfamiliar with. Is it possible to have an example for the staggered grid case ? Or a short illustration with code snippets/idea on how should I start with it ? > > 2. Is it a good idea to create a single matrix corresponding to both the > > domains A and B (particularly in view of the difficulty it might bring in > > solving with some preconditioners using schur fieldsplit) ? > > I would start with a single matrix. Use MatSetValuesLocal() in assembly > so that you can easily switch to a different format. > > > 3. I'm trying to put both domains in a single matrix to avoid the > > difficulty I would have if I want to consider only the domain A. In this > > case I would need a traction free boundary condition on the irregular > > boundary of domain A, and it seems a bit too challenging for me to > > incorporate it with the staggered grid. If there is an idea to implement > > this and if you think this could be more suitable than the approach in 2 > > above, I would like to learn about that too! > > Complexity of implementing boundary conditions on staggered grids is one > reason some people turn to other discretization technology, such as > finite elements. > I do not know much about FEM. But some of the reasons why I have avoided it in this particular problem are: (Please correct me on any of the following points if they are wrong) 1. The inputs f1 and f2 are 3D images (in average of size 200^3) that come from other image processing pipeline; it's important that I constrain u at each voxel for div(u) = f2 in domain A. I am trying to avoid having to get the meshing from the 3D image(with very detailed structures), then go back to the image from the obtained u again because I have to use the obtained u to warp the image, transport other parameters again with u in the image space and again obtain new f1 and f2 images. Then iterate this few times. -------------- next part -------------- An HTML attachment was scrubbed... URL: From pengxwang at hotmail.com Wed Nov 13 09:24:16 2013 From: pengxwang at hotmail.com (Roc Wang) Date: Wed, 13 Nov 2013 09:24:16 -0600 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: References: , <87zjpcgj2l.fsf@jedbrown.org>, , , , , , Message-ID: Hi, I tried to use -ksp_type bicg, but there was error. It was fine if I use gmres as solver. Doe it mean the matrix cannot be solved by BiCG? Thanks. [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Floating point exception! [0]PETSC ERROR: Infinite or not-a-number generated in norm! [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Petsc Release Version 3.3.0, Patch 6, Mon Feb 11 12:26:34 CST 2013 [0]PETSC ERROR: See docs/changes/index.html for recent updates. [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [0]PETSC ERROR: See docs/index.html for manual pages. [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: ./x.r on a arch-linu named node48.cocoa5 by pzw2 Wed Nov 13 10:09:22 2013 [0]PETSC ERROR: Libraries linked from /home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib [0]PETSC ERROR: Configure run at Tue Nov 12 09:52:45 2013 [0]PETSC ERROR: Configure options --download-f-blas-lapack --with-mpi-dir=/usr/local/OpenMPI-1.6.4_Intel --download-hypre=1 --download-hdf5=1 --download-superlu_dist --download-parmetis --download-metis --download-spai --with-debugging=no [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: VecNorm() line 169 in /home/pzw2/ZSoft/petsc-3.3-p6/src/vec/vec/interface/rvector.c [0]PETSC ERROR: KSPSolve_BiCG() line 107 in /home/pzw2/ZSoft/petsc-3.3-p6/src/ksp/ksp/impls/bicg/bicg.c [0]PETSC ERROR: KSPSolve() line 446 in /home/pzw2/ZSoft/petsc-3.3-p6/src/ksp/ksp/interface/itfunc.c [0]PETSC ERROR: LinearSolver() line 181 in "unknowndirectory/"src/solver.cpp [23]PETSC ERROR: ------------------------------------------------------------------------ [23]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, probably memory access out of range [23]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger [23]PETSC ERROR: or see http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[23]PETSC ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory corruption errors [23]PETSC ERROR: configure using --with-debugging=yes, recompile, link, and run [23]PETSC ERROR: to get more information on the crash. [23]PETSC ERROR: --------------------- Error Message ------------------------------------ [23]PETSC ERROR: Signal received! Date: Tue, 12 Nov 2013 15:34:16 -0600 Subject: Re: [petsc-users] approaches to reduce computing time From: knepley at gmail.com To: pengxwang at hotmail.com CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov On Tue, Nov 12, 2013 at 3:22 PM, Roc Wang wrote: Date: Tue, 12 Nov 2013 14:59:30 -0600 Subject: Re: [petsc-users] approaches to reduce computing time From: knepley at gmail.com To: pengxwang at hotmail.com CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov On Tue, Nov 12, 2013 at 2:48 PM, Roc Wang wrote: Date: Tue, 12 Nov 2013 14:22:35 -0600 Subject: Re: [petsc-users] approaches to reduce computing time From: knepley at gmail.com To: pengxwang at hotmail.com CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov On Tue, Nov 12, 2013 at 2:14 PM, Roc Wang wrote: Thanks Jed, I have questions about load balance and PC type below. > From: jedbrown at mcs.anl.gov > To: pengxwang at hotmail.com; petsc-users at mcs.anl.gov > Subject: Re: [petsc-users] approaches to reduce computing time > Date: Sun, 10 Nov 2013 12:20:18 -0700 > > Roc Wang writes: > > > Hi all, > > > > I am trying to minimize the computing time to solve a large sparse matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to reduce the computing time from two directions: 1 finding a Pre-conditioner to reduce the number of iterations which reduces the time numerically, 2 requesting more cores. > > > > ----For the first method, I tried several methods: > > 1 default KSP and PC, > > 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type jacobi, > > 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, > > 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, > > 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm (PCASM) > > > > The iterations and timing is like the following with 128 cores requested: > > case# iter timing (s) > > 1 1436 816 > > 2 3 12658 > > 3 1069 669.64 > > 4 872 768.12 > > 5 927 513.14 > > > > It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment can help to reduce the iterations but not the timing (comparing case 3 and 4). Second, the PCASM helps a lot. Although the second option is able to reduce iterations, the timing increases very much. Is it because more operations are needed in the PC? > > > > My questions here are: 1. Which direction should I take to select > > -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger > > restart with large augment is better or larger restart with smaller > > augment is better? > > Look at the -log_summary. By increasing the restart, the work in > KSPGMRESOrthog will increase linearly, but the number of iterations > might decrease enough to compensate. There is no general rule here > since it depends on the relative expense of operations for your problem > on your machine. > > > ----For the second method, I tried with -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different number of cores. I found the speedup ratio increases slowly when more than 32 to 64 cores are requested. I searched the milling list archives and found that I am very likely running into the memory bandwidth bottleneck. http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: > > > > # of cores iter timing > > 1 923 19541.83 > > 4 929 5897.06 > > 8 932 4854.72 > > 16 924 1494.33 > > 32 924 1480.88 > > 64 928 686.89 > > 128 927 627.33 > > 256 926 552.93 > > The bandwidth issue has more to do with using multiple cores within a > node rather than between nodes. Likely the above is a load balancing > problem or bad communication. I use DM to manage the distributed data. The DM was created by calling DMDACreate3d() and let PETSc decide the local number of nodes in each direction. To my understand the load of each core is determined at this stage. If the load balance is done when DMDACreate3d() is called and use PETSC_DECIDE option? Or how should make the load balanced after DM is created? We do not have a way to do fine-grained load balancing for the DMDA since it is intended for very simple topologies. You can seeif it is load imbalance from the division by running with a cube that is evenly divisible with a cube number of processes. Matt So, I have nothing to do to make the load balanced if I use DMDA? Would you please take a look at the attached log summary files and give me some suggestions on how to improve the speedup ratio? Thanks. Please try what I suggested above. And it looks like there is a little load imbalance Roc----So if the domain is a cube, then the number of the processors is better to be like 2^3=8, 3^3=9, 4^4 =16, and so on, right? I want you to try this to eliminate load imbalance as a reason for poor speedup. I don't think it is, but we will see. I am also wondering whether the physical boundary type effects the load balance? Since freed node, Dirichlet node and Neumann node has different number of neighbors? VecAXPY 234 1.0 1.0124e+00 3.4 1.26e+08 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 15290 VecAXPY 234 1.0 4.2862e-01 3.6 6.37e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 36115 although it is not limiting the speedup. The time imbalance is really strange. I am guessing other jobs are running on this machine. Roc----The code was run a cluster. There should be other jobs were running. Do you mean those jobs affect the load balance of my job or speed of the cluster? I am just trying to improve the scalability of the code, but really don't know what's the reason that the speedup ratio decreases so quickly? Thanks. Yes, other people running can definitely screw up speedup and cause imbalance. Usually timing runs are made with dedicated time. Your VecAXPY and MatMult are speeding up just fine. It is reductions which are killing your computation.You should switch to a more effective preconditioner, so you can avoid all those dot products. Also, you might try something like BiCG with fewer dot products. Matt Matt > > > My question here is: Is there any other PC can help on both reducing iterations and increasing scalability? Thanks. > > Always send -log_summary with questions like this, but algebraic multigrid is a good place to start. Please take a look at the attached log file, they are for 128 cores and 256 cores, respectively. Based on the log files, what should be done to increase the scalability? Thanks. -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Wed Nov 13 09:40:27 2013 From: knepley at gmail.com (Matthew Knepley) Date: Wed, 13 Nov 2013 09:40:27 -0600 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: References: <87zjpcgj2l.fsf@jedbrown.org> Message-ID: On Wed, Nov 13, 2013 at 9:24 AM, Roc Wang wrote: > Hi, I tried to use -ksp_type bicg, but there was error. It was fine if I > use gmres as solver. Doe it mean the matrix cannot be solved by BiCG? > Thanks. > BiCG can breakdown. You can try -ksp_type bcgstab Matt > [0]PETSC ERROR: --------------------- Error Message > ------------------------------------ > [0]PETSC ERROR: Floating point exception! > [0]PETSC ERROR: Infinite or not-a-number generated in norm! > [0]PETSC ERROR: > ------------------------------------------------------------------------ > [0]PETSC ERROR: Petsc Release Version 3.3.0, Patch 6, Mon Feb 11 12:26:34 > CST 2013 > [0]PETSC ERROR: See docs/changes/index.html for recent updates. > [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > [0]PETSC ERROR: See docs/index.html for manual pages. > [0]PETSC ERROR: > ------------------------------------------------------------------------ > [0]PETSC ERROR: ./x.r on a arch-linu named node48.cocoa5 by pzw2 Wed Nov > 13 10:09:22 2013 > [0]PETSC ERROR: Libraries linked from > /home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib > [0]PETSC ERROR: Configure run at Tue Nov 12 09:52:45 2013 > [0]PETSC ERROR: Configure options --download-f-blas-lapack > --with-mpi-dir=/usr/local/OpenMPI-1.6.4_Intel --download-hypre=1 > --download-hdf5=1 --download-superlu_dist --download-parmetis > --download-metis --download-spai --with-debugging=no > [0]PETSC ERROR: > ------------------------------------------------------------------------ > [0]PETSC ERROR: VecNorm() line 169 in > /home/pzw2/ZSoft/petsc-3.3-p6/src/vec/vec/interface/rvector.c > [0]PETSC ERROR: KSPSolve_BiCG() line 107 in > /home/pzw2/ZSoft/petsc-3.3-p6/src/ksp/ksp/impls/bicg/bicg.c > [0]PETSC ERROR: KSPSolve() line 446 in > /home/pzw2/ZSoft/petsc-3.3-p6/src/ksp/ksp/interface/itfunc.c > [0]PETSC ERROR: LinearSolver() line 181 in > "unknowndirectory/"src/solver.cpp > [23]PETSC ERROR: > ------------------------------------------------------------------------ > [23]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, > probably memory access out of range > [23]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger > [23]PETSC ERROR: or see > http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[23]PETSCERROR: or try > http://valgrind.org on GNU/linux and Apple Mac OS X to find memory > corruption errors > [23]PETSC ERROR: configure using --with-debugging=yes, recompile, link, > and run > [23]PETSC ERROR: to get more information on the crash. > [23]PETSC ERROR: --------------------- Error Message > ------------------------------------ > [23]PETSC ERROR: Signal received! > > ------------------------------ > Date: Tue, 12 Nov 2013 15:34:16 -0600 > Subject: Re: [petsc-users] approaches to reduce computing time > From: knepley at gmail.com > To: pengxwang at hotmail.com > CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov > > On Tue, Nov 12, 2013 at 3:22 PM, Roc Wang wrote: > > > > ------------------------------ > Date: Tue, 12 Nov 2013 14:59:30 -0600 > Subject: Re: [petsc-users] approaches to reduce computing time > From: knepley at gmail.com > To: pengxwang at hotmail.com > CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov > > On Tue, Nov 12, 2013 at 2:48 PM, Roc Wang wrote: > > > > ------------------------------ > Date: Tue, 12 Nov 2013 14:22:35 -0600 > Subject: Re: [petsc-users] approaches to reduce computing time > From: knepley at gmail.com > To: pengxwang at hotmail.com > CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov > > On Tue, Nov 12, 2013 at 2:14 PM, Roc Wang wrote: > > Thanks Jed, > > I have questions about load balance and PC type below. > > > From: jedbrown at mcs.anl.gov > > To: pengxwang at hotmail.com; petsc-users at mcs.anl.gov > > Subject: Re: [petsc-users] approaches to reduce computing time > > Date: Sun, 10 Nov 2013 12:20:18 -0700 > > > > Roc Wang writes: > > > > > Hi all, > > > > > > I am trying to minimize the computing time to solve a large sparse > matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to > reduce the computing time from two directions: 1 finding a Pre-conditioner > to reduce the number of iterations which reduces the time numerically, 2 > requesting more cores. > > > > > > ----For the first method, I tried several methods: > > > 1 default KSP and PC, > > > 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type > jacobi, > > > 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, > > > 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, > > > 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 > -pc_type asm (PCASM) > > > > > > The iterations and timing is like the following with 128 cores > requested: > > > case# iter timing (s) > > > 1 1436 816 > > > 2 3 12658 > > > 3 1069 669.64 > > > 4 872 768.12 > > > 5 927 513.14 > > > > > > It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment > can help to reduce the iterations but not the timing (comparing case 3 and > 4). Second, the PCASM helps a lot. Although the second option is able to > reduce iterations, the timing increases very much. Is it because more > operations are needed in the PC? > > > > > > My questions here are: 1. Which direction should I take to select > > > -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger > > > restart with large augment is better or larger restart with smaller > > > augment is better? > > > > Look at the -log_summary. By increasing the restart, the work in > > KSPGMRESOrthog will increase linearly, but the number of iterations > > might decrease enough to compensate. There is no general rule here > > since it depends on the relative expense of operations for your problem > > on your machine. > > > > > ----For the second method, I tried with -ksp_type lgmres > -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different > number of cores. I found the speedup ratio increases slowly when more than > 32 to 64 cores are requested. I searched the milling list archives and > found that I am very likely running into the memory bandwidth bottleneck. > http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: > > > > > > # of cores iter timing > > > 1 923 19541.83 > > > 4 929 5897.06 > > > 8 932 4854.72 > > > 16 924 1494.33 > > > 32 924 1480.88 > > > 64 928 686.89 > > > 128 927 627.33 > > > 256 926 552.93 > > > > The bandwidth issue has more to do with using multiple cores within a > > node rather than between nodes. Likely the above is a load balancing > > problem or bad communication. > > I use DM to manage the distributed data. The DM was created by calling > DMDACreate3d() and let PETSc decide the local number of nodes in each > direction. To my understand the load of each core is determined at this > stage. If the load balance is done when DMDACreate3d() is called and use > PETSC_DECIDE option? Or how should make the load balanced after DM is > created? > > > We do not have a way to do fine-grained load balancing for the DMDA since > it is intended for very simple topologies. You can see > if it is load imbalance from the division by running with a cube that is > evenly divisible with a cube number of processes. > > Matt > > So, I have nothing to do to make the load balanced if I use DMDA? Would > you please take a look at the attached log summary files and give me some > suggestions on how to improve the speedup ratio? Thanks. > > > Please try what I suggested above. And it looks like there is a little > load imbalance > > Roc----So if the domain is a cube, then the number of the processors is > better to be like 2^3=8, 3^3=9, 4^4 =16, and so on, right? > > > I want you to try this to eliminate load imbalance as a reason for poor > speedup. I don't think it is, but we will see. > > > I am also wondering whether the physical boundary type effects the load > balance? Since freed node, Dirichlet node and Neumann node has different > number of neighbors? > > VecAXPY 234 1.0 1.0124e+00 3.4 1.26e+08 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 15290 > > VecAXPY 234 1.0 4.2862e-01 3.6 6.37e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 36115 > > > although it is not limiting the speedup. The time imbalance is really > strange. I am guessing other jobs are running on this machine. > > > Roc----The code was run a cluster. There should be other jobs were > running. Do you mean those jobs affect the load balance of my job or speed > of the cluster? I am just trying to improve the scalability of the code, > but really don't know what's the reason that the speedup ratio decreases > so quickly? Thanks. > > > Yes, other people running can definitely screw up speedup and cause > imbalance. Usually timing runs are made with dedicated time. > > Your VecAXPY and MatMult are speeding up just fine. It is reductions which > are killing your computation. > You should switch to a more effective preconditioner, so you can avoid all > those dot products. Also, you > might try something like BiCG with fewer dot products. > > Matt > > > Matt > > > > > > > My question here is: Is there any other PC can help on both reducing > iterations and increasing scalability? Thanks. > > > > Always send -log_summary with questions like this, but algebraic > multigrid is a good place to start. > > Please take a look at the attached log file, they are for 128 cores and > 256 cores, respectively. Based on the log files, what should be done to > increase the scalability? Thanks. > > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From pengxwang at hotmail.com Wed Nov 13 11:49:33 2013 From: pengxwang at hotmail.com (Roc Wang) Date: Wed, 13 Nov 2013 11:49:33 -0600 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: References: , <87zjpcgj2l.fsf@jedbrown.org>, , , , , , , , Message-ID: Date: Wed, 13 Nov 2013 09:40:27 -0600 Subject: Re: [petsc-users] approaches to reduce computing time From: knepley at gmail.com To: pengxwang at hotmail.com CC: petsc-users at mcs.anl.gov On Wed, Nov 13, 2013 at 9:24 AM, Roc Wang wrote: Hi, I tried to use -ksp_type bicg, but there was error. It was fine if I use gmres as solver. Doe it mean the matrix cannot be solved by BiCG? Thanks. BiCG can breakdown. You can try -ksp_type bcgstab Thanks, I used -ksp_bcgsl. It worked well. In addition, I also tried to employed the options -ksp_bcgsl_ell 1 and -ksp_bcgsl_cxpoly. But I the records of residual norm are same as without -ksp_bcgsl_cxpoly. Should there be difference between two options? OR, I didn't set the option correctly? The options showed in the following webpages are different. http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/KSP/KSPBCGSL.html#KSPBCGSL -ksp_bcgsl_cxpol http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/KSP/KSPBCGSLSetPol.html#KSPBCGSLSetPol -ksp_bcgsl_cxpoly - use enhanced polynomial . Thanks. Matt [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Floating point exception! [0]PETSC ERROR: Infinite or not-a-number generated in norm! [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Petsc Release Version 3.3.0, Patch 6, Mon Feb 11 12:26:34 CST 2013 [0]PETSC ERROR: See docs/changes/index.html for recent updates. [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [0]PETSC ERROR: See docs/index.html for manual pages. [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: ./x.r on a arch-linu named node48.cocoa5 by pzw2 Wed Nov 13 10:09:22 2013 [0]PETSC ERROR: Libraries linked from /home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib [0]PETSC ERROR: Configure run at Tue Nov 12 09:52:45 2013 [0]PETSC ERROR: Configure options --download-f-blas-lapack --with-mpi-dir=/usr/local/OpenMPI-1.6.4_Intel --download-hypre=1 --download-hdf5=1 --download-superlu_dist --download-parmetis --download-metis --download-spai --with-debugging=no [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: VecNorm() line 169 in /home/pzw2/ZSoft/petsc-3.3-p6/src/vec/vec/interface/rvector.c [0]PETSC ERROR: KSPSolve_BiCG() line 107 in /home/pzw2/ZSoft/petsc-3.3-p6/src/ksp/ksp/impls/bicg/bicg.c [0]PETSC ERROR: KSPSolve() line 446 in /home/pzw2/ZSoft/petsc-3.3-p6/src/ksp/ksp/interface/itfunc.c [0]PETSC ERROR: LinearSolver() line 181 in "unknowndirectory/"src/solver.cpp [23]PETSC ERROR: ------------------------------------------------------------------------ [23]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, probably memory access out of range [23]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger [23]PETSC ERROR: or see http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[23]PETSC ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory corruption errors [23]PETSC ERROR: configure using --with-debugging=yes, recompile, link, and run [23]PETSC ERROR: to get more information on the crash. [23]PETSC ERROR: --------------------- Error Message ------------------------------------ [23]PETSC ERROR: Signal received! Date: Tue, 12 Nov 2013 15:34:16 -0600 Subject: Re: [petsc-users] approaches to reduce computing time From: knepley at gmail.com To: pengxwang at hotmail.com CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov On Tue, Nov 12, 2013 at 3:22 PM, Roc Wang wrote: Date: Tue, 12 Nov 2013 14:59:30 -0600 Subject: Re: [petsc-users] approaches to reduce computing time From: knepley at gmail.com To: pengxwang at hotmail.com CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov On Tue, Nov 12, 2013 at 2:48 PM, Roc Wang wrote: Date: Tue, 12 Nov 2013 14:22:35 -0600 Subject: Re: [petsc-users] approaches to reduce computing time From: knepley at gmail.com To: pengxwang at hotmail.com CC: jedbrown at mcs.anl.gov; petsc-users at mcs.anl.gov On Tue, Nov 12, 2013 at 2:14 PM, Roc Wang wrote: Thanks Jed, I have questions about load balance and PC type below. > From: jedbrown at mcs.anl.gov > To: pengxwang at hotmail.com; petsc-users at mcs.anl.gov > Subject: Re: [petsc-users] approaches to reduce computing time > Date: Sun, 10 Nov 2013 12:20:18 -0700 > > Roc Wang writes: > > > Hi all, > > > > I am trying to minimize the computing time to solve a large sparse matrix. The matrix dimension is with m=321 n=321 and p=321. I am trying to reduce the computing time from two directions: 1 finding a Pre-conditioner to reduce the number of iterations which reduces the time numerically, 2 requesting more cores. > > > > ----For the first method, I tried several methods: > > 1 default KSP and PC, > > 2 -ksp_type fgmres -ksp_gmres_restart 30 -pc_type ksp -ksp_pc_type jacobi, > > 3 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10, > > 4 -ksp_type lgmres -ksp_gmres_restart 50 -ksp_lgmres_augment 10, > > 5 -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm (PCASM) > > > > The iterations and timing is like the following with 128 cores requested: > > case# iter timing (s) > > 1 1436 816 > > 2 3 12658 > > 3 1069 669.64 > > 4 872 768.12 > > 5 927 513.14 > > > > It can be seen that change -ksp_gmres_restart and -ksp_lgmres_augment can help to reduce the iterations but not the timing (comparing case 3 and 4). Second, the PCASM helps a lot. Although the second option is able to reduce iterations, the timing increases very much. Is it because more operations are needed in the PC? > > > > My questions here are: 1. Which direction should I take to select > > -ksp_gmres_restart and -ksp_lgmres_augment? For example, if larger > > restart with large augment is better or larger restart with smaller > > augment is better? > > Look at the -log_summary. By increasing the restart, the work in > KSPGMRESOrthog will increase linearly, but the number of iterations > might decrease enough to compensate. There is no general rule here > since it depends on the relative expense of operations for your problem > on your machine. > > > ----For the second method, I tried with -ksp_type lgmres -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -pc_type asm with different number of cores. I found the speedup ratio increases slowly when more than 32 to 64 cores are requested. I searched the milling list archives and found that I am very likely running into the memory bandwidth bottleneck. http://www.mail-archive.com/petsc-users at mcs.anl.gov/msg19152.html: > > > > # of cores iter timing > > 1 923 19541.83 > > 4 929 5897.06 > > 8 932 4854.72 > > 16 924 1494.33 > > 32 924 1480.88 > > 64 928 686.89 > > 128 927 627.33 > > 256 926 552.93 > > The bandwidth issue has more to do with using multiple cores within a > node rather than between nodes. Likely the above is a load balancing > problem or bad communication. I use DM to manage the distributed data. The DM was created by calling DMDACreate3d() and let PETSc decide the local number of nodes in each direction. To my understand the load of each core is determined at this stage. If the load balance is done when DMDACreate3d() is called and use PETSC_DECIDE option? Or how should make the load balanced after DM is created? We do not have a way to do fine-grained load balancing for the DMDA since it is intended for very simple topologies. You can seeif it is load imbalance from the division by running with a cube that is evenly divisible with a cube number of processes. Matt So, I have nothing to do to make the load balanced if I use DMDA? Would you please take a look at the attached log summary files and give me some suggestions on how to improve the speedup ratio? Thanks. Please try what I suggested above. And it looks like there is a little load imbalance Roc----So if the domain is a cube, then the number of the processors is better to be like 2^3=8, 3^3=9, 4^4 =16, and so on, right? I want you to try this to eliminate load imbalance as a reason for poor speedup. I don't think it is, but we will see. I am also wondering whether the physical boundary type effects the load balance? Since freed node, Dirichlet node and Neumann node has different number of neighbors? VecAXPY 234 1.0 1.0124e+00 3.4 1.26e+08 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 15290 VecAXPY 234 1.0 4.2862e-01 3.6 6.37e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 36115 although it is not limiting the speedup. The time imbalance is really strange. I am guessing other jobs are running on this machine. Roc----The code was run a cluster. There should be other jobs were running. Do you mean those jobs affect the load balance of my job or speed of the cluster? I am just trying to improve the scalability of the code, but really don't know what's the reason that the speedup ratio decreases so quickly? Thanks. Yes, other people running can definitely screw up speedup and cause imbalance. Usually timing runs are made with dedicated time. Your VecAXPY and MatMult are speeding up just fine. It is reductions which are killing your computation.You should switch to a more effective preconditioner, so you can avoid all those dot products. Also, you might try something like BiCG with fewer dot products. Matt Matt > > > My question here is: Is there any other PC can help on both reducing iterations and increasing scalability? Thanks. > > Always send -log_summary with questions like this, but algebraic multigrid is a good place to start. Please take a look at the attached log file, they are for 128 cores and 256 cores, respectively. Based on the log files, what should be done to increase the scalability? Thanks. -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- An embedded and charset-unspecified text was scrubbed... Name: run_bcgl_ell=1_cxpoly.log URL: -------------- next part -------------- An embedded and charset-unspecified text was scrubbed... Name: run_bcgsl_ell=1.log URL: -------------- next part -------------- An embedded and charset-unspecified text was scrubbed... Name: run_bcgsl_ell=1_cxpol.log URL: From jedbrown at mcs.anl.gov Wed Nov 13 12:00:18 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 13 Nov 2013 11:00:18 -0700 Subject: [petsc-users] approaches to reduce computing time In-Reply-To: References: <87zjpcgj2l.fsf@jedbrown.org> Message-ID: <87zjp86v2l.fsf@jedbrown.org> Roc, please use a standard quoting style so that we can make sense of your replies. If you look at this thread, you can see that your replies repeatedly, actively destroy quoted context, and do not trim, so the message becomes unintelligible. http://lists.mcs.anl.gov/pipermail/petsc-users/2013-November/019648.html The Krylov method is taking a very large number of iterations. Working on a better preconditioner is a better use of your time than micro-tuning Krylov parameters. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From Eric.Chamberland at giref.ulaval.ca Wed Nov 13 13:33:42 2013 From: Eric.Chamberland at giref.ulaval.ca (Eric Chamberland) Date: Wed, 13 Nov 2013 14:33:42 -0500 Subject: [petsc-users] Algorithms for counting nonzeros! Message-ID: <5283D416.2080903@giref.ulaval.ca> Hi all, we are in the process of adding new functionalities to our parallel finite-element code and are facing new challenges in computing the number of non-zeros of a sparse matrix. Since we use Petsc, we want to be able to give this information before the first assembly (to pass to Mat{MPI,SEQ}AIJSetPreallocation). Willing to rewrite the nonzeros counting algorithm from scratch if necessary, we asked ourselves a number of questions: 1) Do algorithms in general count the *exact* number of non-zeros? or do they just count a maximum and let Petsc release the memory after the first assembly? 2) How bad is it to reserve too-much non-zeros before the first assembly? 3) Where could we find some hints/articles examples or information about that kind of algorithm (working in parallel of course) ? Here is the context that we have up to now: we can manage a mixture of interpolations (P1, P2, discontinuous interpolations, etc) and of dimensions (scalar, 3D, etc) for all unknowns in a global matrix (example: a P2-P1 [Taylor-Hood] mixed formulation for Stokes problem). Since these parameters are determined at run-time (in fact they are user parameters), we have to think of a general algorithm to compute the "mostly" exact number of non-zeros for each matrix lines. Right now, we have a working, but memory consuming algorithm which computes the *exact* number on non-zeros per line (we exploded the memory for a problem with a little more than 2^31 unknowns with it -- using 64 bits indices of course). In fact, we manage to store the number of neighbors for each type of mesh components (ex: for a vertice, we have the following number of neighbors: vertices, edges, faces, and each type of elements). It is costly and moreover will not handle the new functionalities we want in the code (XFEM for example). Thanks a lot! Eric From knepley at gmail.com Wed Nov 13 13:40:16 2013 From: knepley at gmail.com (Matthew Knepley) Date: Wed, 13 Nov 2013 13:40:16 -0600 Subject: [petsc-users] Algorithms for counting nonzeros! In-Reply-To: <5283D416.2080903@giref.ulaval.ca> References: <5283D416.2080903@giref.ulaval.ca> Message-ID: On Wed, Nov 13, 2013 at 1:33 PM, Eric Chamberland < Eric.Chamberland at giref.ulaval.ca> wrote: > Hi all, > > we are in the process of adding new functionalities to our parallel > finite-element code and are facing new challenges in computing the number > of non-zeros of a sparse matrix. Since we use Petsc, we want to be able to > give this information before the first assembly (to pass to Mat{MPI,SEQ} > AIJSetPreallocation). > > Willing to rewrite the nonzeros counting algorithm from scratch if > necessary, we asked ourselves a number of questions: > > 1) Do algorithms in general count the *exact* number of non-zeros? or do > they just count a maximum and let Petsc release the memory after the first > assembly? > > 2) How bad is it to reserve too-much non-zeros before the first assembly? > > 3) Where could we find some hints/articles examples or information about > that kind of algorithm (working in parallel of course) ? > > Here is the context that we have up to now: we can manage a mixture of > interpolations (P1, P2, discontinuous interpolations, etc) and of > dimensions (scalar, 3D, etc) for all unknowns in a global matrix (example: > a P2-P1 [Taylor-Hood] mixed formulation for Stokes problem). > > Since these parameters are determined at run-time (in fact they are user > parameters), we have to think of a general algorithm to compute the > "mostly" exact number of non-zeros for each matrix lines. > > Right now, we have a working, but memory consuming algorithm which > computes the *exact* number on non-zeros per line (we exploded the memory > for a problem with a little more than 2^31 unknowns with it -- using 64 > bits indices of course). In fact, we manage to store the number of > neighbors for each type of mesh components (ex: for a vertice, we have the > following number of neighbors: vertices, edges, faces, and each type of > elements). It is costly and moreover will not handle the new > functionalities we want in the code (XFEM for example). > DMPlex will do it already. If you are opposed to using that, you can use Jed's simple method which is to "fake" assemble without values, putting all the (i,j) pairs into a hash table. Matt > Thanks a lot! > > Eric > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From balay at mcs.anl.gov Wed Nov 13 13:48:38 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Wed, 13 Nov 2013 13:48:38 -0600 (CST) Subject: [petsc-users] Algorithms for counting nonzeros! In-Reply-To: <5283D416.2080903@giref.ulaval.ca> References: <5283D416.2080903@giref.ulaval.ca> Message-ID: On Wed, 13 Nov 2013, Eric Chamberland wrote: > Hi all, > > we are in the process of adding new functionalities to our parallel > finite-element code and are facing new challenges in computing the number of > non-zeros of a sparse matrix. Since we use Petsc, we want to be able to give > this information before the first assembly (to pass to > Mat{MPI,SEQ}AIJSetPreallocation). > > Willing to rewrite the nonzeros counting algorithm from scratch if necessary, > we asked ourselves a number of questions: > > 1) Do algorithms in general count the *exact* number of non-zeros? or do they > just count a maximum and let Petsc release the memory after the first > assembly? > > 2) How bad is it to reserve too-much non-zeros before the first assembly? If you are counting extra nonzeros - it won't affect the performance [as no additional mallocs will be required]. However the extra memory is not freed at the assembly time. Satish > > 3) Where could we find some hints/articles examples or information about that > kind of algorithm (working in parallel of course) ? > > Here is the context that we have up to now: we can manage a mixture of > interpolations (P1, P2, discontinuous interpolations, etc) and of dimensions > (scalar, 3D, etc) for all unknowns in a global matrix (example: a P2-P1 > [Taylor-Hood] mixed formulation for Stokes problem). > > Since these parameters are determined at run-time (in fact they are user > parameters), we have to think of a general algorithm to compute the "mostly" > exact number of non-zeros for each matrix lines. > > Right now, we have a working, but memory consuming algorithm which computes > the *exact* number on non-zeros per line (we exploded the memory for a problem > with a little more than 2^31 unknowns with it -- using 64 bits indices of > course). In fact, we manage to store the number of neighbors for each type of > mesh components (ex: for a vertice, we have the following number of neighbors: > vertices, edges, faces, and each type of elements). It is costly and moreover > will not handle the new functionalities we want in the code (XFEM for > example). > > Thanks a lot! > > Eric > From anush at bu.edu Wed Nov 13 13:53:07 2013 From: anush at bu.edu (Anush Krishnan) Date: Wed, 13 Nov 2013 14:53:07 -0500 Subject: [petsc-users] Calculating the PETSc index of a DMComposite vector In-Reply-To: <87bo1p9jw5.fsf@jedbrown.org> References: <87habicu41.fsf@jedbrown.org> <877gcd8kxx.fsf@jedbrown.org> <87bo1p9jw5.fsf@jedbrown.org> Message-ID: > > > For memory allocation, I know that every row of C contains exactly 12 > > non-zeros (since I interpolate from a 3x4 grid). But I have no idea if > they > > are in the diagonal or the off-diagonal region of the matrix. For C^T, I > > can't predict beforehand how many non-zeros will be present in each row. > > Most rows will in fact be empty. But if I assume a "nice" solid > boundary, I > > would imagine each velocity grid point should not be influenced by more > > than ~4 body points. > > Can you just make one pass to count and another to insert? > Yes I could do that. I'll try that out. > C * C^T is the influence of particles on each other? Why can't you > compute that analytically (without referencing the grid directly)? > I haven't thought about that. The interpolation coefficients would depend upon the relative positions of the body points with respect to the grid, and the product matrix will depend on their positions with respect to each other. I don't know if this can be done without referencing the grid. I also have a couple of related questions: [1] Suppose my body moves, and the positions of the body points change, will I have to completely deallocate C^T and reallocate the new rows on each process? It's possible that the number of non-zeros on each process can change. What would you suggest as the best strategy to do this? [2] For a two dimensional staggered grid, I would need to interpolate both the x- and y- velocity components. So my system [C][u]=[b] will be: /Cx 0 \ /u\ = /bx\ \0 Cy/ \v/ \by/ Currently, I set up the right hand side as a single vector of size 2*numBodyPoints using VecCreate, and use the global index of the rows when I want to assemble the matrix C. Suppose I made it a DMComposite of bx and by (which I can create as 1-D DAs), how can I reference the correct row of the vector? Also, is it possible to create a composite vector of two vectors that are not distributed arrays? Thank you, Anush -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Wed Nov 13 13:55:27 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 13 Nov 2013 12:55:27 -0700 Subject: [petsc-users] Algorithms for counting nonzeros! In-Reply-To: <5283D416.2080903@giref.ulaval.ca> References: <5283D416.2080903@giref.ulaval.ca> Message-ID: <877gcc6pqo.fsf@jedbrown.org> Eric Chamberland writes: > Hi all, > > we are in the process of adding new functionalities to our parallel > finite-element code and are facing new challenges in computing the > number of non-zeros of a sparse matrix. Since we use Petsc, we want to > be able to give this information before the first assembly (to pass to > Mat{MPI,SEQ}AIJSetPreallocation). > > Willing to rewrite the nonzeros counting algorithm from scratch if > necessary, we asked ourselves a number of questions: > > 1) Do algorithms in general count the *exact* number of non-zeros? or > do they just count a maximum and let Petsc release the memory after the > first assembly? PETSc does not currently use "realloc" to shrink the allocation size. It does, however, compress the used entries up into the top of the array, so the unused part is harmless except insofar as it has been allocated. PETSc developers: we have discussed realloc() in the past and never added it. It would be useful and fairly harmless in this case, since using it does not increase the peak memory usage. It's still important to allocate precisely if you want to be able to use huge pages or run on a system without virtual memory (like Blue Gene), but for normal users, realloc would allow them to allocate twice as much as they will need without consequences (preallocation/assembly comes before preconditioner setup). > 2) How bad is it to reserve too-much non-zeros before the first assembly? > > 3) Where could we find some hints/articles examples or information about > that kind of algorithm (working in parallel of course) ? Here is a clever strategy that often works. http://mail-archive.com/search?l=mid&q=3BEF7FCC-AF4A-461A-B068-9DB01EF94B7A at mcs.anl.gov -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From jedbrown at mcs.anl.gov Wed Nov 13 14:02:42 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 13 Nov 2013 13:02:42 -0700 Subject: [petsc-users] Calculating the PETSc index of a DMComposite vector In-Reply-To: References: <87habicu41.fsf@jedbrown.org> <877gcd8kxx.fsf@jedbrown.org> <87bo1p9jw5.fsf@jedbrown.org> Message-ID: <874n7g6pel.fsf@jedbrown.org> Anush Krishnan writes: >> C * C^T is the influence of particles on each other? Why can't you >> compute that analytically (without referencing the grid directly)? >> > > I haven't thought about that. The interpolation coefficients would depend > upon the relative positions of the body points with respect to the grid, > and the product matrix will depend on their positions with respect to each > other. I don't know if this can be done without referencing the grid. Are you on a structured grid? It seems like you would use fmod to transplant the body point into a region of a reference patch, but I don't see why you need global indices of those grid points. > I also have a couple of related questions: > > [1] Suppose my body moves, and the positions of the body points change, > will I have to completely deallocate C^T and reallocate the new rows on > each process? It's possible that the number of non-zeros on each process > can change. What would you suggest as the best strategy to do this? Just create a new matrix. > [2] For a two dimensional staggered grid, I would need to interpolate both > the x- and y- velocity components. So my system [C][u]=[b] will be: > > /Cx 0 \ /u\ = /bx\ > \0 Cy/ \v/ \by/ > > Currently, I set up the right hand side as a single vector of size > 2*numBodyPoints using VecCreate, and use the global index of the rows when > I want to assemble the matrix C. Suppose I made it a DMComposite of bx and > by (which I can create as 1-D DAs), how can I reference the correct row of > the vector? I would have interlaced the bx and by (dof=2) so you can index it as one vector. Splitting components at collocated points is confusing and bad for performance. > Also, is it possible to create a composite vector of two vectors that are > not distributed arrays? You can do anything with VecScatter. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From iwaddington at gmail.com Wed Nov 13 14:20:21 2013 From: iwaddington at gmail.com (iwaddington .) Date: Wed, 13 Nov 2013 18:20:21 -0200 Subject: [petsc-users] MatSetValuesStencil correct usage Message-ID: Hi everyone, I am having trouble understanding the usage of MatSetValuesStencil. My doubt is in the roles played by the arguments idxm and idxn, which are passed as arguments according to the function documentation. For example, as I was reading example file "src/ksp/ksp/examples/tutorials/ex32.c.html" line 191 I was wondering, ok, we want to insert a 1x5 submatrix inside our global matrix, and we somehow have to pass the grid coordinates of the points we want to insert, so why is the "row" variable needed, the "col" vector seems to contain all the grid coordinates of all the 5 points we are inserting. I know this doubt will sound trivial to some of you, but I am learning Petsc all by myself, so any help is very much appreciated. Thanks a lot. -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Wed Nov 13 14:24:46 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 13 Nov 2013 13:24:46 -0700 Subject: [petsc-users] MatSetValuesStencil correct usage In-Reply-To: References: Message-ID: <871u2k6odt.fsf@jedbrown.org> "iwaddington ." writes: > Hi everyone, I am having trouble understanding the usage of > MatSetValuesStencil. My doubt is in the roles played by the arguments idxm > and idxn, which are passed as arguments according to the function > documentation. For example, as I was reading example file > "src/ksp/ksp/examples/tutorials/ex32.c.html" > line 191 I was wondering, ok, we want to insert a 1x5 submatrix inside our > global matrix, and we somehow have to pass the grid coordinates of the > points we want to insert, so why is the "row" variable needed, the "col" > vector seems to contain all the grid coordinates of all the 5 points we are > inserting. It is assembling a matrix: the action at a point (row) depends on the values at several other points (col[]). -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From bsmith at mcs.anl.gov Wed Nov 13 14:34:48 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Wed, 13 Nov 2013 14:34:48 -0600 Subject: [petsc-users] MatSetValuesStencil correct usage In-Reply-To: References: Message-ID: On Nov 13, 2013, at 2:20 PM, iwaddington . wrote: > Hi everyone, I am having trouble understanding the usage of MatSetValuesStencil. My doubt is in the roles played by the arguments idxm and idxn, which are passed as arguments according to the function documentation. For example, as I was reading example file > "src/ksp/ksp/examples/tutorials/ex32.c.html" line 191 I was wondering, ok, we want to insert a 1x5 submatrix inside our global matrix, and we somehow have to pass the grid coordinates of the points we want to insert, so why is the "row" variable needed, the "col" vector seems to contain all the grid coordinates of all the 5 points we are inserting. \ The col vector only contains the grid coordinates (i,j,k,color) of the 5 columns of the matrix. The row contains the grid coordinates of the single row. The i,j,k,color get mapped to a single location in a vector. Every entry in a matrix has TWO single coordinates into a vector: the I,J coordinates in matrix. So the row i,j,k,color -> I and a column i,j,k,color coordinate -> J. Barry > I know this doubt will sound trivial to some of you, but I am learning Petsc all by myself, so any help is very much appreciated. > > Thanks a lot. From jedbrown at mcs.anl.gov Wed Nov 13 15:06:09 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 13 Nov 2013 14:06:09 -0700 Subject: [petsc-users] solving stokes-like equation in 3D staggered grid in irregular domain with petscsection In-Reply-To: References: <874n7hb7zj.fsf@jedbrown.org> Message-ID: <87y54s57we.fsf@jedbrown.org> Bishesh Khanal writes: > Within A, for now, I can consider mu to be constant, although later if > possible it can be a variable even a tensor to describe anisotropy. But to > start with I want put this a constant. > The original equations start with mu (grad(u) + grad(u)^T) but then > simplifications occur due to div(u) = f2 Rework that step in case of variable mu. > I'm mostly interested in the phenomenon in A with my model, here B is the > extension of the very irregular domain of A to get a cuboid. Here, in B I > release the div(u) = f2 constraint and just put a regularisation to > penalize large deformation. What is of importance here is to compensate the > net volume expansion in domain A by corresponding contraction in domain B > so that the boundaries of the cuboid do not move. It does not particularly > represent any physics except probably that it gives me a velocity field > having a certain divergence field that penalizes big deformations. Okay, sounds like it's already an artificial equation, so you should be able to leave in a normal equation for p, with a big mass matrix on the diagonal, div(mu(grad(u))) - grad(p) = f1 div(u) - c(x) p = f2 c(x) = 0 in domain A and c(x) is large (the inverse of the second Lam? parameter) in domain B. > I do not know much about FEM. But some of the reasons why I have avoided it > in this particular problem are: (Please correct me on any of the following > points if they are wrong) > 1. The inputs f1 and f2 are 3D images (in average of size 200^3) that come > from other image processing pipeline; it's important that I constrain u at > each voxel for div(u) = f2 in domain A. I am trying to avoid having to get > the meshing from the 3D image(with very detailed structures), then go back > to the image from the obtained u again because I have to use the obtained u > to warp the image, transport other parameters again with u in the image > space and again obtain new f1 and f2 images. Then iterate this few times. Okay, there's nothing wrong with that. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From iwaddington at gmail.com Wed Nov 13 15:39:15 2013 From: iwaddington at gmail.com (iwaddington .) Date: Wed, 13 Nov 2013 19:39:15 -0200 Subject: [petsc-users] MatSetValuesStencil correct usage In-Reply-To: References: Message-ID: Thanks very much, now I get it. On Wed, Nov 13, 2013 at 6:34 PM, Barry Smith wrote: > > On Nov 13, 2013, at 2:20 PM, iwaddington . wrote: > > > Hi everyone, I am having trouble understanding the usage of > MatSetValuesStencil. My doubt is in the roles played by the arguments idxm > and idxn, which are passed as arguments according to the function > documentation. For example, as I was reading example file > > "src/ksp/ksp/examples/tutorials/ex32.c.html" line 191 I was wondering, > ok, we want to insert a 1x5 submatrix inside our global matrix, and we > somehow have to pass the grid coordinates of the points we want to insert, > so why is the "row" variable needed, the "col" vector seems to contain all > the grid coordinates of all the 5 points we are inserting. > \ > The col vector only contains the grid coordinates (i,j,k,color) of the > 5 columns of the matrix. The row contains the grid coordinates of the > single row. > > The i,j,k,color get mapped to a single location in a vector. Every > entry in a matrix has TWO single coordinates into a vector: the I,J > coordinates in matrix. So > > the row i,j,k,color -> I and a column i,j,k,color coordinate -> J. > > Barry > > > I know this doubt will sound trivial to some of you, but I am learning > Petsc all by myself, so any help is very much appreciated. > > > > Thanks a lot. > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From anush at bu.edu Wed Nov 13 16:29:08 2013 From: anush at bu.edu (Anush Krishnan) Date: Wed, 13 Nov 2013 17:29:08 -0500 Subject: [petsc-users] Calculating the PETSc index of a DMComposite vector In-Reply-To: <874n7g6pel.fsf@jedbrown.org> References: <87habicu41.fsf@jedbrown.org> <877gcd8kxx.fsf@jedbrown.org> <87bo1p9jw5.fsf@jedbrown.org> <874n7g6pel.fsf@jedbrown.org> Message-ID: > > > [1] Suppose my body moves, and the positions of the body points change, > > will I have to completely deallocate C^T and reallocate the new rows on > > each process? It's possible that the number of non-zeros on each process > > can change. What would you suggest as the best strategy to do this? > > Just create a new matrix. > Is this done using MatDestroy and MatCreate? I would need to do this at every timestep. And suppose C^T was part of a bigger matrix Q = [G C^T], and I only needed to change C^T (which has much fewer columns and non-zeros compared to G) at every time step, is it possible to handle that separately? > I would have interlaced the bx and by (dof=2) so you can index it as one > vector. Splitting components at collocated points is confusing and bad > for performance. > > > Also, is it possible to create a composite vector of two vectors that are > > not distributed arrays? > > You can do anything with VecScatter. > Jed, thanks again for patiently answering my questions. They've been very helpful. Anush -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Wed Nov 13 16:37:38 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 13 Nov 2013 15:37:38 -0700 Subject: [petsc-users] Calculating the PETSc index of a DMComposite vector In-Reply-To: References: <87habicu41.fsf@jedbrown.org> <877gcd8kxx.fsf@jedbrown.org> <87bo1p9jw5.fsf@jedbrown.org> <874n7g6pel.fsf@jedbrown.org> Message-ID: <87k3gc53nx.fsf@jedbrown.org> Anush Krishnan writes: >> >> > [1] Suppose my body moves, and the positions of the body points change, >> > will I have to completely deallocate C^T and reallocate the new rows on >> > each process? It's possible that the number of non-zeros on each process >> > can change. What would you suggest as the best strategy to do this? >> >> Just create a new matrix. >> > > Is this done using MatDestroy and MatCreate? I would need to do this at > every timestep. And suppose C^T was part of a bigger matrix Q = [G C^T], > and I only needed to change C^T (which has much fewer columns and non-zeros > compared to G) at every time step, is it possible to handle that separately? MatCreate is cheap compared to communicating entries, which is usually cheap compared to a solve or a matrix-matrix multiply. If I were writing this and was concerned about performance or memory use, I would assemble Q in one shot, rather than assembling the pieces like you do. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From bisheshkh at gmail.com Thu Nov 14 06:05:27 2013 From: bisheshkh at gmail.com (Bishesh Khanal) Date: Thu, 14 Nov 2013 13:05:27 +0100 Subject: [petsc-users] solving stokes-like equation in 3D staggered grid in irregular domain with petscsection In-Reply-To: <87y54s57we.fsf@jedbrown.org> References: <874n7hb7zj.fsf@jedbrown.org> <87y54s57we.fsf@jedbrown.org> Message-ID: On Wed, Nov 13, 2013 at 10:06 PM, Jed Brown wrote: > Bishesh Khanal writes: > > > Within A, for now, I can consider mu to be constant, although later if > > possible it can be a variable even a tensor to describe anisotropy. But > to > > start with I want put this a constant. > > The original equations start with mu (grad(u) + grad(u)^T) but then > > simplifications occur due to div(u) = f2 > > Rework that step in case of variable mu. > > > I'm mostly interested in the phenomenon in A with my model, here B is the > > extension of the very irregular domain of A to get a cuboid. Here, in B I > > release the div(u) = f2 constraint and just put a regularisation to > > penalize large deformation. What is of importance here is to compensate > the > > net volume expansion in domain A by corresponding contraction in domain B > > so that the boundaries of the cuboid do not move. It does not > particularly > > represent any physics except probably that it gives me a velocity field > > having a certain divergence field that penalizes big deformations. > > Okay, sounds like it's already an artificial equation, so you should be > able to leave in a normal equation for p, with a big mass matrix on the > diagonal, > > div(mu(grad(u))) - grad(p) = f1 > div(u) - c(x) p = f2 > > c(x) = 0 in domain A and c(x) is large (the inverse of the second Lam? > parameter) in domain B. > > Thanks, this looks quite reasonable. I'll try to experiment with it. > > I do not know much about FEM. But some of the reasons why I have avoided > it > > in this particular problem are: (Please correct me on any of the > following > > points if they are wrong) > > 1. The inputs f1 and f2 are 3D images (in average of size 200^3) that > come > > from other image processing pipeline; it's important that I constrain u > at > > each voxel for div(u) = f2 in domain A. I am trying to avoid having to > get > > the meshing from the 3D image(with very detailed structures), then go > back > > to the image from the obtained u again because I have to use the > obtained u > > to warp the image, transport other parameters again with u in the image > > space and again obtain new f1 and f2 images. Then iterate this few times. > > Okay, there's nothing wrong with that. > Thanks for the confirmation. -------------- next part -------------- An HTML attachment was scrubbed... URL: From bisheshkh at gmail.com Thu Nov 14 06:14:13 2013 From: bisheshkh at gmail.com (Bishesh Khanal) Date: Thu, 14 Nov 2013 13:14:13 +0100 Subject: [petsc-users] solving stokes-like equation in 3D staggered grid in irregular domain with petscsection In-Reply-To: References: <874n7hb7zj.fsf@jedbrown.org> Message-ID: On Wed, Nov 13, 2013 at 12:00 AM, Dave May wrote: > Since your using staggered grids, the physical boundary of your domain > will be approximate by a "stair-case" type of boundary. (Correct me if this > is not what you were thinking to do) Thus, imposing traction boundary > conditions on the stair-case boundary should be no more complex than it was > in your standard cube domain with staggered grids. The only exception is > that you have to implement, in a cell-wise manner, the imposition of the > traction condition. The functionality should already exist in your original > cube staggered grid implementation, but possible the implementation of this > boundary condition was done "wall-wise" rather than cell-wise. > > Thanks Dave. I have not implemented traction condition on the cuboid (A U B) walls till now because for the cuboid walls it's a Dirichlet condition. Traction condition would be needed only if I want to solve for the irregularly shaped domain A. I have to solve for many different cases, where the shape of A will keep on changing. (The input for the domain is a 3D binary mask). So it's hard for me to see the generic method to take care of the each boundary cell for the traction condition, but I take your word on its possibility and would discuss it further here if I go in that direction of solving the system for just the domain A. As for now, I would like to solve for both A and B domains in the way Jed suggested. There are few other reasons for this coming from my other image processing pipelines. > > > > 3. I'm trying to put both domains in a single matrix to avoid the >> > difficulty I would have if I want to consider only the domain A. In this >> > case I would need a traction free boundary condition on the irregular >> > boundary of domain A, and it seems a bit too challenging for me to >> > incorporate it with the staggered grid. If there is an idea to implement >> > this and if you think this could be more suitable than the approach in 2 >> > above, I would like to learn about that too! >> >> Complexity of implementing boundary conditions on staggered grids is one >> reason some people turn to other discretization technology, such as >> finite elements. >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Thu Nov 14 12:02:25 2013 From: irving at naml.us (Geoffrey Irving) Date: Thu, 14 Nov 2013 10:02:25 -0800 Subject: [petsc-users] proper way to make petsc throw exceptions on errors? Message-ID: Is there a safe way to make petsc throw exceptions from C++ code, without interfering with the same passage from errors through C code? The existence of PETSC_ERROR_IN_CXX and PetscPushErrorHandler seems promising? Could I push a C++ error handler at the top of a C++ function and pop it and the end, for example? Note that "no" is a fine answer. :) I only want to do this if there's a way to do it cleanly. The alternative is wrapping a CHECK macro around all my calls to petsc from C++, which is far superior to a fancy but broken solution. Thanks, Geoffrey From knepley at gmail.com Thu Nov 14 12:06:23 2013 From: knepley at gmail.com (Matthew Knepley) Date: Thu, 14 Nov 2013 12:06:23 -0600 Subject: [petsc-users] proper way to make petsc throw exceptions on errors? In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 12:02 PM, Geoffrey Irving wrote: > Is there a safe way to make petsc throw exceptions from C++ code, > without interfering with the same passage from errors through C code? > The existence of PETSC_ERROR_IN_CXX and PetscPushErrorHandler seems > promising? Could I push a C++ error handler at the top of a C++ > function and pop it and the end, for example? > PyLith uses: #define PYLITH_CHECK_ERROR(err) do {if (PetscUnlikely(err)) {PetscError(PETSC_COMM_SELF,__LINE__,PETSC_FUNCTION_NAME,__FILE__,err,PETSC_ERROR_IN_CXX,0);}} while(0) #define PYLITH_CHECK_ERROR_MSG(err, msg) \ if (err) { \ PetscError(PETSC_COMM_SELF,__LINE__,__FUNCT__,__FILE__,err,PETSC_ERROR_IN_CXX, 0, " "); \ throw std::runtime_error(msg); } Matt > Note that "no" is a fine answer. :) I only want to do this if there's > a way to do it cleanly. The alternative is wrapping a CHECK macro > around all my calls to petsc from C++, which is far superior to a > fancy but broken solution. > > Thanks, > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Thu Nov 14 12:15:45 2013 From: irving at naml.us (Geoffrey Irving) Date: Thu, 14 Nov 2013 10:15:45 -0800 Subject: [petsc-users] proper way to make petsc throw exceptions on errors? In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 10:06 AM, Matthew Knepley wrote: > On Thu, Nov 14, 2013 at 12:02 PM, Geoffrey Irving wrote: >> >> Is there a safe way to make petsc throw exceptions from C++ code, >> without interfering with the same passage from errors through C code? >> The existence of PETSC_ERROR_IN_CXX and PetscPushErrorHandler seems >> promising? Could I push a C++ error handler at the top of a C++ >> function and pop it and the end, for example? > > > PyLith uses: > > #define PYLITH_CHECK_ERROR(err) do {if (PetscUnlikely(err)) > {PetscError(PETSC_COMM_SELF,__LINE__,PETSC_FUNCTION_NAME,__FILE__,err,PETSC_ERROR_IN_CXX,0);}} > while(0) > > #define PYLITH_CHECK_ERROR_MSG(err, msg) \ > if (err) { \ > > PetscError(PETSC_COMM_SELF,__LINE__,__FUNCT__,__FILE__,err,PETSC_ERROR_IN_CXX, > 0, " "); \ > throw std::runtime_error(msg); } > > Matt Thanks, that'll work. I guess my hope for a magical solution where I didn't have to call the check macro blinded me to the obvious "pass PETSC_ERROR_IN_CXX" myself path. Minor curiosity: why does the _MSG version not use PetscUnlikely? Geoffrey From knepley at gmail.com Thu Nov 14 12:22:31 2013 From: knepley at gmail.com (Matthew Knepley) Date: Thu, 14 Nov 2013 12:22:31 -0600 Subject: [petsc-users] proper way to make petsc throw exceptions on errors? In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 12:15 PM, Geoffrey Irving wrote: > On Thu, Nov 14, 2013 at 10:06 AM, Matthew Knepley > wrote: > > On Thu, Nov 14, 2013 at 12:02 PM, Geoffrey Irving > wrote: > >> > >> Is there a safe way to make petsc throw exceptions from C++ code, > >> without interfering with the same passage from errors through C code? > >> The existence of PETSC_ERROR_IN_CXX and PetscPushErrorHandler seems > >> promising? Could I push a C++ error handler at the top of a C++ > >> function and pop it and the end, for example? > > > > > > PyLith uses: > > > > #define PYLITH_CHECK_ERROR(err) do {if (PetscUnlikely(err)) > > > {PetscError(PETSC_COMM_SELF,__LINE__,PETSC_FUNCTION_NAME,__FILE__,err,PETSC_ERROR_IN_CXX,0);}} > > while(0) > > > > #define PYLITH_CHECK_ERROR_MSG(err, msg) \ > > if (err) { \ > > > > > PetscError(PETSC_COMM_SELF,__LINE__,__FUNCT__,__FILE__,err,PETSC_ERROR_IN_CXX, > > 0, " "); \ > > throw std::runtime_error(msg); } > > > > Matt > > Thanks, that'll work. I guess my hope for a magical solution where I > didn't have to call the check macro blinded me to the obvious "pass > PETSC_ERROR_IN_CXX" myself path. > > Minor curiosity: why does the _MSG version not use PetscUnlikely? I think I wrote the other one later and did not update. Matt > > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Thu Nov 14 14:42:30 2013 From: irving at naml.us (Geoffrey Irving) Date: Thu, 14 Nov 2013 12:42:30 -0800 Subject: [petsc-users] PetscFEM fe vs. feAux vs. feBd Message-ID: What is the difference between the fe and feAux fields in PetscFEM? Do they (and feBd) have to refer to distinct PetscFE objects, or can I reuse one set of finite element spaces? In particular, based on snes ex12, it doesn't look like feBd contains the logic for what the boundary is. This information appears to be extracted by DMPlexComputeResidualFEM and friends. If so, does that mean PetscFEM doesn't have support for several different boundary routines evaluated over different parts of the boundary? I don't necessarily need this support at the moment, but I want to check whether I'm reading the code correctly. Thanks, Geoffrey From dharmareddy84 at gmail.com Thu Nov 14 14:55:41 2013 From: dharmareddy84 at gmail.com (Dharmendar Reddy) Date: Thu, 14 Nov 2013 14:55:41 -0600 Subject: [petsc-users] DMPlexCreateSubMesh Message-ID: Hello, Looks like there is a change in the DMPlexCreateSubMesh interface. My code was running fine with petsc 3.4.3 but when i switched petsc to next. I am getting runtime error due to DMPlexCreateSubMesh.. The interface changed from: ( http://www.mcs.anl.gov/petsc/petsccurrent/src/dm/impls/plex/plexsubmesh.c.html#DMPlexCreateSubmesh ) 2087: *PetscErrorCode DMPlexCreateSubmesh (DM dm, const char vertexLabel[], PetscInt value, DM *subdm)* to ( https://bitbucket.org/petsc/petsc/src/1d94116c89ca7d6e640482ce2684b8b52bac7dff/src/dm/impls/plex/plexsubmesh.c?at=next#cl-2468 ) PetscErrorCode DMPlexCreateSubmesh(DM dm, DMLabel vertexLabel, PetscInt value, DM *subdm) How should i change my code to make use of the new interface ? I have had issue using DMLabel related function from my fortran code. Thanks Reddy -- ----------------------------------------------------- Dharmendar Reddy Palle Graduate Student Microelectronics Research center, University of Texas at Austin, 10100 Burnet Road, Bldg. 160 MER 2.608F, TX 78758-4445 e-mail: dharmareddy84 at gmail.com Phone: +1-512-350-9082 United States of America. Homepage: https://webspace.utexas.edu/~dpr342 -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Thu Nov 14 15:07:16 2013 From: knepley at gmail.com (Matthew Knepley) Date: Thu, 14 Nov 2013 15:07:16 -0600 Subject: [petsc-users] PetscFEM fe vs. feAux vs. feBd In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 2:42 PM, Geoffrey Irving wrote: > What is the difference between the fe and feAux fields in PetscFEM? > Do they (and feBd) have to refer to distinct PetscFE objects, or can I > reuse one set of finite element spaces? > fe contains the fields that we are solving for, feAux contains the fields we will use the in the problem but not solve for, like coefficients. They can obviously come from different spaces. > In particular, based on snes ex12, it doesn't look like feBd contains > the logic for what the boundary is. This information appears to be > extracted by DMPlexComputeResidualFEM and friends. If so, does that > mean PetscFEM doesn't have support for several different boundary > routines evaluated over different parts of the boundary? I don't > necessarily need this support at the moment, but I want to check > whether I'm reading the code correctly. > Yep, right now there is a single boundary marker and set of pointwise functions. I was not going to make a set of them until I needed it, but we do use that in TS ex11. This would work the same way. Matt > Thanks, > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Thu Nov 14 15:10:18 2013 From: knepley at gmail.com (Matthew Knepley) Date: Thu, 14 Nov 2013 15:10:18 -0600 Subject: [petsc-users] DMPlexCreateSubMesh In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 2:55 PM, Dharmendar Reddy wrote: > Hello, > Looks like there is a change in the DMPlexCreateSubMesh > interface. > > My code was running fine with petsc 3.4.3 but when i switched petsc to > next. I am getting runtime error due to DMPlexCreateSubMesh.. > > The interface changed from: > ( > http://www.mcs.anl.gov/petsc/petsccurrent/src/dm/impls/plex/plexsubmesh.c.html#DMPlexCreateSubmesh > ) > > 2087: *PetscErrorCode DMPlexCreateSubmesh (DM dm, const char vertexLabel[], PetscInt value, DM *subdm)* > > > to > ( > https://bitbucket.org/petsc/petsc/src/1d94116c89ca7d6e640482ce2684b8b52bac7dff/src/dm/impls/plex/plexsubmesh.c?at=next#cl-2468 > ) > > PetscErrorCode DMPlexCreateSubmesh(DM dm, DMLabel vertexLabel, PetscInt value, DM *subdm) > > > How should i change my code to make use of the new interface ? I have had issue using DMLabel related function from my fortran code. > > Okay, we should just fix using DMLabel in Fortran. I will write the binding for DMPlexGetLabel(), and I will define the necessary DMLabel stuff for Fortran. I will mail when its pushed. Matt > Thanks > Reddy > -- > ----------------------------------------------------- > Dharmendar Reddy Palle > Graduate Student > Microelectronics Research center, > University of Texas at Austin, > 10100 Burnet Road, Bldg. 160 > MER 2.608F, TX 78758-4445 > e-mail: dharmareddy84 at gmail.com > Phone: +1-512-350-9082 > United States of America. > Homepage: https://webspace.utexas.edu/~dpr342 > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From dharmareddy84 at gmail.com Thu Nov 14 15:16:19 2013 From: dharmareddy84 at gmail.com (Dharmendar Reddy) Date: Thu, 14 Nov 2013 15:16:19 -0600 Subject: [petsc-users] DMPlexCreateSubMesh In-Reply-To: References: Message-ID: Thanks a lot. Reddy On Thu, Nov 14, 2013 at 3:10 PM, Matthew Knepley wrote: > On Thu, Nov 14, 2013 at 2:55 PM, Dharmendar Reddy > wrote: > >> Hello, >> Looks like there is a change in the DMPlexCreateSubMesh >> interface. >> >> My code was running fine with petsc 3.4.3 but when i switched petsc to >> next. I am getting runtime error due to DMPlexCreateSubMesh.. >> >> The interface changed from: >> ( >> http://www.mcs.anl.gov/petsc/petsccurrent/src/dm/impls/plex/plexsubmesh.c.html#DMPlexCreateSubmesh >> ) >> >> 2087: *PetscErrorCode DMPlexCreateSubmesh (DM dm, const char vertexLabel[], PetscInt value, DM *subdm)* >> >> >> to >> ( >> https://bitbucket.org/petsc/petsc/src/1d94116c89ca7d6e640482ce2684b8b52bac7dff/src/dm/impls/plex/plexsubmesh.c?at=next#cl-2468 >> ) >> >> PetscErrorCode DMPlexCreateSubmesh(DM dm, DMLabel vertexLabel, PetscInt value, DM *subdm) >> >> >> How should i change my code to make use of the new interface ? I have had issue using DMLabel related function from my fortran code. >> >> > Okay, we should just fix using DMLabel in Fortran. I will write the > binding for DMPlexGetLabel(), and I will > define the necessary DMLabel stuff for Fortran. I will mail when its > pushed. > > Matt > > >> Thanks >> Reddy >> -- >> ----------------------------------------------------- >> Dharmendar Reddy Palle >> Graduate Student >> Microelectronics Research center, >> University of Texas at Austin, >> 10100 Burnet Road, Bldg. 160 >> MER 2.608F, TX 78758-4445 >> e-mail: dharmareddy84 at gmail.com >> Phone: +1-512-350-9082 >> United States of America. >> Homepage: https://webspace.utexas.edu/~dpr342 >> > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -- ----------------------------------------------------- Dharmendar Reddy Palle Graduate Student Microelectronics Research center, University of Texas at Austin, 10100 Burnet Road, Bldg. 160 MER 2.608F, TX 78758-4445 e-mail: dharmareddy84 at gmail.com Phone: +1-512-350-9082 United States of America. Homepage: https://webspace.utexas.edu/~dpr342 -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Thu Nov 14 19:32:42 2013 From: irving at naml.us (Geoffrey Irving) Date: Thu, 14 Nov 2013 17:32:42 -0800 Subject: [petsc-users] snes ex12 question about DMPlexProjectFunctionLocal Message-ID: If this line in ex12 is part of Dirichlet boundary conditions, why is it done unconditionally even for Neumann? Is this a mistake that fails to test correct Neumann conditions? ierr = DMPlexProjectFunctionLocal(dm, user.fe, user.exactFuncs, INSERT_BC_VALUES, userJ.u);CHKERRQ(ierr); Geoffrey From knepley at gmail.com Thu Nov 14 19:35:08 2013 From: knepley at gmail.com (Matthew Knepley) Date: Thu, 14 Nov 2013 19:35:08 -0600 Subject: [petsc-users] snes ex12 question about DMPlexProjectFunctionLocal In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 7:32 PM, Geoffrey Irving wrote: > If this line in ex12 is part of Dirichlet boundary conditions, why is > it done unconditionally even for Neumann? Is this a mistake that > fails to test correct Neumann conditions? > > ierr = DMPlexProjectFunctionLocal(dm, user.fe, user.exactFuncs, > INSERT_BC_VALUES, userJ.u);CHKERRQ(ierr); > No, its just idempotent in that case. Matt Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Thu Nov 14 19:38:21 2013 From: irving at naml.us (Geoffrey Irving) Date: Thu, 14 Nov 2013 17:38:21 -0800 Subject: [petsc-users] snes ex12 question about DMPlexProjectFunctionLocal In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 5:35 PM, Matthew Knepley wrote: > On Thu, Nov 14, 2013 at 7:32 PM, Geoffrey Irving wrote: >> >> If this line in ex12 is part of Dirichlet boundary conditions, why is >> it done unconditionally even for Neumann? Is this a mistake that >> fails to test correct Neumann conditions? >> >> ierr = DMPlexProjectFunctionLocal(dm, user.fe, user.exactFuncs, >> INSERT_BC_VALUES, userJ.u);CHKERRQ(ierr); > > No, its just idempotent in that case. That's what I meant by fail to test: if there was a bug related to null spaces, it wouldn't show up in this example. Thanks, Geoffrey From mrosso at uci.edu Thu Nov 14 19:49:31 2013 From: mrosso at uci.edu (Michele Rosso) Date: Thu, 14 Nov 2013 17:49:31 -0800 Subject: [petsc-users] KSPSetDM problem Message-ID: <52857DAB.9090709@uci.edu> Hi, I am solving a Poisson equation (cell-centered discretization) with multigrid. I am using a 2D version of ksp/ksp/examples/tutorials/ex34.c as a reference. The following code performs the same calculations as ex34: call PetscInitialize(PETSC_NULL_CHARACTER,ierr) call DMDACreate2d(PETSC_COMM_WORLD, DMDA_BOUNDARY_NONE, DMDA_BOUNDARY_NONE,& & DMDA_STENCIL_STAR, 16, 16, PETSC_DECIDE, PETSC_DECIDE, & & 1, 1, PETSC_NULL_INTEGER, PETSC_NULL_INTEGER,da,ierr) call DMDASetInterpolationType(da, DMDA_Q0, ierr); call DMCreateGlobalVector(da,b,ierr) call create_rhs(da,b) call DMCreateMatrix(da,'aij',A,ierr) call create_matrix(da,A) call VecDuplicate(b,x,ierr) ! Create solver call KSPCreate(PETSC_COMM_WORLD,solver,ierr) call KSPSetDM(solver,da,ierr) call KSPSetDMActive(solver,PETSC_FALSE,ierr) call KSPSetOperators(solver,A,A,SAME_NONZERO_PATTERN,ierr) call KSPSetType(solver,'cg',ierr) call KSPSetNormType(solver,KSP_NORM_UNPRECONDITIONED,ierr) call KSPSetFromOptions(solver,ierr) call KSPSolve(solver,b,x,ierr) rerr = error(da,x) call PetscFinalize(ierr) Both ex34 and my code produce exactly the same results except when I run with -ksp_view -ksp_monitor_true_residual -ksp_converged_reason -ksp_type cg -pc_type mg -pc_mg_levels 4 In this case my code results in the following error, despite KSPSetDMActive is used after KSPSetDM: [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Object is in wrong state! [0]PETSC ERROR: You called KSPSetDM() but did not use DMKSPSetComputeOperators() or KSPSetDMActive(dm,PETSC_FALSE);! [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 [0]PETSC ERROR: See docs/changes/index.html for recent updates. [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [0]PETSC ERROR: See docs/index.html for manual pages. [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: ./test on a linux-gnu-dbg named enterprise-A by mic Thu Nov 14 17:38:23 2013 [0]PETSC ERROR: Libraries linked from /opt/petsc/petsc-3.4.3/linux-gnu-dbg/lib [0]PETSC ERROR: Configure run at Tue Oct 29 22:43:45 2013 [0]PETSC ERROR: Configure options --known-mpi-shared="0 " --known-memcmp-ok --with-debugging="1 " --with-fortran-datatypes --with-shared-libraries=0 --with-dynamic-loading=0 --with-mpi-compilers="1 " --download-blacs="1 " --download-superlu_dist="1 " --download-metis="1 " --download-parmetis="1 " --download-ml=1 PETSC_ARCH=linux-gnu-dbg [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: KSPSetUp() line 230 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c [0]PETSC ERROR: PCSetUp_MG() line 730 in /opt/petsc/petsc-3.4.3/src/ksp/pc/impls/mg/mg.c [0]PETSC ERROR: PCSetUp() line 890 in /opt/petsc/petsc-3.4.3/src/ksp/pc/interface/precon.c [0]PETSC ERROR: KSPSetUp() line 278 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c [0]PETSC ERROR: KSPSolve() line 399 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c Am I missing anything? Thank you, Michele -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Thu Nov 14 20:08:10 2013 From: knepley at gmail.com (Matthew Knepley) Date: Thu, 14 Nov 2013 20:08:10 -0600 Subject: [petsc-users] snes ex12 question about DMPlexProjectFunctionLocal In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 7:38 PM, Geoffrey Irving wrote: > On Thu, Nov 14, 2013 at 5:35 PM, Matthew Knepley > wrote: > > On Thu, Nov 14, 2013 at 7:32 PM, Geoffrey Irving wrote: > >> > >> If this line in ex12 is part of Dirichlet boundary conditions, why is > >> it done unconditionally even for Neumann? Is this a mistake that > >> fails to test correct Neumann conditions? > >> > >> ierr = DMPlexProjectFunctionLocal(dm, user.fe, user.exactFuncs, > >> INSERT_BC_VALUES, userJ.u);CHKERRQ(ierr); > > > > No, its just idempotent in that case. > > That's what I meant by fail to test: if there was a bug related to > null spaces, it wouldn't show up in this example. > I do not understand. If you have Neumann conditions, nothing happens, so it would check that case. Matt > Thanks, > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From bsmith at mcs.anl.gov Thu Nov 14 20:11:19 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Thu, 14 Nov 2013 20:11:19 -0600 Subject: [petsc-users] KSPSetDM problem In-Reply-To: <52857DAB.9090709@uci.edu> References: <52857DAB.9090709@uci.edu> Message-ID: Since you set the DM inactive the PCSetUp_MG() has no code to compute the coarser grid operators. You need to either 1) provide a DMKSPSetComputeOperators () function that will compute the operator for all levels or 2) run with -pc_mg_galerkin to have the coarse grid operators computed by multiplying the fine grid operator by the restriction/interpolation. Barry On Nov 14, 2013, at 7:49 PM, Michele Rosso wrote: > Hi, > > I am solving a Poisson equation (cell-centered discretization) with multigrid. > I am using a 2D version of ksp/ksp/examples/tutorials/ex34.c as a reference. > The following code performs the same calculations as ex34: > > call PetscInitialize(PETSC_NULL_CHARACTER,ierr) > > call DMDACreate2d(PETSC_COMM_WORLD, DMDA_BOUNDARY_NONE, DMDA_BOUNDARY_NONE,& > & DMDA_STENCIL_STAR, 16, 16, PETSC_DECIDE, PETSC_DECIDE, & > & 1, 1, PETSC_NULL_INTEGER, PETSC_NULL_INTEGER,da,ierr) > call DMDASetInterpolationType(da, DMDA_Q0, ierr); > call DMCreateGlobalVector(da,b,ierr) > call create_rhs(da,b) > call DMCreateMatrix(da,'aij',A,ierr) > call create_matrix(da,A) > call VecDuplicate(b,x,ierr) > > ! Create solver > call KSPCreate(PETSC_COMM_WORLD,solver,ierr) > call KSPSetDM(solver,da,ierr) > call KSPSetDMActive(solver,PETSC_FALSE,ierr) > call KSPSetOperators(solver,A,A,SAME_NONZERO_PATTERN,ierr) > call KSPSetType(solver,'cg',ierr) > call KSPSetNormType(solver,KSP_NORM_UNPRECONDITIONED,ierr) > call KSPSetFromOptions(solver,ierr) > call KSPSolve(solver,b,x,ierr) > > rerr = error(da,x) > > call PetscFinalize(ierr) > > > Both ex34 and my code produce exactly the same results except when I run with > > -ksp_view -ksp_monitor_true_residual -ksp_converged_reason -ksp_type cg -pc_type mg -pc_mg_levels 4 > > In this case my code results in the following error, despite KSPSetDMActive is used after KSPSetDM: > > [0]PETSC ERROR: --------------------- Error Message ------------------------------------ > [0]PETSC ERROR: Object is in wrong state! > [0]PETSC ERROR: You called KSPSetDM() but did not use DMKSPSetComputeOperators() or KSPSetDMActive(dm,PETSC_FALSE);! > [0]PETSC ERROR: ------------------------------------------------------------------------ > [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > [0]PETSC ERROR: See docs/changes/index.html for recent updates. > [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > [0]PETSC ERROR: See docs/index.html for manual pages. > [0]PETSC ERROR: ------------------------------------------------------------------------ > [0]PETSC ERROR: ./test on a linux-gnu-dbg named enterprise-A by mic Thu Nov 14 17:38:23 2013 > [0]PETSC ERROR: Libraries linked from /opt/petsc/petsc-3.4.3/linux-gnu-dbg/lib > [0]PETSC ERROR: Configure run at Tue Oct 29 22:43:45 2013 > [0]PETSC ERROR: Configure options --known-mpi-shared="0 " --known-memcmp-ok --with-debugging="1 " --with-fortran-datatypes --with-shared-libraries=0 --with-dynamic-loading=0 --with-mpi-compilers="1 " --download-blacs="1 " --download-superlu_dist="1 " --download-metis="1 " --download-parmetis="1 " --download-ml=1 PETSC_ARCH=linux-gnu-dbg > [0]PETSC ERROR: ------------------------------------------------------------------------ > [0]PETSC ERROR: KSPSetUp() line 230 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c > [0]PETSC ERROR: PCSetUp_MG() line 730 in /opt/petsc/petsc-3.4.3/src/ksp/pc/impls/mg/mg.c > [0]PETSC ERROR: PCSetUp() line 890 in /opt/petsc/petsc-3.4.3/src/ksp/pc/interface/precon.c > [0]PETSC ERROR: KSPSetUp() line 278 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c > [0]PETSC ERROR: KSPSolve() line 399 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c > > > Am I missing anything? > Thank you, > > Michele > > > > > > > From irving at naml.us Thu Nov 14 20:16:24 2013 From: irving at naml.us (Geoffrey Irving) Date: Thu, 14 Nov 2013 18:16:24 -0800 Subject: [petsc-users] snes ex12 question about DMPlexProjectFunctionLocal In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 6:08 PM, Matthew Knepley wrote: > On Thu, Nov 14, 2013 at 7:38 PM, Geoffrey Irving wrote: >> >> On Thu, Nov 14, 2013 at 5:35 PM, Matthew Knepley >> wrote: >> > On Thu, Nov 14, 2013 at 7:32 PM, Geoffrey Irving wrote: >> >> >> >> If this line in ex12 is part of Dirichlet boundary conditions, why is >> >> it done unconditionally even for Neumann? Is this a mistake that >> >> fails to test correct Neumann conditions? >> >> >> >> ierr = DMPlexProjectFunctionLocal(dm, user.fe, user.exactFuncs, >> >> INSERT_BC_VALUES, userJ.u);CHKERRQ(ierr); >> > >> > No, its just idempotent in that case. >> >> That's what I meant by fail to test: if there was a bug related to >> null spaces, it wouldn't show up in this example. > > > I do not understand. If you have Neumann conditions, nothing happens, so > it would check that case. I may be misunderstanding something, but wouldn't nothing happen only if the rest of the code is correct? Geoffrey From knepley at gmail.com Thu Nov 14 20:22:36 2013 From: knepley at gmail.com (Matthew Knepley) Date: Thu, 14 Nov 2013 20:22:36 -0600 Subject: [petsc-users] snes ex12 question about DMPlexProjectFunctionLocal In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 8:16 PM, Geoffrey Irving wrote: > On Thu, Nov 14, 2013 at 6:08 PM, Matthew Knepley > wrote: > > On Thu, Nov 14, 2013 at 7:38 PM, Geoffrey Irving wrote: > >> > >> On Thu, Nov 14, 2013 at 5:35 PM, Matthew Knepley > >> wrote: > >> > On Thu, Nov 14, 2013 at 7:32 PM, Geoffrey Irving > wrote: > >> >> > >> >> If this line in ex12 is part of Dirichlet boundary conditions, why is > >> >> it done unconditionally even for Neumann? Is this a mistake that > >> >> fails to test correct Neumann conditions? > >> >> > >> >> ierr = DMPlexProjectFunctionLocal(dm, user.fe, user.exactFuncs, > >> >> INSERT_BC_VALUES, userJ.u);CHKERRQ(ierr); > >> > > >> > No, its just idempotent in that case. > >> > >> That's what I meant by fail to test: if there was a bug related to > >> null spaces, it wouldn't show up in this example. > > > > > > I do not understand. If you have Neumann conditions, nothing happens, so > > it would check that case. > > I may be misunderstanding something, but wouldn't nothing happen only > if the rest of the code is correct? Nothing would happen as long as you have not set any Dirichlet conditions. Matt > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From mrosso at uci.edu Thu Nov 14 20:29:08 2013 From: mrosso at uci.edu (Michele Rosso) Date: Thu, 14 Nov 2013 18:29:08 -0800 Subject: [petsc-users] KSPSetDM problem In-Reply-To: References: <52857DAB.9090709@uci.edu> Message-ID: <528586F4.5030000@uci.edu> Barry, thank you. I will have to go with option 1) since with 2) the number of iterations is almost the double (I tried with ex34). I am testing the piecewise constant interpolation for periodic BCs we discussed about in a previous conversation: in case I am successful, how do I submit the fix? Michele On 11/14/2013 06:11 PM, Barry Smith wrote: > Since you set the DM inactive the PCSetUp_MG() has no code to compute the coarser grid operators. You need to either > > 1) provide a DMKSPSetComputeOperators () function that will compute the operator for all levels or > > 2) run with -pc_mg_galerkin to have the coarse grid operators computed by multiplying the fine grid operator by the restriction/interpolation. > > Barry > > On Nov 14, 2013, at 7:49 PM, Michele Rosso wrote: > >> Hi, >> >> I am solving a Poisson equation (cell-centered discretization) with multigrid. >> I am using a 2D version of ksp/ksp/examples/tutorials/ex34.c as a reference. >> The following code performs the same calculations as ex34: >> >> call PetscInitialize(PETSC_NULL_CHARACTER,ierr) >> >> call DMDACreate2d(PETSC_COMM_WORLD, DMDA_BOUNDARY_NONE, DMDA_BOUNDARY_NONE,& >> & DMDA_STENCIL_STAR, 16, 16, PETSC_DECIDE, PETSC_DECIDE, & >> & 1, 1, PETSC_NULL_INTEGER, PETSC_NULL_INTEGER,da,ierr) >> call DMDASetInterpolationType(da, DMDA_Q0, ierr); >> call DMCreateGlobalVector(da,b,ierr) >> call create_rhs(da,b) >> call DMCreateMatrix(da,'aij',A,ierr) >> call create_matrix(da,A) >> call VecDuplicate(b,x,ierr) >> >> ! Create solver >> call KSPCreate(PETSC_COMM_WORLD,solver,ierr) >> call KSPSetDM(solver,da,ierr) >> call KSPSetDMActive(solver,PETSC_FALSE,ierr) >> call KSPSetOperators(solver,A,A,SAME_NONZERO_PATTERN,ierr) >> call KSPSetType(solver,'cg',ierr) >> call KSPSetNormType(solver,KSP_NORM_UNPRECONDITIONED,ierr) >> call KSPSetFromOptions(solver,ierr) >> call KSPSolve(solver,b,x,ierr) >> >> rerr = error(da,x) >> >> call PetscFinalize(ierr) >> >> >> Both ex34 and my code produce exactly the same results except when I run with >> >> -ksp_view -ksp_monitor_true_residual -ksp_converged_reason -ksp_type cg -pc_type mg -pc_mg_levels 4 >> >> In this case my code results in the following error, despite KSPSetDMActive is used after KSPSetDM: >> >> [0]PETSC ERROR: --------------------- Error Message ------------------------------------ >> [0]PETSC ERROR: Object is in wrong state! >> [0]PETSC ERROR: You called KSPSetDM() but did not use DMKSPSetComputeOperators() or KSPSetDMActive(dm,PETSC_FALSE);! >> [0]PETSC ERROR: ------------------------------------------------------------------------ >> [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 >> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >> [0]PETSC ERROR: See docs/index.html for manual pages. >> [0]PETSC ERROR: ------------------------------------------------------------------------ >> [0]PETSC ERROR: ./test on a linux-gnu-dbg named enterprise-A by mic Thu Nov 14 17:38:23 2013 >> [0]PETSC ERROR: Libraries linked from /opt/petsc/petsc-3.4.3/linux-gnu-dbg/lib >> [0]PETSC ERROR: Configure run at Tue Oct 29 22:43:45 2013 >> [0]PETSC ERROR: Configure options --known-mpi-shared="0 " --known-memcmp-ok --with-debugging="1 " --with-fortran-datatypes --with-shared-libraries=0 --with-dynamic-loading=0 --with-mpi-compilers="1 " --download-blacs="1 " --download-superlu_dist="1 " --download-metis="1 " --download-parmetis="1 " --download-ml=1 PETSC_ARCH=linux-gnu-dbg >> [0]PETSC ERROR: ------------------------------------------------------------------------ >> [0]PETSC ERROR: KSPSetUp() line 230 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >> [0]PETSC ERROR: PCSetUp_MG() line 730 in /opt/petsc/petsc-3.4.3/src/ksp/pc/impls/mg/mg.c >> [0]PETSC ERROR: PCSetUp() line 890 in /opt/petsc/petsc-3.4.3/src/ksp/pc/interface/precon.c >> [0]PETSC ERROR: KSPSetUp() line 278 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >> [0]PETSC ERROR: KSPSolve() line 399 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >> >> >> Am I missing anything? >> Thank you, >> >> Michele >> >> >> >> >> >> >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From bsmith at mcs.anl.gov Thu Nov 14 20:51:25 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Thu, 14 Nov 2013 20:51:25 -0600 Subject: [petsc-users] KSPSetDM problem In-Reply-To: <528586F4.5030000@uci.edu> References: <52857DAB.9090709@uci.edu> <528586F4.5030000@uci.edu> Message-ID: https://bitbucket.org/petsc/petsc/wiki/Home#markdown-header-contributing-a-small-patch On Nov 14, 2013, at 8:29 PM, Michele Rosso wrote: > Barry, > > thank you. I will have to go with option 1) since with 2) the number of iterations is almost the double (I tried with ex34). > I am testing the piecewise constant interpolation for periodic BCs we discussed about in a previous conversation: in case I am successful, how do I submit the fix? > > Michele > > On 11/14/2013 06:11 PM, Barry Smith wrote: >> Since you set the DM inactive the PCSetUp_MG() has no code to compute the coarser grid operators. You need to either >> >> 1) provide a DMKSPSetComputeOperators () function that will compute the operator for all levels or >> >> 2) run with -pc_mg_galerkin to have the coarse grid operators computed by multiplying the fine grid operator by the restriction/interpolation. >> >> Barry >> >> On Nov 14, 2013, at 7:49 PM, Michele Rosso >> >> wrote: >> >> >>> Hi, >>> >>> I am solving a Poisson equation (cell-centered discretization) with multigrid. >>> I am using a 2D version of ksp/ksp/examples/tutorials/ex34.c as a reference. >>> The following code performs the same calculations as ex34: >>> >>> call PetscInitialize(PETSC_NULL_CHARACTER,ierr) >>> >>> call DMDACreate2d(PETSC_COMM_WORLD, DMDA_BOUNDARY_NONE, DMDA_BOUNDARY_NONE,& >>> & DMDA_STENCIL_STAR, 16, 16, PETSC_DECIDE, PETSC_DECIDE, & >>> & 1, 1, PETSC_NULL_INTEGER, PETSC_NULL_INTEGER,da,ierr) >>> call DMDASetInterpolationType(da, DMDA_Q0, ierr); >>> call DMCreateGlobalVector(da,b,ierr) >>> call create_rhs(da,b) >>> call DMCreateMatrix(da,'aij',A,ierr) >>> call create_matrix(da,A) >>> call VecDuplicate(b,x,ierr) >>> >>> ! Create solver >>> call KSPCreate(PETSC_COMM_WORLD,solver,ierr) >>> call KSPSetDM(solver,da,ierr) >>> call KSPSetDMActive(solver,PETSC_FALSE,ierr) >>> call KSPSetOperators(solver,A,A,SAME_NONZERO_PATTERN,ierr) >>> call KSPSetType(solver,'cg',ierr) >>> call KSPSetNormType(solver,KSP_NORM_UNPRECONDITIONED,ierr) >>> call KSPSetFromOptions(solver,ierr) >>> call KSPSolve(solver,b,x,ierr) >>> >>> rerr = error(da,x) >>> >>> call PetscFinalize(ierr) >>> >>> >>> Both ex34 and my code produce exactly the same results except when I run with >>> >>> -ksp_view -ksp_monitor_true_residual -ksp_converged_reason -ksp_type cg -pc_type mg -pc_mg_levels 4 >>> >>> In this case my code results in the following error, despite KSPSetDMActive is used after KSPSetDM: >>> >>> [0]PETSC ERROR: --------------------- Error Message ------------------------------------ >>> [0]PETSC ERROR: Object is in wrong state! >>> [0]PETSC ERROR: You called KSPSetDM() but did not use DMKSPSetComputeOperators() or KSPSetDMActive(dm,PETSC_FALSE);! >>> [0]PETSC ERROR: ------------------------------------------------------------------------ >>> [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 >>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>> [0]PETSC ERROR: See docs/index.html for manual pages. >>> [0]PETSC ERROR: ------------------------------------------------------------------------ >>> [0]PETSC ERROR: ./test on a linux-gnu-dbg named enterprise-A by mic Thu Nov 14 17:38:23 2013 >>> [0]PETSC ERROR: Libraries linked from /opt/petsc/petsc-3.4.3/linux-gnu-dbg/lib >>> [0]PETSC ERROR: Configure run at Tue Oct 29 22:43:45 2013 >>> [0]PETSC ERROR: Configure options --known-mpi-shared="0 " --known-memcmp-ok --with-debugging="1 " --with-fortran-datatypes --with-shared-libraries=0 --with-dynamic-loading=0 --with-mpi-compilers="1 " --download-blacs="1 " --download-superlu_dist="1 " --download-metis="1 " --download-parmetis="1 " --download-ml=1 PETSC_ARCH=linux-gnu-dbg >>> [0]PETSC ERROR: ------------------------------------------------------------------------ >>> [0]PETSC ERROR: KSPSetUp() line 230 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >>> [0]PETSC ERROR: PCSetUp_MG() line 730 in /opt/petsc/petsc-3.4.3/src/ksp/pc/impls/mg/mg.c >>> [0]PETSC ERROR: PCSetUp() line 890 in /opt/petsc/petsc-3.4.3/src/ksp/pc/interface/precon.c >>> [0]PETSC ERROR: KSPSetUp() line 278 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >>> [0]PETSC ERROR: KSPSolve() line 399 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >>> >>> >>> Am I missing anything? >>> Thank you, >>> >>> Michele >>> >>> >>> >>> >>> >>> >>> >>> >> > From knepley at gmail.com Fri Nov 15 09:01:29 2013 From: knepley at gmail.com (Matthew Knepley) Date: Fri, 15 Nov 2013 09:01:29 -0600 Subject: [petsc-users] DMPlexCreateSubMesh In-Reply-To: References: Message-ID: On Thu, Nov 14, 2013 at 3:16 PM, Dharmendar Reddy wrote: > Thanks a lot. > Okay, I have pushed it and put in a test at src/dm/impls/plex/examples/tutorials/ex1f90.F I needed to update 'bfort' which makes the Fortran bindings. The easiest thing to do is grab it from githubg (petsc/pkg-sowing) and install into your PETSC_ARCH. I will ask Satish to remake the tarball, so if your install does not work you can just reconfigure after deleting externalpackage/sowing-* and PETSC_ARCH/conf/sowing. Thanks, Matt > Reddy > > > On Thu, Nov 14, 2013 at 3:10 PM, Matthew Knepley wrote: > >> On Thu, Nov 14, 2013 at 2:55 PM, Dharmendar Reddy < >> dharmareddy84 at gmail.com> wrote: >> >>> Hello, >>> Looks like there is a change in the DMPlexCreateSubMesh >>> interface. >>> >>> My code was running fine with petsc 3.4.3 but when i switched petsc to >>> next. I am getting runtime error due to DMPlexCreateSubMesh.. >>> >>> The interface changed from: >>> ( >>> http://www.mcs.anl.gov/petsc/petsccurrent/src/dm/impls/plex/plexsubmesh.c.html#DMPlexCreateSubmesh >>> ) >>> >>> 2087: *PetscErrorCode DMPlexCreateSubmesh (DM dm, const char vertexLabel[], PetscInt value, DM *subdm)* >>> >>> >>> to >>> ( >>> https://bitbucket.org/petsc/petsc/src/1d94116c89ca7d6e640482ce2684b8b52bac7dff/src/dm/impls/plex/plexsubmesh.c?at=next#cl-2468 >>> ) >>> >>> PetscErrorCode DMPlexCreateSubmesh(DM dm, DMLabel vertexLabel, PetscInt value, DM *subdm) >>> >>> >>> How should i change my code to make use of the new interface ? I have had issue using DMLabel related function from my fortran code. >>> >>> >> Okay, we should just fix using DMLabel in Fortran. I will write the >> binding for DMPlexGetLabel(), and I will >> define the necessary DMLabel stuff for Fortran. I will mail when its >> pushed. >> >> Matt >> >> >>> Thanks >>> Reddy >>> -- >>> ----------------------------------------------------- >>> Dharmendar Reddy Palle >>> Graduate Student >>> Microelectronics Research center, >>> University of Texas at Austin, >>> 10100 Burnet Road, Bldg. 160 >>> MER 2.608F, TX 78758-4445 >>> e-mail: dharmareddy84 at gmail.com >>> Phone: +1-512-350-9082 >>> United States of America. >>> Homepage: https://webspace.utexas.edu/~dpr342 >>> >> >> >> >> -- >> What most experimenters take for granted before they begin their >> experiments is infinitely more interesting than any results to which their >> experiments lead. >> -- Norbert Wiener >> > > > > -- > ----------------------------------------------------- > Dharmendar Reddy Palle > Graduate Student > Microelectronics Research center, > University of Texas at Austin, > 10100 Burnet Road, Bldg. 160 > MER 2.608F, TX 78758-4445 > e-mail: dharmareddy84 at gmail.com > Phone: +1-512-350-9082 > United States of America. > Homepage: https://webspace.utexas.edu/~dpr342 > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From mrosso at uci.edu Fri Nov 15 17:34:04 2013 From: mrosso at uci.edu (Michele Rosso) Date: Fri, 15 Nov 2013 15:34:04 -0800 Subject: [petsc-users] KSPSetDM problem In-Reply-To: References: <52857DAB.9090709@uci.edu> <528586F4.5030000@uci.edu> Message-ID: <5286AF6C.7010703@uci.edu> Barry, thank you. I have another question: how does petsc compute the restriction (assuming the user does not supply his own restriction)? Does it use a simple transpose of the interpolation operator? Could you point me to the source code where this is done? Thank you, Michele On 11/14/2013 06:51 PM, Barry Smith wrote: > https://bitbucket.org/petsc/petsc/wiki/Home#markdown-header-contributing-a-small-patch > > On Nov 14, 2013, at 8:29 PM, Michele Rosso wrote: > >> Barry, >> >> thank you. I will have to go with option 1) since with 2) the number of iterations is almost the double (I tried with ex34). >> I am testing the piecewise constant interpolation for periodic BCs we discussed about in a previous conversation: in case I am successful, how do I submit the fix? >> >> Michele >> >> On 11/14/2013 06:11 PM, Barry Smith wrote: >>> Since you set the DM inactive the PCSetUp_MG() has no code to compute the coarser grid operators. You need to either >>> >>> 1) provide a DMKSPSetComputeOperators () function that will compute the operator for all levels or >>> >>> 2) run with -pc_mg_galerkin to have the coarse grid operators computed by multiplying the fine grid operator by the restriction/interpolation. >>> >>> Barry >>> >>> On Nov 14, 2013, at 7:49 PM, Michele Rosso >>> >>> wrote: >>> >>> >>>> Hi, >>>> >>>> I am solving a Poisson equation (cell-centered discretization) with multigrid. >>>> I am using a 2D version of ksp/ksp/examples/tutorials/ex34.c as a reference. >>>> The following code performs the same calculations as ex34: >>>> >>>> call PetscInitialize(PETSC_NULL_CHARACTER,ierr) >>>> >>>> call DMDACreate2d(PETSC_COMM_WORLD, DMDA_BOUNDARY_NONE, DMDA_BOUNDARY_NONE,& >>>> & DMDA_STENCIL_STAR, 16, 16, PETSC_DECIDE, PETSC_DECIDE, & >>>> & 1, 1, PETSC_NULL_INTEGER, PETSC_NULL_INTEGER,da,ierr) >>>> call DMDASetInterpolationType(da, DMDA_Q0, ierr); >>>> call DMCreateGlobalVector(da,b,ierr) >>>> call create_rhs(da,b) >>>> call DMCreateMatrix(da,'aij',A,ierr) >>>> call create_matrix(da,A) >>>> call VecDuplicate(b,x,ierr) >>>> >>>> ! Create solver >>>> call KSPCreate(PETSC_COMM_WORLD,solver,ierr) >>>> call KSPSetDM(solver,da,ierr) >>>> call KSPSetDMActive(solver,PETSC_FALSE,ierr) >>>> call KSPSetOperators(solver,A,A,SAME_NONZERO_PATTERN,ierr) >>>> call KSPSetType(solver,'cg',ierr) >>>> call KSPSetNormType(solver,KSP_NORM_UNPRECONDITIONED,ierr) >>>> call KSPSetFromOptions(solver,ierr) >>>> call KSPSolve(solver,b,x,ierr) >>>> >>>> rerr = error(da,x) >>>> >>>> call PetscFinalize(ierr) >>>> >>>> >>>> Both ex34 and my code produce exactly the same results except when I run with >>>> >>>> -ksp_view -ksp_monitor_true_residual -ksp_converged_reason -ksp_type cg -pc_type mg -pc_mg_levels 4 >>>> >>>> In this case my code results in the following error, despite KSPSetDMActive is used after KSPSetDM: >>>> >>>> [0]PETSC ERROR: --------------------- Error Message ------------------------------------ >>>> [0]PETSC ERROR: Object is in wrong state! >>>> [0]PETSC ERROR: You called KSPSetDM() but did not use DMKSPSetComputeOperators() or KSPSetDMActive(dm,PETSC_FALSE);! >>>> [0]PETSC ERROR: ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 >>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >>>> [0]PETSC ERROR: See docs/index.html for manual pages. >>>> [0]PETSC ERROR: ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: ./test on a linux-gnu-dbg named enterprise-A by mic Thu Nov 14 17:38:23 2013 >>>> [0]PETSC ERROR: Libraries linked from /opt/petsc/petsc-3.4.3/linux-gnu-dbg/lib >>>> [0]PETSC ERROR: Configure run at Tue Oct 29 22:43:45 2013 >>>> [0]PETSC ERROR: Configure options --known-mpi-shared="0 " --known-memcmp-ok --with-debugging="1 " --with-fortran-datatypes --with-shared-libraries=0 --with-dynamic-loading=0 --with-mpi-compilers="1 " --download-blacs="1 " --download-superlu_dist="1 " --download-metis="1 " --download-parmetis="1 " --download-ml=1 PETSC_ARCH=linux-gnu-dbg >>>> [0]PETSC ERROR: ------------------------------------------------------------------------ >>>> [0]PETSC ERROR: KSPSetUp() line 230 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >>>> [0]PETSC ERROR: PCSetUp_MG() line 730 in /opt/petsc/petsc-3.4.3/src/ksp/pc/impls/mg/mg.c >>>> [0]PETSC ERROR: PCSetUp() line 890 in /opt/petsc/petsc-3.4.3/src/ksp/pc/interface/precon.c >>>> [0]PETSC ERROR: KSPSetUp() line 278 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >>>> [0]PETSC ERROR: KSPSolve() line 399 in /opt/petsc/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >>>> >>>> >>>> Am I missing anything? >>>> Thank you, >>>> >>>> Michele >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> > From jedbrown at mcs.anl.gov Fri Nov 15 17:49:56 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 15 Nov 2013 16:49:56 -0700 Subject: [petsc-users] KSPSetDM problem In-Reply-To: <5286AF6C.7010703@uci.edu> References: <52857DAB.9090709@uci.edu> <528586F4.5030000@uci.edu> <5286AF6C.7010703@uci.edu> Message-ID: <87li0puswr.fsf@jedbrown.org> Michele Rosso writes: > Barry, > > thank you. > I have another question: how does petsc compute the restriction > (assuming the user does not supply his own restriction)? > Does it use a simple transpose of the interpolation operator? Yes, it uses the same. > Could you point me to the source code where this is done? See the implementation of PCMGGetInterpolation and PCMGGetRestriction. If you only provide one, the other will use it. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From dharmareddy84 at gmail.com Sat Nov 16 16:28:42 2013 From: dharmareddy84 at gmail.com (Dharmendar Reddy) Date: Sat, 16 Nov 2013 16:28:42 -0600 Subject: [petsc-users] DMPlexCreateSubMesh In-Reply-To: References: Message-ID: Alright. I will test it. Thanks Reddy On Fri, Nov 15, 2013 at 9:01 AM, Matthew Knepley wrote: > On Thu, Nov 14, 2013 at 3:16 PM, Dharmendar Reddy > wrote: > >> Thanks a lot. >> > > Okay, I have pushed it and put in a test at > src/dm/impls/plex/examples/tutorials/ex1f90.F > > I needed to update 'bfort' which makes the Fortran bindings. The easiest > thing to do is grab it from > githubg (petsc/pkg-sowing) and install into your PETSC_ARCH. I will ask > Satish to remake the > tarball, so if your install does not work you can just reconfigure after > deleting externalpackage/sowing-* > and PETSC_ARCH/conf/sowing. > > Thanks, > > Matt > > >> Reddy >> >> >> On Thu, Nov 14, 2013 at 3:10 PM, Matthew Knepley wrote: >> >>> On Thu, Nov 14, 2013 at 2:55 PM, Dharmendar Reddy < >>> dharmareddy84 at gmail.com> wrote: >>> >>>> Hello, >>>> Looks like there is a change in the DMPlexCreateSubMesh >>>> interface. >>>> >>>> My code was running fine with petsc 3.4.3 but when i switched petsc to >>>> next. I am getting runtime error due to DMPlexCreateSubMesh.. >>>> >>>> The interface changed from: >>>> ( >>>> http://www.mcs.anl.gov/petsc/petsccurrent/src/dm/impls/plex/plexsubmesh.c.html#DMPlexCreateSubmesh >>>> ) >>>> >>>> 2087: *PetscErrorCode DMPlexCreateSubmesh (DM dm, const char vertexLabel[], PetscInt value, DM *subdm)* >>>> >>>> >>>> to >>>> ( >>>> https://bitbucket.org/petsc/petsc/src/1d94116c89ca7d6e640482ce2684b8b52bac7dff/src/dm/impls/plex/plexsubmesh.c?at=next#cl-2468 >>>> ) >>>> >>>> PetscErrorCode DMPlexCreateSubmesh(DM dm, DMLabel vertexLabel, PetscInt value, DM *subdm) >>>> >>>> >>>> How should i change my code to make use of the new interface ? I have had issue using DMLabel related function from my fortran code. >>>> >>>> >>> Okay, we should just fix using DMLabel in Fortran. I will write the >>> binding for DMPlexGetLabel(), and I will >>> define the necessary DMLabel stuff for Fortran. I will mail when its >>> pushed. >>> >>> Matt >>> >>> >>>> Thanks >>>> Reddy >>>> -- >>>> ----------------------------------------------------- >>>> Dharmendar Reddy Palle >>>> Graduate Student >>>> Microelectronics Research center, >>>> University of Texas at Austin, >>>> 10100 Burnet Road, Bldg. 160 >>>> MER 2.608F, TX 78758-4445 >>>> e-mail: dharmareddy84 at gmail.com >>>> Phone: +1-512-350-9082 >>>> United States of America. >>>> Homepage: https://webspace.utexas.edu/~dpr342 >>>> >>> >>> >>> >>> -- >>> What most experimenters take for granted before they begin their >>> experiments is infinitely more interesting than any results to which their >>> experiments lead. >>> -- Norbert Wiener >>> >> >> >> >> -- >> ----------------------------------------------------- >> Dharmendar Reddy Palle >> Graduate Student >> Microelectronics Research center, >> University of Texas at Austin, >> 10100 Burnet Road, Bldg. 160 >> MER 2.608F, TX 78758-4445 >> e-mail: dharmareddy84 at gmail.com >> Phone: +1-512-350-9082 >> United States of America. >> Homepage: https://webspace.utexas.edu/~dpr342 >> > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -- ----------------------------------------------------- Dharmendar Reddy Palle Graduate Student Microelectronics Research center, University of Texas at Austin, 10100 Burnet Road, Bldg. 160 MER 2.608F, TX 78758-4445 e-mail: dharmareddy84 at gmail.com Phone: +1-512-350-9082 United States of America. Homepage: https://webspace.utexas.edu/~dpr342 -------------- next part -------------- An HTML attachment was scrubbed... URL: From potaman at outlook.com Sun Nov 17 13:42:43 2013 From: potaman at outlook.com (Subramanya Sadasiva) Date: Sun, 17 Nov 2013 14:42:43 -0500 Subject: [petsc-users] Using a previous update to ameliorate jacobian near singularity Message-ID: Hi, I am trying to solve a degenerate cahn hilliard inequality using petsc_DM and virs through libmesh. The problem is that the jacobian for this system has a nearly singular block owing to the \delta t term multiplying the K [ K + M M ] [ \delta phi] = -R_phi [ M K deltat] [\delta mu ] = -R_mu This leads to the unfortunate situation that the time stepping actually fails for very small timesteps. a solution that seems possible is to add a constant mass matrix to the K delta t, making the system [ K + M M ] [ \delta phi] = -R_phi [ M K deltat +M] [\delta mu ] = -R_mu + M \delta mu And using the \delta mu from the previous timestep. I was just wondering if there is any way to get the pre-line search value of the \delta mu? Is it okay to get the \alpha from the linesearch and divide the previous change by the \alpha? I hope I am making sense here. Thanks, Subramanya From knepley at gmail.com Mon Nov 18 08:30:31 2013 From: knepley at gmail.com (Matthew Knepley) Date: Mon, 18 Nov 2013 08:30:31 -0600 Subject: [petsc-users] Using a previous update to ameliorate jacobian near singularity In-Reply-To: References: Message-ID: On Sun, Nov 17, 2013 at 1:42 PM, Subramanya Sadasiva wrote: > Hi, > I am trying to solve a degenerate cahn hilliard inequality using petsc_DM > and virs through libmesh. The problem is that the jacobian for this system > has a nearly singular block owing to the \delta t term multiplying the K > > [ K + M M ] [ \delta phi] = -R_phi > [ M K deltat] [\delta mu ] = -R_mu > > This leads to the unfortunate situation that the time stepping actually > fails for very small timesteps. a solution that seems possible is to add a > constant mass matrix to the K delta t, making the system > > [ K + M M ] [ \delta phi] = -R_phi > [ M K deltat +M] [\delta mu ] = -R_mu + M \delta mu > > And using the \delta mu from the previous timestep. I was just wondering > if there is any way to get the pre-line search value of the \delta mu? Is > it okay to get the \alpha from the linesearch and divide the previous > change by the \alpha? > Why wouldn't you just scale the problem (Jacobi)? Matt > I hope I am making sense here. > Thanks, > Subramanya > > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From asmund.ervik at ntnu.no Mon Nov 18 11:33:29 2013 From: asmund.ervik at ntnu.no (=?ISO-8859-1?Q?=C5smund_Ervik?=) Date: Mon, 18 Nov 2013 18:33:29 +0100 Subject: [petsc-users] Null space for Poisson with zero Neumann BCs Message-ID: <528A4F69.6010004@ntnu.no> Good people of PETSc, I have read in the list archives that setting the null space is the preferred method of ensuring that the pressure Poisson equation is non-singular. Setting the pressure value at one point is not recommended. I have tried this advice using the following code: call MatNullSpaceCreate(PETSC_COMM_WORLD,PETSC_TRUE,0,0,nullspace,ierr) call MatSetNullSpace(A,nullspace,ierr) call MatNullSpaceRemove(nullspace,rhs_vec,PETSC_NULL_OBJECT,ierr) call MatNullSpaceDestroy(nullspace,ierr) after assembling the mat/vec and before KSPSetOperators, KSPSolve etc. With this I get a better looking pressure field, but the runtime is horrible (10x longer) when compared to my old "set the value at some point" method. Also, I see that it takes a long time for some time steps, but then goes fast for others. The Poisson equation should be essentially the same for all these time steps, the velocity field is changing slowly. Below is log_summary for "old method" (first) and "set nullspace" (last). (And yes, I know I'm running an old version of "dev", but I've been too swamped to change to the git setup.) Any suggestions on improving the runtime? Best regards, ?smund ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./run on a arch-linux2-c-opt named ty with 1 processor, by asmunder Mon Nov 18 18:29:26 2013 Using Petsc Development HG revision: 3a41f882cfc717ec37b4c7f6b31f43b10211af66 HG Date: Sun Feb 17 13:07:58 2013 -0600 Max Max/Min Avg Total Time (sec): 3.617e+00 1.00000 3.617e+00 Objects: 4.400e+01 1.00000 4.400e+01 Flops: 2.051e+07 1.00000 2.051e+07 2.051e+07 Flops/sec: 5.670e+06 1.00000 5.670e+06 5.670e+06 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 1.620e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 3.6173e+00 100.0% 2.0508e+07 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 1.610e+02 99.4% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage ThreadCommRunKer 1628 1.0 9.5713e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecDot 234 1.0 5.8794e-04 1.0 1.92e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 9 0 0 0 0 9 0 0 0 3260 VecDotNorm2 117 1.0 5.3787e-04 1.0 1.92e+06 1.0 0.0e+00 0.0e+00 1.2e+02 0 9 0 0 72 0 9 0 0 73 3564 VecNorm 177 1.0 4.3344e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecCopy 60 1.0 1.6022e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 394 1.0 5.6601e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 30 1.0 1.1802e-04 1.0 2.46e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 2082 VecAXPBYCZ 234 1.0 1.3099e-03 1.0 3.83e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 19 0 0 0 0 19 0 0 0 2927 VecWAXPY 234 1.0 1.1146e-03 1.0 1.92e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 9 0 0 0 0 9 0 0 0 1720 VecAssemblyBegin 30 1.0 4.7684e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAssemblyEnd 30 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 264 1.0 5.7216e-03 1.0 1.07e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 52 0 0 0 0 52 0 0 0 1866 MatConvert 30 1.0 1.8499e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 30 1.0 5.0068e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 30 1.0 8.7762e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 30 1.0 2.8610e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSetUp 30 1.0 1.9836e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 4 0 0 0 0 4 0 KSPSolve 30 1.0 2.1702e-01 1.0 2.05e+07 1.0 0.0e+00 0.0e+00 1.6e+02 6100 0 0 97 6100 0 0 98 95 PCSetUp 30 1.0 8.2594e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 2 0 0 0 2 2 0 0 0 2 0 PCApply 294 1.0 1.2338e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 3 0 0 0 0 3 0 0 0 0 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Vector 40 40 1370240 0 Matrix 1 1 313964 0 Krylov Solver 1 1 1288 0 Preconditioner 1 1 1248 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 0 #PETSc Option Table entries: -ksp_type bcgs -log_summary -pc_hypre_type boomeramg -pc_type hypre #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Tue Feb 19 15:18:58 2013 Configure options: --with-threadcomm --with-pthreadclasses --with-openmp --with-cc=icc --with-fc=ifort --with-debugging=0 --with-shared-libraries=1 --download-mpich --download-hypre COPTFLAGS=-O3 FOPTFLAGS=-O3 ----------------------------------------- Libraries compiled on Tue Feb 19 15:18:58 2013 on ty Machine characteristics: Linux-3.7.9-1-ARCH-x86_64-with-glibc2.2.5 Using PETSc directory: /opt/petsc/petsc-dev-ifort Using PETSc arch: arch-linux2-c-opt ----------------------------------------- Using C compiler: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpicc -fPIC -wd1572 -O3 -fopenmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpif90 -fPIC -O3 -fopenmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/include -I/opt/petsc/petsc-dev-ifort/include -I/opt/petsc/petsc-dev-ifort/include -I/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/include ----------------------------------------- Using C linker: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpicc Using Fortran linker: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpif90 Using libraries: -Wl,-rpath,/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -L/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -lpetsc -Wl,-rpath,/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -L/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -lHYPRE -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/compiler/lib/intel64 -L/opt/intel/composer_xe_2013.1.117/compiler/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/ipp/lib/intel64 -L/opt/intel/composer_xe_2013.1.117/ipp/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/mkl/lib/intel64 -L/opt/intel/composer_xe_2013.1.117/mkl/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/tbb/lib/intel64 -L/opt/intel/composer_xe_2013.1.117/tbb/lib/intel64 -Wl,-rpath,/usr/lib/gcc/x86_64-unknown-linux-gnu/4.7.2 -L/usr/lib/gcc/x86_64-unknown-linux-gnu/4.7.2 -lmpichcxx -llapack -lblas -lX11 -lpthread -lmpichf90 -lifport -lifcore -lm -lm -lmpichcxx -ldl -lmpich -lopa -lmpl -lrt -lpthread -limf -lsvml -lirng -lipgo -ldecimal -lcilkrts -lstdc++ -lgcc_s -lirc -lirc_s -ldl ----------------------------------------- ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./run on a arch-linux2-c-opt named ty with 1 processor, by asmunder Mon Nov 18 18:22:38 2013 Using Petsc Development HG revision: 3a41f882cfc717ec37b4c7f6b31f43b10211af66 HG Date: Sun Feb 17 13:07:58 2013 -0600 Max Max/Min Avg Total Time (sec): 9.378e+01 1.00000 9.378e+01 Objects: 7.400e+01 1.00000 7.400e+01 Flops: 1.630e+10 1.00000 1.630e+10 1.630e+10 Flops/sec: 1.738e+08 1.00000 1.738e+08 1.738e+08 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 3.004e+05 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 9.3784e+01 100.0% 1.6300e+10 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 3.004e+05 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage ThreadCommRunKer 1101474 1.0 6.9823e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 7 0 0 0 0 7 0 0 0 0 0 VecDot 200206 1.0 4.7812e-01 1.0 1.64e+09 1.0 0.0e+00 0.0e+00 0.0e+00 1 10 0 0 0 1 10 0 0 0 3430 VecDotNorm2 100103 1.0 4.3580e-01 1.0 1.64e+09 1.0 0.0e+00 0.0e+00 1.0e+05 0 10 0 0 33 0 10 0 0 33 3763 VecNorm 100163 1.0 2.1225e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecCopy 60 1.0 1.6689e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 200366 1.0 2.3949e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 30 1.0 1.1563e-04 1.0 2.46e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2125 VecAXPBYCZ 200206 1.0 1.1021e+00 1.0 3.28e+09 1.0 0.0e+00 0.0e+00 0.0e+00 1 20 0 0 0 1 20 0 0 0 2976 VecWAXPY 200206 1.0 9.3084e-01 1.0 1.64e+09 1.0 0.0e+00 0.0e+00 0.0e+00 1 10 0 0 0 1 10 0 0 0 1762 VecAssemblyBegin 30 1.0 2.8610e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAssemblyEnd 30 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 200236 1.0 4.2707e+00 1.0 8.10e+09 1.0 0.0e+00 0.0e+00 0.0e+00 5 50 0 0 0 5 50 0 0 0 1896 MatConvert 30 1.0 1.9252e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 30 1.0 1.0967e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 30 1.0 8.7786e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 30 1.0 4.0531e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSetUp 30 1.0 1.8668e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 30 1.0 8.9967e+01 1.0 1.63e+10 1.0 0.0e+00 0.0e+00 3.0e+05 96100 0 0100 96100 0 0100 181 PCSetUp 30 1.0 8.3021e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 200266 1.0 8.0928e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 86 0 0 0 0 86 0 0 0 0 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Vector 40 40 1370240 0 Matrix 1 1 313964 0 Matrix Null Space 30 30 18120 0 Krylov Solver 1 1 1288 0 Preconditioner 1 1 1248 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 #PETSc Option Table entries: -ksp_type bcgs -log_summary -pc_hypre_type boomeramg -pc_type hypre #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Tue Feb 19 15:18:58 2013 Configure options: --with-threadcomm --with-pthreadclasses --with-openmp --with-cc=icc --with-fc=ifort --with-debugging=0 --with-shared-libraries=1 --download-mpich --download-hypre COPTFLAGS=-O3 FOPTFLAGS=-O3 ----------------------------------------- Libraries compiled on Tue Feb 19 15:18:58 2013 on ty Machine characteristics: Linux-3.7.9-1-ARCH-x86_64-with-glibc2.2.5 Using PETSc directory: /opt/petsc/petsc-dev-ifort Using PETSc arch: arch-linux2-c-opt ----------------------------------------- Using C compiler: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpicc -fPIC -wd1572 -O3 -fopenmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpif90 -fPIC -O3 -fopenmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/include -I/opt/petsc/petsc-dev-ifort/include -I/opt/petsc/petsc-dev-ifort/include -I/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/include ----------------------------------------- Using C linker: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpicc Using Fortran linker: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpif90 Using libraries: -Wl,-rpath,/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -L/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -lpetsc -Wl,-rpath,/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -L/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -lHYPRE -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/compiler/lib/intel64 -L/opt/intel/composer_xe_2013.1.117/compiler/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/ipp/lib/intel64 -L/opt/intel/composer_xe_2013.1.117/ipp/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/mkl/lib/intel64 -L/opt/intel/composer_xe_2013.1.117/mkl/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/tbb/lib/intel64 -L/opt/intel/composer_xe_2013.1.117/tbb/lib/intel64 -Wl,-rpath,/usr/lib/gcc/x86_64-unknown-linux-gnu/4.7.2 -L/usr/lib/gcc/x86_64-unknown-linux-gnu/4.7.2 -lmpichcxx -llapack -lblas -lX11 -lpthread -lmpichf90 -lifport -lifcore -lm -lm -lmpichcxx -ldl -lmpich -lopa -lmpl -lrt -lpthread -limf -lsvml -lirng -lipgo -ldecimal -lcilkrts -lstdc++ -lgcc_s -lirc -lirc_s -ldl ----------------------------------------- From asmund.ervik at ntnu.no Mon Nov 18 11:56:57 2013 From: asmund.ervik at ntnu.no (=?ISO-8859-1?Q?=C5smund_Ervik?=) Date: Mon, 18 Nov 2013 18:56:57 +0100 Subject: [petsc-users] Null space for Poisson with zero Neumann BCs In-Reply-To: <528A4F69.6010004@ntnu.no> References: <528A4F69.6010004@ntnu.no> Message-ID: <528A54E9.3040105@ntnu.no> Hi again, Never mind this question. It turned out the residual was overly strict for this way of removing the singularity. Taking it back a notch (from 1e-15 to 1e-13) and both methods are equally fast. Dunno why my "dirty" method was faster with a smaller residual. - ?smund On 18. nov. 2013 18:33, ?smund Ervik wrote: > Good people of PETSc, > > I have read in the list archives that setting the null space is the > preferred method of ensuring that the pressure Poisson equation is > non-singular. Setting the pressure value at one point is not recommended. > > I have tried this advice using the following code: > > call MatNullSpaceCreate(PETSC_COMM_WORLD,PETSC_TRUE,0,0,nullspace,ierr) > call MatSetNullSpace(A,nullspace,ierr) > call MatNullSpaceRemove(nullspace,rhs_vec,PETSC_NULL_OBJECT,ierr) > call MatNullSpaceDestroy(nullspace,ierr) > > after assembling the mat/vec and before KSPSetOperators, KSPSolve etc. > > With this I get a better looking pressure field, but the runtime is > horrible (10x longer) when compared to my old "set the value at some > point" method. Also, I see that it takes a long time for some time > steps, but then goes fast for others. The Poisson equation should be > essentially the same for all these time steps, the velocity field is > changing slowly. > > Below is log_summary for "old method" (first) and "set nullspace" > (last). (And yes, I know I'm running an old version of "dev", but I've > been too swamped to change to the git setup.) Any suggestions on > improving the runtime? > > Best regards, > ?smund > > > > ---------------------------------------------- PETSc Performance > Summary: ---------------------------------------------- > > ./run on a arch-linux2-c-opt named ty with 1 processor, by asmunder Mon > Nov 18 18:29:26 2013 > Using Petsc Development HG revision: > 3a41f882cfc717ec37b4c7f6b31f43b10211af66 HG Date: Sun Feb 17 13:07:58 > 2013 -0600 > > Max Max/Min Avg Total > Time (sec): 3.617e+00 1.00000 3.617e+00 > Objects: 4.400e+01 1.00000 4.400e+01 > Flops: 2.051e+07 1.00000 2.051e+07 2.051e+07 > Flops/sec: 5.670e+06 1.00000 5.670e+06 5.670e+06 > MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Reductions: 1.620e+02 1.00000 > > Flop counting convention: 1 flop = 1 real number operation of type > (multiply/divide/add/subtract) > e.g., VecAXPY() for real vectors of length N > --> 2N flops > and VecAXPY() for complex vectors of length > N --> 8N flops > > Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages > --- -- Message Lengths -- -- Reductions -- > Avg %Total Avg %Total counts > %Total Avg %Total counts %Total > 0: Main Stage: 3.6173e+00 100.0% 2.0508e+07 100.0% 0.000e+00 > 0.0% 0.000e+00 0.0% 1.610e+02 99.4% > > ------------------------------------------------------------------------------------------------------------------------ > See the 'Profiling' chapter of the users' manual for details on > interpreting output. > Phase summary info: > Count: number of times phase was executed > Time and Flops: Max - maximum over all processors > Ratio - ratio of maximum to minimum over all processors > Mess: number of messages sent > Avg. len: average message length (bytes) > Reduct: number of global reductions > Global: entire computation > Stage: stages of a computation. Set stages with PetscLogStagePush() > and PetscLogStagePop(). > %T - percent time in this phase %f - percent flops in this > phase > %M - percent messages in this phase %L - percent message > lengths in this phase > %R - percent reductions in this phase > Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time > over all processors) > ------------------------------------------------------------------------------------------------------------------------ > Event Count Time (sec) Flops > --- Global --- --- Stage --- Total > Max Ratio Max Ratio Max Ratio Mess Avg len > Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s > ------------------------------------------------------------------------------------------------------------------------ > > --- Event Stage 0: Main Stage > > ThreadCommRunKer 1628 1.0 9.5713e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecDot 234 1.0 5.8794e-04 1.0 1.92e+06 1.0 0.0e+00 0.0e+00 > 0.0e+00 0 9 0 0 0 0 9 0 0 0 3260 > VecDotNorm2 117 1.0 5.3787e-04 1.0 1.92e+06 1.0 0.0e+00 0.0e+00 > 1.2e+02 0 9 0 0 72 0 9 0 0 73 3564 > VecNorm 177 1.0 4.3344e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecCopy 60 1.0 1.6022e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecSet 394 1.0 5.6601e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecAXPY 30 1.0 1.1802e-04 1.0 2.46e+05 1.0 0.0e+00 0.0e+00 > 0.0e+00 0 1 0 0 0 0 1 0 0 0 2082 > VecAXPBYCZ 234 1.0 1.3099e-03 1.0 3.83e+06 1.0 0.0e+00 0.0e+00 > 0.0e+00 0 19 0 0 0 0 19 0 0 0 2927 > VecWAXPY 234 1.0 1.1146e-03 1.0 1.92e+06 1.0 0.0e+00 0.0e+00 > 0.0e+00 0 9 0 0 0 0 9 0 0 0 1720 > VecAssemblyBegin 30 1.0 4.7684e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecAssemblyEnd 30 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatMult 264 1.0 5.7216e-03 1.0 1.07e+07 1.0 0.0e+00 0.0e+00 > 0.0e+00 0 52 0 0 0 0 52 0 0 0 1866 > MatConvert 30 1.0 1.8499e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyBegin 30 1.0 5.0068e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyEnd 30 1.0 8.7762e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatGetRowIJ 30 1.0 2.8610e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > KSPSetUp 30 1.0 1.9836e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 6.0e+00 0 0 0 0 4 0 0 0 0 4 0 > KSPSolve 30 1.0 2.1702e-01 1.0 2.05e+07 1.0 0.0e+00 0.0e+00 > 1.6e+02 6100 0 0 97 6100 0 0 98 95 > PCSetUp 30 1.0 8.2594e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 4.0e+00 2 0 0 0 2 2 0 0 0 2 0 > PCApply 294 1.0 1.2338e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 3 0 0 0 0 3 0 0 0 0 0 > ------------------------------------------------------------------------------------------------------------------------ > > Memory usage is given in bytes: > > Object Type Creations Destructions Memory Descendants' Mem. > Reports information only for process 0. > > --- Event Stage 0: Main Stage > > Vector 40 40 1370240 0 > Matrix 1 1 313964 0 > Krylov Solver 1 1 1288 0 > Preconditioner 1 1 1248 0 > Viewer 1 0 0 0 > ======================================================================================================================== > Average time to get PetscTime(): 0 > #PETSc Option Table entries: > -ksp_type bcgs > -log_summary > -pc_hypre_type boomeramg > -pc_type hypre > #End of PETSc Option Table entries > Compiled without FORTRAN kernels > Compiled with full precision matrices (default) > sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 > sizeof(PetscScalar) 8 sizeof(PetscInt) 4 > Configure run at: Tue Feb 19 15:18:58 2013 > Configure options: --with-threadcomm --with-pthreadclasses --with-openmp > --with-cc=icc --with-fc=ifort --with-debugging=0 > --with-shared-libraries=1 --download-mpich --download-hypre > COPTFLAGS=-O3 FOPTFLAGS=-O3 > ----------------------------------------- > Libraries compiled on Tue Feb 19 15:18:58 2013 on ty > Machine characteristics: Linux-3.7.9-1-ARCH-x86_64-with-glibc2.2.5 > Using PETSc directory: /opt/petsc/petsc-dev-ifort > Using PETSc arch: arch-linux2-c-opt > ----------------------------------------- > > Using C compiler: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpicc > -fPIC -wd1572 -O3 -fopenmp ${COPTFLAGS} ${CFLAGS} > Using Fortran compiler: > /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpif90 -fPIC -O3 > -fopenmp ${FOPTFLAGS} ${FFLAGS} > ----------------------------------------- > > Using include paths: > -I/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/include > -I/opt/petsc/petsc-dev-ifort/include > -I/opt/petsc/petsc-dev-ifort/include > -I/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/include > ----------------------------------------- > > Using C linker: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpicc > Using Fortran linker: > /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpif90 > Using libraries: > -Wl,-rpath,/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib > -L/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -lpetsc > -Wl,-rpath,/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib > -L/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -lHYPRE > -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/compiler/lib/intel64 > -L/opt/intel/composer_xe_2013.1.117/compiler/lib/intel64 > -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/ipp/lib/intel64 > -L/opt/intel/composer_xe_2013.1.117/ipp/lib/intel64 > -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/mkl/lib/intel64 > -L/opt/intel/composer_xe_2013.1.117/mkl/lib/intel64 > -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/tbb/lib/intel64 > -L/opt/intel/composer_xe_2013.1.117/tbb/lib/intel64 > -Wl,-rpath,/usr/lib/gcc/x86_64-unknown-linux-gnu/4.7.2 > -L/usr/lib/gcc/x86_64-unknown-linux-gnu/4.7.2 -lmpichcxx -llapack -lblas > -lX11 -lpthread -lmpichf90 -lifport -lifcore -lm -lm -lmpichcxx -ldl > -lmpich -lopa -lmpl -lrt -lpthread -limf -lsvml -lirng -lipgo -ldecimal > -lcilkrts -lstdc++ -lgcc_s -lirc -lirc_s -ldl > ----------------------------------------- > > > > > > > > > ---------------------------------------------- PETSc Performance > Summary: ---------------------------------------------- > > ./run on a arch-linux2-c-opt named ty with 1 processor, by asmunder Mon > Nov 18 18:22:38 2013 > Using Petsc Development HG revision: > 3a41f882cfc717ec37b4c7f6b31f43b10211af66 HG Date: Sun Feb 17 13:07:58 > 2013 -0600 > > Max Max/Min Avg Total > Time (sec): 9.378e+01 1.00000 9.378e+01 > Objects: 7.400e+01 1.00000 7.400e+01 > Flops: 1.630e+10 1.00000 1.630e+10 1.630e+10 > Flops/sec: 1.738e+08 1.00000 1.738e+08 1.738e+08 > MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 > MPI Reductions: 3.004e+05 1.00000 > > Flop counting convention: 1 flop = 1 real number operation of type > (multiply/divide/add/subtract) > e.g., VecAXPY() for real vectors of length N > --> 2N flops > and VecAXPY() for complex vectors of length > N --> 8N flops > > Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages > --- -- Message Lengths -- -- Reductions -- > Avg %Total Avg %Total counts > %Total Avg %Total counts %Total > 0: Main Stage: 9.3784e+01 100.0% 1.6300e+10 100.0% 0.000e+00 > 0.0% 0.000e+00 0.0% 3.004e+05 100.0% > > ------------------------------------------------------------------------------------------------------------------------ > See the 'Profiling' chapter of the users' manual for details on > interpreting output. > Phase summary info: > Count: number of times phase was executed > Time and Flops: Max - maximum over all processors > Ratio - ratio of maximum to minimum over all processors > Mess: number of messages sent > Avg. len: average message length (bytes) > Reduct: number of global reductions > Global: entire computation > Stage: stages of a computation. Set stages with PetscLogStagePush() > and PetscLogStagePop(). > %T - percent time in this phase %f - percent flops in this > phase > %M - percent messages in this phase %L - percent message > lengths in this phase > %R - percent reductions in this phase > Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time > over all processors) > ------------------------------------------------------------------------------------------------------------------------ > Event Count Time (sec) Flops > --- Global --- --- Stage --- Total > Max Ratio Max Ratio Max Ratio Mess Avg len > Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s > ------------------------------------------------------------------------------------------------------------------------ > > --- Event Stage 0: Main Stage > > ThreadCommRunKer 1101474 1.0 6.9823e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 7 0 0 0 0 7 0 0 0 0 0 > VecDot 200206 1.0 4.7812e-01 1.0 1.64e+09 1.0 0.0e+00 0.0e+00 > 0.0e+00 1 10 0 0 0 1 10 0 0 0 3430 > VecDotNorm2 100103 1.0 4.3580e-01 1.0 1.64e+09 1.0 0.0e+00 0.0e+00 > 1.0e+05 0 10 0 0 33 0 10 0 0 33 3763 > VecNorm 100163 1.0 2.1225e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecCopy 60 1.0 1.6689e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecSet 200366 1.0 2.3949e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecAXPY 30 1.0 1.1563e-04 1.0 2.46e+05 1.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 2125 > VecAXPBYCZ 200206 1.0 1.1021e+00 1.0 3.28e+09 1.0 0.0e+00 0.0e+00 > 0.0e+00 1 20 0 0 0 1 20 0 0 0 2976 > VecWAXPY 200206 1.0 9.3084e-01 1.0 1.64e+09 1.0 0.0e+00 0.0e+00 > 0.0e+00 1 10 0 0 0 1 10 0 0 0 1762 > VecAssemblyBegin 30 1.0 2.8610e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > VecAssemblyEnd 30 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatMult 200236 1.0 4.2707e+00 1.0 8.10e+09 1.0 0.0e+00 0.0e+00 > 0.0e+00 5 50 0 0 0 5 50 0 0 0 1896 > MatConvert 30 1.0 1.9252e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyBegin 30 1.0 1.0967e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatAssemblyEnd 30 1.0 8.7786e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > MatGetRowIJ 30 1.0 4.0531e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 > KSPSetUp 30 1.0 1.8668e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 6.0e+00 0 0 0 0 0 0 0 0 0 0 0 > KSPSolve 30 1.0 8.9967e+01 1.0 1.63e+10 1.0 0.0e+00 0.0e+00 > 3.0e+05 96100 0 0100 96100 0 0100 181 > PCSetUp 30 1.0 8.3021e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 4.0e+00 0 0 0 0 0 0 0 0 0 0 0 > PCApply 200266 1.0 8.0928e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 > 0.0e+00 86 0 0 0 0 86 0 0 0 0 0 > ------------------------------------------------------------------------------------------------------------------------ > > Memory usage is given in bytes: > > Object Type Creations Destructions Memory Descendants' Mem. > Reports information only for process 0. > > --- Event Stage 0: Main Stage > > Vector 40 40 1370240 0 > Matrix 1 1 313964 0 > Matrix Null Space 30 30 18120 0 > Krylov Solver 1 1 1288 0 > Preconditioner 1 1 1248 0 > Viewer 1 0 0 0 > ======================================================================================================================== > Average time to get PetscTime(): 9.53674e-08 > #PETSc Option Table entries: > -ksp_type bcgs > -log_summary > -pc_hypre_type boomeramg > -pc_type hypre > #End of PETSc Option Table entries > Compiled without FORTRAN kernels > Compiled with full precision matrices (default) > sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 > sizeof(PetscScalar) 8 sizeof(PetscInt) 4 > Configure run at: Tue Feb 19 15:18:58 2013 > Configure options: --with-threadcomm --with-pthreadclasses --with-openmp > --with-cc=icc --with-fc=ifort --with-debugging=0 > --with-shared-libraries=1 --download-mpich --download-hypre > COPTFLAGS=-O3 FOPTFLAGS=-O3 > ----------------------------------------- > Libraries compiled on Tue Feb 19 15:18:58 2013 on ty > Machine characteristics: Linux-3.7.9-1-ARCH-x86_64-with-glibc2.2.5 > Using PETSc directory: /opt/petsc/petsc-dev-ifort > Using PETSc arch: arch-linux2-c-opt > ----------------------------------------- > > Using C compiler: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpicc > -fPIC -wd1572 -O3 -fopenmp ${COPTFLAGS} ${CFLAGS} > Using Fortran compiler: > /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpif90 -fPIC -O3 > -fopenmp ${FOPTFLAGS} ${FFLAGS} > ----------------------------------------- > > Using include paths: > -I/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/include > -I/opt/petsc/petsc-dev-ifort/include > -I/opt/petsc/petsc-dev-ifort/include > -I/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/include > ----------------------------------------- > > Using C linker: /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpicc > Using Fortran linker: > /opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/bin/mpif90 > Using libraries: > -Wl,-rpath,/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib > -L/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -lpetsc > -Wl,-rpath,/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib > -L/opt/petsc/petsc-dev-ifort/arch-linux2-c-opt/lib -lHYPRE > -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/compiler/lib/intel64 > -L/opt/intel/composer_xe_2013.1.117/compiler/lib/intel64 > -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/ipp/lib/intel64 > -L/opt/intel/composer_xe_2013.1.117/ipp/lib/intel64 > -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/mkl/lib/intel64 > -L/opt/intel/composer_xe_2013.1.117/mkl/lib/intel64 > -Wl,-rpath,/opt/intel/composer_xe_2013.1.117/tbb/lib/intel64 > -L/opt/intel/composer_xe_2013.1.117/tbb/lib/intel64 > -Wl,-rpath,/usr/lib/gcc/x86_64-unknown-linux-gnu/4.7.2 > -L/usr/lib/gcc/x86_64-unknown-linux-gnu/4.7.2 -lmpichcxx -llapack -lblas > -lX11 -lpthread -lmpichf90 -lifport -lifcore -lm -lm -lmpichcxx -ldl > -lmpich -lopa -lmpl -lrt -lpthread -limf -lsvml -lirng -lipgo -ldecimal > -lcilkrts -lstdc++ -lgcc_s -lirc -lirc_s -ldl > ----------------------------------------- > > > From irving at naml.us Mon Nov 18 17:55:28 2013 From: irving at naml.us (Geoffrey Irving) Date: Mon, 18 Nov 2013 15:55:28 -0800 Subject: [petsc-users] Understanding parallelism in DMPlexCreateBoxMesh Message-ID: DMPlexCreateSquareBoundary naively seems to 1. Create topology only on rank 0. 2. Set positions on all ranks. In particular, if seems to allocate space proportional to the entire number of vertices on every rank: ierr = PetscSectionGetStorageSize(coordSection, &coordSize);CHKERRQ(ierr); ierr = VecCreate(PetscObjectComm((PetscObject)dm), &coordinates);CHKERRQ(ierr); ierr = VecSetSizes(coordinates, coordSize, PETSC_DETERMINE);CHKERRQ(ierr); How does this code work? It's clearly intended to be callable in a parallel context given that it has a !rank conditional, and works as part of various examples, so I'm clearly missing something. Geoffrey From jedbrown at mcs.anl.gov Mon Nov 18 18:18:04 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Mon, 18 Nov 2013 17:18:04 -0700 Subject: [petsc-users] Understanding parallelism in DMPlexCreateBoxMesh In-Reply-To: References: Message-ID: <87wqk5tfb7.fsf@jedbrown.org> Geoffrey Irving writes: > DMPlexCreateSquareBoundary naively seems to > > 1. Create topology only on rank 0. > 2. Set positions on all ranks. > > In particular, if seems to allocate space proportional to the entire > number of vertices on every rank: > > ierr = PetscSectionGetStorageSize(coordSection, &coordSize);CHKERRQ(ierr); > ierr = VecCreate(PetscObjectComm((PetscObject)dm), > &coordinates);CHKERRQ(ierr); > ierr = VecSetSizes(coordinates, coordSize, PETSC_DETERMINE);CHKERRQ(ierr); > > How does this code work? It's clearly intended to be callable in a > parallel context given that it has a !rank conditional, and works as > part of various examples, so I'm clearly missing something. This just creates the mesh on rank 0. It is usually partitioned and distributed subsequently (see src/snes/examples/tutorials/ex12.c). It's fine for testing, but obviously non-scalable. Now that Matt and I are running some larger tests with DMPlex, this is one of the scalability issues that will be fixed. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From bhatiamanav at gmail.com Tue Nov 19 11:11:31 2013 From: bhatiamanav at gmail.com (Manav Bhatia) Date: Tue, 19 Nov 2013 12:11:31 -0500 Subject: [petsc-users] Euclid/Hypre running slow Message-ID: Hi, I am using ilu(k) preconditioner from Euclid/Hypre on my Mac OS 10.9 box with 12 processors and openmpi v1.7.3 from macports. I find that for 1 CPU (-np 1) the code runs reasonably fast, but when multiple processors are included the solver runs significantly slow (by orders of magnitude). I ran the time profiler on one of the processes spawned by openmpirun, and it seems to be taking a lot of time in MPI_Waitall called from Euclid_dhSetup. This could very well be a Hypre issue, but I wanted to check to see if anynoe on the Petsc list has experience a similar problem. Any suggestions/inputs are welcome. Thanks, Manav -------------- next part -------------- An HTML attachment was scrubbed... URL: From karpeev at mcs.anl.gov Tue Nov 19 11:32:22 2013 From: karpeev at mcs.anl.gov (Dmitry Karpeyev) Date: Tue, 19 Nov 2013 10:32:22 -0700 Subject: [petsc-users] HIRING: PETSc/libMesh/MOOSE developer Message-ID: Dear All, There is a job opportunity at Argonne National Laboratory/University of Chicago. The job is to develop numerical algorithms and supporting code on the interface between PETSc, libMesh and MOOSE. Therefore, familiarity with these libraries is strongly desired. This is not a postdoc position, but is ideal for someone with a Bachelor's or a Master's degree, well versed in numerical programming and looking to do some real work. A good knowledge of C/C++ is required. A background in numerical linear algebra, finite-element methods and numerical algorithms in general is preferred. If you know your way around multigrid, that's even better. The developer will work closely with the PETSc developers at Argonne National Laboratory and the University of Chicago. This position has not been posted yet, but is expected to be up shortly. If you have questions, free to email me directly. My apologies, if you are get this email multiple times. Best regards, Dmitry. -- Dmitry Karpeyev Mathematics and Computer Science Argonne National Laboratory Argonne, Illinois, USA and Computation Institute University of Chicago 5735 S. Ellis Avenue Chicago, IL 60637 ----------------------- Phone: 630-252-1229 Fax: 630-252-5986 -------------- next part -------------- An HTML attachment was scrubbed... URL: From iwaddington at gmail.com Tue Nov 19 14:52:03 2013 From: iwaddington at gmail.com (iwaddington .) Date: Tue, 19 Nov 2013 18:52:03 -0200 Subject: [petsc-users] Removing PETSC_ARCH Message-ID: Hi everybody, do you know the correct procedure to delete a PETSC_ARCH installed in a folder of its own ? Thanks, Italo. -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Tue Nov 19 15:10:20 2013 From: knepley at gmail.com (Matthew Knepley) Date: Tue, 19 Nov 2013 15:10:20 -0600 Subject: [petsc-users] Removing PETSC_ARCH In-Reply-To: References: Message-ID: On Tue, Nov 19, 2013 at 2:52 PM, iwaddington . wrote: > Hi everybody, do you know the correct procedure to delete a PETSC_ARCH > installed in a folder of its own ? Thanks, Italo. > cd $PETSC_DIR rm -rf $PETSC_ARCH Matt -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From mfadams at lbl.gov Tue Nov 19 18:42:48 2013 From: mfadams at lbl.gov (Mark Adams) Date: Tue, 19 Nov 2013 19:42:48 -0500 Subject: [petsc-users] Euclid/Hypre running slow In-Reply-To: References: Message-ID: This sounds like a hypre issues. Euclid/Hypre is not use much AFAIK so it might have some problems that have not been addressed. On Tue, Nov 19, 2013 at 12:11 PM, Manav Bhatia wrote: > Hi, > > I am using ilu(k) preconditioner from Euclid/Hypre on my Mac OS 10.9 > box with 12 processors and openmpi v1.7.3 from macports. > > I find that for 1 CPU (-np 1) the code runs reasonably fast, but when > multiple processors are included the solver runs significantly slow (by > orders of magnitude). I ran the time profiler on one of the processes > spawned by openmpirun, and it seems to be taking a lot of time in > MPI_Waitall called from Euclid_dhSetup. > > This could very well be a Hypre issue, but I wanted to check to see if > anynoe on the Petsc list has experience a similar problem. > > Any suggestions/inputs are welcome. > > Thanks, > Manav > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From mailinglistxf at yahoo.fr Wed Nov 20 05:36:17 2013 From: mailinglistxf at yahoo.fr (mailinglist) Date: Wed, 20 Nov 2013 12:36:17 +0100 Subject: [petsc-users] Same random number Message-ID: <528C9EB1.5010906@yahoo.fr> Hello, I use Petsc random number generator. Every time I launch the code I get same random number, the code is like this: PetscRandom rnd; PetscRandomCreate(PETSC_COMM_WORLD,&rnd); PetscRandomSetType(rnd,PETSCRAND); PetscScalar r; PetscRandomGetValue(rnd,&r); I would like to ask two questions. How could I get different random number? And what's the difference bewteen PETSCRAND and PETSCRAND48? Thank you very much! Feng XING From hzhang at mcs.anl.gov Wed Nov 20 09:52:55 2013 From: hzhang at mcs.anl.gov (Hong Zhang) Date: Wed, 20 Nov 2013 09:52:55 -0600 Subject: [petsc-users] Same random number In-Reply-To: <528C9EB1.5010906@yahoo.fr> References: <528C9EB1.5010906@yahoo.fr> Message-ID: Feng XING, > > I use Petsc random number generator. Every time I launch the code I get > same random number, the code is like this: > > PetscRandom rnd; > PetscRandomCreate(PETSC_COMM_WORLD,&rnd); > PetscRandomSetType(rnd,PETSCRAND); > > PetscScalar r; > PetscRandomGetValue(rnd,&r); > > I would like to ask two questions. How could I get different random number? You can use PetscRandomSetSeed() and PetscRandomSeed(). See an example: petsc/src/dm/impls/plex/examples/tests/ex6.c > And what's the difference bewteen PETSCRAND and PETSCRAND48? > Two different random number generators, see http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/Sys/PetscRandomType.html Hong -------------- next part -------------- An HTML attachment was scrubbed... URL: From sickofcowboys at hotmail.com Wed Nov 20 17:26:50 2013 From: sickofcowboys at hotmail.com (aymeric aymeric) Date: Thu, 21 Nov 2013 00:26:50 +0100 Subject: [petsc-users] how to partition a matrix? Message-ID: Sorry to ask something tat sounds so trivial, but I can't find help in the manual or the examples.In the context of FE (unstructured grid), I am trying to partition a matrix; I am able to follow the example given in section 2.5 of the manual. But what I get is an IS representing the processor for each element... and what am I supposed to do with that? Assuming that I could distribute the mesh, then how am I supposed to create the matrix associated with the partition? I feel like I am missing something obvious, cause it can't be complicated...Thank you for your help -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Wed Nov 20 17:59:42 2013 From: knepley at gmail.com (Matthew Knepley) Date: Wed, 20 Nov 2013 17:59:42 -0600 Subject: [petsc-users] how to partition a matrix? In-Reply-To: References: Message-ID: On Wed, Nov 20, 2013 at 5:26 PM, aymeric aymeric wrote: > Sorry to ask something tat sounds so trivial, but I can't find help in the > manual or the examples. > In the context of FE (unstructured grid), I am trying to partition a > matrix; I am able to follow the example given in section 2.5 of the manual. > But what I get is an IS representing the processor for each element... and > what am I supposed to do with that? Assuming that I could distribute the > mesh, then how am I supposed to create the matrix associated with the > partition? I feel like I am missing something obvious, cause it can't be > complicated... > All standard mesh partitioners give the same output, partition number of each row (vertex). You can use this in MatPermute() for instance, but that is generally expensive. It might help to explain what you want to do. Matt > Thank you for your help > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From jsd1 at rice.edu Wed Nov 20 18:48:06 2013 From: jsd1 at rice.edu (Justin Dong (Me)) Date: Wed, 20 Nov 2013 18:48:06 -0600 Subject: [petsc-users] Troubles installing PETSc Message-ID: <7F374499-AB94-4B8B-BB79-8B143FF1A852@rice.edu> Hi all, I?ve just installed PETSc and it seems that I don?t get any errors. I type ?make? from the command line and this is the end bit of the result: > ========================================== > Building PETSc using CMake with 5 build threads > ========================================== > Re-run cmake file: Makefile older than: ../CMakeLists.txt > -- Configuring done > -- Generating done > -- Build files have been written to: /Users/[?]/Classes/CAAMResearch/petsc-3.4.3/arch-darwin-c-debug > [100%] Built target petsc > Running /usr/bin/dsymutil on libpetsc > ========================================= > Now to check if the libraries are working do: > make PETSC_DIR=/Users/[?]/Classes/CAAMResearch/petsc-3.4.3 PETSC_ARCH=arch-darwin-c-debug test and if I run the command it tells me to after that, I get: > C/C++ example src/snes/examples/tutorials/ex19 run successfully with 1 MPI process > C/C++ example src/snes/examples/tutorials/ex19 run successfully with 2 MPI processes > Fortran example src/snes/examples/tutorials/ex5f run successfully with 1 MPI process > Completed test examples I try this example (http://www.mcs.anl.gov/petsc/documentation/exercises/compiling/index.html) and get this error: > cc ex2.c -o ex2 > ex2.c:13:10: fatal error: 'petscsys.h' file not found > #include > ^ > 1 error generated. > make: *** [ex2] Error 1 I?m working in the PETSc directory so I don?t think it?s an issue with setting the path, but I?m not that great at all of the architecture stuff. If anyone can assist, I?d be grateful. Thanks, Justin -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Wed Nov 20 18:51:13 2013 From: knepley at gmail.com (Matthew Knepley) Date: Wed, 20 Nov 2013 18:51:13 -0600 Subject: [petsc-users] Troubles installing PETSc In-Reply-To: <7F374499-AB94-4B8B-BB79-8B143FF1A852@rice.edu> References: <7F374499-AB94-4B8B-BB79-8B143FF1A852@rice.edu> Message-ID: On Wed, Nov 20, 2013 at 6:48 PM, Justin Dong (Me) wrote: > Hi all, > > I?ve just installed PETSc and it seems that I don?t get any errors. I type > ?make? from the command line and this is the end bit of the result: > > ========================================== > Building PETSc using CMake with 5 build threads > ========================================== > Re-run cmake file: Makefile older than: ../CMakeLists.txt > -- Configuring done > -- Generating done > -- Build files have been written to: > /Users/[?]/Classes/CAAMResearch/petsc-3.4.3/arch-darwin-c-debug > [100%] Built target petsc > Running /usr/bin/dsymutil on libpetsc > ========================================= > Now to check if the libraries are working do: > make PETSC_DIR=/Users/[?]/Classes/CAAMResearch/petsc-3.4.3 > PETSC_ARCH=arch-darwin-c-debug test > > > and if I run the command it tells me to after that, I get: > > C/C++ example src/snes/examples/tutorials/ex19 run successfully with 1 MPI > process > C/C++ example src/snes/examples/tutorials/ex19 run successfully with 2 MPI > processes > Fortran example src/snes/examples/tutorials/ex5f run successfully with 1 > MPI process > Completed test examples > > > I try this example ( > http://www.mcs.anl.gov/petsc/documentation/exercises/compiling/index.html) > and get this error: > > cc ex2.c -o ex2 > *ex2.c:13:10: **fatal error: **'petscsys.h' file not found* > #include > * ^* > 1 error generated. > make: *** [ex2] Error > > There is a chapter in the manual on using PETSc makefiles Thanks Matt > I?m working in the PETSc directory so I don?t think it?s an issue with > setting the path, but I?m not that great at all of the architecture stuff. > If anyone can assist, I?d be grateful. > > Thanks, > Justin > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From jsd1 at rice.edu Wed Nov 20 20:01:17 2013 From: jsd1 at rice.edu (Justin Dong (Me)) Date: Wed, 20 Nov 2013 20:01:17 -0600 Subject: [petsc-users] Troubles installing PETSc In-Reply-To: References: <7F374499-AB94-4B8B-BB79-8B143FF1A852@rice.edu> Message-ID: Hi Matt, I?ve checked out the section on makefiles and am not really sure what I could be doing incorrectly. I?m positive $PETSC_DIR is set correctly. I?m on MAC OS X 10.9 right now and have $PETSC_ARCH set as linux-gnu. According to the example I linked to below, I should just type ?make ex2?, which gives me the error I listed. In the user manual, I try the suggestion ?make PETSC_ARCH=arch ex2? with arch replaced by linux-gnu and I still get the same error. I haven?t messed with the makefile at all, and if I?ve set $PETSC_DIR and $PETSC_ARCH correctly, it should be fine, right? Any other tips? Thanks, Justin On Nov 20, 2013, at 6:51 PM, Matthew Knepley wrote: > On Wed, Nov 20, 2013 at 6:48 PM, Justin Dong (Me) wrote: > Hi all, > > I?ve just installed PETSc and it seems that I don?t get any errors. I type ?make? from the command line and this is the end bit of the result: > >> ========================================== >> Building PETSc using CMake with 5 build threads >> ========================================== >> Re-run cmake file: Makefile older than: ../CMakeLists.txt >> -- Configuring done >> -- Generating done >> -- Build files have been written to: /Users/[?]/Classes/CAAMResearch/petsc-3.4.3/arch-darwin-c-debug >> [100%] Built target petsc >> Running /usr/bin/dsymutil on libpetsc >> ========================================= >> Now to check if the libraries are working do: >> make PETSC_DIR=/Users/[?]/Classes/CAAMResearch/petsc-3.4.3 PETSC_ARCH=arch-darwin-c-debug test > > and if I run the command it tells me to after that, I get: > >> C/C++ example src/snes/examples/tutorials/ex19 run successfully with 1 MPI process >> C/C++ example src/snes/examples/tutorials/ex19 run successfully with 2 MPI processes >> Fortran example src/snes/examples/tutorials/ex5f run successfully with 1 MPI process >> Completed test examples > > I try this example (http://www.mcs.anl.gov/petsc/documentation/exercises/compiling/index.html) and get this error: > >> cc ex2.c -o ex2 >> ex2.c:13:10: fatal error: 'petscsys.h' file not found >> #include >> ^ >> 1 error generated. >> make: *** [ex2] Error > > > There is a chapter in the manual on using PETSc makefiles > > Thanks > > Matt > > I?m working in the PETSc directory so I don?t think it?s an issue with setting the path, but I?m not that great at all of the architecture stuff. If anyone can assist, I?d be grateful. > > Thanks, > Justin > > > > -- > What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. > -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Wed Nov 20 20:26:08 2013 From: knepley at gmail.com (Matthew Knepley) Date: Wed, 20 Nov 2013 20:26:08 -0600 Subject: [petsc-users] Troubles installing PETSc In-Reply-To: References: <7F374499-AB94-4B8B-BB79-8B143FF1A852@rice.edu> Message-ID: On Wed, Nov 20, 2013 at 8:01 PM, Justin Dong (Me) wrote: > Hi Matt, > > I?ve checked out the section on makefiles and am not really sure what I > could be doing incorrectly. I?m positive $PETSC_DIR is set correctly. I?m > on MAC OS X 10.9 right now and have $PETSC_ARCH set as linux-gnu. > > According to the example I linked to below, I should just type ?make ex2?, > which gives me the error I listed. In the user manual, I try the suggestion > ?make PETSC_ARCH=arch ex2? with arch replaced by linux-gnu and I still get > the same error. I haven?t messed with the makefile at all, and if I?ve set > $PETSC_DIR and $PETSC_ARCH correctly, it should be fine, right? > Either it does not see your makefile in that directory or something is wrong with the makefile. The simplest makefile is ex2: ex2.o ${CLINKER} -o ex2 ex2.o ${PETSC_LIB} include ${PETSC_DIR}/${PETSC_ARCH}/conf/variables include ${PETSC_DIR}/${PETSC_ARCH}/conf/rules Matt > Any other tips? > > Thanks, > Justin > > On Nov 20, 2013, at 6:51 PM, Matthew Knepley wrote: > > On Wed, Nov 20, 2013 at 6:48 PM, Justin Dong (Me) wrote: > >> Hi all, >> >> I?ve just installed PETSc and it seems that I don?t get any errors. I >> type ?make? from the command line and this is the end bit of the result: >> >> ========================================== >> Building PETSc using CMake with 5 build threads >> ========================================== >> Re-run cmake file: Makefile older than: ../CMakeLists.txt >> -- Configuring done >> -- Generating done >> -- Build files have been written to: >> /Users/[?]/Classes/CAAMResearch/petsc-3.4.3/arch-darwin-c-debug >> [100%] Built target petsc >> Running /usr/bin/dsymutil on libpetsc >> ========================================= >> Now to check if the libraries are working do: >> make PETSC_DIR=/Users/[?]/Classes/CAAMResearch/petsc-3.4.3 >> PETSC_ARCH=arch-darwin-c-debug test >> >> >> and if I run the command it tells me to after that, I get: >> >> C/C++ example src/snes/examples/tutorials/ex19 run successfully with 1 >> MPI process >> C/C++ example src/snes/examples/tutorials/ex19 run successfully with 2 >> MPI processes >> Fortran example src/snes/examples/tutorials/ex5f run successfully with 1 >> MPI process >> Completed test examples >> >> >> I try this example ( >> http://www.mcs.anl.gov/petsc/documentation/exercises/compiling/index.html) >> and get this error: >> >> cc ex2.c -o ex2 >> *ex2.c:13:10: **fatal error: **'petscsys.h' file not found* >> #include >> * ^* >> 1 error generated. >> make: *** [ex2] Error >> >> > There is a chapter in the manual on using PETSc makefiles > > Thanks > > Matt > > >> I?m working in the PETSc directory so I don?t think it?s an issue with >> setting the path, but I?m not that great at all of the architecture stuff. >> If anyone can assist, I?d be grateful. >> >> Thanks, >> Justin >> > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Wed Nov 20 20:05:03 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 20 Nov 2013 19:05:03 -0700 Subject: [petsc-users] Troubles installing PETSc In-Reply-To: References: <7F374499-AB94-4B8B-BB79-8B143FF1A852@rice.edu> Message-ID: <87r4aay0fk.fsf@jedbrown.org> "Justin Dong (Me)" writes: > Hi Matt, > > I?ve checked out the section on makefiles and am not really sure what > I could be doing incorrectly. I?m positive $PETSC_DIR is set > correctly. I?m on MAC OS X 10.9 right now and have $PETSC_ARCH set as > linux-gnu. Paste your makefile and all output when you try to run make. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From zonexo at gmail.com Thu Nov 21 00:04:24 2013 From: zonexo at gmail.com (TAY wee-beng) Date: Thu, 21 Nov 2013 14:04:24 +0800 Subject: [petsc-users] Modify RHS values at each iteration Message-ID: <528DA268.1030704@gmail.com> Hi, Due to a special case, I need modify RHS values of the system of eqns at each iteration. Hence, what I can do is to solve the eqn using KSPSolve, limit the maxits in KSPSetTolerances to 1, modify the RHS, and then run KSPSolve again. It will loop until it reaches the convergences I specify. My question is if that's the only way, and will it suffer a large drop in speed? Thank you! -- Yours sincerely, TAY wee-beng From jedbrown at mcs.anl.gov Thu Nov 21 00:09:46 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Wed, 20 Nov 2013 23:09:46 -0700 Subject: [petsc-users] Modify RHS values at each iteration In-Reply-To: <528DA268.1030704@gmail.com> References: <528DA268.1030704@gmail.com> Message-ID: <87zjoywaj9.fsf@jedbrown.org> TAY wee-beng writes: > Hi, > > Due to a special case, I need modify RHS values of the system of eqns at > each iteration. > > Hence, what I can do is to solve the eqn using KSPSolve, limit the > maxits in KSPSetTolerances to 1, modify the RHS, and then run KSPSolve > again. > > It will loop until it reaches the convergences I specify. > > My question is if that's the only way, There is almost certainly a better way to formulate your problem. > and will it suffer a large drop in speed? That algorithm (without RHS modification) reduces to Richardson preconditioned by one step of your chosen preconditioned iterative method. This likely converges much worse than a Krylov method. Note that in the form you describe, the preconditioner will be applied twice and the operator once. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From zonexo at gmail.com Thu Nov 21 01:40:48 2013 From: zonexo at gmail.com (TAY wee-beng) Date: Thu, 21 Nov 2013 15:40:48 +0800 Subject: [petsc-users] Modify RHS values at each iteration In-Reply-To: <87zjoywaj9.fsf@jedbrown.org> References: <528DA268.1030704@gmail.com> <87zjoywaj9.fsf@jedbrown.org> Message-ID: <528DB900.6000704@gmail.com> On 21/11/2013 2:09 PM, Jed Brown wrote: > TAY wee-beng writes: > >> Hi, >> >> Due to a special case, I need modify RHS values of the system of eqns at >> each iteration. >> >> Hence, what I can do is to solve the eqn using KSPSolve, limit the >> maxits in KSPSetTolerances to 1, modify the RHS, and then run KSPSolve >> again. >> >> It will loop until it reaches the convergences I specify. >> >> My question is if that's the only way, > There is almost certainly a better way to formulate your problem. Hi, in that case, do you want any suggestion? Is it possible to apply the preconditioner only once at the start, and not at every intermediate step? Thank you! > >> and will it suffer a large drop in speed? > That algorithm (without RHS modification) reduces to Richardson > preconditioned by one step of your chosen preconditioned iterative > method. This likely converges much worse than a Krylov method. > > Note that in the form you describe, the preconditioner will be applied > twice and the operator once. From jedbrown at mcs.anl.gov Thu Nov 21 01:44:13 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 21 Nov 2013 00:44:13 -0700 Subject: [petsc-users] Modify RHS values at each iteration In-Reply-To: <528DB900.6000704@gmail.com> References: <528DA268.1030704@gmail.com> <87zjoywaj9.fsf@jedbrown.org> <528DB900.6000704@gmail.com> Message-ID: <87pppuw65u.fsf@jedbrown.org> TAY wee-beng writes: > Hi, in that case, do you want any suggestion? Is it possible to apply > the preconditioner only once at the start, and not at every intermediate > step? You'll have to explain what you are trying to do. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From jsd1 at rice.edu Thu Nov 21 03:43:59 2013 From: jsd1 at rice.edu (Justin Dong (Me)) Date: Thu, 21 Nov 2013 03:43:59 -0600 Subject: [petsc-users] Troubles installing PETSc In-Reply-To: <87r4aay0fk.fsf@jedbrown.org> References: <7F374499-AB94-4B8B-BB79-8B143FF1A852@rice.edu> <87r4aay0fk.fsf@jedbrown.org> Message-ID: Here is the error message: > $ make ex2 > cc ex2.c -o ex2 > ex2.c:13:10: fatal error: 'petscsys.h' file not found > #include > ^ > 1 error generated. > make: *** [ex2] Error 1 and this is the entire makefile. It?s long but it?s literally a copy paste from the exercises. Also, I?ve fixed the issue with the PETSC_ARCH so it?s set to linux-gnu now. And when I do cd $PETSC_DIR, I?m moved to the correct directory. > ALL: > > CFLAGS = > FFLAGS = -I${PETSC_DIR}/include/finclude > CPPFLAGS = > FPPFLAGS = > LOCDIR = src/sys/examples/tutorials/ > EXAMPLESC = ex1.c ex2.c ex3.c ex4.c ex5.c ex6.c ex7.c ex9.c ex11.c > EXAMPLESF = ex3f.F ex4f.F ex4f90.F90 ex5f90.F90 ex8f90.F90 ex10f90.F90 > MANSEC = Sys > CLEANFILES = binaryoutput binaryoutput.info > > include ${PETSC_DIR}/conf/variables > include ${PETSC_DIR}/conf/rules > > ex1: ex1.o chkopts > -${CLINKER} -o ex1 ex1.o ${PETSC_SYS_LIB} > ${RM} -f ex1.o > > ex2: ex2.o chkopts > -${CLINKER} -o ex2 ex2.o ${PETSC_SYS_LIB} > ${RM} -f ex2.o > > ex3: ex3.o chkopts > -${CLINKER} -o ex3 ex3.o ${PETSC_SYS_LIB} > ${RM} -f ex3.o > > ex3f: ex3f.o chkopts > -${FLINKER} -o ex3f ex3f.o ${PETSC_SYS_LIB} > ${RM} -f ex3f.o > > ex4: ex4.o chkopts > -${CLINKER} -o ex4 ex4.o ${PETSC_SYS_LIB} > ${RM} -f ex4.o > > ex4f: ex4f.o chkopts > -${FLINKER} -o ex4f ex4f.o ${PETSC_SYS_LIB} > ${RM} -f ex4f.o > > ex4f90: ex4f90.o chkopts > -${FLINKER} -o ex4f90 ex4f90.o ${PETSC_SYS_LIB} > ${RM} -f ex4f90.o > > ex5: ex5.o chkopts > -${CLINKER} -o ex5 ex5.o ${PETSC_SYS_LIB} > ${RM} -f ex5.o > > ex5f90: ex5f90.o chkopts > -${FLINKER} -o ex5f90 ex5f90.o ${PETSC_SYS_LIB} > ${RM} ex5f90.o > > ex6: ex6.o chkopts > -${CLINKER} -o ex6 ex6.o ${PETSC_SYS_LIB} > ${RM} -f ex6.o > > ex7: ex7.o chkopts > -${CLINKER} -o ex7 ex7.o ${PETSC_SYS_LIB} > ${RM} -f ex7.o > > ex8f90: ex8f90.o chkopts > -${FLINKER} -o ex8f90 ex8f90.o ${PETSC_SYS_LIB} > ${RM} ex8f90.o > > ex9: ex9.o chkopts > -${CLINKER} -o ex9 ex9.o ${PETSC_SYS_LIB} > ${RM} -f ex9.o > > ex10f90: ex10f90.o chkopts > -${FLINKER} -o ex10f90 ex10f90.o ${PETSC_SYS_LIB} > ${RM} -f ex10f90.o > > ex11: ex11.o chkopts > -${CLINKER} -o ex11 ex11.o ${PETSC_SYS_LIB} > ${RM} -f ex11.o > # > # ------------------------------------------------------------------------ > # > runex1: > -@${MPIEXEC} -n 1 ./ex1 > ex1_1.tmp 2>&1; \ > if (${DIFF} output/ex1_1.out ex1_1.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex1_1, diffs above \n========================================="; fi; \ > ${RM} -f ex1_1.tmp > > runex2: > -@${MPIEXEC} -n 1 ./ex2 > ex2_1.tmp 2>&1; \ > if (${DIFF} output/ex2_1.out ex2_1.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex2_1, diffs above \n========================================="; fi; \ > ${RM} -f ex2_1.tmp > > runex3: > -@${MPIEXEC} -n 1 ./ex3 > > runex3f: > -@${MPIEXEC} -n 1 ./ex3f > > runex4: > -@${MPIEXEC} -n 5 ./ex4 | sort > ex4_1.tmp 2>&1; \ > if (${DIFF} output/ex4_1.out ex4_1.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex4_1, diffs above \n========================================="; fi; \ > ${RM} -f ex4_1.tmp > > runex4f: > -@${MPIEXEC} -n 5 ./ex4f | sort > ex4f_1.tmp 2>&1; \ > if (${DIFF} output/ex4f_1.out ex4f_1.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex4f_1, diffs above \n========================================="; fi; \ > ${RM} -f ex4f_1.tmp > > runex4f90: > -@${MPIEXEC} -n 1 ./ex4f90 > ex4f90_1.tmp 5>&1; \ > if (${DIFF} output/ex4f90_1.out ex4f90_1.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex4f90_1, diffs above \n========================================="; fi; \ > ${RM} -f ex4f90_1.tmp > > runex5: > -@${MPIEXEC} -n 1 ./ex5 -pbag_rho 44 -pbag_do_output true > ex5_1.tmp 5>&1; \ > if (${DIFF} output/ex5_1.out ex5_1.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex5_1, diffs above \n========================================="; fi; \ > ${RM} -f ex5_1.tmp > > runex5f90: > -@${MPIEXEC} -n 1 ./ex5f90 -pbag_rarray 4,5,88 > ex5f90_1.tmp 5>&1; \ > if (${DIFF} output/ex5f90_1.out ex5f90_1.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex5f90_1, diffs above \n========================================="; fi; \ > ${RM} -f ex5f90_1.tmp > > runex6: > -@${MPIEXEC} -n 1 ./ex6 > > runex8f90: > -@${MPIEXEC} -n 1 ./ex8f90 -joe_jeff b456 > ex8f90_1.tmp 8>&1; \ > if (${DIFF} output/ex8f90_1.out ex8f90_1.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex8f90_1, diffs above \n========================================="; fi; \ > ${RM} -f ex8f90_1.tmp > > runex9_1: > - at PETSC_OPTIONS= ${MPIEXEC} -n 1 ./ex9 -f petsc.yml > ex9_1.tmp 5>&1; \ > if (${DIFF} output/ex9_1.out ex9_1.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex9_1, diffs above \n========================================="; fi; \ > ${RM} -f ex9_1.tmp > > runex9_2: > - at PETSC_OPTIONS= ${MPIEXEC} -n 1 ./ex9 -options_file_yaml petsc.yml > ex9_2.tmp 5>&1; \ > if (${DIFF} output/ex9_2.out ex9_2.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex9_2, diffs above \n========================================="; fi; \ > ${RM} -f ex9_2.tmp > > runex10f90_1: > - at PETSC_OPTIONS= ${MPIEXEC} -n 1 ./ex10f90 -f petsc.yml > ex10f90_1.tmp 8>&1; \ > if (${DIFF} output/ex10f90_1.out ex10f90_1.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex10f90_1, diffs above \n========================================="; fi; \ > ${RM} -f ex10f90_1.tmp > > runex10f90_2: > - at PETSC_OPTIONS= ${MPIEXEC} -n 1 ./ex10f90 -options_file_yaml petsc.yml > ex10f90_2.tmp 8>&1; \ > if (${DIFF} output/ex10f90_2.out ex10f90_2.tmp) then true; \ > else echo ${PWD} ; echo "Possible problem with with ex10f90_2, diffs above \n========================================="; fi; \ > ${RM} -f ex10f90_2.tmp > > runex11: > -@${MPIEXEC} -n 1 ./ex11 > > TESTEXAMPLES_C = ex1.PETSc runex1 ex1.rm ex2.PETSc runex2 ex2.rm \ > ex3.PETSc runex3 ex3.rm ex4.PETSc runex4 ex4.rm ex6.PETSc runex6 ex6.rm ex11.PETSc runex11 ex11.rm > TESTEXAMPLES_C_NOCOMPLEX = ex5.PETSc runex5 ex5.rm > TESTEXAMPLES_C_X_MPIUNI = ex1.PETSc runex1 ex1.rm ex2.PETSc runex2 ex2.rm \ > ex3.PETSc runex3 ex3.rm > TESTEXAMPLES_FORTRAN = ex3f.PETSc runex3f ex3f.rm ex4f.PETSc runex4f ex4f.rm > TESTEXAMPLES_F90 = ex4f90.PETSc runex4f90 ex4f90.rm > TESTEXAMPLES_F2003 = ex5f90.PETSc runex5f90 ex5f90.rm ex8f90.PETSc runex8f90 ex8f90.rm > TESTEXAMPLES_YAML = ex9.PETSc runex9_1 runex9_2 ex9.rm ex10f90.PETSc runex10f90_1 runex10f90_2 ex10f90.rm > include ${PETSC_DIR}/conf/test On Nov 20, 2013, at 8:05 PM, Jed Brown wrote: > "Justin Dong (Me)" writes: -------------- next part -------------- An HTML attachment was scrubbed... URL: From zonexo at gmail.com Thu Nov 21 04:06:11 2013 From: zonexo at gmail.com (TAY wee-beng) Date: Thu, 21 Nov 2013 18:06:11 +0800 Subject: [petsc-users] Modify RHS values at each iteration In-Reply-To: <87pppuw65u.fsf@jedbrown.org> References: <528DA268.1030704@gmail.com> <87zjoywaj9.fsf@jedbrown.org> <528DB900.6000704@gmail.com> <87pppuw65u.fsf@jedbrown.org> Message-ID: <528DDB13.7080208@gmail.com> Hi Jed, I need a solve a Poisson equation, which is part of my finite volume CFD code. The RHS terms are calculated from the velocity. In the standard mtd, the RHS terms are calculated from the velocity, which then solves the Poisson eqn to get the pressure. Now, to improve the accuracy of the pressure obtained, a paper suggested obtaining the pressure from the velocity in 1 iteration. Then use the new pressure for the momentum eqn to get a more accurate intermediate velocity. The RHS terms are recomputed again using the new velocity, which is then used to calculate the new pressure. This cycle goes on until the pressure converges to a certain value. Hence, I wonder what is the best way to do it. Thanks! Yours sincerely, TAY wee-beng On 21/11/2013 3:44 PM, Jed Brown wrote: > TAY wee-beng writes: >> Hi, in that case, do you want any suggestion? Is it possible to apply >> the preconditioner only once at the start, and not at every intermediate >> step? > You'll have to explain what you are trying to do. From jedbrown at mcs.anl.gov Thu Nov 21 06:41:49 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 21 Nov 2013 05:41:49 -0700 Subject: [petsc-users] Modify RHS values at each iteration In-Reply-To: <528DDB13.7080208@gmail.com> References: <528DA268.1030704@gmail.com> <87zjoywaj9.fsf@jedbrown.org> <528DB900.6000704@gmail.com> <87pppuw65u.fsf@jedbrown.org> <528DDB13.7080208@gmail.com> Message-ID: <87mwkyvsdu.fsf@jedbrown.org> TAY wee-beng writes: > Hi Jed, > > I need a solve a Poisson equation, which is part of my finite volume CFD > code. > > The RHS terms are calculated from the velocity. > > In the standard mtd, the RHS terms are calculated from the velocity, > which then solves the Poisson eqn to get the pressure. > > Now, to improve the accuracy of the pressure obtained, a paper suggested > obtaining the pressure from the velocity in 1 iteration. Then use the > new pressure for the momentum eqn to get a more accurate intermediate > velocity. The RHS terms are recomputed again using the new velocity, > which is then used to calculate the new pressure. This cycle goes on > until the pressure converges to a certain value. This is a crude iterative method for solving a coupled problem. I recommend that you solve the coupled problem. See the section of the user's manual on solving block problems, and the examples in PETSc that solve Stokes-type problems. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From jedbrown at mcs.anl.gov Thu Nov 21 06:48:48 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 21 Nov 2013 05:48:48 -0700 Subject: [petsc-users] Troubles installing PETSc In-Reply-To: References: <7F374499-AB94-4B8B-BB79-8B143FF1A852@rice.edu> <87r4aay0fk.fsf@jedbrown.org> Message-ID: <87k3g1x6mn.fsf@jedbrown.org> "Justin Dong (Me)" writes: > Here is the error message: > >> $ make ex2 >> cc ex2.c -o ex2 >> ex2.c:13:10: fatal error: 'petscsys.h' file not found >> #include >> ^ >> 1 error generated. >> make: *** [ex2] Error 1 Looks like this make command is not being run from the same directory as the makefile, or perhaps "makefile" is misspelled? Can you build examples in-place within the source tree? $ cd src/sys/examples/tutorials $ make ex2 Then go back to the directory containing your files (ex2.c and makefile) and try $ make -f makefile ex2 You can trim it down to the minimal makefile like Matt suggested, though this should not change anything. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From asmund.ervik at ntnu.no Thu Nov 21 06:52:48 2013 From: asmund.ervik at ntnu.no (=?ISO-8859-1?Q?=C5smund_Ervik?=) Date: Thu, 21 Nov 2013 13:52:48 +0100 Subject: [petsc-users] Modify RHS values at each iteration In-Reply-To: References: Message-ID: <528E0220.8090307@ntnu.no> Hi Tay, I've just been working on ways to get more accurate pressure from the Poisson equation. I can highly recommend this article by Guermond and coworkers: http://dx.doi.org/10.1016/j.cma.2005.10.010 It's so long that it may be scary, but just reading 3.1 and 3.2 gives quite a bit of information already. The article sums up the literature quite nicely. Regards, ?smund Ervik > Date: Thu, 21 Nov 2013 18:06:11 +0800 > From: TAY wee-beng > To: Jed Brown , PETSc list > > Subject: Re: [petsc-users] Modify RHS values at each iteration > Message-ID: <528DDB13.7080208 at gmail.com> > Content-Type: text/plain; charset=ISO-8859-1; format=flowed > > Hi Jed, > > I need a solve a Poisson equation, which is part of my finite volume CFD > code. > > The RHS terms are calculated from the velocity. > > In the standard mtd, the RHS terms are calculated from the velocity, > which then solves the Poisson eqn to get the pressure. > > Now, to improve the accuracy of the pressure obtained, a paper suggested > obtaining the pressure from the velocity in 1 iteration. Then use the > new pressure for the momentum eqn to get a more accurate intermediate > velocity. The RHS terms are recomputed again using the new velocity, > which is then used to calculate the new pressure. This cycle goes on > until the pressure converges to a certain value. > > Hence, I wonder what is the best way to do it. > > Thanks! > > Yours sincerely, > > TAY wee-beng > > On 21/11/2013 3:44 PM, Jed Brown wrote: >> TAY wee-beng writes: >>> Hi, in that case, do you want any suggestion? Is it possible to apply >>> the preconditioner only once at the start, and not at every intermediate >>> step? >> You'll have to explain what you are trying to do. > From einar.sorheim at gmail.com Thu Nov 21 07:41:52 2013 From: einar.sorheim at gmail.com (=?ISO-8859-1?Q?Einar_S=F8rheim?=) Date: Thu, 21 Nov 2013 14:41:52 +0100 Subject: [petsc-users] Problem with GAMG elasticity case:'createdefaultdata' not set(?) need to support NULL data! In-Reply-To: <71DB6D5F-F64E-4351-8B91-43D660E2E271@mcs.anl.gov> References: <8EE4826E-4436-4FE4-84D9-2EBF4C877340@lbl.gov> <1DD67C25-D646-4BD7-B44D-709D505CF267@lbl.gov> <87txggipud.fsf@mcs.anl.gov> <6958C2D4-F7CA-43F7-8E56-A3933E259FA5@lbl.gov> <83ED04D2-DE73-4A70-BDEE-0994317E0051@mcs.anl.gov> <71DB6D5F-F64E-4351-8B91-43D660E2E271@mcs.anl.gov> Message-ID: I am on the windows platform so unfortunatley no valgrind, however have been able to track down the problem in the debugger and it seems like a matrixpointer gets corrupted when using the fortran-c interface in petsc. At some point in the calculation the matrix pointer in MatSetSize gets corrupted when called from fortran, propably in this cast: void PETSC_STDCALL matsetsizes_(Mat A,PetscInt *m,PetscInt *n,PetscInt *M,PetscInt *N, int *__ierr ){ *__ierr = MatSetSizes( (Mat)PetscToPointer((A) ),*m,*n,*M,*N); } when entering the MatSetSize in the debugger the A-matrix pointer is corrupt and I get an access violation on code line: PetscValidHeaderSpecific(A,MAT_CLASSID,1); Any ideas what may be the cause of this? Regards Einar 2013/10/24 Barry Smith > > So apparently no memory leak. > > If you are still getting crashes you will need to run the case that > crashes 1) with valgrind if you haven?t or 2) using > -on_error_attach_debugger or -start_in_debugger to track down the cause of > the crash. > > Barry > > > > On Oct 24, 2013, at 2:09 AM, Einar S?rheim > wrote: > > > Well i I did a rerun without -malloc_log, this is what I got (I have > also attached the log-sumary file): > > > > [0] PetscFinalize(): PetscFinalize() called > > [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > > Summary of Memory Usage in PETSc > > [0]Current space PetscMalloc()ed 45936, max space PetscMalloced() > 1.13728e+009 > > [0]OS cannot compute process memory > > [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850688 > -2080374784 > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > [0] PetscFOpen(): Opening file Log.0 > > [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm > -2080374784 > > [0] Petsc_DelComm_Inner(): Removing reference to PETSc communicator > embedded in a user MPI_Comm -2080374784 > > [0] Petsc_DelComm_Outer(): User MPI_Comm 1140850688 is being freed after > removing reference from inner PETSc comm to this outer comm > > [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 > > [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm > -2080374784 > > [0] Petsc_DelThreadComm(): Deleting thread communicator data in an > MPI_Comm -2080374784 > > [0] Petsc_DelCounter(): Deleting counter data in an MPI_Comm -2080374784 > > > > > > > > 2013/10/23 Barry Smith > > > > Don?t run with the -malloc_log Mark did not suggest using that option. > > > > > > On Oct 23, 2013, at 8:48 AM, Einar S?rheim > wrote: > > > > > I tried running with the following options : > > > -info > > > -malloc > > > -malloc_log > > > -malloc_dump > > > -malloc_debug > > > -memory_info > > > -log_summary log_summary.txt > > > -log > > > but didn't get a lot wiser however in my case would it be more > efficient to keep the ksp solver object, if so should I have one object eg. > for the thermal problem and another one for the mechanical? > > > The output from the end of the run: > > > Total wall clock time actually solving: 856.231474 > > > [0] PetscFinalize(): PetscFinalize() called > > > [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 > -2080374784 max tags = 2147483647 > > > Summary of Memory Usage in PETSc > > > [0]Current space PetscMalloc()ed 45936, max space PetscMalloced() > 1.13728e+009 > > > [0]OS cannot compute process memory > > > [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850688 > -2080374784 > > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > > [0] PetscFOpen(): Opening file Log.0 > > > [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm > -2080374784 > > > [0] Petsc_DelComm_Inner(): Removing reference to PETSc communicator > embedded in a user MPI_Comm -2080374784 > > > [0] Petsc_DelComm_Outer(): User MPI_Comm 1140850688 is being freed > after removing reference from inner PETSc comm to this outer comm > > > [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 > > > [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm > -2080374784 > > > [0] Petsc_DelThreadComm(): Deleting thread communicator data in an > MPI_Comm -2080374784 > > > [0] Petsc_DelCounter(): Deleting counter data in an MPI_Comm > -2080374784 > > > [0]PETSC ERROR: --------------------- Error Message > ------------------------------------ > > > [0]PETSC ERROR: Object is in wrong state! > > > [0]PETSC ERROR: PetscMallocDumpLog() called without call to > PetscMallocSetDumpLog() this is often due to > > > setting the option -malloc_log AFTER > PetscInitialize() with PetscOptionsInsert() or PetscOptionsInsertFile()! > > > [0]PETSC ERROR: > ------------------------------------------------------------------------ > > > [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > > > [0]PETSC ERROR: See docs/changes/index.html for recent updates. > > > [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > > > [0]PETSC ERROR: See docs/index.html for manual pages. > > > [0]PETSC ERROR: > ------------------------------------------------------------------------ > > > [0]PETSC ERROR: D:\Programming\gits\Alsim\SharedModules - > Copy\Source\Release\Alsim_nypetsc.exe on a arch-mswin-c-debug named CMP3 by > admeinar Wed Oct 23 11:54:03 2013 > > > [0]PETSC ERROR: Libraries linked from > /cygdrive/d/Programming/petsc-3.4.3/arch-mswin-c-debug/lib > > > [0]PETSC ERROR: Configure run at Thu Oct 17 08:23:18 2013 > > > [0]PETSC ERROR: Configure options --with-debugging=1 --with-openmp=yes > --with-pthread=no --with-cc="win32fe icl" --with-fc="win32fe ifort" > --with-cxx="win32fe icl" --with-blas-lapack-dir="C:/Program Files > (x86)/Intel/Composer XE 2013/mkl/lib/intel64" --download-superlu > --with-sowing=0 --with-c2html=0 --with-mpi-dir="C:/Program Files > (x86)/Intel/MPI/4.1.0.028/em64t" --useThreads=0 --useThreads=0 > > > [0]PETSC ERROR: > ------------------------------------------------------------------------ > > > [0]PETSC ERROR: PetscMallocDumpLog() line 652 in > src/sys/memory/D:\PROGRA~1\PETSC-~1.3\src\sys\memory\mtr.c > > > [0]PETSC ERROR: PetscFinalize() line 1185 in > src/sys/objects/D:\PROGRA~1\PETSC-~1.3\src\sys\objects\pinit.c > > > > > > > > > 2013/10/19 Mark F. Adams > > > You can try running a problem that finishes with > > > > > > -malloc > > > -malloc_debug > > > -malloc_dump > > > > > > to see if there are any memory issues. It looks like you are > completely destroying the solver object each time and starting over. > Valgrind would be very useful here because it looks like memory is getting > corrupted if the same code runs many times. > > > > > > > > > On Oct 19, 2013, at 3:39 PM, Einar S?rheim > wrote: > > > > > >> yes, I call the given code many times, what i have seen so far is > that changing e.g. the order of destruction statements changes the time of > failure,I will try to debug to get more details > > >> > > >> > > >> 2013/10/18 Mark F. Adams > > >> I would guess there is some memory corruption or leak someplace. Try > running with -on_error_attach_debugger, and try to get some more > information. Do you just repeat a call to the code below many times? Is > this deterministic, does it fail on the same solve each time? > > >> > > >> > > >> On Oct 18, 2013, at 9:30 AM, Einar S?rheim > wrote: > > >> > > >>> Uncommenting KSPSetFromOptions solved my intial problem, however I > got a new problem that occurs after many timesteps, in our code we use > petsc gamg as a solver for several problems(e.g. thermal and mechanical) at > each time step, I have a suspicion it might have something to do with > cleanup after solving, the rror appears after many calls to the petsc gamg > solver interface, the error message is the following: > > >>> > > >>> [0]PETSC ERROR: > ------------------------------------------------------------------------ > > >>> [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation > Violation, probably memory access out of range > > >>> [0]PETSC ERROR: Try option -start_in_debugger or > -on_error_attach_debugger > > >>> [0]PETSC ERROR: or see > http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSCERROR: or try > http://valgrind.org on GNU/linux and Apple Mac OS X to find memory > corruption errors > > >>> [0]PETSC ERROR: likely location of problem given in stack below > > >>> [0]PETSC ERROR: --------------------- Stack Frames > ------------------------------------ > > >>> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not > available, > > >>> [0]PETSC ERROR: INSTEAD the line number of the start of the > function > > >>> [0]PETSC ERROR: is given. > > >>> [0]PETSC ERROR: [0] MatSetNearNullSpace line 7580 > src/mat/interface/D:\PROGRA~1\PETSC-~1.3\src\mat\INTERF~1\matrix.c > > >>> [0]PETSC ERROR: --------------------- Error Message > ------------------------------------ > > >>> [0]PETSC ERROR: Signal received! > > >>> [0]PETSC ERROR: > ------------------------------------------------------------------------ > > >>> [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > > >>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. > > >>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > > >>> [0]PETSC ERROR: See docs/index.html for manual pages. > > >>> [0]PETSC ERROR: > ------------------------------------------------------------------------ > > >>> [0]PETSC ERROR: D:\Programming\gits\Alsim\SharedModules - > Copy\Source\Release\Alsim_nypetsc.exe on a arch-mswin-c-debug named CMP3 by > admeinar Fri Oct 18 12:26:04 2013 > > >>> [0]PETSC ERROR: Libraries linked from > /cygdrive/d/Programming/petsc-3.4.3/arch-mswin-c-debug/lib > > >>> [0]PETSC ERROR: Configure run at Thu Oct 17 08:23:18 2013 > > >>> [0]PETSC ERROR: Configure options --with-debugging=1 > --with-openmp=yes --with-pthread=no --with-cc="win32fe icl" > --with-fc="win32fe ifort" --with-cxx="win32fe icl" > --with-blas-lapack-dir="C:/Program Files (x86)/Intel/Composer XE > 2013/mkl/lib/intel64" --download-superlu --with-sowing=0 --with-c2html=0 > --with-mpi-dir="C:/Program Files (x86)/Intel/MPI/4.1.0.028/em64t" > --useThreads=0 --useThreads=0 > > >>> > > >>> The source code for the petsc interface: > > >>> > > >>> subroutine > petscsolver(i_nodes,r_elements_,i_rowstart_,i_columnindex_,r_x_,r_b_,i_Nnz,i_Precond, > i_Solver > > >>> $ ,coords_,i_nodedof) > > >>> #include > > >>> #include > > >>> #include > > >>> #include > > >>> #include > > >>> #include > > >>> > > >>> integer i_nodes, i_Nnz, i_Precond, i_solver, i_nodedof > > >>> integer i_rowstart_(*),i_columnindex_(*) > > >>> double precision r_elements_(*),r_x_(*), r_b_(*),coords_(*) > > >>> integer, allocatable :: i_indices_(:) > > >>> ! Variables: > > >>> ! > > >>> ! A - matrix that defines linear system > > >>> ! ksp - KSP context > > >>> ! ksp - KSP context > > >>> ! x, b, u - approx solution, RHS, exact solution vectors > > >>> ! > > >>> ! implicit none > > >>> Vec x,u,b,vec_coords > > >>> PC pc > > >>> Mat A,F,Matnull > > >>> KSP ksp > > >>> PetscInt i,j,II,JJ,m,n,i1,imaxiter > > >>> PetscInt Istart,Iend > > >>> PetscInt nsteps,one, neg_one > > >>> PetscErrorCode ierr > > >>> PetscBool flg > > >>> PetscScalar v,x_array_(1), norm, tol, t0,t1 > > >>> PetscOffset i_x > > >>> PetscViewer viewer > > >>> > > >>> > > >>> call PetscTime(t0,ierr) > > >>> write(*,*) "Enter petsc" > > >>> > > >>> n = 1 > > >>> nsteps = 1 > > >>> one = 1 > > >>> neg_one = 1 > > >>> i1 = 1 > > >>> imaxiter = 1200 > > >>> > > >>> > > >>> call MatCreate(PETSC_COMM_WORLD,A,ierr) > > >>> call > MatSetSizes(A,PETSC_DECIDE,PETSC_DECIDE,i_nodes,i_nodes,ierr) > > >>> call MatSetType(A, MATSEQAIJ,ierr) > > >>> call MatSetUp(A,ierr) > > >>> call MatSeqAIJSetPreallocation(A,100,PETSC_NULL_INTEGER,ierr) > > >>> ! The matrix is partitioned by contiguous chunks of rows across the > > >>> ! processors. Determine which rows of the matrix are locally owned. > > >>> > > >>> call MatGetOwnershipRange(A,Istart,Iend,ierr) > > >>> > > >>> > > >>> write(*,*) "Start of matrix fill", i_nodes, Istart, Iend > > >>> do i=1,i_nodes > > >>> ! write(*,*) "Start of loop",i !!, i_indices_(i+1) > > >>> ! i_indices_(i+1) = i > > >>> ! write(*,*) "Start of loop2" > > >>> do j=i_rowstart_(i),i_rowstart_(i+1)-1 > > >>> ! write(*,*) "Start of loop 3" > > >>> v = r_elements_(j) > > >>> ! write(*,*) i1,i,i1,j-1,v > > >>> call > MatSetValues(A,i1,i-1,i1,i_columnindex_(j)-1,v,INSERT_VALUES,ierr) > > >>> if (ierr.gt.0) stop > > >>> end do > > >>> end do > > >>> ! write(*,*) "End of matrix fill" > > >>> > > >>> ! Assemble matrix, using the 2-step process: > > >>> ! MatAssemblyBegin(), MatAssemblyEnd() > > >>> ! Computations can be done while messages are in transition > > >>> ! by placing code between these two statements. > > >>> > > >>> call MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY,ierr) > > >>> call MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY,ierr) > > >>> > > >>> ! Create parallel vectors. > > >>> ! - When using VecCreate(), the parallel partitioning of the vector > > >>> ! is determined by PETSc at runtime. > > >>> ! - Note: We form 1 vector from scratch and then duplicate as > needed. > > >>> ! write(*,*) "Start of vector fill" > > >>> call VecCreate(PETSC_COMM_WORLD,u,ierr) > > >>> call VecSetType(u, VECSEQ,ierr) > > >>> call VecSetSizes(u,PETSC_DECIDE,i_nodes,ierr) > > >>> ! call VecSetFromOptions(u,ierr) > > >>> call VecDuplicate(u,b,ierr) > > >>> call VecDuplicate(b,x,ierr) > > >>> do i=1,i_nodes > > >>> call VecSetValues(x,i1,i-1,r_x_(i),INSERT_VALUES,ierr) > > >>> call VecSetValues(b,i1,i-1,r_b_(i),INSERT_VALUES,ierr) > > >>> enddo > > >>> ! Assemble vector > > >>> > > >>> call VecAssemblyBegin(x,ierr) > > >>> call VecAssemblyEnd(x,ierr) > > >>> call VecAssemblyBegin(b,ierr) > > >>> call VecAssemblyEnd(b,ierr) > > >>> call PetscTime(t1,ierr) > > >>> ! write(*,*) "End of vec fill time spent :", t1-t0 > > >>> ! Create linear solver context > > >>> > > >>> call KSPCreate(PETSC_COMM_WORLD,ksp,ierr) > > >>> > > >>> > > >>> ! Set operators. Here the matrix that defines the linear system > > >>> ! also serves as the preconditioning matrix. > > >>> > > >>> call KSPSetOperators(ksp,A,A,DIFFERENT_NONZERO_PATTERN,ierr) > > >>> > > >>> > > >>> call KSPGetPC(ksp,pc,ierr) > > >>> tol = 1.e-20 > > >>> call > KSPSetTolerances(ksp,tol,PETSC_DEFAULT_DOUBLE_PRECISION,PETSC_DEFAULT_DOUBLE_PRECISION,imaxiter,ierr) > > >>> select case (i_Precond) > > >>> case ( 1 ) > > >>> call PCSetType(pc,PCJACOBI,ierr) > > >>> case ( 2 ) > > >>> call PCSetType(pc,PCILU,ierr) > > >>> call PCFactorSetMatSolverPackage(pc,MATSOLVERSUPERLU,ierr); > > >>> call PCFactorSetUpMatSolverPackage(pc,ierr); > > >>> call PCFactorGetMatrix(pc,F); > > >>> call MatSuperluSetILUDropTol(F,1.e-4); > > >>> case ( 3 ) > > >>> call PCSetType(pc,PCGAMG,ierr) > > >>> ! call PCGAMGInitializePackage() > > >>> ! call PCCreate_GAMG(pc,ierr) > > >>> case DEFAULT > > >>> call PCSetType(pc,PCJACOBI,ierr) > > >>> end select > > >>> > > >>> ! call PCSetType(pc,PCJACOBI,ierr) > > >>> select case (i_solver) > > >>> case ( 0 ) > > >>> call KSPSetType(ksp,KSPCG,ierr) > > >>> case DEFAULT > > >>> call KSPSetType(ksp,KSPCG,ierr) > > >>> end select > > >>> call KSPCGSetType(ksp,KSP_CG_SYMMETRIC,ierr) > > >>> if (i_nodedof==3) then > > >>> write(*,*) "Set coord, number of nodes is:", > i_nodes/i_nodedof > > >>> call > VecCreateSeqWithArray(MPI_COMM_SELF,3,i_nodes,coords_,vec_coords,ierr) > > >>> call MatNullSpaceCreateRigidBody(vec_coords,Matnull,ierr) > > >>> call MatSetNearNullSpace(A,Matnull,ierr) > > >>> call MatNullSpaceDestroy(Matnull,ierr) > > >>> > > >>> write(*,*) "Finish setting null space ierr =", ierr > > >>> ! call PCSetCoordinates( pc, 3,i_nodes/i_nodedof, coords_, > ierr ) > > >>> > > >>> end if > > >>> ! call KSPMonitorSet(ksp, KSPMonitorDefault, PETSC_NULL, > PETSC_NULL,ierr) > > >>> ! Set runtime options, e.g., > > >>> ! -ksp_type -pc_type -ksp_monitor -ksp_rtol > > >>> ! These options will override those specified above as long as > > >>> ! KSPSetFromOptions() is called _after_ any other customization > > >>> ! routines. > > >>> > > >>> call KSPSetFromOptions(ksp,ierr) > > >>> call KSPSetInitialGuessNonzero(ksp,PETSC_TRUE,ierr) > > >>> call KSPSetUp(ksp,ierr) > > >>> call PetscTime(t0,ierr) > > >>> write(*,*) "Time setup", t0-t1 > > >>> ! - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > - - > > >>> ! Solve the linear system > > >>> ! - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > - - > > >>> > > >>> call KSPSolve(ksp,b,x,ierr) > > >>> call PetscTime(t1,ierr) > > >>> write(*,*) "end solve, time used:",t1-t0 > > >>> call KSPGetSolution(ksp,x,ierr) > > >>> call VecGetArray(x,x_array_,i_x,ierr) > > >>> do i=1,i_nodes > > >>> r_x_(i) = x_array_(i_x + i) > > >>> enddo > > >>> call VecRestoreArray(x,x_array_,i_x,ierr) > > >>> > > >>> > > >>> ! Free work space. All PETSc objects should be destroyed when they > > >>> ! are no longer needed. > > >>> ! call PCDestroy(pc,ierr) > > >>> > > >>> call VecDestroy(u,ierr) > > >>> call VecDestroy(x,ierr) > > >>> call VecDestroy(b,ierr) > > >>> call MatDestroy(A,ierr) > > >>> if (i_nodedof==3) then > > >>> call VecDestroy(vec_coords,ierr) > > >>> call MatDestroy(Matnull,ierr) > > >>> endif > > >>> call KSPDestroy(ksp,ierr) > > >>> call PetscTime(t0,ierr) > > >>> write(*,*) "time celan up :", t0-t1 > > >>> > > >>> end > > >>> > > >>> -- > > >>> Einar S?rheim > > >>> > > >> > > >> > > >> > > >> > > >> -- > > >> Einar S?rheim > > > > > > > > > > > > > > > -- > > > Einar S?rheim > > > > > > > > > > -- > > Einar S?rheim > > > > -- Einar S?rheim -------------- next part -------------- An HTML attachment was scrubbed... URL: From balay at mcs.anl.gov Thu Nov 21 11:17:24 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Thu, 21 Nov 2013 11:17:24 -0600 (CST) Subject: [petsc-users] Problem with GAMG elasticity case:'createdefaultdata' not set(?) need to support NULL data! In-Reply-To: References: <8EE4826E-4436-4FE4-84D9-2EBF4C877340@lbl.gov> <1DD67C25-D646-4BD7-B44D-709D505CF267@lbl.gov> <87txggipud.fsf@mcs.anl.gov> <6958C2D4-F7CA-43F7-8E56-A3933E259FA5@lbl.gov> <83ED04D2-DE73-4A70-BDEE-0994317E0051@mcs.anl.gov> <71DB6D5F-F64E-4351-8B91-43D660E2E271@mcs.anl.gov> Message-ID: matsetsizes() is used practically everywhere - so I can't believe there is a bug there. [the fortran interface code autogenerated by bfort]. Do you see the error with the 'first' call to this routine - or some subsequent loop? Perhaps you can try using valgrind equivalent on windows to see if there is memory corruption somewhere else. http://stackoverflow.com/questions/413477/is-there-a-good-valgrind-substitute-for-windows suggests using http://code.google.com/p/drmemory/ Satish On Thu, 21 Nov 2013, Einar S?rheim wrote: > I am on the windows platform so unfortunatley no valgrind, however have > been able to track down the problem in the debugger and it seems like a > matrixpointer gets corrupted when using the fortran-c interface in petsc. > At some point in the calculation the matrix pointer in MatSetSize gets > corrupted when called from fortran, propably in this cast: > > void PETSC_STDCALL matsetsizes_(Mat A,PetscInt *m,PetscInt *n,PetscInt > *M,PetscInt *N, int *__ierr ){ > *__ierr = MatSetSizes( > (Mat)PetscToPointer((A) ),*m,*n,*M,*N); > } > when entering the MatSetSize in the debugger the A-matrix pointer is > corrupt and I get an access violation on code line: > PetscValidHeaderSpecific(A,MAT_CLASSID,1); > Any ideas what may be the cause of this? > Regards > Einar > > > 2013/10/24 Barry Smith > > > > > So apparently no memory leak. > > > > If you are still getting crashes you will need to run the case that > > crashes 1) with valgrind if you haven?t or 2) using > > -on_error_attach_debugger or -start_in_debugger to track down the cause of > > the crash. > > > > Barry > > > > > > > > On Oct 24, 2013, at 2:09 AM, Einar S?rheim > > wrote: > > > > > Well i I did a rerun without -malloc_log, this is what I got (I have > > also attached the log-sumary file): > > > > > > [0] PetscFinalize(): PetscFinalize() called > > > [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 > > -2080374784 max tags = 2147483647 > > > Summary of Memory Usage in PETSc > > > [0]Current space PetscMalloc()ed 45936, max space PetscMalloced() > > 1.13728e+009 > > > [0]OS cannot compute process memory > > > [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850688 > > -2080374784 > > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > > [0] PetscFOpen(): Opening file Log.0 > > > [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm > > -2080374784 > > > [0] Petsc_DelComm_Inner(): Removing reference to PETSc communicator > > embedded in a user MPI_Comm -2080374784 > > > [0] Petsc_DelComm_Outer(): User MPI_Comm 1140850688 is being freed after > > removing reference from inner PETSc comm to this outer comm > > > [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 > > > [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm > > -2080374784 > > > [0] Petsc_DelThreadComm(): Deleting thread communicator data in an > > MPI_Comm -2080374784 > > > [0] Petsc_DelCounter(): Deleting counter data in an MPI_Comm -2080374784 > > > > > > > > > > > > 2013/10/23 Barry Smith > > > > > > Don?t run with the -malloc_log Mark did not suggest using that option. > > > > > > > > > On Oct 23, 2013, at 8:48 AM, Einar S?rheim > > wrote: > > > > > > > I tried running with the following options : > > > > -info > > > > -malloc > > > > -malloc_log > > > > -malloc_dump > > > > -malloc_debug > > > > -memory_info > > > > -log_summary log_summary.txt > > > > -log > > > > but didn't get a lot wiser however in my case would it be more > > efficient to keep the ksp solver object, if so should I have one object eg. > > for the thermal problem and another one for the mechanical? > > > > The output from the end of the run: > > > > Total wall clock time actually solving: 856.231474 > > > > [0] PetscFinalize(): PetscFinalize() called > > > > [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 > > -2080374784 max tags = 2147483647 > > > > Summary of Memory Usage in PETSc > > > > [0]Current space PetscMalloc()ed 45936, max space PetscMalloced() > > 1.13728e+009 > > > > [0]OS cannot compute process memory > > > > [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850688 > > -2080374784 > > > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > > > [0] PetscFOpen(): Opening file Log.0 > > > > [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm > > -2080374784 > > > > [0] Petsc_DelComm_Inner(): Removing reference to PETSc communicator > > embedded in a user MPI_Comm -2080374784 > > > > [0] Petsc_DelComm_Outer(): User MPI_Comm 1140850688 is being freed > > after removing reference from inner PETSc comm to this outer comm > > > > [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 > > > > [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm > > -2080374784 > > > > [0] Petsc_DelThreadComm(): Deleting thread communicator data in an > > MPI_Comm -2080374784 > > > > [0] Petsc_DelCounter(): Deleting counter data in an MPI_Comm > > -2080374784 > > > > [0]PETSC ERROR: --------------------- Error Message > > ------------------------------------ > > > > [0]PETSC ERROR: Object is in wrong state! > > > > [0]PETSC ERROR: PetscMallocDumpLog() called without call to > > PetscMallocSetDumpLog() this is often due to > > > > setting the option -malloc_log AFTER > > PetscInitialize() with PetscOptionsInsert() or PetscOptionsInsertFile()! > > > > [0]PETSC ERROR: > > ------------------------------------------------------------------------ > > > > [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > > > > [0]PETSC ERROR: See docs/changes/index.html for recent updates. > > > > [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > > > > [0]PETSC ERROR: See docs/index.html for manual pages. > > > > [0]PETSC ERROR: > > ------------------------------------------------------------------------ > > > > [0]PETSC ERROR: D:\Programming\gits\Alsim\SharedModules - > > Copy\Source\Release\Alsim_nypetsc.exe on a arch-mswin-c-debug named CMP3 by > > admeinar Wed Oct 23 11:54:03 2013 > > > > [0]PETSC ERROR: Libraries linked from > > /cygdrive/d/Programming/petsc-3.4.3/arch-mswin-c-debug/lib > > > > [0]PETSC ERROR: Configure run at Thu Oct 17 08:23:18 2013 > > > > [0]PETSC ERROR: Configure options --with-debugging=1 --with-openmp=yes > > --with-pthread=no --with-cc="win32fe icl" --with-fc="win32fe ifort" > > --with-cxx="win32fe icl" --with-blas-lapack-dir="C:/Program Files > > (x86)/Intel/Composer XE 2013/mkl/lib/intel64" --download-superlu > > --with-sowing=0 --with-c2html=0 --with-mpi-dir="C:/Program Files > > (x86)/Intel/MPI/4.1.0.028/em64t" --useThreads=0 --useThreads=0 > > > > [0]PETSC ERROR: > > ------------------------------------------------------------------------ > > > > [0]PETSC ERROR: PetscMallocDumpLog() line 652 in > > src/sys/memory/D:\PROGRA~1\PETSC-~1.3\src\sys\memory\mtr.c > > > > [0]PETSC ERROR: PetscFinalize() line 1185 in > > src/sys/objects/D:\PROGRA~1\PETSC-~1.3\src\sys\objects\pinit.c > > > > > > > > > > > > 2013/10/19 Mark F. Adams > > > > You can try running a problem that finishes with > > > > > > > > -malloc > > > > -malloc_debug > > > > -malloc_dump > > > > > > > > to see if there are any memory issues. It looks like you are > > completely destroying the solver object each time and starting over. > > Valgrind would be very useful here because it looks like memory is getting > > corrupted if the same code runs many times. > > > > > > > > > > > > On Oct 19, 2013, at 3:39 PM, Einar S?rheim > > wrote: > > > > > > > >> yes, I call the given code many times, what i have seen so far is > > that changing e.g. the order of destruction statements changes the time of > > failure,I will try to debug to get more details > > > >> > > > >> > > > >> 2013/10/18 Mark F. Adams > > > >> I would guess there is some memory corruption or leak someplace. Try > > running with -on_error_attach_debugger, and try to get some more > > information. Do you just repeat a call to the code below many times? Is > > this deterministic, does it fail on the same solve each time? > > > >> > > > >> > > > >> On Oct 18, 2013, at 9:30 AM, Einar S?rheim > > wrote: > > > >> > > > >>> Uncommenting KSPSetFromOptions solved my intial problem, however I > > got a new problem that occurs after many timesteps, in our code we use > > petsc gamg as a solver for several problems(e.g. thermal and mechanical) at > > each time step, I have a suspicion it might have something to do with > > cleanup after solving, the rror appears after many calls to the petsc gamg > > solver interface, the error message is the following: > > > >>> > > > >>> [0]PETSC ERROR: > > ------------------------------------------------------------------------ > > > >>> [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation > > Violation, probably memory access out of range > > > >>> [0]PETSC ERROR: Try option -start_in_debugger or > > -on_error_attach_debugger > > > >>> [0]PETSC ERROR: or see > > http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSCERROR: or try > > http://valgrind.org on GNU/linux and Apple Mac OS X to find memory > > corruption errors > > > >>> [0]PETSC ERROR: likely location of problem given in stack below > > > >>> [0]PETSC ERROR: --------------------- Stack Frames > > ------------------------------------ > > > >>> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not > > available, > > > >>> [0]PETSC ERROR: INSTEAD the line number of the start of the > > function > > > >>> [0]PETSC ERROR: is given. > > > >>> [0]PETSC ERROR: [0] MatSetNearNullSpace line 7580 > > src/mat/interface/D:\PROGRA~1\PETSC-~1.3\src\mat\INTERF~1\matrix.c > > > >>> [0]PETSC ERROR: --------------------- Error Message > > ------------------------------------ > > > >>> [0]PETSC ERROR: Signal received! > > > >>> [0]PETSC ERROR: > > ------------------------------------------------------------------------ > > > >>> [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > > > >>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. > > > >>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > > > >>> [0]PETSC ERROR: See docs/index.html for manual pages. > > > >>> [0]PETSC ERROR: > > ------------------------------------------------------------------------ > > > >>> [0]PETSC ERROR: D:\Programming\gits\Alsim\SharedModules - > > Copy\Source\Release\Alsim_nypetsc.exe on a arch-mswin-c-debug named CMP3 by > > admeinar Fri Oct 18 12:26:04 2013 > > > >>> [0]PETSC ERROR: Libraries linked from > > /cygdrive/d/Programming/petsc-3.4.3/arch-mswin-c-debug/lib > > > >>> [0]PETSC ERROR: Configure run at Thu Oct 17 08:23:18 2013 > > > >>> [0]PETSC ERROR: Configure options --with-debugging=1 > > --with-openmp=yes --with-pthread=no --with-cc="win32fe icl" > > --with-fc="win32fe ifort" --with-cxx="win32fe icl" > > --with-blas-lapack-dir="C:/Program Files (x86)/Intel/Composer XE > > 2013/mkl/lib/intel64" --download-superlu --with-sowing=0 --with-c2html=0 > > --with-mpi-dir="C:/Program Files (x86)/Intel/MPI/4.1.0.028/em64t" > > --useThreads=0 --useThreads=0 > > > >>> > > > >>> The source code for the petsc interface: > > > >>> > > > >>> subroutine > > petscsolver(i_nodes,r_elements_,i_rowstart_,i_columnindex_,r_x_,r_b_,i_Nnz,i_Precond, > > i_Solver > > > >>> $ ,coords_,i_nodedof) > > > >>> #include > > > >>> #include > > > >>> #include > > > >>> #include > > > >>> #include > > > >>> #include > > > >>> > > > >>> integer i_nodes, i_Nnz, i_Precond, i_solver, i_nodedof > > > >>> integer i_rowstart_(*),i_columnindex_(*) > > > >>> double precision r_elements_(*),r_x_(*), r_b_(*),coords_(*) > > > >>> integer, allocatable :: i_indices_(:) > > > >>> ! Variables: > > > >>> ! > > > >>> ! A - matrix that defines linear system > > > >>> ! ksp - KSP context > > > >>> ! ksp - KSP context > > > >>> ! x, b, u - approx solution, RHS, exact solution vectors > > > >>> ! > > > >>> ! implicit none > > > >>> Vec x,u,b,vec_coords > > > >>> PC pc > > > >>> Mat A,F,Matnull > > > >>> KSP ksp > > > >>> PetscInt i,j,II,JJ,m,n,i1,imaxiter > > > >>> PetscInt Istart,Iend > > > >>> PetscInt nsteps,one, neg_one > > > >>> PetscErrorCode ierr > > > >>> PetscBool flg > > > >>> PetscScalar v,x_array_(1), norm, tol, t0,t1 > > > >>> PetscOffset i_x > > > >>> PetscViewer viewer > > > >>> > > > >>> > > > >>> call PetscTime(t0,ierr) > > > >>> write(*,*) "Enter petsc" > > > >>> > > > >>> n = 1 > > > >>> nsteps = 1 > > > >>> one = 1 > > > >>> neg_one = 1 > > > >>> i1 = 1 > > > >>> imaxiter = 1200 > > > >>> > > > >>> > > > >>> call MatCreate(PETSC_COMM_WORLD,A,ierr) > > > >>> call > > MatSetSizes(A,PETSC_DECIDE,PETSC_DECIDE,i_nodes,i_nodes,ierr) > > > >>> call MatSetType(A, MATSEQAIJ,ierr) > > > >>> call MatSetUp(A,ierr) > > > >>> call MatSeqAIJSetPreallocation(A,100,PETSC_NULL_INTEGER,ierr) > > > >>> ! The matrix is partitioned by contiguous chunks of rows across the > > > >>> ! processors. Determine which rows of the matrix are locally owned. > > > >>> > > > >>> call MatGetOwnershipRange(A,Istart,Iend,ierr) > > > >>> > > > >>> > > > >>> write(*,*) "Start of matrix fill", i_nodes, Istart, Iend > > > >>> do i=1,i_nodes > > > >>> ! write(*,*) "Start of loop",i !!, i_indices_(i+1) > > > >>> ! i_indices_(i+1) = i > > > >>> ! write(*,*) "Start of loop2" > > > >>> do j=i_rowstart_(i),i_rowstart_(i+1)-1 > > > >>> ! write(*,*) "Start of loop 3" > > > >>> v = r_elements_(j) > > > >>> ! write(*,*) i1,i,i1,j-1,v > > > >>> call > > MatSetValues(A,i1,i-1,i1,i_columnindex_(j)-1,v,INSERT_VALUES,ierr) > > > >>> if (ierr.gt.0) stop > > > >>> end do > > > >>> end do > > > >>> ! write(*,*) "End of matrix fill" > > > >>> > > > >>> ! Assemble matrix, using the 2-step process: > > > >>> ! MatAssemblyBegin(), MatAssemblyEnd() > > > >>> ! Computations can be done while messages are in transition > > > >>> ! by placing code between these two statements. > > > >>> > > > >>> call MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY,ierr) > > > >>> call MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY,ierr) > > > >>> > > > >>> ! Create parallel vectors. > > > >>> ! - When using VecCreate(), the parallel partitioning of the vector > > > >>> ! is determined by PETSc at runtime. > > > >>> ! - Note: We form 1 vector from scratch and then duplicate as > > needed. > > > >>> ! write(*,*) "Start of vector fill" > > > >>> call VecCreate(PETSC_COMM_WORLD,u,ierr) > > > >>> call VecSetType(u, VECSEQ,ierr) > > > >>> call VecSetSizes(u,PETSC_DECIDE,i_nodes,ierr) > > > >>> ! call VecSetFromOptions(u,ierr) > > > >>> call VecDuplicate(u,b,ierr) > > > >>> call VecDuplicate(b,x,ierr) > > > >>> do i=1,i_nodes > > > >>> call VecSetValues(x,i1,i-1,r_x_(i),INSERT_VALUES,ierr) > > > >>> call VecSetValues(b,i1,i-1,r_b_(i),INSERT_VALUES,ierr) > > > >>> enddo > > > >>> ! Assemble vector > > > >>> > > > >>> call VecAssemblyBegin(x,ierr) > > > >>> call VecAssemblyEnd(x,ierr) > > > >>> call VecAssemblyBegin(b,ierr) > > > >>> call VecAssemblyEnd(b,ierr) > > > >>> call PetscTime(t1,ierr) > > > >>> ! write(*,*) "End of vec fill time spent :", t1-t0 > > > >>> ! Create linear solver context > > > >>> > > > >>> call KSPCreate(PETSC_COMM_WORLD,ksp,ierr) > > > >>> > > > >>> > > > >>> ! Set operators. Here the matrix that defines the linear system > > > >>> ! also serves as the preconditioning matrix. > > > >>> > > > >>> call KSPSetOperators(ksp,A,A,DIFFERENT_NONZERO_PATTERN,ierr) > > > >>> > > > >>> > > > >>> call KSPGetPC(ksp,pc,ierr) > > > >>> tol = 1.e-20 > > > >>> call > > KSPSetTolerances(ksp,tol,PETSC_DEFAULT_DOUBLE_PRECISION,PETSC_DEFAULT_DOUBLE_PRECISION,imaxiter,ierr) > > > >>> select case (i_Precond) > > > >>> case ( 1 ) > > > >>> call PCSetType(pc,PCJACOBI,ierr) > > > >>> case ( 2 ) > > > >>> call PCSetType(pc,PCILU,ierr) > > > >>> call PCFactorSetMatSolverPackage(pc,MATSOLVERSUPERLU,ierr); > > > >>> call PCFactorSetUpMatSolverPackage(pc,ierr); > > > >>> call PCFactorGetMatrix(pc,F); > > > >>> call MatSuperluSetILUDropTol(F,1.e-4); > > > >>> case ( 3 ) > > > >>> call PCSetType(pc,PCGAMG,ierr) > > > >>> ! call PCGAMGInitializePackage() > > > >>> ! call PCCreate_GAMG(pc,ierr) > > > >>> case DEFAULT > > > >>> call PCSetType(pc,PCJACOBI,ierr) > > > >>> end select > > > >>> > > > >>> ! call PCSetType(pc,PCJACOBI,ierr) > > > >>> select case (i_solver) > > > >>> case ( 0 ) > > > >>> call KSPSetType(ksp,KSPCG,ierr) > > > >>> case DEFAULT > > > >>> call KSPSetType(ksp,KSPCG,ierr) > > > >>> end select > > > >>> call KSPCGSetType(ksp,KSP_CG_SYMMETRIC,ierr) > > > >>> if (i_nodedof==3) then > > > >>> write(*,*) "Set coord, number of nodes is:", > > i_nodes/i_nodedof > > > >>> call > > VecCreateSeqWithArray(MPI_COMM_SELF,3,i_nodes,coords_,vec_coords,ierr) > > > >>> call MatNullSpaceCreateRigidBody(vec_coords,Matnull,ierr) > > > >>> call MatSetNearNullSpace(A,Matnull,ierr) > > > >>> call MatNullSpaceDestroy(Matnull,ierr) > > > >>> > > > >>> write(*,*) "Finish setting null space ierr =", ierr > > > >>> ! call PCSetCoordinates( pc, 3,i_nodes/i_nodedof, coords_, > > ierr ) > > > >>> > > > >>> end if > > > >>> ! call KSPMonitorSet(ksp, KSPMonitorDefault, PETSC_NULL, > > PETSC_NULL,ierr) > > > >>> ! Set runtime options, e.g., > > > >>> ! -ksp_type -pc_type -ksp_monitor -ksp_rtol > > > >>> ! These options will override those specified above as long as > > > >>> ! KSPSetFromOptions() is called _after_ any other customization > > > >>> ! routines. > > > >>> > > > >>> call KSPSetFromOptions(ksp,ierr) > > > >>> call KSPSetInitialGuessNonzero(ksp,PETSC_TRUE,ierr) > > > >>> call KSPSetUp(ksp,ierr) > > > >>> call PetscTime(t0,ierr) > > > >>> write(*,*) "Time setup", t0-t1 > > > >>> ! - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > > - - > > > >>> ! Solve the linear system > > > >>> ! - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > > - - > > > >>> > > > >>> call KSPSolve(ksp,b,x,ierr) > > > >>> call PetscTime(t1,ierr) > > > >>> write(*,*) "end solve, time used:",t1-t0 > > > >>> call KSPGetSolution(ksp,x,ierr) > > > >>> call VecGetArray(x,x_array_,i_x,ierr) > > > >>> do i=1,i_nodes > > > >>> r_x_(i) = x_array_(i_x + i) > > > >>> enddo > > > >>> call VecRestoreArray(x,x_array_,i_x,ierr) > > > >>> > > > >>> > > > >>> ! Free work space. All PETSc objects should be destroyed when they > > > >>> ! are no longer needed. > > > >>> ! call PCDestroy(pc,ierr) > > > >>> > > > >>> call VecDestroy(u,ierr) > > > >>> call VecDestroy(x,ierr) > > > >>> call VecDestroy(b,ierr) > > > >>> call MatDestroy(A,ierr) > > > >>> if (i_nodedof==3) then > > > >>> call VecDestroy(vec_coords,ierr) > > > >>> call MatDestroy(Matnull,ierr) > > > >>> endif > > > >>> call KSPDestroy(ksp,ierr) > > > >>> call PetscTime(t0,ierr) > > > >>> write(*,*) "time celan up :", t0-t1 > > > >>> > > > >>> end > > > >>> > > > >>> -- > > > >>> Einar S?rheim > > > >>> > > > >> > > > >> > > > >> > > > >> > > > >> -- > > > >> Einar S?rheim > > > > > > > > > > > > > > > > > > > > -- > > > > Einar S?rheim > > > > > > > > > > > > > > > -- > > > Einar S?rheim > > > > > > > > > > From john.mousel at gmail.com Thu Nov 21 13:39:39 2013 From: john.mousel at gmail.com (John Mousel) Date: Thu, 21 Nov 2013 13:39:39 -0600 Subject: [petsc-users] GAMG processor reduction Message-ID: I'm solving a Poisson equation with BiCGStab/GAMG. I'm looking at KSPView and I see that the problem is not reducing the number of processors on the coarser grids. My coarse problem is 723x723, and it looks like it's still being solved on 192 processes. This doesn't seem right, and I'm believe this used to trigger processor reduction. I'm running with -pres_ksp_type bcgsl -pres_pc_type gamg -pres_pc_gamg_threshold 0.05 -pres_mg_levels_ksp_type richardson -pres_mg_levels_pc_type sor -pres_mg_coarse_ksp_type richardson -pres_mg_coarse_pc_type sor -pres_mg_coarse_pc_sor_its 5 -pres_pc_gamg_type agg -pres_pc_gamg_agg_nsmooths 1 -pres_pc_gamg_sym_graph true -pres_ksp_monitor -pres_ksp_view -pres_gamg_process_eq_limit 5000 The KSPView output is: KSP Object:(pres_) 192 MPI processes type: bcgsl BCGSL: Ell = 2 BCGSL: Delta = 0 maximum iterations=1000, initial guess is zero tolerances: relative=1e-06, absolute=1e-50, divergence=10000 left preconditioning has attached null space using PRECONDITIONED norm type for convergence test PC Object:(pres_) 192 MPI processes type: gamg MG: type is MULTIPLICATIVE, levels=4 cycles=v Cycles per PCApply=1 Using Galerkin computed coarse grid matrices Coarse grid solver -- level ------------------------------- KSP Object: (pres_mg_coarse_) 192 MPI processes type: preonly maximum iterations=1, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using NONE norm type for convergence test PC Object: (pres_mg_coarse_) 192 MPI processes type: sor SOR: type = local_symmetric, iterations = 5, local iterations = 1, omega = 1 linear system matrix = precond matrix: Mat Object: 192 MPI processes type: mpiaij rows=723, cols=723 total: nonzeros=519221, allocated nonzeros=519221 total number of mallocs used during MatSetValues calls =0 using I-node (on process 0) routines: found 499 nodes, limit used is 5 Down solver (pre-smoother) on level 1 ------------------------------- KSP Object: (pres_mg_levels_1_) 192 MPI processes type: richardson Richardson: damping factor=1 maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (pres_mg_levels_1_) 192 MPI processes type: sor SOR: type = local_symmetric, iterations = 1, local iterations = 1, omega = 1 linear system matrix = precond matrix: Mat Object: 192 MPI processes type: mpiaij rows=15231, cols=15231 total: nonzeros=4604432, allocated nonzeros=4604432 total number of mallocs used during MatSetValues calls =0 using I-node (on process 0) routines: found 33 nodes, limit used is 5 Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 2 ------------------------------- KSP Object: (pres_mg_levels_2_) 192 MPI processes type: richardson Richardson: damping factor=1 maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (pres_mg_levels_2_) 192 MPI processes type: sor SOR: type = local_symmetric, iterations = 1, local iterations = 1, omega = 1 linear system matrix = precond matrix: Mat Object: 192 MPI processes type: mpiaij rows=242341, cols=242341 total: nonzeros=18277529, allocated nonzeros=18277529 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 3 ------------------------------- KSP Object: (pres_mg_levels_3_) 192 MPI processes type: richardson Richardson: damping factor=1 maximum iterations=2 tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using nonzero initial guess using NONE norm type for convergence test PC Object: (pres_mg_levels_3_) 192 MPI processes type: sor SOR: type = local_symmetric, iterations = 1, local iterations = 1, omega = 1 linear system matrix = precond matrix: Mat Object: 192 MPI processes type: mpiaij rows=3285079, cols=3285079 total: nonzeros=48995036, allocated nonzeros=53586621 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Up solver (post-smoother) same as down solver (pre-smoother) linear system matrix = precond matrix: Mat Object: 192 MPI processes type: mpiaij rows=3285079, cols=3285079 total: nonzeros=48995036, allocated nonzeros=53586621 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines John -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Thu Nov 21 15:19:08 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 21 Nov 2013 14:19:08 -0700 Subject: [petsc-users] GAMG processor reduction In-Reply-To: References: Message-ID: <871u29zc4z.fsf@jedbrown.org> John Mousel writes: > I'm solving a Poisson equation with BiCGStab/GAMG. I'm looking at KSPView > and I see that the problem is not reducing the number of processors on the > coarser grids. My coarse problem is 723x723, and it looks like it's still > being solved on 192 processes. This doesn't seem right, and I'm believe > this used to trigger processor reduction. It still uses the big communicator, but all but rank 0 should be empty on the coarsest grid. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From john.mousel at gmail.com Thu Nov 21 15:22:07 2013 From: john.mousel at gmail.com (John Mousel) Date: Thu, 21 Nov 2013 15:22:07 -0600 Subject: [petsc-users] GAMG processor reduction In-Reply-To: <871u29zc4z.fsf@jedbrown.org> References: <871u29zc4z.fsf@jedbrown.org> Message-ID: Thanks Jed. How does this represent itself in the KSPView output? On Thu, Nov 21, 2013 at 3:19 PM, Jed Brown wrote: > John Mousel writes: > > > I'm solving a Poisson equation with BiCGStab/GAMG. I'm looking at KSPView > > and I see that the problem is not reducing the number of processors on > the > > coarser grids. My coarse problem is 723x723, and it looks like it's still > > being solved on 192 processes. This doesn't seem right, and I'm believe > > this used to trigger processor reduction. > > It still uses the big communicator, but all but rank 0 should be empty > on the coarsest grid. > -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Thu Nov 21 15:34:07 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 21 Nov 2013 14:34:07 -0700 Subject: [petsc-users] GAMG processor reduction In-Reply-To: References: <871u29zc4z.fsf@jedbrown.org> Message-ID: <87y54hxwvk.fsf@jedbrown.org> John Mousel writes: > Thanks Jed. How does this represent itself in the KSPView output? I'm afraid it's not there, though you can extract the ownership ranges From code. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From dave.mayhem23 at gmail.com Thu Nov 21 15:43:10 2013 From: dave.mayhem23 at gmail.com (Dave May) Date: Thu, 21 Nov 2013 22:43:10 +0100 Subject: [petsc-users] GAMG processor reduction In-Reply-To: <87y54hxwvk.fsf@jedbrown.org> References: <871u29zc4z.fsf@jedbrown.org> <87y54hxwvk.fsf@jedbrown.org> Message-ID: Is using the "big communicator" really the right way to go? What happens when I call VecNorm() when the local size on most ranks =0.. the global reduction still has to be performed and all ranks in the original communicator associated with the fine get participate. I thought the primary advantage/reason to use less ranks with small distributed systems was to avoid seeing the network latency when there is little computational work. I don't see how using the big communicator avoids this. Am I missing something? Cheers, Dave On Thursday, 21 November 2013, Jed Brown wrote: > John Mousel > writes: > > > Thanks Jed. How does this represent itself in the KSPView output? > > I'm afraid it's not there, though you can extract the ownership ranges > From code. > -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Thu Nov 21 16:06:21 2013 From: knepley at gmail.com (Matthew Knepley) Date: Thu, 21 Nov 2013 16:06:21 -0600 Subject: [petsc-users] GAMG processor reduction In-Reply-To: References: <871u29zc4z.fsf@jedbrown.org> <87y54hxwvk.fsf@jedbrown.org> Message-ID: On Thu, Nov 21, 2013 at 3:43 PM, Dave May wrote: > Is using the "big communicator" really the right way to go? What happens > when I call VecNorm() when the local size on most ranks =0.. the global > reduction still has to be performed and all ranks in the original > communicator associated with the fine get participate. > > I thought the primary advantage/reason to use less ranks with small > distributed systems was to avoid seeing the network latency when there is > little computational work. I don't see how using the big communicator > avoids this. > Its not just this. You do not want to get to the point where you have 1 or < 1 point per process, so you rebalance to put a reasonable number of unknowns per process and leave others empty. You could create a subcomm with only the nonzero procs to use in the solve. Not sure if this is worth it. Matt > Am I missing something? > > Cheers, > Dave > > On Thursday, 21 November 2013, Jed Brown wrote: > >> John Mousel writes: >> >> > Thanks Jed. How does this represent itself in the KSPView output? >> >> I'm afraid it's not there, though you can extract the ownership ranges >> From code. >> > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From dave.mayhem23 at gmail.com Thu Nov 21 16:13:16 2013 From: dave.mayhem23 at gmail.com (Dave May) Date: Thu, 21 Nov 2013 23:13:16 +0100 Subject: [petsc-users] GAMG processor reduction In-Reply-To: References: <871u29zc4z.fsf@jedbrown.org> <87y54hxwvk.fsf@jedbrown.org> Message-ID: I argue it does matter as I've seen runs on 32k cores where a huge amount of time is spent in those global reductions. I can provide an implementation which uses a sub comm (PCSemiRedundant) if someone thinks doing reductions on less cores is beneficial. I'd like to hear from other users running mg like algs with large core counts where they have run into this same issue. Cheers Dave On Thursday, 21 November 2013, Matthew Knepley wrote: > On Thu, Nov 21, 2013 at 3:43 PM, Dave May > > wrote: > >> Is using the "big communicator" really the right way to go? What happens >> when I call VecNorm() when the local size on most ranks =0.. the global >> reduction still has to be performed and all ranks in the original >> communicator associated with the fine get participate. >> >> I thought the primary advantage/reason to use less ranks with small >> distributed systems was to avoid seeing the network latency when there is >> little computational work. I don't see how using the big communicator >> avoids this. >> > > Its not just this. You do not want to get to the point where you have 1 or > < 1 point per process, so > you rebalance to put a reasonable number of unknowns per process and leave > others empty. You > could create a subcomm with only the nonzero procs to use in the solve. > Not sure if this is worth it. > > Matt > > >> Am I missing something? >> >> Cheers, >> Dave >> >> On Thursday, 21 November 2013, Jed Brown wrote: >> >>> John Mousel writes: >>> >>> > Thanks Jed. How does this represent itself in the KSPView output? >>> >>> I'm afraid it's not there, though you can extract the ownership ranges >>> From code. >>> >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From mfadams at lbl.gov Thu Nov 21 16:48:11 2013 From: mfadams at lbl.gov (Mark Adams) Date: Thu, 21 Nov 2013 17:48:11 -0500 Subject: [petsc-users] Problem with GAMG elasticity case:'createdefaultdata' not set(?) need to support NULL data! In-Reply-To: References: <8EE4826E-4436-4FE4-84D9-2EBF4C877340@lbl.gov> <1DD67C25-D646-4BD7-B44D-709D505CF267@lbl.gov> <87txggipud.fsf@mcs.anl.gov> <6958C2D4-F7CA-43F7-8E56-A3933E259FA5@lbl.gov> <83ED04D2-DE73-4A70-BDEE-0994317E0051@mcs.anl.gov> <71DB6D5F-F64E-4351-8B91-43D660E2E271@mcs.anl.gov> Message-ID: It does look like memory is getting corrupted and it just fails first in MatGetSizes. This happens after many calls the same solver routine that does not have any side effects as I recall. Does fail with a different PC like Jacobi? As I recall you checked for memory leaks. You mentioned that it failed at different times when you reordered the Destroy calls. Is this repeatable? (ie, does it fail on the same solve number every time with a given configuration. Could you try commenting out half the Destroys and see if that fixes it (if you don't run out of memory). Then try the other half of the Destroys and if you get a fix go for bisection search. Mark On Thu, Nov 21, 2013 at 12:17 PM, Satish Balay wrote: > matsetsizes() is used practically everywhere - so I can't believe > there is a bug there. [the fortran interface code autogenerated by > bfort]. > > Do you see the error with the 'first' call to this routine - or some > subsequent loop? > > Perhaps you can try using valgrind equivalent on windows to see if > there is memory corruption somewhere else. > > > http://stackoverflow.com/questions/413477/is-there-a-good-valgrind-substitute-for-windows > suggests using http://code.google.com/p/drmemory/ > > Satish > > On Thu, 21 Nov 2013, Einar S?rheim wrote: > > > I am on the windows platform so unfortunatley no valgrind, however have > > been able to track down the problem in the debugger and it seems like a > > matrixpointer gets corrupted when using the fortran-c interface in petsc. > > At some point in the calculation the matrix pointer in MatSetSize gets > > corrupted when called from fortran, propably in this cast: > > > > void PETSC_STDCALL matsetsizes_(Mat A,PetscInt *m,PetscInt *n,PetscInt > > *M,PetscInt *N, int *__ierr ){ > > *__ierr = MatSetSizes( > > (Mat)PetscToPointer((A) ),*m,*n,*M,*N); > > } > > when entering the MatSetSize in the debugger the A-matrix pointer is > > corrupt and I get an access violation on code line: > > PetscValidHeaderSpecific(A,MAT_CLASSID,1); > > Any ideas what may be the cause of this? > > Regards > > Einar > > > > > > 2013/10/24 Barry Smith > > > > > > > > So apparently no memory leak. > > > > > > If you are still getting crashes you will need to run the case that > > > crashes 1) with valgrind if you haven?t or 2) using > > > -on_error_attach_debugger or -start_in_debugger to track down the > cause of > > > the crash. > > > > > > Barry > > > > > > > > > > > > On Oct 24, 2013, at 2:09 AM, Einar S?rheim > > > wrote: > > > > > > > Well i I did a rerun without -malloc_log, this is what I got (I have > > > also attached the log-sumary file): > > > > > > > > [0] PetscFinalize(): PetscFinalize() called > > > > [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 > > > -2080374784 max tags = 2147483647 > > > > Summary of Memory Usage in PETSc > > > > [0]Current space PetscMalloc()ed 45936, max space PetscMalloced() > > > 1.13728e+009 > > > > [0]OS cannot compute process memory > > > > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 > > > -2080374784 > > > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > > > [0] PetscFOpen(): Opening file Log.0 > > > > [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm > > > -2080374784 > > > > [0] Petsc_DelComm_Inner(): Removing reference to PETSc communicator > > > embedded in a user MPI_Comm -2080374784 > > > > [0] Petsc_DelComm_Outer(): User MPI_Comm 1140850688 is being freed > after > > > removing reference from inner PETSc comm to this outer comm > > > > [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 > > > > [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm > > > -2080374784 > > > > [0] Petsc_DelThreadComm(): Deleting thread communicator data in an > > > MPI_Comm -2080374784 > > > > [0] Petsc_DelCounter(): Deleting counter data in an MPI_Comm > -2080374784 > > > > > > > > > > > > > > > > 2013/10/23 Barry Smith > > > > > > > > Don?t run with the -malloc_log Mark did not suggest using that > option. > > > > > > > > > > > > On Oct 23, 2013, at 8:48 AM, Einar S?rheim > > > wrote: > > > > > > > > > I tried running with the following options : > > > > > -info > > > > > -malloc > > > > > -malloc_log > > > > > -malloc_dump > > > > > -malloc_debug > > > > > -memory_info > > > > > -log_summary log_summary.txt > > > > > -log > > > > > but didn't get a lot wiser however in my case would it be more > > > efficient to keep the ksp solver object, if so should I have one > object eg. > > > for the thermal problem and another one for the mechanical? > > > > > The output from the end of the run: > > > > > Total wall clock time actually solving: 856.231474 > > > > > [0] PetscFinalize(): PetscFinalize() called > > > > > [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 > > > -2080374784 max tags = 2147483647 > > > > > Summary of Memory Usage in PETSc > > > > > [0]Current space PetscMalloc()ed 45936, max space PetscMalloced() > > > 1.13728e+009 > > > > > [0]OS cannot compute process memory > > > > > [0] PetscCommDuplicate(): Using internal PETSc communicator > 1140850688 > > > -2080374784 > > > > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > > > > [0] PetscGetHostName(): Rejecting domainname, likely is NIS CMP3. > > > > > [0] PetscFOpen(): Opening file Log.0 > > > > > [0] Petsc_DelViewer(): Removing viewer data attribute in an > MPI_Comm > > > -2080374784 > > > > > [0] Petsc_DelComm_Inner(): Removing reference to PETSc communicator > > > embedded in a user MPI_Comm -2080374784 > > > > > [0] Petsc_DelComm_Outer(): User MPI_Comm 1140850688 is being freed > > > after removing reference from inner PETSc comm to this outer comm > > > > > [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 > > > > > [0] Petsc_DelViewer(): Removing viewer data attribute in an > MPI_Comm > > > -2080374784 > > > > > [0] Petsc_DelThreadComm(): Deleting thread communicator data in an > > > MPI_Comm -2080374784 > > > > > [0] Petsc_DelCounter(): Deleting counter data in an MPI_Comm > > > -2080374784 > > > > > [0]PETSC ERROR: --------------------- Error Message > > > ------------------------------------ > > > > > [0]PETSC ERROR: Object is in wrong state! > > > > > [0]PETSC ERROR: PetscMallocDumpLog() called without call to > > > PetscMallocSetDumpLog() this is often due to > > > > > setting the option -malloc_log AFTER > > > PetscInitialize() with PetscOptionsInsert() or > PetscOptionsInsertFile()! > > > > > [0]PETSC ERROR: > > > > ------------------------------------------------------------------------ > > > > > [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > > > > > [0]PETSC ERROR: See docs/changes/index.html for recent updates. > > > > > [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > > > > > [0]PETSC ERROR: See docs/index.html for manual pages. > > > > > [0]PETSC ERROR: > > > > ------------------------------------------------------------------------ > > > > > [0]PETSC ERROR: D:\Programming\gits\Alsim\SharedModules - > > > Copy\Source\Release\Alsim_nypetsc.exe on a arch-mswin-c-debug named > CMP3 by > > > admeinar Wed Oct 23 11:54:03 2013 > > > > > [0]PETSC ERROR: Libraries linked from > > > /cygdrive/d/Programming/petsc-3.4.3/arch-mswin-c-debug/lib > > > > > [0]PETSC ERROR: Configure run at Thu Oct 17 08:23:18 2013 > > > > > [0]PETSC ERROR: Configure options --with-debugging=1 > --with-openmp=yes > > > --with-pthread=no --with-cc="win32fe icl" --with-fc="win32fe ifort" > > > --with-cxx="win32fe icl" --with-blas-lapack-dir="C:/Program Files > > > (x86)/Intel/Composer XE 2013/mkl/lib/intel64" --download-superlu > > > --with-sowing=0 --with-c2html=0 --with-mpi-dir="C:/Program Files > > > (x86)/Intel/MPI/4.1.0.028/em64t" --useThreads=0 --useThreads=0 > > > > > [0]PETSC ERROR: > > > > ------------------------------------------------------------------------ > > > > > [0]PETSC ERROR: PetscMallocDumpLog() line 652 in > > > src/sys/memory/D:\PROGRA~1\PETSC-~1.3\src\sys\memory\mtr.c > > > > > [0]PETSC ERROR: PetscFinalize() line 1185 in > > > src/sys/objects/D:\PROGRA~1\PETSC-~1.3\src\sys\objects\pinit.c > > > > > > > > > > > > > > > 2013/10/19 Mark F. Adams > > > > > You can try running a problem that finishes with > > > > > > > > > > -malloc > > > > > -malloc_debug > > > > > -malloc_dump > > > > > > > > > > to see if there are any memory issues. It looks like you are > > > completely destroying the solver object each time and starting over. > > > Valgrind would be very useful here because it looks like memory is > getting > > > corrupted if the same code runs many times. > > > > > > > > > > > > > > > On Oct 19, 2013, at 3:39 PM, Einar S?rheim < > einar.sorheim at gmail.com> > > > wrote: > > > > > > > > > >> yes, I call the given code many times, what i have seen so far is > > > that changing e.g. the order of destruction statements changes the > time of > > > failure,I will try to debug to get more details > > > > >> > > > > >> > > > > >> 2013/10/18 Mark F. Adams > > > > >> I would guess there is some memory corruption or leak someplace. > Try > > > running with -on_error_attach_debugger, and try to get some more > > > information. Do you just repeat a call to the code below many times? > Is > > > this deterministic, does it fail on the same solve each time? > > > > >> > > > > >> > > > > >> On Oct 18, 2013, at 9:30 AM, Einar S?rheim < > einar.sorheim at gmail.com> > > > wrote: > > > > >> > > > > >>> Uncommenting KSPSetFromOptions solved my intial problem, > however I > > > got a new problem that occurs after many timesteps, in our code we use > > > petsc gamg as a solver for several problems(e.g. thermal and > mechanical) at > > > each time step, I have a suspicion it might have something to do with > > > cleanup after solving, the rror appears after many calls to the petsc > gamg > > > solver interface, the error message is the following: > > > > >>> > > > > >>> [0]PETSC ERROR: > > > > ------------------------------------------------------------------------ > > > > >>> [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation > > > Violation, probably memory access out of range > > > > >>> [0]PETSC ERROR: Try option -start_in_debugger or > > > -on_error_attach_debugger > > > > >>> [0]PETSC ERROR: or see > > > > http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[0]PETSCERROR: > or try > > > http://valgrind.org on GNU/linux and Apple Mac OS X to find memory > > > corruption errors > > > > >>> [0]PETSC ERROR: likely location of problem given in stack below > > > > >>> [0]PETSC ERROR: --------------------- Stack Frames > > > ------------------------------------ > > > > >>> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not > > > available, > > > > >>> [0]PETSC ERROR: INSTEAD the line number of the start of the > > > function > > > > >>> [0]PETSC ERROR: is given. > > > > >>> [0]PETSC ERROR: [0] MatSetNearNullSpace line 7580 > > > src/mat/interface/D:\PROGRA~1\PETSC-~1.3\src\mat\INTERF~1\matrix.c > > > > >>> [0]PETSC ERROR: --------------------- Error Message > > > ------------------------------------ > > > > >>> [0]PETSC ERROR: Signal received! > > > > >>> [0]PETSC ERROR: > > > > ------------------------------------------------------------------------ > > > > >>> [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > > > > >>> [0]PETSC ERROR: See docs/changes/index.html for recent updates. > > > > >>> [0]PETSC ERROR: See docs/faq.html for hints about trouble > shooting. > > > > >>> [0]PETSC ERROR: See docs/index.html for manual pages. > > > > >>> [0]PETSC ERROR: > > > > ------------------------------------------------------------------------ > > > > >>> [0]PETSC ERROR: D:\Programming\gits\Alsim\SharedModules - > > > Copy\Source\Release\Alsim_nypetsc.exe on a arch-mswin-c-debug named > CMP3 by > > > admeinar Fri Oct 18 12:26:04 2013 > > > > >>> [0]PETSC ERROR: Libraries linked from > > > /cygdrive/d/Programming/petsc-3.4.3/arch-mswin-c-debug/lib > > > > >>> [0]PETSC ERROR: Configure run at Thu Oct 17 08:23:18 2013 > > > > >>> [0]PETSC ERROR: Configure options --with-debugging=1 > > > --with-openmp=yes --with-pthread=no --with-cc="win32fe icl" > > > --with-fc="win32fe ifort" --with-cxx="win32fe icl" > > > --with-blas-lapack-dir="C:/Program Files (x86)/Intel/Composer XE > > > 2013/mkl/lib/intel64" --download-superlu --with-sowing=0 > --with-c2html=0 > > > --with-mpi-dir="C:/Program Files (x86)/Intel/MPI/4.1.0.028/em64t" > > > --useThreads=0 --useThreads=0 > > > > >>> > > > > >>> The source code for the petsc interface: > > > > >>> > > > > >>> subroutine > > > > petscsolver(i_nodes,r_elements_,i_rowstart_,i_columnindex_,r_x_,r_b_,i_Nnz,i_Precond, > > > i_Solver > > > > >>> $ ,coords_,i_nodedof) > > > > >>> #include > > > > >>> #include > > > > >>> #include > > > > >>> #include > > > > >>> #include > > > > >>> #include > > > > >>> > > > > >>> integer i_nodes, i_Nnz, i_Precond, i_solver, i_nodedof > > > > >>> integer i_rowstart_(*),i_columnindex_(*) > > > > >>> double precision r_elements_(*),r_x_(*), r_b_(*),coords_(*) > > > > >>> integer, allocatable :: i_indices_(:) > > > > >>> ! Variables: > > > > >>> ! > > > > >>> ! A - matrix that defines linear system > > > > >>> ! ksp - KSP context > > > > >>> ! ksp - KSP context > > > > >>> ! x, b, u - approx solution, RHS, exact solution vectors > > > > >>> ! > > > > >>> ! implicit none > > > > >>> Vec x,u,b,vec_coords > > > > >>> PC pc > > > > >>> Mat A,F,Matnull > > > > >>> KSP ksp > > > > >>> PetscInt i,j,II,JJ,m,n,i1,imaxiter > > > > >>> PetscInt Istart,Iend > > > > >>> PetscInt nsteps,one, neg_one > > > > >>> PetscErrorCode ierr > > > > >>> PetscBool flg > > > > >>> PetscScalar v,x_array_(1), norm, tol, t0,t1 > > > > >>> PetscOffset i_x > > > > >>> PetscViewer viewer > > > > >>> > > > > >>> > > > > >>> call PetscTime(t0,ierr) > > > > >>> write(*,*) "Enter petsc" > > > > >>> > > > > >>> n = 1 > > > > >>> nsteps = 1 > > > > >>> one = 1 > > > > >>> neg_one = 1 > > > > >>> i1 = 1 > > > > >>> imaxiter = 1200 > > > > >>> > > > > >>> > > > > >>> call MatCreate(PETSC_COMM_WORLD,A,ierr) > > > > >>> call > > > MatSetSizes(A,PETSC_DECIDE,PETSC_DECIDE,i_nodes,i_nodes,ierr) > > > > >>> call MatSetType(A, MATSEQAIJ,ierr) > > > > >>> call MatSetUp(A,ierr) > > > > >>> call > MatSeqAIJSetPreallocation(A,100,PETSC_NULL_INTEGER,ierr) > > > > >>> ! The matrix is partitioned by contiguous chunks of rows across > the > > > > >>> ! processors. Determine which rows of the matrix are locally > owned. > > > > >>> > > > > >>> call MatGetOwnershipRange(A,Istart,Iend,ierr) > > > > >>> > > > > >>> > > > > >>> write(*,*) "Start of matrix fill", i_nodes, Istart, Iend > > > > >>> do i=1,i_nodes > > > > >>> ! write(*,*) "Start of loop",i !!, i_indices_(i+1) > > > > >>> ! i_indices_(i+1) = i > > > > >>> ! write(*,*) "Start of loop2" > > > > >>> do j=i_rowstart_(i),i_rowstart_(i+1)-1 > > > > >>> ! write(*,*) "Start of loop 3" > > > > >>> v = r_elements_(j) > > > > >>> ! write(*,*) i1,i,i1,j-1,v > > > > >>> call > > > MatSetValues(A,i1,i-1,i1,i_columnindex_(j)-1,v,INSERT_VALUES,ierr) > > > > >>> if (ierr.gt.0) stop > > > > >>> end do > > > > >>> end do > > > > >>> ! write(*,*) "End of matrix fill" > > > > >>> > > > > >>> ! Assemble matrix, using the 2-step process: > > > > >>> ! MatAssemblyBegin(), MatAssemblyEnd() > > > > >>> ! Computations can be done while messages are in transition > > > > >>> ! by placing code between these two statements. > > > > >>> > > > > >>> call MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY,ierr) > > > > >>> call MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY,ierr) > > > > >>> > > > > >>> ! Create parallel vectors. > > > > >>> ! - When using VecCreate(), the parallel partitioning of the > vector > > > > >>> ! is determined by PETSc at runtime. > > > > >>> ! - Note: We form 1 vector from scratch and then duplicate as > > > needed. > > > > >>> ! write(*,*) "Start of vector fill" > > > > >>> call VecCreate(PETSC_COMM_WORLD,u,ierr) > > > > >>> call VecSetType(u, VECSEQ,ierr) > > > > >>> call VecSetSizes(u,PETSC_DECIDE,i_nodes,ierr) > > > > >>> ! call VecSetFromOptions(u,ierr) > > > > >>> call VecDuplicate(u,b,ierr) > > > > >>> call VecDuplicate(b,x,ierr) > > > > >>> do i=1,i_nodes > > > > >>> call VecSetValues(x,i1,i-1,r_x_(i),INSERT_VALUES,ierr) > > > > >>> call VecSetValues(b,i1,i-1,r_b_(i),INSERT_VALUES,ierr) > > > > >>> enddo > > > > >>> ! Assemble vector > > > > >>> > > > > >>> call VecAssemblyBegin(x,ierr) > > > > >>> call VecAssemblyEnd(x,ierr) > > > > >>> call VecAssemblyBegin(b,ierr) > > > > >>> call VecAssemblyEnd(b,ierr) > > > > >>> call PetscTime(t1,ierr) > > > > >>> ! write(*,*) "End of vec fill time spent :", t1-t0 > > > > >>> ! Create linear solver context > > > > >>> > > > > >>> call KSPCreate(PETSC_COMM_WORLD,ksp,ierr) > > > > >>> > > > > >>> > > > > >>> ! Set operators. Here the matrix that defines the linear system > > > > >>> ! also serves as the preconditioning matrix. > > > > >>> > > > > >>> call > KSPSetOperators(ksp,A,A,DIFFERENT_NONZERO_PATTERN,ierr) > > > > >>> > > > > >>> > > > > >>> call KSPGetPC(ksp,pc,ierr) > > > > >>> tol = 1.e-20 > > > > >>> call > > > > KSPSetTolerances(ksp,tol,PETSC_DEFAULT_DOUBLE_PRECISION,PETSC_DEFAULT_DOUBLE_PRECISION,imaxiter,ierr) > > > > >>> select case (i_Precond) > > > > >>> case ( 1 ) > > > > >>> call PCSetType(pc,PCJACOBI,ierr) > > > > >>> case ( 2 ) > > > > >>> call PCSetType(pc,PCILU,ierr) > > > > >>> call > PCFactorSetMatSolverPackage(pc,MATSOLVERSUPERLU,ierr); > > > > >>> call PCFactorSetUpMatSolverPackage(pc,ierr); > > > > >>> call PCFactorGetMatrix(pc,F); > > > > >>> call MatSuperluSetILUDropTol(F,1.e-4); > > > > >>> case ( 3 ) > > > > >>> call PCSetType(pc,PCGAMG,ierr) > > > > >>> ! call PCGAMGInitializePackage() > > > > >>> ! call PCCreate_GAMG(pc,ierr) > > > > >>> case DEFAULT > > > > >>> call PCSetType(pc,PCJACOBI,ierr) > > > > >>> end select > > > > >>> > > > > >>> ! call PCSetType(pc,PCJACOBI,ierr) > > > > >>> select case (i_solver) > > > > >>> case ( 0 ) > > > > >>> call KSPSetType(ksp,KSPCG,ierr) > > > > >>> case DEFAULT > > > > >>> call KSPSetType(ksp,KSPCG,ierr) > > > > >>> end select > > > > >>> call KSPCGSetType(ksp,KSP_CG_SYMMETRIC,ierr) > > > > >>> if (i_nodedof==3) then > > > > >>> write(*,*) "Set coord, number of nodes is:", > > > i_nodes/i_nodedof > > > > >>> call > > > VecCreateSeqWithArray(MPI_COMM_SELF,3,i_nodes,coords_,vec_coords,ierr) > > > > >>> call > MatNullSpaceCreateRigidBody(vec_coords,Matnull,ierr) > > > > >>> call MatSetNearNullSpace(A,Matnull,ierr) > > > > >>> call MatNullSpaceDestroy(Matnull,ierr) > > > > >>> > > > > >>> write(*,*) "Finish setting null space ierr =", ierr > > > > >>> ! call PCSetCoordinates( pc, 3,i_nodes/i_nodedof, > coords_, > > > ierr ) > > > > >>> > > > > >>> end if > > > > >>> ! call KSPMonitorSet(ksp, KSPMonitorDefault, PETSC_NULL, > > > PETSC_NULL,ierr) > > > > >>> ! Set runtime options, e.g., > > > > >>> ! -ksp_type -pc_type -ksp_monitor -ksp_rtol > > > > > >>> ! These options will override those specified above as long as > > > > >>> ! KSPSetFromOptions() is called _after_ any other customization > > > > >>> ! routines. > > > > >>> > > > > >>> call KSPSetFromOptions(ksp,ierr) > > > > >>> call KSPSetInitialGuessNonzero(ksp,PETSC_TRUE,ierr) > > > > >>> call KSPSetUp(ksp,ierr) > > > > >>> call PetscTime(t0,ierr) > > > > >>> write(*,*) "Time setup", t0-t1 > > > > >>> ! - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > - - > > > - - > > > > >>> ! Solve the linear system > > > > >>> ! - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > - - > > > - - > > > > >>> > > > > >>> call KSPSolve(ksp,b,x,ierr) > > > > >>> call PetscTime(t1,ierr) > > > > >>> write(*,*) "end solve, time used:",t1-t0 > > > > >>> call KSPGetSolution(ksp,x,ierr) > > > > >>> call VecGetArray(x,x_array_,i_x,ierr) > > > > >>> do i=1,i_nodes > > > > >>> r_x_(i) = x_array_(i_x + i) > > > > >>> enddo > > > > >>> call VecRestoreArray(x,x_array_,i_x,ierr) > > > > >>> > > > > >>> > > > > >>> ! Free work space. All PETSc objects should be destroyed when > they > > > > >>> ! are no longer needed. > > > > >>> ! call PCDestroy(pc,ierr) > > > > >>> > > > > >>> call VecDestroy(u,ierr) > > > > >>> call VecDestroy(x,ierr) > > > > >>> call VecDestroy(b,ierr) > > > > >>> call MatDestroy(A,ierr) > > > > >>> if (i_nodedof==3) then > > > > >>> call VecDestroy(vec_coords,ierr) > > > > >>> call MatDestroy(Matnull,ierr) > > > > >>> endif > > > > >>> call KSPDestroy(ksp,ierr) > > > > >>> call PetscTime(t0,ierr) > > > > >>> write(*,*) "time celan up :", t0-t1 > > > > >>> > > > > >>> end > > > > >>> > > > > >>> -- > > > > >>> Einar S?rheim > > > > >>> > > > > >> > > > > >> > > > > >> > > > > >> > > > > >> -- > > > > >> Einar S?rheim > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > Einar S?rheim > > > > > > > > > > > > > > > > > > > > -- > > > > Einar S?rheim > > > > > > > > > > > > > > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Thu Nov 21 17:10:19 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 21 Nov 2013 16:10:19 -0700 Subject: [petsc-users] GAMG processor reduction In-Reply-To: References: <871u29zc4z.fsf@jedbrown.org> <87y54hxwvk.fsf@jedbrown.org> Message-ID: <87txf5xsf8.fsf@jedbrown.org> Dave May writes: > I argue it does matter as I've seen runs on 32k cores where a huge amount > of time is spent in those global reductions. I can provide an > implementation which uses a sub comm (PCSemiRedundant) if someone thinks > doing reductions on less cores is beneficial. It doesn't matter much on Blue Gene, but is a big deal on older Crays. Aires seems to be in between. The default GAMG configuration doesn't do any reductions in the coarse grid, so the issue is moot. If an iterative coarse solver was used, I think we would be more motivated to put the coarse problem on a subcomm. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From dave.mayhem23 at gmail.com Thu Nov 21 17:17:12 2013 From: dave.mayhem23 at gmail.com (Dave May) Date: Fri, 22 Nov 2013 00:17:12 +0100 Subject: [petsc-users] GAMG processor reduction In-Reply-To: <87txf5xsf8.fsf@jedbrown.org> References: <871u29zc4z.fsf@jedbrown.org> <87y54hxwvk.fsf@jedbrown.org> <87txf5xsf8.fsf@jedbrown.org> Message-ID: Exactly! My tests were performed on Kraken (Cray XT5) and a Cray XE6 and I was using iterative solvers on the coarse grid. On 22 November 2013 00:10, Jed Brown wrote: > Dave May writes: > > > I argue it does matter as I've seen runs on 32k cores where a huge amount > > of time is spent in those global reductions. I can provide an > > implementation which uses a sub comm (PCSemiRedundant) if someone thinks > > doing reductions on less cores is beneficial. > > It doesn't matter much on Blue Gene, but is a big deal on older Crays. > Aires seems to be in between. The default GAMG configuration doesn't do > any reductions in the coarse grid, so the issue is moot. If an > iterative coarse solver was used, I think we would be more motivated to > put the coarse problem on a subcomm. > -------------- next part -------------- An HTML attachment was scrubbed... URL: From iwaddington at gmail.com Thu Nov 21 17:44:34 2013 From: iwaddington at gmail.com (iwaddington .) Date: Thu, 21 Nov 2013 21:44:34 -0200 Subject: [petsc-users] KSPSetComputeRHS and KSPSetComputeOperators Message-ID: HI everybody, I have a doubt concerning how the function KSPSetComputeRHS works, because according to its prototype one of its arguments is a pointer to a function that takes a vector object as argument, not a pointer towards a vector object, so how is it able to set the rhs if the passage of the vector is done by value and not by reference ? Another thing is the functioning of KSPSetComputeOperators. I want to set the linear system matrix and I want the preconditioning matrix to be equal, so can I just set the linar system matrix in the function that is argument of KSPSetComputeOperators, or do I need to make a matcopy in it ? -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Thu Nov 21 18:00:57 2013 From: knepley at gmail.com (Matthew Knepley) Date: Thu, 21 Nov 2013 18:00:57 -0600 Subject: [petsc-users] KSPSetComputeRHS and KSPSetComputeOperators In-Reply-To: References: Message-ID: On Thu, Nov 21, 2013 at 5:44 PM, iwaddington . wrote: > HI everybody, I have a doubt concerning how the function KSPSetComputeRHS > works, because according to its prototype one of its arguments is a pointer > to a function that takes a vector object as argument, not a pointer towards > a vector object, so how is it able to set the rhs if the passage of the > vector is done by value and not by reference ? > The values of the Vec can be set, rather than changing the object pointer itself. > Another thing is the functioning of KSPSetComputeOperators. I want to set > the linear system matrix and I want the preconditioning matrix to be equal, > so can I just set the linar system matrix in the function that is argument > of KSPSetComputeOperators, or do I need to make a matcopy in it ? > No, just give the same argument twice. Matt -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From mfadams at lbl.gov Thu Nov 21 18:10:39 2013 From: mfadams at lbl.gov (Mark Adams) Date: Thu, 21 Nov 2013 19:10:39 -0500 Subject: [petsc-users] GAMG processor reduction In-Reply-To: References: <871u29zc4z.fsf@jedbrown.org> <87y54hxwvk.fsf@jedbrown.org> <87txf5xsf8.fsf@jedbrown.org> Message-ID: '-pc_gamg_verbose 2' will print out stuff where you can see the number of active processors. We have not been super motivated to use subcoms because there are no norms taken on coarse grids, except in the eigen estimator for cheby, unless you use krylov in the solver, which is not usually useful. Jed: GAMG does reduce active processors by default. I'm pretty sure. I assume that if you are using an iterative solver on the coarse grid that you have a singular problem, and the AMG coarse grid space can represent it exactly, so you have a singular matrix one the coarse grid. Mark On Thu, Nov 21, 2013 at 6:17 PM, Dave May wrote: > Exactly! > My tests were performed on Kraken (Cray XT5) and a Cray XE6 and I was > using iterative solvers on the coarse grid. > > > > On 22 November 2013 00:10, Jed Brown wrote: > >> Dave May writes: >> >> > I argue it does matter as I've seen runs on 32k cores where a huge >> amount >> > of time is spent in those global reductions. I can provide an >> > implementation which uses a sub comm (PCSemiRedundant) if someone thinks >> > doing reductions on less cores is beneficial. >> >> It doesn't matter much on Blue Gene, but is a big deal on older Crays. >> Aires seems to be in between. The default GAMG configuration doesn't do >> any reductions in the coarse grid, so the issue is moot. If an >> iterative coarse solver was used, I think we would be more motivated to >> put the coarse problem on a subcomm. >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From iwaddington at gmail.com Thu Nov 21 19:08:49 2013 From: iwaddington at gmail.com (iwaddington .) Date: Thu, 21 Nov 2013 23:08:49 -0200 Subject: [petsc-users] KSPSetComputeRHS and KSPSetComputeOperators In-Reply-To: References: Message-ID: Thanks Matthew, but how exactly do I do that in a user defined function of the form used by KSPSetComputeOperators: #undef __FUNCT__ #define __FUNCT__ "ComputeMatrix" PetscErrorCode ComputeMatrix(KSP ksp, Mat J,Mat jac,MatStructure *str, void *ctx) { /* variables */ PetscFunctionBeginUser; /* Setting the system linear matrix J*/ /* Now I want to set jac=J which is the preconditioning matrix in this part , my first idea was to use matcopy, but I would like something more efficient */ PetscFunctionReturn(0); } On Thu, Nov 21, 2013 at 10:26 PM, Matthew Knepley wrote: > On Thu, Nov 21, 2013 at 6:21 PM, iwaddington . wrote: > >> Sorry Matthew, but I didn't understand what you said, because it is not >> the case of passing arguments to a function, I have to write the function >> which I will later pass to KSPSetComputeOperators, and in the function I >> have to set the system linear matrix and preconditioning matrix. >> > > You can return the same matrix for the system and preconditioner. > > Matt > > >> On Thu, Nov 21, 2013 at 10:00 PM, Matthew Knepley wrote: >> >>> On Thu, Nov 21, 2013 at 5:44 PM, iwaddington . wrote: >>> >>>> HI everybody, I have a doubt concerning how the function >>>> KSPSetComputeRHS works, because according to its prototype one of its >>>> arguments is a pointer to a function that takes a vector object as >>>> argument, not a pointer towards a vector object, so how is it able to set >>>> the rhs if the passage of the vector is done by value and not by reference ? >>>> >>> >>> The values of the Vec can be set, rather than changing the object >>> pointer itself. >>> >>> >>>> Another thing is the functioning of KSPSetComputeOperators. I want to >>>> set the linear system matrix and I want the preconditioning matrix to be >>>> equal, so can I just set the linar system matrix in the function that is >>>> argument of KSPSetComputeOperators, or do I need to make a matcopy in it ? >>>> >>> >>> No, just give the same argument twice. >>> >>> Matt >>> >>> -- >>> What most experimenters take for granted before they begin their >>> experiments is infinitely more interesting than any results to which their >>> experiments lead. >>> -- Norbert Wiener >>> >> >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Thu Nov 21 19:15:22 2013 From: knepley at gmail.com (Matthew Knepley) Date: Thu, 21 Nov 2013 19:15:22 -0600 Subject: [petsc-users] KSPSetComputeRHS and KSPSetComputeOperators In-Reply-To: References: Message-ID: On Thu, Nov 21, 2013 at 7:08 PM, iwaddington . wrote: > Thanks Matthew, but how exactly do I do that in a user defined function of > the form used by KSPSetComputeOperators: > There are a few problems here: 1) Your prototype is wrong: http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/KSP/KSPSetComputeOperators.html 2) Regardless, people generally use the matrix that is passed in. Thus you call MatSetValues(*J, ...) Why are you not doing this? 3) It is customary if you have the same system and preconditioner matrix to assemble into the preconditioner Matt > #undef __FUNCT__ > #define __FUNCT__ "ComputeMatrix" > PetscErrorCode ComputeMatrix(KSP ksp, Mat J,Mat jac,MatStructure *str, > void *ctx) > { > > /* variables */ > > PetscFunctionBeginUser; > > /* Setting the system linear matrix J*/ > > /* Now I want to set jac=J which is the preconditioning matrix in this > part , my first idea was to use matcopy, but I would like something more > efficient */ > > PetscFunctionReturn(0); > > } > > > > > On Thu, Nov 21, 2013 at 10:26 PM, Matthew Knepley wrote: > >> On Thu, Nov 21, 2013 at 6:21 PM, iwaddington . wrote: >> >>> Sorry Matthew, but I didn't understand what you said, because it is not >>> the case of passing arguments to a function, I have to write the function >>> which I will later pass to KSPSetComputeOperators, and in the function >>> I have to set the system linear matrix and preconditioning matrix. >>> >> >> You can return the same matrix for the system and preconditioner. >> >> Matt >> >> >>> On Thu, Nov 21, 2013 at 10:00 PM, Matthew Knepley wrote: >>> >>>> On Thu, Nov 21, 2013 at 5:44 PM, iwaddington . wrote: >>>> >>>>> HI everybody, I have a doubt concerning how the function >>>>> KSPSetComputeRHS works, because according to its prototype one of its >>>>> arguments is a pointer to a function that takes a vector object as >>>>> argument, not a pointer towards a vector object, so how is it able to set >>>>> the rhs if the passage of the vector is done by value and not by reference ? >>>>> >>>> >>>> The values of the Vec can be set, rather than changing the object >>>> pointer itself. >>>> >>>> >>>>> Another thing is the functioning of KSPSetComputeOperators. I want to >>>>> set the linear system matrix and I want the preconditioning matrix to be >>>>> equal, so can I just set the linar system matrix in the function that is >>>>> argument of KSPSetComputeOperators, or do I need to make a matcopy in it ? >>>>> >>>> >>>> No, just give the same argument twice. >>>> >>>> Matt >>>> >>>> -- >>>> What most experimenters take for granted before they begin their >>>> experiments is infinitely more interesting than any results to which their >>>> experiments lead. >>>> -- Norbert Wiener >>>> >>> >>> >> >> >> -- >> What most experimenters take for granted before they begin their >> experiments is infinitely more interesting than any results to which their >> experiments lead. >> -- Norbert Wiener >> > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Thu Nov 21 20:12:32 2013 From: irving at naml.us (Geoffrey Irving) Date: Thu, 21 Nov 2013 18:12:32 -0800 Subject: [petsc-users] orientation semantics for DMPlex Message-ID: What are the orientation semantics of DMPlex? I figured if I listed the edges of each triangle in counterclockwise order it would come out right, but it appears to be right for some triangles and wrong for others. Do I always have to call DMPlexSetConeOrientation? Unfortunately, I haven't yet grasped the documentation for DMPlexSetConeOrientation. Thanks, Geoffrey From jedbrown at mcs.anl.gov Thu Nov 21 20:38:25 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 21 Nov 2013 19:38:25 -0700 Subject: [petsc-users] GAMG processor reduction In-Reply-To: References: <871u29zc4z.fsf@jedbrown.org> <87y54hxwvk.fsf@jedbrown.org> <87txf5xsf8.fsf@jedbrown.org> Message-ID: <87y54h88ke.fsf@jedbrown.org> Mark Adams writes: > '-pc_gamg_verbose 2' will print out stuff where you can see the number of > active processors. We have not been super motivated to use subcoms because > there are no norms taken on coarse grids, except in the eigen estimator for > cheby, unless you use krylov in the solver, which is not usually useful. > Jed: GAMG does reduce active processors by default. I'm pretty sure. Yup, it reduces the active process set, but does not create a subcomm, There are no reductions (inner products or norms) on the coarse grid in the current configuration. Iterative coarse solvers would be a good reason to create a subcomm. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From iwaddington at gmail.com Thu Nov 21 20:47:41 2013 From: iwaddington at gmail.com (iwaddington .) Date: Fri, 22 Nov 2013 00:47:41 -0200 Subject: [petsc-users] KSPSetComputeRHS and KSPSetComputeOperators In-Reply-To: References: Message-ID: On Thu, Nov 21, 2013 at 11:50 PM, iwaddington . wrote: > But still regarding the issue of setting the matrix that will be used to > build the preconditioner, what would be the way to do it in the function > that I described you before , without the use of matcopy to make it equal > to the system matrix set with MatSetValuesStencil? My idea is to use the > conjugate gradient method. Sorry to bother so much, but I am learning petsc > all by myself. > > > > On Thu, Nov 21, 2013 at 11:27 PM, Matthew Knepley wrote: > >> On Thu, Nov 21, 2013 at 7:24 PM, iwaddington . wrote: >> >>> My idea is to use MatSetValuesStencil to set the system matrix, because >>> I am dealing with a dmda, and regarding the prototype, I know that what is >>> written in the web page is different, but when you look further in the >>> src/ksp/ksp/interface/itfunc.cand also the example files you see that my prototype is in fact the right >>> one. >>> >> >> I will fix the docs. Yes, just call MatSetValuesStencil. >> >> Matt >> >> >>> On Thu, Nov 21, 2013 at 11:15 PM, Matthew Knepley wrote: >>> >>>> On Thu, Nov 21, 2013 at 7:08 PM, iwaddington . wrote: >>>> >>>>> Thanks Matthew, but how exactly do I do that in a user defined >>>>> function of the form used by KSPSetComputeOperators: >>>>> >>>> >>>> There are a few problems here: >>>> >>>> 1) Your prototype is wrong: >>>> >>>> >>>> http://www.mcs.anl.gov/petsc/petsc-current/docs/manualpages/KSP/KSPSetComputeOperators.html >>>> >>>> 2) Regardless, people generally use the matrix that is passed in. Thus >>>> you call >>>> >>>> MatSetValues(*J, ...) >>>> >>>> Why are you not doing this? >>>> >>>> 3) It is customary if you have the same system and preconditioner >>>> matrix to assemble into the preconditioner >>>> >>>> Matt >>>> >>>> >>>>> #undef __FUNCT__ >>>>> #define __FUNCT__ "ComputeMatrix" >>>>> PetscErrorCode ComputeMatrix(KSP ksp, Mat J,Mat jac,MatStructure *str, >>>>> void *ctx) >>>>> { >>>>> >>>>> /* variables */ >>>>> >>>>> PetscFunctionBeginUser; >>>>> >>>>> /* Setting the system linear matrix J*/ >>>>> >>>>> /* Now I want to set jac=J which is the preconditioning matrix in this >>>>> part , my first idea was to use matcopy, but I would like something more >>>>> efficient */ >>>>> >>>>> PetscFunctionReturn(0); >>>>> >>>>> } >>>>> >>>>> >>>>> >>>>> >>>>> On Thu, Nov 21, 2013 at 10:26 PM, Matthew Knepley wrote: >>>>> >>>>>> On Thu, Nov 21, 2013 at 6:21 PM, iwaddington . >>>>> > wrote: >>>>>> >>>>>>> Sorry Matthew, but I didn't understand what you said, because it is >>>>>>> not the case of passing arguments to a function, I have to write the >>>>>>> function which I will later pass to KSPSetComputeOperators, and in >>>>>>> the function I have to set the system linear matrix and preconditioning >>>>>>> matrix. >>>>>>> >>>>>> >>>>>> You can return the same matrix for the system and preconditioner. >>>>>> >>>>>> Matt >>>>>> >>>>>> >>>>>>> On Thu, Nov 21, 2013 at 10:00 PM, Matthew Knepley >>>>>> > wrote: >>>>>>> >>>>>>>> On Thu, Nov 21, 2013 at 5:44 PM, iwaddington . < >>>>>>>> iwaddington at gmail.com> wrote: >>>>>>>> >>>>>>>>> HI everybody, I have a doubt concerning how the function >>>>>>>>> KSPSetComputeRHS works, because according to its prototype one of its >>>>>>>>> arguments is a pointer to a function that takes a vector object as >>>>>>>>> argument, not a pointer towards a vector object, so how is it able to set >>>>>>>>> the rhs if the passage of the vector is done by value and not by reference ? >>>>>>>>> >>>>>>>> >>>>>>>> The values of the Vec can be set, rather than changing the object >>>>>>>> pointer itself. >>>>>>>> >>>>>>>> >>>>>>>>> Another thing is the functioning of KSPSetComputeOperators. I want >>>>>>>>> to set the linear system matrix and I want the preconditioning matrix to be >>>>>>>>> equal, so can I just set the linar system matrix in the function that is >>>>>>>>> argument of KSPSetComputeOperators, or do I need to make a matcopy in it ? >>>>>>>>> >>>>>>>> >>>>>>>> No, just give the same argument twice. >>>>>>>> >>>>>>>> Matt >>>>>>>> >>>>>>>> -- >>>>>>>> What most experimenters take for granted before they begin their >>>>>>>> experiments is infinitely more interesting than any results to which their >>>>>>>> experiments lead. >>>>>>>> -- Norbert Wiener >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> What most experimenters take for granted before they begin their >>>>>> experiments is infinitely more interesting than any results to which their >>>>>> experiments lead. >>>>>> -- Norbert Wiener >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> What most experimenters take for granted before they begin their >>>> experiments is infinitely more interesting than any results to which their >>>> experiments lead. >>>> -- Norbert Wiener >>>> >>> >>> >> >> >> -- >> What most experimenters take for granted before they begin their >> experiments is infinitely more interesting than any results to which their >> experiments lead. >> -- Norbert Wiener >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Fri Nov 22 05:08:02 2013 From: knepley at gmail.com (Matthew Knepley) Date: Fri, 22 Nov 2013 05:08:02 -0600 Subject: [petsc-users] orientation semantics for DMPlex In-Reply-To: References: Message-ID: On Thu, Nov 21, 2013 at 8:12 PM, Geoffrey Irving wrote: > What are the orientation semantics of DMPlex? I figured if I listed > the edges of each triangle in counterclockwise order it would come out > right, but it appears to be right for some triangles and wrong for > others. Do I always have to call DMPlexSetConeOrientation? > > Unfortunately, I haven't yet grasped the documentation for > DMPlexSetConeOrientation. > This is because I have kept changing my mind. There is the way it is, and the way I want it to be now. So, the way it is: All cone arrows have an orientation attached. This indicates the order in which the _next_ cone (of that cone point) should be read out. The order is cyclic. A positive integer p means "start from vertex p instead of 0". A negative integer q means "start from -(q+1) and go backward". The important thing to remember here is that this is ordering only the cone points, and is composed with higher level ordering when reading out the full closure is 3D. I have considered changing the semantics for negative values to start at -(coneSize-q+1) since then negation does what you think it should (go backwards from the guy just before), but then all the routines needs to know the coneSize which is not great. The reason that your example does not work is that triangles on opposite sides of an edge must have the opposite orientation for that edge. For example, an edge e = {v0, v1} lies between two triangles. Now the cone of e will always come out {v0, v1}. In triangle 0 that is fine, but in triangle 1 it needs to be {v1, v0} if all triangles are oriented counter-clockwise, so we need orientation -2 for that edge in the cone of triangle 1. Thanks, Matt > Thanks, > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From mpovolot at purdue.edu Fri Nov 22 06:23:34 2013 From: mpovolot at purdue.edu (Michael Povolotskyi) Date: Fri, 22 Nov 2013 12:23:34 +0000 Subject: [petsc-users] question about LU decomposition of a conjugated matrix Message-ID: <528F4CC6.3050901@purdue.edu> Dear developers of Petsc and users, I have the following numerical problem. I need to solve the system Ax = B, where A is a square matrix, x and B are rectangular matrix The system is solved by doing LU decomposition. In the next step I need to solve a system A'y = C, where A' is a complex conjugate of A. My question: is it possible to reuse the LU decomposition of A for doing this? Thank you, Michael. -- Michael Povolotskyi, PhD Research Assistant Professor Network for Computational Nanotechnology 207 S Martin Jischke Drive Purdue University, DLR, room 441-10 West Lafayette, Indiana 47907 phone: +1-765-494-9396 fax: +1-765-496-6026 From irving at naml.us Fri Nov 22 12:11:19 2013 From: irving at naml.us (Geoffrey Irving) Date: Fri, 22 Nov 2013 10:11:19 -0800 Subject: [petsc-users] orientation semantics for DMPlex In-Reply-To: References: Message-ID: On Fri, Nov 22, 2013 at 3:08 AM, Matthew Knepley wrote: > On Thu, Nov 21, 2013 at 8:12 PM, Geoffrey Irving wrote: >> >> What are the orientation semantics of DMPlex? I figured if I listed >> the edges of each triangle in counterclockwise order it would come out >> right, but it appears to be right for some triangles and wrong for >> others. Do I always have to call DMPlexSetConeOrientation? >> >> Unfortunately, I haven't yet grasped the documentation for >> DMPlexSetConeOrientation. > > This is because I have kept changing my mind. There is the way it is, and > the way I want it to be now. So, the way it is: > > All cone arrows have an orientation attached. This indicates the order in > which the _next_ cone (of that cone point) should be read out. The order > is cyclic. A positive integer p means "start from vertex p instead of 0". A > negative integer q means "start from -(q+1) and go backward". Can't the order only be meaningfully described as cyclic only if the child element has dimension 2 (triangles) where the element boundary is topologically S^1? A cyclic order doesn't make any sense for tetrahedra (boundary S^2) or for segments (S^1), both of which have trivial fundamental group. If I have a triangle (0 1 2) pointing to child segment (0 1), wouldn't orientation values 0 and -1 refer to the same traversal? Since only one bit is ideally conveyed here, namely the sign of the permutation required to reorient the child to match the parent, would it be better to have orientations be PetscBool? Is it too late for that kind of API change? > The important thing to remember here is that this is ordering only the cone > points, and is composed with higher level ordering when reading out the > full closure is 3D. I have considered changing the semantics for negative > values to start at -(coneSize-q+1) since then negation does what you think > it should (go backwards from the guy just before), but then all the routines > needs to know the coneSize which is not great. > > The reason that your example does not work is that triangles on opposite sides > of an edge must have the opposite orientation for that edge. For example, an > edge e = {v0, v1} lies between two triangles. Now the cone of e will always come > out {v0, v1}. In triangle 0 that is fine, but in triangle 1 it needs to be {v1, v0} if all > triangles are oriented counter-clockwise, so we need orientation -2 for that edge > in the cone of triangle 1. Thanks, I should be good to go. It might be useful to have a consistency checking routine for users to run to make sure their custom created DMPlex is valid. Geoffrey From jedbrown at mcs.anl.gov Fri Nov 22 12:07:20 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 22 Nov 2013 12:07:20 -0600 Subject: [petsc-users] question about LU decomposition of a conjugated matrix In-Reply-To: <528F4CC6.3050901@purdue.edu> References: <528F4CC6.3050901@purdue.edu> Message-ID: <87wqk071k7.fsf@jedbrown.org> Michael Povolotskyi writes: > Dear developers of Petsc and users, > I have the following numerical problem. > > I need to solve the system Ax = B, where A is a square matrix, x and B > are rectangular matrix > The system is solved by doing LU decomposition. You can use MatLUFactor(), MatMatSolve(). There is no MatMatSolveTranspose() at this time. > In the next step I need to solve a system A'y = C, where A' is a complex > conjugate of A. > My question: is it possible to reuse the LU decomposition of A for doing > this? Not presently; you'd have to use MatHermitianTranspose() and refactor. Support for reuse would not be very difficult to implement. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From mpovolot at purdue.edu Fri Nov 22 07:18:18 2013 From: mpovolot at purdue.edu (Michael Povolotskyi) Date: Fri, 22 Nov 2013 13:18:18 +0000 Subject: [petsc-users] question about LU decomposition of a conjugated matrix In-Reply-To: <87wqk071k7.fsf@jedbrown.org> References: <528F4CC6.3050901@purdue.edu> <87wqk071k7.fsf@jedbrown.org> Message-ID: <528F599A.7070909@purdue.edu> On 11/22/2013 06:07 PM, Jed Brown wrote: > Michael Povolotskyi writes: > >> Dear developers of Petsc and users, >> I have the following numerical problem. >> >> I need to solve the system Ax = B, where A is a square matrix, x and B >> are rectangular matrix >> The system is solved by doing LU decomposition. > You can use MatLUFactor(), MatMatSolve(). There is no > MatMatSolveTranspose() at this time. > >> In the next step I need to solve a system A'y = C, where A' is a complex >> conjugate of A. >> My question: is it possible to reuse the LU decomposition of A for doing >> this? > Not presently; you'd have to use MatHermitianTranspose() and refactor. > Support for reuse would not be very difficult to implement. Thank you. I'm actually interested in implementing it. Could you, please, tell what has to be done in MatMatSolveTranspose()? By transpose I mean transpose and complex conjugation. Michael. From knepley at gmail.com Fri Nov 22 12:22:11 2013 From: knepley at gmail.com (Matthew Knepley) Date: Fri, 22 Nov 2013 12:22:11 -0600 Subject: [petsc-users] orientation semantics for DMPlex In-Reply-To: References: Message-ID: On Fri, Nov 22, 2013 at 12:11 PM, Geoffrey Irving wrote: > On Fri, Nov 22, 2013 at 3:08 AM, Matthew Knepley > wrote: > > On Thu, Nov 21, 2013 at 8:12 PM, Geoffrey Irving wrote: > >> > >> What are the orientation semantics of DMPlex? I figured if I listed > >> the edges of each triangle in counterclockwise order it would come out > >> right, but it appears to be right for some triangles and wrong for > >> others. Do I always have to call DMPlexSetConeOrientation? > >> > >> Unfortunately, I haven't yet grasped the documentation for > >> DMPlexSetConeOrientation. > > > > This is because I have kept changing my mind. There is the way it is, and > > the way I want it to be now. So, the way it is: > > > > All cone arrows have an orientation attached. This indicates the order > in > > which the _next_ cone (of that cone point) should be read out. The order > > is cyclic. A positive integer p means "start from vertex p instead of > 0". A > > negative integer q means "start from -(q+1) and go backward". > > Can't the order only be meaningfully described as cyclic only if the > child element has dimension 2 (triangles) where the element boundary > is topologically S^1? A cyclic order doesn't make any sense for > tetrahedra (boundary S^2) or for segments (S^1), both of which have > trivial fundamental group. If I have a triangle (0 1 2) pointing to > child segment (0 1), wouldn't orientation values 0 and -1 refer to the > same traversal? > There is a problem with the word "orientation", since it does not mean what is traditional in topology. Rather it means something close to a label for the element of the symmetric group associated with the cell. So it really is PetscInt. > Since only one bit is ideally conveyed here, namely the sign of the > permutation required to reorient the child to match the parent, would > it be better to have orientations be PetscBool? Is it too late for > that kind of API change? > > > The important thing to remember here is that this is ordering only the > cone > > points, and is composed with higher level ordering when reading out the > > full closure is 3D. I have considered changing the semantics for negative > > values to start at -(coneSize-q+1) since then negation does what you > think > > it should (go backwards from the guy just before), but then all the > routines > > needs to know the coneSize which is not great. > > > > The reason that your example does not work is that triangles on opposite > sides > > of an edge must have the opposite orientation for that edge. For > example, an > > edge e = {v0, v1} lies between two triangles. Now the cone of e will > always come > > out {v0, v1}. In triangle 0 that is fine, but in triangle 1 it needs to > be {v1, v0} if all > > triangles are oriented counter-clockwise, so we need orientation -2 for > that edge > > in the cone of triangle 1. > > Thanks, I should be good to go. It might be useful to have a > consistency checking routine for users to run to make sure their > custom created DMPlex is valid. I do have DMPlexOrient(). We should probably add DMPlexCheckOrientation() which does exactly the same thing except mismatches are not corrected. Matt > > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Fri Nov 22 12:29:41 2013 From: irving at naml.us (Geoffrey Irving) Date: Fri, 22 Nov 2013 10:29:41 -0800 Subject: [petsc-users] orientation semantics for DMPlex In-Reply-To: References: Message-ID: On Fri, Nov 22, 2013 at 10:22 AM, Matthew Knepley wrote: > On Fri, Nov 22, 2013 at 12:11 PM, Geoffrey Irving wrote: >> >> On Fri, Nov 22, 2013 at 3:08 AM, Matthew Knepley >> wrote: >> > On Thu, Nov 21, 2013 at 8:12 PM, Geoffrey Irving wrote: >> >> >> >> What are the orientation semantics of DMPlex? I figured if I listed >> >> the edges of each triangle in counterclockwise order it would come out >> >> right, but it appears to be right for some triangles and wrong for >> >> others. Do I always have to call DMPlexSetConeOrientation? >> >> >> >> Unfortunately, I haven't yet grasped the documentation for >> >> DMPlexSetConeOrientation. >> > >> > This is because I have kept changing my mind. There is the way it is, >> > and >> > the way I want it to be now. So, the way it is: >> > >> > All cone arrows have an orientation attached. This indicates the order >> > in >> > which the _next_ cone (of that cone point) should be read out. The order >> > is cyclic. A positive integer p means "start from vertex p instead of >> > 0". A >> > negative integer q means "start from -(q+1) and go backward". >> >> Can't the order only be meaningfully described as cyclic only if the >> child element has dimension 2 (triangles) where the element boundary >> is topologically S^1? A cyclic order doesn't make any sense for >> tetrahedra (boundary S^2) or for segments (S^1), both of which have >> trivial fundamental group. If I have a triangle (0 1 2) pointing to >> child segment (0 1), wouldn't orientation values 0 and -1 refer to the >> same traversal? > > > There is a problem with the word "orientation", since it does not mean what > is traditional in topology. Rather it means something close to a label for the > element of the symmetric group associated with the cell. So it really is > PetscInt. If that's the case, the documentation should mention the fact that of the four possible values of the orientation flag for a segment child of a triangle, there are only two distinct meanings. And for tetrahedra children of hypersimplices (not they those are particularly common) both forward and reversed orders refer to the same topological orientation, so the data structure doesn't work for that case at all. Geoffrey >> Since only one bit is ideally conveyed here, namely the sign of the >> permutation required to reorient the child to match the parent, would >> it be better to have orientations be PetscBool? Is it too late for >> that kind of API change? >> >> > The important thing to remember here is that this is ordering only the >> > cone >> > points, and is composed with higher level ordering when reading out the >> > full closure is 3D. I have considered changing the semantics for >> > negative >> > values to start at -(coneSize-q+1) since then negation does what you >> > think >> > it should (go backwards from the guy just before), but then all the >> > routines >> > needs to know the coneSize which is not great. >> > >> > The reason that your example does not work is that triangles on opposite >> > sides >> > of an edge must have the opposite orientation for that edge. For >> > example, an >> > edge e = {v0, v1} lies between two triangles. Now the cone of e will >> > always come >> > out {v0, v1}. In triangle 0 that is fine, but in triangle 1 it needs to >> > be {v1, v0} if all >> > triangles are oriented counter-clockwise, so we need orientation -2 for >> > that edge >> > in the cone of triangle 1. >> >> Thanks, I should be good to go. It might be useful to have a >> consistency checking routine for users to run to make sure their >> custom created DMPlex is valid. > > > I do have DMPlexOrient(). We should probably add DMPlexCheckOrientation() > which > does exactly the same thing except mismatches are not corrected. > > Matt > >> >> >> Geoffrey > > > > > -- > What most experimenters take for granted before they begin their experiments > is infinitely more interesting than any results to which their experiments > lead. > -- Norbert Wiener From irving at naml.us Fri Nov 22 12:31:51 2013 From: irving at naml.us (Geoffrey Irving) Date: Fri, 22 Nov 2013 10:31:51 -0800 Subject: [petsc-users] orientation semantics for DMPlex In-Reply-To: References: Message-ID: On Fri, Nov 22, 2013 at 10:29 AM, Geoffrey Irving wrote: > On Fri, Nov 22, 2013 at 10:22 AM, Matthew Knepley wrote: >> On Fri, Nov 22, 2013 at 12:11 PM, Geoffrey Irving wrote: >>> >>> On Fri, Nov 22, 2013 at 3:08 AM, Matthew Knepley >>> wrote: >>> > On Thu, Nov 21, 2013 at 8:12 PM, Geoffrey Irving wrote: >>> >> >>> >> What are the orientation semantics of DMPlex? I figured if I listed >>> >> the edges of each triangle in counterclockwise order it would come out >>> >> right, but it appears to be right for some triangles and wrong for >>> >> others. Do I always have to call DMPlexSetConeOrientation? >>> >> >>> >> Unfortunately, I haven't yet grasped the documentation for >>> >> DMPlexSetConeOrientation. >>> > >>> > This is because I have kept changing my mind. There is the way it is, >>> > and >>> > the way I want it to be now. So, the way it is: >>> > >>> > All cone arrows have an orientation attached. This indicates the order >>> > in >>> > which the _next_ cone (of that cone point) should be read out. The order >>> > is cyclic. A positive integer p means "start from vertex p instead of >>> > 0". A >>> > negative integer q means "start from -(q+1) and go backward". >>> >>> Can't the order only be meaningfully described as cyclic only if the >>> child element has dimension 2 (triangles) where the element boundary >>> is topologically S^1? A cyclic order doesn't make any sense for >>> tetrahedra (boundary S^2) or for segments (S^1), both of which have >>> trivial fundamental group. If I have a triangle (0 1 2) pointing to >>> child segment (0 1), wouldn't orientation values 0 and -1 refer to the >>> same traversal? >> >> >> There is a problem with the word "orientation", since it does not mean what >> is traditional in topology. Rather it means something close to a label for the >> element of the symmetric group associated with the cell. So it really is >> PetscInt. > > If that's the case, the documentation should mention the fact that of > the four possible values of the orientation flag for a segment child > of a triangle, there are only two distinct meanings. And for > tetrahedra children of hypersimplices (not they those are particularly > common) both forward and reversed orders refer to the same topological > orientation, so the data structure doesn't work for that case at all. Sorry, it does work, it's just that the sign doesn't matter. Shifting by one is an odd permutation on four indices. Sorry for the noise. Geoffrey From jedbrown at mcs.anl.gov Fri Nov 22 12:36:33 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 22 Nov 2013 12:36:33 -0600 Subject: [petsc-users] question about LU decomposition of a conjugated matrix In-Reply-To: <528F599A.7070909@purdue.edu> References: <528F4CC6.3050901@purdue.edu> <87wqk071k7.fsf@jedbrown.org> <528F599A.7070909@purdue.edu> Message-ID: <87r4a8707i.fsf@jedbrown.org> Michael, your time zone is mistakenly set to UTC, though the time is actually set to CST, so your mails appear out of order. Michael Povolotskyi writes: > I'm actually interested in implementing it. > Could you, please, tell what has to be done in MatMatSolveTranspose()? > > By transpose I mean transpose and complex conjugation. Some time ago, Barry proposed adding a function like MatSetTransposeType(A,MAT_TRANSPOSE_HERMITIAN) that would cause MatMultTranspose(), MatSolveTranspose(), etc., to use Hermitian transpose. The thought was that this was ultimately simpler to work since the problem formulation usually dictates whether a complex matrix uses a Hermitian transpose or not. Does that sound reasonable to you? I can see two implementation strategies. One is to go into MatLUFactor_SeqAIJ and set A->ops->transpose to a new function that explicitly transposes the factors. The other is to copy MatMatSolve and modify to be MatMatSolveTranspose, along with an implementation MatMatSolveTranspose_SeqAIJ (and external packages if relevant). -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From jedbrown at mcs.anl.gov Fri Nov 22 12:36:33 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 22 Nov 2013 12:36:33 -0600 Subject: [petsc-users] question about LU decomposition of a conjugated matrix In-Reply-To: <528F599A.7070909@purdue.edu> References: <528F4CC6.3050901@purdue.edu> <87wqk071k7.fsf@jedbrown.org> <528F599A.7070909@purdue.edu> Message-ID: <87r4a8707i.fsf@jedbrown.org> Michael, your time zone is mistakenly set to UTC, though the time is actually set to CST, so your mails appear out of order. Michael Povolotskyi writes: > I'm actually interested in implementing it. > Could you, please, tell what has to be done in MatMatSolveTranspose()? > > By transpose I mean transpose and complex conjugation. Some time ago, Barry proposed adding a function like MatSetTransposeType(A,MAT_TRANSPOSE_HERMITIAN) that would cause MatMultTranspose(), MatSolveTranspose(), etc., to use Hermitian transpose. The thought was that this was ultimately simpler to work since the problem formulation usually dictates whether a complex matrix uses a Hermitian transpose or not. Does that sound reasonable to you? I can see two implementation strategies. One is to go into MatLUFactor_SeqAIJ and set A->ops->transpose to a new function that explicitly transposes the factors. The other is to copy MatMatSolve and modify to be MatMatSolveTranspose, along with an implementation MatMatSolveTranspose_SeqAIJ (and external packages if relevant). -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From mpovolot at purdue.edu Fri Nov 22 08:06:40 2013 From: mpovolot at purdue.edu (Michael Povolotskyi) Date: Fri, 22 Nov 2013 14:06:40 +0000 Subject: [petsc-users] question about LU decomposition of a conjugated matrix In-Reply-To: <87r4a8707i.fsf@jedbrown.org> References: <528F4CC6.3050901@purdue.edu> <87wqk071k7.fsf@jedbrown.org> <528F599A.7070909@purdue.edu> <87r4a8707i.fsf@jedbrown.org> Message-ID: <528F64F0.2060906@purdue.edu> On 11/22/2013 06:36 PM, Jed Brown wrote: > Michael, your time zone is mistakenly set to UTC, though the time is > actually set to CST, so your mails appear out of order. > > Michael Povolotskyi writes: > >> I'm actually interested in implementing it. >> Could you, please, tell what has to be done in MatMatSolveTranspose()? >> >> By transpose I mean transpose and complex conjugation. > Some time ago, Barry proposed adding a function like > MatSetTransposeType(A,MAT_TRANSPOSE_HERMITIAN) that would cause > MatMultTranspose(), MatSolveTranspose(), etc., to use Hermitian > transpose. The thought was that this was ultimately simpler to work > since the problem formulation usually dictates whether a complex matrix > uses a Hermitian transpose or not. Does that sound reasonable to you? > > > I can see two implementation strategies. One is to go into > MatLUFactor_SeqAIJ and set A->ops->transpose to a new function that > explicitly transposes the factors. The other is to copy MatMatSolve and > modify to be MatMatSolveTranspose, along with an implementation > MatMatSolveTranspose_SeqAIJ (and external packages if relevant). Thank you, Jed, I fixed the time zone, sorry for the inconvenience. This sounds very reasonable for me and my colleagues to have MatSetTransposeType(A,MAT_TRANSPOSE_HERMITIAN). Our problems always require Hermitian transpose but I assume there are problems where it is not the case. Actually my goal is to solve the transposed problem with PARDISO package. It looks like it is the fastest option for the problems we usually solve. By the way, it there a chance to have interface to Pardiso in the next PETSc release? My colleague Jose David Bermeo sent the modules for this some time ago for your approval. Michael. From jedbrown at mcs.anl.gov Fri Nov 22 13:24:05 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 22 Nov 2013 13:24:05 -0600 Subject: [petsc-users] question about LU decomposition of a conjugated matrix In-Reply-To: <528F64F0.2060906@purdue.edu> References: <528F4CC6.3050901@purdue.edu> <87wqk071k7.fsf@jedbrown.org> <528F599A.7070909@purdue.edu> <87r4a8707i.fsf@jedbrown.org> <528F64F0.2060906@purdue.edu> Message-ID: <87fvqo6y0a.fsf@jedbrown.org> Michael Povolotskyi writes: > This sounds very reasonable for me and my colleagues to have > MatSetTransposeType(A,MAT_TRANSPOSE_HERMITIAN). > Our problems always require Hermitian transpose but I assume there are > problems where it is not the case. Complex symmetric shows up in frequency-domain wave problems. Methods like Sou-Cheng's CS-MINRES take advantage of that structure. > Actually my goal is to solve the transposed problem with PARDISO > package. It looks like it is the fastest option for the problems we > usually solve. > > By the way, it there a chance to have interface to Pardiso in the next > PETSc release? My colleague Jose David Bermeo sent the modules for this > some time ago for your approval. I was buried in a proposal deadline and then SC13, but I will go through the pull requests that nobody has gotten to. I think the Pardiso contribution is more-or-less ready to merge. For MatMatSolveTranspose, I would recommend doing it in a new branch starting from 'master'. Please include a reference implementation so that we can test without Pardiso (which is not free software). The code should be similar to MatMatSolve_SeqAIJ. Once that works (and has a test), you'll merge the Pardiso branch and add MatMatSolveTranspose_Pardiso. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From karpeev at mcs.anl.gov Fri Nov 22 15:36:20 2013 From: karpeev at mcs.anl.gov (Dmitry Karpeyev) Date: Fri, 22 Nov 2013 14:36:20 -0700 Subject: [petsc-users] question about LU decomposition of a conjugated matrix In-Reply-To: <87fvqo6y0a.fsf@jedbrown.org> References: <528F4CC6.3050901@purdue.edu> <87wqk071k7.fsf@jedbrown.org> <528F599A.7070909@purdue.edu> <87r4a8707i.fsf@jedbrown.org> <528F64F0.2060906@purdue.edu> <87fvqo6y0a.fsf@jedbrown.org> Message-ID: On Fri, Nov 22, 2013 at 12:24 PM, Jed Brown wrote: > Michael Povolotskyi writes: > > > This sounds very reasonable for me and my colleagues to have > > MatSetTransposeType(A,MAT_TRANSPOSE_HERMITIAN). > > Our problems always require Hermitian transpose but I assume there are > > problems where it is not the case. > > Complex symmetric shows up in frequency-domain wave problems. Methods > like Sou-Cheng's CS-MINRES take advantage of that structure. > > > Actually my goal is to solve the transposed problem with PARDISO > > package. It looks like it is the fastest option for the problems we > > usually solve. > > > > By the way, it there a chance to have interface to Pardiso in the next > > PETSc release? My colleague Jose David Bermeo sent the modules for this > > some time ago for your approval. > > I was buried in a proposal deadline and then SC13, but I will go through > the pull requests that nobody has gotten to. I think the Pardiso > contribution is more-or-less ready to merge. For MatMatSolveTranspose, > I would recommend doing it in a new branch starting from 'master'. > Please include a reference implementation so that we can test without > Pardiso (which is not free software). The code should be similar to > MatMatSolve_SeqAIJ. Once that works (and has a test), you'll merge the > Pardiso branch and add MatMatSolveTranspose_Pardiso. > By analogy with MatTransposeMatMult() etc., should this be MatTransposeMatSolve()? Since 'Transpose' applies to the first Mat. Dmitry. -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Fri Nov 22 15:42:35 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 22 Nov 2013 15:42:35 -0600 Subject: [petsc-users] question about LU decomposition of a conjugated matrix In-Reply-To: References: <528F4CC6.3050901@purdue.edu> <87wqk071k7.fsf@jedbrown.org> <528F599A.7070909@purdue.edu> <87r4a8707i.fsf@jedbrown.org> <528F64F0.2060906@purdue.edu> <87fvqo6y0a.fsf@jedbrown.org> Message-ID: <87mwkw5d10.fsf@jedbrown.org> Dmitry Karpeyev writes: > By analogy with MatTransposeMatMult() etc., should this be > MatTransposeMatSolve()? > Since 'Transpose' applies to the first Mat. Sure, just be consistent. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From irving at naml.us Fri Nov 22 17:36:05 2013 From: irving at naml.us (Geoffrey Irving) Date: Fri, 22 Nov 2013 15:36:05 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 Message-ID: I have a duplicate of snes ex12 (FEM Poisson) which works with Dirichlet boundary conditions, but it's breaking for me with Neumann conditions. In particular, with Neumann conditions I get results which explode even though I believe I am setting a constant nullspace. For example, if I use two first order elements (the unit square divided into two triangles), the resulting solution has L2 error = 1.75514e+08 u = [-175513825.75680602, -175513825.66302037, -175513825.48390722, -175513824.84436429] This looks rather a lot like the null space isn't getting through. I am creating the constant nullspace with MatNullSpace null; CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); CHECK(MatSetNullSpace(m,null)); CHECK(MatNullSpaceDestroy(&null)); If I pass "-ksp_view -mat_view", I get the following. The matrix entries seem right (they do indeed have the constant nullspace), and ksp_view shows that a nullspace is attached. Is attaching the nullspace to the matrix with MatSetNullSpace enough, or do I need to additionally attach it to the KSP object? Thanks, Geoffrey --------------------------------------------- Mat Object: 1 MPI processes type: seqaij row 0: (0, 1) (1, -0.5) (2, -0.5) row 1: (0, -0.5) (1, 1) (2, 0) (3, -0.5) row 2: (0, -0.5) (1, 0) (2, 1) (3, -0.5) row 3: (1, -0.5) (2, -0.5) (3, 1) KSP Object: 1 MPI processes type: gmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning has attached null space using PRECONDITIONED norm type for convergence test PC Object: 1 MPI processes type: ilu ILU: out-of-place factorization 0 levels of fill tolerance for zero pivot 2.22045e-14 using diagonal shift on blocks to prevent zero pivot [INBLOCKS] matrix ordering: natural factor fill ratio given 1, needed 1 Factored matrix follows: Mat Object: 1 MPI processes type: seqaij rows=4, cols=4 package used to perform factorization: petsc total: nonzeros=14, allocated nonzeros=14 total number of mallocs used during MatSetValues calls =0 using I-node routines: found 3 nodes, limit used is 5 linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=4, cols=4 total: nonzeros=14, allocated nonzeros=14 total number of mallocs used during MatSetValues calls =0 has attached null space using I-node routines: found 3 nodes, limit used is 5 From knepley at gmail.com Fri Nov 22 17:41:09 2013 From: knepley at gmail.com (Matthew Knepley) Date: Fri, 22 Nov 2013 17:41:09 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving wrote: > I have a duplicate of snes ex12 (FEM Poisson) which works with > Dirichlet boundary conditions, but it's breaking for me with Neumann > conditions. In particular, with Neumann conditions I get results > which explode even though I believe I am setting a constant nullspace. > > For example, if I use two first order elements (the unit square > divided into two triangles), the resulting solution has > > L2 error = 1.75514e+08 > u = [-175513825.75680602, -175513825.66302037, > -175513825.48390722, -175513824.84436429] > > This looks rather a lot like the null space isn't getting through. I > am creating the constant nullspace with > > MatNullSpace null; > CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); > CHECK(MatSetNullSpace(m,null)); > CHECK(MatNullSpaceDestroy(&null)); > > If I pass "-ksp_view -mat_view", I get the following. The matrix > entries seem right (they do indeed have the constant nullspace), and > ksp_view shows that a nullspace is attached. Is attaching the > nullspace to the matrix with MatSetNullSpace enough, or do I need to > additionally attach it to the KSP object? > 1) I always run with -ksp_monitor_true_residual now when debugging. This can give you an idea whether you have a singular PC, which I suspect here. 2) Can you try using -pc_type jacobi? I think ILU might go crazy on a deficient matrix. Thanks, Matt > Thanks, > Geoffrey > > --------------------------------------------- > > Mat Object: 1 MPI processes > type: seqaij > row 0: (0, 1) (1, -0.5) (2, -0.5) > row 1: (0, -0.5) (1, 1) (2, 0) (3, -0.5) > row 2: (0, -0.5) (1, 0) (2, 1) (3, -0.5) > row 3: (1, -0.5) (2, -0.5) (3, 1) > KSP Object: 1 MPI processes > type: gmres > GMRES: restart=30, using Classical (unmodified) Gram-Schmidt > Orthogonalization with no iterative refinement > GMRES: happy breakdown tolerance 1e-30 > maximum iterations=10000, initial guess is zero > tolerances: relative=1e-05, absolute=1e-50, divergence=10000 > left preconditioning > has attached null space > using PRECONDITIONED norm type for convergence test > PC Object: 1 MPI processes > type: ilu > ILU: out-of-place factorization > 0 levels of fill > tolerance for zero pivot 2.22045e-14 > using diagonal shift on blocks to prevent zero pivot [INBLOCKS] > matrix ordering: natural > factor fill ratio given 1, needed 1 > Factored matrix follows: > Mat Object: 1 MPI processes > type: seqaij > rows=4, cols=4 > package used to perform factorization: petsc > total: nonzeros=14, allocated nonzeros=14 > total number of mallocs used during MatSetValues calls =0 > using I-node routines: found 3 nodes, limit used is 5 > linear system matrix = precond matrix: > Mat Object: 1 MPI processes > type: seqaij > rows=4, cols=4 > total: nonzeros=14, allocated nonzeros=14 > total number of mallocs used during MatSetValues calls =0 > has attached null space > using I-node routines: found 3 nodes, limit used is 5 > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Fri Nov 22 18:09:25 2013 From: irving at naml.us (Geoffrey Irving) Date: Fri, 22 Nov 2013 16:09:25 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley wrote: > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving wrote: >> >> I have a duplicate of snes ex12 (FEM Poisson) which works with >> Dirichlet boundary conditions, but it's breaking for me with Neumann >> conditions. In particular, with Neumann conditions I get results >> which explode even though I believe I am setting a constant nullspace. >> >> For example, if I use two first order elements (the unit square >> divided into two triangles), the resulting solution has >> >> L2 error = 1.75514e+08 >> u = [-175513825.75680602, -175513825.66302037, >> -175513825.48390722, -175513824.84436429] >> >> This looks rather a lot like the null space isn't getting through. I >> am creating the constant nullspace with >> >> MatNullSpace null; >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); >> CHECK(MatSetNullSpace(m,null)); >> CHECK(MatNullSpaceDestroy(&null)); >> >> If I pass "-ksp_view -mat_view", I get the following. The matrix >> entries seem right (they do indeed have the constant nullspace), and >> ksp_view shows that a nullspace is attached. Is attaching the >> nullspace to the matrix with MatSetNullSpace enough, or do I need to >> additionally attach it to the KSP object? > > > 1) I always run with -ksp_monitor_true_residual now when debugging. This can > give > you an idea whether you have a singular PC, which I suspect here. > > 2) Can you try using -pc_type jacobi? I think ILU might go crazy on a > deficient matrix. Here are results with -ksp_monitor_true_residual -pc_type none: http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 Both versions converge in 3 iterations for the first SNES iteration, but the relative one starts to churn after that since the residual starts very small. The true residual goes down to 4/3 and stagnates. Is there a convenient way to print out the RHS to see whether it has a component in the nullspace (which seems likely given the true residual stagnation)? I suppose I already do print the result of SNESComputeFunction on the zero vec, which is RHS = [ 6.66666667e-01 1.33333333e+00 6.66666667e-01 1.11022302e-16] Thanks, Geoffrey From jedbrown at mcs.anl.gov Fri Nov 22 18:17:27 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 22 Nov 2013 18:17:27 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: <87wqk03rag.fsf@jedbrown.org> Geoffrey Irving writes: > Here are results with -ksp_monitor_true_residual -pc_type none: > > http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 > http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 Looks like the preconditioner is singular. > Both versions converge in 3 iterations for the first SNES iteration, > but the relative one starts to churn after that since the residual > starts very small. The true residual goes down to 4/3 and stagnates. > Is there a convenient way to print out the RHS to see whether it has a > component in the nullspace (which seems likely given the true residual > stagnation)? -snes_monitor_residual will plot the solution. (This option should be upgraded to support arbitrary output formats.) You can also call VecView from your residual function. > I suppose I already do print the result of SNESComputeFunction on the > zero vec, which is > > RHS = [ 6.66666667e-01 1.33333333e+00 6.66666667e-01 1.11022302e-16] Looks like you need to project out the null space. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From irving at naml.us Fri Nov 22 18:25:25 2013 From: irving at naml.us (Geoffrey Irving) Date: Fri, 22 Nov 2013 16:25:25 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: <87wqk03rag.fsf@jedbrown.org> References: <87wqk03rag.fsf@jedbrown.org> Message-ID: On Fri, Nov 22, 2013 at 4:17 PM, Jed Brown wrote: > Geoffrey Irving writes: >> Here are results with -ksp_monitor_true_residual -pc_type none: >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 >> http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 > > Looks like the preconditioner is singular. The preconditioner is none: -pc_type none. >> Both versions converge in 3 iterations for the first SNES iteration, >> but the relative one starts to churn after that since the residual >> starts very small. The true residual goes down to 4/3 and stagnates. >> Is there a convenient way to print out the RHS to see whether it has a >> component in the nullspace (which seems likely given the true residual >> stagnation)? > > -snes_monitor_residual will plot the solution. (This option should be > upgraded to support arbitrary output formats.) You can also call > VecView from your residual function. > >> I suppose I already do print the result of SNESComputeFunction on the >> zero vec, which is >> >> RHS = [ 6.66666667e-01 1.33333333e+00 6.66666667e-01 1.11022302e-16] > > Looks like you need to project out the null space. Yep, I was hallucinating that analytically consistent implies discretely consistent. Is there a standard way to do this in the context of an SNES, where I'm not computing the residual directly myself? Should I write a wrapper around DMPlexComputeResidualFEM and pass the wrapper to DMSNESSetFunctionLocal, or is there a way to tell SNES about the nullspace directly? Is such a projection happening somewhere in snes ex12? Thanks, Geoffrey From knepley at gmail.com Fri Nov 22 18:25:54 2013 From: knepley at gmail.com (Matthew Knepley) Date: Fri, 22 Nov 2013 18:25:54 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving wrote: > On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley > wrote: > > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving wrote: > >> > >> I have a duplicate of snes ex12 (FEM Poisson) which works with > >> Dirichlet boundary conditions, but it's breaking for me with Neumann > >> conditions. In particular, with Neumann conditions I get results > >> which explode even though I believe I am setting a constant nullspace. > >> > >> For example, if I use two first order elements (the unit square > >> divided into two triangles), the resulting solution has > >> > >> L2 error = 1.75514e+08 > >> u = [-175513825.75680602, -175513825.66302037, > >> -175513825.48390722, -175513824.84436429] > >> > >> This looks rather a lot like the null space isn't getting through. I > >> am creating the constant nullspace with > >> > >> MatNullSpace null; > >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); > >> CHECK(MatSetNullSpace(m,null)); > >> CHECK(MatNullSpaceDestroy(&null)); > >> > >> If I pass "-ksp_view -mat_view", I get the following. The matrix > >> entries seem right (they do indeed have the constant nullspace), and > >> ksp_view shows that a nullspace is attached. Is attaching the > >> nullspace to the matrix with MatSetNullSpace enough, or do I need to > >> additionally attach it to the KSP object? > > > > > > 1) I always run with -ksp_monitor_true_residual now when debugging. This > can > > give > > you an idea whether you have a singular PC, which I suspect here. > > > > 2) Can you try using -pc_type jacobi? I think ILU might go crazy on a > > deficient matrix. > > Here are results with -ksp_monitor_true_residual -pc_type none: > > http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 > http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 > Okay, if you have an inconsistent RHS I do not think that true_residual will work since it uses the unprojected b, but the solve should be fine. Matt > Both versions converge in 3 iterations for the first SNES iteration, > but the relative one starts to churn after that since the residual > starts very small. The true residual goes down to 4/3 and stagnates. > Is there a convenient way to print out the RHS to see whether it has a > component in the nullspace (which seems likely given the true residual > stagnation)? > > I suppose I already do print the result of SNESComputeFunction on the > zero vec, which is > > RHS = [ 6.66666667e-01 1.33333333e+00 6.66666667e-01 > 1.11022302e-16] > > Thanks, > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Fri Nov 22 18:27:21 2013 From: knepley at gmail.com (Matthew Knepley) Date: Fri, 22 Nov 2013 18:27:21 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: <87wqk03rag.fsf@jedbrown.org> Message-ID: On Fri, Nov 22, 2013 at 6:25 PM, Geoffrey Irving wrote: > On Fri, Nov 22, 2013 at 4:17 PM, Jed Brown wrote: > > Geoffrey Irving writes: > >> Here are results with -ksp_monitor_true_residual -pc_type none: > >> > >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 > >> http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 > > > > Looks like the preconditioner is singular. > > The preconditioner is none: -pc_type none. > > >> Both versions converge in 3 iterations for the first SNES iteration, > >> but the relative one starts to churn after that since the residual > >> starts very small. The true residual goes down to 4/3 and stagnates. > >> Is there a convenient way to print out the RHS to see whether it has a > >> component in the nullspace (which seems likely given the true residual > >> stagnation)? > > > > -snes_monitor_residual will plot the solution. (This option should be > > upgraded to support arbitrary output formats.) You can also call > > VecView from your residual function. > > > >> I suppose I already do print the result of SNESComputeFunction on the > >> zero vec, which is > >> > >> RHS = [ 6.66666667e-01 1.33333333e+00 6.66666667e-01 > 1.11022302e-16] > > > > Looks like you need to project out the null space. > > Yep, I was hallucinating that analytically consistent implies > discretely consistent. > > Is there a standard way to do this in the context of an SNES, where > I'm not computing the residual directly myself? Should I write a > wrapper around DMPlexComputeResidualFEM and pass the wrapper to > DMSNESSetFunctionLocal, or is there a way to tell SNES about the > nullspace directly? Is such a projection happening somewhere in snes > ex12? > SNES will do it automatically. You can just call MatNullSpaceRemove(). Matt > Thanks, > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Fri Nov 22 18:42:23 2013 From: irving at naml.us (Geoffrey Irving) Date: Fri, 22 Nov 2013 16:42:23 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley wrote: > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving wrote: >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley >> wrote: >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving wrote: >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which works with >> >> Dirichlet boundary conditions, but it's breaking for me with Neumann >> >> conditions. In particular, with Neumann conditions I get results >> >> which explode even though I believe I am setting a constant nullspace. >> >> >> >> For example, if I use two first order elements (the unit square >> >> divided into two triangles), the resulting solution has >> >> >> >> L2 error = 1.75514e+08 >> >> u = [-175513825.75680602, -175513825.66302037, >> >> -175513825.48390722, -175513824.84436429] >> >> >> >> This looks rather a lot like the null space isn't getting through. I >> >> am creating the constant nullspace with >> >> >> >> MatNullSpace null; >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); >> >> CHECK(MatSetNullSpace(m,null)); >> >> CHECK(MatNullSpaceDestroy(&null)); >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. The matrix >> >> entries seem right (they do indeed have the constant nullspace), and >> >> ksp_view shows that a nullspace is attached. Is attaching the >> >> nullspace to the matrix with MatSetNullSpace enough, or do I need to >> >> additionally attach it to the KSP object? >> > >> > >> > 1) I always run with -ksp_monitor_true_residual now when debugging. This >> > can >> > give >> > you an idea whether you have a singular PC, which I suspect here. >> > >> > 2) Can you try using -pc_type jacobi? I think ILU might go crazy on a >> > deficient matrix. >> >> Here are results with -ksp_monitor_true_residual -pc_type none: >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 >> http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 > > > Okay, if you have an inconsistent RHS I do not think that true_residual will work > since it uses the unprojected b, but the solve should be fine. I still don't understand why the atol version is able to drift so far away from zero mean, even after tens of thousands of iterations. If KSP sees a null space on the matrix, shouldn't it project that null space out of the *linear system* residual and also out of solution on each iteration? Even if it is only projecting out of the solution delta, how can null space errors be accumulating? >> Is there a standard way to do this in the context of an SNES, where >> I'm not computing the residual directly myself? Should I write a >> wrapper around DMPlexComputeResidualFEM and pass the wrapper to >> DMSNESSetFunctionLocal, or is there a way to tell SNES about the >> nullspace directly? Is such a projection happening somewhere in snes >> ex12? > SNES will do it automatically. You can just call MatNullSpaceRemove(). I.e., SNES will call MatNullSpaceRemove automatically, or I should manually in a wrapper around DMSNESSetFunctionLocal? Thanks, Geoffrey From jsd1 at rice.edu Sat Nov 23 00:40:39 2013 From: jsd1 at rice.edu (Justin Dong) Date: Sat, 23 Nov 2013 00:40:39 -0600 Subject: [petsc-users] Compatibility between PETSc typedefs and regular typedefs? Message-ID: I?m still trying to get the hang of PETSc but am adapting pretty well. I was cleaning up some code I wrote today and realized that my function prototype expects a pointer to a double but I accidentally gave it a pointer to a PetscScalar instead. I didn't get any errors during compiling or execution so I was just wondering if the two types are compatible. The same goes for PetscInt and ints. Here's a minimal example below. Is this considered a major error? I'm correcting it anyway but just wanted to satisfy my curiosity. #define A(i,j) A[ j + 5*i ] #define b(i,j) b[ j + 5*i ] void myFunction(double* A, int* b) { int i, j; for (i=0; i<5, ++i) { for (j=0; j<5; ++j) { A(i,j) = sin(i+j); b(i,j) = i+j; } } } int main(int argc, char **argv) { PetscErrorCode ierr; ierr = PetscInitialize(&argc,&argv,(char*)0,help); CHKERRQ(ierr); PetscScalar* A; PetscInt* b; int n = 5; PetscMalloc(sizeof(PetscScalar)*n*n, &A); PetscMalloc(sizeof(PetscInt)*n*n, &b); myFunction(A, b); PetscFree(A); PetscFree(b); ierr = PetscFinalize(); return 0; } Thanks, Justin -------------- next part -------------- An HTML attachment was scrubbed... URL: From dave.mayhem23 at gmail.com Sat Nov 23 03:36:10 2013 From: dave.mayhem23 at gmail.com (Dave May) Date: Sat, 23 Nov 2013 10:36:10 +0100 Subject: [petsc-users] Compatibility between PETSc typedefs and regular typedefs? In-Reply-To: References: Message-ID: If you want your code to be portable, you shouldn't definitely not assume that PetscScalar is the same as double, nor should you assume that PetscInt is the same as int. The reason is that depending on how you configured PETSc, these types can be mapped to different different types. For example, PetscScalar could be mapped to PetscScalar ==> PetscReal == double (default) PetscScalar ==> PetscReal ==> float (--with-precision=single) PetscScalar ==> double precision complex number (--with-sclar-type=complex) whilst PetscInt ==> 32 bit int (default) PetscInt ==> 64 bit int (--with-64bit-indices) Cheers, Dave On 23 November 2013 07:40, Justin Dong wrote: > I?m still trying to get the hang of PETSc but am adapting pretty well. I > was cleaning up some code I wrote today and realized that my function > prototype expects a pointer to a double but I accidentally gave it a > pointer to a PetscScalar instead. I didn't get any errors during compiling > or execution so I was just wondering if the two types are compatible. The > same goes for PetscInt and ints. Here's a minimal example below. Is this > considered a major error? I'm correcting it anyway but just wanted to > satisfy my curiosity. > > > #define A(i,j) A[ j + 5*i ] > #define b(i,j) b[ j + 5*i ] > > void myFunction(double* A, int* b) > { > > int i, j; > > for (i=0; i<5, ++i) > { > for (j=0; j<5; ++j) > { > A(i,j) = sin(i+j); > b(i,j) = i+j; > } > } > > } > > int main(int argc, char **argv) > { > > PetscErrorCode ierr; > > ierr = PetscInitialize(&argc,&argv,(char*)0,help); CHKERRQ(ierr); > PetscScalar* A; > PetscInt* b; > > int n = 5; > > PetscMalloc(sizeof(PetscScalar)*n*n, &A); > PetscMalloc(sizeof(PetscInt)*n*n, &b); > > myFunction(A, b); > > PetscFree(A); > PetscFree(b); > > ierr = PetscFinalize(); > > return 0; > } > > Thanks, > Justin > -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Sat Nov 23 06:43:51 2013 From: knepley at gmail.com (Matthew Knepley) Date: Sat, 23 Nov 2013 06:43:51 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving wrote: > On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley > wrote: > > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving wrote: > >> > >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley > >> wrote: > >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving > wrote: > >> >> > >> >> I have a duplicate of snes ex12 (FEM Poisson) which works with > >> >> Dirichlet boundary conditions, but it's breaking for me with Neumann > >> >> conditions. In particular, with Neumann conditions I get results > >> >> which explode even though I believe I am setting a constant > nullspace. > >> >> > >> >> For example, if I use two first order elements (the unit square > >> >> divided into two triangles), the resulting solution has > >> >> > >> >> L2 error = 1.75514e+08 > >> >> u = [-175513825.75680602, -175513825.66302037, > >> >> -175513825.48390722, -175513824.84436429] > >> >> > >> >> This looks rather a lot like the null space isn't getting through. I > >> >> am creating the constant nullspace with > >> >> > >> >> MatNullSpace null; > >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); > >> >> CHECK(MatSetNullSpace(m,null)); > >> >> CHECK(MatNullSpaceDestroy(&null)); > >> >> > >> >> If I pass "-ksp_view -mat_view", I get the following. The matrix > >> >> entries seem right (they do indeed have the constant nullspace), and > >> >> ksp_view shows that a nullspace is attached. Is attaching the > >> >> nullspace to the matrix with MatSetNullSpace enough, or do I need to > >> >> additionally attach it to the KSP object? > >> > > >> > > >> > 1) I always run with -ksp_monitor_true_residual now when debugging. > This > >> > can > >> > give > >> > you an idea whether you have a singular PC, which I suspect here. > >> > > >> > 2) Can you try using -pc_type jacobi? I think ILU might go crazy on a > >> > deficient matrix. > >> > >> Here are results with -ksp_monitor_true_residual -pc_type none: > >> > >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 > >> http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 > > > > > > Okay, if you have an inconsistent RHS I do not think that true_residual > will work > > since it uses the unprojected b, but the solve should be fine. > > I still don't understand why the atol version is able to drift so far > away from zero mean, even after tens of thousands of iterations. If > KSP sees a null space on the matrix, shouldn't it project that null > space out of the *linear system* residual and also out of solution on > each iteration? Even if it is only projecting out of the solution > delta, how can null space errors be accumulating? Both the KSP and Mat show that the null space is set, so everything should work fine, and at this point its no longer DMPlex that is in control, its standard PETSc. We have reached the limit of usefu talking. Something is obviously wrong with the code, but since this routinely works in PETSc examples. In situations like these I think we need to follow the execution in the debugger to see what is wrong..You can look at Vec values in the debugger using (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n and I look at DMPlex things with (gdb) p ((DM_Plex*) dm->data)->coneSection etc. Matt > >> Is there a standard way to do this in the context of an SNES, where > >> I'm not computing the residual directly myself? Should I write a > >> wrapper around DMPlexComputeResidualFEM and pass the wrapper to > >> DMSNESSetFunctionLocal, or is there a way to tell SNES about the > >> nullspace directly? Is such a projection happening somewhere in snes > >> ex12? > > > SNES will do it automatically. You can just call MatNullSpaceRemove(). > > I.e., SNES will call MatNullSpaceRemove automatically, or I should > manually in a wrapper around DMSNESSetFunctionLocal? > > Thanks, > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Sat Nov 23 11:58:13 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Sat, 23 Nov 2013 11:58:13 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: <87mwkv9f0q.fsf@jedbrown.org> Geoffrey Irving writes: >> SNES will do it automatically. You can just call MatNullSpaceRemove(). > > I.e., SNES will call MatNullSpaceRemove automatically, or I should > manually in a wrapper around DMSNESSetFunctionLocal? If you can't make the SNES residual be consistent, then yes, you'll have to remove it explicitly. If the problem is only that the solution "drifts" after a huge number of solves, you could use SNESLineSearchSetPostCheck or otherwise project out the noise. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From irving at naml.us Sat Nov 23 12:07:57 2013 From: irving at naml.us (Geoffrey Irving) Date: Sat, 23 Nov 2013 10:07:57 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley wrote: > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving wrote: >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley >> wrote: >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving wrote: >> >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley >> >> wrote: >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving >> >> > wrote: >> >> >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which works with >> >> >> Dirichlet boundary conditions, but it's breaking for me with Neumann >> >> >> conditions. In particular, with Neumann conditions I get results >> >> >> which explode even though I believe I am setting a constant >> >> >> nullspace. >> >> >> >> >> >> For example, if I use two first order elements (the unit square >> >> >> divided into two triangles), the resulting solution has >> >> >> >> >> >> L2 error = 1.75514e+08 >> >> >> u = [-175513825.75680602, -175513825.66302037, >> >> >> -175513825.48390722, -175513824.84436429] >> >> >> >> >> >> This looks rather a lot like the null space isn't getting through. >> >> >> I >> >> >> am creating the constant nullspace with >> >> >> >> >> >> MatNullSpace null; >> >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); >> >> >> CHECK(MatSetNullSpace(m,null)); >> >> >> CHECK(MatNullSpaceDestroy(&null)); >> >> >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. The matrix >> >> >> entries seem right (they do indeed have the constant nullspace), and >> >> >> ksp_view shows that a nullspace is attached. Is attaching the >> >> >> nullspace to the matrix with MatSetNullSpace enough, or do I need to >> >> >> additionally attach it to the KSP object? >> >> > >> >> > >> >> > 1) I always run with -ksp_monitor_true_residual now when debugging. >> >> > This >> >> > can >> >> > give >> >> > you an idea whether you have a singular PC, which I suspect here. >> >> > >> >> > 2) Can you try using -pc_type jacobi? I think ILU might go crazy on a >> >> > deficient matrix. >> >> >> >> Here are results with -ksp_monitor_true_residual -pc_type none: >> >> >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 >> >> http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 >> > >> > >> > Okay, if you have an inconsistent RHS I do not think that true_residual >> > will work >> > since it uses the unprojected b, but the solve should be fine. >> >> I still don't understand why the atol version is able to drift so far >> away from zero mean, even after tens of thousands of iterations. If >> KSP sees a null space on the matrix, shouldn't it project that null >> space out of the *linear system* residual and also out of solution on >> each iteration? Even if it is only projecting out of the solution >> delta, how can null space errors be accumulating? > > > Both the KSP and Mat show that the null space is set, so everything should > work fine, > and at this point its no longer DMPlex that is in control, its standard > PETSc. > > We have reached the limit of usefu talking. Something is obviously wrong with the code, > but since this routinely works in PETSc examples. In situations like these I think we need > to follow the execution in the debugger to see what is wrong..You can look at Vec values > in the debugger using > > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n > > and I look at DMPlex things with > > (gdb) p ((DM_Plex*) dm->data)->coneSection > > etc. Thanks, I appreciate the help. It looks like there were at least two different problems: 1. The boundary FE I was creating had the same dimension as the interior FE (instead of codimension 1), due to misreading ex12 even though I had correctly refactored it. I added a dimension consistency check to my code, but I can do this in DMPlexComputeResidualFEM as well to catch future user errors. 2. Even after fixing the dimensions, my boundary functions in PetscFEM are getting x values both inside and completely outside the domain. Almost certainly more user error, but hopefully also something I can add a check for in petsc once I localize it. Geoffrey From knepley at gmail.com Sat Nov 23 12:11:31 2013 From: knepley at gmail.com (Matthew Knepley) Date: Sat, 23 Nov 2013 12:11:31 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving wrote: > On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley > wrote: > > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving wrote: > >> > >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley > >> wrote: > >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving > wrote: > >> >> > >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley > >> >> wrote: > >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving > >> >> > wrote: > >> >> >> > >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which works with > >> >> >> Dirichlet boundary conditions, but it's breaking for me with > Neumann > >> >> >> conditions. In particular, with Neumann conditions I get results > >> >> >> which explode even though I believe I am setting a constant > >> >> >> nullspace. > >> >> >> > >> >> >> For example, if I use two first order elements (the unit square > >> >> >> divided into two triangles), the resulting solution has > >> >> >> > >> >> >> L2 error = 1.75514e+08 > >> >> >> u = [-175513825.75680602, -175513825.66302037, > >> >> >> -175513825.48390722, -175513824.84436429] > >> >> >> > >> >> >> This looks rather a lot like the null space isn't getting through. > >> >> >> I > >> >> >> am creating the constant nullspace with > >> >> >> > >> >> >> MatNullSpace null; > >> >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); > >> >> >> CHECK(MatSetNullSpace(m,null)); > >> >> >> CHECK(MatNullSpaceDestroy(&null)); > >> >> >> > >> >> >> If I pass "-ksp_view -mat_view", I get the following. The matrix > >> >> >> entries seem right (they do indeed have the constant nullspace), > and > >> >> >> ksp_view shows that a nullspace is attached. Is attaching the > >> >> >> nullspace to the matrix with MatSetNullSpace enough, or do I need > to > >> >> >> additionally attach it to the KSP object? > >> >> > > >> >> > > >> >> > 1) I always run with -ksp_monitor_true_residual now when debugging. > >> >> > This > >> >> > can > >> >> > give > >> >> > you an idea whether you have a singular PC, which I suspect > here. > >> >> > > >> >> > 2) Can you try using -pc_type jacobi? I think ILU might go crazy > on a > >> >> > deficient matrix. > >> >> > >> >> Here are results with -ksp_monitor_true_residual -pc_type none: > >> >> > >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 > >> >> http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 > >> > > >> > > >> > Okay, if you have an inconsistent RHS I do not think that > true_residual > >> > will work > >> > since it uses the unprojected b, but the solve should be fine. > >> > >> I still don't understand why the atol version is able to drift so far > >> away from zero mean, even after tens of thousands of iterations. If > >> KSP sees a null space on the matrix, shouldn't it project that null > >> space out of the *linear system* residual and also out of solution on > >> each iteration? Even if it is only projecting out of the solution > >> delta, how can null space errors be accumulating? > > > > > > Both the KSP and Mat show that the null space is set, so everything > should > > work fine, > > and at this point its no longer DMPlex that is in control, its standard > > PETSc. > > > > We have reached the limit of usefu talking. Something is obviously wrong > with the code, > > but since this routinely works in PETSc examples. In situations like > these I think we need > > to follow the execution in the debugger to see what is wrong..You can > look at Vec values > > in the debugger using > > > > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n > > > > and I look at DMPlex things with > > > > (gdb) p ((DM_Plex*) dm->data)->coneSection > > > > etc. > > Thanks, I appreciate the help. It looks like there were at least two > different problems: > > 1. The boundary FE I was creating had the same dimension as the > interior FE (instead of codimension 1), due to misreading ex12 even > though I had correctly refactored it. I added a dimension consistency > check to my code, but I can do this in DMPlexComputeResidualFEM as > well to catch future user errors. > > 2. Even after fixing the dimensions, my boundary functions in PetscFEM > are getting x values both inside and completely outside the domain. > Almost certainly more user error, but hopefully also something I can > add a check for in petsc once I localize it. This could be my bug. The test I have for ex12 is the variable coefficient problem with div (x + y) grad u = f. This seems to check between the analytic and field versions, meaning that the x coming into f1() matches the x I used to make the field, and my exact solution. Matt > > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Sat Nov 23 14:04:18 2013 From: irving at naml.us (Geoffrey Irving) Date: Sat, 23 Nov 2013 12:04:18 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sat, Nov 23, 2013 at 10:11 AM, Matthew Knepley wrote: > On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving wrote: >> >> On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley >> wrote: >> > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving wrote: >> >> >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley >> >> wrote: >> >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving >> >> > wrote: >> >> >> >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley >> >> >> wrote: >> >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving >> >> >> > wrote: >> >> >> >> >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which works with >> >> >> >> Dirichlet boundary conditions, but it's breaking for me with >> >> >> >> Neumann >> >> >> >> conditions. In particular, with Neumann conditions I get results >> >> >> >> which explode even though I believe I am setting a constant >> >> >> >> nullspace. >> >> >> >> >> >> >> >> For example, if I use two first order elements (the unit square >> >> >> >> divided into two triangles), the resulting solution has >> >> >> >> >> >> >> >> L2 error = 1.75514e+08 >> >> >> >> u = [-175513825.75680602, -175513825.66302037, >> >> >> >> -175513825.48390722, -175513824.84436429] >> >> >> >> >> >> >> >> This looks rather a lot like the null space isn't getting >> >> >> >> through. >> >> >> >> I >> >> >> >> am creating the constant nullspace with >> >> >> >> >> >> >> >> MatNullSpace null; >> >> >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); >> >> >> >> CHECK(MatSetNullSpace(m,null)); >> >> >> >> CHECK(MatNullSpaceDestroy(&null)); >> >> >> >> >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. The matrix >> >> >> >> entries seem right (they do indeed have the constant nullspace), >> >> >> >> and >> >> >> >> ksp_view shows that a nullspace is attached. Is attaching the >> >> >> >> nullspace to the matrix with MatSetNullSpace enough, or do I need >> >> >> >> to >> >> >> >> additionally attach it to the KSP object? >> >> >> > >> >> >> > >> >> >> > 1) I always run with -ksp_monitor_true_residual now when >> >> >> > debugging. >> >> >> > This >> >> >> > can >> >> >> > give >> >> >> > you an idea whether you have a singular PC, which I suspect >> >> >> > here. >> >> >> > >> >> >> > 2) Can you try using -pc_type jacobi? I think ILU might go crazy >> >> >> > on a >> >> >> > deficient matrix. >> >> >> >> >> >> Here are results with -ksp_monitor_true_residual -pc_type none: >> >> >> >> >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 >> >> >> http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 >> >> > >> >> > >> >> > Okay, if you have an inconsistent RHS I do not think that >> >> > true_residual >> >> > will work >> >> > since it uses the unprojected b, but the solve should be fine. >> >> >> >> I still don't understand why the atol version is able to drift so far >> >> away from zero mean, even after tens of thousands of iterations. If >> >> KSP sees a null space on the matrix, shouldn't it project that null >> >> space out of the *linear system* residual and also out of solution on >> >> each iteration? Even if it is only projecting out of the solution >> >> delta, how can null space errors be accumulating? >> > >> > >> > Both the KSP and Mat show that the null space is set, so everything >> > should >> > work fine, >> > and at this point its no longer DMPlex that is in control, its standard >> > PETSc. >> > >> > We have reached the limit of usefu talking. Something is obviously wrong >> > with the code, >> > but since this routinely works in PETSc examples. In situations like >> > these I think we need >> > to follow the execution in the debugger to see what is wrong..You can >> > look at Vec values >> > in the debugger using >> > >> > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n >> > >> > and I look at DMPlex things with >> > >> > (gdb) p ((DM_Plex*) dm->data)->coneSection >> > >> > etc. >> >> Thanks, I appreciate the help. It looks like there were at least two >> different problems: >> >> 1. The boundary FE I was creating had the same dimension as the >> interior FE (instead of codimension 1), due to misreading ex12 even >> though I had correctly refactored it. I added a dimension consistency >> check to my code, but I can do this in DMPlexComputeResidualFEM as >> well to catch future user errors. >> >> 2. Even after fixing the dimensions, my boundary functions in PetscFEM >> are getting x values both inside and completely outside the domain. >> Almost certainly more user error, but hopefully also something I can >> add a check for in petsc once I localize it. > > This could be my bug. The test I have for ex12 is the variable coefficient problem > with div (x + y) grad u = f. This seems to check between the analytic and field > versions, meaning that the x coming into f1() matches the x I used to make the > field, and my exact solution. It does seem to happen with stock snes ex12: branch: irving/assert-ex12-in-box1 % mpiexec -host localhost -n 1 /home/irving/petsc/debug/lib/ex12-obj/ex12 -run_type test -refinement_limit 0.0 -bc_type neumann -interpolate 1 -petscspace_order 1 -bd_petscspace_order 1 -show_initial -dm_plex_print_fem 1 -dm_view ::ascii_info_detail ... [0]PETSC ERROR: evaluation at point outside unit box: 0 1.25 I'll trace down why this is happening. Geoffrey From knepley at gmail.com Sat Nov 23 14:20:23 2013 From: knepley at gmail.com (Matthew Knepley) Date: Sat, 23 Nov 2013 14:20:23 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sat, Nov 23, 2013 at 2:04 PM, Geoffrey Irving wrote: > On Sat, Nov 23, 2013 at 10:11 AM, Matthew Knepley > wrote: > > On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving > wrote: > >> > >> On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley > >> wrote: > >> > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving > wrote: > >> >> > >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley > >> >> wrote: > >> >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving > >> >> > wrote: > >> >> >> > >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley < > knepley at gmail.com> > >> >> >> wrote: > >> >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving < > irving at naml.us> > >> >> >> > wrote: > >> >> >> >> > >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which works with > >> >> >> >> Dirichlet boundary conditions, but it's breaking for me with > >> >> >> >> Neumann > >> >> >> >> conditions. In particular, with Neumann conditions I get > results > >> >> >> >> which explode even though I believe I am setting a constant > >> >> >> >> nullspace. > >> >> >> >> > >> >> >> >> For example, if I use two first order elements (the unit square > >> >> >> >> divided into two triangles), the resulting solution has > >> >> >> >> > >> >> >> >> L2 error = 1.75514e+08 > >> >> >> >> u = [-175513825.75680602, -175513825.66302037, > >> >> >> >> -175513825.48390722, -175513824.84436429] > >> >> >> >> > >> >> >> >> This looks rather a lot like the null space isn't getting > >> >> >> >> through. > >> >> >> >> I > >> >> >> >> am creating the constant nullspace with > >> >> >> >> > >> >> >> >> MatNullSpace null; > >> >> >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); > >> >> >> >> CHECK(MatSetNullSpace(m,null)); > >> >> >> >> CHECK(MatNullSpaceDestroy(&null)); > >> >> >> >> > >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. The > matrix > >> >> >> >> entries seem right (they do indeed have the constant > nullspace), > >> >> >> >> and > >> >> >> >> ksp_view shows that a nullspace is attached. Is attaching the > >> >> >> >> nullspace to the matrix with MatSetNullSpace enough, or do I > need > >> >> >> >> to > >> >> >> >> additionally attach it to the KSP object? > >> >> >> > > >> >> >> > > >> >> >> > 1) I always run with -ksp_monitor_true_residual now when > >> >> >> > debugging. > >> >> >> > This > >> >> >> > can > >> >> >> > give > >> >> >> > you an idea whether you have a singular PC, which I suspect > >> >> >> > here. > >> >> >> > > >> >> >> > 2) Can you try using -pc_type jacobi? I think ILU might go crazy > >> >> >> > on a > >> >> >> > deficient matrix. > >> >> >> > >> >> >> Here are results with -ksp_monitor_true_residual -pc_type none: > >> >> >> > >> >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 > >> >> >> http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 > >> >> > > >> >> > > >> >> > Okay, if you have an inconsistent RHS I do not think that > >> >> > true_residual > >> >> > will work > >> >> > since it uses the unprojected b, but the solve should be fine. > >> >> > >> >> I still don't understand why the atol version is able to drift so far > >> >> away from zero mean, even after tens of thousands of iterations. If > >> >> KSP sees a null space on the matrix, shouldn't it project that null > >> >> space out of the *linear system* residual and also out of solution on > >> >> each iteration? Even if it is only projecting out of the solution > >> >> delta, how can null space errors be accumulating? > >> > > >> > > >> > Both the KSP and Mat show that the null space is set, so everything > >> > should > >> > work fine, > >> > and at this point its no longer DMPlex that is in control, its > standard > >> > PETSc. > >> > > >> > We have reached the limit of usefu talking. Something is obviously > wrong > >> > with the code, > >> > but since this routinely works in PETSc examples. In situations like > >> > these I think we need > >> > to follow the execution in the debugger to see what is wrong..You can > >> > look at Vec values > >> > in the debugger using > >> > > >> > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n > >> > > >> > and I look at DMPlex things with > >> > > >> > (gdb) p ((DM_Plex*) dm->data)->coneSection > >> > > >> > etc. > >> > >> Thanks, I appreciate the help. It looks like there were at least two > >> different problems: > >> > >> 1. The boundary FE I was creating had the same dimension as the > >> interior FE (instead of codimension 1), due to misreading ex12 even > >> though I had correctly refactored it. I added a dimension consistency > >> check to my code, but I can do this in DMPlexComputeResidualFEM as > >> well to catch future user errors. > >> > >> 2. Even after fixing the dimensions, my boundary functions in PetscFEM > >> are getting x values both inside and completely outside the domain. > >> Almost certainly more user error, but hopefully also something I can > >> add a check for in petsc once I localize it. > > > > This could be my bug. The test I have for ex12 is the variable > coefficient problem > > with div (x + y) grad u = f. This seems to check between the analytic > and field > > versions, meaning that the x coming into f1() matches the x I used to > make the > > field, and my exact solution. > > It does seem to happen with stock snes ex12: > > branch: irving/assert-ex12-in-box1 > > % mpiexec -host localhost -n 1 > /home/irving/petsc/debug/lib/ex12-obj/ex12 -run_type test > -refinement_limit 0.0 -bc_type neumann -interpolate 1 > -petscspace_order 1 -bd_petscspace_order 1 -show_initial > -dm_plex_print_fem 1 -dm_view ::ascii_info_detail > ... > [0]PETSC ERROR: evaluation at point outside unit box: 0 1.25 > > I'll trace down why this is happening. My first guess is a triangle with backwards edge. This could cause the geometry routines to barf. Matt > > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From mc0710 at gmail.com Sat Nov 23 15:38:55 2013 From: mc0710 at gmail.com (Mani Chandra) Date: Sat, 23 Nov 2013 15:38:55 -0600 Subject: [petsc-users] AOSOA configuration using DMDA Message-ID: Hi, Is it possible to use an Arrays of Structs of Arrays (AOSOA) configuration using DMDAs? Something like struct node { float var1[16], var2[16], var3[16]; } Instead of struct node { float var1, var2, var3; } as is the usual way of using DMDAs. The global grid size of say a 2D grid would then decrease from NxN to (N/16)xN I'm interested in doing this for ease of vectorization as described in http://software.intel.com/en-us/articles/memory-layout-transformations Thanks, Mani -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Sat Nov 23 15:48:26 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Sat, 23 Nov 2013 15:48:26 -0600 Subject: [petsc-users] AOSOA configuration using DMDA In-Reply-To: References: Message-ID: <878uwe94d1.fsf@jedbrown.org> Mani Chandra writes: > Hi, > > Is it possible to use an Arrays of Structs of Arrays (AOSOA) configuration > using DMDAs? Something like > > struct node { > float var1[16], var2[16], var3[16]; > } Yes, you can manually manage this dimension/chunking, and use DMDASetBlockFills() so that the resulting matrix retains proper sparsity. Neighbor exchange will not automatically understand the blocks, and you would have to use a different fringe layout if you want to organize data as AoSoA. > Instead of > > struct node { > float var1, var2, var3; > } > > as is the usual way of using DMDAs. > > The global grid size of say a 2D grid would then decrease from NxN to (N/16)xN > > I'm interested in doing this for ease of vectorization as described in > http://software.intel.com/en-us/articles/memory-layout-transformations Note that sparse iterative methods are overwhelmingly limited by memory bandwidth rather than vectorization, so you'll get no speedup here. Heavy optimization of stencil operations requires either unaligned loads or a "roll" operation, at which point the benefit over register transposition fades. So instead of trying to change the global memory alignment, I recommend packing aligned representations at whichever granularity makes sense (in registers, in L1-cache tiles, etc). Make sure to benchmark the real memory access patterns before leaping to conclusions about optimal memory layout. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From irving at naml.us Sat Nov 23 17:44:33 2013 From: irving at naml.us (Geoffrey Irving) Date: Sat, 23 Nov 2013 15:44:33 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sat, Nov 23, 2013 at 12:20 PM, Matthew Knepley wrote: > On Sat, Nov 23, 2013 at 2:04 PM, Geoffrey Irving wrote: >> >> On Sat, Nov 23, 2013 at 10:11 AM, Matthew Knepley >> wrote: >> > On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving >> > wrote: >> >> >> >> On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley >> >> wrote: >> >> > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving >> >> > wrote: >> >> >> >> >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley >> >> >> wrote: >> >> >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving >> >> >> > wrote: >> >> >> >> >> >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley >> >> >> >> >> >> >> >> wrote: >> >> >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving >> >> >> >> > >> >> >> >> > wrote: >> >> >> >> >> >> >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which works with >> >> >> >> >> Dirichlet boundary conditions, but it's breaking for me with >> >> >> >> >> Neumann >> >> >> >> >> conditions. In particular, with Neumann conditions I get >> >> >> >> >> results >> >> >> >> >> which explode even though I believe I am setting a constant >> >> >> >> >> nullspace. >> >> >> >> >> >> >> >> >> >> For example, if I use two first order elements (the unit >> >> >> >> >> square >> >> >> >> >> divided into two triangles), the resulting solution has >> >> >> >> >> >> >> >> >> >> L2 error = 1.75514e+08 >> >> >> >> >> u = [-175513825.75680602, -175513825.66302037, >> >> >> >> >> -175513825.48390722, -175513824.84436429] >> >> >> >> >> >> >> >> >> >> This looks rather a lot like the null space isn't getting >> >> >> >> >> through. >> >> >> >> >> I >> >> >> >> >> am creating the constant nullspace with >> >> >> >> >> >> >> >> >> >> MatNullSpace null; >> >> >> >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); >> >> >> >> >> CHECK(MatSetNullSpace(m,null)); >> >> >> >> >> CHECK(MatNullSpaceDestroy(&null)); >> >> >> >> >> >> >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. The >> >> >> >> >> matrix >> >> >> >> >> entries seem right (they do indeed have the constant >> >> >> >> >> nullspace), >> >> >> >> >> and >> >> >> >> >> ksp_view shows that a nullspace is attached. Is attaching the >> >> >> >> >> nullspace to the matrix with MatSetNullSpace enough, or do I >> >> >> >> >> need >> >> >> >> >> to >> >> >> >> >> additionally attach it to the KSP object? >> >> >> >> > >> >> >> >> > >> >> >> >> > 1) I always run with -ksp_monitor_true_residual now when >> >> >> >> > debugging. >> >> >> >> > This >> >> >> >> > can >> >> >> >> > give >> >> >> >> > you an idea whether you have a singular PC, which I suspect >> >> >> >> > here. >> >> >> >> > >> >> >> >> > 2) Can you try using -pc_type jacobi? I think ILU might go >> >> >> >> > crazy >> >> >> >> > on a >> >> >> >> > deficient matrix. >> >> >> >> >> >> >> >> Here are results with -ksp_monitor_true_residual -pc_type none: >> >> >> >> >> >> >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol 1e-5 >> >> >> >> http://naml.us/random/laplace-atol.txt # with -ksp_atol 1e-5 >> >> >> > >> >> >> > >> >> >> > Okay, if you have an inconsistent RHS I do not think that >> >> >> > true_residual >> >> >> > will work >> >> >> > since it uses the unprojected b, but the solve should be fine. >> >> >> >> >> >> I still don't understand why the atol version is able to drift so >> >> >> far >> >> >> away from zero mean, even after tens of thousands of iterations. If >> >> >> KSP sees a null space on the matrix, shouldn't it project that null >> >> >> space out of the *linear system* residual and also out of solution >> >> >> on >> >> >> each iteration? Even if it is only projecting out of the solution >> >> >> delta, how can null space errors be accumulating? >> >> > >> >> > >> >> > Both the KSP and Mat show that the null space is set, so everything >> >> > should >> >> > work fine, >> >> > and at this point its no longer DMPlex that is in control, its >> >> > standard >> >> > PETSc. >> >> > >> >> > We have reached the limit of usefu talking. Something is obviously >> >> > wrong >> >> > with the code, >> >> > but since this routinely works in PETSc examples. In situations like >> >> > these I think we need >> >> > to follow the execution in the debugger to see what is wrong..You can >> >> > look at Vec values >> >> > in the debugger using >> >> > >> >> > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n >> >> > >> >> > and I look at DMPlex things with >> >> > >> >> > (gdb) p ((DM_Plex*) dm->data)->coneSection >> >> > >> >> > etc. >> >> >> >> Thanks, I appreciate the help. It looks like there were at least two >> >> different problems: >> >> >> >> 1. The boundary FE I was creating had the same dimension as the >> >> interior FE (instead of codimension 1), due to misreading ex12 even >> >> though I had correctly refactored it. I added a dimension consistency >> >> check to my code, but I can do this in DMPlexComputeResidualFEM as >> >> well to catch future user errors. >> >> >> >> 2. Even after fixing the dimensions, my boundary functions in PetscFEM >> >> are getting x values both inside and completely outside the domain. >> >> Almost certainly more user error, but hopefully also something I can >> >> add a check for in petsc once I localize it. >> > >> > This could be my bug. The test I have for ex12 is the variable >> > coefficient problem >> > with div (x + y) grad u = f. This seems to check between the analytic >> > and field >> > versions, meaning that the x coming into f1() matches the x I used to >> > make the >> > field, and my exact solution. >> >> It does seem to happen with stock snes ex12: >> >> branch: irving/assert-ex12-in-box1 >> >> % mpiexec -host localhost -n 1 >> /home/irving/petsc/debug/lib/ex12-obj/ex12 -run_type test >> -refinement_limit 0.0 -bc_type neumann -interpolate 1 >> -petscspace_order 1 -bd_petscspace_order 1 -show_initial >> -dm_plex_print_fem 1 -dm_view ::ascii_info_detail >> ... >> [0]PETSC ERROR: evaluation at point outside unit box: 0 1.25 >> >> I'll trace down why this is happening. > > My first guess is a triangle with backwards edge. This could cause the > geometry routines to barf. I don't think it's edge orientation: it breaks (though at different points) regardless of whether I orient all the edges clockwise or counterclockwise. Also, I would expect bad edge orientation to result in bad normals but not to produce bad quadrature locations (nor bad residuals as long as the user routines don't depend on normal). Specifically, I think the problem is a sign error in DMPlexComputeProjection2Dto1D_Internal. The following patch seems to fix the out of bounds evaluation problem. DMPlexComputeProjection2Dto1D is computing a matrix which maps from the given segment to the canonical segment, and DMPlexComputeLineGeometry_Internal expects a map from the canonical segment to the given segment. snes ex12 passes with and without this change, presumably because the only Neumann test has constants along each box side, and is therefore invariant to this error. Unfortunately, my Laplace test is also invariant to this error, so this bug is unrelated to the earlier problem. Geoffrey -------------------------------------------------- --- a/src/dm/impls/plex/plexgeometry.c +++ b/src/dm/impls/plex/plexgeometry.c @@ -213,8 +213,8 @@ static PetscErrorCode DMPlexComputeProjection2Dto1D_Internal(PetscScalar coords[ const PetscReal r = sqrt(x*x + y*y), c = x/r, s = y/r; PetscFunctionBegin; - R[0] = c; R[1] = s; - R[2] = -s; R[3] = c; + R[0] = c; R[1] = -s; + R[2] = s; R[3] = c; coords[0] = 0.0; coords[1] = r; From knepley at gmail.com Sun Nov 24 07:41:44 2013 From: knepley at gmail.com (Matthew Knepley) Date: Sun, 24 Nov 2013 07:41:44 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sat, Nov 23, 2013 at 5:44 PM, Geoffrey Irving wrote: > On Sat, Nov 23, 2013 at 12:20 PM, Matthew Knepley > wrote: > > On Sat, Nov 23, 2013 at 2:04 PM, Geoffrey Irving wrote: > >> > >> On Sat, Nov 23, 2013 at 10:11 AM, Matthew Knepley > >> wrote: > >> > On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving > >> > wrote: > >> >> > >> >> On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley > >> >> wrote: > >> >> > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving > >> >> > wrote: > >> >> >> > >> >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley < > knepley at gmail.com> > >> >> >> wrote: > >> >> >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving < > irving at naml.us> > >> >> >> > wrote: > >> >> >> >> > >> >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley > >> >> >> >> > >> >> >> >> wrote: > >> >> >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving > >> >> >> >> > > >> >> >> >> > wrote: > >> >> >> >> >> > >> >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which works > with > >> >> >> >> >> Dirichlet boundary conditions, but it's breaking for me with > >> >> >> >> >> Neumann > >> >> >> >> >> conditions. In particular, with Neumann conditions I get > >> >> >> >> >> results > >> >> >> >> >> which explode even though I believe I am setting a constant > >> >> >> >> >> nullspace. > >> >> >> >> >> > >> >> >> >> >> For example, if I use two first order elements (the unit > >> >> >> >> >> square > >> >> >> >> >> divided into two triangles), the resulting solution has > >> >> >> >> >> > >> >> >> >> >> L2 error = 1.75514e+08 > >> >> >> >> >> u = [-175513825.75680602, -175513825.66302037, > >> >> >> >> >> -175513825.48390722, -175513824.84436429] > >> >> >> >> >> > >> >> >> >> >> This looks rather a lot like the null space isn't getting > >> >> >> >> >> through. > >> >> >> >> >> I > >> >> >> >> >> am creating the constant nullspace with > >> >> >> >> >> > >> >> >> >> >> MatNullSpace null; > >> >> >> >> >> > CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); > >> >> >> >> >> CHECK(MatSetNullSpace(m,null)); > >> >> >> >> >> CHECK(MatNullSpaceDestroy(&null)); > >> >> >> >> >> > >> >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. The > >> >> >> >> >> matrix > >> >> >> >> >> entries seem right (they do indeed have the constant > >> >> >> >> >> nullspace), > >> >> >> >> >> and > >> >> >> >> >> ksp_view shows that a nullspace is attached. Is attaching > the > >> >> >> >> >> nullspace to the matrix with MatSetNullSpace enough, or do I > >> >> >> >> >> need > >> >> >> >> >> to > >> >> >> >> >> additionally attach it to the KSP object? > >> >> >> >> > > >> >> >> >> > > >> >> >> >> > 1) I always run with -ksp_monitor_true_residual now when > >> >> >> >> > debugging. > >> >> >> >> > This > >> >> >> >> > can > >> >> >> >> > give > >> >> >> >> > you an idea whether you have a singular PC, which I > suspect > >> >> >> >> > here. > >> >> >> >> > > >> >> >> >> > 2) Can you try using -pc_type jacobi? I think ILU might go > >> >> >> >> > crazy > >> >> >> >> > on a > >> >> >> >> > deficient matrix. > >> >> >> >> > >> >> >> >> Here are results with -ksp_monitor_true_residual -pc_type none: > >> >> >> >> > >> >> >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol > 1e-5 > >> >> >> >> http://naml.us/random/laplace-atol.txt # with -ksp_atol > 1e-5 > >> >> >> > > >> >> >> > > >> >> >> > Okay, if you have an inconsistent RHS I do not think that > >> >> >> > true_residual > >> >> >> > will work > >> >> >> > since it uses the unprojected b, but the solve should be fine. > >> >> >> > >> >> >> I still don't understand why the atol version is able to drift so > >> >> >> far > >> >> >> away from zero mean, even after tens of thousands of iterations. > If > >> >> >> KSP sees a null space on the matrix, shouldn't it project that > null > >> >> >> space out of the *linear system* residual and also out of solution > >> >> >> on > >> >> >> each iteration? Even if it is only projecting out of the solution > >> >> >> delta, how can null space errors be accumulating? > >> >> > > >> >> > > >> >> > Both the KSP and Mat show that the null space is set, so everything > >> >> > should > >> >> > work fine, > >> >> > and at this point its no longer DMPlex that is in control, its > >> >> > standard > >> >> > PETSc. > >> >> > > >> >> > We have reached the limit of usefu talking. Something is obviously > >> >> > wrong > >> >> > with the code, > >> >> > but since this routinely works in PETSc examples. In situations > like > >> >> > these I think we need > >> >> > to follow the execution in the debugger to see what is wrong..You > can > >> >> > look at Vec values > >> >> > in the debugger using > >> >> > > >> >> > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n > >> >> > > >> >> > and I look at DMPlex things with > >> >> > > >> >> > (gdb) p ((DM_Plex*) dm->data)->coneSection > >> >> > > >> >> > etc. > >> >> > >> >> Thanks, I appreciate the help. It looks like there were at least two > >> >> different problems: > >> >> > >> >> 1. The boundary FE I was creating had the same dimension as the > >> >> interior FE (instead of codimension 1), due to misreading ex12 even > >> >> though I had correctly refactored it. I added a dimension > consistency > >> >> check to my code, but I can do this in DMPlexComputeResidualFEM as > >> >> well to catch future user errors. > >> >> > >> >> 2. Even after fixing the dimensions, my boundary functions in > PetscFEM > >> >> are getting x values both inside and completely outside the domain. > >> >> Almost certainly more user error, but hopefully also something I can > >> >> add a check for in petsc once I localize it. > >> > > >> > This could be my bug. The test I have for ex12 is the variable > >> > coefficient problem > >> > with div (x + y) grad u = f. This seems to check between the analytic > >> > and field > >> > versions, meaning that the x coming into f1() matches the x I used to > >> > make the > >> > field, and my exact solution. > >> > >> It does seem to happen with stock snes ex12: > >> > >> branch: irving/assert-ex12-in-box1 > >> > >> % mpiexec -host localhost -n 1 > >> /home/irving/petsc/debug/lib/ex12-obj/ex12 -run_type test > >> -refinement_limit 0.0 -bc_type neumann -interpolate 1 > >> -petscspace_order 1 -bd_petscspace_order 1 -show_initial > >> -dm_plex_print_fem 1 -dm_view ::ascii_info_detail > >> ... > >> [0]PETSC ERROR: evaluation at point outside unit box: 0 1.25 > >> > >> I'll trace down why this is happening. > > > > My first guess is a triangle with backwards edge. This could cause the > > geometry routines to barf. > > I don't think it's edge orientation: it breaks (though at different > points) regardless of whether I orient all the edges clockwise or > counterclockwise. Also, I would expect bad edge orientation to result > in bad normals but not to produce bad quadrature locations (nor bad > residuals as long as the user routines don't depend on normal). > > Specifically, I think the problem is a sign error in > DMPlexComputeProjection2Dto1D_Internal. The following patch seems to > fix the out of bounds evaluation problem. > DMPlexComputeProjection2Dto1D is computing a matrix which maps from > the given segment to the canonical segment, and > DMPlexComputeLineGeometry_Internal expects a map from the canonical > segment to the given segment. > > snes ex12 passes with and without this change, presumably because the > only Neumann test has constants along each box side, and is therefore > invariant to this error. > When I replaced my kludge with code using the normal explicitly, I get the same error as you do. You fix is correct, and I checked it into knepley/fix-fem-bd-integrate https://bitbucket.org/petsc/petsc/branch/knepley/fix-fem-bd-integrate along with other fixes that I think get Neumann all the way correct in ex12. > Unfortunately, my Laplace test is also invariant to this error, so > this bug is unrelated to the earlier problem. > It could be one of the other fixes I made. Could you run again? Thanks, Matt > Geoffrey > > -------------------------------------------------- > > --- a/src/dm/impls/plex/plexgeometry.c > +++ b/src/dm/impls/plex/plexgeometry.c > @@ -213,8 +213,8 @@ static PetscErrorCode > DMPlexComputeProjection2Dto1D_Internal(PetscScalar coords[ > const PetscReal r = sqrt(x*x + y*y), c = x/r, s = y/r; > > PetscFunctionBegin; > - R[0] = c; R[1] = s; > - R[2] = -s; R[3] = c; > + R[0] = c; R[1] = -s; > + R[2] = s; R[3] = c; > coords[0] = 0.0; > coords[1] = r; > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From mc0710 at gmail.com Sun Nov 24 17:25:41 2013 From: mc0710 at gmail.com (Mani Chandra) Date: Sun, 24 Nov 2013 17:25:41 -0600 Subject: [petsc-users] AOSOA configuration using DMDA In-Reply-To: <878uwe94d1.fsf@jedbrown.org> References: <878uwe94d1.fsf@jedbrown.org> Message-ID: Could you elaborate a bit on what you mean by packing aligned representations at some granularity? I thought this was what the AOSOA configuration does: packing in variables at the aligned SIMD width. Do you mean loop blocking with each block fitting into the L1 cache? On Sat, Nov 23, 2013 at 3:48 PM, Jed Brown wrote: > Mani Chandra writes: > > > Hi, > > > > Is it possible to use an Arrays of Structs of Arrays (AOSOA) > configuration > > using DMDAs? Something like > > > > struct node { > > float var1[16], var2[16], var3[16]; > > } > > Yes, you can manually manage this dimension/chunking, and use > DMDASetBlockFills() so that the resulting matrix retains proper > sparsity. Neighbor exchange will not automatically understand the > blocks, and you would have to use a different fringe layout if you want > to organize data as AoSoA. > > > Instead of > > > > struct node { > > float var1, var2, var3; > > } > > > > as is the usual way of using DMDAs. > > > > The global grid size of say a 2D grid would then decrease from NxN to > (N/16)xN > > > > I'm interested in doing this for ease of vectorization as described in > > http://software.intel.com/en-us/articles/memory-layout-transformations > > Note that sparse iterative methods are overwhelmingly limited by memory > bandwidth rather than vectorization, so you'll get no speedup here. > Heavy optimization of stencil operations requires either unaligned loads > or a "roll" operation, at which point the benefit over register > transposition fades. So instead of trying to change the global memory > alignment, I recommend packing aligned representations at whichever > granularity makes sense (in registers, in L1-cache tiles, etc). Make > sure to benchmark the real memory access patterns before leaping to > conclusions about optimal memory layout. > -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Sun Nov 24 18:06:04 2013 From: irving at naml.us (Geoffrey Irving) Date: Sun, 24 Nov 2013 16:06:04 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sun, Nov 24, 2013 at 5:41 AM, Matthew Knepley wrote: > On Sat, Nov 23, 2013 at 5:44 PM, Geoffrey Irving wrote: >> >> On Sat, Nov 23, 2013 at 12:20 PM, Matthew Knepley >> wrote: >> > On Sat, Nov 23, 2013 at 2:04 PM, Geoffrey Irving wrote: >> >> >> >> On Sat, Nov 23, 2013 at 10:11 AM, Matthew Knepley >> >> wrote: >> >> > On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving >> >> > wrote: >> >> >> >> >> >> On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley >> >> >> wrote: >> >> >> > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving >> >> >> > wrote: >> >> >> >> >> >> >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley >> >> >> >> >> >> >> >> wrote: >> >> >> >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving >> >> >> >> > >> >> >> >> > wrote: >> >> >> >> >> >> >> >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley >> >> >> >> >> >> >> >> >> >> wrote: >> >> >> >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving >> >> >> >> >> > >> >> >> >> >> > wrote: >> >> >> >> >> >> >> >> >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which works >> >> >> >> >> >> with >> >> >> >> >> >> Dirichlet boundary conditions, but it's breaking for me >> >> >> >> >> >> with >> >> >> >> >> >> Neumann >> >> >> >> >> >> conditions. In particular, with Neumann conditions I get >> >> >> >> >> >> results >> >> >> >> >> >> which explode even though I believe I am setting a constant >> >> >> >> >> >> nullspace. >> >> >> >> >> >> >> >> >> >> >> >> For example, if I use two first order elements (the unit >> >> >> >> >> >> square >> >> >> >> >> >> divided into two triangles), the resulting solution has >> >> >> >> >> >> >> >> >> >> >> >> L2 error = 1.75514e+08 >> >> >> >> >> >> u = [-175513825.75680602, -175513825.66302037, >> >> >> >> >> >> -175513825.48390722, -175513824.84436429] >> >> >> >> >> >> >> >> >> >> >> >> This looks rather a lot like the null space isn't getting >> >> >> >> >> >> through. >> >> >> >> >> >> I >> >> >> >> >> >> am creating the constant nullspace with >> >> >> >> >> >> >> >> >> >> >> >> MatNullSpace null; >> >> >> >> >> >> >> >> >> >> >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); >> >> >> >> >> >> CHECK(MatSetNullSpace(m,null)); >> >> >> >> >> >> CHECK(MatNullSpaceDestroy(&null)); >> >> >> >> >> >> >> >> >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. The >> >> >> >> >> >> matrix >> >> >> >> >> >> entries seem right (they do indeed have the constant >> >> >> >> >> >> nullspace), >> >> >> >> >> >> and >> >> >> >> >> >> ksp_view shows that a nullspace is attached. Is attaching >> >> >> >> >> >> the >> >> >> >> >> >> nullspace to the matrix with MatSetNullSpace enough, or do >> >> >> >> >> >> I >> >> >> >> >> >> need >> >> >> >> >> >> to >> >> >> >> >> >> additionally attach it to the KSP object? >> >> >> >> >> > >> >> >> >> >> > >> >> >> >> >> > 1) I always run with -ksp_monitor_true_residual now when >> >> >> >> >> > debugging. >> >> >> >> >> > This >> >> >> >> >> > can >> >> >> >> >> > give >> >> >> >> >> > you an idea whether you have a singular PC, which I >> >> >> >> >> > suspect >> >> >> >> >> > here. >> >> >> >> >> > >> >> >> >> >> > 2) Can you try using -pc_type jacobi? I think ILU might go >> >> >> >> >> > crazy >> >> >> >> >> > on a >> >> >> >> >> > deficient matrix. >> >> >> >> >> >> >> >> >> >> Here are results with -ksp_monitor_true_residual -pc_type >> >> >> >> >> none: >> >> >> >> >> >> >> >> >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol >> >> >> >> >> 1e-5 >> >> >> >> >> http://naml.us/random/laplace-atol.txt # with -ksp_atol >> >> >> >> >> 1e-5 >> >> >> >> > >> >> >> >> > >> >> >> >> > Okay, if you have an inconsistent RHS I do not think that >> >> >> >> > true_residual >> >> >> >> > will work >> >> >> >> > since it uses the unprojected b, but the solve should be fine. >> >> >> >> >> >> >> >> I still don't understand why the atol version is able to drift so >> >> >> >> far >> >> >> >> away from zero mean, even after tens of thousands of iterations. >> >> >> >> If >> >> >> >> KSP sees a null space on the matrix, shouldn't it project that >> >> >> >> null >> >> >> >> space out of the *linear system* residual and also out of >> >> >> >> solution >> >> >> >> on >> >> >> >> each iteration? Even if it is only projecting out of the >> >> >> >> solution >> >> >> >> delta, how can null space errors be accumulating? >> >> >> > >> >> >> > >> >> >> > Both the KSP and Mat show that the null space is set, so >> >> >> > everything >> >> >> > should >> >> >> > work fine, >> >> >> > and at this point its no longer DMPlex that is in control, its >> >> >> > standard >> >> >> > PETSc. >> >> >> > >> >> >> > We have reached the limit of usefu talking. Something is obviously >> >> >> > wrong >> >> >> > with the code, >> >> >> > but since this routinely works in PETSc examples. In situations >> >> >> > like >> >> >> > these I think we need >> >> >> > to follow the execution in the debugger to see what is wrong..You >> >> >> > can >> >> >> > look at Vec values >> >> >> > in the debugger using >> >> >> > >> >> >> > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n >> >> >> > >> >> >> > and I look at DMPlex things with >> >> >> > >> >> >> > (gdb) p ((DM_Plex*) dm->data)->coneSection >> >> >> > >> >> >> > etc. >> >> >> >> >> >> Thanks, I appreciate the help. It looks like there were at least >> >> >> two >> >> >> different problems: >> >> >> >> >> >> 1. The boundary FE I was creating had the same dimension as the >> >> >> interior FE (instead of codimension 1), due to misreading ex12 even >> >> >> though I had correctly refactored it. I added a dimension >> >> >> consistency >> >> >> check to my code, but I can do this in DMPlexComputeResidualFEM as >> >> >> well to catch future user errors. >> >> >> >> >> >> 2. Even after fixing the dimensions, my boundary functions in >> >> >> PetscFEM >> >> >> are getting x values both inside and completely outside the domain. >> >> >> Almost certainly more user error, but hopefully also something I can >> >> >> add a check for in petsc once I localize it. >> >> > >> >> > This could be my bug. The test I have for ex12 is the variable >> >> > coefficient problem >> >> > with div (x + y) grad u = f. This seems to check between the analytic >> >> > and field >> >> > versions, meaning that the x coming into f1() matches the x I used to >> >> > make the >> >> > field, and my exact solution. >> >> >> >> It does seem to happen with stock snes ex12: >> >> >> >> branch: irving/assert-ex12-in-box1 >> >> >> >> % mpiexec -host localhost -n 1 >> >> /home/irving/petsc/debug/lib/ex12-obj/ex12 -run_type test >> >> -refinement_limit 0.0 -bc_type neumann -interpolate 1 >> >> -petscspace_order 1 -bd_petscspace_order 1 -show_initial >> >> -dm_plex_print_fem 1 -dm_view ::ascii_info_detail >> >> ... >> >> [0]PETSC ERROR: evaluation at point outside unit box: 0 1.25 >> >> >> >> I'll trace down why this is happening. >> > >> > My first guess is a triangle with backwards edge. This could cause the >> > geometry routines to barf. >> >> I don't think it's edge orientation: it breaks (though at different >> points) regardless of whether I orient all the edges clockwise or >> counterclockwise. Also, I would expect bad edge orientation to result >> in bad normals but not to produce bad quadrature locations (nor bad >> residuals as long as the user routines don't depend on normal). >> >> Specifically, I think the problem is a sign error in >> DMPlexComputeProjection2Dto1D_Internal. The following patch seems to >> fix the out of bounds evaluation problem. >> DMPlexComputeProjection2Dto1D is computing a matrix which maps from >> the given segment to the canonical segment, and >> DMPlexComputeLineGeometry_Internal expects a map from the canonical >> segment to the given segment. >> >> snes ex12 passes with and without this change, presumably because the >> only Neumann test has constants along each box side, and is therefore >> invariant to this error. > > > When I replaced my kludge with code using the normal explicitly, I get the > same > error as you do. You fix is correct, and I checked it into > knepley/fix-fem-bd-integrate > > https://bitbucket.org/petsc/petsc/branch/knepley/fix-fem-bd-integrate > > along with other fixes that I think get Neumann all the way correct in ex12. > >> >> Unfortunately, my Laplace test is also invariant to this error, so >> this bug is unrelated to the earlier problem. > > It could be one of the other fixes I made. Could you run again? Much better. All the points and (inward) normals are right, and now my residuals have zero sum as expected. There's still something wrong, since I don't get the exact solution with second order elements, but that's probably an independent problem. I will trace it down. Thanks for the help and fixes! Why did you choose inward pointing normals, by the way? I would have though outward pointing normals are the nearly universal convention. Geoffrey From knepley at gmail.com Sun Nov 24 18:10:29 2013 From: knepley at gmail.com (Matthew Knepley) Date: Sun, 24 Nov 2013 18:10:29 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sun, Nov 24, 2013 at 6:06 PM, Geoffrey Irving wrote: > On Sun, Nov 24, 2013 at 5:41 AM, Matthew Knepley > wrote: > > On Sat, Nov 23, 2013 at 5:44 PM, Geoffrey Irving wrote: > >> > >> On Sat, Nov 23, 2013 at 12:20 PM, Matthew Knepley > >> wrote: > >> > On Sat, Nov 23, 2013 at 2:04 PM, Geoffrey Irving > wrote: > >> >> > >> >> On Sat, Nov 23, 2013 at 10:11 AM, Matthew Knepley > > >> >> wrote: > >> >> > On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving > >> >> > wrote: > >> >> >> > >> >> >> On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley < > knepley at gmail.com> > >> >> >> wrote: > >> >> >> > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving < > irving at naml.us> > >> >> >> > wrote: > >> >> >> >> > >> >> >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley > >> >> >> >> > >> >> >> >> wrote: > >> >> >> >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving > >> >> >> >> > > >> >> >> >> > wrote: > >> >> >> >> >> > >> >> >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley > >> >> >> >> >> > >> >> >> >> >> wrote: > >> >> >> >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving > >> >> >> >> >> > > >> >> >> >> >> > wrote: > >> >> >> >> >> >> > >> >> >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which works > >> >> >> >> >> >> with > >> >> >> >> >> >> Dirichlet boundary conditions, but it's breaking for me > >> >> >> >> >> >> with > >> >> >> >> >> >> Neumann > >> >> >> >> >> >> conditions. In particular, with Neumann conditions I get > >> >> >> >> >> >> results > >> >> >> >> >> >> which explode even though I believe I am setting a > constant > >> >> >> >> >> >> nullspace. > >> >> >> >> >> >> > >> >> >> >> >> >> For example, if I use two first order elements (the unit > >> >> >> >> >> >> square > >> >> >> >> >> >> divided into two triangles), the resulting solution has > >> >> >> >> >> >> > >> >> >> >> >> >> L2 error = 1.75514e+08 > >> >> >> >> >> >> u = [-175513825.75680602, -175513825.66302037, > >> >> >> >> >> >> -175513825.48390722, -175513824.84436429] > >> >> >> >> >> >> > >> >> >> >> >> >> This looks rather a lot like the null space isn't getting > >> >> >> >> >> >> through. > >> >> >> >> >> >> I > >> >> >> >> >> >> am creating the constant nullspace with > >> >> >> >> >> >> > >> >> >> >> >> >> MatNullSpace null; > >> >> >> >> >> >> > >> >> >> >> >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); > >> >> >> >> >> >> CHECK(MatSetNullSpace(m,null)); > >> >> >> >> >> >> CHECK(MatNullSpaceDestroy(&null)); > >> >> >> >> >> >> > >> >> >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. > The > >> >> >> >> >> >> matrix > >> >> >> >> >> >> entries seem right (they do indeed have the constant > >> >> >> >> >> >> nullspace), > >> >> >> >> >> >> and > >> >> >> >> >> >> ksp_view shows that a nullspace is attached. Is > attaching > >> >> >> >> >> >> the > >> >> >> >> >> >> nullspace to the matrix with MatSetNullSpace enough, or > do > >> >> >> >> >> >> I > >> >> >> >> >> >> need > >> >> >> >> >> >> to > >> >> >> >> >> >> additionally attach it to the KSP object? > >> >> >> >> >> > > >> >> >> >> >> > > >> >> >> >> >> > 1) I always run with -ksp_monitor_true_residual now when > >> >> >> >> >> > debugging. > >> >> >> >> >> > This > >> >> >> >> >> > can > >> >> >> >> >> > give > >> >> >> >> >> > you an idea whether you have a singular PC, which I > >> >> >> >> >> > suspect > >> >> >> >> >> > here. > >> >> >> >> >> > > >> >> >> >> >> > 2) Can you try using -pc_type jacobi? I think ILU might go > >> >> >> >> >> > crazy > >> >> >> >> >> > on a > >> >> >> >> >> > deficient matrix. > >> >> >> >> >> > >> >> >> >> >> Here are results with -ksp_monitor_true_residual -pc_type > >> >> >> >> >> none: > >> >> >> >> >> > >> >> >> >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol > >> >> >> >> >> 1e-5 > >> >> >> >> >> http://naml.us/random/laplace-atol.txt # with -ksp_atol > >> >> >> >> >> 1e-5 > >> >> >> >> > > >> >> >> >> > > >> >> >> >> > Okay, if you have an inconsistent RHS I do not think that > >> >> >> >> > true_residual > >> >> >> >> > will work > >> >> >> >> > since it uses the unprojected b, but the solve should be > fine. > >> >> >> >> > >> >> >> >> I still don't understand why the atol version is able to drift > so > >> >> >> >> far > >> >> >> >> away from zero mean, even after tens of thousands of > iterations. > >> >> >> >> If > >> >> >> >> KSP sees a null space on the matrix, shouldn't it project that > >> >> >> >> null > >> >> >> >> space out of the *linear system* residual and also out of > >> >> >> >> solution > >> >> >> >> on > >> >> >> >> each iteration? Even if it is only projecting out of the > >> >> >> >> solution > >> >> >> >> delta, how can null space errors be accumulating? > >> >> >> > > >> >> >> > > >> >> >> > Both the KSP and Mat show that the null space is set, so > >> >> >> > everything > >> >> >> > should > >> >> >> > work fine, > >> >> >> > and at this point its no longer DMPlex that is in control, its > >> >> >> > standard > >> >> >> > PETSc. > >> >> >> > > >> >> >> > We have reached the limit of usefu talking. Something is > obviously > >> >> >> > wrong > >> >> >> > with the code, > >> >> >> > but since this routinely works in PETSc examples. In situations > >> >> >> > like > >> >> >> > these I think we need > >> >> >> > to follow the execution in the debugger to see what is > wrong..You > >> >> >> > can > >> >> >> > look at Vec values > >> >> >> > in the debugger using > >> >> >> > > >> >> >> > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n > >> >> >> > > >> >> >> > and I look at DMPlex things with > >> >> >> > > >> >> >> > (gdb) p ((DM_Plex*) dm->data)->coneSection > >> >> >> > > >> >> >> > etc. > >> >> >> > >> >> >> Thanks, I appreciate the help. It looks like there were at least > >> >> >> two > >> >> >> different problems: > >> >> >> > >> >> >> 1. The boundary FE I was creating had the same dimension as the > >> >> >> interior FE (instead of codimension 1), due to misreading ex12 > even > >> >> >> though I had correctly refactored it. I added a dimension > >> >> >> consistency > >> >> >> check to my code, but I can do this in DMPlexComputeResidualFEM as > >> >> >> well to catch future user errors. > >> >> >> > >> >> >> 2. Even after fixing the dimensions, my boundary functions in > >> >> >> PetscFEM > >> >> >> are getting x values both inside and completely outside the > domain. > >> >> >> Almost certainly more user error, but hopefully also something I > can > >> >> >> add a check for in petsc once I localize it. > >> >> > > >> >> > This could be my bug. The test I have for ex12 is the variable > >> >> > coefficient problem > >> >> > with div (x + y) grad u = f. This seems to check between the > analytic > >> >> > and field > >> >> > versions, meaning that the x coming into f1() matches the x I used > to > >> >> > make the > >> >> > field, and my exact solution. > >> >> > >> >> It does seem to happen with stock snes ex12: > >> >> > >> >> branch: irving/assert-ex12-in-box1 > >> >> > >> >> % mpiexec -host localhost -n 1 > >> >> /home/irving/petsc/debug/lib/ex12-obj/ex12 -run_type test > >> >> -refinement_limit 0.0 -bc_type neumann -interpolate 1 > >> >> -petscspace_order 1 -bd_petscspace_order 1 -show_initial > >> >> -dm_plex_print_fem 1 -dm_view ::ascii_info_detail > >> >> ... > >> >> [0]PETSC ERROR: evaluation at point outside unit box: 0 1.25 > >> >> > >> >> I'll trace down why this is happening. > >> > > >> > My first guess is a triangle with backwards edge. This could cause the > >> > geometry routines to barf. > >> > >> I don't think it's edge orientation: it breaks (though at different > >> points) regardless of whether I orient all the edges clockwise or > >> counterclockwise. Also, I would expect bad edge orientation to result > >> in bad normals but not to produce bad quadrature locations (nor bad > >> residuals as long as the user routines don't depend on normal). > >> > >> Specifically, I think the problem is a sign error in > >> DMPlexComputeProjection2Dto1D_Internal. The following patch seems to > >> fix the out of bounds evaluation problem. > >> DMPlexComputeProjection2Dto1D is computing a matrix which maps from > >> the given segment to the canonical segment, and > >> DMPlexComputeLineGeometry_Internal expects a map from the canonical > >> segment to the given segment. > >> > >> snes ex12 passes with and without this change, presumably because the > >> only Neumann test has constants along each box side, and is therefore > >> invariant to this error. > > > > > > When I replaced my kludge with code using the normal explicitly, I get > the > > same > > error as you do. You fix is correct, and I checked it into > > knepley/fix-fem-bd-integrate > > > > https://bitbucket.org/petsc/petsc/branch/knepley/fix-fem-bd-integrate > > > > along with other fixes that I think get Neumann all the way correct in > ex12. > > > >> > >> Unfortunately, my Laplace test is also invariant to this error, so > >> this bug is unrelated to the earlier problem. > > > > It could be one of the other fixes I made. Could you run again? > > Much better. All the points and (inward) normals are right, and now > my residuals have zero sum as expected. There's still something > wrong, since I don't get the exact solution with second order > elements, but that's probably an independent problem. I will trace it > down. Thanks for the help and fixes! > > Why did you choose inward pointing normals, by the way? I would have > though outward pointing normals are the nearly universal convention. That is a bug fixed in that branch. Did you try next? Thanks Matt > > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Sun Nov 24 18:18:10 2013 From: irving at naml.us (Geoffrey Irving) Date: Sun, 24 Nov 2013 16:18:10 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sun, Nov 24, 2013 at 4:10 PM, Matthew Knepley wrote: > On Sun, Nov 24, 2013 at 6:06 PM, Geoffrey Irving wrote: >> >> On Sun, Nov 24, 2013 at 5:41 AM, Matthew Knepley >> wrote: >> > On Sat, Nov 23, 2013 at 5:44 PM, Geoffrey Irving wrote: >> >> >> >> On Sat, Nov 23, 2013 at 12:20 PM, Matthew Knepley >> >> wrote: >> >> > On Sat, Nov 23, 2013 at 2:04 PM, Geoffrey Irving >> >> > wrote: >> >> >> >> >> >> On Sat, Nov 23, 2013 at 10:11 AM, Matthew Knepley >> >> >> >> >> >> wrote: >> >> >> > On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving >> >> >> > wrote: >> >> >> >> >> >> >> >> On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley >> >> >> >> >> >> >> >> wrote: >> >> >> >> > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving >> >> >> >> > >> >> >> >> > wrote: >> >> >> >> >> >> >> >> >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley >> >> >> >> >> >> >> >> >> >> wrote: >> >> >> >> >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving >> >> >> >> >> > >> >> >> >> >> > wrote: >> >> >> >> >> >> >> >> >> >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley >> >> >> >> >> >> >> >> >> >> >> >> wrote: >> >> >> >> >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving >> >> >> >> >> >> > >> >> >> >> >> >> > wrote: >> >> >> >> >> >> >> >> >> >> >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which >> >> >> >> >> >> >> works >> >> >> >> >> >> >> with >> >> >> >> >> >> >> Dirichlet boundary conditions, but it's breaking for me >> >> >> >> >> >> >> with >> >> >> >> >> >> >> Neumann >> >> >> >> >> >> >> conditions. In particular, with Neumann conditions I >> >> >> >> >> >> >> get >> >> >> >> >> >> >> results >> >> >> >> >> >> >> which explode even though I believe I am setting a >> >> >> >> >> >> >> constant >> >> >> >> >> >> >> nullspace. >> >> >> >> >> >> >> >> >> >> >> >> >> >> For example, if I use two first order elements (the unit >> >> >> >> >> >> >> square >> >> >> >> >> >> >> divided into two triangles), the resulting solution has >> >> >> >> >> >> >> >> >> >> >> >> >> >> L2 error = 1.75514e+08 >> >> >> >> >> >> >> u = [-175513825.75680602, -175513825.66302037, >> >> >> >> >> >> >> -175513825.48390722, -175513824.84436429] >> >> >> >> >> >> >> >> >> >> >> >> >> >> This looks rather a lot like the null space isn't >> >> >> >> >> >> >> getting >> >> >> >> >> >> >> through. >> >> >> >> >> >> >> I >> >> >> >> >> >> >> am creating the constant nullspace with >> >> >> >> >> >> >> >> >> >> >> >> >> >> MatNullSpace null; >> >> >> >> >> >> >> >> >> >> >> >> >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); >> >> >> >> >> >> >> CHECK(MatSetNullSpace(m,null)); >> >> >> >> >> >> >> CHECK(MatNullSpaceDestroy(&null)); >> >> >> >> >> >> >> >> >> >> >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. >> >> >> >> >> >> >> The >> >> >> >> >> >> >> matrix >> >> >> >> >> >> >> entries seem right (they do indeed have the constant >> >> >> >> >> >> >> nullspace), >> >> >> >> >> >> >> and >> >> >> >> >> >> >> ksp_view shows that a nullspace is attached. Is >> >> >> >> >> >> >> attaching >> >> >> >> >> >> >> the >> >> >> >> >> >> >> nullspace to the matrix with MatSetNullSpace enough, or >> >> >> >> >> >> >> do >> >> >> >> >> >> >> I >> >> >> >> >> >> >> need >> >> >> >> >> >> >> to >> >> >> >> >> >> >> additionally attach it to the KSP object? >> >> >> >> >> >> > >> >> >> >> >> >> > >> >> >> >> >> >> > 1) I always run with -ksp_monitor_true_residual now when >> >> >> >> >> >> > debugging. >> >> >> >> >> >> > This >> >> >> >> >> >> > can >> >> >> >> >> >> > give >> >> >> >> >> >> > you an idea whether you have a singular PC, which I >> >> >> >> >> >> > suspect >> >> >> >> >> >> > here. >> >> >> >> >> >> > >> >> >> >> >> >> > 2) Can you try using -pc_type jacobi? I think ILU might >> >> >> >> >> >> > go >> >> >> >> >> >> > crazy >> >> >> >> >> >> > on a >> >> >> >> >> >> > deficient matrix. >> >> >> >> >> >> >> >> >> >> >> >> Here are results with -ksp_monitor_true_residual -pc_type >> >> >> >> >> >> none: >> >> >> >> >> >> >> >> >> >> >> >> http://naml.us/random/laplace-rtol.txt # with -ksp_rtol >> >> >> >> >> >> 1e-5 >> >> >> >> >> >> http://naml.us/random/laplace-atol.txt # with -ksp_atol >> >> >> >> >> >> 1e-5 >> >> >> >> >> > >> >> >> >> >> > >> >> >> >> >> > Okay, if you have an inconsistent RHS I do not think that >> >> >> >> >> > true_residual >> >> >> >> >> > will work >> >> >> >> >> > since it uses the unprojected b, but the solve should be >> >> >> >> >> > fine. >> >> >> >> >> >> >> >> >> >> I still don't understand why the atol version is able to drift >> >> >> >> >> so >> >> >> >> >> far >> >> >> >> >> away from zero mean, even after tens of thousands of >> >> >> >> >> iterations. >> >> >> >> >> If >> >> >> >> >> KSP sees a null space on the matrix, shouldn't it project that >> >> >> >> >> null >> >> >> >> >> space out of the *linear system* residual and also out of >> >> >> >> >> solution >> >> >> >> >> on >> >> >> >> >> each iteration? Even if it is only projecting out of the >> >> >> >> >> solution >> >> >> >> >> delta, how can null space errors be accumulating? >> >> >> >> > >> >> >> >> > >> >> >> >> > Both the KSP and Mat show that the null space is set, so >> >> >> >> > everything >> >> >> >> > should >> >> >> >> > work fine, >> >> >> >> > and at this point its no longer DMPlex that is in control, its >> >> >> >> > standard >> >> >> >> > PETSc. >> >> >> >> > >> >> >> >> > We have reached the limit of usefu talking. Something is >> >> >> >> > obviously >> >> >> >> > wrong >> >> >> >> > with the code, >> >> >> >> > but since this routinely works in PETSc examples. In situations >> >> >> >> > like >> >> >> >> > these I think we need >> >> >> >> > to follow the execution in the debugger to see what is >> >> >> >> > wrong..You >> >> >> >> > can >> >> >> >> > look at Vec values >> >> >> >> > in the debugger using >> >> >> >> > >> >> >> >> > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n >> >> >> >> > >> >> >> >> > and I look at DMPlex things with >> >> >> >> > >> >> >> >> > (gdb) p ((DM_Plex*) dm->data)->coneSection >> >> >> >> > >> >> >> >> > etc. >> >> >> >> >> >> >> >> Thanks, I appreciate the help. It looks like there were at least >> >> >> >> two >> >> >> >> different problems: >> >> >> >> >> >> >> >> 1. The boundary FE I was creating had the same dimension as the >> >> >> >> interior FE (instead of codimension 1), due to misreading ex12 >> >> >> >> even >> >> >> >> though I had correctly refactored it. I added a dimension >> >> >> >> consistency >> >> >> >> check to my code, but I can do this in DMPlexComputeResidualFEM >> >> >> >> as >> >> >> >> well to catch future user errors. >> >> >> >> >> >> >> >> 2. Even after fixing the dimensions, my boundary functions in >> >> >> >> PetscFEM >> >> >> >> are getting x values both inside and completely outside the >> >> >> >> domain. >> >> >> >> Almost certainly more user error, but hopefully also something I >> >> >> >> can >> >> >> >> add a check for in petsc once I localize it. >> >> >> > >> >> >> > This could be my bug. The test I have for ex12 is the variable >> >> >> > coefficient problem >> >> >> > with div (x + y) grad u = f. This seems to check between the >> >> >> > analytic >> >> >> > and field >> >> >> > versions, meaning that the x coming into f1() matches the x I used >> >> >> > to >> >> >> > make the >> >> >> > field, and my exact solution. >> >> >> >> >> >> It does seem to happen with stock snes ex12: >> >> >> >> >> >> branch: irving/assert-ex12-in-box1 >> >> >> >> >> >> % mpiexec -host localhost -n 1 >> >> >> /home/irving/petsc/debug/lib/ex12-obj/ex12 -run_type test >> >> >> -refinement_limit 0.0 -bc_type neumann -interpolate 1 >> >> >> -petscspace_order 1 -bd_petscspace_order 1 -show_initial >> >> >> -dm_plex_print_fem 1 -dm_view ::ascii_info_detail >> >> >> ... >> >> >> [0]PETSC ERROR: evaluation at point outside unit box: 0 1.25 >> >> >> >> >> >> I'll trace down why this is happening. >> >> > >> >> > My first guess is a triangle with backwards edge. This could cause >> >> > the >> >> > geometry routines to barf. >> >> >> >> I don't think it's edge orientation: it breaks (though at different >> >> points) regardless of whether I orient all the edges clockwise or >> >> counterclockwise. Also, I would expect bad edge orientation to result >> >> in bad normals but not to produce bad quadrature locations (nor bad >> >> residuals as long as the user routines don't depend on normal). >> >> >> >> Specifically, I think the problem is a sign error in >> >> DMPlexComputeProjection2Dto1D_Internal. The following patch seems to >> >> fix the out of bounds evaluation problem. >> >> DMPlexComputeProjection2Dto1D is computing a matrix which maps from >> >> the given segment to the canonical segment, and >> >> DMPlexComputeLineGeometry_Internal expects a map from the canonical >> >> segment to the given segment. >> >> >> >> snes ex12 passes with and without this change, presumably because the >> >> only Neumann test has constants along each box side, and is therefore >> >> invariant to this error. >> > >> > >> > When I replaced my kludge with code using the normal explicitly, I get >> > the >> > same >> > error as you do. You fix is correct, and I checked it into >> > knepley/fix-fem-bd-integrate >> > >> > https://bitbucket.org/petsc/petsc/branch/knepley/fix-fem-bd-integrate >> > >> > along with other fixes that I think get Neumann all the way correct in >> > ex12. >> > >> >> >> >> Unfortunately, my Laplace test is also invariant to this error, so >> >> this bug is unrelated to the earlier problem. >> > >> > It could be one of the other fixes I made. Could you run again? >> >> Much better. All the points and (inward) normals are right, and now >> my residuals have zero sum as expected. There's still something >> wrong, since I don't get the exact solution with second order >> elements, but that's probably an independent problem. I will trace it >> down. Thanks for the help and fixes! >> >> Why did you choose inward pointing normals, by the way? I would have >> though outward pointing normals are the nearly universal convention. > > That is a bug fixed in that branch. Did you try next? Oops, I had misread that commit message as "Use inward pointing normal" rather than "Used inward pointing normal". I get outward normals if I reorient the outer boundary edges in my DMPlex to be clockwise, which I guess is what you intended. Closer and closer! Geoffrey From knepley at gmail.com Sun Nov 24 18:21:29 2013 From: knepley at gmail.com (Matthew Knepley) Date: Sun, 24 Nov 2013 18:21:29 -0600 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sun, Nov 24, 2013 at 6:18 PM, Geoffrey Irving wrote: > On Sun, Nov 24, 2013 at 4:10 PM, Matthew Knepley > wrote: > > On Sun, Nov 24, 2013 at 6:06 PM, Geoffrey Irving wrote: > >> > >> On Sun, Nov 24, 2013 at 5:41 AM, Matthew Knepley > >> wrote: > >> > On Sat, Nov 23, 2013 at 5:44 PM, Geoffrey Irving > wrote: > >> >> > >> >> On Sat, Nov 23, 2013 at 12:20 PM, Matthew Knepley > > >> >> wrote: > >> >> > On Sat, Nov 23, 2013 at 2:04 PM, Geoffrey Irving > >> >> > wrote: > >> >> >> > >> >> >> On Sat, Nov 23, 2013 at 10:11 AM, Matthew Knepley > >> >> >> > >> >> >> wrote: > >> >> >> > On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving < > irving at naml.us> > >> >> >> > wrote: > >> >> >> >> > >> >> >> >> On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley > >> >> >> >> > >> >> >> >> wrote: > >> >> >> >> > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving > >> >> >> >> > > >> >> >> >> > wrote: > >> >> >> >> >> > >> >> >> >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley > >> >> >> >> >> > >> >> >> >> >> wrote: > >> >> >> >> >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving > >> >> >> >> >> > > >> >> >> >> >> > wrote: > >> >> >> >> >> >> > >> >> >> >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley > >> >> >> >> >> >> > >> >> >> >> >> >> wrote: > >> >> >> >> >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving > >> >> >> >> >> >> > > >> >> >> >> >> >> > wrote: > >> >> >> >> >> >> >> > >> >> >> >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which > >> >> >> >> >> >> >> works > >> >> >> >> >> >> >> with > >> >> >> >> >> >> >> Dirichlet boundary conditions, but it's breaking for > me > >> >> >> >> >> >> >> with > >> >> >> >> >> >> >> Neumann > >> >> >> >> >> >> >> conditions. In particular, with Neumann conditions I > >> >> >> >> >> >> >> get > >> >> >> >> >> >> >> results > >> >> >> >> >> >> >> which explode even though I believe I am setting a > >> >> >> >> >> >> >> constant > >> >> >> >> >> >> >> nullspace. > >> >> >> >> >> >> >> > >> >> >> >> >> >> >> For example, if I use two first order elements (the > unit > >> >> >> >> >> >> >> square > >> >> >> >> >> >> >> divided into two triangles), the resulting solution > has > >> >> >> >> >> >> >> > >> >> >> >> >> >> >> L2 error = 1.75514e+08 > >> >> >> >> >> >> >> u = [-175513825.75680602, -175513825.66302037, > >> >> >> >> >> >> >> -175513825.48390722, -175513824.84436429] > >> >> >> >> >> >> >> > >> >> >> >> >> >> >> This looks rather a lot like the null space isn't > >> >> >> >> >> >> >> getting > >> >> >> >> >> >> >> through. > >> >> >> >> >> >> >> I > >> >> >> >> >> >> >> am creating the constant nullspace with > >> >> >> >> >> >> >> > >> >> >> >> >> >> >> MatNullSpace null; > >> >> >> >> >> >> >> > >> >> >> >> >> >> >> > CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); > >> >> >> >> >> >> >> CHECK(MatSetNullSpace(m,null)); > >> >> >> >> >> >> >> CHECK(MatNullSpaceDestroy(&null)); > >> >> >> >> >> >> >> > >> >> >> >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. > >> >> >> >> >> >> >> The > >> >> >> >> >> >> >> matrix > >> >> >> >> >> >> >> entries seem right (they do indeed have the constant > >> >> >> >> >> >> >> nullspace), > >> >> >> >> >> >> >> and > >> >> >> >> >> >> >> ksp_view shows that a nullspace is attached. Is > >> >> >> >> >> >> >> attaching > >> >> >> >> >> >> >> the > >> >> >> >> >> >> >> nullspace to the matrix with MatSetNullSpace enough, > or > >> >> >> >> >> >> >> do > >> >> >> >> >> >> >> I > >> >> >> >> >> >> >> need > >> >> >> >> >> >> >> to > >> >> >> >> >> >> >> additionally attach it to the KSP object? > >> >> >> >> >> >> > > >> >> >> >> >> >> > > >> >> >> >> >> >> > 1) I always run with -ksp_monitor_true_residual now > when > >> >> >> >> >> >> > debugging. > >> >> >> >> >> >> > This > >> >> >> >> >> >> > can > >> >> >> >> >> >> > give > >> >> >> >> >> >> > you an idea whether you have a singular PC, which I > >> >> >> >> >> >> > suspect > >> >> >> >> >> >> > here. > >> >> >> >> >> >> > > >> >> >> >> >> >> > 2) Can you try using -pc_type jacobi? I think ILU might > >> >> >> >> >> >> > go > >> >> >> >> >> >> > crazy > >> >> >> >> >> >> > on a > >> >> >> >> >> >> > deficient matrix. > >> >> >> >> >> >> > >> >> >> >> >> >> Here are results with -ksp_monitor_true_residual -pc_type > >> >> >> >> >> >> none: > >> >> >> >> >> >> > >> >> >> >> >> >> http://naml.us/random/laplace-rtol.txt # with > -ksp_rtol > >> >> >> >> >> >> 1e-5 > >> >> >> >> >> >> http://naml.us/random/laplace-atol.txt # with > -ksp_atol > >> >> >> >> >> >> 1e-5 > >> >> >> >> >> > > >> >> >> >> >> > > >> >> >> >> >> > Okay, if you have an inconsistent RHS I do not think that > >> >> >> >> >> > true_residual > >> >> >> >> >> > will work > >> >> >> >> >> > since it uses the unprojected b, but the solve should be > >> >> >> >> >> > fine. > >> >> >> >> >> > >> >> >> >> >> I still don't understand why the atol version is able to > drift > >> >> >> >> >> so > >> >> >> >> >> far > >> >> >> >> >> away from zero mean, even after tens of thousands of > >> >> >> >> >> iterations. > >> >> >> >> >> If > >> >> >> >> >> KSP sees a null space on the matrix, shouldn't it project > that > >> >> >> >> >> null > >> >> >> >> >> space out of the *linear system* residual and also out of > >> >> >> >> >> solution > >> >> >> >> >> on > >> >> >> >> >> each iteration? Even if it is only projecting out of the > >> >> >> >> >> solution > >> >> >> >> >> delta, how can null space errors be accumulating? > >> >> >> >> > > >> >> >> >> > > >> >> >> >> > Both the KSP and Mat show that the null space is set, so > >> >> >> >> > everything > >> >> >> >> > should > >> >> >> >> > work fine, > >> >> >> >> > and at this point its no longer DMPlex that is in control, > its > >> >> >> >> > standard > >> >> >> >> > PETSc. > >> >> >> >> > > >> >> >> >> > We have reached the limit of usefu talking. Something is > >> >> >> >> > obviously > >> >> >> >> > wrong > >> >> >> >> > with the code, > >> >> >> >> > but since this routinely works in PETSc examples. In > situations > >> >> >> >> > like > >> >> >> >> > these I think we need > >> >> >> >> > to follow the execution in the debugger to see what is > >> >> >> >> > wrong..You > >> >> >> >> > can > >> >> >> >> > look at Vec values > >> >> >> >> > in the debugger using > >> >> >> >> > > >> >> >> >> > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n > >> >> >> >> > > >> >> >> >> > and I look at DMPlex things with > >> >> >> >> > > >> >> >> >> > (gdb) p ((DM_Plex*) dm->data)->coneSection > >> >> >> >> > > >> >> >> >> > etc. > >> >> >> >> > >> >> >> >> Thanks, I appreciate the help. It looks like there were at > least > >> >> >> >> two > >> >> >> >> different problems: > >> >> >> >> > >> >> >> >> 1. The boundary FE I was creating had the same dimension as the > >> >> >> >> interior FE (instead of codimension 1), due to misreading ex12 > >> >> >> >> even > >> >> >> >> though I had correctly refactored it. I added a dimension > >> >> >> >> consistency > >> >> >> >> check to my code, but I can do this in DMPlexComputeResidualFEM > >> >> >> >> as > >> >> >> >> well to catch future user errors. > >> >> >> >> > >> >> >> >> 2. Even after fixing the dimensions, my boundary functions in > >> >> >> >> PetscFEM > >> >> >> >> are getting x values both inside and completely outside the > >> >> >> >> domain. > >> >> >> >> Almost certainly more user error, but hopefully also something > I > >> >> >> >> can > >> >> >> >> add a check for in petsc once I localize it. > >> >> >> > > >> >> >> > This could be my bug. The test I have for ex12 is the variable > >> >> >> > coefficient problem > >> >> >> > with div (x + y) grad u = f. This seems to check between the > >> >> >> > analytic > >> >> >> > and field > >> >> >> > versions, meaning that the x coming into f1() matches the x I > used > >> >> >> > to > >> >> >> > make the > >> >> >> > field, and my exact solution. > >> >> >> > >> >> >> It does seem to happen with stock snes ex12: > >> >> >> > >> >> >> branch: irving/assert-ex12-in-box1 > >> >> >> > >> >> >> % mpiexec -host localhost -n 1 > >> >> >> /home/irving/petsc/debug/lib/ex12-obj/ex12 -run_type test > >> >> >> -refinement_limit 0.0 -bc_type neumann -interpolate 1 > >> >> >> -petscspace_order 1 -bd_petscspace_order 1 -show_initial > >> >> >> -dm_plex_print_fem 1 -dm_view ::ascii_info_detail > >> >> >> ... > >> >> >> [0]PETSC ERROR: evaluation at point outside unit box: 0 1.25 > >> >> >> > >> >> >> I'll trace down why this is happening. > >> >> > > >> >> > My first guess is a triangle with backwards edge. This could cause > >> >> > the > >> >> > geometry routines to barf. > >> >> > >> >> I don't think it's edge orientation: it breaks (though at different > >> >> points) regardless of whether I orient all the edges clockwise or > >> >> counterclockwise. Also, I would expect bad edge orientation to > result > >> >> in bad normals but not to produce bad quadrature locations (nor bad > >> >> residuals as long as the user routines don't depend on normal). > >> >> > >> >> Specifically, I think the problem is a sign error in > >> >> DMPlexComputeProjection2Dto1D_Internal. The following patch seems to > >> >> fix the out of bounds evaluation problem. > >> >> DMPlexComputeProjection2Dto1D is computing a matrix which maps from > >> >> the given segment to the canonical segment, and > >> >> DMPlexComputeLineGeometry_Internal expects a map from the canonical > >> >> segment to the given segment. > >> >> > >> >> snes ex12 passes with and without this change, presumably because the > >> >> only Neumann test has constants along each box side, and is therefore > >> >> invariant to this error. > >> > > >> > > >> > When I replaced my kludge with code using the normal explicitly, I get > >> > the > >> > same > >> > error as you do. You fix is correct, and I checked it into > >> > knepley/fix-fem-bd-integrate > >> > > >> > > https://bitbucket.org/petsc/petsc/branch/knepley/fix-fem-bd-integrate > >> > > >> > along with other fixes that I think get Neumann all the way correct in > >> > ex12. > >> > > >> >> > >> >> Unfortunately, my Laplace test is also invariant to this error, so > >> >> this bug is unrelated to the earlier problem. > >> > > >> > It could be one of the other fixes I made. Could you run again? > >> > >> Much better. All the points and (inward) normals are right, and now > >> my residuals have zero sum as expected. There's still something > >> wrong, since I don't get the exact solution with second order > >> elements, but that's probably an independent problem. I will trace it > >> down. Thanks for the help and fixes! > >> > >> Why did you choose inward pointing normals, by the way? I would have > >> though outward pointing normals are the nearly universal convention. > > > > That is a bug fixed in that branch. Did you try next? > > Oops, I had misread that commit message as "Use inward pointing > normal" rather than "Used inward pointing normal". I get outward > normals if I reorient the outer boundary edges in my DMPlex to be > clockwise, which I guess is what you intended. > > Closer and closer! Hmm, no. I use counter-clockwise ordering too. Now I do not understand what is going on. Something is wrong. Matt > > Geoffrey > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Sun Nov 24 18:27:38 2013 From: irving at naml.us (Geoffrey Irving) Date: Sun, 24 Nov 2013 16:27:38 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sun, Nov 24, 2013 at 4:21 PM, Matthew Knepley wrote: > On Sun, Nov 24, 2013 at 6:18 PM, Geoffrey Irving wrote: >> >> On Sun, Nov 24, 2013 at 4:10 PM, Matthew Knepley >> wrote: >> > On Sun, Nov 24, 2013 at 6:06 PM, Geoffrey Irving wrote: >> >> >> >> On Sun, Nov 24, 2013 at 5:41 AM, Matthew Knepley >> >> wrote: >> >> > On Sat, Nov 23, 2013 at 5:44 PM, Geoffrey Irving >> >> > wrote: >> >> >> >> >> >> On Sat, Nov 23, 2013 at 12:20 PM, Matthew Knepley >> >> >> >> >> >> wrote: >> >> >> > On Sat, Nov 23, 2013 at 2:04 PM, Geoffrey Irving >> >> >> > wrote: >> >> >> >> >> >> >> >> On Sat, Nov 23, 2013 at 10:11 AM, Matthew Knepley >> >> >> >> >> >> >> >> wrote: >> >> >> >> > On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving >> >> >> >> > >> >> >> >> > wrote: >> >> >> >> >> >> >> >> >> >> On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley >> >> >> >> >> >> >> >> >> >> wrote: >> >> >> >> >> > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving >> >> >> >> >> > >> >> >> >> >> > wrote: >> >> >> >> >> >> >> >> >> >> >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley >> >> >> >> >> >> >> >> >> >> >> >> wrote: >> >> >> >> >> >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving >> >> >> >> >> >> > >> >> >> >> >> >> > wrote: >> >> >> >> >> >> >> >> >> >> >> >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley >> >> >> >> >> >> >> >> >> >> >> >> >> >> wrote: >> >> >> >> >> >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving >> >> >> >> >> >> >> > >> >> >> >> >> >> >> > wrote: >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which >> >> >> >> >> >> >> >> works >> >> >> >> >> >> >> >> with >> >> >> >> >> >> >> >> Dirichlet boundary conditions, but it's breaking for >> >> >> >> >> >> >> >> me >> >> >> >> >> >> >> >> with >> >> >> >> >> >> >> >> Neumann >> >> >> >> >> >> >> >> conditions. In particular, with Neumann conditions I >> >> >> >> >> >> >> >> get >> >> >> >> >> >> >> >> results >> >> >> >> >> >> >> >> which explode even though I believe I am setting a >> >> >> >> >> >> >> >> constant >> >> >> >> >> >> >> >> nullspace. >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> For example, if I use two first order elements (the >> >> >> >> >> >> >> >> unit >> >> >> >> >> >> >> >> square >> >> >> >> >> >> >> >> divided into two triangles), the resulting solution >> >> >> >> >> >> >> >> has >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> L2 error = 1.75514e+08 >> >> >> >> >> >> >> >> u = [-175513825.75680602, -175513825.66302037, >> >> >> >> >> >> >> >> -175513825.48390722, -175513824.84436429] >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> This looks rather a lot like the null space isn't >> >> >> >> >> >> >> >> getting >> >> >> >> >> >> >> >> through. >> >> >> >> >> >> >> >> I >> >> >> >> >> >> >> >> am creating the constant nullspace with >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> MatNullSpace null; >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); >> >> >> >> >> >> >> >> CHECK(MatSetNullSpace(m,null)); >> >> >> >> >> >> >> >> CHECK(MatNullSpaceDestroy(&null)); >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. >> >> >> >> >> >> >> >> The >> >> >> >> >> >> >> >> matrix >> >> >> >> >> >> >> >> entries seem right (they do indeed have the constant >> >> >> >> >> >> >> >> nullspace), >> >> >> >> >> >> >> >> and >> >> >> >> >> >> >> >> ksp_view shows that a nullspace is attached. Is >> >> >> >> >> >> >> >> attaching >> >> >> >> >> >> >> >> the >> >> >> >> >> >> >> >> nullspace to the matrix with MatSetNullSpace enough, >> >> >> >> >> >> >> >> or >> >> >> >> >> >> >> >> do >> >> >> >> >> >> >> >> I >> >> >> >> >> >> >> >> need >> >> >> >> >> >> >> >> to >> >> >> >> >> >> >> >> additionally attach it to the KSP object? >> >> >> >> >> >> >> > >> >> >> >> >> >> >> > >> >> >> >> >> >> >> > 1) I always run with -ksp_monitor_true_residual now >> >> >> >> >> >> >> > when >> >> >> >> >> >> >> > debugging. >> >> >> >> >> >> >> > This >> >> >> >> >> >> >> > can >> >> >> >> >> >> >> > give >> >> >> >> >> >> >> > you an idea whether you have a singular PC, which >> >> >> >> >> >> >> > I >> >> >> >> >> >> >> > suspect >> >> >> >> >> >> >> > here. >> >> >> >> >> >> >> > >> >> >> >> >> >> >> > 2) Can you try using -pc_type jacobi? I think ILU >> >> >> >> >> >> >> > might >> >> >> >> >> >> >> > go >> >> >> >> >> >> >> > crazy >> >> >> >> >> >> >> > on a >> >> >> >> >> >> >> > deficient matrix. >> >> >> >> >> >> >> >> >> >> >> >> >> >> Here are results with -ksp_monitor_true_residual >> >> >> >> >> >> >> -pc_type >> >> >> >> >> >> >> none: >> >> >> >> >> >> >> >> >> >> >> >> >> >> http://naml.us/random/laplace-rtol.txt # with >> >> >> >> >> >> >> -ksp_rtol >> >> >> >> >> >> >> 1e-5 >> >> >> >> >> >> >> http://naml.us/random/laplace-atol.txt # with >> >> >> >> >> >> >> -ksp_atol >> >> >> >> >> >> >> 1e-5 >> >> >> >> >> >> > >> >> >> >> >> >> > >> >> >> >> >> >> > Okay, if you have an inconsistent RHS I do not think that >> >> >> >> >> >> > true_residual >> >> >> >> >> >> > will work >> >> >> >> >> >> > since it uses the unprojected b, but the solve should be >> >> >> >> >> >> > fine. >> >> >> >> >> >> >> >> >> >> >> >> I still don't understand why the atol version is able to >> >> >> >> >> >> drift >> >> >> >> >> >> so >> >> >> >> >> >> far >> >> >> >> >> >> away from zero mean, even after tens of thousands of >> >> >> >> >> >> iterations. >> >> >> >> >> >> If >> >> >> >> >> >> KSP sees a null space on the matrix, shouldn't it project >> >> >> >> >> >> that >> >> >> >> >> >> null >> >> >> >> >> >> space out of the *linear system* residual and also out of >> >> >> >> >> >> solution >> >> >> >> >> >> on >> >> >> >> >> >> each iteration? Even if it is only projecting out of the >> >> >> >> >> >> solution >> >> >> >> >> >> delta, how can null space errors be accumulating? >> >> >> >> >> > >> >> >> >> >> > >> >> >> >> >> > Both the KSP and Mat show that the null space is set, so >> >> >> >> >> > everything >> >> >> >> >> > should >> >> >> >> >> > work fine, >> >> >> >> >> > and at this point its no longer DMPlex that is in control, >> >> >> >> >> > its >> >> >> >> >> > standard >> >> >> >> >> > PETSc. >> >> >> >> >> > >> >> >> >> >> > We have reached the limit of usefu talking. Something is >> >> >> >> >> > obviously >> >> >> >> >> > wrong >> >> >> >> >> > with the code, >> >> >> >> >> > but since this routinely works in PETSc examples. In >> >> >> >> >> > situations >> >> >> >> >> > like >> >> >> >> >> > these I think we need >> >> >> >> >> > to follow the execution in the debugger to see what is >> >> >> >> >> > wrong..You >> >> >> >> >> > can >> >> >> >> >> > look at Vec values >> >> >> >> >> > in the debugger using >> >> >> >> >> > >> >> >> >> >> > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n >> >> >> >> >> > >> >> >> >> >> > and I look at DMPlex things with >> >> >> >> >> > >> >> >> >> >> > (gdb) p ((DM_Plex*) dm->data)->coneSection >> >> >> >> >> > >> >> >> >> >> > etc. >> >> >> >> >> >> >> >> >> >> Thanks, I appreciate the help. It looks like there were at >> >> >> >> >> least >> >> >> >> >> two >> >> >> >> >> different problems: >> >> >> >> >> >> >> >> >> >> 1. The boundary FE I was creating had the same dimension as >> >> >> >> >> the >> >> >> >> >> interior FE (instead of codimension 1), due to misreading ex12 >> >> >> >> >> even >> >> >> >> >> though I had correctly refactored it. I added a dimension >> >> >> >> >> consistency >> >> >> >> >> check to my code, but I can do this in >> >> >> >> >> DMPlexComputeResidualFEM >> >> >> >> >> as >> >> >> >> >> well to catch future user errors. >> >> >> >> >> >> >> >> >> >> 2. Even after fixing the dimensions, my boundary functions in >> >> >> >> >> PetscFEM >> >> >> >> >> are getting x values both inside and completely outside the >> >> >> >> >> domain. >> >> >> >> >> Almost certainly more user error, but hopefully also something >> >> >> >> >> I >> >> >> >> >> can >> >> >> >> >> add a check for in petsc once I localize it. >> >> >> >> > >> >> >> >> > This could be my bug. The test I have for ex12 is the variable >> >> >> >> > coefficient problem >> >> >> >> > with div (x + y) grad u = f. This seems to check between the >> >> >> >> > analytic >> >> >> >> > and field >> >> >> >> > versions, meaning that the x coming into f1() matches the x I >> >> >> >> > used >> >> >> >> > to >> >> >> >> > make the >> >> >> >> > field, and my exact solution. >> >> >> >> >> >> >> >> It does seem to happen with stock snes ex12: >> >> >> >> >> >> >> >> branch: irving/assert-ex12-in-box1 >> >> >> >> >> >> >> >> % mpiexec -host localhost -n 1 >> >> >> >> /home/irving/petsc/debug/lib/ex12-obj/ex12 -run_type test >> >> >> >> -refinement_limit 0.0 -bc_type neumann -interpolate 1 >> >> >> >> -petscspace_order 1 -bd_petscspace_order 1 -show_initial >> >> >> >> -dm_plex_print_fem 1 -dm_view ::ascii_info_detail >> >> >> >> ... >> >> >> >> [0]PETSC ERROR: evaluation at point outside unit box: 0 1.25 >> >> >> >> >> >> >> >> I'll trace down why this is happening. >> >> >> > >> >> >> > My first guess is a triangle with backwards edge. This could cause >> >> >> > the >> >> >> > geometry routines to barf. >> >> >> >> >> >> I don't think it's edge orientation: it breaks (though at different >> >> >> points) regardless of whether I orient all the edges clockwise or >> >> >> counterclockwise. Also, I would expect bad edge orientation to >> >> >> result >> >> >> in bad normals but not to produce bad quadrature locations (nor bad >> >> >> residuals as long as the user routines don't depend on normal). >> >> >> >> >> >> Specifically, I think the problem is a sign error in >> >> >> DMPlexComputeProjection2Dto1D_Internal. The following patch seems >> >> >> to >> >> >> fix the out of bounds evaluation problem. >> >> >> DMPlexComputeProjection2Dto1D is computing a matrix which maps from >> >> >> the given segment to the canonical segment, and >> >> >> DMPlexComputeLineGeometry_Internal expects a map from the canonical >> >> >> segment to the given segment. >> >> >> >> >> >> snes ex12 passes with and without this change, presumably because >> >> >> the >> >> >> only Neumann test has constants along each box side, and is >> >> >> therefore >> >> >> invariant to this error. >> >> > >> >> > >> >> > When I replaced my kludge with code using the normal explicitly, I >> >> > get >> >> > the >> >> > same >> >> > error as you do. You fix is correct, and I checked it into >> >> > knepley/fix-fem-bd-integrate >> >> > >> >> > >> >> > https://bitbucket.org/petsc/petsc/branch/knepley/fix-fem-bd-integrate >> >> > >> >> > along with other fixes that I think get Neumann all the way correct >> >> > in >> >> > ex12. >> >> > >> >> >> >> >> >> Unfortunately, my Laplace test is also invariant to this error, so >> >> >> this bug is unrelated to the earlier problem. >> >> > >> >> > It could be one of the other fixes I made. Could you run again? >> >> >> >> Much better. All the points and (inward) normals are right, and now >> >> my residuals have zero sum as expected. There's still something >> >> wrong, since I don't get the exact solution with second order >> >> elements, but that's probably an independent problem. I will trace it >> >> down. Thanks for the help and fixes! >> >> >> >> Why did you choose inward pointing normals, by the way? I would have >> >> though outward pointing normals are the nearly universal convention. >> > >> > That is a bug fixed in that branch. Did you try next? >> >> Oops, I had misread that commit message as "Use inward pointing >> normal" rather than "Used inward pointing normal". I get outward >> normals if I reorient the outer boundary edges in my DMPlex to be >> clockwise, which I guess is what you intended. >> >> Closer and closer! > > > Hmm, no. I use counter-clockwise ordering too. Now I do not understand what > is going on. Something is wrong. Closer look coming up. I could easily be wrong. I hate arithmetic in Z_2. Geoffrey From irving at naml.us Sun Nov 24 18:35:23 2013 From: irving at naml.us (Geoffrey Irving) Date: Sun, 24 Nov 2013 16:35:23 -0800 Subject: [petsc-users] neumann failure in my version of snes ex12 In-Reply-To: References: Message-ID: On Sun, Nov 24, 2013 at 4:27 PM, Geoffrey Irving wrote: > On Sun, Nov 24, 2013 at 4:21 PM, Matthew Knepley wrote: >> On Sun, Nov 24, 2013 at 6:18 PM, Geoffrey Irving wrote: >>> >>> On Sun, Nov 24, 2013 at 4:10 PM, Matthew Knepley >>> wrote: >>> > On Sun, Nov 24, 2013 at 6:06 PM, Geoffrey Irving wrote: >>> >> >>> >> On Sun, Nov 24, 2013 at 5:41 AM, Matthew Knepley >>> >> wrote: >>> >> > On Sat, Nov 23, 2013 at 5:44 PM, Geoffrey Irving >>> >> > wrote: >>> >> >> >>> >> >> On Sat, Nov 23, 2013 at 12:20 PM, Matthew Knepley >>> >> >> >>> >> >> wrote: >>> >> >> > On Sat, Nov 23, 2013 at 2:04 PM, Geoffrey Irving >>> >> >> > wrote: >>> >> >> >> >>> >> >> >> On Sat, Nov 23, 2013 at 10:11 AM, Matthew Knepley >>> >> >> >> >>> >> >> >> wrote: >>> >> >> >> > On Sat, Nov 23, 2013 at 12:07 PM, Geoffrey Irving >>> >> >> >> > >>> >> >> >> > wrote: >>> >> >> >> >> >>> >> >> >> >> On Sat, Nov 23, 2013 at 4:43 AM, Matthew Knepley >>> >> >> >> >> >>> >> >> >> >> wrote: >>> >> >> >> >> > On Fri, Nov 22, 2013 at 6:42 PM, Geoffrey Irving >>> >> >> >> >> > >>> >> >> >> >> > wrote: >>> >> >> >> >> >> >>> >> >> >> >> >> On Fri, Nov 22, 2013 at 4:25 PM, Matthew Knepley >>> >> >> >> >> >> >>> >> >> >> >> >> wrote: >>> >> >> >> >> >> > On Fri, Nov 22, 2013 at 6:09 PM, Geoffrey Irving >>> >> >> >> >> >> > >>> >> >> >> >> >> > wrote: >>> >> >> >> >> >> >> >>> >> >> >> >> >> >> On Fri, Nov 22, 2013 at 3:41 PM, Matthew Knepley >>> >> >> >> >> >> >> >>> >> >> >> >> >> >> wrote: >>> >> >> >> >> >> >> > On Fri, Nov 22, 2013 at 5:36 PM, Geoffrey Irving >>> >> >> >> >> >> >> > >>> >> >> >> >> >> >> > wrote: >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> >> >> I have a duplicate of snes ex12 (FEM Poisson) which >>> >> >> >> >> >> >> >> works >>> >> >> >> >> >> >> >> with >>> >> >> >> >> >> >> >> Dirichlet boundary conditions, but it's breaking for >>> >> >> >> >> >> >> >> me >>> >> >> >> >> >> >> >> with >>> >> >> >> >> >> >> >> Neumann >>> >> >> >> >> >> >> >> conditions. In particular, with Neumann conditions I >>> >> >> >> >> >> >> >> get >>> >> >> >> >> >> >> >> results >>> >> >> >> >> >> >> >> which explode even though I believe I am setting a >>> >> >> >> >> >> >> >> constant >>> >> >> >> >> >> >> >> nullspace. >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> >> >> For example, if I use two first order elements (the >>> >> >> >> >> >> >> >> unit >>> >> >> >> >> >> >> >> square >>> >> >> >> >> >> >> >> divided into two triangles), the resulting solution >>> >> >> >> >> >> >> >> has >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> >> >> L2 error = 1.75514e+08 >>> >> >> >> >> >> >> >> u = [-175513825.75680602, -175513825.66302037, >>> >> >> >> >> >> >> >> -175513825.48390722, -175513824.84436429] >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> >> >> This looks rather a lot like the null space isn't >>> >> >> >> >> >> >> >> getting >>> >> >> >> >> >> >> >> through. >>> >> >> >> >> >> >> >> I >>> >> >> >> >> >> >> >> am creating the constant nullspace with >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> >> >> MatNullSpace null; >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> >> >> CHECK(MatNullSpaceCreate(comm(),PETSC_TRUE,0,0,&null)); >>> >> >> >> >> >> >> >> CHECK(MatSetNullSpace(m,null)); >>> >> >> >> >> >> >> >> CHECK(MatNullSpaceDestroy(&null)); >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> >> >> If I pass "-ksp_view -mat_view", I get the following. >>> >> >> >> >> >> >> >> The >>> >> >> >> >> >> >> >> matrix >>> >> >> >> >> >> >> >> entries seem right (they do indeed have the constant >>> >> >> >> >> >> >> >> nullspace), >>> >> >> >> >> >> >> >> and >>> >> >> >> >> >> >> >> ksp_view shows that a nullspace is attached. Is >>> >> >> >> >> >> >> >> attaching >>> >> >> >> >> >> >> >> the >>> >> >> >> >> >> >> >> nullspace to the matrix with MatSetNullSpace enough, >>> >> >> >> >> >> >> >> or >>> >> >> >> >> >> >> >> do >>> >> >> >> >> >> >> >> I >>> >> >> >> >> >> >> >> need >>> >> >> >> >> >> >> >> to >>> >> >> >> >> >> >> >> additionally attach it to the KSP object? >>> >> >> >> >> >> >> > >>> >> >> >> >> >> >> > >>> >> >> >> >> >> >> > 1) I always run with -ksp_monitor_true_residual now >>> >> >> >> >> >> >> > when >>> >> >> >> >> >> >> > debugging. >>> >> >> >> >> >> >> > This >>> >> >> >> >> >> >> > can >>> >> >> >> >> >> >> > give >>> >> >> >> >> >> >> > you an idea whether you have a singular PC, which >>> >> >> >> >> >> >> > I >>> >> >> >> >> >> >> > suspect >>> >> >> >> >> >> >> > here. >>> >> >> >> >> >> >> > >>> >> >> >> >> >> >> > 2) Can you try using -pc_type jacobi? I think ILU >>> >> >> >> >> >> >> > might >>> >> >> >> >> >> >> > go >>> >> >> >> >> >> >> > crazy >>> >> >> >> >> >> >> > on a >>> >> >> >> >> >> >> > deficient matrix. >>> >> >> >> >> >> >> >>> >> >> >> >> >> >> Here are results with -ksp_monitor_true_residual >>> >> >> >> >> >> >> -pc_type >>> >> >> >> >> >> >> none: >>> >> >> >> >> >> >> >>> >> >> >> >> >> >> http://naml.us/random/laplace-rtol.txt # with >>> >> >> >> >> >> >> -ksp_rtol >>> >> >> >> >> >> >> 1e-5 >>> >> >> >> >> >> >> http://naml.us/random/laplace-atol.txt # with >>> >> >> >> >> >> >> -ksp_atol >>> >> >> >> >> >> >> 1e-5 >>> >> >> >> >> >> > >>> >> >> >> >> >> > >>> >> >> >> >> >> > Okay, if you have an inconsistent RHS I do not think that >>> >> >> >> >> >> > true_residual >>> >> >> >> >> >> > will work >>> >> >> >> >> >> > since it uses the unprojected b, but the solve should be >>> >> >> >> >> >> > fine. >>> >> >> >> >> >> >>> >> >> >> >> >> I still don't understand why the atol version is able to >>> >> >> >> >> >> drift >>> >> >> >> >> >> so >>> >> >> >> >> >> far >>> >> >> >> >> >> away from zero mean, even after tens of thousands of >>> >> >> >> >> >> iterations. >>> >> >> >> >> >> If >>> >> >> >> >> >> KSP sees a null space on the matrix, shouldn't it project >>> >> >> >> >> >> that >>> >> >> >> >> >> null >>> >> >> >> >> >> space out of the *linear system* residual and also out of >>> >> >> >> >> >> solution >>> >> >> >> >> >> on >>> >> >> >> >> >> each iteration? Even if it is only projecting out of the >>> >> >> >> >> >> solution >>> >> >> >> >> >> delta, how can null space errors be accumulating? >>> >> >> >> >> > >>> >> >> >> >> > >>> >> >> >> >> > Both the KSP and Mat show that the null space is set, so >>> >> >> >> >> > everything >>> >> >> >> >> > should >>> >> >> >> >> > work fine, >>> >> >> >> >> > and at this point its no longer DMPlex that is in control, >>> >> >> >> >> > its >>> >> >> >> >> > standard >>> >> >> >> >> > PETSc. >>> >> >> >> >> > >>> >> >> >> >> > We have reached the limit of usefu talking. Something is >>> >> >> >> >> > obviously >>> >> >> >> >> > wrong >>> >> >> >> >> > with the code, >>> >> >> >> >> > but since this routinely works in PETSc examples. In >>> >> >> >> >> > situations >>> >> >> >> >> > like >>> >> >> >> >> > these I think we need >>> >> >> >> >> > to follow the execution in the debugger to see what is >>> >> >> >> >> > wrong..You >>> >> >> >> >> > can >>> >> >> >> >> > look at Vec values >>> >> >> >> >> > in the debugger using >>> >> >> >> >> > >>> >> >> >> >> > (gdb) p ((Vec_Seq*) b-.data)->array[0]@v->map.n >>> >> >> >> >> > >>> >> >> >> >> > and I look at DMPlex things with >>> >> >> >> >> > >>> >> >> >> >> > (gdb) p ((DM_Plex*) dm->data)->coneSection >>> >> >> >> >> > >>> >> >> >> >> > etc. >>> >> >> >> >> >>> >> >> >> >> Thanks, I appreciate the help. It looks like there were at >>> >> >> >> >> least >>> >> >> >> >> two >>> >> >> >> >> different problems: >>> >> >> >> >> >>> >> >> >> >> 1. The boundary FE I was creating had the same dimension as >>> >> >> >> >> the >>> >> >> >> >> interior FE (instead of codimension 1), due to misreading ex12 >>> >> >> >> >> even >>> >> >> >> >> though I had correctly refactored it. I added a dimension >>> >> >> >> >> consistency >>> >> >> >> >> check to my code, but I can do this in >>> >> >> >> >> DMPlexComputeResidualFEM >>> >> >> >> >> as >>> >> >> >> >> well to catch future user errors. >>> >> >> >> >> >>> >> >> >> >> 2. Even after fixing the dimensions, my boundary functions in >>> >> >> >> >> PetscFEM >>> >> >> >> >> are getting x values both inside and completely outside the >>> >> >> >> >> domain. >>> >> >> >> >> Almost certainly more user error, but hopefully also something >>> >> >> >> >> I >>> >> >> >> >> can >>> >> >> >> >> add a check for in petsc once I localize it. >>> >> >> >> > >>> >> >> >> > This could be my bug. The test I have for ex12 is the variable >>> >> >> >> > coefficient problem >>> >> >> >> > with div (x + y) grad u = f. This seems to check between the >>> >> >> >> > analytic >>> >> >> >> > and field >>> >> >> >> > versions, meaning that the x coming into f1() matches the x I >>> >> >> >> > used >>> >> >> >> > to >>> >> >> >> > make the >>> >> >> >> > field, and my exact solution. >>> >> >> >> >>> >> >> >> It does seem to happen with stock snes ex12: >>> >> >> >> >>> >> >> >> branch: irving/assert-ex12-in-box1 >>> >> >> >> >>> >> >> >> % mpiexec -host localhost -n 1 >>> >> >> >> /home/irving/petsc/debug/lib/ex12-obj/ex12 -run_type test >>> >> >> >> -refinement_limit 0.0 -bc_type neumann -interpolate 1 >>> >> >> >> -petscspace_order 1 -bd_petscspace_order 1 -show_initial >>> >> >> >> -dm_plex_print_fem 1 -dm_view ::ascii_info_detail >>> >> >> >> ... >>> >> >> >> [0]PETSC ERROR: evaluation at point outside unit box: 0 1.25 >>> >> >> >> >>> >> >> >> I'll trace down why this is happening. >>> >> >> > >>> >> >> > My first guess is a triangle with backwards edge. This could cause >>> >> >> > the >>> >> >> > geometry routines to barf. >>> >> >> >>> >> >> I don't think it's edge orientation: it breaks (though at different >>> >> >> points) regardless of whether I orient all the edges clockwise or >>> >> >> counterclockwise. Also, I would expect bad edge orientation to >>> >> >> result >>> >> >> in bad normals but not to produce bad quadrature locations (nor bad >>> >> >> residuals as long as the user routines don't depend on normal). >>> >> >> >>> >> >> Specifically, I think the problem is a sign error in >>> >> >> DMPlexComputeProjection2Dto1D_Internal. The following patch seems >>> >> >> to >>> >> >> fix the out of bounds evaluation problem. >>> >> >> DMPlexComputeProjection2Dto1D is computing a matrix which maps from >>> >> >> the given segment to the canonical segment, and >>> >> >> DMPlexComputeLineGeometry_Internal expects a map from the canonical >>> >> >> segment to the given segment. >>> >> >> >>> >> >> snes ex12 passes with and without this change, presumably because >>> >> >> the >>> >> >> only Neumann test has constants along each box side, and is >>> >> >> therefore >>> >> >> invariant to this error. >>> >> > >>> >> > >>> >> > When I replaced my kludge with code using the normal explicitly, I >>> >> > get >>> >> > the >>> >> > same >>> >> > error as you do. You fix is correct, and I checked it into >>> >> > knepley/fix-fem-bd-integrate >>> >> > >>> >> > >>> >> > https://bitbucket.org/petsc/petsc/branch/knepley/fix-fem-bd-integrate >>> >> > >>> >> > along with other fixes that I think get Neumann all the way correct >>> >> > in >>> >> > ex12. >>> >> > >>> >> >> >>> >> >> Unfortunately, my Laplace test is also invariant to this error, so >>> >> >> this bug is unrelated to the earlier problem. >>> >> > >>> >> > It could be one of the other fixes I made. Could you run again? >>> >> >>> >> Much better. All the points and (inward) normals are right, and now >>> >> my residuals have zero sum as expected. There's still something >>> >> wrong, since I don't get the exact solution with second order >>> >> elements, but that's probably an independent problem. I will trace it >>> >> down. Thanks for the help and fixes! >>> >> >>> >> Why did you choose inward pointing normals, by the way? I would have >>> >> though outward pointing normals are the nearly universal convention. >>> > >>> > That is a bug fixed in that branch. Did you try next? >>> >>> Oops, I had misread that commit message as "Use inward pointing >>> normal" rather than "Used inward pointing normal". I get outward >>> normals if I reorient the outer boundary edges in my DMPlex to be >>> clockwise, which I guess is what you intended. >>> >>> Closer and closer! >> >> >> Hmm, no. I use counter-clockwise ordering too. Now I do not understand what >> is going on. Something is wrong. > > Closer look coming up. I could easily be wrong. I hate arithmetic in Z_2. Sorry for the noise: everything is inconsistent, I'd just forgotten that boundary halfedges in a corner mesh are counterclockwise around the *outside*, not the inside. I get outward pointing normals if my DMPlex has counterclockwise boundary edges. Geoffrey From iwaddington at gmail.com Sun Nov 24 18:36:36 2013 From: iwaddington at gmail.com (iwaddington .) Date: Sun, 24 Nov 2013 22:36:36 -0200 Subject: [petsc-users] PetscOptionsBegin Message-ID: Hi everybody, I compiled the file src/ksp/ksp/examples/tutorials/ex29.c.htmlin order to see how PetscOptionsBegin and PetscOptionsEnd work, but it looks like nothing happens. There are no options popping on the terminal for me to set. Thanks. -------------- next part -------------- An HTML attachment was scrubbed... URL: From iwaddington at gmail.com Sun Nov 24 20:47:47 2013 From: iwaddington at gmail.com (iwaddington .) Date: Mon, 25 Nov 2013 00:47:47 -0200 Subject: [petsc-users] PetscDrawSetSave Message-ID: Hi, I am using PetscDrawSetSave to save some contour plots in gif files, but it is not saving the contour scales along with the images. Do you know a way to save the contour scales along with the plots ? Thanks. -------------- next part -------------- An HTML attachment was scrubbed... URL: From bsmith at mcs.anl.gov Sun Nov 24 20:50:33 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Sun, 24 Nov 2013 20:50:33 -0600 Subject: [petsc-users] PetscOptionsBegin In-Reply-To: References: Message-ID: Currently the options do not ?pop up? for you to fill in. You can run with -help and it will list the options that are presented and you can set them when you run the program from the command line arguments to the program. Barry On Nov 24, 2013, at 6:36 PM, iwaddington . wrote: > Hi everybody, I compiled the file src/ksp/ksp/examples/tutorials/ex29.c.html in order to see how PetscOptionsBegin and PetscOptionsEnd work, but it looks like nothing happens. There are no options popping on the terminal for me to set. Thanks. From bsmith at mcs.anl.gov Sun Nov 24 22:27:11 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Sun, 24 Nov 2013 22:27:11 -0600 Subject: [petsc-users] PetscDrawSetSave In-Reply-To: References: Message-ID: We didn?t have support for saving these. I have just added it. You can access it in the ?next? branch of the PETSc repository, see https://bitbucket.org/petsc/petsc/wiki/Home Run the code with -draw_save -popup_draw_save (I had named those extra windows popup windows) Please let us know if this does not work. Barry On Nov 24, 2013, at 8:47 PM, iwaddington . wrote: > Hi, I am using PetscDrawSetSave to save some contour plots in gif files, but it is not saving the contour scales along with the images. Do you know a way to save the contour scales along with the plots ? Thanks. From jedbrown at mcs.anl.gov Sun Nov 24 22:30:24 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Sun, 24 Nov 2013 22:30:24 -0600 Subject: [petsc-users] PetscOptionsBegin In-Reply-To: References: Message-ID: <87siul5cin.fsf@jedbrown.org> "iwaddington ." writes: > Hi everybody, I compiled the file > src/ksp/ksp/examples/tutorials/ex29.c.htmlin > order to see how PetscOptionsBegin and PetscOptionsEnd work, but it > looks like nothing happens. There are no options popping on the terminal > for me to set. Run the example with -help. ./ex29 -help -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From laurent.berenguer at gmail.com Mon Nov 25 05:51:00 2013 From: laurent.berenguer at gmail.com (Laurent Berenguer) Date: Mon, 25 Nov 2013 12:51:00 +0100 Subject: [petsc-users] subksp type of PCGASM Message-ID: Hi everyone, I would like to use the Schwarz method with a few processors per subdomain, that is to say with parallel subksps. PCGASM seems to do this if I look at the example ksp/examples/tutorials/ex8g.c. The problem is that I didn't manage to choose the subksp and the subpc. I tried the example ex8g.c with the option -user_set_subdomains_solver to set explicitly KSPBCGS as subksp, and PCJACOBI as subpc and I get the following errors: [0]PETSC ERROR: Null argument, when expecting valid pointer! [0]PETSC ERROR: Null Object: Parameter # 1! ... [0]PETSC ERROR: VecScatterBegin() line 1616 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/vec/vec/utils/vscat.c ... [0]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c Obviously, I am doing something wrong. Do you have any idea/suggestion? Thank you, Laurent Berenguer PS : You can find the full files in attachment, I used runex8g_2 -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Null argument, when expecting valid pointer! [0]PETSC ERROR: Null Object: Parameter # 1! [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 [0]PETSC ERROR: See docs/changes/index.html for recent updates. [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [0]PETSC ERROR: See docs/index.html for manual pages. [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: ./ex8g on a gnu-debug named trielen by berenguer Mon Nov 25 11:39:08 2013 [0]PETSC ERROR: Libraries linked from /home/cdcsp/berenguer/PETSC_INSTALL/lib [0]PETSC ERROR: Configure run at Mon Nov 25 09:58:03 2013 [0]PETSC ERROR: Configure options --prefix=/home/cdcsp/berenguer/PETSC_INSTALL --PETSC_DIR=/home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3 --with-mpi-dir=/home/cdcsp/berenguer/mpichinstall/ --with-shared-libraries=0 --with-scalar-type=real --with-precision=double --with-debugging=yes --download-f-blas-lapack=1 --with-valgrind=1 [0]PETSC ERROR: ------------------------------------------------------------------------ [0]PETSC ERROR: VecScatterBegin() line 1616 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/vec/vec/utils/vscat.c [0]PETSC ERROR: MatMult_MPIAIJ() line 1111 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/impls/aij/mpi/mpiaij.c [0]PETSC ERROR: MatMult() line 2179 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/interface/matrix.c [0]PETSC ERROR: PCApplyBAorAB() line 679 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c [0]PETSC ERROR: KSP_PCApplyBAorAB() line 257 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h [0]PETSC ERROR: KSPGMRESCycle() line 160 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c [0]PETSC ERROR: KSPSolve_GMRES() line 240 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c [0]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c [0]PETSC ERROR: PCApply_GASM() line 535 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/impls/gasm/gasm.c [0]PETSC ERROR: PCApply() line 442 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c [0]PETSC ERROR: KSP_PCApply() line 227 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h [0]PETSC ERROR: KSPInitialResidual() line 64 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itres.c [1]PETSC ERROR: --------------------- Error Message ------------------------------------ [1]PETSC ERROR: Null argument, when expecting valid pointer! [1]PETSC ERROR: Null Object: Parameter # 1! [1]PETSC ERROR: ------------------------------------------------------------------------ [1]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 [1]PETSC ERROR: See docs/changes/index.html for recent updates. [1]PETSC ERROR: See docs/faq.html for hints about trouble shooting. [1]PETSC ERROR: See docs/index.html for manual pages. [1]PETSC ERROR: ------------------------------------------------------------------------ [1]PETSC ERROR: ./ex8g on a gnu-debug named trielen by berenguer Mon Nov 25 11:39:08 2013 [1]PETSC ERROR: Libraries linked from /home/cdcsp/berenguer/PETSC_INSTALL/lib [1]PETSC ERROR: Configure run at Mon Nov 25 09:58:03 2013 [1]PETSC ERROR: Configure options --prefix=/home/cdcsp/berenguer/PETSC_INSTALL --PETSC_DIR=/home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3 --with-mpi-dir=/home/cdcsp/berenguer/mpichinstall/ --with-shared-libraries=0 --with-scalar-type=real --with-precision=double --with-debugging=yes --download-f-blas-lapack=1 --with-valgrind=1 [1]PETSC ERROR: ------------------------------------------------------------------------ [1]PETSC ERROR: VecScatterBegin() line 1616 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/vec/vec/utils/vscat.c [1]PETSC ERROR: MatMult_MPIAIJ() line 1111 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/impls/aij/mpi/mpiaij.c [1]PETSC ERROR: MatMult() line 2179 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/interface/matrix.c [1]PETSC ERROR: PCApplyBAorAB() line 679 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c [1]PETSC ERROR: KSP_PCApplyBAorAB() line 257 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h [1]PETSC ERROR: KSPGMRESCycle() line 160 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c [1]PETSC ERROR: KSPSolve_GMRES() line 240 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c [1]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c [1]PETSC ERROR: PCApply_GASM() line 535 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/impls/gasm/gasm.c [1]PETSC ERROR: PCApply() line 442 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c [1]PETSC ERROR: KSP_PCApply() line 227 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h [1]PETSC ERROR: KSPInitialResidual() line 64 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itres.c [1]PETSC ERROR: KSPSolve_GMRES() line 239 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c [1]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c [1]PETSC ERROR: main() line 272 in src/ksp/ksp/examples/tutorials/ex8g.c application called MPI_Abort(MPI_COMM_WORLD, 85) - process 1 [0]PETSC ERROR: KSPSolve_GMRES() line 239 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c [0]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c [0]PETSC ERROR: main() line 272 in src/ksp/ksp/examples/tutorials/ex8g.c application called MPI_Abort(MPI_COMM_WORLD, 85) - process 0 make: [runex8g_2] Erreur 85 (ignor?e) -------------- next part -------------- A non-text attachment was scrubbed... Name: ex8g.c Type: text/x-csrc Size: 12784 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: makefile Type: application/octet-stream Size: 69713 bytes Desc: not available URL: From info at cinaledlighting.com Mon Nov 25 05:01:34 2013 From: info at cinaledlighting.com (Daniel) Date: Mon, 25 Nov 2013 19:01:34 +0800 (CST) Subject: [petsc-users] ETL/cETL led PetrolStation canopy Message-ID: <1cd8be29.11a21.1428eebe6fa.Coremail.info@cinaledlighting.com> Thank you , Sir/Madam , We have LED lights with highest safety, highest reliability and highest quality . Their application is Gas station , Petrol station , chemical industry, Car parking , etc . 60W, 80W, 100W, 120W and 150W are available . For Highest Safety , our LED light pass almost all international safety standards such as CE, ETL/cETL(USA&Canada&Mexico),SAA(Australia&New Zealand), EN62471(Photobiological Safety), ATEX(Explosion proof), IP66,RoHS, PSE(Japan), FCC , For Highest Quality , we use Best LED(Original CREE LED) , Best Driver?Taiwan Meanwell driver?and best Housing . I send you the ETL/cETL certificate and technical specification. Please check the attachment . If you need more information such as Techincal data, IES file, certificates,etc, please feel free to contact me . CINA LED LIGHTING TECH LIMITED is a professional factory for LED lighting, which specialized in R&D, manufacturing and selling LED lighting . We run our factory under ISO9001 and ERP and RoHS mark system. In order to provide SAFE and HEALTHY LED products , we pass CE , ROHS , ATEX, TUV , PSE , FCC , EN 62471 , IP65 , IP66, SAA, ETL/cETL,etc . Thanks to its ambitious and creative team and our customers? trust, the company started to grow and soon open serveral branch offices in Japan , Israel , Netherlands and USA . By providing excellent sales and consultancy support , CINA develops customers? satisfaction and trust and maintains a solid customer base. In the list of CINA?s customers are presented some of ours parterners like ? Hilton ,KFC , IKEA and Mercedes Benz ,THORN LIGHTING, SHELL , INCOLL, AURORA, BP,etc High Standard, High Quality , High Lumens, High CRI and 5-year Warranty ,we can do it ? ? -------------------------- Take care and have fun ! Daniel Yu General Manager CINA LED LIGHTING TECH LIMITED 8/F ,WanJunHui Building, BaoAn , ShenZhen,China TEL: +86-755-29197481 FAX: +86-755-29197482 Mobile: +86 158 1862 4784 Email: info at cinaledlighting.com or info at cinaled.com MSN: danilyu_86 at hotmail.com SKYPE: hidaniel007 Website: www.cinaLEDlighting.com or www.cinaLED.com -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: ETL&cETL Certificate .pdf Type: application/pdf Size: 314679 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: Technical data 150W Mini340 LED Canopy light.pdf Type: application/pdf Size: 413510 bytes Desc: not available URL: From jedbrown at mcs.anl.gov Mon Nov 25 07:41:02 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Mon, 25 Nov 2013 07:41:02 -0600 Subject: [petsc-users] subksp type of PCGASM In-Reply-To: References: Message-ID: <87eh6461ld.fsf@jedbrown.org> Laurent Berenguer writes: > Hi everyone, > > I would like to use the Schwarz method with a few processors per subdomain, > that is to say with parallel subksps. PCGASM seems to do this if I look at > the example ksp/examples/tutorials/ex8g.c. The problem is that I didn't > manage to choose the subksp and the subpc. I tried the example ex8g.c with > the option -user_set_subdomains_solver to set explicitly KSPBCGS as > subksp, How exactly did you set this subksp? Did you modify the code or change the command line arguments? Did the test run without errors when you don't modify anything? > and PCJACOBI as subpc and I get the following errors: > > [0]PETSC ERROR: Null argument, when expecting valid pointer! > [0]PETSC ERROR: Null Object: Parameter # 1! > ... > [0]PETSC ERROR: VecScatterBegin() line 1616 in > /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/vec/vec/utils/vscat.c > ... > [0]PETSC ERROR: KSPSolve() line 441 in > /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c > > > Obviously, I am doing something wrong. Do you have any idea/suggestion? > > Thank you, > Laurent Berenguer > > PS : You can find the full files in attachment, I used runex8g_2 > [0]PETSC ERROR: --------------------- Error Message ------------------------------------ > [0]PETSC ERROR: Null argument, when expecting valid pointer! > [0]PETSC ERROR: Null Object: Parameter # 1! > [0]PETSC ERROR: ------------------------------------------------------------------------ > [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > [0]PETSC ERROR: See docs/changes/index.html for recent updates. > [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > [0]PETSC ERROR: See docs/index.html for manual pages. > [0]PETSC ERROR: ------------------------------------------------------------------------ > [0]PETSC ERROR: ./ex8g on a gnu-debug named trielen by berenguer Mon Nov 25 11:39:08 2013 > [0]PETSC ERROR: Libraries linked from /home/cdcsp/berenguer/PETSC_INSTALL/lib > [0]PETSC ERROR: Configure run at Mon Nov 25 09:58:03 2013 > [0]PETSC ERROR: Configure options --prefix=/home/cdcsp/berenguer/PETSC_INSTALL --PETSC_DIR=/home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3 --with-mpi-dir=/home/cdcsp/berenguer/mpichinstall/ --with-shared-libraries=0 --with-scalar-type=real --with-precision=double --with-debugging=yes --download-f-blas-lapack=1 --with-valgrind=1 > [0]PETSC ERROR: ------------------------------------------------------------------------ > [0]PETSC ERROR: VecScatterBegin() line 1616 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/vec/vec/utils/vscat.c > [0]PETSC ERROR: MatMult_MPIAIJ() line 1111 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/impls/aij/mpi/mpiaij.c > [0]PETSC ERROR: MatMult() line 2179 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/interface/matrix.c > [0]PETSC ERROR: PCApplyBAorAB() line 679 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c > [0]PETSC ERROR: KSP_PCApplyBAorAB() line 257 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h > [0]PETSC ERROR: KSPGMRESCycle() line 160 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c > [0]PETSC ERROR: KSPSolve_GMRES() line 240 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c > [0]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c > [0]PETSC ERROR: PCApply_GASM() line 535 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/impls/gasm/gasm.c > [0]PETSC ERROR: PCApply() line 442 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c > [0]PETSC ERROR: KSP_PCApply() line 227 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h > [0]PETSC ERROR: KSPInitialResidual() line 64 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itres.c > [1]PETSC ERROR: --------------------- Error Message ------------------------------------ > [1]PETSC ERROR: Null argument, when expecting valid pointer! > [1]PETSC ERROR: Null Object: Parameter # 1! > [1]PETSC ERROR: ------------------------------------------------------------------------ > [1]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 > [1]PETSC ERROR: See docs/changes/index.html for recent updates. > [1]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > [1]PETSC ERROR: See docs/index.html for manual pages. > [1]PETSC ERROR: ------------------------------------------------------------------------ > [1]PETSC ERROR: ./ex8g on a gnu-debug named trielen by berenguer Mon Nov 25 11:39:08 2013 > [1]PETSC ERROR: Libraries linked from /home/cdcsp/berenguer/PETSC_INSTALL/lib > [1]PETSC ERROR: Configure run at Mon Nov 25 09:58:03 2013 > [1]PETSC ERROR: Configure options --prefix=/home/cdcsp/berenguer/PETSC_INSTALL --PETSC_DIR=/home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3 --with-mpi-dir=/home/cdcsp/berenguer/mpichinstall/ --with-shared-libraries=0 --with-scalar-type=real --with-precision=double --with-debugging=yes --download-f-blas-lapack=1 --with-valgrind=1 > [1]PETSC ERROR: ------------------------------------------------------------------------ > [1]PETSC ERROR: VecScatterBegin() line 1616 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/vec/vec/utils/vscat.c > [1]PETSC ERROR: MatMult_MPIAIJ() line 1111 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/impls/aij/mpi/mpiaij.c > [1]PETSC ERROR: MatMult() line 2179 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/interface/matrix.c > [1]PETSC ERROR: PCApplyBAorAB() line 679 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c > [1]PETSC ERROR: KSP_PCApplyBAorAB() line 257 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h > [1]PETSC ERROR: KSPGMRESCycle() line 160 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c > [1]PETSC ERROR: KSPSolve_GMRES() line 240 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c > [1]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c > [1]PETSC ERROR: PCApply_GASM() line 535 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/impls/gasm/gasm.c > [1]PETSC ERROR: PCApply() line 442 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c > [1]PETSC ERROR: KSP_PCApply() line 227 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h > [1]PETSC ERROR: KSPInitialResidual() line 64 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itres.c > [1]PETSC ERROR: KSPSolve_GMRES() line 239 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c > [1]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c > [1]PETSC ERROR: main() line 272 in src/ksp/ksp/examples/tutorials/ex8g.c > application called MPI_Abort(MPI_COMM_WORLD, 85) - process 1 > [0]PETSC ERROR: KSPSolve_GMRES() line 239 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c > [0]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c > [0]PETSC ERROR: main() line 272 in src/ksp/ksp/examples/tutorials/ex8g.c > application called MPI_Abort(MPI_COMM_WORLD, 85) - process 0 > make: [runex8g_2] Erreur 85 (ignor?e) > > static char help[] = "Illustrates use of the preconditioner GASM.\n\ > The Additive Schwarz Method for solving a linear system in parallel with KSP. The\n\ > code indicates the procedure for setting user-defined subdomains. Input\n\ > parameters include:\n\ > -M: Number of mesh points in the x direction\n\ > -N: Number of mesh points in the y direction\n\ > -user_set_subdomain_solvers: User explicitly sets subdomain solvers\n\ > -user_set_subdomains: Use the user-provided subdomain partitioning routine\n\ > With -user_set_subdomains on, the following options are meaningful:\n\ > -Mdomains: Number of subdomains in the x direction \n\ > -Ndomains: Number of subdomains in the y direction \n\ > -overlap: Size of domain overlap in terms of the number of mesh lines in x and y\n\ > General useful options:\n\ > -pc_gasm_print_subdomains: Print the index sets defining the subdomains\n\ > \n"; > > /* > Note: This example focuses on setting the subdomains for the GASM > preconditioner for a problem on a 2D rectangular grid. See ex1.c > and ex2.c for more detailed comments on the basic usage of KSP > (including working with matrices and vectors). > > The GASM preconditioner is fully parallel. The user-space routine > CreateSubdomains2D that computes the domain decomposition is also parallel > and attempts to generate both subdomains straddling processors and multiple > domains per processor. > > > This matrix in this linear system arises from the discretized Laplacian, > and thus is not very interesting in terms of experimenting with variants > of the GASM preconditioner. > */ > > /*T > Concepts: KSP^Additive Schwarz Method (GASM) with user-defined subdomains > Processors: n > T*/ > > /* > Include "petscksp.h" so that we can use KSP solvers. Note that this file > automatically includes: > petscsys.h - base PETSc routines petscvec.h - vectors > petscmat.h - matrices > petscis.h - index sets petscksp.h - Krylov subspace methods > petscviewer.h - viewers petscpc.h - preconditioners > */ > #include > > > > #undef __FUNCT__ > #define __FUNCT__ "main" > int main(int argc,char **args) > { > Vec x,b,u; /* approx solution, RHS, exact solution */ > Mat A; /* linear system matrix */ > KSP ksp; /* linear solver context */ > PC pc; /* PC context */ > IS *inneris,*outeris; /* array of index sets that define the subdomains */ > PetscInt overlap = 1; /* width of subdomain overlap */ > PetscInt Nsub; /* number of subdomains */ > PetscInt m = 15,n = 17; /* mesh dimensions in x- and y- directions */ > PetscInt M = 2,N = 1; /* number of subdomains in x- and y- directions */ > PetscInt i,j,Ii,J,Istart,Iend; > PetscErrorCode ierr; > PetscMPIInt size; > PetscBool flg; > PetscBool user_set_subdomains = PETSC_FALSE; > PetscScalar v, one = 1.0; > PetscReal e; > > PetscInitialize(&argc,&args,(char*)0,help); > ierr = MPI_Comm_size(PETSC_COMM_WORLD,&size);CHKERRQ(ierr); > ierr = PetscOptionsGetInt(NULL,"-M",&m,NULL);CHKERRQ(ierr); > ierr = PetscOptionsGetInt(NULL,"-N",&n,NULL);CHKERRQ(ierr); > ierr = PetscOptionsGetBool(NULL,"-user_set_subdomains",&user_set_subdomains,NULL);CHKERRQ(ierr); > ierr = PetscOptionsGetInt(NULL,"-Mdomains",&M,NULL);CHKERRQ(ierr); > ierr = PetscOptionsGetInt(NULL,"-Ndomains",&N,NULL);CHKERRQ(ierr); > ierr = PetscOptionsGetInt(NULL,"-overlap",&overlap,NULL);CHKERRQ(ierr); > > /* ------------------------------------------------------------------- > Compute the matrix and right-hand-side vector that define > the linear system, Ax = b. > ------------------------------------------------------------------- */ > > /* > Assemble the matrix for the five point stencil, YET AGAIN > */ > ierr = MatCreate(PETSC_COMM_WORLD,&A);CHKERRQ(ierr); > ierr = MatSetSizes(A,PETSC_DECIDE,PETSC_DECIDE,m*n,m*n);CHKERRQ(ierr); > ierr = MatSetFromOptions(A);CHKERRQ(ierr); > ierr = MatSetUp(A);CHKERRQ(ierr); > ierr = MatGetOwnershipRange(A,&Istart,&Iend);CHKERRQ(ierr); > for (Ii=Istart; Ii v = -1.0; i = Ii/n; j = Ii - i*n; > if (i>0) {J = Ii - n; ierr = MatSetValues(A,1,&Ii,1,&J,&v,INSERT_VALUES);CHKERRQ(ierr);} > if (i if (j>0) {J = Ii - 1; ierr = MatSetValues(A,1,&Ii,1,&J,&v,INSERT_VALUES);CHKERRQ(ierr);} > if (j v = 4.0; ierr = MatSetValues(A,1,&Ii,1,&Ii,&v,INSERT_VALUES);CHKERRQ(ierr); > } > ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); > ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); > > /* > Create and set vectors > */ > ierr = VecCreate(PETSC_COMM_WORLD,&b);CHKERRQ(ierr); > ierr = VecSetSizes(b,PETSC_DECIDE,m*n);CHKERRQ(ierr); > ierr = VecSetFromOptions(b);CHKERRQ(ierr); > ierr = VecDuplicate(b,&u);CHKERRQ(ierr); > ierr = VecDuplicate(b,&x);CHKERRQ(ierr); > ierr = VecSet(u,one);CHKERRQ(ierr); > ierr = MatMult(A,u,b);CHKERRQ(ierr); > > /* > Create linear solver context > */ > ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); > > /* > Set operators. Here the matrix that defines the linear system > also serves as the preconditioning matrix. > */ > ierr = KSPSetOperators(ksp,A,A,DIFFERENT_NONZERO_PATTERN);CHKERRQ(ierr); > > /* > Set the default preconditioner for this program to be GASM > */ > ierr = KSPGetPC(ksp,&pc);CHKERRQ(ierr); > ierr = PCSetType(pc,PCGASM);CHKERRQ(ierr); > > /* ------------------------------------------------------------------- > Define the problem decomposition > ------------------------------------------------------------------- */ > > /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > Basic method, should be sufficient for the needs of many users. > - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > > Set the overlap, using the default PETSc decomposition via > PCGASMSetOverlap(pc,overlap); > Could instead use the option -pc_gasm_overlap > > Set the total number of blocks via -pc_gasm_blocks > Note: The GASM default is to use 1 block per processor. To > experiment on a single processor with various overlaps, you > must specify use of multiple blocks! > */ > > /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > More advanced method, setting user-defined subdomains > - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > > Firstly, create index sets that define the subdomains. The utility > routine PCGASMCreateSubdomains2D() is a simple example, which partitions > the 2D grid into MxN subdomains with an optional overlap. > More generally, the user should write a custom routine for a particular > problem geometry. > > Then call PCGASMSetLocalSubdomains() with resulting index sets > to set the subdomains for the GASM preconditioner. > */ > > if (!user_set_subdomains) { /* basic version */ > ierr = PCGASMSetOverlap(pc,overlap);CHKERRQ(ierr); > } else { /* advanced version */ > ierr = PCGASMCreateSubdomains2D(pc, m,n,M,N,1,overlap,&Nsub,&inneris,&outeris);CHKERRQ(ierr); > ierr = PCGASMSetSubdomains(pc,Nsub,inneris,outeris);CHKERRQ(ierr); > flg = PETSC_FALSE; > ierr = PetscOptionsGetBool(NULL,"-subdomain_view",&flg,NULL);CHKERRQ(ierr); > if (flg) { > ierr = PetscPrintf(PETSC_COMM_SELF,"Nmesh points: %D x %D; subdomain partition: %D x %D; overlap: %D; Nsub: %D\n",m,n,M,N,overlap,Nsub);CHKERRQ(ierr); > ierr = PetscPrintf(PETSC_COMM_SELF,"Outer IS:\n");CHKERRQ(ierr); > for (i=0; i ierr = PetscPrintf(PETSC_COMM_SELF," outer IS[%d]\n",i);CHKERRQ(ierr); > ierr = ISView(outeris[i],PETSC_VIEWER_STDOUT_SELF);CHKERRQ(ierr); > } > ierr = PetscPrintf(PETSC_COMM_SELF,"Inner IS:\n");CHKERRQ(ierr); > for (i=0; i ierr = PetscPrintf(PETSC_COMM_SELF," inner IS[%d]\n",i);CHKERRQ(ierr); > ierr = ISView(inneris[i],PETSC_VIEWER_STDOUT_SELF);CHKERRQ(ierr); > } > } > } > > /* ------------------------------------------------------------------- > Set the linear solvers for the subblocks > ------------------------------------------------------------------- */ > > /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > Basic method, should be sufficient for the needs of most users. > - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > > By default, the GASM preconditioner uses the same solver on each > block of the problem. To set the same solver options on all blocks, > use the prefix -sub before the usual PC and KSP options, e.g., > -sub_pc_type -sub_ksp_type -sub_ksp_rtol 1.e-4 > > - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > Advanced method, setting different solvers for various blocks. > - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - > > Note that each block's KSP context is completely independent of > the others, and the full range of uniprocessor KSP options is > available for each block. > > - Use PCGASMGetSubKSP() to extract the array of KSP contexts for > the local blocks. > - See ex7.c for a simple example of setting different linear solvers > for the individual blocks for the block Jacobi method (which is > equivalent to the GASM method with zero overlap). > */ > > flg = PETSC_FALSE; > ierr = PetscOptionsGetBool(NULL,"-user_set_subdomain_solvers",&flg,NULL);CHKERRQ(ierr); > //flg=PETSC_TRUE; > if (flg) { > KSP *subksp; /* array of KSP contexts for local subblocks */ > PetscInt nlocal,first; /* number of local subblocks, first local subblock */ > PC subpc; /* PC context for subblock */ > PetscBool isasm; > > ierr = PetscPrintf(PETSC_COMM_WORLD,"User explicitly sets subdomain solvers.\n");CHKERRQ(ierr); > > /* > Set runtime options > */ > ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); > > /* > Flag an error if PCTYPE is changed from the runtime options > */ > ierr = PetscObjectTypeCompare((PetscObject)pc,PCGASM,&isasm);CHKERRQ(ierr); > if (!isasm) SETERRQ(PETSC_COMM_WORLD,1,"Cannot Change the PCTYPE when manually changing the subdomain solver settings"); > > /* > Call KSPSetUp() to set the block Jacobi data structures (including > creation of an internal KSP context for each block). > > Note: KSPSetUp() MUST be called before PCGASMGetSubKSP(). > */ > ierr = KSPSetUp(ksp);CHKERRQ(ierr); > > /* > Extract the array of KSP contexts for the local blocks > */ > ierr = PCGASMGetSubKSP(pc,&nlocal,&first,&subksp);CHKERRQ(ierr); > /* > Loop over the local blocks, setting various KSP options > for each block. > */ > for (i=0; i ierr = KSPGetPC(subksp[i],&subpc);CHKERRQ(ierr); > ierr = PCSetType(subpc,PCJACOBI);CHKERRQ(ierr); > ierr = KSPSetType(subksp[i],KSPGMRES);CHKERRQ(ierr); > ierr = KSPSetTolerances(subksp[i],1.e-7,PETSC_DEFAULT,PETSC_DEFAULT,PETSC_DEFAULT);CHKERRQ(ierr); > } > } else { > /* > Set runtime options > */ > ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); > } > > /* ------------------------------------------------------------------- > Solve the linear system > ------------------------------------------------------------------- */ > > ierr = KSPSolve(ksp,b,x);CHKERRQ(ierr); > > /* ------------------------------------------------------------------- > Compare result to the exact solution > ------------------------------------------------------------------- */ > ierr = VecAXPY(x,-1.0,u);CHKERRQ(ierr); > ierr = VecNorm(x,NORM_INFINITY, &e);CHKERRQ(ierr); > > flg = PETSC_FALSE; > ierr = PetscOptionsGetBool(NULL,"-print_error",&flg,NULL);CHKERRQ(ierr); > if (flg) { > ierr = PetscPrintf(PETSC_COMM_WORLD, "Infinity norm of the error: %G\n", e);CHKERRQ(ierr); > } > > /* > Free work space. All PETSc objects should be destroyed when they > are no longer needed. > */ > > if (user_set_subdomains) { > ierr = PCGASMDestroySubdomains(Nsub, inneris, outeris);CHKERRQ(ierr); > } > ierr = KSPDestroy(&ksp);CHKERRQ(ierr); > ierr = VecDestroy(&u);CHKERRQ(ierr); > ierr = VecDestroy(&x);CHKERRQ(ierr); > ierr = VecDestroy(&b);CHKERRQ(ierr); > ierr = MatDestroy(&A);CHKERRQ(ierr); > ierr = PetscFinalize(); > return 0; > } -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From laurent.berenguer at gmail.com Mon Nov 25 08:44:17 2013 From: laurent.berenguer at gmail.com (Laurent Berenguer) Date: Mon, 25 Nov 2013 15:44:17 +0100 Subject: [petsc-users] subksp type of PCGASM In-Reply-To: <87eh6461ld.fsf@jedbrown.org> References: <87eh6461ld.fsf@jedbrown.org> Message-ID: >How exactly did you set this subksp? Did you modify the code or change >the command line arguments? Did the test run without errors when you >don't modify anything? I tried both methods, the error is the same in both cases : - I added the option -user_set_subdomain_solvers and modified the code from ierr = PCSetType(subpc,PCILU);CHKERRQ(ierr); to ierr = PCSetType(subpc,PCJACOBI);CHKERRQ(ierr); - Or I only added -sub_ksp_type gmres without changing the code. The example runs correctly when I change nothing. But I'm not sure what is the default subksp. When I change nothing in the code but it seems to be KSPPREONLY. 2013/11/25 Jed Brown : > Laurent Berenguer writes: > >> Hi everyone, >> >> I would like to use the Schwarz method with a few processors per subdomain, >> that is to say with parallel subksps. PCGASM seems to do this if I look at >> the example ksp/examples/tutorials/ex8g.c. The problem is that I didn't >> manage to choose the subksp and the subpc. I tried the example ex8g.c with >> the option -user_set_subdomains_solver to set explicitly KSPBCGS as >> subksp, > > How exactly did you set this subksp? Did you modify the code or change > the command line arguments? Did the test run without errors when you > don't modify anything? > >> and PCJACOBI as subpc and I get the following errors: >> >> [0]PETSC ERROR: Null argument, when expecting valid pointer! >> [0]PETSC ERROR: Null Object: Parameter # 1! >> ... >> [0]PETSC ERROR: VecScatterBegin() line 1616 in >> /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/vec/vec/utils/vscat.c >> ... >> [0]PETSC ERROR: KSPSolve() line 441 in >> /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >> >> >> Obviously, I am doing something wrong. Do you have any idea/suggestion? >> >> Thank you, >> Laurent Berenguer >> >> PS : You can find the full files in attachment, I used runex8g_2 >> [0]PETSC ERROR: --------------------- Error Message ------------------------------------ >> [0]PETSC ERROR: Null argument, when expecting valid pointer! >> [0]PETSC ERROR: Null Object: Parameter # 1! >> [0]PETSC ERROR: ------------------------------------------------------------------------ >> [0]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 >> [0]PETSC ERROR: See docs/changes/index.html for recent updates. >> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >> [0]PETSC ERROR: See docs/index.html for manual pages. >> [0]PETSC ERROR: ------------------------------------------------------------------------ >> [0]PETSC ERROR: ./ex8g on a gnu-debug named trielen by berenguer Mon Nov 25 11:39:08 2013 >> [0]PETSC ERROR: Libraries linked from /home/cdcsp/berenguer/PETSC_INSTALL/lib >> [0]PETSC ERROR: Configure run at Mon Nov 25 09:58:03 2013 >> [0]PETSC ERROR: Configure options --prefix=/home/cdcsp/berenguer/PETSC_INSTALL --PETSC_DIR=/home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3 --with-mpi-dir=/home/cdcsp/berenguer/mpichinstall/ --with-shared-libraries=0 --with-scalar-type=real --with-precision=double --with-debugging=yes --download-f-blas-lapack=1 --with-valgrind=1 >> [0]PETSC ERROR: ------------------------------------------------------------------------ >> [0]PETSC ERROR: VecScatterBegin() line 1616 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/vec/vec/utils/vscat.c >> [0]PETSC ERROR: MatMult_MPIAIJ() line 1111 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/impls/aij/mpi/mpiaij.c >> [0]PETSC ERROR: MatMult() line 2179 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/interface/matrix.c >> [0]PETSC ERROR: PCApplyBAorAB() line 679 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c >> [0]PETSC ERROR: KSP_PCApplyBAorAB() line 257 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h >> [0]PETSC ERROR: KSPGMRESCycle() line 160 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c >> [0]PETSC ERROR: KSPSolve_GMRES() line 240 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c >> [0]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >> [0]PETSC ERROR: PCApply_GASM() line 535 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/impls/gasm/gasm.c >> [0]PETSC ERROR: PCApply() line 442 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c >> [0]PETSC ERROR: KSP_PCApply() line 227 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h >> [0]PETSC ERROR: KSPInitialResidual() line 64 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itres.c >> [1]PETSC ERROR: --------------------- Error Message ------------------------------------ >> [1]PETSC ERROR: Null argument, when expecting valid pointer! >> [1]PETSC ERROR: Null Object: Parameter # 1! >> [1]PETSC ERROR: ------------------------------------------------------------------------ >> [1]PETSC ERROR: Petsc Release Version 3.4.3, Oct, 15, 2013 >> [1]PETSC ERROR: See docs/changes/index.html for recent updates. >> [1]PETSC ERROR: See docs/faq.html for hints about trouble shooting. >> [1]PETSC ERROR: See docs/index.html for manual pages. >> [1]PETSC ERROR: ------------------------------------------------------------------------ >> [1]PETSC ERROR: ./ex8g on a gnu-debug named trielen by berenguer Mon Nov 25 11:39:08 2013 >> [1]PETSC ERROR: Libraries linked from /home/cdcsp/berenguer/PETSC_INSTALL/lib >> [1]PETSC ERROR: Configure run at Mon Nov 25 09:58:03 2013 >> [1]PETSC ERROR: Configure options --prefix=/home/cdcsp/berenguer/PETSC_INSTALL --PETSC_DIR=/home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3 --with-mpi-dir=/home/cdcsp/berenguer/mpichinstall/ --with-shared-libraries=0 --with-scalar-type=real --with-precision=double --with-debugging=yes --download-f-blas-lapack=1 --with-valgrind=1 >> [1]PETSC ERROR: ------------------------------------------------------------------------ >> [1]PETSC ERROR: VecScatterBegin() line 1616 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/vec/vec/utils/vscat.c >> [1]PETSC ERROR: MatMult_MPIAIJ() line 1111 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/impls/aij/mpi/mpiaij.c >> [1]PETSC ERROR: MatMult() line 2179 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/mat/interface/matrix.c >> [1]PETSC ERROR: PCApplyBAorAB() line 679 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c >> [1]PETSC ERROR: KSP_PCApplyBAorAB() line 257 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h >> [1]PETSC ERROR: KSPGMRESCycle() line 160 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c >> [1]PETSC ERROR: KSPSolve_GMRES() line 240 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c >> [1]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >> [1]PETSC ERROR: PCApply_GASM() line 535 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/impls/gasm/gasm.c >> [1]PETSC ERROR: PCApply() line 442 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/pc/interface/precon.c >> [1]PETSC ERROR: KSP_PCApply() line 227 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/include/petsc-private/kspimpl.h >> [1]PETSC ERROR: KSPInitialResidual() line 64 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itres.c >> [1]PETSC ERROR: KSPSolve_GMRES() line 239 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c >> [1]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >> [1]PETSC ERROR: main() line 272 in src/ksp/ksp/examples/tutorials/ex8g.c >> application called MPI_Abort(MPI_COMM_WORLD, 85) - process 1 >> [0]PETSC ERROR: KSPSolve_GMRES() line 239 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/impls/gmres/gmres.c >> [0]PETSC ERROR: KSPSolve() line 441 in /home/cdcsp/berenguer/PETSC_INSTALL/petsc-3.4.3/src/ksp/ksp/interface/itfunc.c >> [0]PETSC ERROR: main() line 272 in src/ksp/ksp/examples/tutorials/ex8g.c >> application called MPI_Abort(MPI_COMM_WORLD, 85) - process 0 >> make: [runex8g_2] Erreur 85 (ignor?e) >> >> static char help[] = "Illustrates use of the preconditioner GASM.\n\ >> The Additive Schwarz Method for solving a linear system in parallel with KSP. The\n\ >> code indicates the procedure for setting user-defined subdomains. Input\n\ >> parameters include:\n\ >> -M: Number of mesh points in the x direction\n\ >> -N: Number of mesh points in the y direction\n\ >> -user_set_subdomain_solvers: User explicitly sets subdomain solvers\n\ >> -user_set_subdomains: Use the user-provided subdomain partitioning routine\n\ >> With -user_set_subdomains on, the following options are meaningful:\n\ >> -Mdomains: Number of subdomains in the x direction \n\ >> -Ndomains: Number of subdomains in the y direction \n\ >> -overlap: Size of domain overlap in terms of the number of mesh lines in x and y\n\ >> General useful options:\n\ >> -pc_gasm_print_subdomains: Print the index sets defining the subdomains\n\ >> \n"; >> >> /* >> Note: This example focuses on setting the subdomains for the GASM >> preconditioner for a problem on a 2D rectangular grid. See ex1.c >> and ex2.c for more detailed comments on the basic usage of KSP >> (including working with matrices and vectors). >> >> The GASM preconditioner is fully parallel. The user-space routine >> CreateSubdomains2D that computes the domain decomposition is also parallel >> and attempts to generate both subdomains straddling processors and multiple >> domains per processor. >> >> >> This matrix in this linear system arises from the discretized Laplacian, >> and thus is not very interesting in terms of experimenting with variants >> of the GASM preconditioner. >> */ >> >> /*T >> Concepts: KSP^Additive Schwarz Method (GASM) with user-defined subdomains >> Processors: n >> T*/ >> >> /* >> Include "petscksp.h" so that we can use KSP solvers. Note that this file >> automatically includes: >> petscsys.h - base PETSc routines petscvec.h - vectors >> petscmat.h - matrices >> petscis.h - index sets petscksp.h - Krylov subspace methods >> petscviewer.h - viewers petscpc.h - preconditioners >> */ >> #include >> >> >> >> #undef __FUNCT__ >> #define __FUNCT__ "main" >> int main(int argc,char **args) >> { >> Vec x,b,u; /* approx solution, RHS, exact solution */ >> Mat A; /* linear system matrix */ >> KSP ksp; /* linear solver context */ >> PC pc; /* PC context */ >> IS *inneris,*outeris; /* array of index sets that define the subdomains */ >> PetscInt overlap = 1; /* width of subdomain overlap */ >> PetscInt Nsub; /* number of subdomains */ >> PetscInt m = 15,n = 17; /* mesh dimensions in x- and y- directions */ >> PetscInt M = 2,N = 1; /* number of subdomains in x- and y- directions */ >> PetscInt i,j,Ii,J,Istart,Iend; >> PetscErrorCode ierr; >> PetscMPIInt size; >> PetscBool flg; >> PetscBool user_set_subdomains = PETSC_FALSE; >> PetscScalar v, one = 1.0; >> PetscReal e; >> >> PetscInitialize(&argc,&args,(char*)0,help); >> ierr = MPI_Comm_size(PETSC_COMM_WORLD,&size);CHKERRQ(ierr); >> ierr = PetscOptionsGetInt(NULL,"-M",&m,NULL);CHKERRQ(ierr); >> ierr = PetscOptionsGetInt(NULL,"-N",&n,NULL);CHKERRQ(ierr); >> ierr = PetscOptionsGetBool(NULL,"-user_set_subdomains",&user_set_subdomains,NULL);CHKERRQ(ierr); >> ierr = PetscOptionsGetInt(NULL,"-Mdomains",&M,NULL);CHKERRQ(ierr); >> ierr = PetscOptionsGetInt(NULL,"-Ndomains",&N,NULL);CHKERRQ(ierr); >> ierr = PetscOptionsGetInt(NULL,"-overlap",&overlap,NULL);CHKERRQ(ierr); >> >> /* ------------------------------------------------------------------- >> Compute the matrix and right-hand-side vector that define >> the linear system, Ax = b. >> ------------------------------------------------------------------- */ >> >> /* >> Assemble the matrix for the five point stencil, YET AGAIN >> */ >> ierr = MatCreate(PETSC_COMM_WORLD,&A);CHKERRQ(ierr); >> ierr = MatSetSizes(A,PETSC_DECIDE,PETSC_DECIDE,m*n,m*n);CHKERRQ(ierr); >> ierr = MatSetFromOptions(A);CHKERRQ(ierr); >> ierr = MatSetUp(A);CHKERRQ(ierr); >> ierr = MatGetOwnershipRange(A,&Istart,&Iend);CHKERRQ(ierr); >> for (Ii=Istart; Ii> v = -1.0; i = Ii/n; j = Ii - i*n; >> if (i>0) {J = Ii - n; ierr = MatSetValues(A,1,&Ii,1,&J,&v,INSERT_VALUES);CHKERRQ(ierr);} >> if (i> if (j>0) {J = Ii - 1; ierr = MatSetValues(A,1,&Ii,1,&J,&v,INSERT_VALUES);CHKERRQ(ierr);} >> if (j> v = 4.0; ierr = MatSetValues(A,1,&Ii,1,&Ii,&v,INSERT_VALUES);CHKERRQ(ierr); >> } >> ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >> ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); >> >> /* >> Create and set vectors >> */ >> ierr = VecCreate(PETSC_COMM_WORLD,&b);CHKERRQ(ierr); >> ierr = VecSetSizes(b,PETSC_DECIDE,m*n);CHKERRQ(ierr); >> ierr = VecSetFromOptions(b);CHKERRQ(ierr); >> ierr = VecDuplicate(b,&u);CHKERRQ(ierr); >> ierr = VecDuplicate(b,&x);CHKERRQ(ierr); >> ierr = VecSet(u,one);CHKERRQ(ierr); >> ierr = MatMult(A,u,b);CHKERRQ(ierr); >> >> /* >> Create linear solver context >> */ >> ierr = KSPCreate(PETSC_COMM_WORLD,&ksp);CHKERRQ(ierr); >> >> /* >> Set operators. Here the matrix that defines the linear system >> also serves as the preconditioning matrix. >> */ >> ierr = KSPSetOperators(ksp,A,A,DIFFERENT_NONZERO_PATTERN);CHKERRQ(ierr); >> >> /* >> Set the default preconditioner for this program to be GASM >> */ >> ierr = KSPGetPC(ksp,&pc);CHKERRQ(ierr); >> ierr = PCSetType(pc,PCGASM);CHKERRQ(ierr); >> >> /* ------------------------------------------------------------------- >> Define the problem decomposition >> ------------------------------------------------------------------- */ >> >> /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - >> Basic method, should be sufficient for the needs of many users. >> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - >> >> Set the overlap, using the default PETSc decomposition via >> PCGASMSetOverlap(pc,overlap); >> Could instead use the option -pc_gasm_overlap >> >> Set the total number of blocks via -pc_gasm_blocks >> Note: The GASM default is to use 1 block per processor. To >> experiment on a single processor with various overlaps, you >> must specify use of multiple blocks! >> */ >> >> /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - >> More advanced method, setting user-defined subdomains >> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - >> >> Firstly, create index sets that define the subdomains. The utility >> routine PCGASMCreateSubdomains2D() is a simple example, which partitions >> the 2D grid into MxN subdomains with an optional overlap. >> More generally, the user should write a custom routine for a particular >> problem geometry. >> >> Then call PCGASMSetLocalSubdomains() with resulting index sets >> to set the subdomains for the GASM preconditioner. >> */ >> >> if (!user_set_subdomains) { /* basic version */ >> ierr = PCGASMSetOverlap(pc,overlap);CHKERRQ(ierr); >> } else { /* advanced version */ >> ierr = PCGASMCreateSubdomains2D(pc, m,n,M,N,1,overlap,&Nsub,&inneris,&outeris);CHKERRQ(ierr); >> ierr = PCGASMSetSubdomains(pc,Nsub,inneris,outeris);CHKERRQ(ierr); >> flg = PETSC_FALSE; >> ierr = PetscOptionsGetBool(NULL,"-subdomain_view",&flg,NULL);CHKERRQ(ierr); >> if (flg) { >> ierr = PetscPrintf(PETSC_COMM_SELF,"Nmesh points: %D x %D; subdomain partition: %D x %D; overlap: %D; Nsub: %D\n",m,n,M,N,overlap,Nsub);CHKERRQ(ierr); >> ierr = PetscPrintf(PETSC_COMM_SELF,"Outer IS:\n");CHKERRQ(ierr); >> for (i=0; i> ierr = PetscPrintf(PETSC_COMM_SELF," outer IS[%d]\n",i);CHKERRQ(ierr); >> ierr = ISView(outeris[i],PETSC_VIEWER_STDOUT_SELF);CHKERRQ(ierr); >> } >> ierr = PetscPrintf(PETSC_COMM_SELF,"Inner IS:\n");CHKERRQ(ierr); >> for (i=0; i> ierr = PetscPrintf(PETSC_COMM_SELF," inner IS[%d]\n",i);CHKERRQ(ierr); >> ierr = ISView(inneris[i],PETSC_VIEWER_STDOUT_SELF);CHKERRQ(ierr); >> } >> } >> } >> >> /* ------------------------------------------------------------------- >> Set the linear solvers for the subblocks >> ------------------------------------------------------------------- */ >> >> /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - >> Basic method, should be sufficient for the needs of most users. >> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - >> >> By default, the GASM preconditioner uses the same solver on each >> block of the problem. To set the same solver options on all blocks, >> use the prefix -sub before the usual PC and KSP options, e.g., >> -sub_pc_type -sub_ksp_type -sub_ksp_rtol 1.e-4 >> >> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - >> Advanced method, setting different solvers for various blocks. >> - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - >> >> Note that each block's KSP context is completely independent of >> the others, and the full range of uniprocessor KSP options is >> available for each block. >> >> - Use PCGASMGetSubKSP() to extract the array of KSP contexts for >> the local blocks. >> - See ex7.c for a simple example of setting different linear solvers >> for the individual blocks for the block Jacobi method (which is >> equivalent to the GASM method with zero overlap). >> */ >> >> flg = PETSC_FALSE; >> ierr = PetscOptionsGetBool(NULL,"-user_set_subdomain_solvers",&flg,NULL);CHKERRQ(ierr); >> //flg=PETSC_TRUE; >> if (flg) { >> KSP *subksp; /* array of KSP contexts for local subblocks */ >> PetscInt nlocal,first; /* number of local subblocks, first local subblock */ >> PC subpc; /* PC context for subblock */ >> PetscBool isasm; >> >> ierr = PetscPrintf(PETSC_COMM_WORLD,"User explicitly sets subdomain solvers.\n");CHKERRQ(ierr); >> >> /* >> Set runtime options >> */ >> ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); >> >> /* >> Flag an error if PCTYPE is changed from the runtime options >> */ >> ierr = PetscObjectTypeCompare((PetscObject)pc,PCGASM,&isasm);CHKERRQ(ierr); >> if (!isasm) SETERRQ(PETSC_COMM_WORLD,1,"Cannot Change the PCTYPE when manually changing the subdomain solver settings"); >> >> /* >> Call KSPSetUp() to set the block Jacobi data structures (including >> creation of an internal KSP context for each block). >> >> Note: KSPSetUp() MUST be called before PCGASMGetSubKSP(). >> */ >> ierr = KSPSetUp(ksp);CHKERRQ(ierr); >> >> /* >> Extract the array of KSP contexts for the local blocks >> */ >> ierr = PCGASMGetSubKSP(pc,&nlocal,&first,&subksp);CHKERRQ(ierr); >> /* >> Loop over the local blocks, setting various KSP options >> for each block. >> */ >> for (i=0; i> ierr = KSPGetPC(subksp[i],&subpc);CHKERRQ(ierr); >> ierr = PCSetType(subpc,PCJACOBI);CHKERRQ(ierr); >> ierr = KSPSetType(subksp[i],KSPGMRES);CHKERRQ(ierr); >> ierr = KSPSetTolerances(subksp[i],1.e-7,PETSC_DEFAULT,PETSC_DEFAULT,PETSC_DEFAULT);CHKERRQ(ierr); >> } >> } else { >> /* >> Set runtime options >> */ >> ierr = KSPSetFromOptions(ksp);CHKERRQ(ierr); >> } >> >> /* ------------------------------------------------------------------- >> Solve the linear system >> ------------------------------------------------------------------- */ >> >> ierr = KSPSolve(ksp,b,x);CHKERRQ(ierr); >> >> /* ------------------------------------------------------------------- >> Compare result to the exact solution >> ------------------------------------------------------------------- */ >> ierr = VecAXPY(x,-1.0,u);CHKERRQ(ierr); >> ierr = VecNorm(x,NORM_INFINITY, &e);CHKERRQ(ierr); >> >> flg = PETSC_FALSE; >> ierr = PetscOptionsGetBool(NULL,"-print_error",&flg,NULL);CHKERRQ(ierr); >> if (flg) { >> ierr = PetscPrintf(PETSC_COMM_WORLD, "Infinity norm of the error: %G\n", e);CHKERRQ(ierr); >> } >> >> /* >> Free work space. All PETSc objects should be destroyed when they >> are no longer needed. >> */ >> >> if (user_set_subdomains) { >> ierr = PCGASMDestroySubdomains(Nsub, inneris, outeris);CHKERRQ(ierr); >> } >> ierr = KSPDestroy(&ksp);CHKERRQ(ierr); >> ierr = VecDestroy(&u);CHKERRQ(ierr); >> ierr = VecDestroy(&x);CHKERRQ(ierr); >> ierr = VecDestroy(&b);CHKERRQ(ierr); >> ierr = MatDestroy(&A);CHKERRQ(ierr); >> ierr = PetscFinalize(); >> return 0; >> } -- Laurent Berenguer From jedbrown at mcs.anl.gov Mon Nov 25 09:07:43 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Mon, 25 Nov 2013 09:07:43 -0600 Subject: [petsc-users] subksp type of PCGASM In-Reply-To: References: <87eh6461ld.fsf@jedbrown.org> Message-ID: <8761rg5xkw.fsf@jedbrown.org> Laurent Berenguer writes: >>How exactly did you set this subksp? Did you modify the code or change >>the command line arguments? Did the test run without errors when you >>don't modify anything? > > I tried both methods, the error is the same in both cases : > - I added the option -user_set_subdomain_solvers and modified the code > from > ierr = PCSetType(subpc,PCILU);CHKERRQ(ierr); > to > ierr = PCSetType(subpc,PCJACOBI);CHKERRQ(ierr); > - Or I only added -sub_ksp_type gmres without changing the code. > > The example runs correctly when I change nothing. But I'm not sure > what is the default subksp. When I change nothing in the code but it > seems to be KSPPREONLY. It should say if you run with -ksp_view. PCGASM is Dmitry's project and I don't think it has gotten much use yet. I'll leave it to him to explain what's going on here. If your cases are supposed to work, they should be added as tests. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From bsmith at mcs.anl.gov Mon Nov 25 11:38:00 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Mon, 25 Nov 2013 11:38:00 -0600 Subject: [petsc-users] PetscDrawSetSave In-Reply-To: References: Message-ID: <760A3E72-5146-456A-BF0B-BF18B31DC65A@mcs.anl.gov> On Nov 25, 2013, at 9:21 AM, iwaddington . wrote: > Thanks Barry, so I just have to download the 3 files you changed and run the configure and make steps again am I right ? You can do that but it would be better to use git to access the repository, instructions are in the wiki I sent you. The reason to use git to access the repository is because it is then trivial to access new changes and improvements in the future. If you always try to do it by downloading individual files you will find it is cumbersome and prone to error. Barry > > > On Mon, Nov 25, 2013 at 2:27 AM, Barry Smith wrote: > > We didn?t have support for saving these. I have just added it. You can access it in the ?next? branch of the PETSc repository, see https://bitbucket.org/petsc/petsc/wiki/Home > > Run the code with -draw_save -popup_draw_save (I had named those extra windows popup windows) > > Please let us know if this does not work. > > Barry > > On Nov 24, 2013, at 8:47 PM, iwaddington . wrote: > > > Hi, I am using PetscDrawSetSave to save some contour plots in gif files, but it is not saving the contour scales along with the images. Do you know a way to save the contour scales along with the plots ? Thanks. > > From davidblogueur at gmail.com Tue Nov 26 05:21:38 2013 From: davidblogueur at gmail.com (David Laroche) Date: Tue, 26 Nov 2013 12:21:38 +0100 Subject: [petsc-users] =?utf-8?q?=5BDL=5D_Une_vid=C3=A9o_gratuite_sur_la_r?= =?utf-8?q?ecette_du_bonheur_-_mettez_des_lunettes_3D_pour_cette_vid=C3=A9?= =?utf-8?q?o?= Message-ID: <439922a40269c91f3a6bf486e4bbe30e@smtp5.ymlpserver.net> -------------------------------------------------------------------------------- Cette newsletter vous a ?t? envoy?e au format graphique HTML. Si vous lisez cette version, alors votre logiciel de messagerie pr?f?re les e-mails au format texte. Vous pouvez lire la version originale en ligne: http://ymlp265.net/zN0Sny -------------------------------------------------------------------------------- Bonjour, Pour bien d?marrer cette magnifique journ?e, je vous envoie une des derni?res vid?os des WakeUpCalls nomm?e "la recette du bonheur". Dans cette vid?o vous d?couvrirez comment utiliser la puissance de votre cerveau pour ?tre PLUS heureux. Les WakeUpCalls est une s?rie de vid?o gratuite quotidienne pour passer ? l'action et faire passer sa vie au niveau sup?rieur. Cliquez ici pour d?couvrir "la recette du bonheur". ( http://lefacteurcle.com/recettebonheur ) ps : vous pouvez aussi cliquer sur l'image David, Conf?rencier & Formateur de la m?thode Laroche Strat?gies pratiques pour gagner confiance en soi _____________________________ Unsubscribe / Change Profile: http://ymlp265.net/ugbhewuugsgeshwygmggjeuhhb Powered par YourMailingListProvider -------------- next part -------------- An HTML attachment was scrubbed... URL: From epscodes at gmail.com Tue Nov 26 10:59:31 2013 From: epscodes at gmail.com (Xiangdong) Date: Tue, 26 Nov 2013 11:59:31 -0500 Subject: [petsc-users] tmpdir: change the path of temporary directory created during installation Message-ID: Hi everyone, During the petsc installation, it created a folder under /tmp/petsc-XYZ and ran some binary test. Since executing binaries is not allowed in /tmp on my workstation, Is there a variable so that I can set this tmpdir to another path? Thank you. Best, Xiangdong -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Tue Nov 26 11:02:18 2013 From: knepley at gmail.com (Matthew Knepley) Date: Tue, 26 Nov 2013 11:02:18 -0600 Subject: [petsc-users] tmpdir: change the path of temporary directory created during installation In-Reply-To: References: Message-ID: On Tue, Nov 26, 2013 at 10:59 AM, Xiangdong wrote: > Hi everyone, > > During the petsc installation, it created a folder under /tmp/petsc-XYZ > and ran some binary test. Since executing binaries is not allowed in /tmp > on my workstation, Is there a variable so that I can set this tmpdir to > another path? > You can set the TMPDIR env variable. Matt > Thank you. > > Best, > Xiangdong > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Tue Nov 26 11:02:42 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Tue, 26 Nov 2013 11:02:42 -0600 Subject: [petsc-users] tmpdir: change the path of temporary directory created during installation In-Reply-To: References: Message-ID: <87bo17p03x.fsf@jedbrown.org> Xiangdong writes: > Hi everyone, > > During the petsc installation, it created a folder under /tmp/petsc-XYZ and > ran some binary test. Since executing binaries is not allowed in /tmp on my > workstation, Is there a variable so that I can set this tmpdir to another > path? export TMPDIR=/place/to/put/tmp/files -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From epscodes at gmail.com Tue Nov 26 11:44:38 2013 From: epscodes at gmail.com (Xiangdong) Date: Tue, 26 Nov 2013 12:44:38 -0500 Subject: [petsc-users] tmpdir: change the path of temporary directory created during installation In-Reply-To: <87bo17p03x.fsf@jedbrown.org> References: <87bo17p03x.fsf@jedbrown.org> Message-ID: Thank you Matt and Jed. By the way, how can I find a list of all the environment variables that petsc takes? Thanks. Best, Xiangdong On Tue, Nov 26, 2013 at 12:02 PM, Jed Brown wrote: > Xiangdong writes: > > > Hi everyone, > > > > During the petsc installation, it created a folder under /tmp/petsc-XYZ > and > > ran some binary test. Since executing binaries is not allowed in /tmp on > my > > workstation, Is there a variable so that I can set this tmpdir to another > > path? > > export TMPDIR=/place/to/put/tmp/files > -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Tue Nov 26 11:53:11 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Tue, 26 Nov 2013 11:53:11 -0600 Subject: [petsc-users] tmpdir: change the path of temporary directory created during installation In-Reply-To: References: <87bo17p03x.fsf@jedbrown.org> Message-ID: <878uwboxrs.fsf@jedbrown.org> Xiangdong writes: > By the way, how can I find a list of all the environment variables that > petsc takes? Thanks. This is not a PETSc thing, it's POSIX mkstemp/mkdtemp. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From knepley at gmail.com Tue Nov 26 11:57:03 2013 From: knepley at gmail.com (Matthew Knepley) Date: Tue, 26 Nov 2013 11:57:03 -0600 Subject: [petsc-users] tmpdir: change the path of temporary directory created during installation In-Reply-To: <878uwboxrs.fsf@jedbrown.org> References: <87bo17p03x.fsf@jedbrown.org> <878uwboxrs.fsf@jedbrown.org> Message-ID: On Tue, Nov 26, 2013 at 11:53 AM, Jed Brown wrote: > Xiangdong writes: > > By the way, how can I find a list of all the environment variables that > > petsc takes? Thanks. > > This is not a PETSc thing, it's POSIX mkstemp/mkdtemp. > We use two env vars to indicate the build: PETSC_DIR = Root of the source tree PETSC_ARCH = Label for the build (has default if not defined) We use one env var to input options for a run: PETSC_OPTIONS = Matt -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From karpeev at mcs.anl.gov Tue Nov 26 12:33:17 2013 From: karpeev at mcs.anl.gov (Dmitry Karpeyev) Date: Tue, 26 Nov 2013 12:33:17 -0600 Subject: [petsc-users] subksp type of PCGASM In-Reply-To: <8761rg5xkw.fsf@jedbrown.org> References: <87eh6461ld.fsf@jedbrown.org> <8761rg5xkw.fsf@jedbrown.org> Message-ID: Thanks for reporting this. There appears to be a bug in MatGetSubMatricesParallel(). I'll look into it and add a test for the case above. In the meantime, maybe you could tell me more about what you are trying to do? PCGASM: it is rather experimental and doesn't get much use ordinarily. Depending on your particular use case, we can either try improving PCGASM, or find a different way to accomplish the same thing. Thanks. Dmitry. On Mon, Nov 25, 2013 at 9:07 AM, Jed Brown wrote: > Laurent Berenguer writes: > > >>How exactly did you set this subksp? Did you modify the code or change > >>the command line arguments? Did the test run without errors when you > >>don't modify anything? > > > > I tried both methods, the error is the same in both cases : > > - I added the option -user_set_subdomain_solvers and modified the code > > from > > ierr = PCSetType(subpc,PCILU);CHKERRQ(ierr); > > to > > ierr = PCSetType(subpc,PCJACOBI);CHKERRQ(ierr); > > - Or I only added -sub_ksp_type gmres without changing the code. > > > > The example runs correctly when I change nothing. But I'm not sure > > what is the default subksp. When I change nothing in the code but it > > seems to be KSPPREONLY. > > It should say if you run with -ksp_view. > > PCGASM is Dmitry's project and I don't think it has gotten much use yet. > I'll leave it to him to explain what's going on here. If your cases are > supposed to work, they should be added as tests. > -------------- next part -------------- An HTML attachment was scrubbed... URL: From karpeev at mcs.anl.gov Tue Nov 26 14:51:39 2013 From: karpeev at mcs.anl.gov (Dmitry Karpeyev) Date: Tue, 26 Nov 2013 14:51:39 -0600 Subject: [petsc-users] HIRING: Postdoctoral position in computational soft matter science Message-ID: Dear All, My apologies in advance for a misuse of this list. However, I wanted to draw your attention to a new postdoctoral position at Argonne National Laboratory, which may be of interest to some of you. This job involves research into efficient algorithms for numerical modeling and simulation of solutions of charged particles in confined geometries. Experience with PETSc, libMesh or MOOSE will be a big plus here, hence, this message. The formal position description can be found here: http://web.anl.gov/jobsearch/detail.jsp?userreqid=321577+MCS&lsBrowse=POSTDOC In plainer terms, however, we are interested in the quantitative understanding of the dynamics of DNA molecules suspended in solutions of varying ionic strength, and confined to channels of 10-100nm linear cross-section, with charged walls. If you know your way around FEM models of Stokesian fluids in nontrivial geometries, have experience with electrostatic force calculations, and are interested in making these algorithms run fast, this may be the job for you. We are interested in accelerating these simulations to be able to resolve the long-time fluctuation dynamics of the DNA molecules translocating (i.e., flopping their way through) the nanochannels. Some modeling issues will likely need to be addressed as well: at various ionic strengths the finite-size effects of ions in the solution may become significant near the channel walls or the solute (DNA) molecules. Thus, an interest and experience in particle-to-continuum coupling is a plus. A working knowledge of C/C++ is required. Familiarity with the GPU/multicore/manycore architectures is preferred. Experience with the PETSc library (http://www.mcs.anl.gov/petsc) is a huge plus. Ditto libMesh and MOOSE. A successful candidate will work closely with the University of Chicago scientists at the Computation Institute as well as the Institute for Molecular Engineering. If you have questions, please feel free to contact me directly. I apologize, if you are getting this message more than once. Best regards, Dmitry. -- Dmitry Karpeyev Mathematics and Computer Science Argonne National Laboratory Argonne, Illinois, USA and Computation Institute University of Chicago 5735 S. Ellis Avenue Chicago, IL 60637 ----------------------- Phone: 630-252-1229 Fax: 630-252-5986 -------------- next part -------------- An HTML attachment was scrubbed... URL: From lu_qin_2000 at yahoo.com Tue Nov 26 15:08:57 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Tue, 26 Nov 2013 13:08:57 -0800 (PST) Subject: [petsc-users] Specify -O2 for PETSc build in Windows Message-ID: <1385500137.41277.YahooMailNeo@web160216.mail.bf1.yahoo.com> Hello, I?was trying to build PETSc using Intel compiler with -O2 optimization option in Windows-7, but somehow it was override by -O3 during make. My configure options are as follows: ./configure --configModules=PETSc.Configure --optionsModule=PETSc.compilerOptions --with-cc="win32fe icl -O2" --with-fc="win32fe ifort -O2" --with-cxx="win32fe icl" --with-petsc-arch=arch-win64-release-with-hypre-o2 --prefix=/cygdrive/c/cygwin_cache/petsc-3.4.2-release-win64-with-hypre-o2 --with-blas-lapack-dir="/cygdrive/c/Program Files (x86)/intel/Compiler/11.1/051/mkl/em64t/lib" --with-mpi-dir="/cygdrive/c/Program Files/mpich2x64" --with-debugging=0 --with-x=0 --with-x11=0 --with-xt=0 --with-shared-libraries=0 --useThreads=0 Then the make gave messages such as:? snesdnest.c libfast in: /cygdrive/c/cygwin_cache/petsc-3.4.2/src/snes/interface/ftn-auto icl: command line warning #10120: overriding '/O2' with '/O3' ? How can I force the make to use -O2? ? Many thanks for your suggestions, Qin?? From balay at mcs.anl.gov Tue Nov 26 15:13:26 2013 From: balay at mcs.anl.gov (Satish Balay) Date: Tue, 26 Nov 2013 15:13:26 -0600 (CST) Subject: [petsc-users] Specify -O2 for PETSc build in Windows In-Reply-To: <1385500137.41277.YahooMailNeo@web160216.mail.bf1.yahoo.com> References: <1385500137.41277.YahooMailNeo@web160216.mail.bf1.yahoo.com> Message-ID: You have to set the optimzation flags with COPTFLAGS/FOPTFLAGS/CXXOPTFLAGS options. The defaults are '-O3 -QxW' for intel compilers so you might want to use: ./configure COPTFLAGS='-O2 -QxW' CXXOPTFLAGS='-O2 -QxW' FOPTFLAGS='-O2 -QxW' ..... Satish On Tue, 26 Nov 2013, Qin Lu wrote: > Hello, > > I?was trying to build PETSc using Intel compiler with -O2 optimization option in Windows-7, but somehow it was override by -O3 during make. My configure options are as follows: > > ./configure --configModules=PETSc.Configure --optionsModule=PETSc.compilerOptions --with-cc="win32fe icl -O2" --with-fc="win32fe ifort -O2" --with-cxx="win32fe icl" --with-petsc-arch=arch-win64-release-with-hypre-o2 --prefix=/cygdrive/c/cygwin_cache/petsc-3.4.2-release-win64-with-hypre-o2 --with-blas-lapack-dir="/cygdrive/c/Program Files (x86)/intel/Compiler/11.1/051/mkl/em64t/lib" --with-mpi-dir="/cygdrive/c/Program Files/mpich2x64" --with-debugging=0 --with-x=0 --with-x11=0 --with-xt=0 --with-shared-libraries=0 --useThreads=0 > > Then the make gave messages such as:? > snesdnest.c > libfast in: /cygdrive/c/cygwin_cache/petsc-3.4.2/src/snes/interface/ftn-auto > icl: command line warning #10120: overriding '/O2' with '/O3' > ? > How can I force the make to use -O2? > ? > Many thanks for your suggestions, > Qin?? > From rlmackie862 at gmail.com Tue Nov 26 15:45:22 2013 From: rlmackie862 at gmail.com (Randall Mackie) Date: Tue, 26 Nov 2013 13:45:22 -0800 Subject: [petsc-users] printing from some processors to a file Message-ID: I am trying to print a character string to a file from one or more processors in a communicator. I thought that I could do this using PetscViewerASCIISynchronizedPrintf, but it prints to the screen instead of to the file opened as a viewer. The attached simple program illustrates the issue. If I remove the (if rank == 1) and therefore call PetscViewerASCIISynchronizedPrintf from every process, it works as expected. If called from only 1 process, it prints to the screen. I thought, from the documentation, that it was not collective, but maybe I am not understanding it's usage. Randy M. -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: test.F90 Type: application/octet-stream Size: 615 bytes Desc: not available URL: From bsmith at mcs.anl.gov Tue Nov 26 16:15:50 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Tue, 26 Nov 2013 16:15:50 -0600 Subject: [petsc-users] printing from some processors to a file In-Reply-To: References: Message-ID: <9BE155FF-78B6-475C-BCE0-45F2DDB6400F@mcs.anl.gov> Randy, Thanks for reporting this with a full test case. It is our bug. The implementation keeps the file pointer in a global variable on process 0 that is set in the first call to the PetscViewerASCIISynchronizedPrintf() or in PetscSynchronizedFPrintf() . The reason for the ugly global is that PetscSynchronizedFlush() can be applied to any number of file pointers. But if the SynchronizedPrintf() is never called on process 0 the file pointer global never gets set so it defaults to stdout. It may be fixable by having PetscSynchronizedFlush() take a file pointer argument (makes sense). I?ll try to fix it, until make sure process 0 prints at least one thing and it will work ok. Barry On Nov 26, 2013, at 3:45 PM, Randall Mackie wrote: > I am trying to print a character string to a file from one or more processors in a communicator. I thought that I could do this using PetscViewerASCIISynchronizedPrintf, but it prints to the screen instead of to the file opened as a viewer. The attached simple program illustrates the issue. If I remove the (if rank == 1) and therefore call PetscViewerASCIISynchronizedPrintf from every process, it works as expected. If called from only 1 process, it prints to the screen. I thought, from the documentation, that it was not collective, but maybe I am not understanding it's usage. > > > Randy M. > From lu_qin_2000 at yahoo.com Tue Nov 26 17:04:19 2013 From: lu_qin_2000 at yahoo.com (Qin Lu) Date: Tue, 26 Nov 2013 15:04:19 -0800 (PST) Subject: [petsc-users] Specify -O2 for PETSc build in Windows In-Reply-To: References: <1385500137.41277.YahooMailNeo@web160216.mail.bf1.yahoo.com> Message-ID: <1385507059.46518.YahooMailNeo@web160204.mail.bf1.yahoo.com> Satish, ? It works. Thanks?a lot! ? Qin On Tuesday, November 26, 2013 3:13 PM, Satish Balay wrote: You have to set the optimzation flags with COPTFLAGS/FOPTFLAGS/CXXOPTFLAGS options. The defaults are '-O3 -QxW' for intel compilers so you might want to use: ./configure COPTFLAGS='-O2 -QxW' CXXOPTFLAGS='-O2 -QxW' FOPTFLAGS='-O2 -QxW' ..... Satish On Tue, 26 Nov 2013, Qin Lu wrote: > Hello, > > I?was trying to build PETSc using Intel compiler with -O2 optimization option in Windows-7, but somehow it was override by -O3 during make. My configure options are as follows: > > ./configure --configModules=PETSc.Configure --optionsModule=PETSc.compilerOptions --with-cc="win32fe icl -O2" --with-fc="win32fe ifort -O2" --with-cxx="win32fe icl" --with-petsc-arch=arch-win64-release-with-hypre-o2 --prefix=/cygdrive/c/cygwin_cache/petsc-3.4.2-release-win64-with-hypre-o2 --with-blas-lapack-dir="/cygdrive/c/Program Files (x86)/intel/Compiler/11.1/051/mkl/em64t/lib" --with-mpi-dir="/cygdrive/c/Program Files/mpich2x64" --with-debugging=0 --with-x=0 --with-x11=0 --with-xt=0 --with-shared-libraries=0 --useThreads=0 > > Then the make gave messages such as:? > snesdnest.c > libfast in: /cygdrive/c/cygwin_cache/petsc-3.4.2/src/snes/interface/ftn-auto > icl: command line warning #10120: overriding '/O2' with '/O3' > ? > How can I force the make to use -O2? > ? > Many thanks for your suggestions, > Qin?? > From bsmith at mcs.anl.gov Tue Nov 26 22:29:49 2013 From: bsmith at mcs.anl.gov (Barry Smith) Date: Tue, 26 Nov 2013 22:29:49 -0600 Subject: [petsc-users] printing from some processors to a file In-Reply-To: References: Message-ID: <5F66DE9C-D584-4ABF-9274-2BF0A8854AC0@mcs.anl.gov> I have fixed this in the next branch of PETSc https://bitbucket.org/petsc/petsc/wiki/Home Because it changes APIs I cannot merge it into the current release (though it doesn?t change APIs for your code). Thanks for reporting the problem, Barry On Nov 26, 2013, at 3:45 PM, Randall Mackie wrote: > I am trying to print a character string to a file from one or more processors in a communicator. I thought that I could do this using PetscViewerASCIISynchronizedPrintf, but it prints to the screen instead of to the file opened as a viewer. The attached simple program illustrates the issue. If I remove the (if rank == 1) and therefore call PetscViewerASCIISynchronizedPrintf from every process, it works as expected. If called from only 1 process, it prints to the screen. I thought, from the documentation, that it was not collective, but maybe I am not understanding it's usage. > > > Randy M. > From jsd1 at rice.edu Wed Nov 27 02:12:47 2013 From: jsd1 at rice.edu (Justin Dong (Me)) Date: Wed, 27 Nov 2013 03:12:47 -0500 Subject: [petsc-users] Floating point exception, only occurs when assembling "large" matrices Message-ID: I am assembling a global mass matrix on a mesh consisting of 8192 elements and 3 basis functions per element (so the global dimension is 24,576. For some reason, when assembling this matrix I get tons of floating point exceptions everywhere: [0]PETSC ERROR: --------------------- Error Message ------------------------------------ [0]PETSC ERROR: Floating point exception! [0]PETSC ERROR: Inserting nan at matrix entry (0,0)! I get this error for every 3rd diagonal entry of the matrix, but the problem is bigger than that I suspect. In my computation of the local 3x3 element mass matrices, printing out the values gives all zeros (which is what the entries are initialized to be in my code). I?m at a loss for how to debug this problem. My first thought was that there is an error in the mesh, but I?m certain now that this is not the case since the mesh is generated using the exact same routine that generates all of my coarser meshes before this one that fails. For example, the mesh that is one refinement level below this one has 2048 elements and works completely fine. This is how I am creating the global matrix: MatCreateSeqAIJ(PETSC_COMM_SELF, NLoc*nElems, NLoc*nElems, NLoc, PETSC_NULL, &Mglobal); where I allocate NLoc = 3 non-zero entries per row in this case. -------------- next part -------------- An HTML attachment was scrubbed... URL: From rupp at mcs.anl.gov Wed Nov 27 02:22:57 2013 From: rupp at mcs.anl.gov (Karl Rupp) Date: Wed, 27 Nov 2013 09:22:57 +0100 Subject: [petsc-users] Floating point exception, only occurs when assembling "large" matrices In-Reply-To: References: Message-ID: <5295ABE1.8020302@mcs.anl.gov> Hi Justin, did you run your code through Valgrind? I suspect that this is caused by some memory corruption, particularly as you claim that smaller matrix sizes are fine. Also, do you check all the error codes returned by PETSc routines? Also, are you sure about the nonzeropattern of your matrix? From your description it sounds like you are solving 8192 decoupled problems of size 3x3, while for e.g. typical finite element applications you get 3x3 local mass matrices per cell, with the total number of degrees of freedom given by the number of vertices. If the above doesn't help: Any chance of sending us the relevant source code? Best regards, Karli On 11/27/2013 09:12 AM, Justin Dong (Me) wrote: > I am assembling a global mass matrix on a mesh consisting of 8192 > elements and 3 basis functions per element (so the global dimension is > 24,576. For some reason, when assembling this matrix I get tons of > floating point exceptions everywhere: > > [0]PETSC ERROR: --------------------- Error Message > ------------------------------------ > [0]PETSC ERROR: Floating point exception! > [0]PETSC ERROR: Inserting nan at matrix entry (0,0)! > > I get this error for every 3rd diagonal entry of the matrix, but the > problem is bigger than that I suspect. In my computation of the local > 3x3 element mass matrices, printing out the values gives all zeros > (which is what the entries are initialized to be in my code). > > I?m at a loss for how to debug this problem. My first thought was that > there is an error in the mesh, but I?m certain now that this is not the > case since the mesh is generated using the exact same routine that > generates all of my coarser meshes before this one that fails. For > example, the mesh that is one refinement level below this one has 2048 > elements and works completely fine. This is how I am creating the global > matrix: > > MatCreateSeqAIJ(PETSC_COMM_SELF, NLoc*nElems, NLoc*nElems, NLoc, > PETSC_NULL, &Mglobal); > > where I allocate NLoc = 3 non-zero entries per row in this case. From jsd1 at rice.edu Wed Nov 27 03:19:41 2013 From: jsd1 at rice.edu (Justin Dong (Me)) Date: Wed, 27 Nov 2013 04:19:41 -0500 Subject: [petsc-users] Floating point exception, only occurs when assembling "large" matrices In-Reply-To: <5295ABE1.8020302@mcs.anl.gov> References: <5295ABE1.8020302@mcs.anl.gov> Message-ID: <69A6F0C7-A99B-4500-A66E-CBEDAA7C9362@rice.edu> Hi Karl, Unfortunately, I don?t have access to valgrind as I?m on Mac OS X 10.9 right now. I?m only examining the global mass matrix right now, which will be block diagonal in 3x3 blocks if I?m not mistaken. The larger application is for DG FEM, which definitely has a different non-zero pattern and more than 3 non-zero entries per row. For the global matrix overall, my allocation of the non-zero entries per rough is just a rough estimate. As I understand it, I don?t have to be exact when specifying this number to MatCreateSeqAIJ. Attached below is a minimal code that should run and produce the problem. I hope it will compile for you. I added comments where necessary. The variable ?refs? in main controls the mesh, and on my machine it works up to and including refs=5. For refs=6 is where I get the errors. It?s a bit lengthy but oh well. -------------- next part -------------- A non-text attachment was scrubbed... Name: MinimalFEMCode.zip Type: application/zip Size: 212030 bytes Desc: not available URL: -------------- next part -------------- On Nov 27, 2013, at 3:22 AM, Karl Rupp wrote: > Hi Justin, > > did you run your code through Valgrind? I suspect that this is caused by some memory corruption, particularly as you claim that smaller matrix sizes are fine. Also, do you check all the error codes returned by PETSc routines? > > Also, are you sure about the nonzeropattern of your matrix? From your description it sounds like you are solving 8192 decoupled problems of size 3x3, while for e.g. typical finite element applications you get 3x3 local mass matrices per cell, with the total number of degrees of freedom given by the number of vertices. > > If the above doesn't help: Any chance of sending us the relevant source code? > > Best regards, > Karli > > > On 11/27/2013 09:12 AM, Justin Dong (Me) wrote: >> I am assembling a global mass matrix on a mesh consisting of 8192 >> elements and 3 basis functions per element (so the global dimension is >> 24,576. For some reason, when assembling this matrix I get tons of >> floating point exceptions everywhere: >> >> [0]PETSC ERROR: --------------------- Error Message >> ------------------------------------ >> [0]PETSC ERROR: Floating point exception! >> [0]PETSC ERROR: Inserting nan at matrix entry (0,0)! >> >> I get this error for every 3rd diagonal entry of the matrix, but the >> problem is bigger than that I suspect. In my computation of the local >> 3x3 element mass matrices, printing out the values gives all zeros >> (which is what the entries are initialized to be in my code). >> >> I?m at a loss for how to debug this problem. My first thought was that >> there is an error in the mesh, but I?m certain now that this is not the >> case since the mesh is generated using the exact same routine that >> generates all of my coarser meshes before this one that fails. For >> example, the mesh that is one refinement level below this one has 2048 >> elements and works completely fine. This is how I am creating the global >> matrix: >> >> MatCreateSeqAIJ(PETSC_COMM_SELF, NLoc*nElems, NLoc*nElems, NLoc, >> PETSC_NULL, &Mglobal); >> >> where I allocate NLoc = 3 non-zero entries per row in this case. > From rupp at mcs.anl.gov Wed Nov 27 07:13:11 2013 From: rupp at mcs.anl.gov (Karl Rupp) Date: Wed, 27 Nov 2013 14:13:11 +0100 Subject: [petsc-users] Floating point exception, only occurs when assembling "large" matrices In-Reply-To: <69A6F0C7-A99B-4500-A66E-CBEDAA7C9362@rice.edu> References: <5295ABE1.8020302@mcs.anl.gov> <69A6F0C7-A99B-4500-A66E-CBEDAA7C9362@rice.edu> Message-ID: <5295EFE7.7020007@mcs.anl.gov> Hi Justin, > Unfortunately, I don?t have access to valgrind as I?m on Mac OS X 10.9 right now. I?m only examining the global mass matrix right now, which will be block diagonal in 3x3 blocks if I?m not mistaken. The larger application is for DG FEM, which definitely has a different non-zero pattern and more than 3 non-zero entries per row. For the global matrix overall, my allocation of the non-zero entries per rough is just a rough estimate. As I understand it, I don?t have to be exact when specifying this number to MatCreateSeqAIJ. This is correct, even though it is not ideal in terms of performance. It's usually better to be a bit pessimistic and avoid reallocations this way. > Attached below is a minimal code that should run and produce the problem. I hope it will compile for you. I added comments where necessary. The variable ?refs? in main controls the mesh, and on my machine it works up to and including refs=5. For refs=6 is where I get the errors. It?s a bit lengthy but oh well. The code is valgrind-clean on my machine, but it only writes zeros to the matrix (these are the values in Mlocal) for refs=6. Maybe you have an overflow or a spurious integer operation somewhere? It seems to me that something goes wrong in DG_local_mass() already. Best regards, Karli > > On Nov 27, 2013, at 3:22 AM, Karl Rupp wrote: > >> Hi Justin, >> >> did you run your code through Valgrind? I suspect that this is caused by some memory corruption, particularly as you claim that smaller matrix sizes are fine. Also, do you check all the error codes returned by PETSc routines? >> >> Also, are you sure about the nonzeropattern of your matrix? From your description it sounds like you are solving 8192 decoupled problems of size 3x3, while for e.g. typical finite element applications you get 3x3 local mass matrices per cell, with the total number of degrees of freedom given by the number of vertices. >> >> If the above doesn't help: Any chance of sending us the relevant source code? >> >> Best regards, >> Karli >> >> >> On 11/27/2013 09:12 AM, Justin Dong (Me) wrote: >>> I am assembling a global mass matrix on a mesh consisting of 8192 >>> elements and 3 basis functions per element (so the global dimension is >>> 24,576. For some reason, when assembling this matrix I get tons of >>> floating point exceptions everywhere: >>> >>> [0]PETSC ERROR: --------------------- Error Message >>> ------------------------------------ >>> [0]PETSC ERROR: Floating point exception! >>> [0]PETSC ERROR: Inserting nan at matrix entry (0,0)! >>> >>> I get this error for every 3rd diagonal entry of the matrix, but the >>> problem is bigger than that I suspect. In my computation of the local >>> 3x3 element mass matrices, printing out the values gives all zeros >>> (which is what the entries are initialized to be in my code). >>> >>> I?m at a loss for how to debug this problem. My first thought was that >>> there is an error in the mesh, but I?m certain now that this is not the >>> case since the mesh is generated using the exact same routine that >>> generates all of my coarser meshes before this one that fails. For >>> example, the mesh that is one refinement level below this one has 2048 >>> elements and works completely fine. This is how I am creating the global >>> matrix: >>> >>> MatCreateSeqAIJ(PETSC_COMM_SELF, NLoc*nElems, NLoc*nElems, NLoc, >>> PETSC_NULL, &Mglobal); >>> >>> where I allocate NLoc = 3 non-zero entries per row in this case. >> > From rlmackie862 at gmail.com Wed Nov 27 10:24:44 2013 From: rlmackie862 at gmail.com (Randall Mackie) Date: Wed, 27 Nov 2013 08:24:44 -0800 Subject: [petsc-users] printing from some processors to a file In-Reply-To: <5F66DE9C-D584-4ABF-9274-2BF0A8854AC0@mcs.anl.gov> References: <5F66DE9C-D584-4ABF-9274-2BF0A8854AC0@mcs.anl.gov> Message-ID: Hi Barry, Thanks for fixing this. The workaround you suggested yesterday works perfectly well for me at this point, and I'll look at this fix in the next release. Randy M. On Tue, Nov 26, 2013 at 8:29 PM, Barry Smith wrote: > > I have fixed this in the next branch of PETSc > https://bitbucket.org/petsc/petsc/wiki/Home > > Because it changes APIs I cannot merge it into the current release > (though it doesn?t change APIs for your code). > > Thanks for reporting the problem, > > Barry > On Nov 26, 2013, at 3:45 PM, Randall Mackie wrote: > > > I am trying to print a character string to a file from one or more > processors in a communicator. I thought that I could do this using > PetscViewerASCIISynchronizedPrintf, but it prints to the screen instead of > to the file opened as a viewer. The attached simple program illustrates the > issue. If I remove the (if rank == 1) and therefore call > PetscViewerASCIISynchronizedPrintf from every process, it works as > expected. If called from only 1 process, it prints to the screen. I > thought, from the documentation, that it was not collective, but maybe I am > not understanding it's usage. > > > > > > Randy M. > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From irving at naml.us Wed Nov 27 13:36:02 2013 From: irving at naml.us (Geoffrey Irving) Date: Wed, 27 Nov 2013 11:36:02 -0800 Subject: [petsc-users] best way to ignore a constant in FEM error evaluation Message-ID: I may be missing something, but I can't think of a clean way to use DMPlexComputeL2Diff but ignore a constant shift in a Neumann problem. It seems there's a missing routine that would make this easy, but I'm not entirely sure which routine that is. Mathematically, I have a Vec x and an exact solution f, and want to compute fc = int_D f xc = int_D x error = sqrt int_D (x-f-(xc-fc)/volume)^2 where int_D means integrate over the domain. The most general option would be to make a generic FE integral routine can it twice as volume = int_D 1 shift = (int_D (x - f))/volume error = sqrt int_D (x-f-shift) where I can could either compute volume myself, with another call to the integration routine, or with a new special function. Thoughts? Thanks, Geoffrey From irving at naml.us Wed Nov 27 13:38:28 2013 From: irving at naml.us (Geoffrey Irving) Date: Wed, 27 Nov 2013 11:38:28 -0800 Subject: [petsc-users] best way to ignore a constant in FEM error evaluation In-Reply-To: References: Message-ID: On Wed, Nov 27, 2013 at 11:36 AM, Geoffrey Irving wrote: > I may be missing something, but I can't think of a clean way to use > DMPlexComputeL2Diff but ignore a constant shift in a Neumann problem. > It seems there's a missing routine that would make this easy, but I'm > not entirely sure which routine that is. Mathematically, I have a Vec > x and an exact solution f, and want to compute > > fc = int_D f > xc = int_D x > error = sqrt int_D (x-f-(xc-fc)/volume)^2 > > where int_D means integrate over the domain. > > The most general option would be to make a generic FE integral routine > can it twice as > > volume = int_D 1 > shift = (int_D (x - f))/volume > error = sqrt int_D (x-f-shift) > > where I can could either compute volume myself, with another call to > the integration routine, or with a new special function. > > Thoughts? Oops, I suppose the obvious solution is to just make more PetscFEMs to do the integrals I want. Geoffrey From nico.schloemer at gmail.com Thu Nov 28 09:22:16 2013 From: nico.schloemer at gmail.com (=?ISO-8859-1?Q?Nico_Schl=F6mer?=) Date: Thu, 28 Nov 2013 16:22:16 +0100 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! Message-ID: Hi all, I'm toying around with PETSc's field split capabilities for the 2D-Stokes problem, and similar to ex42 I call // set of the operators // set of a bunch of options PetscOptionsSetValue("-pc_type", "fieldsplit"); PC pc; ierr = KSPGetPC(*_ksp, &pc); const PetscInt ufields[] = {0,1}, pfields[] = {2}; ierr = PCFieldSplitSetBlockSize(pc,3); PCFieldSplitSetFields(pc,"u",2,ufields,ufields); PCFieldSplitSetFields(pc,"p",1,pfields,pfields); When finally solving, however, I'm getting the error message [0]PETSC ERROR: Petsc has generated inconsistent data! [0]PETSC ERROR: Unhandled case, must have at least two fields, not 1! which confuses me since seem to set two fields u, p above. Any hints on what the problem could be? Cheers, Nico From jedbrown at mcs.anl.gov Thu Nov 28 09:25:41 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 28 Nov 2013 09:25:41 -0600 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! In-Reply-To: References: Message-ID: <87pppkeefe.fsf@jedbrown.org> Nico Schl?mer writes: > Hi all, > > I'm toying around with PETSc's field split capabilities for the > 2D-Stokes problem, and similar to ex42 I call > > // set of the operators > // set of a bunch of options > PetscOptionsSetValue("-pc_type", "fieldsplit"); > PC pc; > ierr = KSPGetPC(*_ksp, &pc); > const PetscInt ufields[] = {0,1}, pfields[] = {2}; > ierr = PCFieldSplitSetBlockSize(pc,3); > PCFieldSplitSetFields(pc,"u",2,ufields,ufields); > PCFieldSplitSetFields(pc,"p",1,pfields,pfields); > > When finally solving, however, I'm getting the error message > > [0]PETSC ERROR: Petsc has generated inconsistent data! > [0]PETSC ERROR: Unhandled case, must have at least two fields, not 1! Short answer: use PCFieldSplitSetIS(). From the man page: Notes: Use PCFieldSplitSetIS() to set a completely general set of indices as a field. The PCFieldSplitSetFields() is for defining fields as strided blocks. For example, if the block size is three then one can define a field as 0, or 1 or 2 or 0,1 or 0,2 or 1,2 which mean 0xx3xx6xx9xx12 ... x1xx4xx7xx ... xx2xx5xx8xx.. 01x34x67x... 0x1x3x5x7.. x12x45x78x.... where the numbered entries indicate what is in the field. This function is called once per split (it creates a new split each time). Solve options for this split will be available under the prefix -fieldsplit_SPLITNAME_. Developer Note: This routine does not actually create the IS representing the split, that is delayed until PCSetUp_FieldSplit(), because information about the vector/matrix layouts may not be available when this routine is called. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From nico.schloemer at gmail.com Thu Nov 28 10:28:18 2013 From: nico.schloemer at gmail.com (=?ISO-8859-1?Q?Nico_Schl=F6mer?=) Date: Thu, 28 Nov 2013 17:28:18 +0100 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! In-Reply-To: <87pppkeefe.fsf@jedbrown.org> References: <87pppkeefe.fsf@jedbrown.org> Message-ID: Hi Jed, thanks for the quick answer! I can of course generate the strided index sets (01x34x6..., xx2xx5x...) myself and insert them via PCFieldSplitSetIS(), but I thought that ultimately does the same thing that I would have liked PCFieldSplitSetFields() to do for me. Not true? --Nico On Thu, Nov 28, 2013 at 4:25 PM, Jed Brown wrote: > Nico Schl?mer writes: > >> Hi all, >> >> I'm toying around with PETSc's field split capabilities for the >> 2D-Stokes problem, and similar to ex42 I call >> >> // set of the operators >> // set of a bunch of options >> PetscOptionsSetValue("-pc_type", "fieldsplit"); >> PC pc; >> ierr = KSPGetPC(*_ksp, &pc); >> const PetscInt ufields[] = {0,1}, pfields[] = {2}; >> ierr = PCFieldSplitSetBlockSize(pc,3); >> PCFieldSplitSetFields(pc,"u",2,ufields,ufields); >> PCFieldSplitSetFields(pc,"p",1,pfields,pfields); >> >> When finally solving, however, I'm getting the error message >> >> [0]PETSC ERROR: Petsc has generated inconsistent data! >> [0]PETSC ERROR: Unhandled case, must have at least two fields, not 1! > > Short answer: use PCFieldSplitSetIS(). From the man page: > > Notes: Use PCFieldSplitSetIS() to set a completely general set of indices as a field. > > The PCFieldSplitSetFields() is for defining fields as strided blocks. For example, if the block > size is three then one can define a field as 0, or 1 or 2 or 0,1 or 0,2 or 1,2 which mean > 0xx3xx6xx9xx12 ... x1xx4xx7xx ... xx2xx5xx8xx.. 01x34x67x... 0x1x3x5x7.. x12x45x78x.... > where the numbered entries indicate what is in the field. > > This function is called once per split (it creates a new split each time). Solve options > for this split will be available under the prefix -fieldsplit_SPLITNAME_. > > Developer Note: This routine does not actually create the IS representing the split, that is delayed > until PCSetUp_FieldSplit(), because information about the vector/matrix layouts may not be > available when this routine is called. From jedbrown at mcs.anl.gov Thu Nov 28 10:53:02 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Thu, 28 Nov 2013 10:53:02 -0600 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! In-Reply-To: References: <87pppkeefe.fsf@jedbrown.org> Message-ID: <87eh60eadt.fsf@jedbrown.org> Nico Schl?mer writes: > Hi Jed, > > thanks for the quick answer! > I can of course generate the strided index sets (01x34x6..., > xx2xx5x...) myself and insert them via PCFieldSplitSetIS(), but I > thought that ultimately does the same thing that I would have liked > PCFieldSplitSetFields() to do for me. Not true? PCFieldSplitSetFields is something you call only once to identify all the fields. It doesn't let you group multiple strides in one, so you could create "ux", "uy", and "p", but then those would be separate. You can create splits using numbers, -fieldsplit_block_size 3 -fieldsplit_0_fields 0,1 -fieldsplit_1_fields 2 (assuming collocated discretization), but this can be rather opaque to maintain. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From jsd1 at rice.edu Fri Nov 29 01:11:35 2013 From: jsd1 at rice.edu (Justin Dong (Me)) Date: Fri, 29 Nov 2013 02:11:35 -0500 Subject: [petsc-users] PC factor shift type, random results? Message-ID: <57D3B071-4A66-4F97-80A4-F2ABB6185619@rice.edu> Hi all, I?m solving a non-linear advection-diffusion problem in 2D and am testing the routine on various meshes. The non-linearity is handled via Newton?s method. If I consider the unit square [0,1]^2, my results were fine. I tried testing on [-1,1]^2 and got either garbage results or a warning about zero pivots. I thought my Jacobian might be computed incorrectly, but I think that isn?t the case since I applied -pc-factor_shift_type POSITIVE DEFINITE and miraculously got the correct result now. But in applying this factor shift over various trials on the same problem, the results seem to be pretty random. Since I am testing on manufactured solutions, I force the problem to be linear since I have prior knowledge of the exact solution and Newton?s method should converge in one iteration. For instance, this is what the result should be (iterations and errors are for each Newton iteration) > student-113-wless94-099:NonlinearAdvectionDiffusion justindong$ ./main -pc_factor_shift_type POSITIVE_DEFINITE > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Error in L2-norm: > 4.55137e-02 > 3.28222e-03 > 2.24076e-04 > 1.41531e-05 > 8.87205e-07 > > Error in H1-norm: > 4.72542e-01 > 6.63469e-02 > 9.33978e-03 > 1.18329e-03 > 1.48449e-04 > > > Convergence in L2-norm: > 3.793556 > 3.872613 > 3.984794 > 3.995709 > > Convergence in H1-norm: > 2.832342 > 2.828567 > 2.980589 > 2.994760 But then this call gives a completely different result. > student-113-wless94-099:NonlinearAdvectionDiffusion justindong$ ./main -pc_factor_shift_type POSITIVE_DEFINITE > Iteration: 1, Error: 1.116824e-01 > Iteration: 2, Error: 1.788701e-14 > Iterations for convergence: 2 > Iteration: 1, Error: 4.592120e-02 > Iteration: 2, Error: 6.520052e-03 > Iteration: 3, Error: 0.000000e+00 > Iterations for convergence: 3 > Iteration: 1, Error: 4.706179e-02 > Iteration: 2, Error: 0.000000e+00 > Iterations for convergence: 2 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Error in L2-norm: > 1.20600e-01 > 4.33519e-02 > 4.70623e-02 > 1.41531e-05 > 8.87205e-07 > > Error in H1-norm: > 6.04173e-01 > 1.72409e-01 > 1.82071e-01 > 1.18329e-03 > 1.48449e-04 > > > Convergence in L2-norm: > 1.476068 > -0.118478 > 11.699236 > 3.995709 > > Convergence in H1-norm: > 1.809128 > -0.078670 > 7.265561 > 2.994760 Since I?m still just testing various examples using my code, I could just run it until it gives me the correct result, but that seems silly. Is there anyway I can get consistent results? I know I can specify the amount of the shift, but since I?m not well versed in these solvers, I?m not sure what to specify. -------------- next part -------------- An HTML attachment was scrubbed... URL: From knepley at gmail.com Fri Nov 29 07:52:44 2013 From: knepley at gmail.com (Matthew Knepley) Date: Fri, 29 Nov 2013 07:52:44 -0600 Subject: [petsc-users] PC factor shift type, random results? In-Reply-To: <57D3B071-4A66-4F97-80A4-F2ABB6185619@rice.edu> References: <57D3B071-4A66-4F97-80A4-F2ABB6185619@rice.edu> Message-ID: On Fri, Nov 29, 2013 at 1:11 AM, Justin Dong wrote: > Hi all, > > I?m solving a non-linear advection-diffusion problem in 2D and am testing > the routine on various meshes. The non-linearity is handled via Newton?s > method. If I consider the unit square [0,1]^2, my results were fine. I > tried testing on [-1,1]^2 and got either garbage results or a warning about > zero pivots. > > I thought my Jacobian might be computed incorrectly, but I think that > isn?t the case since I applied -pc-factor_shift_type POSITIVE DEFINITE and > miraculously got the > I still think your code has a problem. There should be no indefiniteness with the discretizations i am familiar with. Simplify the problem until you understand everything. Make it linear diffusion on the new grid. Should be exactly the same as the old grid. If you get indefiniteness when adding advection, go to the element that causes it and see what is happening. None of the methods you use are non-deterministic, so I am guessing you have an uninitialized variable. Check with valgrind. Matt > correct result now. But in applying this factor shift over various trials > on the same problem, the results seem to be pretty random. Since I am > testing on manufactured solutions, I force the problem to be linear since I > have prior knowledge of the exact solution and Newton?s method should > converge in one iteration. For instance, this is what the result should be > (iterations and errors are for each Newton iteration) > > student-113-wless94-099:NonlinearAdvectionDiffusion justindong$ ./main > -pc_factor_shift_type POSITIVE_DEFINITE > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Error in L2-norm: > 4.55137e-02 > 3.28222e-03 > 2.24076e-04 > 1.41531e-05 > 8.87205e-07 > > Error in H1-norm: > 4.72542e-01 > 6.63469e-02 > 9.33978e-03 > 1.18329e-03 > 1.48449e-04 > > > Convergence in L2-norm: > 3.793556 > 3.872613 > 3.984794 > 3.995709 > > Convergence in H1-norm: > 2.832342 > 2.828567 > 2.980589 > 2.994760 > > > But then this call gives a completely different result. > > student-113-wless94-099:NonlinearAdvectionDiffusion justindong$ ./main > -pc_factor_shift_type POSITIVE_DEFINITE > Iteration: 1, Error: 1.116824e-01 > Iteration: 2, Error: 1.788701e-14 > Iterations for convergence: 2 > Iteration: 1, Error: 4.592120e-02 > Iteration: 2, Error: 6.520052e-03 > Iteration: 3, Error: 0.000000e+00 > Iterations for convergence: 3 > Iteration: 1, Error: 4.706179e-02 > Iteration: 2, Error: 0.000000e+00 > Iterations for convergence: 2 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Iteration: 1, Error: 0.000000e+00 > Iterations for convergence: 1 > Error in L2-norm: > 1.20600e-01 > 4.33519e-02 > 4.70623e-02 > 1.41531e-05 > 8.87205e-07 > > Error in H1-norm: > 6.04173e-01 > 1.72409e-01 > 1.82071e-01 > 1.18329e-03 > 1.48449e-04 > > > Convergence in L2-norm: > 1.476068 > -0.118478 > 11.699236 > 3.995709 > > Convergence in H1-norm: > 1.809128 > -0.078670 > 7.265561 > 2.994760 > > > > Since I?m still just testing various examples using my code, I could just > run it until it gives me the correct result, but that seems silly. Is there > anyway I can get consistent results? I know I can specify the amount of the > shift, but since I?m not well versed in these solvers, I?m not sure what to > specify. > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From nico.schloemer at gmail.com Fri Nov 29 10:43:12 2013 From: nico.schloemer at gmail.com (=?ISO-8859-1?Q?Nico_Schl=F6mer?=) Date: Fri, 29 Nov 2013 17:43:12 +0100 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! In-Reply-To: <87eh60eadt.fsf@jedbrown.org> References: <87pppkeefe.fsf@jedbrown.org> <87eh60eadt.fsf@jedbrown.org> Message-ID: Hm, I tried ================ *snip* ================ IS is; ierr = ISCreateGeneral(PETSC_COMM_WORLD, u_indices.size(), u_indices.data(), PETSC_USE_POINTER, &is); PCFieldSplitSetIS(pc, "u", is); IS p_is; ierr = ISCreateGeneral(PETSC_COMM_WORLD, p_indices.size(), p_indices.data(), PETSC_USE_POINTER, &p_is); PCFieldSplitSetIS(pc, "p", p_is); ================ *snap* ================ now (where u_indices, p_indices are set to {0,1,3,4,6,7...}, {2,5,8,...}), but I'm still getting ================ *snip* ================ [0]PETSC ERROR: Petsc has generated inconsistent data! [0]PETSC ERROR: Unhandled case, must have at least two fields, not 1! ================ *snap* ================ What would I have to look at to debug this? --Nico On Thu, Nov 28, 2013 at 5:53 PM, Jed Brown wrote: > Nico Schl?mer writes: > >> Hi Jed, >> >> thanks for the quick answer! >> I can of course generate the strided index sets (01x34x6..., >> xx2xx5x...) myself and insert them via PCFieldSplitSetIS(), but I >> thought that ultimately does the same thing that I would have liked >> PCFieldSplitSetFields() to do for me. Not true? > > PCFieldSplitSetFields is something you call only once to identify all > the fields. It doesn't let you group multiple strides in one, so you > could create "ux", "uy", and "p", but then those would be separate. You > can create splits using numbers, -fieldsplit_block_size 3 > -fieldsplit_0_fields 0,1 -fieldsplit_1_fields 2 (assuming collocated > discretization), but this can be rather opaque to maintain. From knepley at gmail.com Fri Nov 29 10:53:28 2013 From: knepley at gmail.com (Matthew Knepley) Date: Fri, 29 Nov 2013 10:53:28 -0600 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! In-Reply-To: References: <87pppkeefe.fsf@jedbrown.org> <87eh60eadt.fsf@jedbrown.org> Message-ID: On Fri, Nov 29, 2013 at 10:43 AM, Nico Schl?mer wrote: > Hm, I tried > > ================ *snip* ================ > IS is; > ierr = ISCreateGeneral(PETSC_COMM_WORLD, u_indices.size(), > u_indices.data(), > PETSC_USE_POINTER, &is); > PCFieldSplitSetIS(pc, "u", is); > IS p_is; > ierr = ISCreateGeneral(PETSC_COMM_WORLD, p_indices.size(), > p_indices.data(), > PETSC_USE_POINTER, &p_is); > PCFieldSplitSetIS(pc, "p", p_is); > ================ *snap* ================ > > now (where u_indices, p_indices are set to {0,1,3,4,6,7...}, > {2,5,8,...}), but I'm still getting > > ================ *snip* ================ > [0]PETSC ERROR: Petsc has generated inconsistent data! > [0]PETSC ERROR: Unhandled case, must have at least two fields, not 1! > ================ *snap* ================ > > What would I have to look at to debug this? > The PC is getting recreated somewhere in the intervening code. At least this makes the most sense to me. I would get in the debugger and break in PCCreate_FieldSplit. Matt > --Nico > > On Thu, Nov 28, 2013 at 5:53 PM, Jed Brown wrote: > > Nico Schl?mer writes: > > > >> Hi Jed, > >> > >> thanks for the quick answer! > >> I can of course generate the strided index sets (01x34x6..., > >> xx2xx5x...) myself and insert them via PCFieldSplitSetIS(), but I > >> thought that ultimately does the same thing that I would have liked > >> PCFieldSplitSetFields() to do for me. Not true? > > > > PCFieldSplitSetFields is something you call only once to identify all > > the fields. It doesn't let you group multiple strides in one, so you > > could create "ux", "uy", and "p", but then those would be separate. You > > can create splits using numbers, -fieldsplit_block_size 3 > > -fieldsplit_0_fields 0,1 -fieldsplit_1_fields 2 (assuming collocated > > discretization), but this can be rather opaque to maintain. > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Fri Nov 29 10:56:37 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 29 Nov 2013 10:56:37 -0600 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! In-Reply-To: References: <87pppkeefe.fsf@jedbrown.org> <87eh60eadt.fsf@jedbrown.org> Message-ID: <87fvqfcfju.fsf@jedbrown.org> Nico Schl?mer writes: > Hm, I tried > > ================ *snip* ================ > IS is; > ierr = ISCreateGeneral(PETSC_COMM_WORLD, u_indices.size(), u_indices.data(), > PETSC_USE_POINTER, &is); > PCFieldSplitSetIS(pc, "u", is); > IS p_is; > ierr = ISCreateGeneral(PETSC_COMM_WORLD, p_indices.size(), p_indices.data(), > PETSC_USE_POINTER, &p_is); > PCFieldSplitSetIS(pc, "p", p_is); > ================ *snap* ================ > > now (where u_indices, p_indices are set to {0,1,3,4,6,7...}, > {2,5,8,...}), but I'm still getting > > ================ *snip* ================ > [0]PETSC ERROR: Petsc has generated inconsistent data! > [0]PETSC ERROR: Unhandled case, must have at least two fields, not 1! > ================ *snap* ================ Are you also calling PCFieldSplitSetFields? Always send the entire error message. There are a few examples that do this, e.g., src/snes/examples/tutorials/ex28.c and src/ksp/ksp/examples/tests/ex11.c. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From iwaddington at gmail.com Fri Nov 29 12:13:18 2013 From: iwaddington at gmail.com (iwaddington .) Date: Fri, 29 Nov 2013 16:13:18 -0200 Subject: [petsc-users] Inhomogeneous laplacian Message-ID: Hi, I am trying to optimize a code that solves an inhomogeneous laplacian pde of the form : -div ( K(x) grad(u) ) = f 0 < x,y < 1 where K(x) is a matrix that depends on the position I use finite differences and the values of "f" and "u" are cell centered. The problem is the following, since each interface is common to two cells, and I need the value of K on the interfaces, which is taken as an harmonic mean, I would like to compute these values only once, and then pass them as some user context to KSPSetComputeOperators, because I am using a dmda. Any ideas ? -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Fri Nov 29 12:20:30 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 29 Nov 2013 12:20:30 -0600 Subject: [petsc-users] Inhomogeneous laplacian In-Reply-To: References: Message-ID: <8761rbcbo1.fsf@jedbrown.org> "iwaddington ." writes: > Hi, I am trying to optimize a code that solves an inhomogeneous laplacian > pde of the form : > > -div ( K(x) grad(u) ) = f 0 < x,y < 1 where K(x) is a > matrix that > > depends on the position > > I use finite differences and the values of "f" and "u" are cell centered. > > The problem is the following, since each interface is common to two cells, > and I need the value of K on the interfaces, which is taken as an harmonic > mean, This is one choice, and not a perfect one. > I would like to compute these values only once, and then pass them as > some user context to KSPSetComputeOperators, because I am using a > dmda. Any ideas ? You can create a DMDA to hold those staggered coefficients. Note that it may be faster to just recompute the staggered values from cell-centered values, since it involves a bit less memory bandwidth, and memory bandwidth is the main performance bottleneck for these things. That also makes it easier to choose different flux definitions (e.g., for unaligned anisotropy, you'll need at least a 9-point stencil). -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From iwaddington at gmail.com Fri Nov 29 12:34:02 2013 From: iwaddington at gmail.com (iwaddington .) Date: Fri, 29 Nov 2013 16:34:02 -0200 Subject: [petsc-users] Inhomogeneous laplacian In-Reply-To: <8761rbcbo1.fsf@jedbrown.org> References: <8761rbcbo1.fsf@jedbrown.org> Message-ID: Thanks Jed, so if I understood correctly in fact there is nothing to be done, since I was already recomputing the staggered values when passing from one node to its neighbour. And good Black Friday to you. On Fri, Nov 29, 2013 at 4:20 PM, Jed Brown wrote: > "iwaddington ." writes: > > > Hi, I am trying to optimize a code that solves an inhomogeneous laplacian > > pde of the form : > > > > -div ( K(x) grad(u) ) = f 0 < x,y < 1 where K(x) is > a > > matrix that > > > > depends on the position > > > > I use finite differences and the values of "f" and "u" are cell > centered. > > > > The problem is the following, since each interface is common to two > cells, > > and I need the value of K on the interfaces, which is taken as an > harmonic > > mean, > > This is one choice, and not a perfect one. > > > I would like to compute these values only once, and then pass them as > > some user context to KSPSetComputeOperators, because I am using a > > dmda. Any ideas ? > > You can create a DMDA to hold those staggered coefficients. Note that > it may be faster to just recompute the staggered values from > cell-centered values, since it involves a bit less memory bandwidth, and > memory bandwidth is the main performance bottleneck for these things. > That also makes it easier to choose different flux definitions (e.g., > for unaligned anisotropy, you'll need at least a 9-point stencil). > -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Fri Nov 29 12:43:04 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 29 Nov 2013 12:43:04 -0600 Subject: [petsc-users] Inhomogeneous laplacian In-Reply-To: References: <8761rbcbo1.fsf@jedbrown.org> Message-ID: <8738mfcamf.fsf@jedbrown.org> "iwaddington ." writes: > Thanks Jed, so if I understood correctly in fact there is nothing to be > done, since I was already recomputing the staggered values when passing > from one node to its neighbour. I think if you benchmark and don't do something spuriously wasteful, you'll find that computing the averages is not really a bottleneck. What happens if you replace the harmonic average with an arithmetic average (trivially cheap)? If that's a bottleneck, you can unroll-and-jam your stencil operation and vectorize the division. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From yelpoo at gmail.com Fri Nov 29 15:05:44 2013 From: yelpoo at gmail.com (Jianbo Ye) Date: Fri, 29 Nov 2013 16:05:44 -0500 Subject: [petsc-users] static linking MKL without mpi Message-ID: Hi, I am trying to build static lib petsc by static linking MKL without mpi. Here is a command I use initially: ./configure PETSC_ARCH=linux-gnu-c-static-seq --with-clanguage=cxx -with-fc=0 --with-mpi=0 --with-blas-lapack-dir=$(MKLROOT)/lib/intel64 --download-umfpack --download-blacs --with-shared-libraries=0 --with-dynamic-loading=0 It is however automatically create a linking option which link MKL dynamically. BLAS/LAPACK: -Wl,-rpath,/gpfs/apps/x86_64-rhel5/intel/composer_xe/2013.0.079/mkl/lib/intel64 -L/gpfs/apps/x86_64-rhel5/intel/composer_xe/2013.0.079/mkl/lib/intel64 -lmkl_intel_lp64 -lmkl_sequential -lmkl_core -lpthread -lm But if we want to static link MKL, using option like -Wl,--start-group $(MKLROOT)/lib/intel64/libmkl_intel_lp64.a $(MKLROOT)/lib/intel64/libmkl_core.a $(MKLROOT)/lib/intel64/libmkl_sequential.a -Wl,--end-group -lpthread -lm it is suggested to use -with-blas-lapack-lib But I don't really know how to do that, it constantly gives error: invalid libraries Could anyone help on this? Thanks. Jianbo -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Fri Nov 29 15:26:35 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 29 Nov 2013 15:26:35 -0600 Subject: [petsc-users] static linking MKL without mpi In-Reply-To: References: Message-ID: <87pppic31w.fsf@jedbrown.org> Jianbo Ye writes: > Hi, I am trying to build static lib petsc by static linking MKL without > mpi. Here is a command I use initially: > > ./configure PETSC_ARCH=linux-gnu-c-static-seq --with-clanguage=cxx > -with-fc=0 --with-mpi=0 --with-blas-lapack-dir=$(MKLROOT)/lib/intel64 > --download-umfpack --download-blacs --with-shared-libraries=0 > --with-dynamic-loading=0 > > It is however automatically create a linking option which link MKL > dynamically. > > BLAS/LAPACK: > -Wl,-rpath,/gpfs/apps/x86_64-rhel5/intel/composer_xe/2013.0.079/mkl/lib/intel64 > -L/gpfs/apps/x86_64-rhel5/intel/composer_xe/2013.0.079/mkl/lib/intel64 > -lmkl_intel_lp64 -lmkl_sequential -lmkl_core -lpthread -lm > > > But if we want to static link MKL, using option like > > -Wl,--start-group $(MKLROOT)/lib/intel64/libmkl_intel_lp64.a > $(MKLROOT)/lib/intel64/libmkl_core.a > $(MKLROOT)/lib/intel64/libmkl_sequential.a -Wl,--end-group -lpthread -lm > > it is suggested to use -with-blas-lapack-lib > But I don't really know how to do that, it constantly gives error: invalid > libraries You have to tell us exactly what you are passing, and send configure.log. Did you try the following? --with-blas-lapack-lib="-Wl,--start-group $(MKLROOT)/lib/intel64/libmkl_intel_lp64.a $(MKLROOT)/lib/intel64/libmkl_core.a $(MKLROOT)/lib/intel64/libmkl_sequential.a -Wl,--end-group -lpthread -lm" -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From mfadams at lbl.gov Fri Nov 29 21:33:22 2013 From: mfadams at lbl.gov (Mark Adams) Date: Fri, 29 Nov 2013 22:33:22 -0500 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! In-Reply-To: References: Message-ID: > > > > [0]PETSC ERROR: Petsc has generated inconsistent data! > [0]PETSC ERROR: Unhandled case, must have at least two fields, not 1 This can happen if you give fieldsplit options to a non-fieldsplit matrix/PC. So so calling XXXSetFromOptions with a KSP/PC and an AIJ Mat (ie, non-nested Mat) and using a parameter like '-pc_fieldsplit_type schur', will give you this error. -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Fri Nov 29 21:42:00 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Fri, 29 Nov 2013 21:42:00 -0600 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! In-Reply-To: References: Message-ID: <87d2liblo7.fsf@jedbrown.org> Mark Adams writes: >> [0]PETSC ERROR: Petsc has generated inconsistent data! >> [0]PETSC ERROR: Unhandled case, must have at least two fields, not 1 > > > This can happen if you give fieldsplit options to a non-fieldsplit > matrix/PC. So so calling XXXSetFromOptions with a KSP/PC and an AIJ Mat > (ie, non-nested Mat) and using a parameter like '-pc_fieldsplit_type > schur', will give you this error. It should not error. FieldSplit existed years before MatNest and there are lots of example that use it without MatNest (which is merely an optimization, sometimes). If you are getting this error, we need a complete stack trace and instructions for how to reproduce (ideally using a PETSc example). -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From gideon.simpson at gmail.com Sat Nov 30 14:19:42 2013 From: gideon.simpson at gmail.com (Gideon Simpson) Date: Sat, 30 Nov 2013 15:19:42 -0500 Subject: [petsc-users] petsc vector management for monte carlo Message-ID: <76D0CDF7-BEBB-4447-B22F-D98A392F5BAE@gmail.com> I'm migrating some monte carlo code over to petsc, where the realizations will be done serially, but distributed across the cluster. I'm inclined to involve petsc in this for its data management, and had a couple of questions. If I plant to store the (scalar) results of the monte carlo simulations in a Petsc vector v, what is the recommended method for putting the results into this vector? Currently, I have implemented: VecGetOwnershipRange(sim_data, &local_start, &local_end); VecGetArray(sim_data,& sim_data_local); k=0; for(i=local_start;i References: <76D0CDF7-BEBB-4447-B22F-D98A392F5BAE@gmail.com> Message-ID: <87vbz9aati.fsf@jedbrown.org> Gideon Simpson writes: > I'm migrating some monte carlo code over to petsc, where the realizations will be done serially, but distributed across the cluster. I'm inclined to involve petsc in this for its data management, and had a couple of questions. > > If I plant to store the (scalar) results of the monte carlo simulations in a Petsc vector v, what is the recommended method for putting the results into this vector? Currently, I have implemented: > > VecGetOwnershipRange(sim_data, &local_start, &local_end); > > VecGetArray(sim_data,& sim_data_local); > k=0; > > for(i=local_start;i > sim_result = mc_sim_cod(mc_args); > > sim_data_local[k] = sim_result; > k++; > > } > VecRestoreArray(sim_data, & sim_data_local); > > Is this a *good* solution, or should I be using SetValue routines? The above is fine. > Also, I have MC simulations which, instead of returning a single > scalar, return some n-dimensional array of real values. In that case, > I was thinking to make sim_data a petsc matrix. If each result was > n-dimensional, how would I initialize the matrix to hold M > realizations of this n dimensional matrix? ensuring that each set of > n was contiguous on the local machine? Mat is really for linear operators, not a container for data that you want to interpret as 2-dimensional. DMDA can do the latter, especially if you want to distribute. > Alternatively, this could be an n*M vector, but again, I'd need to > ensure each chunk of n was located on the same machine? Just set the local sizes of the Vec such that the entire subset is local. See the middle argument to VecSetSizes(). > I presume here, I would use SetValue routines to insert each chunk of > data. You could, but writing directly into the array is also fine. > Also, this may have to do with it being the newest matlab, 2013b, but when I try to load petsc vectors i receive the following error: > > Error using load > Number of columns on line 2 of ASCII file sim_data.out > must be the same as previous lines. > > where I constructed the sim_data.out file using the lines: > > PetscViewerASCIIOpen(PETSC_COMM_WORLD, "sim_data.out", &output_viewer); > PetscViewerSetFormat(output_viewer, PETSC_VIEWER_ASCII_MATLAB); > VecView(sim_data,output_viewer); > PetscViewerDestroy(&output_viewer); > > Looking at the ascii file, it looks fine. Did MATLAB change their ASCII format? Anyway, it is much better to write binary output and read with PetscBinaryRead (bin/matlab/PetscBinaryRead.m). -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From gideon.simpson at gmail.com Sat Nov 30 14:56:54 2013 From: gideon.simpson at gmail.com (Gideon Simpson) Date: Sat, 30 Nov 2013 15:56:54 -0500 Subject: [petsc-users] petsc vector management for monte carlo In-Reply-To: <87vbz9aati.fsf@jedbrown.org> References: <76D0CDF7-BEBB-4447-B22F-D98A392F5BAE@gmail.com> <87vbz9aati.fsf@jedbrown.org> Message-ID: Thanks Jed, see a few follow up comments below, On Nov 30, 2013, at 3:34 PM, Jed Brown wrote: > Gideon Simpson writes: > >> I'm migrating some monte carlo code over to petsc, where the realizations will be done serially, but distributed across the cluster. I'm inclined to involve petsc in this for its data management, and had a couple of questions. >> >> If I plant to store the (scalar) results of the monte carlo simulations in a Petsc vector v, what is the recommended method for putting the results into this vector? Currently, I have implemented: >> >> VecGetOwnershipRange(sim_data, &local_start, &local_end); >> >> VecGetArray(sim_data,& sim_data_local); >> k=0; >> >> for(i=local_start;i> >> sim_result = mc_sim_cod(mc_args); >> >> sim_data_local[k] = sim_result; >> k++; >> >> } >> VecRestoreArray(sim_data, & sim_data_local); >> >> Is this a *good* solution, or should I be using SetValue routines? > > The above is fine. > >> Also, I have MC simulations which, instead of returning a single >> scalar, return some n-dimensional array of real values. In that case, >> I was thinking to make sim_data a petsc matrix. If each result was >> n-dimensional, how would I initialize the matrix to hold M >> realizations of this n dimensional matrix? ensuring that each set of >> n was contiguous on the local machine? > > Mat is really for linear operators, not a container for data that you > want to interpret as 2-dimensional. DMDA can do the latter, especially > if you want to distribute. > >> Alternatively, this could be an n*M vector, but again, I'd need to >> ensure each chunk of n was located on the same machine? > > Just set the local sizes of the Vec such that the entire subset is > local. See the middle argument to VecSetSizes(). If each realization generated a size n output, is there something slicker than this: VecGetOwnershipRange(sim_data, &local_start, &local_end); VecGetArray(sim_data,& sim_data_local); k=0; m_local = (local_end - local_start-1)/n; for(i=0;i< m_local;i++){ mc_sim_cod(mc_args, &sim_result); for(k = 0;k >> I presume here, I would use SetValue routines to insert each chunk of >> data. > > You could, but writing directly into the array is also fine. > >> Also, this may have to do with it being the newest matlab, 2013b, but when I try to load petsc vectors i receive the following error: >> >> Error using load >> Number of columns on line 2 of ASCII file sim_data.out >> must be the same as previous lines. >> >> where I constructed the sim_data.out file using the lines: >> >> PetscViewerASCIIOpen(PETSC_COMM_WORLD, "sim_data.out", &output_viewer); >> PetscViewerSetFormat(output_viewer, PETSC_VIEWER_ASCII_MATLAB); >> VecView(sim_data,output_viewer); >> PetscViewerDestroy(&output_viewer); >> >> Looking at the ascii file, it looks fine. > > Did MATLAB change their ASCII format? Here's a quick test on 2013b: v = linspace(0,1,5); save('test.out', 'v', '-ascii'); test.out looks like: 0.0000000e+00 2.5000000e-01 5.0000000e-01 7.5000000e-01 1.0000000e+00 > > Anyway, it is much better to write binary output and read with > PetscBinaryRead (bin/matlab/PetscBinaryRead.m). From knepley at gmail.com Sat Nov 30 15:02:57 2013 From: knepley at gmail.com (Matthew Knepley) Date: Sat, 30 Nov 2013 15:02:57 -0600 Subject: [petsc-users] petsc vector management for monte carlo In-Reply-To: References: <76D0CDF7-BEBB-4447-B22F-D98A392F5BAE@gmail.com> <87vbz9aati.fsf@jedbrown.org> Message-ID: On Sat, Nov 30, 2013 at 2:56 PM, Gideon Simpson wrote: > Thanks Jed, see a few follow up comments below, > On Nov 30, 2013, at 3:34 PM, Jed Brown wrote: > > > Gideon Simpson writes: > > > >> I'm migrating some monte carlo code over to petsc, where the > realizations will be done serially, but distributed across the cluster. > I'm inclined to involve petsc in this for its data management, and had a > couple of questions. > >> > >> If I plant to store the (scalar) results of the monte carlo simulations > in a Petsc vector v, what is the recommended method for putting the results > into this vector? Currently, I have implemented: > >> > >> VecGetOwnershipRange(sim_data, &local_start, &local_end); > >> > >> VecGetArray(sim_data,& sim_data_local); > >> k=0; > >> > >> for(i=local_start;i >> > >> sim_result = mc_sim_cod(mc_args); > >> > >> sim_data_local[k] = sim_result; > >> k++; > >> > >> } > >> VecRestoreArray(sim_data, & sim_data_local); > >> > >> Is this a *good* solution, or should I be using SetValue routines? > > > > The above is fine. > > > >> Also, I have MC simulations which, instead of returning a single > >> scalar, return some n-dimensional array of real values. In that case, > >> I was thinking to make sim_data a petsc matrix. If each result was > >> n-dimensional, how would I initialize the matrix to hold M > >> realizations of this n dimensional matrix? ensuring that each set of > >> n was contiguous on the local machine? > > > > Mat is really for linear operators, not a container for data that you > > want to interpret as 2-dimensional. DMDA can do the latter, especially > > if you want to distribute. > > > >> Alternatively, this could be an n*M vector, but again, I'd need to > >> ensure each chunk of n was located on the same machine? > > > > Just set the local sizes of the Vec such that the entire subset is > > local. See the middle argument to VecSetSizes(). > > If each realization generated a size n output, is there something slicker > than this: > > VecGetOwnershipRange(sim_data, &local_start, &local_end); > > VecGetArray(sim_data,& sim_data_local); > k=0; > > m_local = (local_end - local_start-1)/n; > > for(i=0;i< m_local;i++){ > > mc_sim_cod(mc_args, &sim_result); > > for(k = 0;k > sim_data_local[i * n + k] = sim_result[k]; > } > > } > VecRestoreArray(sim_data, & sim_data_local); > > I see expressions like sim_data_local[i * n + k] to handle the indexing > as bit awkward. > I guess it depends on what you want to do. You could have n independent serial vectors. If you want to maintain the global vector, you can always define the pointer at the top of the loop. Thanks, Matt > > > >> I presume here, I would use SetValue routines to insert each chunk of > >> data. > > > > You could, but writing directly into the array is also fine. > > > >> Also, this may have to do with it being the newest matlab, 2013b, but > when I try to load petsc vectors i receive the following error: > >> > >> Error using load > >> Number of columns on line 2 of ASCII file sim_data.out > >> must be the same as previous lines. > >> > >> where I constructed the sim_data.out file using the lines: > >> > >> PetscViewerASCIIOpen(PETSC_COMM_WORLD, "sim_data.out", &output_viewer); > >> PetscViewerSetFormat(output_viewer, PETSC_VIEWER_ASCII_MATLAB); > >> VecView(sim_data,output_viewer); > >> PetscViewerDestroy(&output_viewer); > >> > >> Looking at the ascii file, it looks fine. > > > > Did MATLAB change their ASCII format? > > Here's a quick test on 2013b: > v = linspace(0,1,5); > save('test.out', 'v', '-ascii'); > > test.out looks like: > 0.0000000e+00 2.5000000e-01 5.0000000e-01 7.5000000e-01 > 1.0000000e+00 > > > > > > Anyway, it is much better to write binary output and read with > > PetscBinaryRead (bin/matlab/PetscBinaryRead.m). > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Sat Nov 30 15:12:42 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Sat, 30 Nov 2013 15:12:42 -0600 Subject: [petsc-users] petsc vector management for monte carlo In-Reply-To: References: <76D0CDF7-BEBB-4447-B22F-D98A392F5BAE@gmail.com> <87vbz9aati.fsf@jedbrown.org> Message-ID: <87pppha911.fsf@jedbrown.org> Gideon Simpson writes: > If each realization generated a size n output, is there something slicker than this: > > VecGetOwnershipRange(sim_data, &local_start, &local_end); > > VecGetArray(sim_data,& sim_data_local); > k=0; > > m_local = (local_end - local_start-1)/n; local_end is one past the last owned entry. > for(i=0;i< m_local;i++){ > > mc_sim_cod(mc_args, &sim_result); > > for(k = 0;k > sim_data_local[i * n + k] = sim_result[k]; > } > > } You can drop the result directly into the array: for (i=local_start; i> Did MATLAB change their ASCII format? > > Here's a quick test on 2013b: > v = linspace(0,1,5); > save('test.out', 'v', '-ascii'); > > test.out looks like: > 0.0000000e+00 2.5000000e-01 5.0000000e-01 7.5000000e-01 1.0000000e+00 ASCII_MATLAB output is meant to be sourced: v = [ 0.0000000e+00 2.5000000e-01 5.0000000e-01 7.5000000e-01 1.0000000e+00 ]; >> Anyway, it is much better to write binary output and read with >> PetscBinaryRead (bin/matlab/PetscBinaryRead.m). But seriously, use binary. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From gideon.simpson at gmail.com Sat Nov 30 15:17:31 2013 From: gideon.simpson at gmail.com (Gideon Simpson) Date: Sat, 30 Nov 2013 16:17:31 -0500 Subject: [petsc-users] petsc vector management for monte carlo In-Reply-To: <87pppha911.fsf@jedbrown.org> References: <76D0CDF7-BEBB-4447-B22F-D98A392F5BAE@gmail.com> <87vbz9aati.fsf@jedbrown.org> <87pppha911.fsf@jedbrown.org> Message-ID: <9899E8FF-0DCD-4800-A343-F9E357662BD6@gmail.com> Thanks guys, and I will switch over to binary. Just saying, that's what 2013b is doing. -gideon On Nov 30, 2013, at 4:12 PM, Jed Brown wrote: > Gideon Simpson writes: > >> If each realization generated a size n output, is there something slicker than this: >> >> VecGetOwnershipRange(sim_data, &local_start, &local_end); >> >> VecGetArray(sim_data,& sim_data_local); >> k=0; >> >> m_local = (local_end - local_start-1)/n; > > local_end is one past the last owned entry. > >> for(i=0;i< m_local;i++){ >> >> mc_sim_cod(mc_args, &sim_result); >> >> for(k = 0;k> >> sim_data_local[i * n + k] = sim_result[k]; >> } >> >> } > > You can drop the result directly into the array: > > for (i=local_start; i mc_sim_cod(mc_args, sim_data_local+(i-local_start)); > } > >>> Did MATLAB change their ASCII format? >> >> Here's a quick test on 2013b: >> v = linspace(0,1,5); >> save('test.out', 'v', '-ascii'); >> >> test.out looks like: >> 0.0000000e+00 2.5000000e-01 5.0000000e-01 7.5000000e-01 1.0000000e+00 > > ASCII_MATLAB output is meant to be sourced: > > v = [ > 0.0000000e+00 > 2.5000000e-01 > 5.0000000e-01 > 7.5000000e-01 > 1.0000000e+00 > ]; > >>> Anyway, it is much better to write binary output and read with >>> PetscBinaryRead (bin/matlab/PetscBinaryRead.m). > > But seriously, use binary. -------------- next part -------------- An HTML attachment was scrubbed... URL: From jedbrown at mcs.anl.gov Sat Nov 30 15:28:46 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Sat, 30 Nov 2013 15:28:46 -0600 Subject: [petsc-users] petsc vector management for monte carlo In-Reply-To: <9899E8FF-0DCD-4800-A343-F9E357662BD6@gmail.com> References: <76D0CDF7-BEBB-4447-B22F-D98A392F5BAE@gmail.com> <87vbz9aati.fsf@jedbrown.org> <87pppha911.fsf@jedbrown.org> <9899E8FF-0DCD-4800-A343-F9E357662BD6@gmail.com> Message-ID: <87mwkla8a9.fsf@jedbrown.org> Gideon Simpson writes: > Thanks guys, and I will switch over to binary. Just saying, that's what 2013b is doing. I don't think the ASCII_MATLAB format was ever intended to be loaded with "load". You should be able to source it. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: From mfadams at lbl.gov Sat Nov 30 17:26:21 2013 From: mfadams at lbl.gov (Mark Adams) Date: Sat, 30 Nov 2013 18:26:21 -0500 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! In-Reply-To: <87d2liblo7.fsf@jedbrown.org> References: <87d2liblo7.fsf@jedbrown.org> Message-ID: > > > It should not error. I tried putting what I thought was the problem in ex54 and you are correct it did not cause a problem. I've attached a good and bad "petsc.rc" file from my app and an output file with this error. Perhaps you can see how this broke. -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: tt.tar Type: application/x-tar Size: 112640 bytes Desc: not available URL: From jedbrown at mcs.anl.gov Sat Nov 30 17:33:19 2013 From: jedbrown at mcs.anl.gov (Jed Brown) Date: Sat, 30 Nov 2013 17:33:19 -0600 Subject: [petsc-users] fieldsplit: Unhandled case, must have at least two fields, not 1! In-Reply-To: References: <87d2liblo7.fsf@jedbrown.org> Message-ID: <87haata2io.fsf@jedbrown.org> Mark Adams writes: > I tried putting what I thought was the problem in ex54 and you are correct > it did not cause a problem. Okay. > I've attached a good and bad "petsc.rc" file from my app and an output file > with this error. Perhaps you can see how this broke. I don't have your app so I don't know what that code looks like. The two options files are very different. Presumably you can run on one process and use a debugger to see what happens in PCFieldSplitSetDefaults. If you think PETSc is mishandling this, please add a test to PETSc (perhaps a small addition to an existing test) to demonstrate the issue. -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: application/pgp-signature Size: 835 bytes Desc: not available URL: