Code structuring - Communicator

Barry Smith bsmith at
Fri May 9 14:07:10 CDT 2008

    There are many ways to do this, most of them involve using MPI to  
construct subcommunicators
for the various sub parallel tasks. You very likely want to keep  
PetscInitialize() at
the very beginning of the program; you would not write the calls in  
terms of
PETSC_COMM_WORLD or MPI_COMM_WORLD, rather you would use the
subcommunicators to create the objects.

    An alternative approach is to look at the manual page for  
PetscOpenMPMerge(), PetscOpenMPRun(),
PetscOpenMPNew() in petsc-dev. These allow a simple master-worker  
model of parallelism
with PETSc with a bunch of masters that can work together (instead of  
just one master) and each
master controls a bunch of workers. The code in src/ksp/pc/impls/ 
openmp uses this code.

Note that OpenMP has NOTHING to do with OpenMP the standard. Also I  
don't really have
any support for Fortran, I hope you use C/C++. Comments welcome. It  
sounds like this matches
what you need. It's pretty cool,  but underdeveloped.


On May 9, 2008, at 12:46 PM, Amit.Itagi at wrote:

> Hi,
> I have a question about the Petsc communicator. I have a petsc program
> "foo" which essentially runs in parallel and gives me  
> y=f(x1,x2,...), where
> y is an output parameter and xi's are input parameters. Suppose, I  
> want to
> run a parallel optimizer for the input parameters. I am looking at the
> following functionality. I submit the optimizer job on 16 processors  
> (using
> "mpiexec -np 16 progName"). The optimizer should then submit 4 runs of
> "foo", each running parallely on 4 processors. "foo" will be written  
> as a
> function and not as a main program in this case. How can I get this
> functionality using Petsc ? Should PetscInitialize be called in the
> optimizer, or in each foo run ? If PetscInitialize is called in the
> optimizer, is there a way to make the foo function run only on a  
> subset of
> the 16 processors ?
> May be, I haven't done a good job of explaining my problem. Let me  
> know if
> you need any clarifications.
> Thanks
> Rgds,
> Amit

More information about the petsc-users mailing list