Thanks, this I had no chance to know. Is also true for MatPartitioningCreate? Which one should be run collectively and which one only on master? My code is along this line:
MatCreateMPIAdj MatPartitioningCreate MatPartitioningSetAdjacency MatPartitioningSetNParts MatPartitioningApply Many thanks! Dominik On Sat, Oct 22, 2011 at 4:05 PM, Barry Smith <bsmith at mcs.anl.gov> wrote: > > ? You need to create the MPIAdj matrix with PETSC_COMM_SELF ? not COMM_WORLD > > ? Barry > > On Oct 22, 2011, at 8:45 AM, Dominik Szczerba wrote: > >> I do not seem to find any guidance on this subject in the docu. >> I built petsc with chaco and party to attempt sequential partitioning, >> where parmetis fails. >> However, I get the error: >> >> [0]PETSC ERROR: No support for this operation for this object type! >> [0]PETSC ERROR: Distributed matrix format MPIAdj is not supported for >> sequential partitioners! >> >> >> I do not seem to find a sequential equivalent of MatCreateMPIAdj... >> Are there any examples how to perform partitioning sequentially? >> My mesh/graph is located entirely on master. >> >> Thanks a lot, >> Dominik > >
