Barry,

Thank you for the explanation! I finally got where I was wrong - I mixed up ghost nodes and cells, so that trying to have 0 ghost cells I set up 0 ghost nodes which was wrong as you pointed out.

Best regards,
Mikhail


On 07/15/2015 07:19 PM, Barry Smith wrote:
   Mikhail

DMDAGetElements - Gets an array containing the indices (in local coordinates)
                  of all the local elements

   Since it returns the "local indices" of the vertices of the elements it does not make 
sense for a stencil width of zero (with parallel DMDA) since the vertex of the last element on each 
process is actually owned by the next process so it is a ghost vertex on the process but with a 
stencil width of zero there are no ghost vertices hence no "local index" to represent 
that vertex.

   I have changed the code to return an error instead of "garbage" values if 
the stencil width is zero.

   So you should just use a stencil width of 1 if you want to do finite 
elements (or finite differences actually also :-). Stencil width of 0 is only 
for strange situations and doesn't make much sense for PDEs.

   Thanks for reporting the problem.

   Barry


On Jul 15, 2015, at 10:46 AM, Mikhail Artemyev <[email protected]> 
wrote:

Dear all,

Here is a minimal working example that I'm testing:

#include "petscsys.h"
#include "petscdmda.h"

int main(int argc, char **argv)
{
  PetscInitialize(&argc, &argv, 0, 0);

  PetscInt rank;
  MPI_Comm_rank(PETSC_COMM_WORLD, &rank);

  DM da;
  DMDACreate1d(PETSC_COMM_WORLD, DM_BOUNDARY_NONE, -9, 1, 0, NULL, &da);

  PetscInt n_local_cells, n_cell_nodes;
  const PetscInt *local_indices;
  DMDAGetElements(da, &n_local_cells, &n_cell_nodes, &local_indices);

  PetscSynchronizedPrintf(PETSC_COMM_WORLD,
                          "rank %d n_local_cells %d n_cell_nodes %d\n",
                          rank, n_local_cells, n_cell_nodes);
  PetscSynchronizedFlush(PETSC_COMM_WORLD, PETSC_STDOUT);

  DMDARestoreElements(da, &n_local_cells, &n_cell_nodes, &local_indices);

  PetscFinalize();
  return 0;
}


I believe it creates a 1D DM object, and outputs the number of local cells 
assigned to each process.

Here is what I have as an output:

$ mpirun -np 1 ./test
rank 0 n_local_cells 8 n_cell_nodes 2     // OK

$ mpirun -np 2 ./test
rank 0 n_local_cells 4 n_cell_nodes 2     // OK
rank 1 n_local_cells 3 n_cell_nodes 2     // I expected 4 local cells here

$ mpirun -np 4 ./test
rank 0 n_local_cells 2 n_cell_nodes 2     // OK
rank 1 n_local_cells 1 n_cell_nodes 2     // I expected 2 local cells here
rank 2 n_local_cells 1 n_cell_nodes 2     // I expected 2 local cells here
rank 3 n_local_cells 1 n_cell_nodes 2     // I expected 2 local cells here


What am I missing?

Thank you.

Best,
Mikhail



Reply via email to