Mikhail
DMDAGetElements - Gets an array containing the indices (in local coordinates)
of all the local elements
Since it returns the "local indices" of the vertices of the elements it does
not make sense for a stencil width of zero (with parallel DMDA) since the
vertex of the last element on each process is actually owned by the next
process so it is a ghost vertex on the process but with a stencil width of zero
there are no ghost vertices hence no "local index" to represent that vertex.
I have changed the code to return an error instead of "garbage" values if the
stencil width is zero.
So you should just use a stencil width of 1 if you want to do finite elements
(or finite differences actually also :-). Stencil width of 0 is only for
strange situations and doesn't make much sense for PDEs.
Thanks for reporting the problem.
Barry
> On Jul 15, 2015, at 10:46 AM, Mikhail Artemyev <[email protected]>
> wrote:
>
> Dear all,
>
> Here is a minimal working example that I'm testing:
>
> #include "petscsys.h"
> #include "petscdmda.h"
>
> int main(int argc, char **argv)
> {
> PetscInitialize(&argc, &argv, 0, 0);
>
> PetscInt rank;
> MPI_Comm_rank(PETSC_COMM_WORLD, &rank);
>
> DM da;
> DMDACreate1d(PETSC_COMM_WORLD, DM_BOUNDARY_NONE, -9, 1, 0, NULL, &da);
>
> PetscInt n_local_cells, n_cell_nodes;
> const PetscInt *local_indices;
> DMDAGetElements(da, &n_local_cells, &n_cell_nodes, &local_indices);
>
> PetscSynchronizedPrintf(PETSC_COMM_WORLD,
> "rank %d n_local_cells %d n_cell_nodes %d\n",
> rank, n_local_cells, n_cell_nodes);
> PetscSynchronizedFlush(PETSC_COMM_WORLD, PETSC_STDOUT);
>
> DMDARestoreElements(da, &n_local_cells, &n_cell_nodes, &local_indices);
>
> PetscFinalize();
> return 0;
> }
>
>
> I believe it creates a 1D DM object, and outputs the number of local cells
> assigned to each process.
>
> Here is what I have as an output:
>
> $ mpirun -np 1 ./test
> rank 0 n_local_cells 8 n_cell_nodes 2 // OK
>
> $ mpirun -np 2 ./test
> rank 0 n_local_cells 4 n_cell_nodes 2 // OK
> rank 1 n_local_cells 3 n_cell_nodes 2 // I expected 4 local cells here
>
> $ mpirun -np 4 ./test
> rank 0 n_local_cells 2 n_cell_nodes 2 // OK
> rank 1 n_local_cells 1 n_cell_nodes 2 // I expected 2 local cells here
> rank 2 n_local_cells 1 n_cell_nodes 2 // I expected 2 local cells here
> rank 3 n_local_cells 1 n_cell_nodes 2 // I expected 2 local cells here
>
>
> What am I missing?
>
> Thank you.
>
> Best,
> Mikhail
>
>