Hi Barry, thanks for the clarification. I am running my code on a different (much slower) machine right now, and from the initial results it seems so far that Matt Knepley's suspicion of having a bad network could be correct. But i need to do a couple more runs.
thanks mat On Tuesday 22 August 2006 13:34, Barry Smith wrote: > Mat, > > This will not effect load balance or anything like that. > > When you pass a communicator like MPI_COMM_WORLD to PETSc > we don't actually use that communicator (because you might > be using it and there may be tag collisions etc). So instead > we store our own communicator inside the MPI_COMM_WORLD as an > attribute, this message is just telling us we are accessing > the inner communicator. > > Barry > > On Thu, 17 Aug 2006, Matt Funk wrote: > > Hi, > > > > i was wondering what the message: > > 'PetscCommDuplicate Using internal PETSc communicator 92 170' > > means exactly. I still have issues with PETSc when running 1 vs 2 procs > > w.r.t. the loadbalance. > > However, when run on 2 vs 4 the balance seems to be almost perfect. > > Then the option of a screwed up network was suggested to me, but since > > the 4vs 2 proc case is ok, it seems not necessarily to be the case. > > > > Maybe somebody can tell me what it means? > > > > thanks > > mat
