my apologies...!! I am using METIS for partitioning the domain. On Wed, Jan 18, 2012 at 12:11 AM, Matthew Knepley <knepley at gmail.com> wrote:
> On Tue, Jan 17, 2012 at 12:26 PM, Rahul Praghanmor <praghanmor at > gmail.com>wrote: > >> Dear Sir, >> I am working on a parallel unstructured finite volume >> solver.The solver is efficiently running in parallel using gauss seidel >> linear solver.The matrix is sparse and stored by CSR format.Now I want to >> implement a PETSc library to make the convergence faster.But I am going >> through a major problem as discussed below. >> If I partitioned a big domain say rectangular duct into 4 zones >> using parMetis.Each zone is solved in separate processor as fairly solved >> by gauss seidel linear solver.But I want to solve these zones by PETSc.How >> to do that?How to form a matrix with global numbering which is required >> format for PETSc to form a matrix?Does it necessarily important to form a >> global matrix? very few information available for assembling a matrix from >> unstructured finite volume method in PETSc. >> > > If you already run ParMetis, just number the rows it puts on each process > consecutively. > > Matt > > >> Thankx and regards, >> Rahul. >> >> > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > -- Zeus Numerix Pvt. Ltd. I2IT Campus, P-14 Rajiv Gandhi Infotech Park, Phase-1 Hinjewadi Pune -57 0ff. +91 64731511/9965 -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20120119/750da5ae/attachment.htm>
