Hello, I have encountered some problems. Here are some of my configurations.
OS Version and Type: Linux daihuanhe-Aspire-A315-55G 5.15.0-89-generic
#99~20.04.1-Ubuntu SMP Thu Nov 2 15:16:47 UTC 2023 x86_64 x86_64 x86_64
GNU/Linux
PETSc Version: #define PETSC_VERSION_RELEASE 1
The snes options are not relevant since the parts of a PCFIELDSPLIT are
always linear problems.
By default PCFIELDSPLIT uses a KSP type of preonly on each split (that is
it applies the preconditioner exactly once inside the PCApply_FieldSplit()
hence the -fieldsplit_*_ksp_ options are
Hello, I have encountered some problems.Here are some of my
configurations.
OS Version and Type: Linux daihuanhe-Aspire-A315-55G 5.15.0-89-generic
#99~20.04.1-Ubuntu SMP Thu Nov 2 15:16:47 UTC 2023 x86_64 x86_64 x86_64
GNU/Linux
PETSc Version: #define PETSC_VERSION_RELEASE 1
Perhaps you need
> make PETSC_DIR=~/asd/petsc-3.19.3 PETSC_ARCH=arch-mswin-c-opt all
> On Jul 24, 2023, at 1:11 PM, Константин via petsc-users
> wrote:
>
> Good evening. After configuring petsc I had to write this comand on cygwin64.
> $ make PETSC_DIR=/home/itugr/asd/petsc-3.19.3
Good evening. After configuring petsc I had to write this comand on cygwin64.
$ make PETSC_DIR=/home/itugr/asd/petsc-3.19.3 PETSC_ARCH=arch-mswin-c-opt all
But I have such problem
makefile:26: /home/itugr/asd/petsc-3.19.3/lib/petsc/conf/rules.utils: No such
file or directory
make[1]: *** No rule
On Tue, Jul 11, 2023 at 3:58 PM Константин via petsc-users <
petsc-users@mcs.anl.gov> wrote:
> Hello, I'm trying to build petsc on windows. And when I make it I have
> such problem
>
>
Did you run configure first?
Thanks,
Matt
> --
> Константин
>
--
What most experimenters take for
Hello, I'm trying to build petsc on windows. And when I make it I have such
problem
--
Константин
Please keep on list,
On Thu, Feb 17, 2022 at 12:36 PM Bojan Niceno <
bojan.niceno.scient...@gmail.com> wrote:
> Dear Mark,
>
> Sorry for mistakenly calling you Adam before.
>
> I was thinking about the o_nnz as you suggested, but then something else
> occurred to me. So, I determine the d_nnz
On Thu, Feb 17, 2022 at 11:46 AM Bojan Niceno <
bojan.niceno.scient...@gmail.com> wrote:
> Dear all,
>
>
> I am experiencing difficulties when using PETSc in parallel in an
> unstructured CFD code. It uses CRS format to store its matrices. I use
> the following sequence of PETSc call in the
Dear all,
I am experiencing difficulties when using PETSc in parallel in an
unstructured CFD code. It uses CRS format to store its matrices. I use
the following sequence of PETSc call in the hope to get PETSc solving my
linear systems in parallel. Before I continue, I would just like to say
Thank you, I'll try that.
Best,
Li
On Wed, Dec 4, 2019 at 5:34 AM Smith, Barry F. wrote:
>
> From the code:
>
> if (snes->lagjacobian == -2) {
> snes->lagjacobian = -1;
>
> ierr = PetscInfo(snes,"Recomputing Jacobian/preconditioner because lag
> is -2 (means compute Jacobian, but
From the code:
if (snes->lagjacobian == -2) {
snes->lagjacobian = -1;
ierr = PetscInfo(snes,"Recomputing Jacobian/preconditioner because lag is
-2 (means compute Jacobian, but then never again) \n");CHKERRQ(ierr);
} else if (snes->lagjacobian == -1) {
ierr =
> On Dec 2, 2019, at 2:30 PM, Li Luo wrote:
>
> -snes_mf fails to converge in my case, but -ds_snes_mf_operator works,
> when the original analytic matrix is still used as the preconditioner.
> The timing is several times greater than using the analytic matrix for both
> Jacobian and
Please send a run with optimization turned on (--with-debugging=0 in
./configure) and -log_view without the actual timing information we are just
guessing where the time is spent.
If your problem has a natural block size then using baij should be a bit
faster than aij, but not
Thank you very much! It looks forming an analytic Jacobian is the only
choice.
Best,
Li
On Mon, Dec 2, 2019 at 3:21 PM Matthew Knepley wrote:
> On Mon, Dec 2, 2019 at 4:04 AM Li Luo wrote:
>
>> Thank you for your reply.
>>
>> The matrix is small with only 67500 rows, but is relatively dense
On Mon, Dec 2, 2019 at 4:04 AM Li Luo wrote:
> Thank you for your reply.
>
> The matrix is small with only 67500 rows, but is relatively dense since a
> second-order discontinuous Galerkin FEM is used, nonzeros=23,036,400.
>
This is very dense, 0.5% fill or 340 nonzeros per row.
> The number
Thank you for your reply.
The matrix is small with only 67500 rows, but is relatively dense since a
second-order discontinuous Galerkin FEM is used, nonzeros=23,036,400.
The number of colors is 539 as shown by using -mat_fd_coloring_view:
MatFDColoring Object: 64 MPI processes
type not yet set
How many colors is it requiring? And how long is the MatGetColoring()
taking? Are you running in parallel? The MatGetColoring() MATCOLORINGSL uses a
sequential coloring algorithm so if your matrix is large and parallel the
coloring will take a long time. The parallel colorings are
Mark Adams writes:
> FM matrices are slow and meant for debugging mostly (I thought,
> although the docs just give this warning if coloring is not available).
>
> I would check the timings from -log_view and verify that the time is spent
> in MatFDColoringApply. Running with -info should print
FM matrices are slow and meant for debugging mostly (I thought,
although the docs just give this warning if coloring is not available).
I would check the timings from -log_view and verify that the time is spent
in MatFDColoringApply. Running with -info should print the number of colors
(C). The
Dear Developers,
I tried to use the multi-color finite-difference (MC-FD) method for
constructing the Jacobians. However, I find it is very slow compared to the
exact Jacobian.
My implementation of MC-FD Jacobian is posted below, would you please check
whether I am correct? Anything missed? Thank
Can you send configure.log from this build?
Satish
On Thu, 28 Feb 2019, DAFNAKIS PANAGIOTIS via petsc-users wrote:
> Hi everybody,
>
> I am trying to install PETSc version 3.10.3 on OSX Sierra 10.13.6 with the
> following configure options
> ./configure --CC=mpicc --CXX=mpicxx --FC=mpif90
Hi everybody,
I am trying to install PETSc version 3.10.3 on OSX Sierra
10.13.6 with the following configure options
./configure --CC=mpicc --CXX=mpicxx --FC=mpif90
--PETSC_ARCH=sierra-dbg --with-debugging=1
--download-hypre=1 --with-x=0
however I am getting the following error messages
Message-
> From: Smith, Barry F. [mailto:bsm...@mcs.anl.gov]
> Sent: Thursday, 1 November 2018 9:28 AM
> To: Wenjin Xing
> Cc: petsc-users@mcs.anl.gov
> Subject: Re: [petsc-users] (no subject)
>
>
>This option only works with AIJ matrices; you must be using
This option only works with AIJ matrices; you must be using either BAIJ or
SBAIJ matrices? (or a shell matrix)
Barry
> On Oct 31, 2018, at 5:45 AM, Wenjin Xing via petsc-users
> wrote:
>
> My issue is summarized in the picture and posted in the link
>
My issue is summarized in the picture and posted in the link
https://scicomp.stackexchange.com/questions/30458/what-does-the-error-this-matrix-type-does-not-have-a-find-zero-diagonals-define?noredirect=1#comment56074_30458
[cid:image001.png@01D4715E.DAED5B40]
Kind regards
Wenjin
Thanks. I think I find the right way.
Wayne
On Fri, Sep 16, 2016 at 11:33 AM, Ji Zhang wrote:
> Thanks for your warm help. Could you please show me some necessary
> functions or a simple demo code?
>
>
> Wayne
>
> On Fri, Sep 16, 2016 at 10:32 AM, Barry Smith
Thanks for your warm help. Could you please show me some necessary
functions or a simple demo code?
Wayne
On Fri, Sep 16, 2016 at 10:32 AM, Barry Smith wrote:
>
> You should create your small m_ij matrices as just dense two dimensional
> arrays and then set them into the
You should create your small m_ij matrices as just dense two dimensional
arrays and then set them into the big M matrix. Do not create the small dense
matrices as PETSc matrices.
Barry
> On Sep 15, 2016, at 9:21 PM, Ji Zhang wrote:
>
> I'm so apologize for the
I'm so apologize for the ambiguity. Let me clarify it.
I'm trying to simulation interactions among different bodies. Now I have
calculated the interaction between two of them and stored in the sub-matrix
m_ij. What I want to do is to consider the whole interaction and construct
all sub-matrices
On Thu, Sep 15, 2016 at 4:23 AM, Ji Zhang wrote:
> Thanks Matt. It works well for signal core. But is there any solution if I
> need a MPI program?
>
It unclear what the stuff below would mean in parallel.
If you want to assemble several blocks of a parallel matrix that looks
Thanks Matt. It works well for signal core. But is there any solution if I
need a MPI program?
Thanks.
Wayne
On Tue, Sep 13, 2016 at 9:30 AM, Matthew Knepley wrote:
> On Mon, Sep 12, 2016 at 8:24 PM, Ji Zhang wrote:
>
>> Dear all,
>>
>> I'm using petsc4py
On Mon, Sep 12, 2016 at 8:24 PM, Ji Zhang wrote:
> Dear all,
>
> I'm using petsc4py and now face some problems.
> I have a number of small petsc dense matrices mij, and I want to construct
> them to a big matrix M like this:
>
> [ m11 m12 m13 ]
> M = | m21
Dear all,
I'm using petsc4py and now face some problems.
I have a number of small petsc dense matrices mij, and I want to construct
them to a big matrix M like this:
[ m11 m12 m13 ]
M = | m21 m22 m23 | ,
[ m31 m32 m33 ]
How could I do it effectively?
Now I'm
Re,
Makes sense to read the documentation, I will try with another
preconditioners.
Thanks for the support.
2016-06-02 18:15 GMT+02:00 Matthew Knepley :
> On Thu, Jun 2, 2016 at 11:10 AM, neok m4700 wrote:
>
>> Hi Satish,
>>
>> Thanks for the
On Thu, Jun 2, 2016 at 11:10 AM, neok m4700 wrote:
> Hi Satish,
>
> Thanks for the correction.
>
> The error message is now slightly different, but the result is the same
> (serial runs fine, parallel with mpirun fails with following error):
>
Now the error is correct. You
Hi Satish,
Thanks for the correction.
The error message is now slightly different, but the result is the same
(serial runs fine, parallel with mpirun fails with following error):
[0] KSPSolve() line 599 in <...>/src/ksp/ksp/interface/itfunc.c
[0] KSPSetUp() line 390 in
with petsc-master - you would have to use petsc4py-master.
i.e try petsc-eab7b92 with petsc4py-6e8e093
Satish
On Thu, 2 Jun 2016, neok m4700 wrote:
> Hi Matthew,
>
> I've rebuilt petsc // petsc4py with following versions:
>
> 3.7.0 // 3.7.0 => same runtime error
> 00c67f3 // 3.7.1 => fails
Hi Matthew,
I've rebuilt petsc // petsc4py with following versions:
3.7.0 // 3.7.0 => same runtime error
00c67f3 // 3.7.1 => fails to build petsc4py (error below)
00c67f3 // 6e8e093 => same as above
f1b0812 (latest commit) // 6e8e093 (latest commit) => same as above
In file included from
On Thu, Jun 2, 2016 at 9:12 AM, neok m4700 wrote:
> Hi,
>
> I built petsc 3.7.1 and petsc4py 3.7.0 (with openmpi 1.10.2) and ran the
> examples in the demo directory.
>
I believe this was fixed in 'master':
Hi,
I built petsc 3.7.1 and petsc4py 3.7.0 (with openmpi 1.10.2) and ran the
examples in the demo directory.
$ python test_mat_ksp.py
=> runs as expected (serial)
$ mpiexec -np 2 python test_mat_ksp.py
=> fails with the following output:
Traceback (most recent call last):
File
On Mon, May 4, 2015 at 5:57 PM, Reza Yaghmaie reza.yaghma...@gmail.com
wrote:
Dear Matt,
The initial guess was zero for all cases of SNES solvers. The initial
jacobian was identity for all cases. The system is small and is ran
sequentially.
I have to add that I use FDColoring routine for
Dear Matt,
The initial guess was zero for all cases of SNES solvers. The initial
jacobian was identity for all cases. The system is small and is ran
sequentially.
I have to add that I use FDColoring routine for the jacobian as well.
Regards,
Ray
On Monday, May 4, 2015, Matthew Knepley
On Mon, May 4, 2015 at 2:11 PM, Reza Yaghmaie reza.yaghma...@gmail.com
wrote:
Dear PETSC representatives,
I am solving a nonlinear problem with SNESNGMRES and it converges faster
with less iterations compared to otehr SNES methods. Any idea why that is
the case?
It is impossible to tell
Dear Matt,
Actually the initial jacobian was identity. Regular SNES converges in 48
iterations, GMRES in 19, NCG in 67,...
Do you think SNESQN with the basiclineseach was the problem for divergence?
If I use SNESQN by default should not it converge with initial identity
jacobian?
Best regards,
On Mon, May 4, 2015 at 5:41 PM, Reza Yaghmaie reza.yaghma...@gmail.com
wrote:
Dear Matt,
Actually the initial jacobian was identity. Regular SNES converges in 48
iterations, GMRES in 19, NCG in 67,...
Do you think SNESQN with the basiclineseach was the problem for divergence?
If I use
Dear PETSC representatives,
I am solving a nonlinear problem with SNESNGMRES and it converges faster
with less iterations compared to otehr SNES methods. Any idea why that is
the case?
Also SNESQN diverges quickly. I tried to use SNESLINESEARCHBASIC for the
linesearch option and nothing changes.
I am moving my code to have the DM create the matrix and am hitting this
error in DMGetLocalToGlobalMapping:
if (!dm-ops-getlocaltoglobalmapping)
SETERRQ(PetscObjectComm((PetscObject)dm),PETSC_ERR_SUP,DM can not create
LocalToGlobalMapping);
This happens when I create the matrix like this:
Hello,
I wanted to see how well Umfpack performs in PETSc but i encountered a
problem regarding the matrix type used and the way is distributed. I'm
trying to solve for multiple frequencies in parallel, where system matrix
A = omega*E-C, and omega is a frequency dependent scalar. Can i send
Yes, just assemble the same sequential matrices on each rank. To do this,
create the matrix using the communicator PETSC_COMM_SELF
Cheers
Dave
On Monday, 23 June 2014, Bogdan Dita bog...@lmn.pub.ro wrote:
Hello,
I wanted to see how well Umfpack performs in PETSc but i encountered a
And use the same sequential communicator when you create the KSP on each
rank
On Monday, 23 June 2014, Dave May dave.mayhe...@gmail.com wrote:
Yes, just assemble the same sequential matrices on each rank. To do this,
create the matrix using the communicator PETSC_COMM_SELF
Cheers
Dave
ok,
so considering performance on MIC
can the library MAGMA be used as an alternate to Viennacl for PETSc or
FEniCS?
http://www.nics.tennessee.edu/files/pdf/hpcss/04_03_LinearAlgebraPar.pdf
(from slide 37 onwards)
MAGMA seems to have sparse version which i think is doing all that any
sparse
Hi,
so considering performance on MIC
can the library MAGMA be used as an alternate to Viennacl for PETSc or
FEniCS?
No, there is no interface to MAGMA in PETSc yet. Contributions are
always welcome, yet it is not our priority to come up with an interface
of our own. I don't think it will
Dear sir,
Im a masters student from Indian Institute of Technology delhi. Im
working on PETSc.. for performance, which is my area of interest. Can you
please help me in knowing 'How to run PETSc on MIC' .That would be of
great help to me.
*thanks and regards..*
*M. Sai Rajeswar*
*M-tech
Hi,
Im a masters student from Indian Institute of Technology delhi. Im
working on PETSc.. for performance, which is my area of interest. Can
you please help me in knowing 'How to run PETSc on MIC' .That would
be of great help to me.
my experience is that 'performance' and 'MIC' for
Hello,
In my application I like to use
PetscErrorCode KSPSetOperators(KSP ksp,Mat Amat,Mat Pmat,MatStructure flag)
and have Pmat different from Amat
if Amat = L + D + U
then Pmat = Amat - L* - U* + rowsum(L* + U*)
where L* and U* is a portion of L and U and rowsum(L* + U*) will add into D
I
On Fri, Feb 21, 2014 at 12:34 PM, Chung-Kan Huang ckhua...@gmail.comwrote:
Hello,
In my application I like to use
PetscErrorCode KSPSetOperators(KSP ksp,Mat Amat,Mat Pmat,MatStructure
flag)
and have Pmat different from Amat
if Amat = L + D + U
then Pmat = Amat - L* - U* + rowsum(L* +
I'm attempting to implement PETSc for the parallel solution of a large sparse
matrix in a computer program I work with. The program is written in Fortran. I
currently have been able to get the program working using PETSc and giving
correct results, but the issue is its performance. It is much
I am writting a parallel unstructured FVM solver for fluid flow problem,
but i met some problems when i compute the gradient of a variable using Gauss's
theorem.
? ? As we know,we need the values of a variable at all its neighboring cells,
but if one of my neighbouring cell is not in the
On Sun, Mar 3, 2013 at 4:39 AM, ztdep at yahoo.com.cn wrote:
I am writting a parallel unstructured FVM solver for fluid flow
problem, but i met some problems when i compute the gradient of a variable
using Gauss's theorem.
As we know,we need the values of a variable at all its
On Sun, Mar 3, 2013 at 8:14 AM, Matthew Knepley knepley at gmail.com wrote:
You can see our example of a parallel unstructured FVM solver in TS ex11
in petsc-dev.
Specifically, I can see several ways to do this. You could partition
faces, and have ghost cells, which we can handles. Or
you
Hi Barry,
I wrote the following piece according to your suggestions. Currently it
does nothing but creates a vector with 1 at 1th position, 2 at 2th and so
on. But I made it serial, i.e. (n+1)th place is computed using the value of
nth place. My question, did I do it correctly, i.e. is it safe or
On Sat, Dec 29, 2012 at 1:40 AM, amlan barua abarua at iit.edu wrote:
Hi Barry,
I wrote the following piece according to your suggestions. Currently it
does nothing but creates a vector with 1 at 1th position, 2 at 2th and so
on. But I made it serial, i.e. (n+1)th place is computed using the
On Sat, Dec 29, 2012 at 11:15 PM, amlan barua abarua at iit.edu wrote:
Hi,
I am actually trying to implement a 'parallel' ordinary differential
equation solver.
For proper functioning of the algorithm (the name is parareal), I need to
implement a simple recurrence relation of the form x[i+1]
Hello!
I have a few simple questions about PETSc about functions that I can't seem
to find in the documentation:
1) Is there a way to automatically create a matrix in which all elements
are the same scalar value a, e.g. something like ones(m,n) in Matlab?
2) Is there an equivalent to Matlab .*
On Fri, Dec 28, 2012 at 4:01 PM, Jelena Slivka slivkaje at gmail.com wrote:
Hello!
I have a few simple questions about PETSc about functions that I can't
seem to find in the documentation:
1) Is there a way to automatically create a matrix in which all elements
are the same scalar value a,
Hi,
Is there an analogue of VecScatterCreateToZero for DA vectors? The DMDA
object has more than one degrees of freedom.
If there isn't any, should I use an IS object to do the scattering?
Amlan
-- next part --
An HTML attachment was scrubbed...
URL:
On Thu, Dec 27, 2012 at 1:22 AM, amlan barua abarua at iit.edu wrote:
Hi,
Is there an analogue of VecScatterCreateToZero for DA vectors? The DMDA
object has more than one degrees of freedom.
If there isn't any, should I use an IS object to do the scattering?
I do not understand. Why can't
ierr = DMDACreateNaturalVector(da,natural);CHKERRQ(ierr);
ierr = DMDAGlobalToNaturalBegin(da,xin,INSERT_VALUES,natural);CHKERRQ(ierr);
ierr = DMDAGlobalToNaturalEnd(da,xin,INSERT_VALUES,natural);CHKERRQ(ierr);
Now do VecScatterCreateToZero() from natural and the vector will be in the
On Dec 27, 2012, at 10:34 AM, amlan barua abarua at iit.edu wrote:
I think I can use VecSetValues, is that right?
Yes you could do that. But since you are using a DMDA you could also use
DMGetLocalVector(), DMGlobalToLocalBegin/End() followed by DMDAVecGetArray() to
access the ghost
Dear Developers,
I found the out of core parameter in the external solver MUMPS is important
when using with cygwin because the default directory /tmp/ simply does not
exist in windows. So you may consider putting:
? ierr = PetscOptionsString(-ooc_tmpdir, The Out of Core directory, None,
Pham Van:
Added to petsc-dev
http://petsc.cs.iit.edu/petsc/petsc-dev/rev/c3403225fb48
Let us know if there is any problem with it.
Thanks for your contribution!
Hong
Dear Developers,
I found the out of core parameter in the external solver MUMPS is
important when using with cygwin because the
dr. ir. Christiaan Klaij
CFD Researcher
Research Development
E mailto:C.Klaij at marin.nl
T +31 317 49 33 44
MARIN
2, Haagsteeg, P.O. Box 28, 6700 AA Wageningen, The Netherlands
T +31 317 49 39 11, F +31 317 49 32 45, I www.marin.nl
[0]PETSC ERROR: Unhandled case, must have at least two fields!
You can use PCFieldSplitSetIS(). The implementation could check whether a
MatNest is being used and automatically set the splits if it is, but you
would have to call PCFieldSplitSetIS() later when you wanted to assemble
into an AIJ
On Mon, Jan 30, 2012 at 06:51, Klaij, Christiaan C.Klaij at marin.nl wrote:
Then what would be the best way to create the IS in this case?
Can it somehow be deduced from the separate blocks?
The ISs define the row space of the blocks inside the global matrix.
-- next part
I am still trying to push the Windows build. I am using the separately
compiled blas/lapack to see now a configure error:
---
External package hypre does not work on Microsoft Windows
On Wed, 9 Nov 2011, Dominik Szczerba wrote:
I am still trying to push the Windows build. I am using the separately
compiled blas/lapack to see now a configure error:
---
External package hypre does not work on
[petsc-users] SLEPc generalized eigenvalue problem question?
I create 2 matrices using:
MatCreateSeqDense(PETSC_COMM_SELF, n, n, Ka, A);
MatCreateSeqDense(PETSC_COMM_SELF, n, n, Kb, B);
These matrices are 99% zeros ( 16,016,004 entries and 18660 non-zeros).
They are symmetric and real. Their
On Mon, Sep 5, 2011 at 4:08 PM, Kristen Eisenberg
kristen.eisenberg at yahoo.com wrote:
[petsc-users] SLEPc generalized eigenvalue problem question?
I create 2 matrices using:
MatCreateSeqDense(PETSC_COMM_SELF, n, n, Ka, A);
MatCreateSeqDense(PETSC_COMM_SELF, n, n, Kb, B);
These matrices
On Jul 28, 2011, at 10:57 PM, Debao Shao wrote:
Dear all:
I'm using PETSC iterative solver(PCILU KSPGMRES) in OOQP, but when the to
be solved matrix is ill-posed, say condition number ~ 1e+4, then, KSPGMRES
will exceed its maximal iteration number(default as 1), while when I
,
Debao
-Original Message-
From: petsc-users-bounces at mcs.anl.gov
[mailto:petsc-users-boun...@mcs.anl.gov] On Behalf Of Barry Smith
Sent: Friday, July 29, 2011 12:11 PM
To: PETSc users list
Subject: Re: [petsc-users] (no subject)
On Jul 28, 2011, at 10:57 PM, Debao Shao wrote:
Dear all
Dear all:
I'm using PETSC iterative solver(PCILU KSPGMRES) in OOQP, but when the to be
solved matrix is ill-posed, say condition number ~ 1e+4, then, KSPGMRES will
exceed its maximal iteration number(default as 1), while when I checked the
same data with cholmod direct solver, it can
On Tue, Feb 1, 2011 at 8:01 AM, Hung Thanh Nguyen
hung.thanh.nguyen at petrell.no wrote:
I am trying to install PETSc on Windows. I am using Intel C compiler and
Intel MKL. I will use whichever MPI library is most appropriate (MPICH?).
I have tried to follow the installation
I am trying to install PETSc on Windows. I am using Intel C compiler and Intel
MKL. I will use whichever MPI library is most appropriate (MPICH?).
I have tried to follow the installation instructions: installed cygwin and try
to run configure in the cygwin shell. But I get some error regarding
If using free-format use extension .F90 not .F
On 10/08/2010 09:33 PM, Peter Wang wrote:
I am trying to modify the example code in
{PETSc_Dir}\src\vec\vec\examples\tests\ex19f.F . Only three lines are
added into the original code. However, if the three lines are added,
there is error
I am trying to modify the example code in
{PETSc_Dir}\src\vec\vec\examples\tests\ex19f.F . Only three lines are added
into the original code. However, if the three lines are added, there is error
coming out when it is compiled. Why it cannot be compiled when the lines are
added? I am
Fortran fixed format requires that line not begin before the 7th column
Barry
On Oct 8, 2010, at 9:33 PM, Peter Wang wrote:
I am trying to modify the example code in
{PETSc_Dir}\src\vec\vec\examples\tests\ex19f.F . Only three lines are added
into the original code. However, if the
87 matches
Mail list logo