Hi Mark,

I have had a look at these, and I could only find these ones that looked 
relevant:

-pc_hypre_boomeramg_print_statistics: Print statistics (None)
  -pc_hypre_boomeramg_print_debug: Print debug information (None)

-pc_hypre_boomeramg_print_statistics prints information about the V-cycle 
(number of levels, Dofs per level etc.) It also gave this information (which is 
something I can potentially work with?)

Proc = 0    Level = 0    Coarsen Time = 0.022089
Proc = 0    Level = 0    Build Interp Time = 0.138844
Proc = 0    Level = 0    Build Coarse Operator Time = 0.354488
Proc = 0    Level = 1    Coarsen Time = 0.009334
Proc = 0    Level = 1    Build Interp Time = 0.074119
Proc = 0    Level = 1    Build Coarse Operator Time = 0.097702
Proc = 0    Level = 2    Coarsen Time = 0.004301
Proc = 0    Level = 2    Build Interp Time = 0.035835
Proc = 0    Level = 2    Build Coarse Operator Time = 0.030501
Proc = 0    Level = 3    Coarsen Time = 0.001876
Proc = 0    Level = 3    Build Interp Time = 0.014711
Proc = 0    Level = 3    Build Coarse Operator Time = 0.008681
Proc = 0    Level = 4    Coarsen Time = 0.000557
Proc = 0    Level = 4    Build Interp Time = 0.004307
Proc = 0    Level = 4    Build Coarse Operator Time = 0.002373
Proc = 0    Level = 5    Coarsen Time = 0.000268
Proc = 0    Level = 5    Build Interp Time = 0.001061
Proc = 0    Level = 5    Build Coarse Operator Time = 0.000589
Proc = 0    Level = 6    Coarsen Time = 0.000149
Proc = 0    Level = 6    Build Interp Time = 0.000339
Proc = 0    Level = 6    Build Coarse Operator Time = 0.000206
Proc = 0    Level = 7    Coarsen Time = 0.000090
Proc = 0    Level = 7    Build Interp Time = 0.000148
Proc = 0    Level = 7    Build Coarse Operator Time = 0.000085
Proc = 0    Level = 8    Coarsen Time = 0.000054
Proc = 0    Level = 8    Build Interp Time = 0.000100
Proc = 0    Level = 8    Build Coarse Operator Time = 0.000053

I have not tried -pc_hypre_boomeramg_print_debug yet.
Think I can get the total coarsen time by summing up the time from all the 
levels here. I am still trying to understand how to get the time spent to solve 
the problem at the coarsest level.

Best
Parv

From: Mark Adams <[email protected]>
Sent: Wednesday, April 10, 2024 2:01 PM
To: Khurana, Parv <[email protected]>
Cc: [email protected]
Subject: Re: [petsc-users] Exposing further detail in -log_view for Hypre withe 
PETSc

This email from [email protected]<mailto:[email protected]> originates from outside 
Imperial. Do not click on links and attachments unless you recognise the 
sender. If you trust the sender, add them to your safe senders 
list<https://urldefense.us/v3/__https://spam.ic.ac.uk/SpamConsole/Senders.aspx__;!!G_uCfscf7eWS!ed9RkXQqDh1kRdnHw6WLZnxuX_1FqD8owPJ9aEVg03JkpWQewQVck8IP7DBq-GvoauWVZ3mYePX6Myc8UdEXTsMz4PvWKVOl$
 > to disable email stamping for this address.


I believe there is an option to get hypre to print its performance data.
Run with -help and grep on "pc_hypre" and look for something that looks like a 
logging or view parameter.

Mark

On Wed, Apr 10, 2024 at 7:49 AM Khurana, Parv 
<[email protected]<mailto:[email protected]>> wrote:
Hello PETSc users community, Thank you in advance for your help as always. I am 
using BoomerAMG from Hypre via PETSc as a part of preconditioner in my software 
(Nektar++). I am trying to understand the time profiling information that is 
printed
ZjQcmQRYFpfptBannerStart
This Message Is From an External Sender
This message came from outside your organization.

ZjQcmQRYFpfptBannerEnd
Hello PETSc users community,

Thank you in advance for your help as always.

I am using BoomerAMG from Hypre via PETSc as a part of preconditioner in my 
software (Nektar++). I am trying to understand the time profiling information 
that is printed using the -log_view option.

I want to understand how much time is spent in the smoothening step vs the time 
to solve on the coarsest grid I reach. The output that I get from -log_view 
(pasted below) gives me information of the KSPSolve and MatMult, but I think I 
need more granular time information to see a further breakdown of time spent 
within my routines. I would like to hear if anyone has any recommendations on 
obtaining this information?

Best
Parv Khurana


PETSc database options used for solve:

-ksp_monitor # (source: file)
-ksp_type preonly # (source: file)
-log_view # (source: file)
-pc_hypre_boomeramg_coarsen_type hmis # (source: file)
-pc_hypre_boomeramg_grid_sweeps_all 2 # (source: file)
-pc_hypre_boomeramg_interp_type ext+i # (source: file)
-pc_hypre_boomeramg_max_iter 1 # (source: file)
-pc_hypre_boomeramg_P_max 2 # (source: file)
-pc_hypre_boomeramg_print_debug 1 # (source: file)
-pc_hypre_boomeramg_print_statistics 1 # (source: file)
-pc_hypre_boomeramg_relax_type_all sor/jacobi # (source: file)
-pc_hypre_boomeramg_strong_threshold 0.7 # (source: file)
-pc_hypre_boomeramg_truncfactor 0.3 # (source: file)
-pc_hypre_type boomeramg # (source: file)
-pc_type hypre # (source: file)


PETSc log_view output:

------------------------------------------------------------------------------------------------------------------------
Event                Count      Time (sec)     Flop                             
 --- Global ---  --- Stage ----  Total
                   Max Ratio  Max     Ratio   Max  Ratio  Mess   AvgLen  Reduct 
 %T %F %M %L %R  %T %F %M %L %R Mflop/s
------------------------------------------------------------------------------------------------------------------------

--- Event Stage 0: Main Stage

BuildTwoSided          1 1.0 9.6900e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
MatMult               14 1.0 1.6315e-01 1.0 1.65e+08 1.0 0.0e+00 0.0e+00 
0.0e+00  0 86  0  0  0   0 86  0  0  0  1011
MatConvert             1 1.0 4.3092e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
MatAssemblyBegin       3 1.0 3.1680e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
MatAssemblyEnd         3 1.0 9.4178e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
MatGetRowIJ            2 1.0 1.1630e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
MatSetPreallCOO        1 1.0 3.2132e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
MatSetValuesCOO        1 1.0 2.9956e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecNorm               28 1.0 1.3981e-02 1.0 2.10e+07 1.0 0.0e+00 0.0e+00 
0.0e+00  0 11  0  0  0   0 11  0  0  0  1499
VecSet                13 1.0 6.5185e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecAYPX               14 1.0 7.1511e-03 1.0 5.24e+06 1.0 0.0e+00 0.0e+00 
0.0e+00  0  3  0  0  0   0  3  0  0  0   733
VecAssemblyBegin      14 1.0 1.3998e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecAssemblyEnd        14 1.0 4.2560e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecScatterBegin       14 1.0 8.2761e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecScatterEnd         14 1.0 4.4665e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
SFSetGraph             1 1.0 6.5993e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
SFSetUp                1 1.0 7.9212e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
SFPack                14 1.0 5.8690e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
SFUnpack              14 1.0 4.3370e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
KSPSetUp               1 1.0 2.4910e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
KSPSolve              14 1.0 2.1922e+00 1.0 1.91e+08 1.0 0.0e+00 0.0e+00 
0.0e+00  0 100  0  0  0   0 100  0  0  0    87
PCSetUp                1 1.0 1.3165e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
PCApply               14 1.0 1.9990e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 
0.0e+00  0  0  0  0  0   0  0  0  0  0     0
------------------------------------------------------------------------------------------------------------------------

Reply via email to