Hi Mark, I have had a look at these, and I could only find these ones that looked relevant:
-pc_hypre_boomeramg_print_statistics: Print statistics (None) -pc_hypre_boomeramg_print_debug: Print debug information (None) -pc_hypre_boomeramg_print_statistics prints information about the V-cycle (number of levels, Dofs per level etc.) It also gave this information (which is something I can potentially work with?) Proc = 0 Level = 0 Coarsen Time = 0.022089 Proc = 0 Level = 0 Build Interp Time = 0.138844 Proc = 0 Level = 0 Build Coarse Operator Time = 0.354488 Proc = 0 Level = 1 Coarsen Time = 0.009334 Proc = 0 Level = 1 Build Interp Time = 0.074119 Proc = 0 Level = 1 Build Coarse Operator Time = 0.097702 Proc = 0 Level = 2 Coarsen Time = 0.004301 Proc = 0 Level = 2 Build Interp Time = 0.035835 Proc = 0 Level = 2 Build Coarse Operator Time = 0.030501 Proc = 0 Level = 3 Coarsen Time = 0.001876 Proc = 0 Level = 3 Build Interp Time = 0.014711 Proc = 0 Level = 3 Build Coarse Operator Time = 0.008681 Proc = 0 Level = 4 Coarsen Time = 0.000557 Proc = 0 Level = 4 Build Interp Time = 0.004307 Proc = 0 Level = 4 Build Coarse Operator Time = 0.002373 Proc = 0 Level = 5 Coarsen Time = 0.000268 Proc = 0 Level = 5 Build Interp Time = 0.001061 Proc = 0 Level = 5 Build Coarse Operator Time = 0.000589 Proc = 0 Level = 6 Coarsen Time = 0.000149 Proc = 0 Level = 6 Build Interp Time = 0.000339 Proc = 0 Level = 6 Build Coarse Operator Time = 0.000206 Proc = 0 Level = 7 Coarsen Time = 0.000090 Proc = 0 Level = 7 Build Interp Time = 0.000148 Proc = 0 Level = 7 Build Coarse Operator Time = 0.000085 Proc = 0 Level = 8 Coarsen Time = 0.000054 Proc = 0 Level = 8 Build Interp Time = 0.000100 Proc = 0 Level = 8 Build Coarse Operator Time = 0.000053 I have not tried -pc_hypre_boomeramg_print_debug yet. Think I can get the total coarsen time by summing up the time from all the levels here. I am still trying to understand how to get the time spent to solve the problem at the coarsest level. Best Parv From: Mark Adams <[email protected]> Sent: Wednesday, April 10, 2024 2:01 PM To: Khurana, Parv <[email protected]> Cc: [email protected] Subject: Re: [petsc-users] Exposing further detail in -log_view for Hypre withe PETSc This email from [email protected]<mailto:[email protected]> originates from outside Imperial. Do not click on links and attachments unless you recognise the sender. If you trust the sender, add them to your safe senders list<https://urldefense.us/v3/__https://spam.ic.ac.uk/SpamConsole/Senders.aspx__;!!G_uCfscf7eWS!ed9RkXQqDh1kRdnHw6WLZnxuX_1FqD8owPJ9aEVg03JkpWQewQVck8IP7DBq-GvoauWVZ3mYePX6Myc8UdEXTsMz4PvWKVOl$ > to disable email stamping for this address. I believe there is an option to get hypre to print its performance data. Run with -help and grep on "pc_hypre" and look for something that looks like a logging or view parameter. Mark On Wed, Apr 10, 2024 at 7:49 AM Khurana, Parv <[email protected]<mailto:[email protected]>> wrote: Hello PETSc users community, Thank you in advance for your help as always. I am using BoomerAMG from Hypre via PETSc as a part of preconditioner in my software (Nektar++). I am trying to understand the time profiling information that is printed ZjQcmQRYFpfptBannerStart This Message Is From an External Sender This message came from outside your organization. ZjQcmQRYFpfptBannerEnd Hello PETSc users community, Thank you in advance for your help as always. I am using BoomerAMG from Hypre via PETSc as a part of preconditioner in my software (Nektar++). I am trying to understand the time profiling information that is printed using the -log_view option. I want to understand how much time is spent in the smoothening step vs the time to solve on the coarsest grid I reach. The output that I get from -log_view (pasted below) gives me information of the KSPSolve and MatMult, but I think I need more granular time information to see a further breakdown of time spent within my routines. I would like to hear if anyone has any recommendations on obtaining this information? Best Parv Khurana PETSc database options used for solve: -ksp_monitor # (source: file) -ksp_type preonly # (source: file) -log_view # (source: file) -pc_hypre_boomeramg_coarsen_type hmis # (source: file) -pc_hypre_boomeramg_grid_sweeps_all 2 # (source: file) -pc_hypre_boomeramg_interp_type ext+i # (source: file) -pc_hypre_boomeramg_max_iter 1 # (source: file) -pc_hypre_boomeramg_P_max 2 # (source: file) -pc_hypre_boomeramg_print_debug 1 # (source: file) -pc_hypre_boomeramg_print_statistics 1 # (source: file) -pc_hypre_boomeramg_relax_type_all sor/jacobi # (source: file) -pc_hypre_boomeramg_strong_threshold 0.7 # (source: file) -pc_hypre_boomeramg_truncfactor 0.3 # (source: file) -pc_hypre_type boomeramg # (source: file) -pc_type hypre # (source: file) PETSc log_view output: ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flop --- Global --- --- Stage ---- Total Max Ratio Max Ratio Max Ratio Mess AvgLen Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage BuildTwoSided 1 1.0 9.6900e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 14 1.0 1.6315e-01 1.0 1.65e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 86 0 0 0 0 86 0 0 0 1011 MatConvert 1 1.0 4.3092e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 3 1.0 3.1680e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 3 1.0 9.4178e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 2 1.0 1.1630e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatSetPreallCOO 1 1.0 3.2132e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatSetValuesCOO 1 1.0 2.9956e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNorm 28 1.0 1.3981e-02 1.0 2.10e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 11 0 0 0 0 11 0 0 0 1499 VecSet 13 1.0 6.5185e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAYPX 14 1.0 7.1511e-03 1.0 5.24e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 3 0 0 0 0 3 0 0 0 733 VecAssemblyBegin 14 1.0 1.3998e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAssemblyEnd 14 1.0 4.2560e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 14 1.0 8.2761e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterEnd 14 1.0 4.4665e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 SFSetGraph 1 1.0 6.5993e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 SFSetUp 1 1.0 7.9212e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 SFPack 14 1.0 5.8690e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 SFUnpack 14 1.0 4.3370e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSetUp 1 1.0 2.4910e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 14 1.0 2.1922e+00 1.0 1.91e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 100 0 0 0 0 100 0 0 0 87 PCSetUp 1 1.0 1.3165e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 14 1.0 1.9990e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 ------------------------------------------------------------------------------------------------------------------------
