That's great. Thanks for creating this great piece of software!
Amin
On Wed, Mar 25, 2020 at 5:56 PM Matthew Knepley wrote:
> On Wed, Mar 25, 2020 at 5:41 PM Amin Sadeghi
> wrote:
>
>> Junchao, thank you for doing the experiment, I guess TACC Frontera nodes
>> have hi
ate. You are seeing about 50% of perfect speedup at 16 process. I
>>> would expect that you will see something similar with streams. Without
>>> knowing your machine, your results look typical.
>>>
>>> On Wed, Mar 25, 2020 at 1:05 PM Amin Sadeghi
>>> wro
milar with streams. Without
> knowing your machine, your results look typical.
>
> On Wed, Mar 25, 2020 at 1:05 PM Amin Sadeghi
> wrote:
>
>> Hi,
>>
>> I ran KSP example 45 on a single node with 32 cores and 125GB memory
>> using 1, 16 and 32 MPI processes. Here's
Hi,
I ran KSP example 45 on a single node with 32 cores and 125GB memory using
1, 16 and 32 MPI processes. Here's a comparison of the time spent during
KSP.solve:
- 1 MPI process: ~98 sec, speedup: 1X
- 16 MPI processes: ~12 sec, speedup: ~8X
- 32 MPI processes: ~11 sec, speedup: ~9X
Since the
Hi,
Recently, I've been playing around with petsc4py to solve a battery simulation,
which takes too long to solve using scipy solvers. I also have access to an HPC
cluster with a few nodes, each with a dozen CPU cores. However, I can't seem to
get any further speedup past 4 processors. Very