Hi Brice,
Hi Samuel,
it seems I have done poor job explaining how I'm using hwloc-distrib. Let
me repair it.
On 128 core system for example, we do run series of parallel jobs:
1 job
2 jobs
4 jobs
8 jobs
12 jobs
and so on upto 128 jobs.
Parallel jobs are synchronized via semaphores and we
Brice Goglin, le Thu 29 Aug 2013 09:58:17 +0200, a écrit :
> Anyway, reversing the loop just move the core you don't want to the end of the
> list. But if you use the entire list, you end up using the exact same cores.
He wants that, yes.
Samuel
MAX_COUNT is topology->level_nbobjects[from_depth] -1
On Wed, Aug 28, 2013 at 4:20 PM, Jiri Hladky wrote:
>
> On Tue, Aug 27, 2013 at 7:57 PM, Brice Goglin wrote:
>>
>> You just explained why I don't like weights. Some people will want to
>> ignore
Le 27/08/2013 18:15, Jiri Hladky a écrit :
> using the weights looks like a good solution to me. However, we need
> to think if and how we should propagate the weight via the upper
> levels of the hierarchy. So for example if you have a socket with 4
> cores and cores 0&1 and 2&3 share L2 cache
Hi Brice,
hi Chris,
using the weights looks like a good solution to me. However, we need to
think if and how we should propagate the weight via the upper levels of the
hierarchy. So for example if you have a socket with 4 cores and cores 0&1
and 2&3 share L2 cache than the question is if and how
Le 27/08/2013 05:07, Christopher Samuel a écrit :
> On 27/08/13 00:07, Brice Goglin wrote:
>
> > But there's a more general problem here, some people may want
> > something similar for other cases. I need to think about it.
>
> Something like a sort order perhaps, combined with some method to
>
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1
On 27/08/13 00:07, Brice Goglin wrote:
> But there's a more general problem here, some people may want
> something similar for other cases. I need to think about it.
Something like a sort order perhaps, combined with some method to
exclude or weight
There are actually two issues here. We can easily reverse the output of
hwloc-distrib so that you get the first PU last. But you also need a
--single that doesn't always keep the first bit. We could add a
hwloc_bitmap_singlify_last() (and we'd have to define what to return
when the bitmap is
Hi Brice,
hi all,
I'm using hwloc-distrib to distribute the jobs on the big boxes (see the
attached topology). We use it to measure the performance under the various
loads. The issue is that hwloc-distrib always starts from PU:0. This is a
problem since usually PU:0 handles most of the interrupts