Hello, Alexey!

I suggest that we decide what we can do within ticket [1].

Add "rebalanceId" and "topVer" related to rebalance to all messages.

Add statistical information to a log message:
[2020-05-06 20:56:37,044][INFO ][...] Completed rebalancing [rebalanceId=1, 
grp=cache1,
supplier=94a3fcbc-18d5-4c64-b0ab-4313aba00001, partitions=5, entries=100, 
duration=12ms,
bytesRcvd=5M, topVer=AffinityTopologyVersion [topVer=3, minorTopVer=1],
progress=1/2]

Add a message to log after rebalancing for all cache groups is complete:
[2020-05-06 20:56:36,999][INFO ][...] Completed rebalance chain:
[rebalanceId=2, partitions=10, entries=200, duration=50ms, bytesRcvd=10M]

Any comments or suggestions?

[1] - https://issues.apache.org/jira/browse/IGNITE-12080

20.05.2020, 23:08, "ткаленко кирилл" <tkalkir...@yandex.ru>:
> Hello, Alexey! Unfortunately, my response was delayed.
>
> Point 2: You can do as you suggested, I think it is still worth specifying 
> how many partitions were obtained.
>
> [2020-05-06 20:56:37,044][INFO ][...] Completed rebalancing [grp=cache1,
> supplier=94a3fcbc-18d5-4c64-b0ab-4313aba00001, partitions=5, entries=100, 
> duration=12ms,
> bytesRcvd=5M, topVer=AffinityTopologyVersion [topVer=3, minorTopVer=1],
> progress=1/2]
>
> Point 3: is It "rebalanceId"?
>
> Point 5: I think we can output a summary for each supplier, so as not to keep 
> it in mind.
>
> [2020-05-06 20:56:36,999][INFO ][...] Completed rebalance chain:
> [rebalanceId=2, [supplier=94a3fcbc-18d5-4c64-b0ab-4313aba00001, partitions=5, 
> entries=100, duration=12ms, bytesRcvd=5M],
> [supplier=94a3fcbc-18d5-4c64-b0ab-4313aba00002, partitions=5, entries=100, 
> duration=12ms, bytesRcvd=5M]]
>
> I can add "rebalanceId" to each message that you gave at above.
>
> A detailed message will help us understand how correctly the suppliers were 
> selected.
>
> 06.05.2020, 22:08, "Alexei Scherbakov" <alexey.scherbak...@gmail.com>:
>>  Hello.
>>
>>  Let's look at existing rebalancing log for a single group:
>>
>>  [2020-05-06 20:56:36,999][INFO ][...] Rebalancing scheduled
>>  [order=[ignite-sys-cache, cache1, cache2, default],
>>  top=AffinityTopologyVersion [topVer=3, minorTopVer=1],
>>  evt=DISCOVERY_CUSTOM_EVT, node=9d9edb7b-eb01-47a1-8ff9-fef715d00002]
>>  ...
>>  [2020-05-06 20:56:37,034][INFO ][...] Prepared rebalancing [grp=cache1,
>>  mode=ASYNC, supplier=94a3fcbc-18d5-4c64-b0ab-4313aba00001,
>>  partitionsCount=11, topVer=AffinityTopologyVersion [topVer=3,
>>  minorTopVer=1]]
>>  [2020-05-06 20:56:37,036][INFO ][...] Prepared rebalancing [grp=cache1,
>>  mode=ASYNC, supplier=b3f3aeeb-5fa0-42f7-a74e-cf39fa500000,
>>  partitionsCount=10, topVer=AffinityTopologyVersion [topVer=3,
>>  minorTopVer=1]]
>>  [2020-05-06 20:56:37,036][INFO ][...] Starting rebalance routine [cache1,
>>  topVer=AffinityTopologyVersion [topVer=3, minorTopVer=1],
>>  supplier=94a3fcbc-18d5-4c64-b0ab-4313aba00001, fullPartitions=[1, 5, 7, 9,
>>  11, 13, 15, 23, 27, 29, 31], histPartitions=[]]
>>  [2020-05-06 20:56:37,037][INFO ][...] Starting rebalance routine [cache1,
>>  topVer=AffinityTopologyVersion [topVer=3, minorTopVer=1],
>>  supplier=b3f3aeeb-5fa0-42f7-a74e-cf39fa500000, fullPartitions=[6, 8, 10,
>>  16, 18, 20, 22, 24, 26, 28], histPartitions=[]]
>>  [2020-05-06 20:56:37,044][INFO ][...] Completed rebalancing [grp=cache1,
>>  supplier=94a3fcbc-18d5-4c64-b0ab-4313aba00001,
>>  topVer=AffinityTopologyVersion [topVer=3, minorTopVer=1], progress=1/2]
>>  [2020-05-06 20:56:37,046][INFO ][...] Completed (final) rebalancing
>>  [grp=cache1, supplier=b3f3aeeb-5fa0-42f7-a74e-cf39fa500000,
>>  topVer=AffinityTopologyVersion [topVer=3, minorTopVer=1], progress=2/2]
>>  [2020-05-06 20:56:37,048][INFO ][...] Completed rebalance future:
>>  RebalanceFuture [grp=CacheGroupContext [grp=cache1],
>>  topVer=AffinityTopologyVersion [topVer=3, minorTopVer=1], rebalanceId=2,
>>  routines=2]
>>
>>  From these logs I'm already can get answers to 1 and 4.
>>  The logs look concise and easy to read and understand, and should
>>  remain what way.
>>
>>  But I think some proposed improvements can be done here without harm.
>>
>>  2. OK, let's add it to supplier info per cache with additional info:
>>
>>  [2020-05-06 20:56:37,044][INFO ][...] Completed rebalancing [grp=cache1,
>>  supplier=94a3fcbc-18d5-4c64-b0ab-4313aba00001, entries=100, duration=12ms,
>>  bytesRcvd=5M, topVer=AffinityTopologyVersion [topVer=3, minorTopVer=1],
>>  progress=1/2]
>>
>>  3. This information is already printed to log.
>>
>>  5. OK, let's add a line which is printed after all groups in the chain are
>>  rebalanced or cancelled with a summary:
>>
>>  [2020-05-06 20:56:36,999][INFO ][...] Completed rebalance chain:
>>  [rebalanceId=2, entries=200, duration=50ms, bytesRcvd=10M]
>>
>>  Another thing I would suggest: add rebalanceId to all messages for ease of
>>  grepping and to distinguish multiple rebalances for a same topology version.
>>
>>  Regarding the "detailed" message - I wasn't able to figure out any case
>>  where it is useful. Not sure if we need it at all.
>>  Kirill Tkalenko, can you give me an example?
>>
>>  ср, 6 мая 2020 г. в 21:01, Maxim Muzafarov <mmu...@apache.org>:
>>
>>>   Kirill,
>>>
>>>   Thank you for raising this topic. It's true that the rebalance process
>>>   still requires additional information for analyzing issues. Please,
>>>   don't think that I'm against your changes :-)
>>>
>>>   * My short answer. *
>>>
>>>   We won't do performance analysis on the production environment. Each
>>>   time we need performance analysis it will be done on a test
>>>   environment with verbose logging enabled. Thus I suggest moving these
>>>   changes to a separate `profiling` module and extend the logging much
>>>   more without any ышяу limitations. The same as these [2] [3]
>>>   activities do.
>>>
>>>   Let's keep the `core` module as simple as possible.
>>>   Let's design the right API for accessing rebalance internals for
>>>   profiling tools.
>>>   Can you, please, remove all changes from your PR [6] which are not
>>>   related to the proposed topic? (e.g. variable renamings).
>>>
>>>   * The long answer. *
>>>
>>>   Here are my thoughts on this. There are two different types of issues
>>>   in the rebalance process. The first case must be covered by daily
>>>   monitoring subsystems, the second case must be covered by additional
>>>   profiling tools.
>>>   1. errors during the rebalancing (e.g. rebalance not happens when 
>>> required)
>>>   2. performance rebalancing issues (e.g. rebalance is slow)
>>>
>>>   Daily monitoring tools (JMX, Logging) are always turned on and
>>>   shouldn't require additional systems resources themselves. Since these
>>>   metrics must be lightweight any internal aggregation machinery is not
>>>   used on them. All these metrics are collected from each node
>>>   independently. Please, take a look at this issue [1] which covers most
>>>   of your questions mentioned above.
>>>
>>>   For all available metrics, we can configure LogExporterSpi, so they
>>>   will be available in logs.
>>>
>>>   > 1)How long was the balance(divided by supplier)?
>>>   rebalancingStartTime, rebalancingEndTime already exists for cache groups
>>>   [4].
>>>   We can add the same for suppliers.
>>>
>>>   > 2)How many records and bytes per supplier were rebalanced?
>>>   We already have rebalancingReceivedKeys, rebalancingReceivedBytes [4]
>>>   We will have rebalancingExpectedKeys [5].
>>>   We can add a new metric per cache keys, per supplier.
>>>
>>>   > 3)How many times did the rebalance restart?
>>>   rebalancingLastCancelledTime [4] metric exists.
>>>   Do we need to keep historical data on it?
>>>
>>>   > 4)Which partitions were rebalanced and from which nodes did they receive
>>>   them?
>>>   Let's print this information to log prior to the rebalance process
>>>   starts. This will be helpful information and do not require a lot of
>>>   changes.
>>>
>>>   > 5)When did rebalance for all cache groups end?
>>>   This metric may be aggregated from rebalancingEndTime [4] by pulling
>>>   it from all nodes for all caches.
>>>
>>>   Performance rebalancing issues are related to profiling tools. They
>>>   may require additional system resources and definitely require a
>>>   dedicated environment for tests. We can't do performance analysis on
>>>   production environments due to performance impact.
>>>   I see some disadvantages of adding such tools to production code:
>>>   - verbose logging may affect performance.
>>>   - the problematic process may become even worse if an automatic
>>>   threshold suddenly turns on.
>>>   - new code changes will require additional efforts to keep logging
>>>   up-to-date.
>>>
>>>   [1] https://issues.apache.org/jira/browse/IGNITE-12183
>>>   [2] https://issues.apache.org/jira/browse/IGNITE-12666
>>>   [3]
>>>   
>>> https://cwiki.apache.org/confluence/display/IGNITE/Cluster+performance+profiling+tool
>>>   [4] https://issues.apache.org/jira/browse/IGNITE-12193
>>>   [5] https://issues.apache.org/jira/browse/IGNITE-12194
>>>   [6] https://github.com/apache/ignite/pull/7705/files
>>>
>>>   On Wed, 6 May 2020 at 19:50, Ivan Rakov <ivan.glu...@gmail.com> wrote:
>>>   >
>>>   > Hi,
>>>   >
>>>   > IGNITE_WRITE_REBALANCE_PARTITION_DISTRIBUTION_THRESHOLD - threshold
>>>   > > duration rebalance of cache group after which partitions distribution
>>>   is
>>>   > > output, set in milliseconds, default value is 10 minutes.
>>>   >
>>>   > Does it mean that if the rebalancing process took less than 10 minutes,
>>>   > only a short version of the message (with supplier statistics) will show
>>>   up?
>>>   >
>>>   > In general, I have no objections.
>>>   >
>>>   >
>>>   > On Mon, May 4, 2020 at 10:38 AM ткаленко кирилл <tkalkir...@yandex.ru>
>>>   > wrote:
>>>   >
>>>   > > Hi, Igniters!
>>>   > >
>>>   > > I'd like to share a new small feature in AI [1].
>>>   > >
>>>   > > Current rebalance logging does not allow you to quickly answer
>>>   following
>>>   > > questions:
>>>   > > 1)How long was the balance(divided by supplier)?
>>>   > > 2)How many records and bytes per supplier were rebalanced?
>>>   > > 3)How many times did rebalance restart?
>>>   > > 4)Which partitions were rebalanced and from which nodes did they
>>>   receive
>>>   > > them?
>>>   > > 5)When did rebalance for all cache groups end?
>>>   > >
>>>   > > What you can see in logs now:
>>>   > >
>>>   > > 1)Starting rebalance with order of cache groups.
>>>   > > Rebalancing scheduled [order=[ignite-sys-cache, grp1, grp0],
>>>   > > top=AffinityTopologyVersion [topVer=2, minorTopVer=0], force=false,
>>>   > > evt=NODE_JOINED, node=c2146a04-dc23-4bc9-870d-dfbb55c00001]
>>>   > >
>>>   > > 2)Start rebalance of cache group from a specific supplier, specifying
>>>   > > partition ids and mode - historical or full.
>>>   > > Starting rebalance routine [ignite-sys-cache,
>>>   > > topVer=AffinityTopologyVersion [topVer=2, minorTopVer=0],
>>>   > > supplier=8c525892-703b-4fc4-b28b-b2f139700000, fullPartitions=[0-99],
>>>   > > histPartitions=[]]
>>>   > >
>>>   > > 3)Getting partial or complete partitions of cache group.
>>>   > > Completed rebalancing [grp=ignite-sys-cache,
>>>   > > supplier=8c525892-703b-4fc4-b28b-b2f139700000,
>>>   > > topVer=AffinityTopologyVersion [topVer=5, minorTopVer=0], 
>>> progress=1/2]
>>>   > > Completed (final) rebalancing [grp=ignite-sys-cache,
>>>   > > supplier=c2146a04-dc23-4bc9-870d-dfbb55c00001,
>>>   > > topVer=AffinityTopologyVersion [topVer=5, minorTopVer=0], 
>>> progress=2/2]
>>>   > >
>>>   > > 4)End rebalance of cache group.
>>>   > > Completed rebalance future: RebalanceFuture [grp=CacheGroupContext
>>>   > > [grp=ignite-sys-cache], topVer=AffinityTopologyVersion [topVer=2,
>>>   > > minorTopVer=0], rebalanceId=1, routines=1, receivedBytes=1200,
>>>   > > receivedKeys=0, partitionsLeft=0, startTime=1588519707607, endTime=-1,
>>>   > > lastCancelledTime=-1]
>>>   > >
>>>   > > Rebalance statistics:
>>>   > >
>>>   > > To speed up rebalance analysis, statistics will be output for each
>>>   cache
>>>   > > group and total for all cache groups.
>>>   > > If duration rebalance for cache group is greater than threshold value,
>>>   > > partition distribution is output.
>>>   > > Statistics will you to analyze duration of the balance for each
>>>   supplier
>>>   > > to understand which of them has been transmitting data for longest
>>>   time.
>>>   > >
>>>   > > System properties are used to output statistics:
>>>   > >
>>>   > > IGNITE_QUIET - to output statistics, value must be false;
>>>   > > IGNITE_WRITE_REBALANCE_PARTITION_DISTRIBUTION_THRESHOLD - threshold
>>>   > > duration rebalance of cache group after which partitions distribution
>>>   is
>>>   > > output, set in milliseconds, default value is 10 minutes.
>>>   > >
>>>   > > Statistics examples:
>>>   > >
>>>   > > Successful full and historical rebalance of group cache, without
>>>   > > partitions distribution.
>>>   > > Rebalance information per cache group (successful rebalance):
>>>   [id=3181548,
>>>   > > name=grp1, startTime=2020-04-13 10:55:16,117, finishTime=2020-04-13
>>>   > > 10:55:16,127, d=10 ms, restarted=0] Supplier statistics: [nodeId=0,
>>>   p=5,
>>>   > > d=10 ms] [nodeId=1, p=5, d=10 ms] Aliases: p - partitions, e -
>>>   entries, b -
>>>   > > bytes, d - duration, h - historical, nodeId mapping
>>>   > > (nodeId=id,consistentId) [0=rebalancing.RebalanceStatisticsTest1]
>>>   > > [1=rebalancing.RebalanceStatisticsTest0]
>>>   > > Rebalance information per cache group (successful rebalance):
>>>   [id=3181547,
>>>   > > name=grp0, startTime=2020-04-13 15:01:44,000, finishTime=2020-04-13
>>>   > > 15:01:44,116, d=116 ms, restarted=0] Supplier statistics: [nodeId=0,
>>>   hp=10,
>>>   > > he=300, hb=30267, d=116 ms] Aliases: p - partitions, e - entries, b -
>>>   > > bytes, d - duration, h - historical, nodeId mapping
>>>   > > (nodeId=id,consistentId) [0=rebalancing.RebalanceStatisticsTest0]
>>>   > >
>>>   > > Successful full and historical rebalance of group cache, with
>>>   partitions
>>>   > > distribution.
>>>   > > Rebalance information per cache group (successful rebalance):
>>>   [id=3181548,
>>>   > > name=grp1, startTime=2020-04-13 10:55:16,117, finishTime=2020-04-13
>>>   > > 10:55:16,127, d=10 ms, restarted=0] Supplier statistics: [nodeId=0,
>>>   p=5,
>>>   > > d=10 ms] [nodeId=1, p=5, d=10 ms] Aliases: p - partitions, e -
>>>   entries, b -
>>>   > > bytes, d - duration, h - historical, nodeId mapping
>>>   > > (nodeId=id,consistentId) [0=rebalancing.RebalanceStatisticsTest1]
>>>   > > [1=rebalancing.RebalanceStatisticsTest0] Rebalance duration was 
>>> greater
>>>   > > than 5 ms, printing detailed information about partitions distribution
>>>   > > (threshold can be changed by setting number of milliseconds into
>>>   > > IGNITE_WRITE_REBALANCE_PARTITION_DISTRIBUTION_THRESHOLD) 0 =
>>>   > > [0,bu,su],[1,bu],[2,pr,su] 1 = [0,bu,su],[1,bu],[2,pr,su] 2 =
>>>   > > [0,bu,su],[1,bu],[2,pr,su] 3 = [0,bu,su],[1,bu],[2,pr,su] 4 =
>>>   > > [0,bu,su],[1,bu],[2,pr,su] 5 = [0,bu,su],[1,bu],[2,pr,su] 6 =
>>>   > > [0,bu,su],[1,bu],[2,pr,su] 7 = [0,bu,su],[1,bu],[2,pr,su] 8 =
>>>   > > [0,bu,su],[1,bu],[2,pr,su] 9 = [0,bu,su],[1,bu],[2,pr,su] Aliases: pr 
>>> -
>>>   > > primary, bu - backup, su - supplier node, h - historical, nodeId
>>>   mapping
>>>   > > (nodeId=id,consistentId) [0=rebalancing.RebalanceStatisticsTest1]
>>>   > > [1=rebalancing.RebalanceStatisticsTest2]
>>>   > > [2=rebalancing.RebalanceStatisticsTest0]
>>>   > > Rebalance information per cache group (successful rebalance):
>>>   [id=3181547,
>>>   > > name=grp0, startTime=2020-04-13 15:01:44,000, finishTime=2020-04-13
>>>   > > 15:01:44,116, d=116 ms, restarted=0] Supplier statistics: [nodeId=0,
>>>   hp=10,
>>>   > > he=300, hb=30267, d=116 ms] Aliases: p - partitions, e - entries, b -
>>>   > > bytes, d - duration, h - historical, nodeId mapping
>>>   > > (nodeId=id,consistentId) [0=rebalancing.RebalanceStatisticsTest0]
>>>   Rebalance
>>>   > > duration was greater than 100 ms, printing detailed information about
>>>   > > partitions distribution (threshold can be changed by setting number of
>>>   > > milliseconds into
>>>   IGNITE_WRITE_REBALANCE_PARTITION_DISTRIBUTION_THRESHOLD)
>>>   > > 0 = [0,pr,su],[1,bu],[2,bu] h 1 = [0,bu,su],[1,bu],[2,pr] h 2 =
>>>   > > [0,pr,su],[1,bu],[2,bu] h 3 = [0,bu,su],[1,bu],[2,pr] h 4 =
>>>   > > [0,pr,su],[1,bu],[2,bu] h 5 = [0,pr,su],[1,bu],[2,bu] h 6 =
>>>   > > [0,bu,su],[1,bu],[2,pr] h 7 = [0,pr,su],[1,bu],[2,bu] h 8 =
>>>   > > [0,pr,su],[1,bu],[2,bu] h 9 = [0,pr,su],[1,bu],[2,bu] h Aliases: pr -
>>>   > > primary, bu - backup, su - supplier node, h - historical, nodeId
>>>   mapping
>>>   > > (nodeId=id,consistentId) [0=rebalancing.RebalanceStatisticsTest0]
>>>   > > [1=rebalancing.RebalanceStatisticsTest2]
>>>   > > [2=rebalancing.RebalanceStatisticsTest1]
>>>   > >
>>>   > > Interrupted rebalance of group cache.
>>>   > > Rebalance information per cache group (interrupted rebalance):
>>>   > > [id=644280849, name=default2, startTime=2020-04-13 14:55:24,969,
>>>   > > finishTime=2020-04-13 14:55:24,969, d=0 ms, restarted=0]
>>>   > >
>>>   > > Total full and historical rebalance for all cache groups.
>>>   > > Rebalance total information (including successful and not rebalances):
>>>   > > [startTime=2020-04-13 10:55:18,726, finishTime=2020-04-13 
>>> 10:55:18,780,
>>>   > > d=54 ms] Supplier statistics: [nodeId=0, p=60, e=250, b=25000, d=54 
>>> ms]
>>>   > > [nodeId=1, p=60, e=250, b=24945, d=54 ms] Aliases: p - partitions, e -
>>>   > > entries, b - bytes, d - duration, h - historical, nodeId mapping
>>>   > > (nodeId=id,consistentId) [0=rebalancing.RebalanceStatisticsTest1]
>>>   > > [1=rebalancing.RebalanceStatisticsTest0]
>>>   > > Rebalance total information (including successful and not rebalances):
>>>   > > [startTime=2020-04-13 15:01:43,822, finishTime=2020-04-13 
>>> 15:01:44,116,
>>>   > > d=294 ms] Supplier statistics: [nodeId=0, hp=20, he=500, hb=50445,
>>>   d=294
>>>   > > ms] Aliases: p - partitions, e - entries, b - bytes, d - duration, h -
>>>   > > historical, nodeId mapping (nodeId=id,consistentId)
>>>   > > [0=rebalancing.RebalanceStatisticsTest0]
>>>   > >
>>>   > > [1] - https://issues.apache.org/jira/browse/IGNITE-12080
>>
>>  --
>>
>>  Best regards,
>>  Alexei Scherbakov

Reply via email to