Andrew Wong created KUDU-3168:
---------------------------------
Summary: Script or document how to identify tables that suffer
from KUDU-1400
Key: KUDU-3168
URL: https://issues.apache.org/jira/browse/KUDU-3168
Project: Kudu
Issue Type: Task
Components: compaction, documentation
Affects Versions: 1.8.0
Reporter: Andrew Wong
In older versions of Kudu, tables may suffer from KUDU-1400 when workloads are
sequential and have a low ingest rate (e.g. KBs per minute). Today, the way to
identify issues is to notice that scans for a specific tablet are slow for the
amount of data being scanned, looking at that tablet's rowset layout diagram,
and noting that a large number of rowsets. The guidance then is usually to
rewrite the table with a higher ingest rate (e.g. through an Impala CTAS) and
that typically solves the issue for that table.
Users may want ways to identify tables (not tablets) that suffer from this
issue so they can be rewritten. It would be nice to document how to do this
using existing tooling (or make a script available).
The {{kudu fs list}} tool with the {{rowset-id}} column filter seems like it
would be useful in identifying the number of rowsets there are and how large
they are. If used with {{kudu table list}} users ought to easily identify
affected tables.
In later versions of Kudu, the {{num_rowsets_on_disk}} metric should be useful
in identifying such cases (compute {{tablet size / num_rowsets_on_disk}} and
compare against 32MB, the current default target DRS size).
--
This message was sent by Atlassian Jira
(v8.3.4#803005)