[
https://issues.apache.org/jira/browse/TAJO-283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13855821#comment-13855821
]
Min Zhou edited comment on TAJO-283 at 12/23/13 6:39 PM:
---------------------------------------------------------
[~jihoonson]
That's absolutely a good question!
I have thought about this problem. Firstly, we should figure out how large the
number of partitions is acceptable. From my experience, MySQL works well if we
insert thousands of rows in a time, even tens of thousands are still
acceptable. But if the order of magnitude grows to hundreds of thousands , even
millions or more, MySQL would be very slow when inserting&retrieving those
records.
When we are using HASH partition, since we can defined the buckets number of
hash function, I think the number is under control. Normally it should be tens
or hundreds . For RANGE and LIST partition, it works as well due to the
partitions is enumerable. The worst situation I think is when we are using
COLUMN partitions on a table, which is quite similar with hive's dynamic
partition list below.
{noformat}
CREATE TABLE dst_tbl (key int, value string) PARTITIONED BY (col1 string, col2
it) AS
SELECT key, value, col1, col2 FROM src_tbl
{noformat}
Query users always have no knowledge about this table's value distribution. If
the table is with high cardinality (a.k.a with so many distinct values), that
should be a disaster for the below area
1. The number of files/directories on hdfs would be very large, big pressure
for HDFS namenode's memory
2. As you mentioned, this would be a big problem for catalog.
Acutally, due to the above reasons. In Alibaba.com, my previous employer, which
has one of the largest single hadoop cluster in the world, we disabled dynamic
partitioning. I think you should run into the same problem when you are using
column partitioning. I don't know why you guys decide to support such feature,
could you give me some background about it? How can we benefit from column
partitions?
[~hyunsik]
It's good to know tajo will support indexes. I saw the binary search tree
index in the branch. Actually, I am considering about adding lucene index into
tajo, through which we can implements an online BA system on the top of tajo
like senseidb. We can do aggregations on billions of rows with only a few
milliseconds. If I implement it, we can put tajo into production in linkedin,
my current employer.
[~hyunsik] [~jihoonson]
Thank you. Merry Christmas!
Min
was (Author: coderplay):
[~jihoonson]
That's absolutely a good question!
I have thought about this problem. Firstly, we should figure out how large the
number of partitions is acceptable. From my experience, MySQL works well if we
insert thousands of rows in a time, even tens of thousands are still
acceptable. But if the order of magnitude grows to hundreds of thousands , even
millions or more, MySQL would be very slow when inserting&retrieving those
records.
When we are using HASH partition, since we can defined the buckets number of
hash function, I think the number is under control. Normally it should be tens
or hundreds . For RANGE and LIST partition, it works as well due to the
partitions is enumerable. The worst situation I think is when we are using
COLUMN partitions on a table, which is quite similar with hive's dynamic
partition list below.
{noformat}
CREATE TABLE dst_tbl (key int, value string) PARTITIONED BY (col1 string, col2
it) AS
SELECT key, value, col1, col2 FROM src_tbl
{noformat}
Query users always have no knowledge about this table's value distribution. If
the table is with high cardinality (a.k.a with so many distinct values), that
should be a disaster for the below area
1. The number of files/directories on hdfs would be very large, big pressure
for HDFS namenode's memory
2. As you mentioned, this would be a big problem for catalog.
Acutally, due to the above reasons. In Alibaba.com, my previous employer, which
has one of the largest single hadoop cluster in the world, we disabled dynamic
partitioning. I think you should run into the same problem when you are using
column partitioning. I don't know why you guys decide to support such feature,
could you give me some background about it? How can we benefit from column
partitions?
[~hyunsik] [~jihoonson]
Thank you. Merry Christmats!
Min
> Add Table Partitioning
> ----------------------
>
> Key: TAJO-283
> URL: https://issues.apache.org/jira/browse/TAJO-283
> Project: Tajo
> Issue Type: New Feature
> Components: catalog, physical operator, planner/optimizer
> Reporter: Hyunsik Choi
> Assignee: Hyunsik Choi
> Fix For: 0.8-incubating
>
>
> Table partitioning gives many facilities to maintain large tables. First of
> all, it enables the data management system to prune many input data which are
> actually not necessary. In addition, it gives the system more optimization
> opportunities that exploit the physical layouts.
> Basically, Tajo should follow the RDBMS-style partitioning system, including
> range, list, hash, and so on. In order to keep Hive compatibility, we need to
> add Hive partition type that does not exists in existing DBMS systems.
--
This message was sent by Atlassian JIRA
(v6.1.5#6160)