hi Li, Yes,i got the partition folder as you say, but under the partition folder ,there are many small file just like following picture, How to merge then automatically after jobs done.
thanks ChenXingYu ------------------ Original ------------------ From: "Jacky Li"<jacky.li...@qq.com>; Date: Tue, Jun 5, 2018 08:43 PM To: "dev"<dev@carbondata.apache.org>; Subject: Re: carbondata partitioned by date generate many small files Hi, There is a testcase in StandardPartitionTableQueryTestCase used date column as partition column, if you run that testcase, the partition folder generated looks like following picture. Are you getting similar folders? Regards, Jacky 在 2018年6月5日,下午2:49,陈星宇 <chenxingyu...@keruyun.com> 写道: hi carbondata team, i am using carbondata 1.3.1 to create table and import data, generated many small files and spark job is very slow, i suspected the number of file is related to the number of spark job . but if i decrease the jobs, job will fail because of outofmemory. see my ddl as below: create table xx.xx( dept_name string, xx . . . ) PARTITIONED BY (xxx date) STORED BY 'carbondata' TBLPROPERTIES('SORT_COLUMNS'='xxx,xxx,xxx ,xxx,xxx') please give some advice. thanks ChenXingYu