Hi, It seems attachments are missing.Can you attach them again.
Regards, Ravindra. On 30 March 2017 at 08:02, a <ww...@163.com> wrote: > Hello! > > *Test result:* > When I load csv data into carbondata table 3 times,the executors > distributed unevenly。My purpose > <http://www.baidu.com/link?url=7rAmHkI2HPR9Hz-TG3467RHKqj_V1KLkZC_mMn3HW6HbyCQR1aDIDSiAZdAZGWEda5ZonK2CFcNh_wXtsSW0YVa_n0NK-dBg3708mv1qeXm> > is > one node one task,but the result is some node has 2 task and some node has > no task。 > See the load data 1.png,data 2.png,data 3.png。 > The carbondata data.PNG is the data structure in hadoop. > > I load 4 0000 0000 records into carbondata table takes 2629s seconds,its > too long。 > > *Question:* > How can i make the executors distributed evenly ? > > The environment: > spark2.1+carbondata1.1,there are 7 datanodes. > > *./bin/spark-shell \--master yarn \--deploy-mode client > \--num-executors n \ (the first time is 7(result in load data 1.png),the > second time is 6(result in load data 2.png),the three time is 8(result in > load data3.png))--executor-cores 10 \--executor-memory 40G \--driver-memory > 8G \* > > carbon.properties > ######## DataLoading Configuration ######## > carbon.sort.file.buffer.size=20 > carbon.graph.rowset.size=10000 > carbon.number.of.cores.while.loading=10 > carbon.sort.size=50000 > carbon.number.of.cores.while.compacting=10 > carbon.number.of.cores=10 > > Best regards! > > > > > > > -- Thanks & Regards, Ravi