[
https://issues.apache.org/jira/browse/CARBONDATA-1281?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16104607#comment-16104607
]
xuchuanyin commented on CARBONDATA-1281:
----------------------------------------
Here I will provide the configuration used in my test for others to reference.
# ENV
3 HUAWEI RH2288 nodes, each has 24 Cores([email protected]), 256GB MEM, 11
Disks(SAS)
# USE CASE
88Billion Recods with CSV format
340+ columns per record
NO Dictionary column
TABLE_BLOCKSIZE 64
INVERTED_INDEX about 9 columns
# CONF
parameter value origin-value
carbon.number.of.cores 20
carbon.number.of.cores.while.loading 14
sort.inmemory.size.inmb 2048 1024
offheap.sort.chunk.size.inmb 128 64
carbon.sort.intermediate.files.limit 20 20
carbon.sort.file.buffer.size 50 20
carbon.use.local.dir true false
carbon.use.multiple.dir true false
# RESULT
Using `LOAD DATA INPATH `, the loading cost about 6min
Observing the NMON, each disk IO usage is quite average.
> Disk hotspot found during data loading
> --------------------------------------
>
> Key: CARBONDATA-1281
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1281
> Project: CarbonData
> Issue Type: Improvement
> Components: core, data-load
> Affects Versions: 1.1.0
> Reporter: xuchuanyin
> Assignee: xuchuanyin
> Fix For: 1.2.0
>
> Time Spent: 17.5h
> Remaining Estimate: 0h
>
> # Scenario
> Currently we have done a massive data loading. The input data is about 71GB
> in CSV format,and have about 88million records. When using carbondata, we do
> not use any dictionary encoding. Our testing environment has three nodes and
> each of them have 11 disks as yarn executor directory. We submit the loading
> command through JDBCServer.The JDBCServer instance have three executors in
> total, one on each node respectively. The loading takes about 10minutes
> (+-3min vary from each time).
> We have observed the nmon information during the loading and find:
> 1. lots of CPU waits in the first half of loading;
> 2. only one single disk has many writes and almost reaches its bottleneck
> (Avg. 80M/s, Max. 150M/s on SAS Disk)
> 3. the other disks are quite idel
> # Analyze
> When do data loading, carbondata read and sort data locally(default scope)
> and write the temp files to local disk. In my case, there is only one
> executor in one node, so carbondata write all the temp file to one
> disk(container directory or yarn local directory), thus resulting into single
> disk hotspot.
> # Modification
> We should support multiple directory for writing temp files to avoid disk
> hotspot.
> Ps: I have improved this in my environment and the result is pretty
> optimistic: the loading takes about 6minutes (10 minutes before improving).
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)