??????Carbon??????????????????????????????????????????????????????????????????????????????????
??????????Carbon??????????????????????????????????




------------------ Original ------------------
From:&nbsp;"chenliang6136"<[email protected]&gt;;
Date:&nbsp;Tue, Apr 14, 2020 08:36 PM
To:&nbsp;"dev"<[email protected]&gt;;

Subject:&nbsp;Re: Carbon over-use cluster resources



OK, thank you feedbacked this issue, let us look into it.

Regards
Liang


Manhua Jiang wrote
&gt; Hi All,
&gt; Recently, I found carbon over-use cluster resources. Generally the design
&gt; of carbon work flow does not act as common spark task which only do one
&gt; small work in one thread, but the task has its mind/logic.
&gt; 
&gt; For example,
&gt; 1.launch carbon with --num-executors=1 but set
&gt; carbon.number.of.cores.while.loading=10;
&gt; 2.no_sort table with multi-block input, N Iterator
&gt; <CarbonRowBatch&gt;
&gt;&nbsp; for example, carbon will start N tasks in parallel. And in each task 
the
&gt; CarbonFactDataHandlerColumnar has model.getNumberOfCores() (let's say C)
&gt; in ProducerPool. Totally launch N*C threads; ==&gt;This is the case makes 
me
&gt; take this as serious problem. To many threads stucks the executor to send
&gt; heartbeat and be killed.
&gt; 
&gt; So, the over-use is related to usage of threadpool.
&gt; 
&gt; This would affect the cluster overall resource usage and may lead to wrong
&gt; performance results.
&gt; 
&gt; I hope this get your notice while fixing or writing new codes.





--
Sent from: 
http://apache-carbondata-dev-mailing-list-archive.1130556.n5.nabble.com/

Reply via email to