Dear Sab ,
I must appreciate your kind reply very much, it would be much helpful.
On Monday, December 21, 2015 8:49 PM, Sabarish Sasidharan
wrote:
collect() will bring everything to driver and is costly. Instead of using
collect() + parallelize, you could use rdd1.checkpoint() along with a more
efficient action like rdd1.count(). This you can do within the for loop.
-Do
you want to apply checkpoint to cut out the lineage of DAG , however, as
tested, it seemed that checkpoint is more costlythan collect ...
Hopefully you are using the Kryo serializer already.
This would be all right. From your experience , is Kryo improve efficiency
obviously ...
RegardsSab
On Mon, Dec 21, 2015 at 5:51 PM, Zhiliang Zhu
wrote:
Dear All.
I have some kind of iteration job, that is, the next stag's input would be the
previous stag's output , and it must do quite lots of times of iteration.
JavaRDD rdd1 = //rdd1 may be with one or more
partitions
for (int i=0, JavaRDD rdd2 = rdd1; i < N; ++i) { JavaRDD rdd3 =
rdd2.map(new MapName1(...)); // 1 rdd4 = rdd3.map(new MapName2());
// 2
List list = rdd4.collect(); //however, N is very big, then
this line will be VERY MUCH COST
//Would checkpoint be used in the rdd which will be generated after lots of
steps.//here rdd2 or rdd1 seemed not proper to checkpoint
rdd2 = jsc.parallelize(list, M).cache();}
Is there way to properly improve the run speed?
In fact, I would like to apply spark to mathematica optimization by genetic
algorithm , in the above codes, rdd would be the Vector lines storing ,1 is to count fitness number, and 2 is to breed and variate .To
get good solution, the iteration number will be big (for example more than 1000
) ...
Thanks in advance!Zhiliang
On Monday, December 21, 2015 7:44 PM, Zhiliang Zhu
wrote:
Dear All,
I need to iterator some job / rdd quite a lot of times, but just lost in the
problem of spark only accept to call around 350 number of map before it meets
one action Function , besides, dozens of action will obviously increase the run
time.Is there any proper way ...
As tested, there is piece of codes as follows:
..
83 int count = 0; 84 JavaRDD dataSet = jsc.parallelize(list,
1).cache(); //with only 1 partition 85 int m = 350; 86
JavaRDD r = dataSet.cache(); 87 JavaRDD t = null; 88 89
for(int j=0; j < m; ++j) { //outer loop to temporarily convert the rdd r to t
90 if(null != t) { 91 r = t; 92 } //inner loop
to call map 350 times , if m is much more than 350 (for instance, around 400),
then the job will throw exception message "15/12/21 19:36:17
ERROR yarn.ApplicationMaster: User class threw exception:
java.lang.StackOverflowError java.lang.StackOverflowError") 93 for(int
i=0; i < m; ++i) { 94 r = r.map(new Function() { 95
@Override 96 public Integer call(Integer integer) { 97
double x = Math.random() * 2 - 1; 98 double y = Math.random()
* 2 - 1; 99 return (x * x + y * y < 1) ? 1 : 0;100 }101
});
104 }105106 List lt = r.collect(); //then collect this rdd
to get another rdd, however, dozens of action Function as collect is VERY MUCH
COST107 t = jsc.parallelize(lt, 1).cache();108109 }110..
Thanks very much in advance!Zhiliang
Thanks in advance !
--
Architect - Big Data
Ph: +91 99805 99458
Manthan Systems | Company of the year - Analytics (2014 Frost and Sullivan
India ICT)+++