Re: TreeReduce Functionality in Spark

2015-06-04 Thread DB Tsai
tor(aggregatePartition(it))) >> >>> var numPartitions = partiallyAggregated.partitions.length >> >>> val scale = math.max(math.ceil(math.pow(numPartitions, 1.0 / >> >>> depth)).toInt, 2) >> >>> //

Re: TreeReduce Functionality in Spark

2015-06-04 Thread Raghav Shankar
7;t help reduce > >>> // the wall-clock time, we stop tree aggregation. > >>> while (numPartitions > scale + numPartitions / scale) { > >>> numPartitions /= scale > >>> val curNumPartitions

Re: TreeReduce Functionality in Spark

2015-06-04 Thread DB Tsai
partiallyAggregated = partiallyAggregated.mapPartitionsWithIndex >>> { >>> (i, iter) => iter.map((i % curNumPartitions, _)) >>> }.reduceByKey(new HashPartitioner(curNumPartitions), >>> cleanCombOp).values >>> } >>&g

Re: TreeReduce Functionality in Spark

2015-06-04 Thread Raghav Shankar
urNumPartitions), >> cleanCombOp).values >> } >> partiallyAggregated.reduce(cleanCombOp) >> >> I am completely lost about what is happening in this function. I would >> greatly appreciate some sort of explanation. >> >> >> >> >&

Re: TreeReduce Functionality in Spark

2015-06-04 Thread Reza Zadeh
cleanCombOp).values > } > partiallyAggregated.reduce(cleanCombOp) > > I am completely lost about what is happening in this function. I would > greatly appreciate some sort of explanation. > > > > > -- > View this message in context: > h

TreeReduce Functionality in Spark

2015-06-03 Thread raggy
tly appreciate some sort of explanation. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/TreeReduce-Functionality-in-Spark-tp23147.html Sent from the Apache Spark User List mailing list archive at Nabble.com. -