You do them sequentially in code; Spark will take care of combining them in execution. so something like: foo.map(fcn to [K1, V1]).reduceByKey(fcn from (V1, V1) to V1).map(fcn from (K1, V1) to (K2, V2))
On Wed, Jul 23, 2014 at 11:22 AM, soumick86 <sdasgu...@dstsystems.com> wrote: > How can I transform the mapper key at the reducer output. The functions I > have encountered are combineByKey, reduceByKey, etc that work on the values > and not on the key. For example below, this is what I want to achieve but > seems like I can only have K1 and not K2: > > Mapper----->(K1,V1)----->Reducer----->(K2,V2) > > I must be missing something. Is there a class/method available? Also I am > using the Java API > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Have-different-reduce-key-than-mapper-key-tp10503.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > -- Nathan Kronenfeld Senior Visualization Developer Oculus Info Inc 2 Berkeley Street, Suite 600, Toronto, Ontario M5A 4J5 Phone: +1-416-203-3003 x 238 Email: nkronenf...@oculusinfo.com