related issue)
>>>> so bottom line - how to join big rdd with smaller rdd without
>>>> reshuffling
>>>> big rdd over and over again?
>>>> As soon as I've saved this big rdd and reloaded it from disk I want that
>>>> every other rdd will be
; big rdd over and over again?
>>> As soon as I've saved this big rdd and reloaded it from disk I want that
>>> every other rdd will be partitioned and collocated by the same
>>> "partitioner"(which is absent for hadooprdd) ... somehow, so that only
>>
f small rdds and then join, however I'm not sure it's possible to
>> implement this filter without join.
>>
>> any ideas would be appreciated,
>> thanks in advance
>> Igor
>>
>>
>>
>> --
>> View this message in context:
>> htt
oin, however I'm not sure it's possible to
> implement this filter without join.
>
> any ideas would be appreciated,
> thanks in advance
> Igor
>
>
>
> --
> View this message in context:
>
somehow split baseBlock into 2 parts with filter by
keys of small rdds and then join, however I'm not sure it's possible to
implement this filter without join.
any ideas would be appreciated,
thanks in advance
Igor
--
View this message in context:
http://apache-spark-user-list.1001560.n3