hi,

I have googled and find similar question without good answer,
http://stackoverflow.com/questions/24520225/writing-to-hadoop-distributed-file-system-multiple-times-with-spark

in short, I would like to separate raw data and divide by some key, for
example, create date, and put the in directory named by date, so that I can
easily access portion of data later.

for now I have to extract all keys and then filter by key and save to file
repeatly. are there any good way to do this?  or maybe I shouldn't do such
thing?

Reply via email to