hi, I have googled and find similar question without good answer, http://stackoverflow.com/questions/24520225/writing-to-hadoop-distributed-file-system-multiple-times-with-spark
in short, I would like to separate raw data and divide by some key, for example, create date, and put the in directory named by date, so that I can easily access portion of data later. for now I have to extract all keys and then filter by key and save to file repeatly. are there any good way to do this? or maybe I shouldn't do such thing?