wren ng thornton ha scritto:
Manlio Perillo wrote:
Since ratings for each customers are parsed "at the same time", using a plain list would consume a lot of memory, since stream fusion can only be executed at the end of the parsing.

On the other hand, when I want to group ratings by movies, stream fusion seems to work fine.

[...]

For the problem as you've discussed it, I'd suggest a different approach: You can't fit all the data into memory at once, so you shouldn't try to. You should write one program that takes in the per-movie grouping of data and produces a per-user file as output.

Well, creating 480189 files in a directory is not a very nice thing to do to a normal file system.

I should arrange files in directory, but then this starts to become too complex.

The solution I'm using now just works.
It takes about 950 MB of memory and 35 minutes, but it's not a big problem since:
1) Once loaded, I can serialize the data in binary format
2) I think that the program can be parallelized, parsing
   subsets of the files in N threads, and then merging the maps.

   Using this method, should optimize array copying.

   The problem is that unionWith seems to be lazy, and there is no no
   strict variant; I'm not sure.

Then have your second program read in the reorganized data and do fusion et al.

This reduces the problem to just writing the PerMovie -> PerUser program. Since you still can't fit all the data into memory, that means you can't hope to write the per-user file in one go.

The data *do* fit into memory, fortunately.

> [...]

Best of luck.



Thanks  Manlio
_______________________________________________
Haskell-Cafe mailing list
Haskell-Cafe@haskell.org
http://www.haskell.org/mailman/listinfo/haskell-cafe

Reply via email to