First post on the list! I like the ideas behind camlistore and I'm toying with an idea where camlistore might be the backend for my collection of photos/videos including a representation of the process from originally snapped photos into whatever version(s) I end up with as the final result. The process could be combination of photos to HDR/panoramas, just tuning colors or GIMPing it up in other ways. From what I understand camlistore will offer me the ability to see all versions of a given file by means of claims to a permanode which would actually replace the camliContent / bytes of the file. That helps a lot for single file projects. Additionally I would need to make some sort of grouping concept, which I guess I can do with attrs. Somehow. Anyway that's what I'll try to accomplish, if you should have any suggestions feel free to post them! :)
I've followed camlistore for a while but it's only the last week that I've actually installed it and it's now been working a few days while I'm trying to dump my entire photos folder into it. I have some questions that have crossed my mind as I'm seeing how it works. I'm using the 0.9 release now. - The blobs are adressed by their SHA1 (etc) digest. There is a theoretical possibility that two different blobs have the same digest, right? Is that handled somehow? - I'm using FUSE/cammount and my system seems to max out at about 700 KB/s transferring to camlistore. It seems to be transferring the big blobs, I guess the actual data blobs, quite fast, then it's spending some time creating smaller blobs (seems to be the bytes-indexes) after that. Seems it's maxing out the CPU as it's being slow (quite old hardware). Finally it spends some more time "packing file". I will try the monthly build too, but are there any typical things I can do to speed it up? - I see that most blobs from a big file will be around ~64KB but their sizes vary a lot. Also among the first chunks it seems typical to see one of 200KB+. What is the strategy to chunk up the file? I read somewhere that it's up to the client so I guess maybe the question is how cammount does it. With regard to deduplication I suppose the that strategy can be quite important, like if some header was changed in length but the rest of the file was the same, it could mean with ideal chunks only 1 chunk (the header chunk) was new and the rest unchanged, but with less ideal chunking all chunks would be new... Sorry for the random ramble. I hope somewhere down the line I'll get more into both go and camlistore and be able to contribute and do something with my mentioned idea. -- You received this message because you are subscribed to the Google Groups "Camlistore" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. For more options, visit https://groups.google.com/d/optout.
