First post on the list!

I like the ideas behind camlistore and I'm toying with an idea where 
camlistore might be the backend for my collection of photos/videos 
including a representation of the process from originally snapped photos 
into whatever version(s) I end up with as the final result. The process 
could be combination of photos to HDR/panoramas, just tuning colors or 
GIMPing it up in other ways. From what I understand camlistore will offer 
me the ability to see all versions of a given file by means of claims to a 
permanode which would actually replace the camliContent / bytes of the 
file. That helps a lot for single file projects. Additionally I would need 
to make some sort of grouping concept, which I guess I can do with attrs. 
Somehow. Anyway that's what I'll try to accomplish, if you should have any 
suggestions feel free to post them! :)

I've followed camlistore for a while but it's only the last week that I've 
actually installed it and it's now been working a few days while I'm trying 
to dump my entire photos folder into it. I have some questions that have 
crossed my mind as I'm seeing how it works. I'm using the 0.9 release now.

- The blobs are adressed by their SHA1 (etc) digest. There is a theoretical 
possibility that two different blobs have the same digest, right? Is that 
handled somehow?

- I'm using FUSE/cammount and my system seems to max out at about 700 KB/s 
transferring to camlistore. It seems to be transferring the big blobs, I 
guess the actual data blobs, quite fast, then it's spending some time 
creating smaller blobs (seems to be the bytes-indexes) after that. Seems 
it's maxing out the CPU as it's being slow (quite old hardware). Finally it 
spends some more time "packing file". I will try the monthly build too, but 
are there any typical things I can do to speed it up?

- I see that most blobs from a big file will be around ~64KB but their 
sizes vary a lot. Also among the first chunks it seems typical to see one 
of 200KB+. What is the strategy to chunk up the file? I read somewhere that 
it's up to the client so I guess maybe the question is how cammount does 
it. With regard to deduplication I suppose the that strategy can be quite 
important, like if some header was changed in length but the rest of the 
file was the same, it could mean with ideal chunks only 1 chunk (the header 
chunk) was new and the rest unchanged, but with less ideal chunking all 
chunks would be new...

Sorry for the random ramble. I hope somewhere down the line I'll get more 
into both go and camlistore and be able to contribute and do something with 
my mentioned idea.









-- 
You received this message because you are subscribed to the Google Groups 
"Camlistore" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to