I forgot the last part... To know how to use the datastore: http://wiki.apache.org/jackrabbit/DataStore
Greetings! 2012/6/24 Francisco Carriedo Scher <[email protected]> > Don't worry, Jackrabbit can deal with such files properly. Just use the > datastore (saves the files in a path you specify in the file system, more > efficient by far than any RDMS handling large blobs) and you are done! > > About getting resources out of the repository there is no problem neither, > the file will be streamed (no serious memory issues). > > Hope that helps! > > Regards. > > > 2012/6/9 Mark Herman <[email protected]> > >> >> Di Simone Maria Nicole wrote >> > >> > Hi everybody, >> > is there a limit or any best pratice for node dimensions? >> > In my project someone would like to store very big documents (1GB) but >> > I don't agree with this idea. >> > Is there any raccomandation about this topic? >> > >> >> What persistence manager do you plan on using? The default uses whatever >> file system you're on, which can probably handle 1gb files without an >> issue. >> I wouldn't be surprised if some sql server implementations don't react >> well >> to being loaded with a bunch of 1gb files. >> >> >> Do you expect lucene to be indexing the content of this file? In my >> experience, the indexer is fairly non-invasive so I wouldn't expect it to >> hurt the server. I've never thrown something at it that big though. >> >> Either way you should watch the jvm memory, not sure how much jr needs to >> store in memory or if it can stream straight to the persistence manager. >> >> -- >> View this message in context: >> http://jackrabbit.510166.n4.nabble.com/Question-Is-there-any-limit-on-node-dimension-tp4655240p4655338.html >> Sent from the Jackrabbit - Users mailing list archive at Nabble.com. >> > >
