Re: Error in S3 Garbage Collection

2020-11-03 Thread Tanvi Shah
Hi, I think this fetch size limit should be able to change through the jackrabbit api also. I think the task should be taken by jackrabbit team to make it configurable. Also I needed to understand is there some another provision through which S3 garbage collection could be initiated

Re: Error in S3 Garbage Collection

2020-11-02 Thread Julian Reschke
needed to understand is there some another provision through which S3 garbage collection could be initiated for such huge Database. I don't think there's a way to get it running before fixing the OOM in the scan phase first. Best regards, Julian

Re: Error in S3 Garbage Collection

2020-10-23 Thread Tanvi Shah
How is it possible with jackrabbit oak api version 1.22.2? From: Julian Reschke Sent: 22 October 2020 13:28 To: Tanvi Shah; oak-dev@jackrabbit.apache.org Subject: Re: Error in S3 Garbage Collection [External - Use Caution] Am 22.10.2020 um 07:48 schrieb

Re: Error in S3 Garbage Collection

2020-10-23 Thread Julian Reschke
Am 23.10.2020 um 11:51 schrieb Tanvi Shah: How is it possible with jackrabbit oak api version 1.22.2? ... You can "svn co https://svn.apache.org/repos/asf/jackrabbit/oak/tags/jackrabbit-oak-1.22.2/;, modify the code, then rebuild. (that said, I'd update to the latest, that is 1.22.5 or the

Re: Error in S3 Garbage Collection

2020-10-22 Thread Tanvi Shah
org Subject: Re: Error in S3 Garbage Collection [External - Use Caution] So, what *exactly* did you try? And was the outcome exactly the same as before? Best regards, Julian Am 21.10.2020 um 15:20 schrieb Tanvi Shah: > Hi Julian, > Tried by adding 'defaultRowFetchSize' but it is sti

Re: Error in S3 Garbage Collection

2020-10-22 Thread Julian Reschke
Am 22.10.2020 um 07:48 schrieb Tanvi Shah: Hi , So I tried using jdbc:postgresql://host:port/database?defaultRowFetchSize=10 And this didn't work and I again got Out Of Memory exception. When i debbuged I found that

Re: Error in S3 Garbage Collection

2020-10-21 Thread Julian Reschke
. From: Julian Reschke Sent: 21 October 2020 16:45 To: Tanvi Shah; oak-dev@jackrabbit.apache.org Subject: Re: Error in S3 Garbage Collection [External - Use Caution] Am 21.10.2020 um 12:41 schrieb Tanvi Shah: Hi, I couldn't find the option to set the limit for Query in S3 GC.(<ht

Re: Error in S3 Garbage Collection

2020-10-21 Thread Tanvi Shah
Hi Julian, Tried by adding 'defaultRowFetchSize' but it is still taking by default INT value. Thanks. From: Julian Reschke Sent: 21 October 2020 16:45 To: Tanvi Shah; oak-dev@jackrabbit.apache.org Subject: Re: Error in S3 Garbage Collection [External

Re: Error in S3 Garbage Collection

2020-10-21 Thread Tanvi Shah
t.AbstractIterator.hasNext(AbstractIterator.java:138) at org.apache.jackrabbit.oak.plugins.document.DocumentBlobReferenceRetriever.collectReferences(DocumentBlobReferenceRetriever.java:49) So I wanted to ask is there any other way how can we restrict this and successfully run S3 G

Re: Error in S3 Garbage Collection

2020-10-21 Thread Julian Reschke
Am 21.10.2020 um 12:41 schrieb Tanvi Shah: Hi, I couldn't find the option to set the limit for Query in S3 GC.(>). ​... Citing

Re: Error in S3 Garbage Collection

2020-10-19 Thread Julian Reschke
Am 19.10.2020 um 16:17 schrieb Julian Reschke: Am 19.10.2020 um 08:37 schrieb Tanvi Shah: Hi , We have started Se Garbage Collection and S3 size is more than 2 TB. We are facing memory issues while executing GC even when we have given 11 GB of memory to the application. Code is: final

Re: Error in S3 Garbage Collection

2020-10-19 Thread Julian Reschke
Am 19.10.2020 um 08:37 schrieb Tanvi Shah: Hi , We have started Se Garbage Collection and S3 size is more than 2 TB. We are facing memory issues while executing GC even when we have given 11 GB of memory to the application. Code is: final MarkSweepGarbageCollector gc =

Error in S3 Garbage Collection

2020-10-19 Thread Tanvi Shah
ractIterator.hasNext(AbstractIterator.java:138) at org.apache.jackrabbit.oak.plugins.document.DocumentBlobReferenceRetriever.collectReferences(DocumentBlobReferenceRetriever.java:49) So I wanted to ask is there any other way how can we restrict this and successfully run S3 Garbage

Re: S3 Garbage Collection

2020-07-31 Thread Julian Reschke
Am 31.07.2020 um 12:57 schrieb Tanvi Shah: Hi, We tried running S3 garbage collection on Live environment with oak 1.22.3 version. Here the data in S3 is more than 2.5TB and blob ids are more than 20 million. While running garbage collection we get following logs : Starting Blob garbage

RE: S3 Garbage Collection

2020-07-31 Thread Tanvi Shah
Hi, We tried running S3 garbage collection on Live environment with oak 1.22.3 version. Here the data in S3 is more than 2.5TB and blob ids are more than 20 million. While running garbage collection we get following logs : Starting Blob garbage collection with markOnly [false] No content length

Re: S3 Garbage Collection

2020-07-22 Thread Amit Jain
ah > Sent: 09 July 2020 16:18 > To: 'oak-dev@jackrabbit.apache.org' > Subject: S3 Garbage Collection > > Hi, > We have implemented the S3 garbage Collection with jackrabbit oak library > as > documentNodeStore.createBlobGarbageCollector( seconds, > reposit

RE: S3 Garbage Collection

2020-07-22 Thread Tanvi Shah
Any update on this? From: Tanvi Shah Sent: 09 July 2020 16:18 To: 'oak-dev@jackrabbit.apache.org' Subject: S3 Garbage Collection Hi, We have implemented the S3 garbage Collection with jackrabbit oak library as documentNodeStore.createBlobGarbageCollector( seconds

S3 Garbage Collection

2020-07-09 Thread Tanvi Shah
Hi, We have implemented the S3 garbage Collection with jackrabbit oak library as documentNodeStore.createBlobGarbageCollector( seconds, repository.toString(), wb, new DefaultStatisticsProvider(Executors.newScheduledThreadPool(1))). collectGarbage