Hi Tim, There is no time bound SLA provided by AWS when a given binary would be successfully replicated to destination S3 bucket. There would be cases of missing binaries if mongo nodes sync faster than S3 replication. Also S3 replication works between a given pair of buckets. So one S3 bucket can replicate to a single S3 destination bucket.
I think we can implement a tiered S3Datastore which writes/reads to/from multiple S3 buckets. The tiered S3DS first tries to read from same-region bucket and if not found than fallback to cross-geo buckets. > Has this been tested already ? Generally, wdyt ? No. I suggest to first test cross geo mongo deployment with single S3 bucket. There shouldn't be functional issue in using single S3 bucket. Few customers use single shared S3 bucket between non-clustered cross-geo jackrabbit2 repositories in production. Thanks, -shashank -----Original Message----- From: maret.timot...@gmail.com [mailto:maret.timot...@gmail.com] On Behalf Of Timothée Maret Sent: Monday, June 29, 2015 4:05 PM To: oak-dev@jackrabbit.apache.org Subject: S3DataStore leverage Cross-Region Replication Hi, In a cross region setup using the S3 data store, it may make sense to leverage the Cross-Region auto replication of S3 buckets [0,1]. In order to avoid data replication issues it would make sense IMO to allow configuring the S3DataStore with two S3 buckets, one for writing and one for reading. The writing bucket would be shared among all instance (from all regions) while the reading bucket would be in each region (thus decreasing the latency). The writing bucket would auto replicate to the reading buckets. Has this been tested already ? Generally, wdyt ? Regards, Timothee [0] https://aws.amazon.com/blogs/aws/new-cross-region-replication-for-amazon-s3/ [1] https://docs.aws.amazon.com/AmazonS3/latest/dev/crr.html