I don't understand, why are you caching some pieces of data on both servers?
And if it's not a problem for you to write a piece of data to two servers at once, why is it a problem to do deletion in the exact same way? You will have synchronization issues with both writing and deleting if you do it this way though, so why do it all? Why not use memcached the way it's supposed to be used, bunch the servers you have into a single cluster, and if you need more cache capacity, add servers to the cluster? /Henrik On Thu, Oct 11, 2012 at 5:12 PM, Kiran Kumar <[email protected]> wrote: > > > On Thursday, 11 October 2012 20:28:06 UTC+5:30, Kiran Kumar wrote: >> >> I am working on a heavy traffic web site , where there will be GB's of >> data written per minute into our Memcache . So we have decided to use two >> separate instances of Memcache for the application . >> > > > > >> >> Memcache1 Memcache2 >> \ / >> \ / >> \ / >> \ / >> \ / >> \ / >> CustomerData >> >> So right now as per the set up , there are two Memcache instances for a >> single application . >> >> Now my question is , once we recive a value inside the application , >> which writes/sets to both the Memcache instances , assume that if a key is >> read one of the instance of Memcache - 1 , i need to delete the same key on >> the other instance of memcahce also at the same time , so taht they will be >> in sync with each other . >> >> As per the code point of view once a value is read from Memcache , i am >> deleting that key . >> > > So conclusion is , it writes to both the instances and read reads from > any one . >
