Jason,

It's predecessor did, Lucandra. But Solandra is a new approach that manages 
shards of documents across the cluster for you and uses solrs distributed 
search to query indexes. 

Jake

On Mar 9, 2011, at 5:15 PM, Jason Rutherglen <jason.rutherg...@gmail.com> wrote:

> Doesn't Solandra partition by term instead of document?
> 
> On Wed, Mar 9, 2011 at 2:13 PM, Smiley, David W. <dsmi...@mitre.org> wrote:
>> I was just about to jump in this conversation to mention Solandra and go 
>> fig, Solandra's committer comes in. :-)   It was nice to meet you at Strata, 
>> Jake.
>> 
>> I haven't dug into the code yet but Solandra strikes me as a killer way to 
>> scale Solr. I'm looking forward to playing with it; particularly looking at 
>> disk requirements and performance measurements.
>> 
>> ~ David Smiley
>> 
>> On Mar 9, 2011, at 3:14 PM, Jake Luciani wrote:
>> 
>>> Hi Otis,
>>> 
>>> Have you considered using Solandra with Quorum writes
>>> to achieve master/master with CA semantics?
>>> 
>>> -Jake
>>> 
>>> 
>>> On Wed, Mar 9, 2011 at 2:48 PM, Otis Gospodnetic <otis_gospodne...@yahoo.com
>>>> wrote:
>>> 
>>>> Hi,
>>>> 
>>>> ---- Original Message ----
>>>> 
>>>>> From: Robert Petersen <rober...@buy.com>
>>>>> 
>>>>> Can't you skip the SAN and keep the indexes locally?  Then you  would
>>>>> have two redundant copies of the index and no lock issues.
>>>> 
>>>> I could, but then I'd have the issue of keeping them in sync, which seems
>>>> more
>>>> fragile.  I think SAN makes things simpler overall.
>>>> 
>>>>> Also, Can't master02 just be a slave to master01 (in the master farm  and
>>>>> separate from the slave farm) until such time as master01 fails?   Then
>>>> 
>>>> No, because it wouldn't be in sync.  It would always be N minutes behind,
>>>> and
>>>> when the primary master fails, the secondary would not have all the docs -
>>>> data
>>>> loss.
>>>> 
>>>>> master02 would start receiving the new documents with an  indexes
>>>>> complete up to the last replication at least and the other slaves  would
>>>>> be directed by LB to poll master02 also...
>>>> 
>>>> Yeah, "complete up to the last replication" is the problem.  It's a data
>>>> gap
>>>> that now needs to be filled somehow.
>>>> 
>>>> Otis
>>>> ----
>>>> Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch
>>>> Lucene ecosystem search :: http://search-lucene.com/
>>>> 
>>>> 
>>>>> -----Original  Message-----
>>>>> From: Otis Gospodnetic [mailto:otis_gospodne...@yahoo.com]
>>>>> Sent: Wednesday, March 09, 2011 9:47 AM
>>>>> To: solr-user@lucene.apache.org
>>>>> Subject:  Re: True master-master fail-over without data gaps (choosing CA
>>>>> in  CAP)
>>>>> 
>>>>> Hi,
>>>>> 
>>>>> 
>>>>> ----- Original Message ----
>>>>>> From: Walter  Underwood <wun...@wunderwood.org>
>>>>> 
>>>>>> On  Mar 9, 2011, at 9:02 AM, Otis Gospodnetic wrote:
>>>>>> 
>>>>>>> You mean  it's  not possible to have 2 masters that are in nearly
>>>>> real-time
>>>>>> sync?
>>>>>>> How  about with DRBD?  I know people use  DRBD to keep 2 Hadoop NNs
>>>>> (their
>>>>>> edit
>>>>>> 
>>>>>>> logs) in  sync to avoid the current NN SPOF, for example, so I'm
>>>>> thinking
>>>>>> this
>>>>>> 
>>>>>>> could be doable with Solr masters, too, no?
>>>>>> 
>>>>>> If you add fault-tolerant, you run into the CAP  Theorem.  Consistency,
>>>>> 
>>>>>> availability, partition: choose two. You cannot have  it  all.
>>>>> 
>>>>> Right, so I'll take Consistency and Availability, and I'll  put my 2
>>>>> masters in
>>>>> the same rack (which has redundant switches, power  supply, etc.) and
>>>>> thus
>>>>> minimize/avoid partitioning.
>>>>> Assuming the above  actually works, I think my Q remains:
>>>>> 
>>>>> How do you set up 2 Solr masters so  they are in near real-time sync?
>>>>> DRBD?
>>>>> 
>>>>> But here is maybe a simpler  scenario that more people may be
>>>>> considering:
>>>>> 
>>>>> Imagine 2 masters on 2  different servers in 1 rack, pointing to the same
>>>>> index
>>>>> on the shared  storage (SAN) that also happens to live in the same rack.
>>>>> 2 Solr masters are  behind 1 LB VIP that indexer talks to.
>>>>> The VIP is configured so that all  requests always get routed to the
>>>>> primary
>>>>> master (because only 1 master  can be modifying an index at a time),
>>>>> except when
>>>>> this primary is down,  in which case the requests are sent to the
>>>>> secondary
>>>>> master.
>>>>> 
>>>>> So in  this case my Q is around automation of this, around Lucene index
>>>>> locks,
>>>>> around the need for manual intervention, and such.
>>>>> Concretely, if you  have these 2 master instances, the primary master has
>>>>> the
>>>>> Lucene index  lock in the index dir.  When the secondary master needs to
>>>>> take
>>>>> over  (i.e., when it starts receiving documents via LB), it needs to be
>>>>> able to
>>>>> write to that same index.  But what if that lock is still around?   One
>>>>> could use
>>>>> the Native lock to make the lock disappear if the primary  master's JVM
>>>>> exited
>>>>> unexpectedly, and in that case everything *should*  work and be
>>>>> completely
>>>>> transparent, right?  That is, the secondary  will start getting new docs,
>>>>> it will
>>>>> use its IndexWriter to write to that  same shared index, which won't be
>>>>> locked
>>>>> for writes because the lock is  gone, and everyone will be happy.  Did I
>>>>> miss
>>>>> something important  here?
>>>>> 
>>>>> Assuming the above is correct, what if the lock is *not* gone  because
>>>>> the
>>>>> primary master's JVM is actually not dead, although maybe  unresponsive,
>>>>> so LB
>>>>> thinks the primary master is dead.  Then the LB  will route indexing
>>>>> requests to
>>>>> the secondary master, which will attempt  to write to the index, but be
>>>>> denied
>>>>> because of the lock.  So a  human needs to jump in, remove the lock, and
>>>>> manually
>>>>> reindex failed docs  if the upstream component doesn't buffer docs that
>>>>> failed to
>>>>> get indexed  and doesn't retry indexing them automatically.  Is this
>>>>> correct or
>>>>> is there a way to avoid humans  here?
>>>>> 
>>>>> Thanks,
>>>>> Otis
>>>>> ----
>>>>> Sematext :: http://sematext.com/ :: Solr -  Lucene - Nutch
>>>>> Lucene ecosystem search :: http://search-lucene.com/
>>>>> 
>>>> 
>>> 
>>> 
>>> 
>>> --
>>> http://twitter.com/tjake
>> 
>> 

Reply via email to