Hmmm...all of that looks pretty normal... Did a commit somehow fail on the other machine? When you view the stats for the update handler, are there a lot of pending adds for on of the nodes? Do the commit counts match across nodes?
You can also query an individual node with distrib=false to check that. If you build is a month old, I'd honestly recommend you try upgrading as well. - Mark On Feb 27, 2012, at 3:34 PM, Matthew Parker wrote: > Here is most of the cluster state: > > Connected to Zookeeper > localhost:2181, localhost: 2182, localhost:2183 > > /(v=0 children=7) "" > /CONFIGS(v=0, children=1) > /CONFIGURATION(v=0 children=25) > <<<<< all the configuration files, velocity info, xslt, etc. >>>>> > /NODE_STATES(v=0 children=4) > MACHINE1:8083_SOLR (v=121)"[{"shard_id":"shard1", > "state":"active","core":"","collection":"collection1","node_name:"..." > MACHINE1:8082_SOLR (v=101)"[{"shard_id":"shard2", > "state":"active","core":"","collection":"collection1","node_name:"..." > MACHINE1:8081_SOLR (v=92)"[{"shard_id":"shard1", > "state":"active","core":"","collection":"collection1","node_name:"..." > MACHINE1:8084_SOLR (v=73)"[{"shard_id":"shard2", > "state":"active","core":"","collection":"collection1","node_name:"..." > /ZOOKEEPER (v-0 children=1) > QUOTA(v=0) > > /CLUSTERSTATE.JSON(V=272)"{"collection1":{"shard1":{MACHINE1:8081_solr_":{shard_id":"shard1","leader":"true","..." > /LIVE_NODES (v=0 children=4) > MACHINE1:8083_SOLR(ephemeral v=0) > MACHINE1:8082_SOLR(ephemeral v=0) > MACHINE1:8081_SOLR(ephemeral v=0) > MACHINE1:8084_SOLR(ephemeral v=0) > /COLLECTIONS (v=1 children=1) > COLLECTION1(v=0 children=2)"{"configName":"configuration1"}" > LEADER_ELECT(v=0 children=2) > SHARD1(V=0 children=1) > ELECTION(v=0 children=2) > > 87186203314552835-MACHINE1:8081_SOLR_-N_0000000096(ephemeral v=0) > > 87186203314552836-MACHINE1:8083_SOLR_-N_0000000084(ephemeral v=0) > SHARD2(v=0 children=1) > ELECTION(v=0 children=2) > > 231301391392833539-MACHINE1:8084_SOLR_-N_0000000085(ephemeral v=0) > > 159243797356740611-MACHINE1:8082_SOLR_-N_0000000084(ephemeral v=0) > LEADERS (v=0 children=2) > SHARD1 (ephemeral > v=0)"{"core":"","node_name":"MACHINE1:8081_solr","base_url":" > http://MACHINE1:8081/solr"}" > SHARD2 (ephemeral > v=0)"{"core":"","node_name":"MACHINE1:8082_solr","base_url":" > http://MACHINE1:8082/solr"}" > /OVERSEER_ELECT (v=0 children=2) > ELECTION (v=0 children=4) > 231301391392833539-MACHINE1:8084_SOLR_-N_0000000251(ephemeral v=0) > 87186203314552835-MACHINE1:8081_SOLR_-N_0000000248(ephemeral v=0) > 159243797356740611-MACHINE1:8082_SOLR_-N_0000000250(ephemeral v=0) > 87186203314552836-MACHINE1:8083_SOLR_-N_0000000249(ephemeral v=0) > LEADER (emphemeral > v=0)"{"id":"87186203314552835-MACHINE1:8081_solr-n_000000248"}" > > > > On Mon, Feb 27, 2012 at 2:47 PM, Mark Miller <markrmil...@gmail.com> wrote: > >> >> On Feb 27, 2012, at 2:22 PM, Matthew Parker wrote: >> >>> Thanks for your reply Mark. >>> >>> I believe the build was towards the begining of the month. The >>> solr.spec.version is 4.0.0.2012.01.10.38.09 >>> >>> I cannot access the clusterstate.json contents. I clicked on it a couple >> of >>> times, but nothing happens. Is that stored on disk somewhere? >> >> Are you using the new admin UI? That has recently been updated to work >> better with cloud - it had some troubles not too long ago. If you are, you >> should trying using the old admin UI's zookeeper page - that should show >> the cluster state. >> >> That being said, there has been a lot of bug fixes over the past month - >> so you may just want to update to a recent version. >> >>> >>> I configured a custom request handler to calculate an unique document id >>> based on the file's url. >>> >>> On Mon, Feb 27, 2012 at 1:13 PM, Mark Miller <markrmil...@gmail.com> >> wrote: >>> >>>> Hey Matt - is your build recent? >>>> >>>> Can you visit the cloud/zookeeper page in the admin and send the >> contents >>>> of the clusterstate.json node? >>>> >>>> Are you using a custom index chain or anything out of the ordinary? >>>> >>>> >>>> - Mark >>>> >>>> On Feb 27, 2012, at 12:26 PM, Matthew Parker wrote: >>>> >>>>> TWIMC: >>>>> >>>>> Environment >>>>> ========= >>>>> Apache SOLR rev-1236154 >>>>> Apache Zookeeper 3.3.4 >>>>> Windows 7 >>>>> JDK 1.6.0_23.b05 >>>>> >>>>> I have built a SOLR Cloud instance with 4 nodes using the embeded Jetty >>>>> servers. >>>>> >>>>> I created a 3 node zookeeper ensemble to manage the solr configuration >>>> data. >>>>> >>>>> All the instances run on one server so I've had to move ports around >> for >>>>> the various applications. >>>>> >>>>> I start the 3 zookeeper nodes. >>>>> >>>>> I started the first instance of solr cloud with the parameter to have >> two >>>>> shards. >>>>> >>>>> The start the remaining 3 solr nodes. >>>>> >>>>> The system comes up fine. No errors thrown. >>>>> >>>>> I can view the solr cloud console and I can see the SOLR configuration >>>>> files managed by ZooKeeper. >>>>> >>>>> I published data into the SOLR Cloud instances from SharePoint using >>>> Apache >>>>> Manifold 0.4-incubating. Manifold is setup to publish the data into >>>>> collection1, which is the only collection defined in the cluster. >>>>> >>>>> When I query the data from collection1 as per the solr wiki, the >> results >>>>> are inconsistent. Sometimes all the results are there, other times >>>> nothing >>>>> comes back at all. >>>>> >>>>> It seems to be having an issue auto replicating the data across the >>>> cloud. >>>>> >>>>> Is there some specific setting I might have missed? Based upon what I >>>> read, >>>>> I thought that SOLR cloud would take care of distributing and >> replicating >>>>> the data automatically. Do you have to tell it what shard to publish >> the >>>>> data into as well? >>>>> >>>>> Any help would be appreciated. >>>>> >>>>> Thanks, >>>>> >>>>> Matt >>>>> >>>>> ------------------------------ >>>>> This e-mail and any files transmitted with it may be proprietary. >>>> Please note that any views or opinions presented in this e-mail are >> solely >>>> those of the author and do not necessarily represent those of Apogee >>>> Integration. >>>> >>>> - Mark Miller >>>> lucidimagination.com >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>> >>> Matt >>> >>> ------------------------------ >>> This e-mail and any files transmitted with it may be proprietary. >> Please note that any views or opinions presented in this e-mail are solely >> those of the author and do not necessarily represent those of Apogee >> Integration. >> >> - Mark Miller >> lucidimagination.com >> >> >> >> >> >> >> >> >> >> >> >> > > ------------------------------ > This e-mail and any files transmitted with it may be proprietary. Please > note that any views or opinions presented in this e-mail are solely those of > the author and do not necessarily represent those of Apogee Integration. - Mark Miller lucidimagination.com