this is part of the "different replica types" capability, there are NRT (the only type available prior to 7x), PULL and TLOG which would have different names. I don't know of any way to switch it off.
As far as moving the data, here's a little known trick: Use the replication API to issue a fetchindexk, see: https://lucene.apache.org/solr/guide/6_6/index-replication.html As long as the target cluster can "see" the source cluster via http, this should work. This is entirely outside SolrCloud and ZooKeeper is not involved. This would even work with, say, one side being stand-alone and the other being SolrCloud (not that you want to do that, just illustrating it's not part of SolrCloud)... So you'd specify something like: http://target_node:port/solr/core_name/replication?command=fetchindex&masterUrl=http://source_node:port/solr/core_name "core_name" in these cases is what appears in the "cores" dropdown on the admin UI page. You do not have to shut Solr down at all on either end to use this, although last I knew the target node would not serve queries while this was happening. An alternative is to not hard-code the names in your copy script, rather look at the information in ZooKeeper for your source and target information, you could do this by using the CLUSTERSTATUS collections API call. Best, Erick On Tue, Mar 6, 2018 at 6:47 AM, Patrick Schemitz <p...@solute.de> wrote: > Hi List, > > so I'm running a bunch of SolrCloud clusters (each cluster is: 8 shards > on 2 servers, with 4 instances per server, no replicas, i.e. 1 shard per > instance). > > Building the index afresh takes 15+ hours, so when I have to deploy a new > index, I build it once, on one cluster, and then copy (scp) over the > data/<main_index>/index directories (shutting down the Solr instances first). > > I could get Solr 6.5.1 to number the shard/replica directories nicely via > the createNodeSet and createNodeSet.shuffle options: > > Solr 6.5.1 /var/lib/solr: > > Server node 1: > instance00/data/main_index_shard1_replica1 > instance01/data/main_index_shard2_replica1 > instance02/data/main_index_shard3_replica1 > instance03/data/main_index_shard4_replica1 > > Server node 2: > instance00/data/main_index_shard5_replica1 > instance01/data/main_index_shard6_replica1 > instance02/data/main_index_shard7_replica1 > instance03/data/main_index_shard8_replica1 > > However, while attempting to upgrade to 7.2.1, this numbering has changed: > > Solr 7.2.1 /var/lib/solr: > > Server node 1: > instance00/data/main_index_shard1_replica_n1 > instance01/data/main_index_shard2_replica_n2 > instance02/data/main_index_shard3_replica_n4 > instance03/data/main_index_shard4_replica_n6 > > Server node 2: > instance00/data/main_index_shard5_replica_n8 > instance01/data/main_index_shard6_replica_n10 > instance02/data/main_index_shard7_replica_n12 > instance03/data/main_index_shard8_replica_n14 > > This new numbering breaks my copy script, and furthermode, I'm worried > as to what happens when the numbering is different among target clusters. > > How can I switch this back to the old numbering scheme? > > Side note: is there a recommended way of doing this? Is the > backup/restore mechanism suitable for this? The ref guide is kind of terse > here. > > Thanks in advance, > > Ciao, Patrick