Re: Unable to perform search query after changing uniqueKey

2015-03-30 Thread Mostafa Gomaa
Hi Zheng,

It's possible that there's a problem with your schema.xml. Are all fields
defined and have appropriate options enabled?

Regards,

Mostafa.

On Mon, Mar 30, 2015 at 7:49 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com
wrote:

 Hi Erick,

 I've tried that, and removed the data directory from both the shards. But
 the same problem still occurs, so we probably can rule out the memory
 issue.

 Regards,
 Edwin

 On 30 March 2015 at 12:39, Erick Erickson erickerick...@gmail.com wrote:

  I meant shut down Solr and physically remove the entire data
  directory. Not saying this is the cure, but it can't hurt to rule out
  the index having memory...
 
  Best,
  Erick
 
  On Sun, Mar 29, 2015 at 6:35 PM, Zheng Lin Edwin Yeo
  edwinye...@gmail.com wrote:
   Hi Erick,
  
   I used the following query to delete all the index.
  
   http://localhost:8983/solr/update?stream.body=
  deletequery*:*/query/delete
  http://localhost:8983/solr/update?stream.body=commit/
  
  
   Or is it better to physically delete the entire data directory?
  
  
   Regards,
   Edwin
  
  
   On 28 March 2015 at 02:27, Erick Erickson erickerick...@gmail.com
  wrote:
  
   You say you re-indexed, did you _completely_ remove the data directory
   first, i.e. the parent of the index and, maybe, tlog directories?
   I've occasionally seen remnants of old definitions pollute the new
   one, and since the uniqueKey key is so fundamental I can see it
   being a problem.
  
   Best,
   Erick
  
   On Fri, Mar 27, 2015 at 1:42 AM, Andrea Gazzarini 
  a.gazzar...@gmail.com
   wrote:
Hi Edwin,
please provide some other detail about your context, (e.g. complete
stacktrace, query you're issuing)
   
Best,
Andrea
   
   
On 03/27/2015 09:38 AM, Zheng Lin Edwin Yeo wrote:
   
Hi everyone,
   
I've changed my uniqueKey to another name, instead of using id, on
  the
schema.xml.
   
However, after I have done the indexing (the indexing is
 successful),
   I'm
not able to perform a search query on it. I gives the error
java.lang.NullPointerException.
   
Is there other place which I need to configure, besides changing
 the
uniqueKey field in scheam.xml?
   
Regards,
Edwin
   
   
  
 



Solr went on recovery multiple time.

2015-03-30 Thread sthita
Solr server crashes 2-3 times a day .  When i looked into log i can see some
recovery failure. 

I got these errors on my solr log

WARN  - 2015-03-30 16:03:51.137; org.apache.solr.cloud.RecoveryStrategy;
Stopping recovery for zkNodeName=searcher:8980_solr_rn0core=rn0
ERROR - 2015-03-30 16:03:56.121; org.apache.solr.update.UpdateLog; Error
inspecting tlog tlog{file=/solr/rn0/data/tlog/tlog.709
refcount=2}
ERROR - 2015-03-30 16:03:56.121; org.apache.solr.update.UpdateLog; Error
inspecting tlog tlog{file=/solr/rn0/data/tlog/tlog.710
refcount=2}
WARN  - 2015-03-30 16:04:00.242; org.apache.solr.core.SolrCore; [rn0]
PERFORMANCE WARNING: Overlapping onDeckSearchers=2

I googled so many places, suggestions are related to my commit . 

My commit configuration on solrconfig.xml : 


updateHandler class=solr.DirectUpdateHandler2

autoCommit
maxTime60/maxTime
openSearcherfalse/openSearcher
/autoCommit 

updateLog
str name=dir${solr.data.dir:}/str
/updateLog
/updateHandler


Why is it happening ? Any idea/solution will be appreciated. 





--
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr-went-on-recovery-multiple-time-tp4196249.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Restart solr failed after applied the patch in https://issues.apache.org/jira/browse/SOLR-6359

2015-03-30 Thread Ramkumar R. Aiyengar
I doubt this has anything to do with the patch. Do you observe the same
behaviour if you reduce the values for the config to defaults? (100, 10)
On 30 Mar 2015 09:51, forest_soup tanglin0...@gmail.com wrote:

 https://issues.apache.org/jira/browse/SOLR-6359

 I also posted the questions to the JIRA ticket.

 We have a SolrCloud with 5 solr servers of Solr 4.7.0. There are one
 collection with 80 shards(2 replicas per shard) on those 5 servers. And we
 made a patch by merge the patch
 (https://issues.apache.org/jira/secure/attachment/12702473/SOLR-6359.patch
 )
 to 4.7.0 stream. And after applied the patch to our servers with the config
 changing uploaded to ZooKeeper, we did a restart on one of the 5 solr
 server, we met some issues on that server. Below is the details -
 The solrconfig.xml we changed:
 updateLog
 str name=dir$
 {solr.ulog.dir:}
 /str
 int name=numRecordsToKeep1/int
 int name=maxNumLogsToKeep100/int
 /updateLog

 After we restarted one solr server without other 4 servers are running, we
 met below exceptions in the restarted one:
 ERROR - 2015-03-16 20:48:48.214; org.apache.solr.common.SolrException;
 org.apache.solr.common.SolrException: Exception writing document id
 Q049bGx0bWFpbDIxL089bGxwX3VzMQ==41703656!B68BF5EC5A4A650D85257E0A00724A3B
 to
 the index; possible analysis error.
 at

 org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:164)
 at

 org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:69)
 at

 org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:51)
 at

 org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:703)
 at

 org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:857)
 at

 org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:556)
 at

 org.apache.solr.handler.loader.JavabinLoader$1.update(JavabinLoader.java:96)
 at

 org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readOuterMostDocIterator(JavaBinUpdateRequestCodec.java:166)
 at

 org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readIterator(JavaBinUpdateRequestCodec.java:136)
 at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:225)
 at

 org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readNamedList(JavaBinUpdateRequestCodec.java:121)
 at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:190)
 at
 org.apache.solr.common.util.JavaBinCodec.unmarshal(JavaBinCodec.java:116)
 at

 org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec.unmarshal(JavaBinUpdateRequestCodec.java:173)
 at

 org.apache.solr.handler.loader.JavabinLoader.parseAndLoadDocs(JavabinLoader.java:106)
 at org.apache.solr.handler.loader.JavabinLoader.load(JavabinLoader.java:58)
 at

 org.apache.solr.handler.UpdateRequestHandler$1.load(UpdateRequestHandler.java:92)
 at

 org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:74)
 at

 org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
 at org.apache.solr.core.SolrCore.execute(SolrCore.java:1916)
 at

 org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:780)
 at

 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:427)
 at

 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:217)
 at

 org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
 at

 org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
 at

 org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:220)
 at

 org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:122)
 at

 org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:171)
 at

 org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
 at

 org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:116)
 at
 org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:408)
 at

 org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1040)
 at

 org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:607)
 at

 org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:314)
 at

 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1156)
 at

 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:626)
 at

 org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61)
 at java.lang.Thread.run(Thread.java:804)
 Caused by: org.apache.lucene.store.AlreadyClosedException: this IndexWriter
 is closed
 at 

Re: Installing the auto-phrase-tokenfilter

2015-03-30 Thread luismart
Many thanks Andrea, I followed your guidance:

the following site helped me to install and configure Ant in Windows
http://madhukaudantha.blogspot.com.es/2010/06/installing-ant-for-windows-7.html 

Then after cloning the repository
https://github.com/LucidWorks/auto-phrase-tokenfilter I just run the ant
command and got the AutoPhraseTokenFilter.jar file needed.

I will now try to make it work in Solr.







--
View this message in context: 
http://lucene.472066.n3.nabble.com/Installing-the-auto-phrase-tokenfilter-tp4195466p4196241.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Same schema.xml is loaded for different cores in SolrCloud

2015-03-30 Thread Zheng Lin Edwin Yeo
I've roughly know what is the problem from here.
http://stackoverflow.com/questions/23338324/zookeeper-multiple-collection-different-schema

However, I couldn't find the zoo_data directory in all of my solr folder.
What could be the problem or where is the directory supposed to be located?

Regards,
Edwin


On 30 March 2015 at 11:56, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote:

 Hi everyone,

 I've created a SolrCloud with multiple core, and I have different
 schema.xml for each of the core. However, when I start Solr, there's only
 one version of the schema.xml that is loaded onto Solr. Regardless of which
 core I go to, the schema.xml that is shown is the first one which I have
 loaded.

 What I did was, I have 3 cores: logmill, collection1 and collection2.
 Each of the core has 2 shrads: shard1 and shard2

 I first started the Solr with shard1 using the following command:
 java -Dcollection.configName=logmill -DzkRun -DnumShards=2
 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar

 After that I start shard2 using the following command:
 java -Dcollection.configName=logmill -DzkRun -DnumShards=2
 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar

 All the schema.xml loaded are from logmill core, even for the collection1
 and collection2.

 Even after I change the command to start shard1 with the following
 command, all the schema.xml are still from logmill
 java -Dcollection.configName=collection1 -DzkRun
 -DnumShards=2 -Dbootstrap_confdir=./solr/collection1/conf -jar start.jar


 How do I get Solr to read the different schema.xml for the different cores?

 Regards,
 Edwin



Re: Unable to perform search query after changing uniqueKey

2015-03-30 Thread Zheng Lin Edwin Yeo
Hi Mostafa,

Yes, I've defined all the fields in schema.xml. It is able to work on the
version without SolrCloud, but it is not working for the one with SolrCloud.
Both of them are using the same schema.xml.

Regards,
Edwin



On 30 March 2015 at 14:34, Mostafa Gomaa mostafa.goma...@gmail.com wrote:

 Hi Zheng,

 It's possible that there's a problem with your schema.xml. Are all fields
 defined and have appropriate options enabled?

 Regards,

 Mostafa.

 On Mon, Mar 30, 2015 at 7:49 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com
 
 wrote:

  Hi Erick,
 
  I've tried that, and removed the data directory from both the shards. But
  the same problem still occurs, so we probably can rule out the memory
  issue.
 
  Regards,
  Edwin
 
  On 30 March 2015 at 12:39, Erick Erickson erickerick...@gmail.com
 wrote:
 
   I meant shut down Solr and physically remove the entire data
   directory. Not saying this is the cure, but it can't hurt to rule out
   the index having memory...
  
   Best,
   Erick
  
   On Sun, Mar 29, 2015 at 6:35 PM, Zheng Lin Edwin Yeo
   edwinye...@gmail.com wrote:
Hi Erick,
   
I used the following query to delete all the index.
   
http://localhost:8983/solr/update?stream.body=
   deletequery*:*/query/delete
   http://localhost:8983/solr/update?stream.body=commit/
   
   
Or is it better to physically delete the entire data directory?
   
   
Regards,
Edwin
   
   
On 28 March 2015 at 02:27, Erick Erickson erickerick...@gmail.com
   wrote:
   
You say you re-indexed, did you _completely_ remove the data
 directory
first, i.e. the parent of the index and, maybe, tlog
 directories?
I've occasionally seen remnants of old definitions pollute the new
one, and since the uniqueKey key is so fundamental I can see it
being a problem.
   
Best,
Erick
   
On Fri, Mar 27, 2015 at 1:42 AM, Andrea Gazzarini 
   a.gazzar...@gmail.com
wrote:
 Hi Edwin,
 please provide some other detail about your context, (e.g.
 complete
 stacktrace, query you're issuing)

 Best,
 Andrea


 On 03/27/2015 09:38 AM, Zheng Lin Edwin Yeo wrote:

 Hi everyone,

 I've changed my uniqueKey to another name, instead of using id,
 on
   the
 schema.xml.

 However, after I have done the indexing (the indexing is
  successful),
I'm
 not able to perform a search query on it. I gives the error
 java.lang.NullPointerException.

 Is there other place which I need to configure, besides changing
  the
 uniqueKey field in scheam.xml?

 Regards,
 Edwin


   
  
 



Restart solr failed after applied the patch in https://issues.apache.org/jira/browse/SOLR-6359

2015-03-30 Thread forest_soup
https://issues.apache.org/jira/browse/SOLR-6359

I also posted the questions to the JIRA ticket.

We have a SolrCloud with 5 solr servers of Solr 4.7.0. There are one
collection with 80 shards(2 replicas per shard) on those 5 servers. And we
made a patch by merge the patch
(https://issues.apache.org/jira/secure/attachment/12702473/SOLR-6359.patch)
to 4.7.0 stream. And after applied the patch to our servers with the config
changing uploaded to ZooKeeper, we did a restart on one of the 5 solr
server, we met some issues on that server. Below is the details - 
The solrconfig.xml we changed:
updateLog
str name=dir$
{solr.ulog.dir:}
/str
int name=numRecordsToKeep1/int
int name=maxNumLogsToKeep100/int
/updateLog

After we restarted one solr server without other 4 servers are running, we
met below exceptions in the restarted one:
ERROR - 2015-03-16 20:48:48.214; org.apache.solr.common.SolrException;
org.apache.solr.common.SolrException: Exception writing document id
Q049bGx0bWFpbDIxL089bGxwX3VzMQ==41703656!B68BF5EC5A4A650D85257E0A00724A3B to
the index; possible analysis error.
at
org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:164)
at
org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:69)
at
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:51)
at
org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:703)
at
org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:857)
at
org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:556)
at
org.apache.solr.handler.loader.JavabinLoader$1.update(JavabinLoader.java:96)
at
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readOuterMostDocIterator(JavaBinUpdateRequestCodec.java:166)
at
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readIterator(JavaBinUpdateRequestCodec.java:136)
at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:225)
at
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readNamedList(JavaBinUpdateRequestCodec.java:121)
at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:190)
at org.apache.solr.common.util.JavaBinCodec.unmarshal(JavaBinCodec.java:116)
at
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec.unmarshal(JavaBinUpdateRequestCodec.java:173)
at
org.apache.solr.handler.loader.JavabinLoader.parseAndLoadDocs(JavabinLoader.java:106)
at org.apache.solr.handler.loader.JavabinLoader.load(JavabinLoader.java:58)
at
org.apache.solr.handler.UpdateRequestHandler$1.load(UpdateRequestHandler.java:92)
at
org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:74)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1916)
at
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:780)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:427)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:217)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:220)
at
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:122)
at
org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:171)
at
org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
at
org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:116)
at
org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:408)
at
org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1040)
at
org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:607)
at
org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:314)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1156)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:626)
at
org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61)
at java.lang.Thread.run(Thread.java:804)
Caused by: org.apache.lucene.store.AlreadyClosedException: this IndexWriter
is closed
at org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:645)
at org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:659)
at org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:1525)
at
org.apache.solr.update.DirectUpdateHandler2.addDoc0(DirectUpdateHandler2.java:236)
at
org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:160)
... 37 

Date Time datatypes?

2015-03-30 Thread Bram Van Dam

Howdy folks,

Is there any way index only the date and time portions of a datetime field?

A Date is really a period of 24hrs, starting at 00:00 in said date's 
time zone. It would be useful if there was a way to search for documents 
of a certain date with these semantics.


As for times, I'd like to be able to do queries like time:[17:00 TO 
18:00]. I suppose I could accomplish that by resetting the date portion 
to some bogus value, but then my facet/range values will contain that 
bogus date.


I suppose the alternative is to create my own data types. Extending 
PrimitiveFieldType doesn't seem too hairy but maybe I'm overlooking some 
of the complexity...


Thanks a bunch,

 - Bram


Re: Optimize SolrCloud without downtime

2015-03-30 Thread Pavel Hladik
Hi,

thanks for reply. We have a lot of deleted docs cause we have to reindex all
records from time to time, changing some important parameters.. When we do
update, it means create and delete.

Our deleted docs do not disappear by merging segments. I see our deleted
docs are almost the same number after two weeks and always around 40
segments. I ask our developers and they confirmed we use default settings
for merging policy. Please, do you have some tips and tricks for efficient
setting for merging segments?

Our HW is RAID5 SAS disks and 40GB of memory dedicated for Solr.

Regards,
Pavel





--
View this message in context: 
http://lucene.472066.n3.nabble.com/Optimize-SolrCloud-without-downtime-tp4195170p4196273.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Restart solr failed after applied the patch in https://issues.apache.org/jira/browse/SOLR-6359

2015-03-30 Thread forest_soup
Yes, I also doubt the patch. I restore the patch with original .jar file,
there is no that issue.



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Restart-solr-failed-after-applied-the-patch-in-https-issues-apache-org-jira-browse-SOLR-6359-tp4196251p4196278.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Korean script conversion

2015-03-30 Thread Benson Margulies
Why do you think that this is a good idea? Hanja are used for special
purposes; they are not trivally convertable to Hanjul due to ambiguity, and
it's not at all clear that a typical search user wants to treat them as
equivalent.

On Sun, Mar 29, 2015 at 1:52 AM, Eyal Naamati 
eyal.naam...@exlibrisgroup.com wrote:

  Hi,



 We are starting to index records in Korean. Korean text can be written in
 two scripts: Han characters (Chinese) and Hangul characters (Korean).

 We are looking for some solr filter or another built in solr component
 that converts between Han and Hangul characters (transliteration).

 I know there is the ICUTransformFilterFactory that can convert between
 Japanese or chinese scripts, for example:

 filter class=*solr.ICUTransformFilterFactory* id=*Katakana- Hiragana*
 / for Japanese script conversions

 So far I couldn't find anything readymade for Korean scripts, but perhaps
 someone knows of one?



 Thanks!

 Eyal Naamati
 Alma Developer
 Tel: +972-2-6499313
 Mobile: +972-547915255
 eyal.naam...@exlibrisgroup.com
 [image: Description: Description: Description: Description:
 C://signature/exlibris.jpg]
 www.exlibrisgroup.com





Re: Restart solr failed after applied the patch in https://issues.apache.org/jira/browse/SOLR-6359

2015-03-30 Thread forest_soup
But if the value can only be 100,10, is there any difference with no that
patch? Can we enlarge those 2 values? Thanks!



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Restart-solr-failed-after-applied-the-patch-in-https-issues-apache-org-jira-browse-SOLR-6359-tp4196251p4196280.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: SOLR Index in shared/Network folder

2015-03-30 Thread Erick Erickson
First examine whether you can reduce the amount of data you keep
around, field norms, stored fields, etc. Here's a place to start:
http://stackoverflow.com/questions/10080881/solr-index-size-reduction

I have heard of people doing what you suggest, but be _very_ careful
that you don't accidentally write to the index from the slave.

I always have to ask. You're talking about reducing disk space because
you have 100's of G indexes. Is it really worth the engineering
effort to save the disk space? Terabyte hard drives are  $100. Buy
two for the cost of untangling a single corrupt index because someone
(and someone will) caused the slaves to write to the index rather than
the master.

FWIW,
Erick

On Sun, Mar 29, 2015 at 10:42 PM, abhi Abhishek abhi26...@gmail.com wrote:
 Hello,
  Thanks for the suggestions. My aim is to reduce the disk space usage.
 I have 1 master with 2 slave configured, where slaves are used for
 searching and master ingests new data replicated to slaves, but as my index
 size is in 100's of GB we see 3x times space overhead. i would like to
 reduce this overhead, can you suggest something for this?

 Thanks in Advance

 Best Regards,
 Abhishek

 On Sat, Mar 28, 2015 at 12:13 AM, Erick Erickson erickerick...@gmail.com
 wrote:

 To pile on: If you're talking about pointing two Solr instances at the
 _same_ index, it doesn't matter whether you are on NFS or not, you'll
 have all sorts of problems. And if this is a SolrCloud installation,
 it's particularly hard to get right.

 Please do not do this unless you have a very good reason, and please
 tell us what the reason is so we can perhaps suggest alternatives.

 Best,
 Erick

 On Fri, Mar 27, 2015 at 8:08 AM, Walter Underwood wun...@wunderwood.org
 wrote:
  Several years ago, I accidentally put Solr indexes on an NFS volume and
 it was 100X slower.
 
  If you have enough RAM, query speed should be OK, but startup time
 (loading indexes into file buffers) could be really long. Indexing could be
 quite slow.
 
  wunder
  Walter Underwood
  wun...@wunderwood.org
  http://observer.wunderwood.org/  (my blog)
 
 
  On Mar 26, 2015, at 11:31 PM, Shawn Heisey apa...@elyograg.org wrote:
 
  On 3/27/2015 12:06 AM, abhi Abhishek wrote:
  Greetings,
   I am trying to use a network shared location as my index
 directory.
  are there any known problems in using a Network File System for
 running a
  SOLR Instance?
 
  It is not recommended.  You will probably need to change the lockType,
  ... the default native probably will not work, and you might need to
  change it to none to get it working ... but that disables an important
  safety mechanism that prevents index corruption.
 
  http://stackoverflow.com/questions/9599529/solr-over-nfs-problems
 
  Thanks,
  Shawn
 
 



Re: Solr went on recovery multiple time.

2015-03-30 Thread Erick Erickson
Sometimes happens when your zookeeper timeout is short (the 15 second
timeout has been too short) and you hit a GC pause that is longer. Try
lengthening your ZK timeout to 45 or 60 seconds.

It would help if you told us about your setup, memory allocated to
Solr, all that stuff, see:

http://wiki.apache.org/solr/UsingMailingLists

Best,
Erick

On Mon, Mar 30, 2015 at 1:43 AM, sthita sthit...@gmail.com wrote:
 Solr server crashes 2-3 times a day .  When i looked into log i can see some
 recovery failure.

 I got these errors on my solr log

 WARN  - 2015-03-30 16:03:51.137; org.apache.solr.cloud.RecoveryStrategy;
 Stopping recovery for zkNodeName=searcher:8980_solr_rn0core=rn0
 ERROR - 2015-03-30 16:03:56.121; org.apache.solr.update.UpdateLog; Error
 inspecting tlog tlog{file=/solr/rn0/data/tlog/tlog.709
 refcount=2}
 ERROR - 2015-03-30 16:03:56.121; org.apache.solr.update.UpdateLog; Error
 inspecting tlog tlog{file=/solr/rn0/data/tlog/tlog.710
 refcount=2}
 WARN  - 2015-03-30 16:04:00.242; org.apache.solr.core.SolrCore; [rn0]
 PERFORMANCE WARNING: Overlapping onDeckSearchers=2

 I googled so many places, suggestions are related to my commit .

 My commit configuration on solrconfig.xml :


 updateHandler class=solr.DirectUpdateHandler2

 autoCommit
 maxTime60/maxTime
 openSearcherfalse/openSearcher
 /autoCommit

 updateLog
 str name=dir${solr.data.dir:}/str
 /updateLog
 /updateHandler


 Why is it happening ? Any idea/solution will be appreciated.





 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Solr-went-on-recovery-multiple-time-tp4196249.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Re: Unable to perform search query after changing uniqueKey

2015-03-30 Thread Erick Erickson
Well, let's see the definition of your ID field, 'cause I'm puzzled.

It's definitely A Bad Thing to have it be any kind of tokenized field
though, but that's a shot in the dark.

Best,
Erick

On Mon, Mar 30, 2015 at 2:17 AM, Zheng Lin Edwin Yeo
edwinye...@gmail.com wrote:
 Hi Mostafa,

 Yes, I've defined all the fields in schema.xml. It is able to work on the
 version without SolrCloud, but it is not working for the one with SolrCloud.
 Both of them are using the same schema.xml.

 Regards,
 Edwin



 On 30 March 2015 at 14:34, Mostafa Gomaa mostafa.goma...@gmail.com wrote:

 Hi Zheng,

 It's possible that there's a problem with your schema.xml. Are all fields
 defined and have appropriate options enabled?

 Regards,

 Mostafa.

 On Mon, Mar 30, 2015 at 7:49 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com
 
 wrote:

  Hi Erick,
 
  I've tried that, and removed the data directory from both the shards. But
  the same problem still occurs, so we probably can rule out the memory
  issue.
 
  Regards,
  Edwin
 
  On 30 March 2015 at 12:39, Erick Erickson erickerick...@gmail.com
 wrote:
 
   I meant shut down Solr and physically remove the entire data
   directory. Not saying this is the cure, but it can't hurt to rule out
   the index having memory...
  
   Best,
   Erick
  
   On Sun, Mar 29, 2015 at 6:35 PM, Zheng Lin Edwin Yeo
   edwinye...@gmail.com wrote:
Hi Erick,
   
I used the following query to delete all the index.
   
http://localhost:8983/solr/update?stream.body=
   deletequery*:*/query/delete
   http://localhost:8983/solr/update?stream.body=commit/
   
   
Or is it better to physically delete the entire data directory?
   
   
Regards,
Edwin
   
   
On 28 March 2015 at 02:27, Erick Erickson erickerick...@gmail.com
   wrote:
   
You say you re-indexed, did you _completely_ remove the data
 directory
first, i.e. the parent of the index and, maybe, tlog
 directories?
I've occasionally seen remnants of old definitions pollute the new
one, and since the uniqueKey key is so fundamental I can see it
being a problem.
   
Best,
Erick
   
On Fri, Mar 27, 2015 at 1:42 AM, Andrea Gazzarini 
   a.gazzar...@gmail.com
wrote:
 Hi Edwin,
 please provide some other detail about your context, (e.g.
 complete
 stacktrace, query you're issuing)

 Best,
 Andrea


 On 03/27/2015 09:38 AM, Zheng Lin Edwin Yeo wrote:

 Hi everyone,

 I've changed my uniqueKey to another name, instead of using id,
 on
   the
 schema.xml.

 However, after I have done the indexing (the indexing is
  successful),
I'm
 not able to perform a search query on it. I gives the error
 java.lang.NullPointerException.

 Is there other place which I need to configure, besides changing
  the
 uniqueKey field in scheam.xml?

 Regards,
 Edwin


   
  
 



Re: SOLR Index in shared/Network folder

2015-03-30 Thread Walter Underwood
I suggest that you do not try to save money on disk space. Disk is cheap.

You will spend weeks of expensive engineering time trying to make this work. 
Once you make it work, it will be slow an unreliable.

300GB Amazon EBS volumes are $180/year, $360/year for SSD. Just spend the money.

wunder
Walter Underwood
wun...@wunderwood.org
http://observer.wunderwood.org/  (my blog)


On Mar 29, 2015, at 10:42 PM, abhi Abhishek abhi26...@gmail.com wrote:

 Hello,
 Thanks for the suggestions. My aim is to reduce the disk space usage.
 I have 1 master with 2 slave configured, where slaves are used for
 searching and master ingests new data replicated to slaves, but as my index
 size is in 100's of GB we see 3x times space overhead. i would like to
 reduce this overhead, can you suggest something for this?
 
 Thanks in Advance
 
 Best Regards,
 Abhishek
 
 On Sat, Mar 28, 2015 at 12:13 AM, Erick Erickson erickerick...@gmail.com
 wrote:
 
 To pile on: If you're talking about pointing two Solr instances at the
 _same_ index, it doesn't matter whether you are on NFS or not, you'll
 have all sorts of problems. And if this is a SolrCloud installation,
 it's particularly hard to get right.
 
 Please do not do this unless you have a very good reason, and please
 tell us what the reason is so we can perhaps suggest alternatives.
 
 Best,
 Erick
 
 On Fri, Mar 27, 2015 at 8:08 AM, Walter Underwood wun...@wunderwood.org
 wrote:
 Several years ago, I accidentally put Solr indexes on an NFS volume and
 it was 100X slower.
 
 If you have enough RAM, query speed should be OK, but startup time
 (loading indexes into file buffers) could be really long. Indexing could be
 quite slow.
 
 wunder
 Walter Underwood
 wun...@wunderwood.org
 http://observer.wunderwood.org/  (my blog)
 
 
 On Mar 26, 2015, at 11:31 PM, Shawn Heisey apa...@elyograg.org wrote:
 
 On 3/27/2015 12:06 AM, abhi Abhishek wrote:
 Greetings,
 I am trying to use a network shared location as my index
 directory.
 are there any known problems in using a Network File System for
 running a
 SOLR Instance?
 
 It is not recommended.  You will probably need to change the lockType,
 ... the default native probably will not work, and you might need to
 change it to none to get it working ... but that disables an important
 safety mechanism that prevents index corruption.
 
 http://stackoverflow.com/questions/9599529/solr-over-nfs-problems
 
 Thanks,
 Shawn
 
 
 



Re: Date Time datatypes?

2015-03-30 Thread Erick Erickson
Hmmm, not OOB. You could index two tdate fields though and in the
second one only index seconds into the day and index the other with
date math (i.e. /DAY) to normalize them. Or a tdate field with just a
long field for the seconds.

You could then use the range faceting to, say, put the docs into hour
buckets, and your time:[17:00 TO 18:00] would turn into something like
time:[61200 TO 64800]. A bit awkward I'll admit, but probably easier
than a custom field type. Or add a field for hours if that fits the
use-case.

FWIW,
Erick

On Mon, Mar 30, 2015 at 4:11 AM, Bram Van Dam bram.van...@intix.eu wrote:
 Howdy folks,

 Is there any way index only the date and time portions of a datetime field?

 A Date is really a period of 24hrs, starting at 00:00 in said date's time
 zone. It would be useful if there was a way to search for documents of a
 certain date with these semantics.

 As for times, I'd like to be able to do queries like time:[17:00 TO 18:00].
 I suppose I could accomplish that by resetting the date portion to some
 bogus value, but then my facet/range values will contain that bogus date.

 I suppose the alternative is to create my own data types. Extending
 PrimitiveFieldType doesn't seem too hairy but maybe I'm overlooking some of
 the complexity...

 Thanks a bunch,

  - Bram


protected phrases - possible?

2015-03-30 Thread Tao, Jing
Hi,

The way our collection is setup, searches for breast cancer are returning 
results for ovarian cancer, or anything that contains either breast or 
cancer.  The reason is, we are searching across multiple fields.  Even though 
I have set a mm value so that if less than 3 terms, ALL terms much 
match...SOLR considers it all matched even though breast was in the title and 
cancer is in the description.

Is there a way to protect certain phrases so that they will not be tokenized?  
I tried using CommonGramsFilterFactory, but having breast cancer in the word 
list did not seem to do anything.  I'm guessing it's because the field is 
tokenized first, so nothing would match that phrase.  If I put breast and 
cancer as separate entries in the word list, I end up with too many 
unnecessary shingles, and breast and cancer are still two of the final 
terms.

I have a feeling CommonGramsFilterFactory is not the right way to handle this.  
What are other options?  Is it better to put all fields in one field, apply mm, 
and proximity boost?

Thanks!
Jing


Re: Same schema.xml is loaded for different cores in SolrCloud

2015-03-30 Thread Erick Erickson
OK, this is a bit confused:

1 You're starting two embedded Zookeepers but they don't know about
each other. So looking for the configsets is a bit confused.
2 There's no need to do the bootstrap thing after the first time. The
_very_ first time you do this it pushes the configs up to Zookeeper,
but after that you should just reference the config name.
3 you specify the config name when you _create_ the collection, not
when you start it up. You may be doing this, but your startup.
4 I think you're confusing shards with collections. The equivalent of
older-style cores would be just single-shard _collections_. Configs
are associated on the collection level, not the shard level as all
shards in a collection are presumed (indeed, _must_) use the same
configuration.

HTH,
Erick

On Mon, Mar 30, 2015 at 2:20 AM, Zheng Lin Edwin Yeo
edwinye...@gmail.com wrote:
 I've roughly know what is the problem from here.
 http://stackoverflow.com/questions/23338324/zookeeper-multiple-collection-different-schema

 However, I couldn't find the zoo_data directory in all of my solr folder.
 What could be the problem or where is the directory supposed to be located?

 Regards,
 Edwin


 On 30 March 2015 at 11:56, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote:

 Hi everyone,

 I've created a SolrCloud with multiple core, and I have different
 schema.xml for each of the core. However, when I start Solr, there's only
 one version of the schema.xml that is loaded onto Solr. Regardless of which
 core I go to, the schema.xml that is shown is the first one which I have
 loaded.

 What I did was, I have 3 cores: logmill, collection1 and collection2.
 Each of the core has 2 shrads: shard1 and shard2

 I first started the Solr with shard1 using the following command:
 java -Dcollection.configName=logmill -DzkRun -DnumShards=2
 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar

 After that I start shard2 using the following command:
 java -Dcollection.configName=logmill -DzkRun -DnumShards=2
 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar

 All the schema.xml loaded are from logmill core, even for the collection1
 and collection2.

 Even after I change the command to start shard1 with the following
 command, all the schema.xml are still from logmill
 java -Dcollection.configName=collection1 -DzkRun
 -DnumShards=2 -Dbootstrap_confdir=./solr/collection1/conf -jar start.jar


 How do I get Solr to read the different schema.xml for the different cores?

 Regards,
 Edwin



RE: Structured and Unstructured data indexing in SolrCloud

2015-03-30 Thread Reitzel, Charles
Hi Vijay, 

The short answer is yes, you can combine almost anything you want into a single 
collection.   But, in addition to working out your queries, you might want work 
out your data life cycle.

In our application, we have comingled the structured and unstructured documents 
into a single collection for initial development purposes.   The only field 
they have in common is the unique ID.Works fine.

In production, however, we see things like query rates, access controls, load 
balancing, availability, shard keys, overall document counts, update frequency, 
etc. will drive us to use separate collections.  For us, the deciding factor is 
less about structured vs. unstructured and more about public vs. private.   
We have developed our app so that splitting the collection will have minimal 
impact by executing separate queries, in parallel, at runtime.   

Of course, your application is different.  YMMV, etc.

hth,
Charlie


-Original Message-
From: Jack Krupansky [mailto:jack.krupan...@gmail.com] 
Sent: Sunday, March 29, 2015 4:26 PM
To: solr-user@lucene.apache.org
Subject: Re: Structured and Unstructured data indexing in SolrCloud

The first step is to work out the queries that you wish to perform - that will 
determine how the data should be organized in the Solr schema.

-- Jack Krupansky

On Sun, Mar 29, 2015 at 4:04 PM, Vijay Bhoomireddy  
vijaya.bhoomire...@whishworks.com wrote:

 Hi,



 We have a requirement where both structured and unstructured data 
 comes into the system. We need to index both of them and then enable 
 search functionality on it. We are using SolrCloud on Hadoop platform. 
 For structured data, we are planning to put the data into HBase and 
 for unstructured, directly into HDFS.



 My question is how to index these sources under a single Solr core? 
 Would that be possible to index both structured and unstructured data 
 under a single core/collection in SolrCloud and then enable search 
 functionality over that index?



 Thanks in advance.


 --
 The contents of this e-mail are confidential and for the exclusive use 
 of the intended recipient. If you receive this e-mail in error please 
 delete it from your system immediately and notify us either by e-mail 
 or telephone. You should not copy, forward or otherwise disclose the 
 content of the e-mail. The views expressed in this communication may 
 not necessarily be the view held by WHISHWORKS.


*
This e-mail may contain confidential or privileged information.
If you are not the intended recipient, please notify the sender immediately and 
then delete it.

TIAA-CREF
*


Re: protected phrases - possible?

2015-03-30 Thread Ahmet Arslan
Hi Jing,

You can boost phrases by pf (phrase fields) parameter. If you don't like this 
solution, you can modify search query at client side. E.g. surround certain 
phrases with quotes. This will force proximity search without interfering with 
tokenisation.

Ahmet


On Monday, March 30, 2015 8:49 PM, Tao, Jing j...@webmd.net wrote:
Hi,

The way our collection is setup, searches for breast cancer are returning 
results for ovarian cancer, or anything that contains either breast or 
cancer.  The reason is, we are searching across multiple fields.  Even though 
I have set a mm value so that if less than 3 terms, ALL terms much 
match...SOLR considers it all matched even though breast was in the title and 
cancer is in the description.

Is there a way to protect certain phrases so that they will not be tokenized?  
I tried using CommonGramsFilterFactory, but having breast cancer in the word 
list did not seem to do anything.  I'm guessing it's because the field is 
tokenized first, so nothing would match that phrase.  If I put breast and 
cancer as separate entries in the word list, I end up with too many 
unnecessary shingles, and breast and cancer are still two of the final 
terms.

I have a feeling CommonGramsFilterFactory is not the right way to handle this.  
What are other options?  Is it better to put all fields in one field, apply mm, 
and proximity boost?

Thanks!
Jing 


NoNode for /clusterstate.json in solr5.0.0 cloud

2015-03-30 Thread Purohit, Sumit
I have a basic Solr 5.0.0 cloud setup after following 
http://lucene.apache.org/solr/quickstart.html

I am trying to read data from spark and index it into solr using following lib:
https://github.com/LucidWorks/spark-solr

I am getting following error when my code try to make request to solr


Exception in thread main org.apache.spark.SparkException: Job aborted due to 
stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost 
task 0.0 in stage 0.0 (TID 0, localhost): 
org.apache.solr.common.cloud.ZooKeeperException:

at 
org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465)

.

..

..

Caused by: org.apache.zookeeper.KeeperException$NoNodeException: 
KeeperErrorCode = NoNode for /clusterstate.json

at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)

at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)

at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)

at org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500)



I am not sure how (and when) to create nodes for /clusterstate.json

I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar



Thanks for the help.

Sumit Purohit



Re: NoNode for /clusterstate.json in solr5.0.0 cloud

2015-03-30 Thread Timothy Potter
Anything in the server-side Solr logs? Also, if you go to the Solr admin
console at http://localhost:8983/solr, do you see the gettingstarted
collection in the cloud panel?



On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov
wrote:

 I have a basic Solr 5.0.0 cloud setup after following
 http://lucene.apache.org/solr/quickstart.html

 I am trying to read data from spark and index it into solr using following
 lib:
 https://github.com/LucidWorks/spark-solr

 I am getting following error when my code try to make request to solr


 Exception in thread main org.apache.spark.SparkException: Job aborted
 due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent
 failure: Lost task 0.0 in stage 0.0 (TID 0, localhost):
 org.apache.solr.common.cloud.ZooKeeperException:

 at
 org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465)

 .

 ..

 ..

 Caused by: org.apache.zookeeper.KeeperException$NoNodeException:
 KeeperErrorCode = NoNode for /clusterstate.json

 at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)

 at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)

 at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)

 at
 org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500)



 I am not sure how (and when) to create nodes for /clusterstate.json

 I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar



 Thanks for the help.

 Sumit Purohit




Re: Optimize SolrCloud without downtime

2015-03-30 Thread Erick Erickson
Hmmm, are you indexing during the time you see the deleted docs not
changing? Because this is very strange. Theoretically, if you reindex
everything, that should result in segments that have _no_ live docs in
them and they should really disappear ASAP.

One way to work around this if we determine that this is a problem in
Solr might be to create a second collection and index to _that_, then
use collection aliasing to use the new index. But that's not
optimal...

Best,
Erick


On Mon, Mar 30, 2015 at 4:35 AM, Pavel Hladik
pavel.hla...@profimedia.cz wrote:
 Hi,

 thanks for reply. We have a lot of deleted docs cause we have to reindex all
 records from time to time, changing some important parameters.. When we do
 update, it means create and delete.

 Our deleted docs do not disappear by merging segments. I see our deleted
 docs are almost the same number after two weeks and always around 40
 segments. I ask our developers and they confirmed we use default settings
 for merging policy. Please, do you have some tips and tricks for efficient
 setting for merging segments?

 Our HW is RAID5 SAS disks and 40GB of memory dedicated for Solr.

 Regards,
 Pavel





 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Optimize-SolrCloud-without-downtime-tp4195170p4196273.html
 Sent from the Solr - User mailing list archive at Nabble.com.


how do you replicate solr-cloud between datacenters?

2015-03-30 Thread Timothy Ehlers
Can you use /replication ??? How would you do this between datacenters?

-- 
Tim Ehlers


RE: NoNode for /clusterstate.json in solr5.0.0 cloud

2015-03-30 Thread Purohit, Sumit
Thanks Tim,

i had to make some changes in my local spark-solr clone to build it for sorl5. 
If its ok, i can commit these to github.

thanks
sumit

From: Timothy Potter [thelabd...@gmail.com]
Sent: Monday, March 30, 2015 2:27 PM
To: solr-user@lucene.apache.org
Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud

Ok, let me upgrade my version of spark-solr to 5 to see what I get ...

On Mon, Mar 30, 2015 at 2:26 PM, Purohit, Sumit sumit.puro...@pnnl.gov
wrote:

 yes there is getting started collection..
 and on admin webpage  console--cloud---tree---/clusterstate.json  shows
 me this table

 version =1
 aversion=0
 children_count=0
 ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901)
 cversion=0
 czxid=32
 ephemeralOwner=0
 mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453)
 mzxid=110
 pzxid=32
 dataLength=2

 children_count=0  seems related to no node error.

 thanks
 sumit
 
 From: Timothy Potter [thelabd...@gmail.com]
 Sent: Monday, March 30, 2015 2:18 PM
 To: solr-user@lucene.apache.org
 Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud

 Anything in the server-side Solr logs? Also, if you go to the Solr admin
 console at http://localhost:8983/solr, do you see the gettingstarted
 collection in the cloud panel?



 On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov
 wrote:

  I have a basic Solr 5.0.0 cloud setup after following
  http://lucene.apache.org/solr/quickstart.html
 
  I am trying to read data from spark and index it into solr using
 following
  lib:
  https://github.com/LucidWorks/spark-solr
 
  I am getting following error when my code try to make request to solr
 
 
  Exception in thread main org.apache.spark.SparkException: Job aborted
  due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent
  failure: Lost task 0.0 in stage 0.0 (TID 0, localhost):
  org.apache.solr.common.cloud.ZooKeeperException:
 
  at
 
 org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465)
 
  .
 
  ..
 
  ..
 
  Caused by: org.apache.zookeeper.KeeperException$NoNodeException:
  KeeperErrorCode = NoNode for /clusterstate.json
 
  at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
 
  at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
 
  at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
 
  at
 
 org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500)
 
 
 
  I am not sure how (and when) to create nodes for /clusterstate.json
 
  I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar
 
 
 
  Thanks for the help.
 
  Sumit Purohit
 
 



RE: NoNode for /clusterstate.json in solr5.0.0 cloud

2015-03-30 Thread Purohit, Sumit
Yes i am using default localhost:9983 and i think my program does connect with 
zookeeper as per logs :


2015-03-30 15:00:51,512 [Executor task launch worker-0] INFO  SolrZkClient  - 
Using default ZkCredentialsProvider
2015-03-30 15:00:51,530 [Executor task launch worker-0] INFO  ConnectionManager 
 - Waiting for client to connect to ZooKeeper
2015-03-30 15:00:51,544 [zkCallback-3-thread-1] INFO  ConnectionManager  - 
Watcher org.apache.solr.common.cloud.ConnectionManager@51ffb98d 
name:ZooKeeperConnection Watcher:localhost:9983/sorl got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
2015-03-30 15:00:51,544 [Executor task launch worker-0] INFO  ConnectionManager 
 - Client is connected to ZooKeeper
2015-03-30 15:00:51,544 [Executor task launch worker-0] INFO  SolrZkClient  - 
Using default ZkACLProvider
2015-03-30 15:00:51,548 [Executor task launch worker-0] INFO  SolrZkClient  - 
makePath: /clusterstate.json
2015-03-30 15:00:51,557 [Executor task launch worker-0] ERROR Executor  - 
Exception in task 0.0 in stage 0.0 (TID 0)

So to upgrade spark-solr project, i had to change all the references to of 
class xxxSolrServer to xxxSolrClient . did you do something similar. 

If you plan to push it to github, i can grab that and compare with my 
environment.  
Thanks for you prompt response(s).


thanks
sumit

From: Timothy Potter [thelabd...@gmail.com]
Sent: Monday, March 30, 2015 2:54 PM
To: solr-user@lucene.apache.org
Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud

I upgraded the spark-solr project to solrj-5.0.0 and was able to index into
the gettingstarted collection using Solr 5.0.0, so seems like it may be
environmental. Almost seems like the spark project is looking at the wrong
ZooKeeper? Are you using the default -zkHost localhost:9983

On Mon, Mar 30, 2015 at 2:32 PM, Purohit, Sumit sumit.puro...@pnnl.gov
wrote:

 Thanks Tim,

 i had to make some changes in my local spark-solr clone to build it for
 sorl5.
 If its ok, i can commit these to github.

 thanks
 sumit
 
 From: Timothy Potter [thelabd...@gmail.com]
 Sent: Monday, March 30, 2015 2:27 PM
 To: solr-user@lucene.apache.org
 Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud

 Ok, let me upgrade my version of spark-solr to 5 to see what I get ...

 On Mon, Mar 30, 2015 at 2:26 PM, Purohit, Sumit sumit.puro...@pnnl.gov
 wrote:

  yes there is getting started collection..
  and on admin webpage  console--cloud---tree---/clusterstate.json
 shows
  me this table
 
  version =1
  aversion=0
  children_count=0
  ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901)
  cversion=0
  czxid=32
  ephemeralOwner=0
  mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453)
  mzxid=110
  pzxid=32
  dataLength=2
 
  children_count=0  seems related to no node error.
 
  thanks
  sumit
  
  From: Timothy Potter [thelabd...@gmail.com]
  Sent: Monday, March 30, 2015 2:18 PM
  To: solr-user@lucene.apache.org
  Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud
 
  Anything in the server-side Solr logs? Also, if you go to the Solr admin
  console at http://localhost:8983/solr, do you see the gettingstarted
  collection in the cloud panel?
 
 
 
  On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov
  wrote:
 
   I have a basic Solr 5.0.0 cloud setup after following
   http://lucene.apache.org/solr/quickstart.html
  
   I am trying to read data from spark and index it into solr using
  following
   lib:
   https://github.com/LucidWorks/spark-solr
  
   I am getting following error when my code try to make request to solr
  
  
   Exception in thread main org.apache.spark.SparkException: Job aborted
   due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent
   failure: Lost task 0.0 in stage 0.0 (TID 0, localhost):
   org.apache.solr.common.cloud.ZooKeeperException:
  
   at
  
 
 org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465)
  
   .
  
   ..
  
   ..
  
   Caused by: org.apache.zookeeper.KeeperException$NoNodeException:
   KeeperErrorCode = NoNode for /clusterstate.json
  
   at
 org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
  
   at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
  
   at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
  
   at
  
 
 org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500)
  
  
  
   I am not sure how (and when) to create nodes for /clusterstate.json
  
   I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar
  
  
  
   Thanks for the help.
  
   Sumit Purohit
  
  
 



Re: NoNode for /clusterstate.json in solr5.0.0 cloud

2015-03-30 Thread Timothy Potter
Ok, let me upgrade my version of spark-solr to 5 to see what I get ...

On Mon, Mar 30, 2015 at 2:26 PM, Purohit, Sumit sumit.puro...@pnnl.gov
wrote:

 yes there is getting started collection..
 and on admin webpage  console--cloud---tree---/clusterstate.json  shows
 me this table

 version =1
 aversion=0
 children_count=0
 ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901)
 cversion=0
 czxid=32
 ephemeralOwner=0
 mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453)
 mzxid=110
 pzxid=32
 dataLength=2

 children_count=0  seems related to no node error.

 thanks
 sumit
 
 From: Timothy Potter [thelabd...@gmail.com]
 Sent: Monday, March 30, 2015 2:18 PM
 To: solr-user@lucene.apache.org
 Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud

 Anything in the server-side Solr logs? Also, if you go to the Solr admin
 console at http://localhost:8983/solr, do you see the gettingstarted
 collection in the cloud panel?



 On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov
 wrote:

  I have a basic Solr 5.0.0 cloud setup after following
  http://lucene.apache.org/solr/quickstart.html
 
  I am trying to read data from spark and index it into solr using
 following
  lib:
  https://github.com/LucidWorks/spark-solr
 
  I am getting following error when my code try to make request to solr
 
 
  Exception in thread main org.apache.spark.SparkException: Job aborted
  due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent
  failure: Lost task 0.0 in stage 0.0 (TID 0, localhost):
  org.apache.solr.common.cloud.ZooKeeperException:
 
  at
 
 org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465)
 
  .
 
  ..
 
  ..
 
  Caused by: org.apache.zookeeper.KeeperException$NoNodeException:
  KeeperErrorCode = NoNode for /clusterstate.json
 
  at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
 
  at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
 
  at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
 
  at
 
 org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500)
 
 
 
  I am not sure how (and when) to create nodes for /clusterstate.json
 
  I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar
 
 
 
  Thanks for the help.
 
  Sumit Purohit
 
 



Re: how do you replicate solr-cloud between datacenters?

2015-03-30 Thread Jack Krupansky
That's an open issue. See:
https://issues.apache.org/jira/browse/SOLR-6273

-- Jack Krupansky

On Mon, Mar 30, 2015 at 5:45 PM, Timothy Ehlers ehle...@gmail.com wrote:

 Can you use /replication ??? How would you do this between datacenters?

 --
 Tim Ehlers



RE: NoNode for /clusterstate.json in solr5.0.0 cloud

2015-03-30 Thread Purohit, Sumit
yes there is getting started collection..
and on admin webpage  console--cloud---tree---/clusterstate.json  shows me 
this table

version =1
aversion=0
children_count=0
ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901)
cversion=0
czxid=32
ephemeralOwner=0
mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453)
mzxid=110
pzxid=32
dataLength=2

children_count=0  seems related to no node error.

thanks
sumit

From: Timothy Potter [thelabd...@gmail.com]
Sent: Monday, March 30, 2015 2:18 PM
To: solr-user@lucene.apache.org
Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud

Anything in the server-side Solr logs? Also, if you go to the Solr admin
console at http://localhost:8983/solr, do you see the gettingstarted
collection in the cloud panel?



On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov
wrote:

 I have a basic Solr 5.0.0 cloud setup after following
 http://lucene.apache.org/solr/quickstart.html

 I am trying to read data from spark and index it into solr using following
 lib:
 https://github.com/LucidWorks/spark-solr

 I am getting following error when my code try to make request to solr


 Exception in thread main org.apache.spark.SparkException: Job aborted
 due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent
 failure: Lost task 0.0 in stage 0.0 (TID 0, localhost):
 org.apache.solr.common.cloud.ZooKeeperException:

 at
 org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465)

 .

 ..

 ..

 Caused by: org.apache.zookeeper.KeeperException$NoNodeException:
 KeeperErrorCode = NoNode for /clusterstate.json

 at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)

 at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)

 at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)

 at
 org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500)



 I am not sure how (and when) to create nodes for /clusterstate.json

 I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar



 Thanks for the help.

 Sumit Purohit




Re: NoNode for /clusterstate.json in solr5.0.0 cloud

2015-03-30 Thread Timothy Potter
I upgraded the spark-solr project to solrj-5.0.0 and was able to index into
the gettingstarted collection using Solr 5.0.0, so seems like it may be
environmental. Almost seems like the spark project is looking at the wrong
ZooKeeper? Are you using the default -zkHost localhost:9983

On Mon, Mar 30, 2015 at 2:32 PM, Purohit, Sumit sumit.puro...@pnnl.gov
wrote:

 Thanks Tim,

 i had to make some changes in my local spark-solr clone to build it for
 sorl5.
 If its ok, i can commit these to github.

 thanks
 sumit
 
 From: Timothy Potter [thelabd...@gmail.com]
 Sent: Monday, March 30, 2015 2:27 PM
 To: solr-user@lucene.apache.org
 Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud

 Ok, let me upgrade my version of spark-solr to 5 to see what I get ...

 On Mon, Mar 30, 2015 at 2:26 PM, Purohit, Sumit sumit.puro...@pnnl.gov
 wrote:

  yes there is getting started collection..
  and on admin webpage  console--cloud---tree---/clusterstate.json
 shows
  me this table
 
  version =1
  aversion=0
  children_count=0
  ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901)
  cversion=0
  czxid=32
  ephemeralOwner=0
  mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453)
  mzxid=110
  pzxid=32
  dataLength=2
 
  children_count=0  seems related to no node error.
 
  thanks
  sumit
  
  From: Timothy Potter [thelabd...@gmail.com]
  Sent: Monday, March 30, 2015 2:18 PM
  To: solr-user@lucene.apache.org
  Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud
 
  Anything in the server-side Solr logs? Also, if you go to the Solr admin
  console at http://localhost:8983/solr, do you see the gettingstarted
  collection in the cloud panel?
 
 
 
  On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov
  wrote:
 
   I have a basic Solr 5.0.0 cloud setup after following
   http://lucene.apache.org/solr/quickstart.html
  
   I am trying to read data from spark and index it into solr using
  following
   lib:
   https://github.com/LucidWorks/spark-solr
  
   I am getting following error when my code try to make request to solr
  
  
   Exception in thread main org.apache.spark.SparkException: Job aborted
   due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent
   failure: Lost task 0.0 in stage 0.0 (TID 0, localhost):
   org.apache.solr.common.cloud.ZooKeeperException:
  
   at
  
 
 org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465)
  
   .
  
   ..
  
   ..
  
   Caused by: org.apache.zookeeper.KeeperException$NoNodeException:
   KeeperErrorCode = NoNode for /clusterstate.json
  
   at
 org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
  
   at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
  
   at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
  
   at
  
 
 org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500)
  
  
  
   I am not sure how (and when) to create nodes for /clusterstate.json
  
   I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar
  
  
  
   Thanks for the help.
  
   Sumit Purohit
  
  
 



Re: Solr went on recovery multiple time.

2015-03-30 Thread sthita
Thanks Erick for the reply. These are the configurations.

1.Solr servers count : 4 (1 master , 3 slaves)
2.My initial and maximum memory allocations to each solr servers are :
-Xms5g 
-Xmx6g
3. 5 running zookeepers 
4. My index size in each solr server around: 48GB
5. Every sec merger add/delete 500 docs to/from solr

yes, i checked the the ZK config timeout is 10sec. I will increase the
timeout to 45sec, Hope that will help. 

ZK config : 

# The number of milliseconds of each tick
tickTime=2000
# The number of ticks that the initial
# synchronization phase can take
initLimit=10
# The number of ticks that can pass between
# sending a request and getting an acknowledgement
syncLimit=5







--
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr-went-on-recovery-multiple-time-tp4196249p4196450.html
Sent from the Solr - User mailing list archive at Nabble.com.


Collapse and Expand behaviour on result with 1 document.

2015-03-30 Thread Derek Poh

Hi

I have a query which return 1 document.
When I add the collapse and expand parameters to it, 
expand=trueexpand.rows=5fq={!collapse%20field=P_SupplierId}, the 
expanded section is empty (lst name=expanded/).


Is this the behaviour of collapse and expand parameters on result which 
contain only 1 document?


-Derek




Re: Unable to perform search query after changing uniqueKey

2015-03-30 Thread Zheng Lin Edwin Yeo
Hi Erick,

I've changed the uniqueKey from id to Item No.

uniqueKeyItem No/uniqueKey


Below are my definitions for both the id and Item No.

field name=id type=string indexed=true stored=true
required=false multiValued=false /
field name=Item No type=text_general indexed=true stored=true/

Regards,
Edwin


On 30 March 2015 at 23:05, Erick Erickson erickerick...@gmail.com wrote:

 Well, let's see the definition of your ID field, 'cause I'm puzzled.

 It's definitely A Bad Thing to have it be any kind of tokenized field
 though, but that's a shot in the dark.

 Best,
 Erick

 On Mon, Mar 30, 2015 at 2:17 AM, Zheng Lin Edwin Yeo
 edwinye...@gmail.com wrote:
  Hi Mostafa,
 
  Yes, I've defined all the fields in schema.xml. It is able to work on the
  version without SolrCloud, but it is not working for the one with
 SolrCloud.
  Both of them are using the same schema.xml.
 
  Regards,
  Edwin
 
 
 
  On 30 March 2015 at 14:34, Mostafa Gomaa mostafa.goma...@gmail.com
 wrote:
 
  Hi Zheng,
 
  It's possible that there's a problem with your schema.xml. Are all
 fields
  defined and have appropriate options enabled?
 
  Regards,
 
  Mostafa.
 
  On Mon, Mar 30, 2015 at 7:49 AM, Zheng Lin Edwin Yeo 
 edwinye...@gmail.com
  
  wrote:
 
   Hi Erick,
  
   I've tried that, and removed the data directory from both the shards.
 But
   the same problem still occurs, so we probably can rule out the
 memory
   issue.
  
   Regards,
   Edwin
  
   On 30 March 2015 at 12:39, Erick Erickson erickerick...@gmail.com
  wrote:
  
I meant shut down Solr and physically remove the entire data
directory. Not saying this is the cure, but it can't hurt to rule
 out
the index having memory...
   
Best,
Erick
   
On Sun, Mar 29, 2015 at 6:35 PM, Zheng Lin Edwin Yeo
edwinye...@gmail.com wrote:
 Hi Erick,

 I used the following query to delete all the index.

 http://localhost:8983/solr/update?stream.body=
deletequery*:*/query/delete
http://localhost:8983/solr/update?stream.body=commit/


 Or is it better to physically delete the entire data directory?


 Regards,
 Edwin


 On 28 March 2015 at 02:27, Erick Erickson 
 erickerick...@gmail.com
wrote:

 You say you re-indexed, did you _completely_ remove the data
  directory
 first, i.e. the parent of the index and, maybe, tlog
  directories?
 I've occasionally seen remnants of old definitions pollute the
 new
 one, and since the uniqueKey key is so fundamental I can see it
 being a problem.

 Best,
 Erick

 On Fri, Mar 27, 2015 at 1:42 AM, Andrea Gazzarini 
a.gazzar...@gmail.com
 wrote:
  Hi Edwin,
  please provide some other detail about your context, (e.g.
  complete
  stacktrace, query you're issuing)
 
  Best,
  Andrea
 
 
  On 03/27/2015 09:38 AM, Zheng Lin Edwin Yeo wrote:
 
  Hi everyone,
 
  I've changed my uniqueKey to another name, instead of using
 id,
  on
the
  schema.xml.
 
  However, after I have done the indexing (the indexing is
   successful),
 I'm
  not able to perform a search query on it. I gives the error
  java.lang.NullPointerException.
 
  Is there other place which I need to configure, besides
 changing
   the
  uniqueKey field in scheam.xml?
 
  Regards,
  Edwin
 
 

   
  
 



Re: Collapse and Expand behaviour on result with 1 document.

2015-03-30 Thread Joel Bernstein
The expanded section will only include groups that have expanded documents.

So, if the document that in the main result set has no documents to expand,
then this is working as expected.



Joel Bernstein
http://joelsolr.blogspot.com/

On Mon, Mar 30, 2015 at 8:43 PM, Derek Poh d...@globalsources.com wrote:

 Hi

 I have a query which return 1 document.
 When I add the collapse and expand parameters to it,
 expand=trueexpand.rows=5fq={!collapse%20field=P_SupplierId}, the
 expanded section is empty (lst name=expanded/).

 Is this the behaviour of collapse and expand parameters on result which
 contain only 1 document?

 -Derek





RE: Solr Unexpected Query Parser Exception

2015-03-30 Thread Reitzel, Charles
Saw that one.  Can't remember for certain, but recall the actual syntax error 
was in a filter query.   It could have been a quoting error or a date math 
error in a range expression.  But, either way, the issue was in the fq.  Using 
edismax.  hth

-Original Message-
From: Jack Krupansky [mailto:jack.krupan...@gmail.com] 
Sent: Friday, March 20, 2015 5:08 PM
To: solr-user@lucene.apache.org
Subject: Re: Solr Unexpected Query Parser Exception

Which query parser are you using? The dismax query parser does not support wild 
cards or *:*.

Either way, the error message is unhelpful - worth filing a Jira.

-- Jack Krupansky

On Fri, Mar 20, 2015 at 7:21 AM, Vishnu Mishra vdil...@gmail.com wrote:

 Hi, I am using solr 4.10.3 and doing distributed shard query. I am 
 getting following syntax exception at regular intervals.


  ERROR org.apache.solr.core.SolrCore  ?
 org.apache.solr.common.SolrException: org.apache.solr.search.SyntaxError:
 Cannot parse '*:*': Encountered EOF at line 1, column 3.
 Was expecting one of:
 NOT ...
 + ...
 - ...
 BAREOPER ...
 ( ...
 * ...
 ^ ...
 QUOTED ...
 TERM ...
 FUZZY_SLOP ...
 PREFIXTERM ...
 WILDTERM ...
 REGEXPTERM ...
 [ ...
 { ...
 PROXY_MF ...
 JOIN_MF ...
 FRQ_CUT_OFF ...
 NUMBER ...
 TERM ...
 * ...


 My query to solr is *:* can anyone explain me why this exception is 
 occurring.



 --
 View this message in context:
 http://lucene.472066.n3.nabble.com/Solr-Unexpected-Query-Parser-Except
 ion-tp4194156.html Sent from the Solr - User mailing list archive at 
 Nabble.com.


*
This e-mail may contain confidential or privileged information.
If you are not the intended recipient, please notify the sender immediately and 
then delete it.

TIAA-CREF
*


RE: NoNode for /clusterstate.json in solr5.0.0 cloud

2015-03-30 Thread Purohit, Sumit
One development difference is that i am trying to use it in a scala project.
I compile spark-solr as jar and use it in my scala project.

Do you think this might be a reason.

thanks
sumit

From: Purohit, Sumit
Sent: Monday, March 30, 2015 3:19 PM
To: solr-user@lucene.apache.org
Subject: RE: NoNode for /clusterstate.json in solr5.0.0 cloud

Yes i am using default localhost:9983 and i think my program does connect with 
zookeeper as per logs :


2015-03-30 15:00:51,512 [Executor task launch worker-0] INFO  SolrZkClient  - 
Using default ZkCredentialsProvider
2015-03-30 15:00:51,530 [Executor task launch worker-0] INFO  ConnectionManager 
 - Waiting for client to connect to ZooKeeper
2015-03-30 15:00:51,544 [zkCallback-3-thread-1] INFO  ConnectionManager  - 
Watcher org.apache.solr.common.cloud.ConnectionManager@51ffb98d 
name:ZooKeeperConnection Watcher:localhost:9983/sorl got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
2015-03-30 15:00:51,544 [Executor task launch worker-0] INFO  ConnectionManager 
 - Client is connected to ZooKeeper
2015-03-30 15:00:51,544 [Executor task launch worker-0] INFO  SolrZkClient  - 
Using default ZkACLProvider
2015-03-30 15:00:51,548 [Executor task launch worker-0] INFO  SolrZkClient  - 
makePath: /clusterstate.json
2015-03-30 15:00:51,557 [Executor task launch worker-0] ERROR Executor  - 
Exception in task 0.0 in stage 0.0 (TID 0)

So to upgrade spark-solr project, i had to change all the references to of 
class xxxSolrServer to xxxSolrClient . did you do something similar.

If you plan to push it to github, i can grab that and compare with my 
environment.
Thanks for you prompt response(s).


thanks
sumit

From: Timothy Potter [thelabd...@gmail.com]
Sent: Monday, March 30, 2015 2:54 PM
To: solr-user@lucene.apache.org
Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud

I upgraded the spark-solr project to solrj-5.0.0 and was able to index into
the gettingstarted collection using Solr 5.0.0, so seems like it may be
environmental. Almost seems like the spark project is looking at the wrong
ZooKeeper? Are you using the default -zkHost localhost:9983

On Mon, Mar 30, 2015 at 2:32 PM, Purohit, Sumit sumit.puro...@pnnl.gov
wrote:

 Thanks Tim,

 i had to make some changes in my local spark-solr clone to build it for
 sorl5.
 If its ok, i can commit these to github.

 thanks
 sumit
 
 From: Timothy Potter [thelabd...@gmail.com]
 Sent: Monday, March 30, 2015 2:27 PM
 To: solr-user@lucene.apache.org
 Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud

 Ok, let me upgrade my version of spark-solr to 5 to see what I get ...

 On Mon, Mar 30, 2015 at 2:26 PM, Purohit, Sumit sumit.puro...@pnnl.gov
 wrote:

  yes there is getting started collection..
  and on admin webpage  console--cloud---tree---/clusterstate.json
 shows
  me this table
 
  version =1
  aversion=0
  children_count=0
  ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901)
  cversion=0
  czxid=32
  ephemeralOwner=0
  mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453)
  mzxid=110
  pzxid=32
  dataLength=2
 
  children_count=0  seems related to no node error.
 
  thanks
  sumit
  
  From: Timothy Potter [thelabd...@gmail.com]
  Sent: Monday, March 30, 2015 2:18 PM
  To: solr-user@lucene.apache.org
  Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud
 
  Anything in the server-side Solr logs? Also, if you go to the Solr admin
  console at http://localhost:8983/solr, do you see the gettingstarted
  collection in the cloud panel?
 
 
 
  On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov
  wrote:
 
   I have a basic Solr 5.0.0 cloud setup after following
   http://lucene.apache.org/solr/quickstart.html
  
   I am trying to read data from spark and index it into solr using
  following
   lib:
   https://github.com/LucidWorks/spark-solr
  
   I am getting following error when my code try to make request to solr
  
  
   Exception in thread main org.apache.spark.SparkException: Job aborted
   due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent
   failure: Lost task 0.0 in stage 0.0 (TID 0, localhost):
   org.apache.solr.common.cloud.ZooKeeperException:
  
   at
  
 
 org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465)
  
   .
  
   ..
  
   ..
  
   Caused by: org.apache.zookeeper.KeeperException$NoNodeException:
   KeeperErrorCode = NoNode for /clusterstate.json
  
   at
 org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
  
   at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
  
   at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783)
  
   at
  
 
 org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500)
  
  
  
   I am not sure how 

Re: Same schema.xml is loaded for different cores in SolrCloud

2015-03-30 Thread Zheng Lin Edwin Yeo
Hi Erick,

I've started shard2 with the following command instead, but it's still the
same problem.
java -DzkHost=localhost:9983 -Djetty.port=8984 -jar start.jar

But you mean for shard1 we do not have to include 
-Dbootstrap_confdir=./solr/logmill/conf for subsequent startup?

Regards,
Edwin


On 31 March 2015 at 00:46, Erick Erickson erickerick...@gmail.com wrote:

 OK, this is a bit confused:

 1 You're starting two embedded Zookeepers but they don't know about
 each other. So looking for the configsets is a bit confused.
 2 There's no need to do the bootstrap thing after the first time. The
 _very_ first time you do this it pushes the configs up to Zookeeper,
 but after that you should just reference the config name.
 3 you specify the config name when you _create_ the collection, not
 when you start it up. You may be doing this, but your startup.
 4 I think you're confusing shards with collections. The equivalent of
 older-style cores would be just single-shard _collections_. Configs
 are associated on the collection level, not the shard level as all
 shards in a collection are presumed (indeed, _must_) use the same
 configuration.

 HTH,
 Erick

 On Mon, Mar 30, 2015 at 2:20 AM, Zheng Lin Edwin Yeo
 edwinye...@gmail.com wrote:
  I've roughly know what is the problem from here.
 
 http://stackoverflow.com/questions/23338324/zookeeper-multiple-collection-different-schema
 
  However, I couldn't find the zoo_data directory in all of my solr folder.
  What could be the problem or where is the directory supposed to be
 located?
 
  Regards,
  Edwin
 
 
  On 30 March 2015 at 11:56, Zheng Lin Edwin Yeo edwinye...@gmail.com
 wrote:
 
  Hi everyone,
 
  I've created a SolrCloud with multiple core, and I have different
  schema.xml for each of the core. However, when I start Solr, there's
 only
  one version of the schema.xml that is loaded onto Solr. Regardless of
 which
  core I go to, the schema.xml that is shown is the first one which I have
  loaded.
 
  What I did was, I have 3 cores: logmill, collection1 and collection2.
  Each of the core has 2 shrads: shard1 and shard2
 
  I first started the Solr with shard1 using the following command:
  java -Dcollection.configName=logmill -DzkRun -DnumShards=2
  -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar
 
  After that I start shard2 using the following command:
  java -Dcollection.configName=logmill -DzkRun -DnumShards=2
  -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar
 
  All the schema.xml loaded are from logmill core, even for the
 collection1
  and collection2.
 
  Even after I change the command to start shard1 with the following
  command, all the schema.xml are still from logmill
  java -Dcollection.configName=collection1 -DzkRun
  -DnumShards=2 -Dbootstrap_confdir=./solr/collection1/conf -jar start.jar
 
 
  How do I get Solr to read the different schema.xml for the different
 cores?
 
  Regards,
  Edwin
 



Re: Unable to perform search query after changing uniqueKey

2015-03-30 Thread Zheng Lin Edwin Yeo
Latest information that I've found for this is that the error only occurs
for shard2.

If I do a search for just shard1, those records that are assigned to shard1
will be able to be displayed. Only when I search for shard2 will the
NullPointerException error occurs. Previously I was doing a search for both
shards.

Is there any settings that I required to do for shard2 in order to solve
this issue? Currently I have not made any changes to the shards since I
created it using
http://localhost:8983/solr/admin/collections?action=CREATEname=nps1numShards=2collection.configName=collection1


Regards,
Edwin

On 31 March 2015 at 09:42, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote:

 Hi Erick,

 I've changed the uniqueKey from id to Item No.

 uniqueKeyItem No/uniqueKey


 Below are my definitions for both the id and Item No.

 field name=id type=string indexed=true stored=true
 required=false multiValued=false /
 field name=Item No type=text_general indexed=true stored=true/

 Regards,
 Edwin


 On 30 March 2015 at 23:05, Erick Erickson erickerick...@gmail.com wrote:

 Well, let's see the definition of your ID field, 'cause I'm puzzled.

 It's definitely A Bad Thing to have it be any kind of tokenized field
 though, but that's a shot in the dark.

 Best,
 Erick

 On Mon, Mar 30, 2015 at 2:17 AM, Zheng Lin Edwin Yeo
 edwinye...@gmail.com wrote:
  Hi Mostafa,
 
  Yes, I've defined all the fields in schema.xml. It is able to work on
 the
  version without SolrCloud, but it is not working for the one with
 SolrCloud.
  Both of them are using the same schema.xml.
 
  Regards,
  Edwin
 
 
 
  On 30 March 2015 at 14:34, Mostafa Gomaa mostafa.goma...@gmail.com
 wrote:
 
  Hi Zheng,
 
  It's possible that there's a problem with your schema.xml. Are all
 fields
  defined and have appropriate options enabled?
 
  Regards,
 
  Mostafa.
 
  On Mon, Mar 30, 2015 at 7:49 AM, Zheng Lin Edwin Yeo 
 edwinye...@gmail.com
  
  wrote:
 
   Hi Erick,
  
   I've tried that, and removed the data directory from both the
 shards. But
   the same problem still occurs, so we probably can rule out the
 memory
   issue.
  
   Regards,
   Edwin
  
   On 30 March 2015 at 12:39, Erick Erickson erickerick...@gmail.com
  wrote:
  
I meant shut down Solr and physically remove the entire data
directory. Not saying this is the cure, but it can't hurt to rule
 out
the index having memory...
   
Best,
Erick
   
On Sun, Mar 29, 2015 at 6:35 PM, Zheng Lin Edwin Yeo
edwinye...@gmail.com wrote:
 Hi Erick,

 I used the following query to delete all the index.

 http://localhost:8983/solr/update?stream.body=
deletequery*:*/query/delete
http://localhost:8983/solr/update?stream.body=commit/


 Or is it better to physically delete the entire data directory?


 Regards,
 Edwin


 On 28 March 2015 at 02:27, Erick Erickson 
 erickerick...@gmail.com
wrote:

 You say you re-indexed, did you _completely_ remove the data
  directory
 first, i.e. the parent of the index and, maybe, tlog
  directories?
 I've occasionally seen remnants of old definitions pollute
 the new
 one, and since the uniqueKey key is so fundamental I can see
 it
 being a problem.

 Best,
 Erick

 On Fri, Mar 27, 2015 at 1:42 AM, Andrea Gazzarini 
a.gazzar...@gmail.com
 wrote:
  Hi Edwin,
  please provide some other detail about your context, (e.g.
  complete
  stacktrace, query you're issuing)
 
  Best,
  Andrea
 
 
  On 03/27/2015 09:38 AM, Zheng Lin Edwin Yeo wrote:
 
  Hi everyone,
 
  I've changed my uniqueKey to another name, instead of using
 id,
  on
the
  schema.xml.
 
  However, after I have done the indexing (the indexing is
   successful),
 I'm
  not able to perform a search query on it. I gives the error
  java.lang.NullPointerException.
 
  Is there other place which I need to configure, besides
 changing
   the
  uniqueKey field in scheam.xml?
 
  Regards,
  Edwin
 
 

   
  
 





Re: Solr went on recovery multiple time.

2015-03-30 Thread William Bell
How big is your index? -Xmx6g is pretty small in general.

On Mon, Mar 30, 2015 at 7:20 PM, sthita sthit...@gmail.com wrote:

 Thanks Erick for the reply. These are the configurations.

 1.Solr servers count : 4 (1 master , 3 slaves)
 2.My initial and maximum memory allocations to each solr servers are :
 -Xms5g
 -Xmx6g
 3. 5 running zookeepers
 4. My index size in each solr server around: 48GB
 5. Every sec merger add/delete 500 docs to/from solr

 yes, i checked the the ZK config timeout is 10sec. I will increase the
 timeout to 45sec, Hope that will help.

 ZK config :

 # The number of milliseconds of each tick
 tickTime=2000
 # The number of ticks that the initial
 # synchronization phase can take
 initLimit=10
 # The number of ticks that can pass between
 # sending a request and getting an acknowledgement
 syncLimit=5







 --
 View this message in context:
 http://lucene.472066.n3.nabble.com/Solr-went-on-recovery-multiple-time-tp4196249p4196450.html
 Sent from the Solr - User mailing list archive at Nabble.com.




-- 
Bill Bell
billnb...@gmail.com
cell 720-256-8076


RE: Korean script conversion

2015-03-30 Thread Eyal Naamati
We only want the conversion Hanja-Hangul, for each Hanja character there 
exists only one Hangul character that can replace it in a Korean text.
The other way around is not convertible. 
We want to allow searching in both scripts and find matches in both scripts.
 Thanks

Eyal Naamati
Alma Developer
Tel: +972-2-6499313
Mobile: +972-547915255
eyal.naam...@exlibrisgroup.com

www.exlibrisgroup.com

-Original Message-
From: Benson Margulies [mailto:bimargul...@gmail.com] 
Sent: Monday, March 30, 2015 1:58 PM
To: solr-user
Subject: Re: Korean script conversion

Why do you think that this is a good idea? Hanja are used for special purposes; 
they are not trivally convertable to Hanjul due to ambiguity, and it's not at 
all clear that a typical search user wants to treat them as equivalent.

On Sun, Mar 29, 2015 at 1:52 AM, Eyal Naamati  eyal.naam...@exlibrisgroup.com 
wrote:

  Hi,



 We are starting to index records in Korean. Korean text can be written 
 in two scripts: Han characters (Chinese) and Hangul characters (Korean).

 We are looking for some solr filter or another built in solr component 
 that converts between Han and Hangul characters (transliteration).

 I know there is the ICUTransformFilterFactory that can convert between 
 Japanese or chinese scripts, for example:

 filter class=*solr.ICUTransformFilterFactory* id=*Katakana- 
 Hiragana* / for Japanese script conversions

 So far I couldn't find anything readymade for Korean scripts, but 
 perhaps someone knows of one?



 Thanks!

 Eyal Naamati
 Alma Developer
 Tel: +972-2-6499313
 Mobile: +972-547915255
 eyal.naam...@exlibrisgroup.com
 [image: Description: Description: Description: Description:
 C://signature/exlibris.jpg]
 www.exlibrisgroup.com





Re: Same schema.xml is loaded for different cores in SolrCloud

2015-03-30 Thread Erick Erickson
By now, I wouldn't particularly trust my setup. I'd blow it away and start over.

bootstrapping is _only_ required to get the configs up to Zookeeper
the first time. In fact I suggest you don't use it at all. Just start
SolrCloud, and use zkcli to push the configs up. Thereafter, create
your collections using the collections API.

Zookeeper is just a central repository for your configs and the
overall state of your cluster. As far as config sets are concerned,
think of the upconfig (or bootstrap) as copying the config files to a
place where they can be found by any random Solr instance that starts
up.

And the same applies to parameters like numShards. It's _only_ used as
a convenience for creating a cluster for demo purposes. Thereafter,
any time you start up that particular cloud, it'll read the old
cluster state and completely ignore the numShards parameter.

Rather than try to untangle what you've done, I'd re-install and work
through the tutorial step-by-step. I think you've jumped ahead and
gotten some things mixed up as far as your cluster state is concerned.

Best,
Erick

On Mon, Mar 30, 2015 at 6:52 PM, Zheng Lin Edwin Yeo
edwinye...@gmail.com wrote:
 Hi Erick,

 I've started shard2 with the following command instead, but it's still the
 same problem.
 java -DzkHost=localhost:9983 -Djetty.port=8984 -jar start.jar

 But you mean for shard1 we do not have to include 
 -Dbootstrap_confdir=./solr/logmill/conf for subsequent startup?

 Regards,
 Edwin


 On 31 March 2015 at 00:46, Erick Erickson erickerick...@gmail.com wrote:

 OK, this is a bit confused:

 1 You're starting two embedded Zookeepers but they don't know about
 each other. So looking for the configsets is a bit confused.
 2 There's no need to do the bootstrap thing after the first time. The
 _very_ first time you do this it pushes the configs up to Zookeeper,
 but after that you should just reference the config name.
 3 you specify the config name when you _create_ the collection, not
 when you start it up. You may be doing this, but your startup.
 4 I think you're confusing shards with collections. The equivalent of
 older-style cores would be just single-shard _collections_. Configs
 are associated on the collection level, not the shard level as all
 shards in a collection are presumed (indeed, _must_) use the same
 configuration.

 HTH,
 Erick

 On Mon, Mar 30, 2015 at 2:20 AM, Zheng Lin Edwin Yeo
 edwinye...@gmail.com wrote:
  I've roughly know what is the problem from here.
 
 http://stackoverflow.com/questions/23338324/zookeeper-multiple-collection-different-schema
 
  However, I couldn't find the zoo_data directory in all of my solr folder.
  What could be the problem or where is the directory supposed to be
 located?
 
  Regards,
  Edwin
 
 
  On 30 March 2015 at 11:56, Zheng Lin Edwin Yeo edwinye...@gmail.com
 wrote:
 
  Hi everyone,
 
  I've created a SolrCloud with multiple core, and I have different
  schema.xml for each of the core. However, when I start Solr, there's
 only
  one version of the schema.xml that is loaded onto Solr. Regardless of
 which
  core I go to, the schema.xml that is shown is the first one which I have
  loaded.
 
  What I did was, I have 3 cores: logmill, collection1 and collection2.
  Each of the core has 2 shrads: shard1 and shard2
 
  I first started the Solr with shard1 using the following command:
  java -Dcollection.configName=logmill -DzkRun -DnumShards=2
  -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar
 
  After that I start shard2 using the following command:
  java -Dcollection.configName=logmill -DzkRun -DnumShards=2
  -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar
 
  All the schema.xml loaded are from logmill core, even for the
 collection1
  and collection2.
 
  Even after I change the command to start shard1 with the following
  command, all the schema.xml are still from logmill
  java -Dcollection.configName=collection1 -DzkRun
  -DnumShards=2 -Dbootstrap_confdir=./solr/collection1/conf -jar start.jar
 
 
  How do I get Solr to read the different schema.xml for the different
 cores?
 
  Regards,
  Edwin