Re: Unable to perform search query after changing uniqueKey
Hi Zheng, It's possible that there's a problem with your schema.xml. Are all fields defined and have appropriate options enabled? Regards, Mostafa. On Mon, Mar 30, 2015 at 7:49 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I've tried that, and removed the data directory from both the shards. But the same problem still occurs, so we probably can rule out the memory issue. Regards, Edwin On 30 March 2015 at 12:39, Erick Erickson erickerick...@gmail.com wrote: I meant shut down Solr and physically remove the entire data directory. Not saying this is the cure, but it can't hurt to rule out the index having memory... Best, Erick On Sun, Mar 29, 2015 at 6:35 PM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I used the following query to delete all the index. http://localhost:8983/solr/update?stream.body= deletequery*:*/query/delete http://localhost:8983/solr/update?stream.body=commit/ Or is it better to physically delete the entire data directory? Regards, Edwin On 28 March 2015 at 02:27, Erick Erickson erickerick...@gmail.com wrote: You say you re-indexed, did you _completely_ remove the data directory first, i.e. the parent of the index and, maybe, tlog directories? I've occasionally seen remnants of old definitions pollute the new one, and since the uniqueKey key is so fundamental I can see it being a problem. Best, Erick On Fri, Mar 27, 2015 at 1:42 AM, Andrea Gazzarini a.gazzar...@gmail.com wrote: Hi Edwin, please provide some other detail about your context, (e.g. complete stacktrace, query you're issuing) Best, Andrea On 03/27/2015 09:38 AM, Zheng Lin Edwin Yeo wrote: Hi everyone, I've changed my uniqueKey to another name, instead of using id, on the schema.xml. However, after I have done the indexing (the indexing is successful), I'm not able to perform a search query on it. I gives the error java.lang.NullPointerException. Is there other place which I need to configure, besides changing the uniqueKey field in scheam.xml? Regards, Edwin
Solr went on recovery multiple time.
Solr server crashes 2-3 times a day . When i looked into log i can see some recovery failure. I got these errors on my solr log WARN - 2015-03-30 16:03:51.137; org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for zkNodeName=searcher:8980_solr_rn0core=rn0 ERROR - 2015-03-30 16:03:56.121; org.apache.solr.update.UpdateLog; Error inspecting tlog tlog{file=/solr/rn0/data/tlog/tlog.709 refcount=2} ERROR - 2015-03-30 16:03:56.121; org.apache.solr.update.UpdateLog; Error inspecting tlog tlog{file=/solr/rn0/data/tlog/tlog.710 refcount=2} WARN - 2015-03-30 16:04:00.242; org.apache.solr.core.SolrCore; [rn0] PERFORMANCE WARNING: Overlapping onDeckSearchers=2 I googled so many places, suggestions are related to my commit . My commit configuration on solrconfig.xml : updateHandler class=solr.DirectUpdateHandler2 autoCommit maxTime60/maxTime openSearcherfalse/openSearcher /autoCommit updateLog str name=dir${solr.data.dir:}/str /updateLog /updateHandler Why is it happening ? Any idea/solution will be appreciated. -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-went-on-recovery-multiple-time-tp4196249.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Restart solr failed after applied the patch in https://issues.apache.org/jira/browse/SOLR-6359
I doubt this has anything to do with the patch. Do you observe the same behaviour if you reduce the values for the config to defaults? (100, 10) On 30 Mar 2015 09:51, forest_soup tanglin0...@gmail.com wrote: https://issues.apache.org/jira/browse/SOLR-6359 I also posted the questions to the JIRA ticket. We have a SolrCloud with 5 solr servers of Solr 4.7.0. There are one collection with 80 shards(2 replicas per shard) on those 5 servers. And we made a patch by merge the patch (https://issues.apache.org/jira/secure/attachment/12702473/SOLR-6359.patch ) to 4.7.0 stream. And after applied the patch to our servers with the config changing uploaded to ZooKeeper, we did a restart on one of the 5 solr server, we met some issues on that server. Below is the details - The solrconfig.xml we changed: updateLog str name=dir$ {solr.ulog.dir:} /str int name=numRecordsToKeep1/int int name=maxNumLogsToKeep100/int /updateLog After we restarted one solr server without other 4 servers are running, we met below exceptions in the restarted one: ERROR - 2015-03-16 20:48:48.214; org.apache.solr.common.SolrException; org.apache.solr.common.SolrException: Exception writing document id Q049bGx0bWFpbDIxL089bGxwX3VzMQ==41703656!B68BF5EC5A4A650D85257E0A00724A3B to the index; possible analysis error. at org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:164) at org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:69) at org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:51) at org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:703) at org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:857) at org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:556) at org.apache.solr.handler.loader.JavabinLoader$1.update(JavabinLoader.java:96) at org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readOuterMostDocIterator(JavaBinUpdateRequestCodec.java:166) at org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readIterator(JavaBinUpdateRequestCodec.java:136) at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:225) at org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readNamedList(JavaBinUpdateRequestCodec.java:121) at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:190) at org.apache.solr.common.util.JavaBinCodec.unmarshal(JavaBinCodec.java:116) at org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec.unmarshal(JavaBinUpdateRequestCodec.java:173) at org.apache.solr.handler.loader.JavabinLoader.parseAndLoadDocs(JavabinLoader.java:106) at org.apache.solr.handler.loader.JavabinLoader.load(JavabinLoader.java:58) at org.apache.solr.handler.UpdateRequestHandler$1.load(UpdateRequestHandler.java:92) at org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:74) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1916) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:780) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:427) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:217) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:220) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:122) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:171) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:116) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:408) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1040) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:607) at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:314) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1156) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:626) at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) at java.lang.Thread.run(Thread.java:804) Caused by: org.apache.lucene.store.AlreadyClosedException: this IndexWriter is closed at
Re: Installing the auto-phrase-tokenfilter
Many thanks Andrea, I followed your guidance: the following site helped me to install and configure Ant in Windows http://madhukaudantha.blogspot.com.es/2010/06/installing-ant-for-windows-7.html Then after cloning the repository https://github.com/LucidWorks/auto-phrase-tokenfilter I just run the ant command and got the AutoPhraseTokenFilter.jar file needed. I will now try to make it work in Solr. -- View this message in context: http://lucene.472066.n3.nabble.com/Installing-the-auto-phrase-tokenfilter-tp4195466p4196241.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Same schema.xml is loaded for different cores in SolrCloud
I've roughly know what is the problem from here. http://stackoverflow.com/questions/23338324/zookeeper-multiple-collection-different-schema However, I couldn't find the zoo_data directory in all of my solr folder. What could be the problem or where is the directory supposed to be located? Regards, Edwin On 30 March 2015 at 11:56, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi everyone, I've created a SolrCloud with multiple core, and I have different schema.xml for each of the core. However, when I start Solr, there's only one version of the schema.xml that is loaded onto Solr. Regardless of which core I go to, the schema.xml that is shown is the first one which I have loaded. What I did was, I have 3 cores: logmill, collection1 and collection2. Each of the core has 2 shrads: shard1 and shard2 I first started the Solr with shard1 using the following command: java -Dcollection.configName=logmill -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar After that I start shard2 using the following command: java -Dcollection.configName=logmill -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar All the schema.xml loaded are from logmill core, even for the collection1 and collection2. Even after I change the command to start shard1 with the following command, all the schema.xml are still from logmill java -Dcollection.configName=collection1 -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/collection1/conf -jar start.jar How do I get Solr to read the different schema.xml for the different cores? Regards, Edwin
Re: Unable to perform search query after changing uniqueKey
Hi Mostafa, Yes, I've defined all the fields in schema.xml. It is able to work on the version without SolrCloud, but it is not working for the one with SolrCloud. Both of them are using the same schema.xml. Regards, Edwin On 30 March 2015 at 14:34, Mostafa Gomaa mostafa.goma...@gmail.com wrote: Hi Zheng, It's possible that there's a problem with your schema.xml. Are all fields defined and have appropriate options enabled? Regards, Mostafa. On Mon, Mar 30, 2015 at 7:49 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I've tried that, and removed the data directory from both the shards. But the same problem still occurs, so we probably can rule out the memory issue. Regards, Edwin On 30 March 2015 at 12:39, Erick Erickson erickerick...@gmail.com wrote: I meant shut down Solr and physically remove the entire data directory. Not saying this is the cure, but it can't hurt to rule out the index having memory... Best, Erick On Sun, Mar 29, 2015 at 6:35 PM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I used the following query to delete all the index. http://localhost:8983/solr/update?stream.body= deletequery*:*/query/delete http://localhost:8983/solr/update?stream.body=commit/ Or is it better to physically delete the entire data directory? Regards, Edwin On 28 March 2015 at 02:27, Erick Erickson erickerick...@gmail.com wrote: You say you re-indexed, did you _completely_ remove the data directory first, i.e. the parent of the index and, maybe, tlog directories? I've occasionally seen remnants of old definitions pollute the new one, and since the uniqueKey key is so fundamental I can see it being a problem. Best, Erick On Fri, Mar 27, 2015 at 1:42 AM, Andrea Gazzarini a.gazzar...@gmail.com wrote: Hi Edwin, please provide some other detail about your context, (e.g. complete stacktrace, query you're issuing) Best, Andrea On 03/27/2015 09:38 AM, Zheng Lin Edwin Yeo wrote: Hi everyone, I've changed my uniqueKey to another name, instead of using id, on the schema.xml. However, after I have done the indexing (the indexing is successful), I'm not able to perform a search query on it. I gives the error java.lang.NullPointerException. Is there other place which I need to configure, besides changing the uniqueKey field in scheam.xml? Regards, Edwin
Restart solr failed after applied the patch in https://issues.apache.org/jira/browse/SOLR-6359
https://issues.apache.org/jira/browse/SOLR-6359 I also posted the questions to the JIRA ticket. We have a SolrCloud with 5 solr servers of Solr 4.7.0. There are one collection with 80 shards(2 replicas per shard) on those 5 servers. And we made a patch by merge the patch (https://issues.apache.org/jira/secure/attachment/12702473/SOLR-6359.patch) to 4.7.0 stream. And after applied the patch to our servers with the config changing uploaded to ZooKeeper, we did a restart on one of the 5 solr server, we met some issues on that server. Below is the details - The solrconfig.xml we changed: updateLog str name=dir$ {solr.ulog.dir:} /str int name=numRecordsToKeep1/int int name=maxNumLogsToKeep100/int /updateLog After we restarted one solr server without other 4 servers are running, we met below exceptions in the restarted one: ERROR - 2015-03-16 20:48:48.214; org.apache.solr.common.SolrException; org.apache.solr.common.SolrException: Exception writing document id Q049bGx0bWFpbDIxL089bGxwX3VzMQ==41703656!B68BF5EC5A4A650D85257E0A00724A3B to the index; possible analysis error. at org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:164) at org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:69) at org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:51) at org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:703) at org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:857) at org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:556) at org.apache.solr.handler.loader.JavabinLoader$1.update(JavabinLoader.java:96) at org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readOuterMostDocIterator(JavaBinUpdateRequestCodec.java:166) at org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readIterator(JavaBinUpdateRequestCodec.java:136) at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:225) at org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readNamedList(JavaBinUpdateRequestCodec.java:121) at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:190) at org.apache.solr.common.util.JavaBinCodec.unmarshal(JavaBinCodec.java:116) at org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec.unmarshal(JavaBinUpdateRequestCodec.java:173) at org.apache.solr.handler.loader.JavabinLoader.parseAndLoadDocs(JavabinLoader.java:106) at org.apache.solr.handler.loader.JavabinLoader.load(JavabinLoader.java:58) at org.apache.solr.handler.UpdateRequestHandler$1.load(UpdateRequestHandler.java:92) at org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:74) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1916) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:780) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:427) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:217) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:220) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:122) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:171) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:116) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:408) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1040) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:607) at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:314) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1156) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:626) at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) at java.lang.Thread.run(Thread.java:804) Caused by: org.apache.lucene.store.AlreadyClosedException: this IndexWriter is closed at org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:645) at org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:659) at org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:1525) at org.apache.solr.update.DirectUpdateHandler2.addDoc0(DirectUpdateHandler2.java:236) at org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:160) ... 37
Date Time datatypes?
Howdy folks, Is there any way index only the date and time portions of a datetime field? A Date is really a period of 24hrs, starting at 00:00 in said date's time zone. It would be useful if there was a way to search for documents of a certain date with these semantics. As for times, I'd like to be able to do queries like time:[17:00 TO 18:00]. I suppose I could accomplish that by resetting the date portion to some bogus value, but then my facet/range values will contain that bogus date. I suppose the alternative is to create my own data types. Extending PrimitiveFieldType doesn't seem too hairy but maybe I'm overlooking some of the complexity... Thanks a bunch, - Bram
Re: Optimize SolrCloud without downtime
Hi, thanks for reply. We have a lot of deleted docs cause we have to reindex all records from time to time, changing some important parameters.. When we do update, it means create and delete. Our deleted docs do not disappear by merging segments. I see our deleted docs are almost the same number after two weeks and always around 40 segments. I ask our developers and they confirmed we use default settings for merging policy. Please, do you have some tips and tricks for efficient setting for merging segments? Our HW is RAID5 SAS disks and 40GB of memory dedicated for Solr. Regards, Pavel -- View this message in context: http://lucene.472066.n3.nabble.com/Optimize-SolrCloud-without-downtime-tp4195170p4196273.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Restart solr failed after applied the patch in https://issues.apache.org/jira/browse/SOLR-6359
Yes, I also doubt the patch. I restore the patch with original .jar file, there is no that issue. -- View this message in context: http://lucene.472066.n3.nabble.com/Restart-solr-failed-after-applied-the-patch-in-https-issues-apache-org-jira-browse-SOLR-6359-tp4196251p4196278.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Korean script conversion
Why do you think that this is a good idea? Hanja are used for special purposes; they are not trivally convertable to Hanjul due to ambiguity, and it's not at all clear that a typical search user wants to treat them as equivalent. On Sun, Mar 29, 2015 at 1:52 AM, Eyal Naamati eyal.naam...@exlibrisgroup.com wrote: Hi, We are starting to index records in Korean. Korean text can be written in two scripts: Han characters (Chinese) and Hangul characters (Korean). We are looking for some solr filter or another built in solr component that converts between Han and Hangul characters (transliteration). I know there is the ICUTransformFilterFactory that can convert between Japanese or chinese scripts, for example: filter class=*solr.ICUTransformFilterFactory* id=*Katakana- Hiragana* / for Japanese script conversions So far I couldn't find anything readymade for Korean scripts, but perhaps someone knows of one? Thanks! Eyal Naamati Alma Developer Tel: +972-2-6499313 Mobile: +972-547915255 eyal.naam...@exlibrisgroup.com [image: Description: Description: Description: Description: C://signature/exlibris.jpg] www.exlibrisgroup.com
Re: Restart solr failed after applied the patch in https://issues.apache.org/jira/browse/SOLR-6359
But if the value can only be 100,10, is there any difference with no that patch? Can we enlarge those 2 values? Thanks! -- View this message in context: http://lucene.472066.n3.nabble.com/Restart-solr-failed-after-applied-the-patch-in-https-issues-apache-org-jira-browse-SOLR-6359-tp4196251p4196280.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SOLR Index in shared/Network folder
First examine whether you can reduce the amount of data you keep around, field norms, stored fields, etc. Here's a place to start: http://stackoverflow.com/questions/10080881/solr-index-size-reduction I have heard of people doing what you suggest, but be _very_ careful that you don't accidentally write to the index from the slave. I always have to ask. You're talking about reducing disk space because you have 100's of G indexes. Is it really worth the engineering effort to save the disk space? Terabyte hard drives are $100. Buy two for the cost of untangling a single corrupt index because someone (and someone will) caused the slaves to write to the index rather than the master. FWIW, Erick On Sun, Mar 29, 2015 at 10:42 PM, abhi Abhishek abhi26...@gmail.com wrote: Hello, Thanks for the suggestions. My aim is to reduce the disk space usage. I have 1 master with 2 slave configured, where slaves are used for searching and master ingests new data replicated to slaves, but as my index size is in 100's of GB we see 3x times space overhead. i would like to reduce this overhead, can you suggest something for this? Thanks in Advance Best Regards, Abhishek On Sat, Mar 28, 2015 at 12:13 AM, Erick Erickson erickerick...@gmail.com wrote: To pile on: If you're talking about pointing two Solr instances at the _same_ index, it doesn't matter whether you are on NFS or not, you'll have all sorts of problems. And if this is a SolrCloud installation, it's particularly hard to get right. Please do not do this unless you have a very good reason, and please tell us what the reason is so we can perhaps suggest alternatives. Best, Erick On Fri, Mar 27, 2015 at 8:08 AM, Walter Underwood wun...@wunderwood.org wrote: Several years ago, I accidentally put Solr indexes on an NFS volume and it was 100X slower. If you have enough RAM, query speed should be OK, but startup time (loading indexes into file buffers) could be really long. Indexing could be quite slow. wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ (my blog) On Mar 26, 2015, at 11:31 PM, Shawn Heisey apa...@elyograg.org wrote: On 3/27/2015 12:06 AM, abhi Abhishek wrote: Greetings, I am trying to use a network shared location as my index directory. are there any known problems in using a Network File System for running a SOLR Instance? It is not recommended. You will probably need to change the lockType, ... the default native probably will not work, and you might need to change it to none to get it working ... but that disables an important safety mechanism that prevents index corruption. http://stackoverflow.com/questions/9599529/solr-over-nfs-problems Thanks, Shawn
Re: Solr went on recovery multiple time.
Sometimes happens when your zookeeper timeout is short (the 15 second timeout has been too short) and you hit a GC pause that is longer. Try lengthening your ZK timeout to 45 or 60 seconds. It would help if you told us about your setup, memory allocated to Solr, all that stuff, see: http://wiki.apache.org/solr/UsingMailingLists Best, Erick On Mon, Mar 30, 2015 at 1:43 AM, sthita sthit...@gmail.com wrote: Solr server crashes 2-3 times a day . When i looked into log i can see some recovery failure. I got these errors on my solr log WARN - 2015-03-30 16:03:51.137; org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for zkNodeName=searcher:8980_solr_rn0core=rn0 ERROR - 2015-03-30 16:03:56.121; org.apache.solr.update.UpdateLog; Error inspecting tlog tlog{file=/solr/rn0/data/tlog/tlog.709 refcount=2} ERROR - 2015-03-30 16:03:56.121; org.apache.solr.update.UpdateLog; Error inspecting tlog tlog{file=/solr/rn0/data/tlog/tlog.710 refcount=2} WARN - 2015-03-30 16:04:00.242; org.apache.solr.core.SolrCore; [rn0] PERFORMANCE WARNING: Overlapping onDeckSearchers=2 I googled so many places, suggestions are related to my commit . My commit configuration on solrconfig.xml : updateHandler class=solr.DirectUpdateHandler2 autoCommit maxTime60/maxTime openSearcherfalse/openSearcher /autoCommit updateLog str name=dir${solr.data.dir:}/str /updateLog /updateHandler Why is it happening ? Any idea/solution will be appreciated. -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-went-on-recovery-multiple-time-tp4196249.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Unable to perform search query after changing uniqueKey
Well, let's see the definition of your ID field, 'cause I'm puzzled. It's definitely A Bad Thing to have it be any kind of tokenized field though, but that's a shot in the dark. Best, Erick On Mon, Mar 30, 2015 at 2:17 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Mostafa, Yes, I've defined all the fields in schema.xml. It is able to work on the version without SolrCloud, but it is not working for the one with SolrCloud. Both of them are using the same schema.xml. Regards, Edwin On 30 March 2015 at 14:34, Mostafa Gomaa mostafa.goma...@gmail.com wrote: Hi Zheng, It's possible that there's a problem with your schema.xml. Are all fields defined and have appropriate options enabled? Regards, Mostafa. On Mon, Mar 30, 2015 at 7:49 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I've tried that, and removed the data directory from both the shards. But the same problem still occurs, so we probably can rule out the memory issue. Regards, Edwin On 30 March 2015 at 12:39, Erick Erickson erickerick...@gmail.com wrote: I meant shut down Solr and physically remove the entire data directory. Not saying this is the cure, but it can't hurt to rule out the index having memory... Best, Erick On Sun, Mar 29, 2015 at 6:35 PM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I used the following query to delete all the index. http://localhost:8983/solr/update?stream.body= deletequery*:*/query/delete http://localhost:8983/solr/update?stream.body=commit/ Or is it better to physically delete the entire data directory? Regards, Edwin On 28 March 2015 at 02:27, Erick Erickson erickerick...@gmail.com wrote: You say you re-indexed, did you _completely_ remove the data directory first, i.e. the parent of the index and, maybe, tlog directories? I've occasionally seen remnants of old definitions pollute the new one, and since the uniqueKey key is so fundamental I can see it being a problem. Best, Erick On Fri, Mar 27, 2015 at 1:42 AM, Andrea Gazzarini a.gazzar...@gmail.com wrote: Hi Edwin, please provide some other detail about your context, (e.g. complete stacktrace, query you're issuing) Best, Andrea On 03/27/2015 09:38 AM, Zheng Lin Edwin Yeo wrote: Hi everyone, I've changed my uniqueKey to another name, instead of using id, on the schema.xml. However, after I have done the indexing (the indexing is successful), I'm not able to perform a search query on it. I gives the error java.lang.NullPointerException. Is there other place which I need to configure, besides changing the uniqueKey field in scheam.xml? Regards, Edwin
Re: SOLR Index in shared/Network folder
I suggest that you do not try to save money on disk space. Disk is cheap. You will spend weeks of expensive engineering time trying to make this work. Once you make it work, it will be slow an unreliable. 300GB Amazon EBS volumes are $180/year, $360/year for SSD. Just spend the money. wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ (my blog) On Mar 29, 2015, at 10:42 PM, abhi Abhishek abhi26...@gmail.com wrote: Hello, Thanks for the suggestions. My aim is to reduce the disk space usage. I have 1 master with 2 slave configured, where slaves are used for searching and master ingests new data replicated to slaves, but as my index size is in 100's of GB we see 3x times space overhead. i would like to reduce this overhead, can you suggest something for this? Thanks in Advance Best Regards, Abhishek On Sat, Mar 28, 2015 at 12:13 AM, Erick Erickson erickerick...@gmail.com wrote: To pile on: If you're talking about pointing two Solr instances at the _same_ index, it doesn't matter whether you are on NFS or not, you'll have all sorts of problems. And if this is a SolrCloud installation, it's particularly hard to get right. Please do not do this unless you have a very good reason, and please tell us what the reason is so we can perhaps suggest alternatives. Best, Erick On Fri, Mar 27, 2015 at 8:08 AM, Walter Underwood wun...@wunderwood.org wrote: Several years ago, I accidentally put Solr indexes on an NFS volume and it was 100X slower. If you have enough RAM, query speed should be OK, but startup time (loading indexes into file buffers) could be really long. Indexing could be quite slow. wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ (my blog) On Mar 26, 2015, at 11:31 PM, Shawn Heisey apa...@elyograg.org wrote: On 3/27/2015 12:06 AM, abhi Abhishek wrote: Greetings, I am trying to use a network shared location as my index directory. are there any known problems in using a Network File System for running a SOLR Instance? It is not recommended. You will probably need to change the lockType, ... the default native probably will not work, and you might need to change it to none to get it working ... but that disables an important safety mechanism that prevents index corruption. http://stackoverflow.com/questions/9599529/solr-over-nfs-problems Thanks, Shawn
Re: Date Time datatypes?
Hmmm, not OOB. You could index two tdate fields though and in the second one only index seconds into the day and index the other with date math (i.e. /DAY) to normalize them. Or a tdate field with just a long field for the seconds. You could then use the range faceting to, say, put the docs into hour buckets, and your time:[17:00 TO 18:00] would turn into something like time:[61200 TO 64800]. A bit awkward I'll admit, but probably easier than a custom field type. Or add a field for hours if that fits the use-case. FWIW, Erick On Mon, Mar 30, 2015 at 4:11 AM, Bram Van Dam bram.van...@intix.eu wrote: Howdy folks, Is there any way index only the date and time portions of a datetime field? A Date is really a period of 24hrs, starting at 00:00 in said date's time zone. It would be useful if there was a way to search for documents of a certain date with these semantics. As for times, I'd like to be able to do queries like time:[17:00 TO 18:00]. I suppose I could accomplish that by resetting the date portion to some bogus value, but then my facet/range values will contain that bogus date. I suppose the alternative is to create my own data types. Extending PrimitiveFieldType doesn't seem too hairy but maybe I'm overlooking some of the complexity... Thanks a bunch, - Bram
protected phrases - possible?
Hi, The way our collection is setup, searches for breast cancer are returning results for ovarian cancer, or anything that contains either breast or cancer. The reason is, we are searching across multiple fields. Even though I have set a mm value so that if less than 3 terms, ALL terms much match...SOLR considers it all matched even though breast was in the title and cancer is in the description. Is there a way to protect certain phrases so that they will not be tokenized? I tried using CommonGramsFilterFactory, but having breast cancer in the word list did not seem to do anything. I'm guessing it's because the field is tokenized first, so nothing would match that phrase. If I put breast and cancer as separate entries in the word list, I end up with too many unnecessary shingles, and breast and cancer are still two of the final terms. I have a feeling CommonGramsFilterFactory is not the right way to handle this. What are other options? Is it better to put all fields in one field, apply mm, and proximity boost? Thanks! Jing
Re: Same schema.xml is loaded for different cores in SolrCloud
OK, this is a bit confused: 1 You're starting two embedded Zookeepers but they don't know about each other. So looking for the configsets is a bit confused. 2 There's no need to do the bootstrap thing after the first time. The _very_ first time you do this it pushes the configs up to Zookeeper, but after that you should just reference the config name. 3 you specify the config name when you _create_ the collection, not when you start it up. You may be doing this, but your startup. 4 I think you're confusing shards with collections. The equivalent of older-style cores would be just single-shard _collections_. Configs are associated on the collection level, not the shard level as all shards in a collection are presumed (indeed, _must_) use the same configuration. HTH, Erick On Mon, Mar 30, 2015 at 2:20 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: I've roughly know what is the problem from here. http://stackoverflow.com/questions/23338324/zookeeper-multiple-collection-different-schema However, I couldn't find the zoo_data directory in all of my solr folder. What could be the problem or where is the directory supposed to be located? Regards, Edwin On 30 March 2015 at 11:56, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi everyone, I've created a SolrCloud with multiple core, and I have different schema.xml for each of the core. However, when I start Solr, there's only one version of the schema.xml that is loaded onto Solr. Regardless of which core I go to, the schema.xml that is shown is the first one which I have loaded. What I did was, I have 3 cores: logmill, collection1 and collection2. Each of the core has 2 shrads: shard1 and shard2 I first started the Solr with shard1 using the following command: java -Dcollection.configName=logmill -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar After that I start shard2 using the following command: java -Dcollection.configName=logmill -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar All the schema.xml loaded are from logmill core, even for the collection1 and collection2. Even after I change the command to start shard1 with the following command, all the schema.xml are still from logmill java -Dcollection.configName=collection1 -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/collection1/conf -jar start.jar How do I get Solr to read the different schema.xml for the different cores? Regards, Edwin
RE: Structured and Unstructured data indexing in SolrCloud
Hi Vijay, The short answer is yes, you can combine almost anything you want into a single collection. But, in addition to working out your queries, you might want work out your data life cycle. In our application, we have comingled the structured and unstructured documents into a single collection for initial development purposes. The only field they have in common is the unique ID.Works fine. In production, however, we see things like query rates, access controls, load balancing, availability, shard keys, overall document counts, update frequency, etc. will drive us to use separate collections. For us, the deciding factor is less about structured vs. unstructured and more about public vs. private. We have developed our app so that splitting the collection will have minimal impact by executing separate queries, in parallel, at runtime. Of course, your application is different. YMMV, etc. hth, Charlie -Original Message- From: Jack Krupansky [mailto:jack.krupan...@gmail.com] Sent: Sunday, March 29, 2015 4:26 PM To: solr-user@lucene.apache.org Subject: Re: Structured and Unstructured data indexing in SolrCloud The first step is to work out the queries that you wish to perform - that will determine how the data should be organized in the Solr schema. -- Jack Krupansky On Sun, Mar 29, 2015 at 4:04 PM, Vijay Bhoomireddy vijaya.bhoomire...@whishworks.com wrote: Hi, We have a requirement where both structured and unstructured data comes into the system. We need to index both of them and then enable search functionality on it. We are using SolrCloud on Hadoop platform. For structured data, we are planning to put the data into HBase and for unstructured, directly into HDFS. My question is how to index these sources under a single Solr core? Would that be possible to index both structured and unstructured data under a single core/collection in SolrCloud and then enable search functionality over that index? Thanks in advance. -- The contents of this e-mail are confidential and for the exclusive use of the intended recipient. If you receive this e-mail in error please delete it from your system immediately and notify us either by e-mail or telephone. You should not copy, forward or otherwise disclose the content of the e-mail. The views expressed in this communication may not necessarily be the view held by WHISHWORKS. * This e-mail may contain confidential or privileged information. If you are not the intended recipient, please notify the sender immediately and then delete it. TIAA-CREF *
Re: protected phrases - possible?
Hi Jing, You can boost phrases by pf (phrase fields) parameter. If you don't like this solution, you can modify search query at client side. E.g. surround certain phrases with quotes. This will force proximity search without interfering with tokenisation. Ahmet On Monday, March 30, 2015 8:49 PM, Tao, Jing j...@webmd.net wrote: Hi, The way our collection is setup, searches for breast cancer are returning results for ovarian cancer, or anything that contains either breast or cancer. The reason is, we are searching across multiple fields. Even though I have set a mm value so that if less than 3 terms, ALL terms much match...SOLR considers it all matched even though breast was in the title and cancer is in the description. Is there a way to protect certain phrases so that they will not be tokenized? I tried using CommonGramsFilterFactory, but having breast cancer in the word list did not seem to do anything. I'm guessing it's because the field is tokenized first, so nothing would match that phrase. If I put breast and cancer as separate entries in the word list, I end up with too many unnecessary shingles, and breast and cancer are still two of the final terms. I have a feeling CommonGramsFilterFactory is not the right way to handle this. What are other options? Is it better to put all fields in one field, apply mm, and proximity boost? Thanks! Jing
NoNode for /clusterstate.json in solr5.0.0 cloud
I have a basic Solr 5.0.0 cloud setup after following http://lucene.apache.org/solr/quickstart.html I am trying to read data from spark and index it into solr using following lib: https://github.com/LucidWorks/spark-solr I am getting following error when my code try to make request to solr Exception in thread main org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465) . .. .. Caused by: org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /clusterstate.json at org.apache.zookeeper.KeeperException.create(KeeperException.java:111) at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783) at org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500) I am not sure how (and when) to create nodes for /clusterstate.json I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar Thanks for the help. Sumit Purohit
Re: NoNode for /clusterstate.json in solr5.0.0 cloud
Anything in the server-side Solr logs? Also, if you go to the Solr admin console at http://localhost:8983/solr, do you see the gettingstarted collection in the cloud panel? On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: I have a basic Solr 5.0.0 cloud setup after following http://lucene.apache.org/solr/quickstart.html I am trying to read data from spark and index it into solr using following lib: https://github.com/LucidWorks/spark-solr I am getting following error when my code try to make request to solr Exception in thread main org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465) . .. .. Caused by: org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /clusterstate.json at org.apache.zookeeper.KeeperException.create(KeeperException.java:111) at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783) at org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500) I am not sure how (and when) to create nodes for /clusterstate.json I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar Thanks for the help. Sumit Purohit
Re: Optimize SolrCloud without downtime
Hmmm, are you indexing during the time you see the deleted docs not changing? Because this is very strange. Theoretically, if you reindex everything, that should result in segments that have _no_ live docs in them and they should really disappear ASAP. One way to work around this if we determine that this is a problem in Solr might be to create a second collection and index to _that_, then use collection aliasing to use the new index. But that's not optimal... Best, Erick On Mon, Mar 30, 2015 at 4:35 AM, Pavel Hladik pavel.hla...@profimedia.cz wrote: Hi, thanks for reply. We have a lot of deleted docs cause we have to reindex all records from time to time, changing some important parameters.. When we do update, it means create and delete. Our deleted docs do not disappear by merging segments. I see our deleted docs are almost the same number after two weeks and always around 40 segments. I ask our developers and they confirmed we use default settings for merging policy. Please, do you have some tips and tricks for efficient setting for merging segments? Our HW is RAID5 SAS disks and 40GB of memory dedicated for Solr. Regards, Pavel -- View this message in context: http://lucene.472066.n3.nabble.com/Optimize-SolrCloud-without-downtime-tp4195170p4196273.html Sent from the Solr - User mailing list archive at Nabble.com.
how do you replicate solr-cloud between datacenters?
Can you use /replication ??? How would you do this between datacenters? -- Tim Ehlers
RE: NoNode for /clusterstate.json in solr5.0.0 cloud
Thanks Tim, i had to make some changes in my local spark-solr clone to build it for sorl5. If its ok, i can commit these to github. thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:27 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud Ok, let me upgrade my version of spark-solr to 5 to see what I get ... On Mon, Mar 30, 2015 at 2:26 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: yes there is getting started collection.. and on admin webpage console--cloud---tree---/clusterstate.json shows me this table version =1 aversion=0 children_count=0 ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901) cversion=0 czxid=32 ephemeralOwner=0 mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453) mzxid=110 pzxid=32 dataLength=2 children_count=0 seems related to no node error. thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:18 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud Anything in the server-side Solr logs? Also, if you go to the Solr admin console at http://localhost:8983/solr, do you see the gettingstarted collection in the cloud panel? On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: I have a basic Solr 5.0.0 cloud setup after following http://lucene.apache.org/solr/quickstart.html I am trying to read data from spark and index it into solr using following lib: https://github.com/LucidWorks/spark-solr I am getting following error when my code try to make request to solr Exception in thread main org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465) . .. .. Caused by: org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /clusterstate.json at org.apache.zookeeper.KeeperException.create(KeeperException.java:111) at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783) at org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500) I am not sure how (and when) to create nodes for /clusterstate.json I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar Thanks for the help. Sumit Purohit
RE: NoNode for /clusterstate.json in solr5.0.0 cloud
Yes i am using default localhost:9983 and i think my program does connect with zookeeper as per logs : 2015-03-30 15:00:51,512 [Executor task launch worker-0] INFO SolrZkClient - Using default ZkCredentialsProvider 2015-03-30 15:00:51,530 [Executor task launch worker-0] INFO ConnectionManager - Waiting for client to connect to ZooKeeper 2015-03-30 15:00:51,544 [zkCallback-3-thread-1] INFO ConnectionManager - Watcher org.apache.solr.common.cloud.ConnectionManager@51ffb98d name:ZooKeeperConnection Watcher:localhost:9983/sorl got event WatchedEvent state:SyncConnected type:None path:null path:null type:None 2015-03-30 15:00:51,544 [Executor task launch worker-0] INFO ConnectionManager - Client is connected to ZooKeeper 2015-03-30 15:00:51,544 [Executor task launch worker-0] INFO SolrZkClient - Using default ZkACLProvider 2015-03-30 15:00:51,548 [Executor task launch worker-0] INFO SolrZkClient - makePath: /clusterstate.json 2015-03-30 15:00:51,557 [Executor task launch worker-0] ERROR Executor - Exception in task 0.0 in stage 0.0 (TID 0) So to upgrade spark-solr project, i had to change all the references to of class xxxSolrServer to xxxSolrClient . did you do something similar. If you plan to push it to github, i can grab that and compare with my environment. Thanks for you prompt response(s). thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:54 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud I upgraded the spark-solr project to solrj-5.0.0 and was able to index into the gettingstarted collection using Solr 5.0.0, so seems like it may be environmental. Almost seems like the spark project is looking at the wrong ZooKeeper? Are you using the default -zkHost localhost:9983 On Mon, Mar 30, 2015 at 2:32 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: Thanks Tim, i had to make some changes in my local spark-solr clone to build it for sorl5. If its ok, i can commit these to github. thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:27 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud Ok, let me upgrade my version of spark-solr to 5 to see what I get ... On Mon, Mar 30, 2015 at 2:26 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: yes there is getting started collection.. and on admin webpage console--cloud---tree---/clusterstate.json shows me this table version =1 aversion=0 children_count=0 ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901) cversion=0 czxid=32 ephemeralOwner=0 mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453) mzxid=110 pzxid=32 dataLength=2 children_count=0 seems related to no node error. thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:18 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud Anything in the server-side Solr logs? Also, if you go to the Solr admin console at http://localhost:8983/solr, do you see the gettingstarted collection in the cloud panel? On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: I have a basic Solr 5.0.0 cloud setup after following http://lucene.apache.org/solr/quickstart.html I am trying to read data from spark and index it into solr using following lib: https://github.com/LucidWorks/spark-solr I am getting following error when my code try to make request to solr Exception in thread main org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465) . .. .. Caused by: org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /clusterstate.json at org.apache.zookeeper.KeeperException.create(KeeperException.java:111) at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783) at org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500) I am not sure how (and when) to create nodes for /clusterstate.json I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar Thanks for the help. Sumit Purohit
Re: NoNode for /clusterstate.json in solr5.0.0 cloud
Ok, let me upgrade my version of spark-solr to 5 to see what I get ... On Mon, Mar 30, 2015 at 2:26 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: yes there is getting started collection.. and on admin webpage console--cloud---tree---/clusterstate.json shows me this table version =1 aversion=0 children_count=0 ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901) cversion=0 czxid=32 ephemeralOwner=0 mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453) mzxid=110 pzxid=32 dataLength=2 children_count=0 seems related to no node error. thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:18 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud Anything in the server-side Solr logs? Also, if you go to the Solr admin console at http://localhost:8983/solr, do you see the gettingstarted collection in the cloud panel? On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: I have a basic Solr 5.0.0 cloud setup after following http://lucene.apache.org/solr/quickstart.html I am trying to read data from spark and index it into solr using following lib: https://github.com/LucidWorks/spark-solr I am getting following error when my code try to make request to solr Exception in thread main org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465) . .. .. Caused by: org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /clusterstate.json at org.apache.zookeeper.KeeperException.create(KeeperException.java:111) at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783) at org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500) I am not sure how (and when) to create nodes for /clusterstate.json I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar Thanks for the help. Sumit Purohit
Re: how do you replicate solr-cloud between datacenters?
That's an open issue. See: https://issues.apache.org/jira/browse/SOLR-6273 -- Jack Krupansky On Mon, Mar 30, 2015 at 5:45 PM, Timothy Ehlers ehle...@gmail.com wrote: Can you use /replication ??? How would you do this between datacenters? -- Tim Ehlers
RE: NoNode for /clusterstate.json in solr5.0.0 cloud
yes there is getting started collection.. and on admin webpage console--cloud---tree---/clusterstate.json shows me this table version =1 aversion=0 children_count=0 ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901) cversion=0 czxid=32 ephemeralOwner=0 mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453) mzxid=110 pzxid=32 dataLength=2 children_count=0 seems related to no node error. thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:18 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud Anything in the server-side Solr logs? Also, if you go to the Solr admin console at http://localhost:8983/solr, do you see the gettingstarted collection in the cloud panel? On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: I have a basic Solr 5.0.0 cloud setup after following http://lucene.apache.org/solr/quickstart.html I am trying to read data from spark and index it into solr using following lib: https://github.com/LucidWorks/spark-solr I am getting following error when my code try to make request to solr Exception in thread main org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465) . .. .. Caused by: org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /clusterstate.json at org.apache.zookeeper.KeeperException.create(KeeperException.java:111) at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783) at org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500) I am not sure how (and when) to create nodes for /clusterstate.json I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar Thanks for the help. Sumit Purohit
Re: NoNode for /clusterstate.json in solr5.0.0 cloud
I upgraded the spark-solr project to solrj-5.0.0 and was able to index into the gettingstarted collection using Solr 5.0.0, so seems like it may be environmental. Almost seems like the spark project is looking at the wrong ZooKeeper? Are you using the default -zkHost localhost:9983 On Mon, Mar 30, 2015 at 2:32 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: Thanks Tim, i had to make some changes in my local spark-solr clone to build it for sorl5. If its ok, i can commit these to github. thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:27 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud Ok, let me upgrade my version of spark-solr to 5 to see what I get ... On Mon, Mar 30, 2015 at 2:26 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: yes there is getting started collection.. and on admin webpage console--cloud---tree---/clusterstate.json shows me this table version =1 aversion=0 children_count=0 ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901) cversion=0 czxid=32 ephemeralOwner=0 mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453) mzxid=110 pzxid=32 dataLength=2 children_count=0 seems related to no node error. thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:18 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud Anything in the server-side Solr logs? Also, if you go to the Solr admin console at http://localhost:8983/solr, do you see the gettingstarted collection in the cloud panel? On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: I have a basic Solr 5.0.0 cloud setup after following http://lucene.apache.org/solr/quickstart.html I am trying to read data from spark and index it into solr using following lib: https://github.com/LucidWorks/spark-solr I am getting following error when my code try to make request to solr Exception in thread main org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465) . .. .. Caused by: org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /clusterstate.json at org.apache.zookeeper.KeeperException.create(KeeperException.java:111) at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783) at org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500) I am not sure how (and when) to create nodes for /clusterstate.json I am using solr 5.0.0, sorlj5.0.0 spark-core_2.10_2.12.jar Thanks for the help. Sumit Purohit
Re: Solr went on recovery multiple time.
Thanks Erick for the reply. These are the configurations. 1.Solr servers count : 4 (1 master , 3 slaves) 2.My initial and maximum memory allocations to each solr servers are : -Xms5g -Xmx6g 3. 5 running zookeepers 4. My index size in each solr server around: 48GB 5. Every sec merger add/delete 500 docs to/from solr yes, i checked the the ZK config timeout is 10sec. I will increase the timeout to 45sec, Hope that will help. ZK config : # The number of milliseconds of each tick tickTime=2000 # The number of ticks that the initial # synchronization phase can take initLimit=10 # The number of ticks that can pass between # sending a request and getting an acknowledgement syncLimit=5 -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-went-on-recovery-multiple-time-tp4196249p4196450.html Sent from the Solr - User mailing list archive at Nabble.com.
Collapse and Expand behaviour on result with 1 document.
Hi I have a query which return 1 document. When I add the collapse and expand parameters to it, expand=trueexpand.rows=5fq={!collapse%20field=P_SupplierId}, the expanded section is empty (lst name=expanded/). Is this the behaviour of collapse and expand parameters on result which contain only 1 document? -Derek
Re: Unable to perform search query after changing uniqueKey
Hi Erick, I've changed the uniqueKey from id to Item No. uniqueKeyItem No/uniqueKey Below are my definitions for both the id and Item No. field name=id type=string indexed=true stored=true required=false multiValued=false / field name=Item No type=text_general indexed=true stored=true/ Regards, Edwin On 30 March 2015 at 23:05, Erick Erickson erickerick...@gmail.com wrote: Well, let's see the definition of your ID field, 'cause I'm puzzled. It's definitely A Bad Thing to have it be any kind of tokenized field though, but that's a shot in the dark. Best, Erick On Mon, Mar 30, 2015 at 2:17 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Mostafa, Yes, I've defined all the fields in schema.xml. It is able to work on the version without SolrCloud, but it is not working for the one with SolrCloud. Both of them are using the same schema.xml. Regards, Edwin On 30 March 2015 at 14:34, Mostafa Gomaa mostafa.goma...@gmail.com wrote: Hi Zheng, It's possible that there's a problem with your schema.xml. Are all fields defined and have appropriate options enabled? Regards, Mostafa. On Mon, Mar 30, 2015 at 7:49 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I've tried that, and removed the data directory from both the shards. But the same problem still occurs, so we probably can rule out the memory issue. Regards, Edwin On 30 March 2015 at 12:39, Erick Erickson erickerick...@gmail.com wrote: I meant shut down Solr and physically remove the entire data directory. Not saying this is the cure, but it can't hurt to rule out the index having memory... Best, Erick On Sun, Mar 29, 2015 at 6:35 PM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I used the following query to delete all the index. http://localhost:8983/solr/update?stream.body= deletequery*:*/query/delete http://localhost:8983/solr/update?stream.body=commit/ Or is it better to physically delete the entire data directory? Regards, Edwin On 28 March 2015 at 02:27, Erick Erickson erickerick...@gmail.com wrote: You say you re-indexed, did you _completely_ remove the data directory first, i.e. the parent of the index and, maybe, tlog directories? I've occasionally seen remnants of old definitions pollute the new one, and since the uniqueKey key is so fundamental I can see it being a problem. Best, Erick On Fri, Mar 27, 2015 at 1:42 AM, Andrea Gazzarini a.gazzar...@gmail.com wrote: Hi Edwin, please provide some other detail about your context, (e.g. complete stacktrace, query you're issuing) Best, Andrea On 03/27/2015 09:38 AM, Zheng Lin Edwin Yeo wrote: Hi everyone, I've changed my uniqueKey to another name, instead of using id, on the schema.xml. However, after I have done the indexing (the indexing is successful), I'm not able to perform a search query on it. I gives the error java.lang.NullPointerException. Is there other place which I need to configure, besides changing the uniqueKey field in scheam.xml? Regards, Edwin
Re: Collapse and Expand behaviour on result with 1 document.
The expanded section will only include groups that have expanded documents. So, if the document that in the main result set has no documents to expand, then this is working as expected. Joel Bernstein http://joelsolr.blogspot.com/ On Mon, Mar 30, 2015 at 8:43 PM, Derek Poh d...@globalsources.com wrote: Hi I have a query which return 1 document. When I add the collapse and expand parameters to it, expand=trueexpand.rows=5fq={!collapse%20field=P_SupplierId}, the expanded section is empty (lst name=expanded/). Is this the behaviour of collapse and expand parameters on result which contain only 1 document? -Derek
RE: Solr Unexpected Query Parser Exception
Saw that one. Can't remember for certain, but recall the actual syntax error was in a filter query. It could have been a quoting error or a date math error in a range expression. But, either way, the issue was in the fq. Using edismax. hth -Original Message- From: Jack Krupansky [mailto:jack.krupan...@gmail.com] Sent: Friday, March 20, 2015 5:08 PM To: solr-user@lucene.apache.org Subject: Re: Solr Unexpected Query Parser Exception Which query parser are you using? The dismax query parser does not support wild cards or *:*. Either way, the error message is unhelpful - worth filing a Jira. -- Jack Krupansky On Fri, Mar 20, 2015 at 7:21 AM, Vishnu Mishra vdil...@gmail.com wrote: Hi, I am using solr 4.10.3 and doing distributed shard query. I am getting following syntax exception at regular intervals. ERROR org.apache.solr.core.SolrCore ? org.apache.solr.common.SolrException: org.apache.solr.search.SyntaxError: Cannot parse '*:*': Encountered EOF at line 1, column 3. Was expecting one of: NOT ... + ... - ... BAREOPER ... ( ... * ... ^ ... QUOTED ... TERM ... FUZZY_SLOP ... PREFIXTERM ... WILDTERM ... REGEXPTERM ... [ ... { ... PROXY_MF ... JOIN_MF ... FRQ_CUT_OFF ... NUMBER ... TERM ... * ... My query to solr is *:* can anyone explain me why this exception is occurring. -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-Unexpected-Query-Parser-Except ion-tp4194156.html Sent from the Solr - User mailing list archive at Nabble.com. * This e-mail may contain confidential or privileged information. If you are not the intended recipient, please notify the sender immediately and then delete it. TIAA-CREF *
RE: NoNode for /clusterstate.json in solr5.0.0 cloud
One development difference is that i am trying to use it in a scala project. I compile spark-solr as jar and use it in my scala project. Do you think this might be a reason. thanks sumit From: Purohit, Sumit Sent: Monday, March 30, 2015 3:19 PM To: solr-user@lucene.apache.org Subject: RE: NoNode for /clusterstate.json in solr5.0.0 cloud Yes i am using default localhost:9983 and i think my program does connect with zookeeper as per logs : 2015-03-30 15:00:51,512 [Executor task launch worker-0] INFO SolrZkClient - Using default ZkCredentialsProvider 2015-03-30 15:00:51,530 [Executor task launch worker-0] INFO ConnectionManager - Waiting for client to connect to ZooKeeper 2015-03-30 15:00:51,544 [zkCallback-3-thread-1] INFO ConnectionManager - Watcher org.apache.solr.common.cloud.ConnectionManager@51ffb98d name:ZooKeeperConnection Watcher:localhost:9983/sorl got event WatchedEvent state:SyncConnected type:None path:null path:null type:None 2015-03-30 15:00:51,544 [Executor task launch worker-0] INFO ConnectionManager - Client is connected to ZooKeeper 2015-03-30 15:00:51,544 [Executor task launch worker-0] INFO SolrZkClient - Using default ZkACLProvider 2015-03-30 15:00:51,548 [Executor task launch worker-0] INFO SolrZkClient - makePath: /clusterstate.json 2015-03-30 15:00:51,557 [Executor task launch worker-0] ERROR Executor - Exception in task 0.0 in stage 0.0 (TID 0) So to upgrade spark-solr project, i had to change all the references to of class xxxSolrServer to xxxSolrClient . did you do something similar. If you plan to push it to github, i can grab that and compare with my environment. Thanks for you prompt response(s). thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:54 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud I upgraded the spark-solr project to solrj-5.0.0 and was able to index into the gettingstarted collection using Solr 5.0.0, so seems like it may be environmental. Almost seems like the spark project is looking at the wrong ZooKeeper? Are you using the default -zkHost localhost:9983 On Mon, Mar 30, 2015 at 2:32 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: Thanks Tim, i had to make some changes in my local spark-solr clone to build it for sorl5. If its ok, i can commit these to github. thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:27 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud Ok, let me upgrade my version of spark-solr to 5 to see what I get ... On Mon, Mar 30, 2015 at 2:26 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: yes there is getting started collection.. and on admin webpage console--cloud---tree---/clusterstate.json shows me this table version =1 aversion=0 children_count=0 ctimeFri= Mar 27 19:20:21 UTC 2015 (1427484021901) cversion=0 czxid=32 ephemeralOwner=0 mtime=Fri Mar 27 19:20:36 UTC 2015 (1427484036453) mzxid=110 pzxid=32 dataLength=2 children_count=0 seems related to no node error. thanks sumit From: Timothy Potter [thelabd...@gmail.com] Sent: Monday, March 30, 2015 2:18 PM To: solr-user@lucene.apache.org Subject: Re: NoNode for /clusterstate.json in solr5.0.0 cloud Anything in the server-side Solr logs? Also, if you go to the Solr admin console at http://localhost:8983/solr, do you see the gettingstarted collection in the cloud panel? On Mon, Mar 30, 2015 at 1:12 PM, Purohit, Sumit sumit.puro...@pnnl.gov wrote: I have a basic Solr 5.0.0 cloud setup after following http://lucene.apache.org/solr/quickstart.html I am trying to read data from spark and index it into solr using following lib: https://github.com/LucidWorks/spark-solr I am getting following error when my code try to make request to solr Exception in thread main org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): org.apache.solr.common.cloud.ZooKeeperException: at org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:465) . .. .. Caused by: org.apache.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /clusterstate.json at org.apache.zookeeper.KeeperException.create(KeeperException.java:111) at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:783) at org.apache.solr.common.cloud.SolrZkClient$10.execute(SolrZkClient.java:500) I am not sure how
Re: Same schema.xml is loaded for different cores in SolrCloud
Hi Erick, I've started shard2 with the following command instead, but it's still the same problem. java -DzkHost=localhost:9983 -Djetty.port=8984 -jar start.jar But you mean for shard1 we do not have to include -Dbootstrap_confdir=./solr/logmill/conf for subsequent startup? Regards, Edwin On 31 March 2015 at 00:46, Erick Erickson erickerick...@gmail.com wrote: OK, this is a bit confused: 1 You're starting two embedded Zookeepers but they don't know about each other. So looking for the configsets is a bit confused. 2 There's no need to do the bootstrap thing after the first time. The _very_ first time you do this it pushes the configs up to Zookeeper, but after that you should just reference the config name. 3 you specify the config name when you _create_ the collection, not when you start it up. You may be doing this, but your startup. 4 I think you're confusing shards with collections. The equivalent of older-style cores would be just single-shard _collections_. Configs are associated on the collection level, not the shard level as all shards in a collection are presumed (indeed, _must_) use the same configuration. HTH, Erick On Mon, Mar 30, 2015 at 2:20 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: I've roughly know what is the problem from here. http://stackoverflow.com/questions/23338324/zookeeper-multiple-collection-different-schema However, I couldn't find the zoo_data directory in all of my solr folder. What could be the problem or where is the directory supposed to be located? Regards, Edwin On 30 March 2015 at 11:56, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi everyone, I've created a SolrCloud with multiple core, and I have different schema.xml for each of the core. However, when I start Solr, there's only one version of the schema.xml that is loaded onto Solr. Regardless of which core I go to, the schema.xml that is shown is the first one which I have loaded. What I did was, I have 3 cores: logmill, collection1 and collection2. Each of the core has 2 shrads: shard1 and shard2 I first started the Solr with shard1 using the following command: java -Dcollection.configName=logmill -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar After that I start shard2 using the following command: java -Dcollection.configName=logmill -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar All the schema.xml loaded are from logmill core, even for the collection1 and collection2. Even after I change the command to start shard1 with the following command, all the schema.xml are still from logmill java -Dcollection.configName=collection1 -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/collection1/conf -jar start.jar How do I get Solr to read the different schema.xml for the different cores? Regards, Edwin
Re: Unable to perform search query after changing uniqueKey
Latest information that I've found for this is that the error only occurs for shard2. If I do a search for just shard1, those records that are assigned to shard1 will be able to be displayed. Only when I search for shard2 will the NullPointerException error occurs. Previously I was doing a search for both shards. Is there any settings that I required to do for shard2 in order to solve this issue? Currently I have not made any changes to the shards since I created it using http://localhost:8983/solr/admin/collections?action=CREATEname=nps1numShards=2collection.configName=collection1 Regards, Edwin On 31 March 2015 at 09:42, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I've changed the uniqueKey from id to Item No. uniqueKeyItem No/uniqueKey Below are my definitions for both the id and Item No. field name=id type=string indexed=true stored=true required=false multiValued=false / field name=Item No type=text_general indexed=true stored=true/ Regards, Edwin On 30 March 2015 at 23:05, Erick Erickson erickerick...@gmail.com wrote: Well, let's see the definition of your ID field, 'cause I'm puzzled. It's definitely A Bad Thing to have it be any kind of tokenized field though, but that's a shot in the dark. Best, Erick On Mon, Mar 30, 2015 at 2:17 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Mostafa, Yes, I've defined all the fields in schema.xml. It is able to work on the version without SolrCloud, but it is not working for the one with SolrCloud. Both of them are using the same schema.xml. Regards, Edwin On 30 March 2015 at 14:34, Mostafa Gomaa mostafa.goma...@gmail.com wrote: Hi Zheng, It's possible that there's a problem with your schema.xml. Are all fields defined and have appropriate options enabled? Regards, Mostafa. On Mon, Mar 30, 2015 at 7:49 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I've tried that, and removed the data directory from both the shards. But the same problem still occurs, so we probably can rule out the memory issue. Regards, Edwin On 30 March 2015 at 12:39, Erick Erickson erickerick...@gmail.com wrote: I meant shut down Solr and physically remove the entire data directory. Not saying this is the cure, but it can't hurt to rule out the index having memory... Best, Erick On Sun, Mar 29, 2015 at 6:35 PM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I used the following query to delete all the index. http://localhost:8983/solr/update?stream.body= deletequery*:*/query/delete http://localhost:8983/solr/update?stream.body=commit/ Or is it better to physically delete the entire data directory? Regards, Edwin On 28 March 2015 at 02:27, Erick Erickson erickerick...@gmail.com wrote: You say you re-indexed, did you _completely_ remove the data directory first, i.e. the parent of the index and, maybe, tlog directories? I've occasionally seen remnants of old definitions pollute the new one, and since the uniqueKey key is so fundamental I can see it being a problem. Best, Erick On Fri, Mar 27, 2015 at 1:42 AM, Andrea Gazzarini a.gazzar...@gmail.com wrote: Hi Edwin, please provide some other detail about your context, (e.g. complete stacktrace, query you're issuing) Best, Andrea On 03/27/2015 09:38 AM, Zheng Lin Edwin Yeo wrote: Hi everyone, I've changed my uniqueKey to another name, instead of using id, on the schema.xml. However, after I have done the indexing (the indexing is successful), I'm not able to perform a search query on it. I gives the error java.lang.NullPointerException. Is there other place which I need to configure, besides changing the uniqueKey field in scheam.xml? Regards, Edwin
Re: Solr went on recovery multiple time.
How big is your index? -Xmx6g is pretty small in general. On Mon, Mar 30, 2015 at 7:20 PM, sthita sthit...@gmail.com wrote: Thanks Erick for the reply. These are the configurations. 1.Solr servers count : 4 (1 master , 3 slaves) 2.My initial and maximum memory allocations to each solr servers are : -Xms5g -Xmx6g 3. 5 running zookeepers 4. My index size in each solr server around: 48GB 5. Every sec merger add/delete 500 docs to/from solr yes, i checked the the ZK config timeout is 10sec. I will increase the timeout to 45sec, Hope that will help. ZK config : # The number of milliseconds of each tick tickTime=2000 # The number of ticks that the initial # synchronization phase can take initLimit=10 # The number of ticks that can pass between # sending a request and getting an acknowledgement syncLimit=5 -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-went-on-recovery-multiple-time-tp4196249p4196450.html Sent from the Solr - User mailing list archive at Nabble.com. -- Bill Bell billnb...@gmail.com cell 720-256-8076
RE: Korean script conversion
We only want the conversion Hanja-Hangul, for each Hanja character there exists only one Hangul character that can replace it in a Korean text. The other way around is not convertible. We want to allow searching in both scripts and find matches in both scripts. Thanks Eyal Naamati Alma Developer Tel: +972-2-6499313 Mobile: +972-547915255 eyal.naam...@exlibrisgroup.com www.exlibrisgroup.com -Original Message- From: Benson Margulies [mailto:bimargul...@gmail.com] Sent: Monday, March 30, 2015 1:58 PM To: solr-user Subject: Re: Korean script conversion Why do you think that this is a good idea? Hanja are used for special purposes; they are not trivally convertable to Hanjul due to ambiguity, and it's not at all clear that a typical search user wants to treat them as equivalent. On Sun, Mar 29, 2015 at 1:52 AM, Eyal Naamati eyal.naam...@exlibrisgroup.com wrote: Hi, We are starting to index records in Korean. Korean text can be written in two scripts: Han characters (Chinese) and Hangul characters (Korean). We are looking for some solr filter or another built in solr component that converts between Han and Hangul characters (transliteration). I know there is the ICUTransformFilterFactory that can convert between Japanese or chinese scripts, for example: filter class=*solr.ICUTransformFilterFactory* id=*Katakana- Hiragana* / for Japanese script conversions So far I couldn't find anything readymade for Korean scripts, but perhaps someone knows of one? Thanks! Eyal Naamati Alma Developer Tel: +972-2-6499313 Mobile: +972-547915255 eyal.naam...@exlibrisgroup.com [image: Description: Description: Description: Description: C://signature/exlibris.jpg] www.exlibrisgroup.com
Re: Same schema.xml is loaded for different cores in SolrCloud
By now, I wouldn't particularly trust my setup. I'd blow it away and start over. bootstrapping is _only_ required to get the configs up to Zookeeper the first time. In fact I suggest you don't use it at all. Just start SolrCloud, and use zkcli to push the configs up. Thereafter, create your collections using the collections API. Zookeeper is just a central repository for your configs and the overall state of your cluster. As far as config sets are concerned, think of the upconfig (or bootstrap) as copying the config files to a place where they can be found by any random Solr instance that starts up. And the same applies to parameters like numShards. It's _only_ used as a convenience for creating a cluster for demo purposes. Thereafter, any time you start up that particular cloud, it'll read the old cluster state and completely ignore the numShards parameter. Rather than try to untangle what you've done, I'd re-install and work through the tutorial step-by-step. I think you've jumped ahead and gotten some things mixed up as far as your cluster state is concerned. Best, Erick On Mon, Mar 30, 2015 at 6:52 PM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi Erick, I've started shard2 with the following command instead, but it's still the same problem. java -DzkHost=localhost:9983 -Djetty.port=8984 -jar start.jar But you mean for shard1 we do not have to include -Dbootstrap_confdir=./solr/logmill/conf for subsequent startup? Regards, Edwin On 31 March 2015 at 00:46, Erick Erickson erickerick...@gmail.com wrote: OK, this is a bit confused: 1 You're starting two embedded Zookeepers but they don't know about each other. So looking for the configsets is a bit confused. 2 There's no need to do the bootstrap thing after the first time. The _very_ first time you do this it pushes the configs up to Zookeeper, but after that you should just reference the config name. 3 you specify the config name when you _create_ the collection, not when you start it up. You may be doing this, but your startup. 4 I think you're confusing shards with collections. The equivalent of older-style cores would be just single-shard _collections_. Configs are associated on the collection level, not the shard level as all shards in a collection are presumed (indeed, _must_) use the same configuration. HTH, Erick On Mon, Mar 30, 2015 at 2:20 AM, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: I've roughly know what is the problem from here. http://stackoverflow.com/questions/23338324/zookeeper-multiple-collection-different-schema However, I couldn't find the zoo_data directory in all of my solr folder. What could be the problem or where is the directory supposed to be located? Regards, Edwin On 30 March 2015 at 11:56, Zheng Lin Edwin Yeo edwinye...@gmail.com wrote: Hi everyone, I've created a SolrCloud with multiple core, and I have different schema.xml for each of the core. However, when I start Solr, there's only one version of the schema.xml that is loaded onto Solr. Regardless of which core I go to, the schema.xml that is shown is the first one which I have loaded. What I did was, I have 3 cores: logmill, collection1 and collection2. Each of the core has 2 shrads: shard1 and shard2 I first started the Solr with shard1 using the following command: java -Dcollection.configName=logmill -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar After that I start shard2 using the following command: java -Dcollection.configName=logmill -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/logmill/conf -jar start.jar All the schema.xml loaded are from logmill core, even for the collection1 and collection2. Even after I change the command to start shard1 with the following command, all the schema.xml are still from logmill java -Dcollection.configName=collection1 -DzkRun -DnumShards=2 -Dbootstrap_confdir=./solr/collection1/conf -jar start.jar How do I get Solr to read the different schema.xml for the different cores? Regards, Edwin