[jira] [Updated] (SOLR-4018) Dispatcher to optionally write QTime and Hits HTTP header
[ https://issues.apache.org/jira/browse/SOLR-4018?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-4018: Description: SolrDispatchFilter should be able to write QTime and Hits HTTP headers via configuration. {code} {code} was:SolrDispatchFilter should be able to write QTime and Hits HTTP headers via configuration. > Dispatcher to optionally write QTime and Hits HTTP header > - > > Key: SOLR-4018 > URL: https://issues.apache.org/jira/browse/SOLR-4018 > Project: Solr > Issue Type: Improvement >Affects Versions: 4.0 >Reporter: Markus Jelsma > Fix For: 4.1, 5.0 > > > SolrDispatchFilter should be able to write QTime and Hits HTTP headers via > configuration. > {code} > > multipartUploadLimitInKB="2048000" /> > > > > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-4018) Dispatcher to optionally write QTime and Hits HTTP header
[ https://issues.apache.org/jira/browse/SOLR-4018?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-4018: Attachment: SOLR-4018-trunk-1.patch Here's a patch for trunk adding two fields to SolrConfig and checking them and adding headers in SolrDispatchFilter. > Dispatcher to optionally write QTime and Hits HTTP header > - > > Key: SOLR-4018 > URL: https://issues.apache.org/jira/browse/SOLR-4018 > Project: Solr > Issue Type: Improvement >Affects Versions: 4.0 >Reporter: Markus Jelsma > Fix For: 4.1, 5.0 > > Attachments: SOLR-4018-trunk-1.patch > > > SolrDispatchFilter should be able to write QTime and Hits HTTP headers via > configuration. > {code} > > multipartUploadLimitInKB="2048000" /> > > > > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-4018) Dispatcher to optionally write QTime and Hits HTTP header
Markus Jelsma created SOLR-4018: --- Summary: Dispatcher to optionally write QTime and Hits HTTP header Key: SOLR-4018 URL: https://issues.apache.org/jira/browse/SOLR-4018 Project: Solr Issue Type: Improvement Affects Versions: 4.0 Reporter: Markus Jelsma Fix For: 4.1, 5.0 SolrDispatchFilter should be able to write QTime and Hits HTTP headers via configuration. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3966) LangID not to log WARN
[ https://issues.apache.org/jira/browse/SOLR-3966?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3966: Attachment: SOLR-3966-trunk-1.patch > LangID not to log WARN > -- > > Key: SOLR-3966 > URL: https://issues.apache.org/jira/browse/SOLR-3966 > Project: Solr > Issue Type: Improvement >Affects Versions: 4.0 >Reporter: Markus Jelsma > Fix For: 4.1, 5.0 > > Attachments: SOLR-3966-trunk-1.patch > > > The LangID UpdateProcessor emits the warning below for documents that do not > contain an input field. The level should go to DEBUG or be removed. It is not > uncommon to see a log full of these messages just because not all documents > contain all the fields we're mapping. > {code}Oct 19, 2012 11:23:43 AM > org.apache.solr.update.processor.LanguageIdentifierUpdateProcessor process > WARNING: Document does not contain input field . Skipping > this{code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3966) LangID not to log WARN
[ https://issues.apache.org/jira/browse/SOLR-3966?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3966: Attachment: (was: SOLR-3966-trunk-1.patch) > LangID not to log WARN > -- > > Key: SOLR-3966 > URL: https://issues.apache.org/jira/browse/SOLR-3966 > Project: Solr > Issue Type: Improvement >Affects Versions: 4.0 >Reporter: Markus Jelsma > Fix For: 4.1, 5.0 > > Attachments: SOLR-3966-trunk-1.patch > > > The LangID UpdateProcessor emits the warning below for documents that do not > contain an input field. The level should go to DEBUG or be removed. It is not > uncommon to see a log full of these messages just because not all documents > contain all the fields we're mapping. > {code}Oct 19, 2012 11:23:43 AM > org.apache.solr.update.processor.LanguageIdentifierUpdateProcessor process > WARNING: Document does not contain input field . Skipping > this{code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3966) LangID not to log WARN
[ https://issues.apache.org/jira/browse/SOLR-3966?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3966: Attachment: SOLR-3966-trunk-1.patch Patch removing the warning. > LangID not to log WARN > -- > > Key: SOLR-3966 > URL: https://issues.apache.org/jira/browse/SOLR-3966 > Project: Solr > Issue Type: Improvement >Affects Versions: 4.0 >Reporter: Markus Jelsma > Fix For: 4.1, 5.0 > > Attachments: SOLR-3966-trunk-1.patch > > > The LangID UpdateProcessor emits the warning below for documents that do not > contain an input field. The level should go to DEBUG or be removed. It is not > uncommon to see a log full of these messages just because not all documents > contain all the fields we're mapping. > {code}Oct 19, 2012 11:23:43 AM > org.apache.solr.update.processor.LanguageIdentifierUpdateProcessor process > WARNING: Document does not contain input field . Skipping > this{code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3966) LangID not to log WARN
Markus Jelsma created SOLR-3966: --- Summary: LangID not to log WARN Key: SOLR-3966 URL: https://issues.apache.org/jira/browse/SOLR-3966 Project: Solr Issue Type: Improvement Affects Versions: 4.0 Reporter: Markus Jelsma Fix For: 4.1, 5.0 The LangID UpdateProcessor emits the warning below for documents that do not contain an input field. The level should go to DEBUG or be removed. It is not uncommon to see a log full of these messages just because not all documents contain all the fields we're mapping. {code}Oct 19, 2012 11:23:43 AM org.apache.solr.update.processor.LanguageIdentifierUpdateProcessor process WARNING: Document does not contain input field . Skipping this{code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3705) hl.alternateField does not support glob
[ https://issues.apache.org/jira/browse/SOLR-3705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13478929#comment-13478929 ] Markus Jelsma commented on SOLR-3705: - I took me a while to get back to this. The problem is that we have over 20 full text fields per document of which only one contains the text, they're all language specific fields like content_en, content_de etc. We use hl.fl=content_* to get highlighted snippets for whatever field is matched by the main query. But a document can also be matched on a non-content field so the highlighter won't find a snippet for the content field. We though that if we could glob the alternate field as well, it would be a simple mechanism to get a snippet from an alternate field that is any of the content_* fields. > hl.alternateField does not support glob > --- > > Key: SOLR-3705 > URL: https://issues.apache.org/jira/browse/SOLR-3705 > Project: Solr > Issue Type: Improvement > Components: highlighter >Affects Versions: 4.0-ALPHA >Reporter: Markus Jelsma >Priority: Minor > Fix For: 5.0 > > > Unlike hl.fl, the hl.alternateField does not support * to match field globs. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3925) Expose SpanFirst in eDismax
[ https://issues.apache.org/jira/browse/SOLR-3925?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3925: Attachment: SOLR-3925-trunk-1.patch > Expose SpanFirst in eDismax > --- > > Key: SOLR-3925 > URL: https://issues.apache.org/jira/browse/SOLR-3925 > Project: Solr > Issue Type: Improvement > Components: query parsers >Affects Versions: 4.0-BETA > Environment: solr-spec 5.0.0.2012.10.09.19.29.59 > solr-impl 5.0-SNAPSHOT 1366361:1396116M - markus - 2012-10-09 19:29:59 >Reporter: Markus Jelsma > Fix For: 4.1, 5.0 > > Attachments: SOLR-3925-trunk-1.patch > > > Expose Lucene's SpanFirst capability in Solr's extended Dismax query parser. > This issue adds the SF-parameter (SpanFirst) and takes a FIELD~DISTANCE^BOOST > formatted value. > For example, sf=title~5^2 will give a boost of 2 if one of the normal > clauses, originally generated for automatic phrase queries, is located within > five positions from the field's start. > Unit test is included and all tests pass. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3925) Expose SpanFirst in eDismax
Markus Jelsma created SOLR-3925: --- Summary: Expose SpanFirst in eDismax Key: SOLR-3925 URL: https://issues.apache.org/jira/browse/SOLR-3925 Project: Solr Issue Type: Improvement Components: query parsers Affects Versions: 4.0-BETA Environment: solr-spec 5.0.0.2012.10.09.19.29.59 solr-impl 5.0-SNAPSHOT 1366361:1396116M - markus - 2012-10-09 19:29:59 Reporter: Markus Jelsma Fix For: 4.1, 5.0 Attachments: SOLR-3925-trunk-1.patch Expose Lucene's SpanFirst capability in Solr's extended Dismax query parser. This issue adds the SF-parameter (SpanFirst) and takes a FIELD~DISTANCE^BOOST formatted value. For example, sf=title~5^2 will give a boost of 2 if one of the normal clauses, originally generated for automatic phrase queries, is located within five positions from the field's start. Unit test is included and all tests pass. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Closed] (LUCENE-4470) Expose SpanFirst in eDismax
[ https://issues.apache.org/jira/browse/LUCENE-4470?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma closed LUCENE-4470. - Resolution: Invalid Fix Version/s: (was: 5.0) (was: 4.1) Accidentally added to Lucene. I'll close and open in the Solr project. Sorry. > Expose SpanFirst in eDismax > --- > > Key: LUCENE-4470 > URL: https://issues.apache.org/jira/browse/LUCENE-4470 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/queryparser >Affects Versions: 4.0-BETA > Environment: solr-spec 5.0.0.2012.10.09.19.29.59 > solr-impl 5.0-SNAPSHOT 1366361:1396116M - markus - 2012-10-09 19:29:59 >Reporter: Markus Jelsma > Attachments: SOLR-4470-trunk-1.patch > > > Expose Lucene's SpanFirst capability in Solr's extended Dismax query parser. > This issue adds the SF-parameter (SpanFirst) and takes a FIELD~DISTANCE^BOOST > formatted value. > For example, sf=title~5^2 will give a boost of 2 if one of the normal > clauses, originally generated for automatic phrase queries, is located within > five positions from the field's start. > Unit test is included and all tests pass. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-4470) Expose SpanFirst in eDismax
[ https://issues.apache.org/jira/browse/LUCENE-4470?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated LUCENE-4470: -- Attachment: SOLR-4470-trunk-1.patch > Expose SpanFirst in eDismax > --- > > Key: LUCENE-4470 > URL: https://issues.apache.org/jira/browse/LUCENE-4470 > Project: Lucene - Core > Issue Type: Improvement > Components: modules/queryparser >Affects Versions: 4.0-BETA > Environment: solr-spec 5.0.0.2012.10.09.19.29.59 > solr-impl 5.0-SNAPSHOT 1366361:1396116M - markus - 2012-10-09 19:29:59 >Reporter: Markus Jelsma > Fix For: 4.1, 5.0 > > Attachments: SOLR-4470-trunk-1.patch > > > Expose Lucene's SpanFirst capability in Solr's extended Dismax query parser. > This issue adds the SF-parameter (SpanFirst) and takes a FIELD~DISTANCE^BOOST > formatted value. > For example, sf=title~5^2 will give a boost of 2 if one of the normal > clauses, originally generated for automatic phrase queries, is located within > five positions from the field's start. > Unit test is included and all tests pass. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-4470) Expose SpanFirst in eDismax
Markus Jelsma created LUCENE-4470: - Summary: Expose SpanFirst in eDismax Key: LUCENE-4470 URL: https://issues.apache.org/jira/browse/LUCENE-4470 Project: Lucene - Core Issue Type: Improvement Components: modules/queryparser Affects Versions: 4.0-BETA Environment: solr-spec 5.0.0.2012.10.09.19.29.59 solr-impl 5.0-SNAPSHOT 1366361:1396116M - markus - 2012-10-09 19:29:59 Reporter: Markus Jelsma Fix For: 4.1, 5.0 Expose Lucene's SpanFirst capability in Solr's extended Dismax query parser. This issue adds the SF-parameter (SpanFirst) and takes a FIELD~DISTANCE^BOOST formatted value. For example, sf=title~5^2 will give a boost of 2 if one of the normal clauses, originally generated for automatic phrase queries, is located within five positions from the field's start. Unit test is included and all tests pass. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) Solr Cloud sometimes skipped peersync attempt and replicated instead due to tlog flags not being cleared when no updates were buffered during a previous replication.
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13458679#comment-13458679 ] Markus Jelsma commented on SOLR-3685: - {quote}So what we're seeing here is the mmapped nodes use more RES and SHR than the NIO node. VIRT is as expected. I'll change another node to NIO and keep them running again for the next few days and keep sending documents and firing queries.{quote} there is still an issue with mmap and high RES opposed to NIOFS but the actual issue is already resolved. I'll open a new issue. > Solr Cloud sometimes skipped peersync attempt and replicated instead due to > tlog flags not being cleared when no updates were buffered during a previous > replication. > - > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Assignee: Yonik Seeley >Priority: Critical > Fix For: 4.0, 5.0 > > Attachments: info.log, oom-killer.log, pmap.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3808) Extraction contrib to utilize Boilerpipe
[ https://issues.apache.org/jira/browse/SOLR-3808?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13455680#comment-13455680 ] Markus Jelsma commented on SOLR-3808: - Hi - in Apache Nutch i keep the loaded extractors in a static hashmap. The content handlers have to be wrapped like this and the extractor implementation has to be passed to the BoilerpipeContentHandler constructor, it doesn't use configuration to find an extractor. > Extraction contrib to utilize Boilerpipe > > > Key: SOLR-3808 > URL: https://issues.apache.org/jira/browse/SOLR-3808 > Project: Solr > Issue Type: Improvement > Components: contrib - Solr Cell (Tika extraction) >Reporter: Markus Jelsma >Priority: Minor > Attachments: SOLR-3808-trunk-1.patch > > > Solr's extraction contrib uses Tika for document parsing and should be able > te use Boilerpipe. Tika comes with Boilerpipe, a library capable of removing > boilerplate text from HTML pages. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3808) Extraction contrib to utilize Boilerpipe
[ https://issues.apache.org/jira/browse/SOLR-3808?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3808: Attachment: SOLR-3808-trunk-1.patch Here's a patch for trunk. It introduces the boilerpipe parameter which takes a names Boilerpipe extractor such as ArticleExtractor, CanolaExtractor or KeepEverythingExtractor. It also comes with a unit test. The test HTML file's extracted content will not contain the word `footer` if the ArticleExtractor is used. > Extraction contrib to utilize Boilerpipe > > > Key: SOLR-3808 > URL: https://issues.apache.org/jira/browse/SOLR-3808 > Project: Solr > Issue Type: Improvement > Components: contrib - Solr Cell (Tika extraction) >Reporter: Markus Jelsma >Priority: Minor > Fix For: 4.0 > > Attachments: SOLR-3808-trunk-1.patch > > > Solr's extraction contrib uses Tika for document parsing and should be able > te use Boilerpipe. Tika comes with Boilerpipe, a library capable of removing > boilerplate text from HTML pages. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3808) Extraction contrib to utilize Boilerpipe
Markus Jelsma created SOLR-3808: --- Summary: Extraction contrib to utilize Boilerpipe Key: SOLR-3808 URL: https://issues.apache.org/jira/browse/SOLR-3808 Project: Solr Issue Type: Improvement Components: contrib - Solr Cell (Tika extraction) Reporter: Markus Jelsma Priority: Minor Fix For: 4.0 Solr's extraction contrib uses Tika for document parsing and should be able te use Boilerpipe. Tika comes with Boilerpipe, a library capable of removing boilerplate text from HTML pages. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3782) A leader going down while updates are coming in can cause shard inconsistency.
[ https://issues.apache.org/jira/browse/SOLR-3782?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13447644#comment-13447644 ] Markus Jelsma commented on SOLR-3782: - hi - this is in CHANGES but is it resolved? > A leader going down while updates are coming in can cause shard inconsistency. > -- > > Key: SOLR-3782 > URL: https://issues.apache.org/jira/browse/SOLR-3782 > Project: Solr > Issue Type: Bug > Components: SolrCloud >Reporter: Mark Miller >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > > Harpoon into the head of the great whale I have been chasing for a couple > weeks now. > ChaosMonkey test was exposing this. > Turns out the problem was the solr cmd distrib executor - when closing the > leader CoreContainer, we would close the zkController while updates can still > flow through the distrib executor. The result was that we would send updates > from the leader briefly even though there was a new leader. > I had suspected something similar to this at one point in the hunt and > started adding some defensive state checks that we wanted to add anyway. I > don't think they caught all of this issue due to the limited tightness one of > the state checks can get to (checking the cloudstate leader from a replica > against the leader indicated by the request). > So the answer is to finally work out how to stop the solr cmd distrib > executor - because we need to stop it before closing zkController and giving > up our role as leader. > I've worked that all out and the issue no longer seems to be a problem. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) Solr Cloud sometimes skipped peersync attempt and replicated instead due to tlog flags not being cleared when no updates were buffered during a previous replication.
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13437790#comment-13437790 ] Markus Jelsma commented on SOLR-3685: - To my surprise the RES for all nodes except the NIOFS node increased slowly over the past three days and were still increasing today. The mmapped nodes used sometimes up to three times the Xmx and, for some reason, about 1/2 Xmx of shared memory. We just restarted all nodes with 9 using mmap and one using NIO, after restart the mmapped nodes immediately start to use a lot more RES than the NIO node. The NIO node also uses much less shared memory. Perhaps what i've seen before with NIO also crashing was due to some other issue. So what we're seeing here is the mmapped nodes use more RES and SHR than the NIO node. VIRT is as expected. I'll change another node to NIO and keep them running again for the next few days and keep sending documents and firing queries. All nodes are using august 20th trunk from now on. > Solr Cloud sometimes skipped peersync attempt and replicated instead due to > tlog flags not being cleared when no updates were buffered during a previous > replication. > - > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Assignee: Yonik Seeley >Priority: Critical > Fix For: 4.0, 5.0 > > Attachments: info.log, oom-killer.log, pmap.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3685) Solr Cloud sometimes skipped peersync attempt and replicated instead due to tlog flags not being cleared when no updates were buffered during a previous replication.
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3685: Attachment: pmap.log Here's the pmap for one node. Heap Xmx is still 256M. I also just noticed this node still has OpenJDK 6 running instead of Sun Java 6 like the other nodes. Despite that difference the memory consumption is equal. I'll also restart a node with NIOFS but i still expect memory to increase as with mmap. > Solr Cloud sometimes skipped peersync attempt and replicated instead due to > tlog flags not being cleared when no updates were buffered during a previous > replication. > - > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Assignee: Yonik Seeley >Priority: Critical > Fix For: 4.0, 5.0 > > Attachments: info.log, oom-killer.log, pmap.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-3685) Solr Cloud sometimes skipped peersync attempt and replicated instead due to tlog flags not being cleared when no updates were buffered during a previous replication
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13436214#comment-13436214 ] Markus Jelsma edited comment on SOLR-3685 at 8/17/12 5:48 AM: -- We didn't think mmap could be the cause but nevertheless we tried that once on a smaller cluster and got a lot of memory consumption again, after which it got killed. I can see if i can run one or two of the nodes with NIOFS but let the other run with mmap. We don't automatically restart cores so it should run fine if we temporarily change the config in zookeeper and restart two nodes. edit: each core has a ~2.5GB index. was (Author: markus17): We didn't think mmap could be the cause but nevertheless we tried that once on a smaller cluster and got a lot of memory consumption again, after which it got killed. I can see if i can run one or two of the nodes with NIOFS but let the other run with mmap. We don't automatically restart cores so it should run fine if we temporarily change the config in zookeeper and restart two nodes. > Solr Cloud sometimes skipped peersync attempt and replicated instead due to > tlog flags not being cleared when no updates were buffered during a previous > replication. > - > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Assignee: Yonik Seeley >Priority: Critical > Fix For: 4.0, 5.0 > > Attachments: info.log, oom-killer.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) Solr Cloud sometimes skipped peersync attempt and replicated instead due to tlog flags not being cleared when no updates were buffered during a previous replication.
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13436214#comment-13436214 ] Markus Jelsma commented on SOLR-3685: - We didn't think mmap could be the cause but nevertheless we tried that once on a smaller cluster and got a lot of memory consumption again, after which it got killed. I can see if i can run one or two of the nodes with NIOFS but let the other run with mmap. We don't automatically restart cores so it should run fine if we temporarily change the config in zookeeper and restart two nodes. > Solr Cloud sometimes skipped peersync attempt and replicated instead due to > tlog flags not being cleared when no updates were buffered during a previous > replication. > - > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Assignee: Yonik Seeley >Priority: Critical > Fix For: 4.0, 5.0 > > Attachments: info.log, oom-killer.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3685) Solr Cloud sometimes skipped peersync attempt and replicated instead due to tlog flags not being cleared when no updates were buffered during a previous replication.
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3685: Attachment: oom-killer.log Here's the relevant part of syslog for a node where Tomcat is killed by the OS. There is 1G or available RAM, no configured swap and the heap size is 256MB. The node has two running cores. The off-heap RES memory for the Java process sometimes gets so large that Linux decides to kill it. > Solr Cloud sometimes skipped peersync attempt and replicated instead due to > tlog flags not being cleared when no updates were buffered during a previous > replication. > - > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Assignee: Yonik Seeley >Priority: Critical > Fix For: 4.0, 5.0 > > Attachments: info.log, oom-killer.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) Solr Cloud sometimes skipped peersync attempt and replicated instead due to tlog flags not being cleared when no updates were buffered during a previous replication.
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13436194#comment-13436194 ] Markus Jelsma commented on SOLR-3685: - One node also got rsyslogd killed but the other survived. I assume the OOMkiller output of Linux is what you refer to? > Solr Cloud sometimes skipped peersync attempt and replicated instead due to > tlog flags not being cleared when no updates were buffered during a previous > replication. > - > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Assignee: Yonik Seeley >Priority: Critical > Fix For: 4.0, 5.0 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) Solr Cloud sometimes skipped peersync attempt and replicated instead due to tlog flags not being cleared when no updates were buffered during a previous replication.
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13436182#comment-13436182 ] Markus Jelsma commented on SOLR-3685: - Finally! Two nodes failed again and got killed by the OS. All nodes have a lot of off-heap RES memory, sometimes 3x higher than the heap which is a meager 256MB. Got a name suggestion for the memory issue? I'll open one tomorrow and link to this one. > Solr Cloud sometimes skipped peersync attempt and replicated instead due to > tlog flags not being cleared when no updates were buffered during a previous > replication. > - > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Assignee: Yonik Seeley >Priority: Critical > Fix For: 4.0, 5.0 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-2789) Add showItems to LRUCache, display extra details about cached data for both LRUCache and FastLRUCache
[ https://issues.apache.org/jira/browse/SOLR-2789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13434079#comment-13434079 ] Markus Jelsma commented on SOLR-2789: - This issue is not resolved but it's documented as-if it was: http://wiki.apache.org/solr/SolrCaching#showItems > Add showItems to LRUCache, display extra details about cached data for both > LRUCache and FastLRUCache > - > > Key: SOLR-2789 > URL: https://issues.apache.org/jira/browse/SOLR-2789 > Project: Solr > Issue Type: Improvement > Components: search >Affects Versions: 3.4 >Reporter: Eric Pugh >Priority: Minor > Attachments: cache_details.patch > > > I noticed that the showItems parameter for cache configuration applies to > FastLRUCache but not LRUCache. So I added it. I also noticed that the key > data about an item that is QueryResult cache hit wasn't very useful, but was > for document and filter cache, so added specific support for QueryResultKey. > This patch should probably be merged with the work in SOLR-1893, which is a > proposed refactor. I'm happy to merge these two into a single patch if > someone things that is a good idea. > The output for a queryresult cache hit now looks like: > 0.66 > 43 > 0 > sort=null)">org.apache.solr.search.DocSlice@1e66a917 > org.apache.solr.search.DocSlice@7bfcb845 > I've also updated the example solrconfig.xml in order to better show off the > showItems parameter. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1238) exception in solrJ when authentication is used
[ https://issues.apache.org/jira/browse/SOLR-1238?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13430701#comment-13430701 ] Markus Jelsma commented on SOLR-1238: - We've seen this issue happening over the past few years with or without authentication using SolrJ. Perhaps this issue could be renamed and marked for current Solr versions if applicable. I can't remember seeing this exception when using Curl to load data. > exception in solrJ when authentication is used > -- > > Key: SOLR-1238 > URL: https://issues.apache.org/jira/browse/SOLR-1238 > Project: Solr > Issue Type: Bug > Components: clients - java >Affects Versions: 1.3 >Reporter: Noble Paul >Priority: Minor > Attachments: SOLR-1238.patch > > > see the thread http://markmail.org/thread/w36ih2fnphbubian > {code} > I am facing getting error when I am using Authentication in Solr. I > followed Wiki. The error doesnot appear when I searching. Below is the > code snippet and the error. > Please note I am using Solr 1.4 Development build from SVN. >HttpClient client=new HttpClient(); >AuthScope scope = new > AuthScope(AuthScope.ANY_HOST,AuthScope.ANY_PORT,null, null); >client.getState().setCredentials(scope,new > UsernamePasswordCredentials("guest", "guest")); >SolrServer server =new > CommonsHttpSolrServer("http://localhost:8983/solr",client); >SolrInputDocument doc1=new SolrInputDocument(); >//Add fields to the document >doc1.addField("employeeid", "1237"); >doc1.addField("employeename", "Ann"); >doc1.addField("employeeunit", "etc"); >doc1.addField("employeedoj", "1995-11-31T23:59:59Z"); >server.add(doc1); > Exception in thread "main" > org.apache.solr.client.solrj.SolrServerException: > org.apache.commons.httpclient.ProtocolException: Unbuffered entity > enclosing request can not be repeated. >at > org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:468) >at > org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:242) >at > org.apache.solr.client.solrj.request.UpdateRequest.process(UpdateRequest.java:259) >at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:63) >at test.SolrAuthenticationTest.(SolrAuthenticationTest.java:49) >at test.SolrAuthenticationTest.main(SolrAuthenticationTest.java:113) > Caused by: org.apache.commons.httpclient.ProtocolException: Unbuffered > entity enclosing request can not be repeated. >at > org.apache.commons.httpclient.methods.EntityEnclosingMethod.writeRequestBody(EntityEnclosingMethod.java:487) >at > org.apache.commons.httpclient.HttpMethodBase.writeRequest(HttpMethodBase.java:2114) >at > org.apache.commons.httpclient.HttpMethodBase.execute(HttpMethodBase.java:1096) >at > org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:398) >at > org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:171) >at > org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397) >at > org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:323) >at > org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:415) >... 5 more. > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) solrcloud crashes on startup due to excessive memory consumption
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13430438#comment-13430438 ] Markus Jelsma commented on SOLR-3685: - When exactly? Do you have an issue? > solrcloud crashes on startup due to excessive memory consumption > > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Priority: Critical > Fix For: 4.1 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) solrcloud crashes on startup due to excessive memory consumption
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13429132#comment-13429132 ] Markus Jelsma commented on SOLR-3685: - Each node has two cores and allow only one warming searcher at any time. The problem is triggered on start up after graceful shutdown as well as a hard power off. I've seen it happening not only when the whole cluster if restarted (i don't think i've ever done that) but just one node of the 6 shard 2 replica test cluster. The attached log is of one node being restarted out of the whole cluster. Could the off-heap RAM be part of data being sent over the wire? We've worked around the problem for now by getting more RAM. > solrcloud crashes on startup due to excessive memory consumption > > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Priority: Critical > Fix For: 4.1 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3473) Distributed deduplication broken
[ https://issues.apache.org/jira/browse/SOLR-3473?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3473: Attachment: SOLR-3473-trunk-2.patch Hello - Could the deleteByQuery issue you mention be fixed with SOLR-3473? I've attached an updated patch for today's trunk. The previous patch was missing the signature field but i added it to one schema. Now other tests seem to fail because they don't see the sig field but do use the update chain. Anyway, it seems the BasicDistributedZkTest passes but i'm not very sure, there's too much log output but it doesn't fail. > Distributed deduplication broken > > > Key: SOLR-3473 > URL: https://issues.apache.org/jira/browse/SOLR-3473 > Project: Solr > Issue Type: Bug > Components: SolrCloud, update >Affects Versions: 4.0-ALPHA >Reporter: Markus Jelsma > Fix For: 4.0 > > Attachments: SOLR-3473-trunk-2.patch, SOLR-3473.patch, SOLR-3473.patch > > > Solr's deduplication via the SignatureUpdateProcessor is broken for > distributed updates on SolrCloud. > Mark Miller: > {quote} > Looking again at the SignatureUpdateProcessor code, I think that indeed this > won't currently work with distrib updates. Could you file a JIRA issue for > that? The problem is that we convert update commands into solr documents - > and that can cause a loss of info if an update proc modifies the update > command. > I think the reason that you see a multiple values error when you try the > other order is because of the lack of a document clone (the other issue I > mentioned a few emails back). Addressing that won't solve your issue though - > we have to come up with a way to propagate the currently lost info on the > update command. > {quote} > Please see the ML thread for the full discussion: > http://lucene.472066.n3.nabble.com/SolrCloud-deduplication-td3984657.html -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3705) hl.alternateField does not support glob
Markus Jelsma created SOLR-3705: --- Summary: hl.alternateField does not support glob Key: SOLR-3705 URL: https://issues.apache.org/jira/browse/SOLR-3705 Project: Solr Issue Type: Improvement Components: highlighter Affects Versions: 4.0-ALPHA Reporter: Markus Jelsma Priority: Minor Fix For: 5.0 Unlike hl.fl, the hl.alternateField does not support * to match field globs. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) solrcloud crashes on startup due to excessive memory consumption
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13424785#comment-13424785 ] Markus Jelsma commented on SOLR-3685: - Hi, 1. Yes, but we allow only one searcher at the same time to be warmed. This resource usage also belongs to the Java heap, it cannot cause 5x as much heap being allocated. 2. Yes, i'll open a new issue and refer to this. 3. Well, in some logs i clearly see a core is attempting to download and judging from the multiple index directories it's true. I am very sure no updates have been added to the cluster for a long time yet it still attempts to recover. Below is a core recovering. {code} 2012-07-30 09:48:36,970 INFO [solr.cloud.ZkController] - [main] - : We are http://nl2.index.openindex.io:8080/solr/openindex_a/ and leader is http://nl1.index.openindex.io:8080/solr/openindex_a/ 2012-07-30 09:48:36,970 INFO [solr.cloud.ZkController] - [main] - : No LogReplay needed for core=openindex_a baseURL=http://nl2.index.openindex.io:8080/solr 2012-07-30 09:48:36,970 INFO [solr.cloud.ZkController] - [main] - : Core needs to recover:openindex_a {code} Something noteworthy may be that for some reasons the index versions of all cores and their replica's don't match. After a restart the generation of a core is also different while it shouldn't have changed. The size in bytes is also slightly different (~20 bytes). The main thing that's concerning that Solr consumes 5x the allocated heap space in the RESident memory. Caches and such are in the heap and the MMapped index dir should be in VIRTual memory and not cause the kernel to kill the process. I'm not yet sure what's going on here. Also, according to Uwe virtual memory should not be more than 2-3 times index size. In our case we see ~800Mb virtual memory for two 26Mb cores right after start up. We have only allocated 98Mb to the heap for now and this is enough for such a small index. > solrcloud crashes on startup due to excessive memory consumption > > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Priority: Critical > Fix For: 4.1 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) solrcloud crashes on startup due to excessive memory consumption
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13423841#comment-13423841 ] Markus Jelsma commented on SOLR-3685: - Ok, i have increased my DocumentCache again to reproduce the problem and configured from -XX:MaxDirectMemorySize=100m to 10m but RES is still climbing at the same rate as before so no change. We don't use Tika only Zookeeper. About virtual memory. That also climbes to ~800Mb which is many times more than the index size. There are no pending commits or merges right after start up. There may be some cloud replication related process that eats the RAM. Thanks > solrcloud crashes on startup due to excessive memory consumption > > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Priority: Critical > Fix For: 4.1 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) solrcloud crashes on startup due to excessive memory consumption
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13423801#comment-13423801 ] Markus Jelsma commented on SOLR-3685: - Hi - i don't look at virtual memory but RESident memory. My Solr install here will eat up to 512MB RESIDENT MEMORY and is killed by the OS. The virtual memory will then be almost 800MB, while both indexes are just 27MB in size. This sounds a lot of VIRT and RES for a tiny index and tiny heap. Also, Solr will run fine and fast with just 100MB of memory, the index is still very small. Thanks > solrcloud crashes on startup due to excessive memory consumption > > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Priority: Critical > Fix For: 4.1 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) solrcloud crashes on startup due to excessive memory consumption
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13423794#comment-13423794 ] Markus Jelsma commented on SOLR-3685: - Java 1.6.0-26 64bit, just as Linux. I should also note now that i made an error in the configuration. I thought i had reduced the DocumentCache size to 64 but the node it was testing on had a size of 1024 configured and redistributed the config over the cluster via config bootstrap. This still leaves the problem that Solr itself should run out of memory and not the OS as the cache is part of the heap. It also should clean old index directories. So this issue may consist of multiple problems. > solrcloud crashes on startup due to excessive memory consumption > > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Priority: Critical > Fix For: 4.1 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) solrcloud crashes on startup due to excessive memory consumption
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13423786#comment-13423786 ] Markus Jelsma commented on SOLR-3685: - I should have added this. I allocate just 98MB to the heap and 32 to the permgen so there just 130MB allocated. > solrcloud crashes on startup due to excessive memory consumption > > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Priority: Critical > Fix For: 4.1 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3685) solrcloud crashes on startup due to excessive memory consumption
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13423782#comment-13423782 ] Markus Jelsma commented on SOLR-3685: - I forgot to add that it doesn't matter if updates are sent to the cluster. A node will start to replicate on startup when it's update to date as well and crash subsequently. > solrcloud crashes on startup due to excessive memory consumption > > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Priority: Critical > Fix For: 4.1 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3685) solrcloud crashes on startup due to excessive memory consumption
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3685: Attachment: info.log Here's a log for a node where the Java process is being killed by the OS. I can reproduce this consistently. > solrcloud crashes on startup due to excessive memory consumption > > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Priority: Critical > Fix For: 4.1 > > Attachments: info.log > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3685) solrcloud crashes on startup due to excessive memory consumption
[ https://issues.apache.org/jira/browse/SOLR-3685?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3685: Summary: solrcloud crashes on startup due to excessive memory consumption (was: Solr ) > solrcloud crashes on startup due to excessive memory consumption > > > Key: SOLR-3685 > URL: https://issues.apache.org/jira/browse/SOLR-3685 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 4.0-ALPHA > Environment: Debian GNU/Linux Squeeze 64bit > Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 >Reporter: Markus Jelsma >Priority: Critical > Fix For: 4.1 > > > There's a serious problem with restarting nodes, not cleaning old or unused > index directories and sudden replication and Java being killed by the OS due > to excessive memory allocation. Since SOLR-1781 was fixed index directories > get cleaned up when a node is being restarted cleanly, however, old or unused > index directories still pile up if Solr crashes or is being killed by the OS, > happening here. > We have a six-node 64-bit Linux test cluster with each node having two > shards. There's 512MB RAM available and no swap. Each index is roughly 27MB > so about 50MB per node, this fits easily and works fine. However, if a node > is being restarted, Solr will consistently crash because it immediately eats > up all RAM. If swap is enabled Solr will eat an additional few 100MB's right > after start up. > This cannot be solved by restarting Solr, it will just crash again and leave > index directories in place until the disk is full. The only way i can restart > a node safely is to delete the index directories and have it replicate from > another node. If i then restart the node it will crash almost consistently. > I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3685) Solr
Markus Jelsma created SOLR-3685: --- Summary: Solr Key: SOLR-3685 URL: https://issues.apache.org/jira/browse/SOLR-3685 Project: Solr Issue Type: Bug Components: replication (java), SolrCloud Affects Versions: 4.0-ALPHA Environment: Debian GNU/Linux Squeeze 64bit Solr 5.0-SNAPSHOT 1365667M - markus - 2012-07-25 19:09:43 Reporter: Markus Jelsma Priority: Critical Fix For: 4.1 There's a serious problem with restarting nodes, not cleaning old or unused index directories and sudden replication and Java being killed by the OS due to excessive memory allocation. Since SOLR-1781 was fixed index directories get cleaned up when a node is being restarted cleanly, however, old or unused index directories still pile up if Solr crashes or is being killed by the OS, happening here. We have a six-node 64-bit Linux test cluster with each node having two shards. There's 512MB RAM available and no swap. Each index is roughly 27MB so about 50MB per node, this fits easily and works fine. However, if a node is being restarted, Solr will consistently crash because it immediately eats up all RAM. If swap is enabled Solr will eat an additional few 100MB's right after start up. This cannot be solved by restarting Solr, it will just crash again and leave index directories in place until the disk is full. The only way i can restart a node safely is to delete the index directories and have it replicate from another node. If i then restart the node it will crash almost consistently. I'll attach a log of one of the nodes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13423779#comment-13423779 ] Markus Jelsma commented on SOLR-1781: - There's still a problem with old index directories not being cleaned up and strange replication on start up. I'll write to the ML for this, the problem is likely larger than just cleaning up. > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13423022#comment-13423022 ] Markus Jelsma commented on SOLR-1781: - Yes, the problem no longer occurs! Great work! Thanks > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13422526#comment-13422526 ] Markus Jelsma commented on SOLR-1781: - It seems this fixes the issue. I'll double check tomorrow! > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13422308#comment-13422308 ] Markus Jelsma commented on SOLR-1781: - Ok, I purged the logs, enabled info and started a tomcat. New indexes are created shortly after : 2012-07-25 10:13:36,125 WARN [solr.core.SolrCore] - [main] - : New index directory detected: old=null new=/opt/solr/cores/openindex_b/data/index.20120725101231289 I'll send it right now. > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13422135#comment-13422135 ] Markus Jelsma commented on SOLR-1781: - Hi, I'll restart one node with two cores. {code} #cat cores/openindex_b/data/index.properties #index properties #Wed Jul 25 09:58:26 UTC 2012 index=index.20120725095644707 {code} {code} # du -h cores/ 4.0Kcores/lib 46M cores/openindex_b/data/index.20120725095644707 404Kcores/openindex_b/data/tlog 46M cores/openindex_b/data 46M cores/openindex_b 98M cores/openindex_a/data/index.20120725095843731 124Kcores/openindex_a/data/tlog 98M cores/openindex_a/data 98M cores/openindex_a 144Mcores/ {code} 2012-07-25 10:01:09,176 WARN [solr.core.SolrCore] - [main] - : New index directory detected: old=null new=/opt/solr/cores/openindex_b/data/index.20120725095644707 ... 2012-07-25 10:01:17,303 WARN [solr.core.SolrCore] - [main] - : New index directory detected: old=null new=/opt/solr/cores/openindex_a/data/index.20120725095843731 ... 2012-07-25 10:01:55,016 WARN [solr.core.SolrCore] - [RecoveryThread] - : New index directory detected: old=/opt/solr/cores/openindex_b/data/index.20120725095644707 new=/opt/solr/cores/openindex_b/data/index.20120725100120496 ... 2012-07-25 10:03:35,236 WARN [solr.core.SolrCore] - [RecoveryThread] - : New index directory detected: old=/opt/solr/cores/openindex_a/data/index.20120725100220706 new=/opt/solr/cores/openindex_a/data/index.20120725100321897 {code} # du -h cores/ 4.0Kcores/lib 46M cores/openindex_b/data/index.20120725095644707 404Kcores/openindex_b/data/tlog 46M cores/openindex_b/data/index.20120725100120496 91M cores/openindex_b/data 91M cores/openindex_b 98M cores/openindex_a/data/index.20120725100321897 98M cores/openindex_a/data/index.20120725100220706 124Kcores/openindex_a/data/tlog 196Mcores/openindex_a/data 196Mcores/openindex_a 287Mcores/ {code} A few minutes later we still have multiple index directories. No updates have been sent to the cluster during this whole scenario. Each time another directory appears it comes with a lot of I/O, on these RAM limited machines it's almost trashing because of the additional directory. It does not create another directory on each restart but sometimes does, it restarted the same machine again and now i have three dirs for each core. I'll turn on INFO logging for the node and restart it again without deleting the surpluss dirs. The master and slave versions are still the same. {code} # du -h cores/ 4.0Kcores/lib 46M cores/openindex_b/data/index.20120725100813961 42M cores/openindex_b/data/index.20120725101349376 46M cores/openindex_b/data/index.20120725095644707 46M cores/openindex_b/data/index.20120725101231289 404Kcores/openindex_b/data/tlog 46M cores/openindex_b/data/index.20120725100120496 223Mcores/openindex_b/data 223Mcores/openindex_b 98M cores/openindex_a/data/index.20120725101252920 98M cores/openindex_a/data/index.20120725100220706 124Kcores/openindex_a/data/tlog 196Mcores/openindex_a/data 196Mcores/openindex_a 418Mcores/ {code} Maybe it cannot find the current index directory on start up (in my case). 2012-07-25 10:13:36,125 WARN [solr.core.SolrCore] - [main] - : New index directory detected: old=null new=/opt/solr/cores/openindex_b/data/index.20120725101231289 2012-07-25 10:13:45,840 WARN [solr.core.SolrCore] - [main] - : New index directory detected: old=null new=/opt/solr/cores/openindex_a/data/index.20120725101252920 2012-07-25 10:15:41,393 WARN [solr.core.SolrCore] - [RecoveryThread] - : New index directory detected: old=/opt/solr/cores/openindex_b/data/index.20120725101231289 new=/opt/solr/cores/openindex_b/data/index.20120725101349376 2012-07-25 10:15:46,895 WARN [solr.cloud.RecoveryStrategy] - [main-EventThread] - : Stopping recovery for core openindex_b zkNodeName=nl2.index.openindex.io:8080_solr_openindex_b 2012-07-25 10:15:46,952 WARN [solr.core.SolrCore] - [RecoveryThread] - : [openindex_a] Error opening new searcher. exceeded limit of maxWarmingSearchers=1, try again later. 2012-07-25 10:15:47,298 ERROR [solr.cloud.RecoveryStrategy] - [RecoveryThread] - : Error while trying to recover. org.apache.solr.common.SolrException: Error opening new searcher. exceeded limit of maxWarmingSearchers=1, try again later. at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1365) at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:1157) at org.apache.solr.update.DirectUpdateHandler2.commit(DirectUpdateHandler2.java:560) at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:316) at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:210) 2012-07-25 10:15:47,299 ERROR [solr.cloud.RecoveryStrategy] - [RecoveryThre
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13421675#comment-13421675 ] Markus Jelsma commented on SOLR-1781: - Besides search and index only the occasional restart when i change some config or deploy a new build. Sometimes i need to start ZK 3.4 again because it died for some reason. Restarting Tomcat a few times in a row may be a clue here. I'll check again tomorrow if whether it's consistent. > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13421565#comment-13421565 ] Markus Jelsma commented on SOLR-1781: - Hi, I've deleted both today and yesterday indexes of more than a few hours old. Some are being removed indeed and some persist. I just restarted all nodes (introduced new fieldTypes and one field) and at least one node has three index directories. Others had two, some just one. Not a single node has a `unable to delete` string in the logs. Thanks > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13421323#comment-13421323 ] Markus Jelsma commented on SOLR-1781: - One of the nodes ended up with two index directories today. Later some other nodes also didn't clean up after they got restarted. > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13420495#comment-13420495 ] Markus Jelsma commented on SOLR-1781: - The problem is resolved. The `bad` node created several new index.[0-9] directories, even with this patch, and caused high I/O. I deleted the complete data directory so also the index.properties file. It loaded its index from the other nodes and no longer created many index dirs. Thanks > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13419385#comment-13419385 ] Markus Jelsma commented on SOLR-1781: - Strange indeed. I can/could replicate it on one machine consistently and not on others. Machines weren't upgraded at the same time to prevent cluster downtime. I'll check back monday, there are two other machines left to upgrade plus the bad node. > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13419127#comment-13419127 ] Markus Jelsma commented on SOLR-1781: - Log sent. This node has two shards on it and executed 2x 512 warmup queries which adds up. It won't talk to ZK (tail of the log). Restarting the node with an 18th's build works fine. Did it three times today. Thanks > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13419114#comment-13419114 ] Markus Jelsma commented on SOLR-1781: - The node will never respond to HTTP requests, all ZK connections time out, very high resource consumption. I'll try provide a log snippet soon. I tried running today's build several times but one specific node refuses to `come online`. Another node did well and runs today's build. I cannot attach a file to a resolved issue. Send over mail? > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13419032#comment-13419032 ] Markus Jelsma commented on SOLR-1781: - Hi, is the core reloading still part of this? I get a lot of firstSearcher events on a test node now and it won't get online. Going back to July 18th (before this patch) build works fine. Other nodes won't come online with a build from the 19th (after this patch). > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0, 5.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch, > SOLR-1781.patch, SOLR-1781.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3644) debug 0.0 process time with distributed search
Markus Jelsma created SOLR-3644: --- Summary: debug 0.0 process time with distributed search Key: SOLR-3644 URL: https://issues.apache.org/jira/browse/SOLR-3644 Project: Solr Issue Type: Bug Components: SearchComponents - other, SolrCloud Affects Versions: 4.0-ALPHA Environment: 5.0.0.2012.07.18.15.28.59 Reporter: Markus Jelsma Priority: Minor Fix For: 4.1 With debugQuery enabled we usually see processTime information for all search components. With distributed search only the processTime for the query, highlight and debug components are non-zero. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13416318#comment-13416318 ] Markus Jelsma commented on SOLR-1781: - Perhaps they could be cleaned up on core start or after some time has passed? > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Mark Miller > Fix For: 4.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13416183#comment-13416183 ] Markus Jelsma commented on SOLR-1781: - We don't have that, i should have included it in my comment. All servers run Debian GNU/Linux 6.0 and the cloud test cluster always runs with a very recent build from trunk. > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java), SolrCloud >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Noble Paul > Fix For: 4.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1781) Replication index directories not always cleaned up
[ https://issues.apache.org/jira/browse/SOLR-1781?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13416005#comment-13416005 ] Markus Jelsma commented on SOLR-1781: - This happens almost daily on our SolrCloud (trunk) test cluster, we sometimes see four surpluss index directories created in a day. > Replication index directories not always cleaned up > --- > > Key: SOLR-1781 > URL: https://issues.apache.org/jira/browse/SOLR-1781 > Project: Solr > Issue Type: Bug > Components: replication (java) >Affects Versions: 1.4 > Environment: Windows Server 2003 R2, Java 6b18 >Reporter: Terje Sten Bjerkseth >Assignee: Noble Paul > Fix For: 4.0 > > Attachments: > 0001-Replication-does-not-always-clean-up-old-directories.patch > > > We had the same problem as someone described in > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201001.mbox/%3c222a518d-ddf5-4fc8-a02a-74d4f232b...@snooth.com%3e. > A partial copy of that message: > We're using the new replication and it's working pretty well. There's > one detail I'd like to get some more information about. > As the replication works, it creates versions of the index in the data > directory. Originally we had index/, but now there are dated versions > such as index.20100127044500/, which are the replicated versions. > Each copy is sized in the vicinity of 65G. With our current hard drive > it's fine to have two around, but 3 gets a little dicey. Sometimes > we're finding that the replication doesn't always clean up after > itself. I would like to understand this better, or to not have this > happen. It could be a configuration issue. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3488) Create a Collections API for SolrCloud
[ https://issues.apache.org/jira/browse/SOLR-3488?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13412927#comment-13412927 ] Markus Jelsma commented on SOLR-3488: - Thanks for claryfing, it makes sense. About the downtime on core reload, a load balancer pinging Solr's admin/ping handler will definately mark the node as down; the ping request will time out for up to a few seconds or even longer in case of many firstSearcher events. > Create a Collections API for SolrCloud > -- > > Key: SOLR-3488 > URL: https://issues.apache.org/jira/browse/SOLR-3488 > Project: Solr > Issue Type: New Feature > Components: SolrCloud >Reporter: Mark Miller >Assignee: Mark Miller > Fix For: 4.0 > > Attachments: SOLR-3488.patch, SOLR-3488.patch, SOLR-3488.patch, > SOLR-3488_2.patch > > -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3488) Create a Collections API for SolrCloud
[ https://issues.apache.org/jira/browse/SOLR-3488?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13412874#comment-13412874 ] Markus Jelsma commented on SOLR-3488: - Is it intended for a collection RELOAD action to reload all collection cores immediately? That implies downtime i assume? > Create a Collections API for SolrCloud > -- > > Key: SOLR-3488 > URL: https://issues.apache.org/jira/browse/SOLR-3488 > Project: Solr > Issue Type: New Feature > Components: SolrCloud >Reporter: Mark Miller >Assignee: Mark Miller > Fix For: 4.0 > > Attachments: SOLR-3488.patch, SOLR-3488.patch, SOLR-3488.patch, > SOLR-3488_2.patch > > -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-3564) SpellcheckCollator NPE with timeAllowed set
[ https://issues.apache.org/jira/browse/SOLR-3564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13409271#comment-13409271 ] Markus Jelsma edited comment on SOLR-3564 at 7/9/12 8:55 AM: - It it in some cases also sent over the wire in response to a legitimate request: edit: this is actually triggered by something else. was (Author: markus17): It it in some cases also sent over the wire in response to a legitimate request: {code} {"response":{"numFound":0,"start":0,"maxScore":0.0,"docs":[]},"facet_counts":{"facet_queries":{},"facet_fields":{"type":{},"host":{},"cat":{}},"facet_dates":{},"facet_ranges":{}},"highlighting":{},"error":{"trace":"java.lang.NullPointerException\n\tat org.apache.solr.handler.component.SpellCheckComponent.finishStage(SpellCheckComponent.java:297)\n\tat org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:315)\n\tat org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)\n\tat org.apache.solr.core.SolrCore.execute(SolrCore.java:1599)\n\tat org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:442) {code} > SpellcheckCollator NPE with timeAllowed set > --- > > Key: SOLR-3564 > URL: https://issues.apache.org/jira/browse/SOLR-3564 > Project: Solr > Issue Type: Bug > Components: spellchecker >Affects Versions: 4.0 > Environment: 5.0-SNAPSHOT 1352525M - markus - 2012-06-21 15:23:39 >Reporter: Markus Jelsma >Priority: Minor > Fix For: 4.0 > > > If the query running time is exceeded during collation checking the > SpellcheckCollator throws the following NPE: > {code} > 2012-06-21 14:34:12,875 WARN [solr.spelling.SpellCheckCollator] - > [http-8080-exec-28] - : Exception trying to re-query to check if a spell > check possi > bility would return any hits. > java.lang.NullPointerException > at > org.apache.solr.handler.component.ResponseBuilder.setResult(ResponseBuilder.java:399) > at > org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:412) > at > org.apache.solr.spelling.SpellCheckCollator.collate(SpellCheckCollator.java:112) > at > org.apache.solr.handler.component.SpellCheckComponent.addCollationsToResponse(SpellCheckComponent.java:203) > at > org.apache.solr.handler.component.SpellCheckComponent.process(SpellCheckComponent.java:180) > at > org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:204) > at > org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129) > at org.apache.solr.core.SolrCore.execute(SolrCore.java:1561) > > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3564) SpellcheckCollator NPE with timeAllowed set
[ https://issues.apache.org/jira/browse/SOLR-3564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13409271#comment-13409271 ] Markus Jelsma commented on SOLR-3564: - It it in some cases also sent over the wire in response to a legitimate request: {code} {"response":{"numFound":0,"start":0,"maxScore":0.0,"docs":[]},"facet_counts":{"facet_queries":{},"facet_fields":{"type":{},"host":{},"cat":{}},"facet_dates":{},"facet_ranges":{}},"highlighting":{},"error":{"trace":"java.lang.NullPointerException\n\tat org.apache.solr.handler.component.SpellCheckComponent.finishStage(SpellCheckComponent.java:297)\n\tat org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:315)\n\tat org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129)\n\tat org.apache.solr.core.SolrCore.execute(SolrCore.java:1599)\n\tat org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:442) {code} > SpellcheckCollator NPE with timeAllowed set > --- > > Key: SOLR-3564 > URL: https://issues.apache.org/jira/browse/SOLR-3564 > Project: Solr > Issue Type: Bug > Components: spellchecker >Affects Versions: 4.0 > Environment: 5.0-SNAPSHOT 1352525M - markus - 2012-06-21 15:23:39 >Reporter: Markus Jelsma >Priority: Minor > Fix For: 4.0 > > > If the query running time is exceeded during collation checking the > SpellcheckCollator throws the following NPE: > {code} > 2012-06-21 14:34:12,875 WARN [solr.spelling.SpellCheckCollator] - > [http-8080-exec-28] - : Exception trying to re-query to check if a spell > check possi > bility would return any hits. > java.lang.NullPointerException > at > org.apache.solr.handler.component.ResponseBuilder.setResult(ResponseBuilder.java:399) > at > org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:412) > at > org.apache.solr.spelling.SpellCheckCollator.collate(SpellCheckCollator.java:112) > at > org.apache.solr.handler.component.SpellCheckComponent.addCollationsToResponse(SpellCheckComponent.java:203) > at > org.apache.solr.handler.component.SpellCheckComponent.process(SpellCheckComponent.java:180) > at > org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:204) > at > org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129) > at org.apache.solr.core.SolrCore.execute(SolrCore.java:1561) > > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3564) SpellcheckCollator NPE with timeAllowed set
Markus Jelsma created SOLR-3564: --- Summary: SpellcheckCollator NPE with timeAllowed set Key: SOLR-3564 URL: https://issues.apache.org/jira/browse/SOLR-3564 Project: Solr Issue Type: Bug Components: spellchecker Affects Versions: 4.0 Environment: 5.0-SNAPSHOT 1352525M - markus - 2012-06-21 15:23:39 Reporter: Markus Jelsma Fix For: 4.0 If the query running time is exceeded during collation checking the SpellcheckCollator throws the following NPE: {code} 2012-06-21 14:34:12,875 WARN [solr.spelling.SpellCheckCollator] - [http-8080-exec-28] - : Exception trying to re-query to check if a spell check possi bility would return any hits. java.lang.NullPointerException at org.apache.solr.handler.component.ResponseBuilder.setResult(ResponseBuilder.java:399) at org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:412) at org.apache.solr.spelling.SpellCheckCollator.collate(SpellCheckCollator.java:112) at org.apache.solr.handler.component.SpellCheckComponent.addCollationsToResponse(SpellCheckComponent.java:203) at org.apache.solr.handler.component.SpellCheckComponent.process(SpellCheckComponent.java:180) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:204) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1561) {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3557) Avoid NPE for distributed request when shards.tolerant=true
[ https://issues.apache.org/jira/browse/SOLR-3557?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13396638#comment-13396638 ] Markus Jelsma commented on SOLR-3557: - Patch works fine! > Avoid NPE for distributed request when shards.tolerant=true > --- > > Key: SOLR-3557 > URL: https://issues.apache.org/jira/browse/SOLR-3557 > Project: Solr > Issue Type: Improvement > Components: search >Reporter: Ryan McKinley >Assignee: Ryan McKinley >Priority: Minor > Fix For: 4.0 > > Attachments: SOLR-3557-tolerant-faceting.patch > > > If a shard fails to return and shards.tolerant=true, the faceting module will > get a null pointer. We should avoid that NPE -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3518) No `hits` in SolrResp. NamedList if distrib=true
[ https://issues.apache.org/jira/browse/SOLR-3518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3518: Attachment: SOLR-3518-4.0-1.patch Patch for trunk adding the `hits` field to the SolrQueryResponse's NamedList. It's only returned in the final response, not in intermediate shardrequests in a distributed search. Most likely not a good solution but it seems to work fine for now. Please improve. > No `hits` in SolrResp. NamedList if distrib=true > > > Key: SOLR-3518 > URL: https://issues.apache.org/jira/browse/SOLR-3518 > Project: Solr > Issue Type: Bug > Components: SolrCloud >Affects Versions: 4.0 > Environment: 5.0-SNAPSHOT 1346798 - markus - 2012-06-06 11:38:15 >Reporter: Markus Jelsma >Priority: Minor > Fix For: 4.0 > > Attachments: SOLR-3518-4.0-1.patch > > > The hits field in the NamedList obtained from SolrQueryResponse.toLog() is > not available for distrib=true requests. The hits fields is also not written > to the log. > See also:: > http://lucene.472066.n3.nabble.com/SolrDispatchFilter-no-hits-in-response-NamedList-if-distrib-true-td3987751.html -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3525) Per-field similarity should display used impl. in debug output broken
Markus Jelsma created SOLR-3525: --- Summary: Per-field similarity should display used impl. in debug output broken Key: SOLR-3525 URL: https://issues.apache.org/jira/browse/SOLR-3525 Project: Solr Issue Type: Bug Components: search Affects Versions: 4.0 Reporter: Markus Jelsma Priority: Minor Fix For: 4.0 When using per-field similarity debugQuery should display the used similarity implementation for each match. Right now it's broken and displays empty brackets: 112.33515 = (MATCH) weight(content:blah in 273) [], result of: -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3518) No `hits` in SolrResp. NamedList if distrib=true
Markus Jelsma created SOLR-3518: --- Summary: No `hits` in SolrResp. NamedList if distrib=true Key: SOLR-3518 URL: https://issues.apache.org/jira/browse/SOLR-3518 Project: Solr Issue Type: Bug Components: SolrCloud Affects Versions: 4.0 Environment: 5.0-SNAPSHOT 1346798 - markus - 2012-06-06 11:38:15 Reporter: Markus Jelsma Priority: Minor Fix For: 4.0 The hits field in the NamedList obtained from SolrQueryResponse.toLog() is not available for distrib=true requests. The hits fields is also not written to the log. See also:: http://lucene.472066.n3.nabble.com/SolrDispatchFilter-no-hits-in-response-NamedList-if-distrib-true-td3987751.html -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3238) Sequel of Admin UI
[ https://issues.apache.org/jira/browse/SOLR-3238?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13281607#comment-13281607 ] Markus Jelsma commented on SOLR-3238: - I think it would also be useful to display the shard information in the core overview page such as its ID and whether it is a leader. > Sequel of Admin UI > -- > > Key: SOLR-3238 > URL: https://issues.apache.org/jira/browse/SOLR-3238 > Project: Solr > Issue Type: Improvement > Components: web gui >Affects Versions: 4.0 >Reporter: Stefan Matheis (steffkes) >Assignee: Stefan Matheis (steffkes) > Fix For: 4.0 > > Attachments: SOLR-3238.patch, SOLR-3238.patch, SOLR-3238.patch, > solradminbug.png > > > Catch-All Issue for all upcoming Bugs/Reports/Suggestions on the Admin UI -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3473) Distributed deduplication broken
[ https://issues.apache.org/jira/browse/SOLR-3473?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13280264#comment-13280264 ] Markus Jelsma commented on SOLR-3473: - That makes sense indeed. To work around the problem of having the digest field as ID, could it not simply issue a deleteByQuery for the digest prior to adding it? Would that cause significant overhead for very large systems with many updates? We would, from Nutch' point of view, certainly want to avoid changing the ID from URL to digest. > Distributed deduplication broken > > > Key: SOLR-3473 > URL: https://issues.apache.org/jira/browse/SOLR-3473 > Project: Solr > Issue Type: Bug > Components: SolrCloud, update >Affects Versions: 4.0 >Reporter: Markus Jelsma > Fix For: 4.0 > > > Solr's deduplication via the SignatureUpdateProcessor is broken for > distributed updates on SolrCloud. > Mark Miller: > {quote} > Looking again at the SignatureUpdateProcessor code, I think that indeed this > won't currently work with distrib updates. Could you file a JIRA issue for > that? The problem is that we convert update commands into solr documents - > and that can cause a loss of info if an update proc modifies the update > command. > I think the reason that you see a multiple values error when you try the > other order is because of the lack of a document clone (the other issue I > mentioned a few emails back). Addressing that won't solve your issue though - > we have to come up with a way to propagate the currently lost info on the > update command. > {quote} > Please see the ML thread for the full discussion: > http://lucene.472066.n3.nabble.com/SolrCloud-deduplication-td3984657.html -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3473) Distributed deduplication broken
Markus Jelsma created SOLR-3473: --- Summary: Distributed deduplication broken Key: SOLR-3473 URL: https://issues.apache.org/jira/browse/SOLR-3473 Project: Solr Issue Type: Bug Components: SolrCloud, update Affects Versions: 4.0 Reporter: Markus Jelsma Fix For: 4.0 Solr's deduplication via the SignatureUpdateProcessor is broken for distributed updates on SolrCloud. Mark Miller: {quote} Looking again at the SignatureUpdateProcessor code, I think that indeed this won't currently work with distrib updates. Could you file a JIRA issue for that? The problem is that we convert update commands into solr documents - and that can cause a loss of info if an update proc modifies the update command. I think the reason that you see a multiple values error when you try the other order is because of the lack of a document clone (the other issue I mentioned a few emails back). Addressing that won't solve your issue though - we have to come up with a way to propagate the currently lost info on the update command. {quote} Please see the ML thread for the full discussion: http://lucene.472066.n3.nabble.com/SolrCloud-deduplication-td3984657.html -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3457) Spellchecker always incorrectly spelled
[ https://issues.apache.org/jira/browse/SOLR-3457?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-3457: Attachment: SOLR-3457-4.0-1.patch Patch for trunk. It seems the isCorrectlySpelled flag is not correctly initialized. In the example samsun is incorrectly spelled, has freqInfo and zero suggestions so it's never set to true. > Spellchecker always incorrectly spelled > --- > > Key: SOLR-3457 > URL: https://issues.apache.org/jira/browse/SOLR-3457 > Project: Solr > Issue Type: Bug > Components: spellchecker >Affects Versions: 4.0 > Environment: solr-spec 4.0.0.2012.05.15.11.42.06 > solr-impl 4.0-SNAPSHOT 1338601 - markus - 2012-05-15 11:42:06 > lucene-spec 4.0-SNAPSHOT > lucene-impl 4.0-SNAPSHOT 1338601 - markus - 2012-05-15 10:51:02 >Reporter: Markus Jelsma > Attachments: SOLR-3457-4.0-1.patch > > > correctlySpelled is always false with default configuration, example config > and example documents: > http://localhost:8983/solr/collection1/browse?wt=xml&spellcheck.extendedResults=true&q=samsung > {code} > > >false > > > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3238) Sequel of Admin UI
[ https://issues.apache.org/jira/browse/SOLR-3238?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13276647#comment-13276647 ] Markus Jelsma commented on SOLR-3238: - There's a small issue on the analysis page. After submitting the form with whitespace in the fields they are printed again still being url encoded with + for whitespace. > Sequel of Admin UI > -- > > Key: SOLR-3238 > URL: https://issues.apache.org/jira/browse/SOLR-3238 > Project: Solr > Issue Type: Improvement > Components: web gui >Affects Versions: 4.0 >Reporter: Stefan Matheis (steffkes) >Assignee: Stefan Matheis (steffkes) > Fix For: 4.0 > > Attachments: SOLR-3238.patch, SOLR-3238.patch, SOLR-3238.patch, > solradminbug.png > > > Catch-All Issue for all upcoming Bugs/Reports/Suggestions on the Admin UI -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3457) Spellchecker always incorrectly spelled
Markus Jelsma created SOLR-3457: --- Summary: Spellchecker always incorrectly spelled Key: SOLR-3457 URL: https://issues.apache.org/jira/browse/SOLR-3457 Project: Solr Issue Type: Bug Components: spellchecker Affects Versions: 4.0 Environment: solr-spec 4.0.0.2012.05.15.11.42.06 solr-impl 4.0-SNAPSHOT 1338601 - markus - 2012-05-15 11:42:06 lucene-spec 4.0-SNAPSHOT lucene-impl 4.0-SNAPSHOT 1338601 - markus - 2012-05-15 10:51:02 Reporter: Markus Jelsma correctlySpelled is always false with default configuration, example config and example documents: http://localhost:8983/solr/collection1/browse?wt=xml&spellcheck.extendedResults=true&q=samsung {code} false {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3221) Make Shard handler threadpool configurable
[ https://issues.apache.org/jira/browse/SOLR-3221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13273514#comment-13273514 ] Markus Jelsma commented on SOLR-3221: - I would agree that latency is preferred as default. > Make Shard handler threadpool configurable > -- > > Key: SOLR-3221 > URL: https://issues.apache.org/jira/browse/SOLR-3221 > Project: Solr > Issue Type: Improvement >Affects Versions: 3.6, 4.0 >Reporter: Greg Bowyer >Assignee: Erick Erickson > Labels: distributed, http, shard > Fix For: 3.6, 4.0 > > Attachments: SOLR-3221-3x_branch.patch, SOLR-3221-3x_branch.patch, > SOLR-3221-3x_branch.patch, SOLR-3221-3x_branch.patch, > SOLR-3221-3x_branch.patch, SOLR-3221-trunk.patch, SOLR-3221-trunk.patch, > SOLR-3221-trunk.patch, SOLR-3221-trunk.patch, SOLR-3221-trunk.patch > > > From profiling of monitor contention, as well as observations of the > 95th and 99th response times for nodes that perform distributed search > (or ‟aggregator‟ nodes) it would appear that the HttpShardHandler code > currently does a suboptimal job of managing outgoing shard level > requests. > Presently the code contained within lucene 3.5's SearchHandler and > Lucene trunk / 3x's ShardHandlerFactory create arbitrary threads in > order to service distributed search requests. This is done presently to > limit the size of the threadpool such that it does not consume resources > in deployment configurations that do not use distributed search. > This unfortunately has two impacts on the response time if the node > coordinating the distribution is under high load. > The usage of the MaxConnectionsPerHost configuration option results in > aggressive activity on semaphores within HttpCommons, it has been > observed that the aggregator can have a response time far greater than > that of the searchers. The above monitor contention would appear to > suggest that in some cases its possible for liveness issues to occur and > for simple queries to be starved of resources simply due to a lack of > attention from the viewpoint of context switching. > With, as mentioned above the http commons connection being hotly > contended > The fair, queue based configuration eliminates this, at the cost of > throughput. > This patch aims to make the threadpool largely configurable allowing for > those using solr to choose the throughput vs latency balance they > desire. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1979) Create LanguageIdentifierUpdateProcessor
[ https://issues.apache.org/jira/browse/SOLR-1979?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13102578#comment-13102578 ] Markus Jelsma commented on SOLR-1979: - Hi. This is not what i understood from reading the wiki doc. Will the update processor skip detection with these settings? It's rather costly on many docs. Anyway, this is great work already, thanks! > Create LanguageIdentifierUpdateProcessor > > > Key: SOLR-1979 > URL: https://issues.apache.org/jira/browse/SOLR-1979 > Project: Solr > Issue Type: New Feature > Components: update >Reporter: Jan Høydahl >Assignee: Jan Høydahl >Priority: Minor > Labels: UpdateProcessor > Fix For: 3.5 > > Attachments: SOLR-1979.patch, SOLR-1979.patch, SOLR-1979.patch, > SOLR-1979.patch, SOLR-1979.patch, SOLR-1979.patch, SOLR-1979.patch > > > Language identification from document fields, and mapping of field names to > language-specific fields based on detected language. > Wrap the Tika LanguageIdentifier in an UpdateProcessor. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1979) Create LanguageIdentifierUpdateProcessor
[ https://issues.apache.org/jira/browse/SOLR-1979?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13102520#comment-13102520 ] Markus Jelsma commented on SOLR-1979: - Hi Jan, Can we also use the mapping feature without detection? Our detection is done in a Nutch cluster so we already identified many millions of docs. Thanks > Create LanguageIdentifierUpdateProcessor > > > Key: SOLR-1979 > URL: https://issues.apache.org/jira/browse/SOLR-1979 > Project: Solr > Issue Type: New Feature > Components: update >Reporter: Jan Høydahl >Assignee: Jan Høydahl >Priority: Minor > Labels: UpdateProcessor > Fix For: 3.5 > > Attachments: SOLR-1979.patch, SOLR-1979.patch, SOLR-1979.patch, > SOLR-1979.patch, SOLR-1979.patch, SOLR-1979.patch, SOLR-1979.patch > > > Language identification from document fields, and mapping of field names to > language-specific fields based on detected language. > Wrap the Tika LanguageIdentifier in an UpdateProcessor. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-1863) spellchecker leaks file on core reload
[ https://issues.apache.org/jira/browse/SOLR-1863?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13089390#comment-13089390 ] Markus Jelsma commented on SOLR-1863: - Is this still an issue in 3.3 or 4.x? > spellchecker leaks file on core reload > -- > > Key: SOLR-1863 > URL: https://issues.apache.org/jira/browse/SOLR-1863 > Project: Solr > Issue Type: Bug > Components: spellchecker > Environment: linux i386 (ubuntu 8.04) >Reporter: Arne de Bruijn > Attachments: SOLR-1863.patch > > > When reloading a core of a multicore solr 1.4.0 instance with > /admin/cores?action=RELOAD&core=name an extra reference to the spellchecker > cfs file appears in the list of open files of the java process. A forced gc > (with jconsole) does not help. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-2689) !frange with query($qq) sets score=1.0f for all returned documents
[ https://issues.apache.org/jira/browse/SOLR-2689?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13078467#comment-13078467 ] Markus Jelsma commented on SOLR-2689: - You are right, it's because both examples use one search term and thus all have the same score. It shows when not all scores are identical when you use multiple terms. I'll provide a better description and example next week when i'll get back. > !frange with query($qq) sets score=1.0f for all returned documents > -- > > Key: SOLR-2689 > URL: https://issues.apache.org/jira/browse/SOLR-2689 > Project: Solr > Issue Type: Bug > Components: search >Affects Versions: 3.4 >Reporter: Markus Jelsma > Fix For: 3.4, 4.0 > > > Consider the following queries, both query the default field for 'test' and > return the document digest and score (i don't seem to be able get only score, > fl=score returns all fields): > This is a normal query and yields normal results with proper scores: > {code} > q=test&fl=digest,score > {code} > {code} > > − > > 4.952673 > c48e784f06a051d89f20b72194b0dcf0 > > − > > 4.952673 > 7f78a504b8cbd86c6cdbf2aa2c4ae5e3 > > − > > 4.952673 > 0f7fefa6586ceda42fc1f095d460aa17 > > {code} > This query uses frange with query() to limit the number of returned > documents. When using multiple search terms i can indeed cut-off the result > set but in the end all returned documents have score=1.0f. The final result > set cannot be sorted by score anymore. The result set seems to be returned in > the order of Lucene docId's. > {code} > q={!frange l=1.23}query($qq)&qq=test&fl=digest,score > {code} > {code} > > − > > 1.0 > c48e784f06a051d89f20b72194b0dcf0 > > − > > 1.0 > 7f78a504b8cbd86c6cdbf2aa2c4ae5e3 > > − > > 1.0 > 0f7fefa6586ceda42fc1f095d460aa17 > > {code} -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-2689) !frange with query($qq) sets score=1.0f for all returned documents
!frange with query($qq) sets score=1.0f for all returned documents -- Key: SOLR-2689 URL: https://issues.apache.org/jira/browse/SOLR-2689 Project: Solr Issue Type: Bug Components: search Affects Versions: 3.4 Reporter: Markus Jelsma Fix For: 3.4, 4.0 Consider the following queries, both query the default field for 'test' and return the document digest and score (i don't seem to be able get only score, fl=score returns all fields): This is a normal query and yields normal results with proper scores: {code} q=test&fl=digest,score {code} {code} − 4.952673 c48e784f06a051d89f20b72194b0dcf0 − 4.952673 7f78a504b8cbd86c6cdbf2aa2c4ae5e3 − 4.952673 0f7fefa6586ceda42fc1f095d460aa17 {code} This query uses frange with query() to limit the number of returned documents. When using multiple search terms i can indeed cut-off the result set but in the end all returned documents have score=1.0f. The final result set cannot be sorted by score anymore. The result set seems to be returned in the order of Lucene docId's. {code} q={!frange l=1.23}query($qq)&qq=test&fl=digest,score {code} {code} − 1.0 c48e784f06a051d89f20b72194b0dcf0 − 1.0 7f78a504b8cbd86c6cdbf2aa2c4ae5e3 − 1.0 0f7fefa6586ceda42fc1f095d460aa17 {code} -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-2555) Always incorrectly spelled with onlyMorePopular enabled
[ https://issues.apache.org/jira/browse/SOLR-2555?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-2555: Fix Version/s: 3.4 > Always incorrectly spelled with onlyMorePopular enabled > --- > > Key: SOLR-2555 > URL: https://issues.apache.org/jira/browse/SOLR-2555 > Project: Solr > Issue Type: Bug > Components: spellchecker >Affects Versions: 3.1 >Reporter: Markus Jelsma > Fix For: 3.4 > > > The spellcheck component will always mark the term(s) as incorrectly spelled > when onlyMorePopular=true, regardless of the term being actually spelled > correctly. > The onlyMorePopular setting can produce collations while the term(s) are > correctly spelled. This is fine behaviour. The problem is that is also marks > the term(s) as incorrectly spelled when they're actually in the spellcheck > index. > See also this thread: > http://lucene.472066.n3.nabble.com/correctlySpelled-and-onlyMorePopular-in-3-1-td2975773.html#a2984201 -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-2556) Spellcheck component not returned with numeric queries
[ https://issues.apache.org/jira/browse/SOLR-2556?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-2556: Fix Version/s: 3.4 > Spellcheck component not returned with numeric queries > -- > > Key: SOLR-2556 > URL: https://issues.apache.org/jira/browse/SOLR-2556 > Project: Solr > Issue Type: Bug > Components: spellchecker >Affects Versions: 3.1 >Reporter: Markus Jelsma > Fix For: 3.4 > > > The spell check component's output is not written when sending queries that > consist of numbers only. Clients depending on the availability of the > spellcheck output need to check if the output is actually there. > See also: > http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201105.mbox/%3c201105301607.41956.markus.jel...@openindex.io%3E -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-2662) QueryResultCache is obligatory
[ https://issues.apache.org/jira/browse/SOLR-2662?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13069524#comment-13069524 ] Markus Jelsma commented on SOLR-2662: - The following may be related. Below the config, just an attempt to disable the query cache. {code} {code} Here's what happens when looking at the stats: {code} Concurrent LRU Cache(maxSize=2, initialSize=0, minSize=1, acceptableSize=1, cleanupThread=false) {code} Strange, it's actually in use and it even works! {code} lookups : 41 hits : 0 hitratio : 0.00 inserts : 0 evictions : 0 size : 0 warmupTime : 0 cumulative_lookups : 145 cumulative_hits : 2 cumulative_hitratio : 0.01 cumulative_inserts : 2 cumulative_evictions : 0 {code} > QueryResultCache is obligatory > -- > > Key: SOLR-2662 > URL: https://issues.apache.org/jira/browse/SOLR-2662 > Project: Solr > Issue Type: Bug > Components: search >Affects Versions: 3.3 >Reporter: Markus Jelsma >Priority: Minor > Fix For: 3.4, 4.0 > > > When the queryResultCache is not defined in the configuration, the start > parameter increments the rows parameter. Start + rows is returned and start > is always 0. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-2662) QueryResultCache is obligatory
QueryResultCache is obligatory -- Key: SOLR-2662 URL: https://issues.apache.org/jira/browse/SOLR-2662 Project: Solr Issue Type: Bug Components: search Affects Versions: 3.3 Reporter: Markus Jelsma Priority: Minor Fix For: 3.4 When the queryResultCache is not defined in the configuration, the start parameter increments the rows parameter. Start + rows is returned and start is always 0. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-2545) Allow equals-sign in key of external file field
[ https://issues.apache.org/jira/browse/SOLR-2545?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13066149#comment-13066149 ] Markus Jelsma commented on SOLR-2545: - Great work! Thanks! > Allow equals-sign in key of external file field > --- > > Key: SOLR-2545 > URL: https://issues.apache.org/jira/browse/SOLR-2545 > Project: Solr > Issue Type: Bug > Components: Schema and Analysis >Reporter: Markus Jelsma >Assignee: Hoss Man >Priority: Minor > Fix For: 3.4, 4.0 > > Attachments: SOLR-2545.patch > > > The external file field doesn't allow an equals-sign in the key. Instead of > going through the hassle of escaping, this patch just uses lastIndexOf to get > the float value. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-2545) Allow equals-sign in key of external file field
[ https://issues.apache.org/jira/browse/SOLR-2545?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-2545: Fix Version/s: 3.4 3.2.1 3.1.1 > Allow equals-sign in key of external file field > --- > > Key: SOLR-2545 > URL: https://issues.apache.org/jira/browse/SOLR-2545 > Project: Solr > Issue Type: Improvement > Components: Schema and Analysis >Affects Versions: 3.1 >Reporter: Markus Jelsma >Priority: Minor > Fix For: 3.1.1, 3.2.1, 3.4 > > Attachments: SOLR-2545.patch > > > The external file field doesn't allow an equals-sign in the key. Instead of > going through the hassle of escaping, this patch just uses lastIndexOf to get > the float value. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-2555) Always incorrectly spelled with onlyMorePopular enabled
[ https://issues.apache.org/jira/browse/SOLR-2555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13041298#comment-13041298 ] Markus Jelsma commented on SOLR-2555: - It's not the suggester we use but the plain old spellcheck component. As far is i remember (checked last week) suggester doesn't return a correctlySpelled parameter. Spellchecker also doesn't take a LookupImpl parameter (according to the wiki). It's an IndexBasedSpellchecker. > Always incorrectly spelled with onlyMorePopular enabled > --- > > Key: SOLR-2555 > URL: https://issues.apache.org/jira/browse/SOLR-2555 > Project: Solr > Issue Type: Bug > Components: spellchecker >Affects Versions: 3.1 >Reporter: Markus Jelsma > > The spellcheck component will always mark the term(s) as incorrectly spelled > when onlyMorePopular=true, regardless of the term being actually spelled > correctly. > The onlyMorePopular setting can produce collations while the term(s) are > correctly spelled. This is fine behaviour. The problem is that is also marks > the term(s) as incorrectly spelled when they're actually in the spellcheck > index. > See also this thread: > http://lucene.472066.n3.nabble.com/correctlySpelled-and-onlyMorePopular-in-3-1-td2975773.html#a2984201 -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-2556) Spellcheck component not returned with numeric queries
Spellcheck component not returned with numeric queries -- Key: SOLR-2556 URL: https://issues.apache.org/jira/browse/SOLR-2556 Project: Solr Issue Type: Bug Components: spellchecker Affects Versions: 3.1 Reporter: Markus Jelsma The spell check component's output is not written when sending queries that consist of numbers only. Clients depending on the availability of the spellcheck output need to check if the output is actually there. See also: http://mail-archives.apache.org/mod_mbox/lucene-solr-user/201105.mbox/%3c201105301607.41956.markus.jel...@openindex.io%3E -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-2555) Always incorrectly spelled with onlyMorePopular enabled
Always incorrectly spelled with onlyMorePopular enabled --- Key: SOLR-2555 URL: https://issues.apache.org/jira/browse/SOLR-2555 Project: Solr Issue Type: Bug Components: spellchecker Affects Versions: 3.1 Reporter: Markus Jelsma The spellcheck component will always mark the term(s) as incorrectly spelled when onlyMorePopular=true, regardless of the term being actually spelled correctly. The onlyMorePopular setting can produce collations while the term(s) are correctly spelled. This is fine behaviour. The problem is that is also marks the term(s) as incorrectly spelled when they're actually in the spellcheck index. See also this thread: http://lucene.472066.n3.nabble.com/correctlySpelled-and-onlyMorePopular-in-3-1-td2975773.html#a2984201 -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-2105) Rename RequestHandler param 'update.processor' to 'update.chain'.
[ https://issues.apache.org/jira/browse/SOLR-2105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13039360#comment-13039360 ] Markus Jelsma commented on SOLR-2105: - Excellent job for printing the deprecation warning, i seem to have overlooked this issue! > Rename RequestHandler param 'update.processor' to 'update.chain'. > - > > Key: SOLR-2105 > URL: https://issues.apache.org/jira/browse/SOLR-2105 > Project: Solr > Issue Type: Improvement > Components: update >Affects Versions: 1.4.1 >Reporter: Jan Høydahl >Assignee: Mark Miller >Priority: Minor > Fix For: 3.2, 4.0 > > Attachments: SOLR-2105.patch, SOLR-2105.patch, SOLR-2105.patch > > > Today we reference a custom updateRequestProcessorChain using the update > request parameter "update.processor". > See > http://wiki.apache.org/solr/SolrConfigXml#UpdateRequestProcessorChain_section > This is confusing, since what we are really referencing is not an > UpdateProcessor, but an updateRequestProcessorChain. > I propose that "update.processor" is renamed as "update.chain" or similar -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-2545) Allow equals-sign in key of external file field
[ https://issues.apache.org/jira/browse/SOLR-2545?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-2545: Attachment: SOLR-2545.patch > Allow equals-sign in key of external file field > --- > > Key: SOLR-2545 > URL: https://issues.apache.org/jira/browse/SOLR-2545 > Project: Solr > Issue Type: Improvement > Components: Schema and Analysis >Affects Versions: 3.1 >Reporter: Markus Jelsma >Priority: Minor > Attachments: SOLR-2545.patch > > > The external file field doesn't allow an equals-sign in the key. Instead of > going through the hassle of escaping, this patch just uses lastIndexOf to get > the float value. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-2545) Allow equals-sign in key of external file field
Allow equals-sign in key of external file field --- Key: SOLR-2545 URL: https://issues.apache.org/jira/browse/SOLR-2545 Project: Solr Issue Type: Improvement Components: Schema and Analysis Affects Versions: 3.1 Reporter: Markus Jelsma Priority: Minor Attachments: SOLR-2545.patch The external file field doesn't allow an equals-sign in the key. Instead of going through the hassle of escaping, this patch just uses lastIndexOf to get the float value. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-2442) Different cores use the same admin-extra.html
Different cores use the same admin-extra.html - Key: SOLR-2442 URL: https://issues.apache.org/jira/browse/SOLR-2442 Project: Solr Issue Type: Bug Components: web gui Affects Versions: 1.4.1 Reporter: Markus Jelsma Priority: Minor Solr loads a single (or overwrites) admin-extra.html for all cores. The core specified last in solr.xml is used to load the admin-extra.html file. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Updated: (SOLR-2327) java.lang.ArithmeticException: / by zero with queryResultCache size=0 and queryResultWindowSize=0
[ https://issues.apache.org/jira/browse/SOLR-2327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-2327: Fix Version/s: 3.1 Priority: Minor (was: Major) Affects Version/s: 3.1 > java.lang.ArithmeticException: / by zero with queryResultCache size=0 and > queryResultWindowSize=0 > - > > Key: SOLR-2327 > URL: https://issues.apache.org/jira/browse/SOLR-2327 > Project: Solr > Issue Type: Bug > Components: search >Affects Versions: 1.4.1, 3.1 >Reporter: Markus Jelsma >Priority: Minor > Fix For: 3.1 > > > With the following configuration: > autowarmCount="0"/> > 0 > The following exception will occur: > 2011-01-21 13:48:13,599 ERROR [solr.core.SolrCore] - [http-8080-1] - : > java.lang.ArithmeticException: / by zero > at > org.apache.solr.search.SolrIndexSearcher.getDocListC(SolrIndexSearcher.java:833) > at > org.apache.solr.search.SolrIndexSearcher.search(SolrIndexSearcher.java:341) > at > org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:182) > at > org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:195) > at > org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) > at org.apache.solr.core.SolrCore.execute(SolrCore.java:1317) > at > org.apache.solr.handler.PingRequestHandler.handleRequestBody(PingRequestHandler.java:63) > at > org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) > at org.apache.solr.core.SolrCore.execute(SolrCore.java:1317) > at > org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) > at > org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235) > at > org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) > at > org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233) > at > org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191) > at > org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127) > at > org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) > at > org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) > at > org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298) > at > org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:857) > at > org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:588) > at > org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) > at java.lang.Thread.run(Thread.java:619) -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Created: (SOLR-2327) java.lang.ArithmeticException: / by zero with queryResultCache size=0 and queryResultWindowSize=0
java.lang.ArithmeticException: / by zero with queryResultCache size=0 and queryResultWindowSize=0 - Key: SOLR-2327 URL: https://issues.apache.org/jira/browse/SOLR-2327 Project: Solr Issue Type: Bug Components: search Affects Versions: 1.4.1 Reporter: Markus Jelsma With the following configuration: 0 The following exception will occur: 2011-01-21 13:48:13,599 ERROR [solr.core.SolrCore] - [http-8080-1] - : java.lang.ArithmeticException: / by zero at org.apache.solr.search.SolrIndexSearcher.getDocListC(SolrIndexSearcher.java:833) at org.apache.solr.search.SolrIndexSearcher.search(SolrIndexSearcher.java:341) at org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:182) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:195) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1317) at org.apache.solr.handler.PingRequestHandler.handleRequestBody(PingRequestHandler.java:63) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1317) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:857) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:588) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) at java.lang.Thread.run(Thread.java:619) -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Updated: (SOLR-2323) Solr should clean old replication temp dirs
[ https://issues.apache.org/jira/browse/SOLR-2323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-2323: Description: In a high commit rate environment (polling < 10s and commits every minute) the shutdown/restart of a slave can result in old temp directories laying around, filling up disk space as we go on. This happens with the following scenario: 1. master has index version 2 2. slave downloads files for version 2 to index.2 temp directory 3. slave is shutdown 4. master increments to version 3 5. slave is started 6. slave downloads files for version 3 to index.3 temp directory The result is index.2 temp directory not getting deleted by any process. This is very annoying in such an environment where nodes are restarted frequently (for whatever reason). Working around the problem can be done by either manually deleting the temp directories between shutdown and startup or by calling the disablepoll command followed by an abortfetch command which will (after a long wait) finally purge the temp directory. See this thread: http://www.mail-archive.com/solr-user@lucene.apache.org/msg45120.html was: In a high commit rate environment (polling < 10s and commits every minute) the shutdown/restart of a slave can result in old temp directories laying around, filling up disk space as we go on. This happens with the following scenario: 1. master has index version 2 2. slave downloads files for version 2 to index.2 temp directory 3. slave is shutdown 4. master increments to version 3 5. slave is started 6. slave downloads files for version 3 to index.2 temp directory The result is index.2 temp directory not getting deleted by any process. This is very annoying in such an environment where nodes are restarted frequently (for whatever reason). Working around the problem can be done by either manually deleting the temp directories between shutdown and startup or by calling the disablepoll command followed by an abortfetch command which will (after a long wait) finally purge the temp directory. See this thread: http://www.mail-archive.com/solr-user@lucene.apache.org/msg45120.html > Solr should clean old replication temp dirs > --- > > Key: SOLR-2323 > URL: https://issues.apache.org/jira/browse/SOLR-2323 > Project: Solr > Issue Type: Bug > Components: replication (java) >Affects Versions: 1.4.1 >Reporter: Markus Jelsma > > In a high commit rate environment (polling < 10s and commits every minute) > the shutdown/restart of a slave can result in old temp directories laying > around, filling up disk space as we go on. This happens with the following > scenario: > 1. master has index version 2 > 2. slave downloads files for version 2 to index.2 temp directory > 3. slave is shutdown > 4. master increments to version 3 > 5. slave is started > 6. slave downloads files for version 3 to index.3 temp directory > The result is index.2 temp directory not getting deleted by any process. This > is very annoying in such an environment where nodes are restarted frequently > (for whatever reason). Working around the problem can be done by either > manually deleting the temp directories between shutdown and startup or by > calling the disablepoll command followed by an abortfetch command which will > (after a long wait) finally purge the temp directory. > See this thread: > http://www.mail-archive.com/solr-user@lucene.apache.org/msg45120.html -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Updated: (SOLR-2323) Solr should clean old replication temp dirs
[ https://issues.apache.org/jira/browse/SOLR-2323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Markus Jelsma updated SOLR-2323: Description: In a high commit rate environment (polling < 10s and commits every minute) the shutdown/restart of a slave can result in old temp directories laying around, filling up disk space as we go on. This happens with the following scenario: 1. master has index version 2 2. slave downloads files for version 2 to index.2 temp directory 3. slave is shutdown 4. master increments to version 3 5. slave is started 6. slave downloads files for version 3 to index.2 temp directory The result is index.2 temp directory not getting deleted by any process. This is very annoying in such an environment where nodes are restarted frequently (for whatever reason). Working around the problem can be done by either manually deleting the temp directories between shutdown and startup or by calling the disablepoll command followed by an abortfetch command which will (after a long wait) finally purge the temp directory. See this thread: http://www.mail-archive.com/solr-user@lucene.apache.org/msg45120.html was: In a high commit rate environment (polling < 10s and commits every minute) the shutdown/restart of a slave can result in old temp directories laying around, filling up disk space as we go on. This happens with the following scenario: 1. master has index version 2 2. slave downloads files for version 2 to index.2 temp directory 3. slave is shutdown 4. master increments to version 3 5. slave downloads files for version 3 to index.2 temp directory The result is index.2 temp directory not getting deleted by any process. This is very annoying in such an environment where nodes are restarted frequently (for whatever reason). Working around the problem can be done by either manually deleting the temp directories between shutdown and startup or by calling the disablepoll command followed by an abortfetch command which will (after a long wait) finally purge the temp directory. See this thread: http://www.mail-archive.com/solr-user@lucene.apache.org/msg45120.html > Solr should clean old replication temp dirs > --- > > Key: SOLR-2323 > URL: https://issues.apache.org/jira/browse/SOLR-2323 > Project: Solr > Issue Type: Bug > Components: replication (java) >Affects Versions: 1.4.1 >Reporter: Markus Jelsma > > In a high commit rate environment (polling < 10s and commits every minute) > the shutdown/restart of a slave can result in old temp directories laying > around, filling up disk space as we go on. This happens with the following > scenario: > 1. master has index version 2 > 2. slave downloads files for version 2 to index.2 temp directory > 3. slave is shutdown > 4. master increments to version 3 > 5. slave is started > 6. slave downloads files for version 3 to index.2 temp directory > The result is index.2 temp directory not getting deleted by any process. This > is very annoying in such an environment where nodes are restarted frequently > (for whatever reason). Working around the problem can be done by either > manually deleting the temp directories between shutdown and startup or by > calling the disablepoll command followed by an abortfetch command which will > (after a long wait) finally purge the temp directory. > See this thread: > http://www.mail-archive.com/solr-user@lucene.apache.org/msg45120.html -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Created: (SOLR-2323) Solr should clean old replication temp dirs
Solr should clean old replication temp dirs --- Key: SOLR-2323 URL: https://issues.apache.org/jira/browse/SOLR-2323 Project: Solr Issue Type: Bug Components: replication (java) Affects Versions: 1.4.1 Reporter: Markus Jelsma In a high commit rate environment (polling < 10s and commits every minute) the shutdown/restart of a slave can result in old temp directories laying around, filling up disk space as we go on. This happens with the following scenario: 1. master has index version 2 2. slave downloads files for version 2 to index.2 temp directory 3. slave is shutdown 4. master increments to version 3 5. slave downloads files for version 3 to index.2 temp directory The result is index.2 temp directory not getting deleted by any process. This is very annoying in such an environment where nodes are restarted frequently (for whatever reason). Working around the problem can be done by either manually deleting the temp directories between shutdown and startup or by calling the disablepoll command followed by an abortfetch command which will (after a long wait) finally purge the temp directory. See this thread: http://www.mail-archive.com/solr-user@lucene.apache.org/msg45120.html -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Commented: (SOLR-2277) Update with add and delete combined fails
[ https://issues.apache.org/jira/browse/SOLR-2277?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12969786#action_12969786 ] Markus Jelsma commented on SOLR-2277: - Ah, an undocumented feature. I've added this to the wiki. I assume this ticket can be closed? http://wiki.apache.org/solr/UpdateXmlMessages#Add_and_delete_in_a_single_batch > Update with add and delete combined fails > - > > Key: SOLR-2277 > URL: https://issues.apache.org/jira/browse/SOLR-2277 > Project: Solr > Issue Type: Bug > Components: update >Affects Versions: 1.4.1 >Reporter: Markus Jelsma > > The following curl command: > curl http://127.0.0.1:8983/solr/update/?commit=true -H "Content-Type: > text/xml" --data-binary ' name="id">171234'; > will trigger the following exception in Solr 1.4.1: > Dec 9, 2010 12:51:22 PM org.apache.solr.common.SolrException log > SEVERE: org.apache.solr.common.SolrException: Illegal to have multiple roots > (start tag in epilog?). > at [row,col {unknown-source}]: [47,2] > at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:72) > at > org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:54) > at > org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) > at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) > at > org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) > at > org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1089) > at > org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:365) > at > org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216) > at > org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181) > at > org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:712) > at > org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:405) > at > org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:211) > at > org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114) > at > org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:139) > at org.mortbay.jetty.Server.handle(Server.java:285) > at > org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:502) > at > org.mortbay.jetty.HttpConnection$RequestHandler.content(HttpConnection.java:835) > at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:641) > at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:202) > at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:378) > at > org.mortbay.jetty.bio.SocketConnector$Connection.run(SocketConnector.java:226) > at > org.mortbay.thread.BoundedThreadPool$PoolThread.run(BoundedThreadPool.java:442) > Caused by: com.ctc.wstx.exc.WstxParsingException: Illegal to have multiple > roots (start tag in epilog?). > at [row,col {unknown-source}]: [47,2] > at > com.ctc.wstx.sr.StreamScanner.constructWfcException(StreamScanner.java:630) > at > com.ctc.wstx.sr.StreamScanner.throwParseError(StreamScanner.java:461) > at > com.ctc.wstx.sr.BasicStreamReader.handleExtraRoot(BasicStreamReader.java:2155) > at > com.ctc.wstx.sr.BasicStreamReader.nextFromProlog(BasicStreamReader.java:2070) > at com.ctc.wstx.sr.BasicStreamReader.next(BasicStreamReader.java:1071) > at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:90) > at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69) > ... 22 more -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Created: (SOLR-2278) PHPSerialized fails with Solr spatial
PHPSerialized fails with Solr spatial - Key: SOLR-2278 URL: https://issues.apache.org/jira/browse/SOLR-2278 Project: Solr Issue Type: Bug Components: Response Writers Affects Versions: 1.4.1 Reporter: Markus Jelsma Solr throws a java.lang.IllegalArgumentException: Map size must not be negative exception when using the PHP Serialized response writer with JTeam SolrSpatial plugin in front. At first it may seem a bug in the plugin but according to some posts in the mailing list thread ( http://lucene.472066.n3.nabble.com/Map-size-must-not-be-negative-with-spatial-results-php-serialized-td2039782.html ) it just might be a bug in Solr. The only way to reproduce the issue that i know of is using using LocalParams to set spatial parameters and having the spatial search component activated as last-components. If the query yields no results, the exception won't show up. distance 1 1 60 ad_latitude ad_longitude _tier_ In the request handler: geodistance query: http://localhost:8983/solr/search?q={!spatial%20lat=51.9562%20long=6.02606%20radius=432%20unit=km}auto&wt=php -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Created: (SOLR-2277) Update with add and delete combined fails
Update with add and delete combined fails - Key: SOLR-2277 URL: https://issues.apache.org/jira/browse/SOLR-2277 Project: Solr Issue Type: Bug Components: update Affects Versions: 1.4.1 Reporter: Markus Jelsma The following curl command: curl http://127.0.0.1:8983/solr/update/?commit=true -H "Content-Type: text/xml" --data-binary '171234'; will trigger the following exception in Solr 1.4.1: Dec 9, 2010 12:51:22 PM org.apache.solr.common.SolrException log SEVERE: org.apache.solr.common.SolrException: Illegal to have multiple roots (start tag in epilog?). at [row,col {unknown-source}]: [47,2] at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:72) at org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:54) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) at org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1089) at org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:365) at org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216) at org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181) at org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:712) at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:405) at org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:211) at org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114) at org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:139) at org.mortbay.jetty.Server.handle(Server.java:285) at org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:502) at org.mortbay.jetty.HttpConnection$RequestHandler.content(HttpConnection.java:835) at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:641) at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:202) at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:378) at org.mortbay.jetty.bio.SocketConnector$Connection.run(SocketConnector.java:226) at org.mortbay.thread.BoundedThreadPool$PoolThread.run(BoundedThreadPool.java:442) Caused by: com.ctc.wstx.exc.WstxParsingException: Illegal to have multiple roots (start tag in epilog?). at [row,col {unknown-source}]: [47,2] at com.ctc.wstx.sr.StreamScanner.constructWfcException(StreamScanner.java:630) at com.ctc.wstx.sr.StreamScanner.throwParseError(StreamScanner.java:461) at com.ctc.wstx.sr.BasicStreamReader.handleExtraRoot(BasicStreamReader.java:2155) at com.ctc.wstx.sr.BasicStreamReader.nextFromProlog(BasicStreamReader.java:2070) at com.ctc.wstx.sr.BasicStreamReader.next(BasicStreamReader.java:1071) at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:90) at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69) ... 22 more -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] Commented: (SOLR-1752) SolrJ fails with exception when passing document ADD and DELETEs in the same request using XML request writer (but not binary request writer)
[ https://issues.apache.org/jira/browse/SOLR-1752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12969704#action_12969704 ] Markus Jelsma commented on SOLR-1752: - This isn't limited to SolrJ. The following curl command will trigger the same error in 1.4.1 curl http://127.0.0.1:8983/solr/update/?commit=true -H "Content-Type: text/xml" --data-binary '171234'; > SolrJ fails with exception when passing document ADD and DELETEs in the same > request using XML request writer (but not binary request writer) > - > > Key: SOLR-1752 > URL: https://issues.apache.org/jira/browse/SOLR-1752 > Project: Solr > Issue Type: Bug > Components: clients - java, update >Affects Versions: 1.4 >Reporter: Jayson Minard >Assignee: Shalin Shekhar Mangar >Priority: Blocker > > Add this test to SolrExampleTests.java and it will fail when using the XML > Request Writer (now default), but not if you change the SolrExampleJettyTest > to use the BinaryRequestWriter. > {code} > public void testAddDeleteInSameRequest() throws Exception { > SolrServer server = getSolrServer(); > SolrInputDocument doc3 = new SolrInputDocument(); > doc3.addField( "id", "id3", 1.0f ); > doc3.addField( "name", "doc3", 1.0f ); > doc3.addField( "price", 10 ); > UpdateRequest up = new UpdateRequest(); > up.add( doc3 ); > up.deleteById("id001"); > up.setWaitFlush(false); > up.setWaitSearcher(false); > up.process( server ); > } > {code} > terminates with exception: > {code} > Feb 3, 2010 8:55:34 AM org.apache.solr.common.SolrException log > SEVERE: org.apache.solr.common.SolrException: Illegal to have multiple roots > (start tag in epilog?). > at [row,col {unknown-source}]: [1,125] > at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:72) > at > org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:54) > at > org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) > at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) > at > org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) > at > org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1089) > at > org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:365) > at > org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181) > at > org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:712) > at > org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:139) > at org.mortbay.jetty.Server.handle(Server.java:285) > at > org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:502) > at > org.mortbay.jetty.HttpConnection$RequestHandler.content(HttpConnection.java:835) > at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:723) > at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:202) > at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:378) > at > org.mortbay.jetty.bio.SocketConnector$Connection.run(SocketConnector.java:226) > at > org.mortbay.thread.BoundedThreadPool$PoolThread.run(BoundedThreadPool.java:442) > Caused by: com.ctc.wstx.exc.WstxParsingException: Illegal to have multiple > roots (start tag in epilog?). > at [row,col {unknown-source}]: [1,125] > at > com.ctc.wstx.sr.StreamScanner.constructWfcException(StreamScanner.java:630) > at com.ctc.wstx.sr.StreamScanner.throwParseError(StreamScanner.java:461) > at > com.ctc.wstx.sr.BasicStreamReader.handleExtraRoot(BasicStreamReader.java:2155) > at > com.ctc.wstx.sr.BasicStreamReader.nextFromProlog(BasicStreamReader.java:2070) > at > com.ctc.wstx.sr.BasicStreamReader.nextFromTree(BasicStreamReader.java:2647) > at com.ctc.wstx.sr.BasicStreamReader.next(BasicStreamReader.java:1019) > at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:90) > at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69) > ... 18 more > {code} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org