[jira] [Updated] (LUCENE-7700) Move throughput control and merge aborting out of IndexWriter's core?

2017-03-09 Thread Dawid Weiss (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-7700?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Weiss updated LUCENE-7700:

Description: 
Here is a bit of a background:
- I wanted to implement a custom merging strategy that would have a custom i/o 
flow control (global),
- currently, the CMS is tightly coupled with a few classes -- MergeRateLimiter, 
OneMerge, IndexWriter.

Looking at the code it seems to me that everything with respect to I/O control 
could be nicely pulled out into classes that explicitly control the merging 
process, that is only MergePolicy and MergeScheduler. By default, one could 
even run without any additional I/O accounting overhead (which is currently in 
there, even if one doesn't use the CMS's throughput control).

Such refactoring would also give a chance to nicely move things where they 
belong -- job aborting into OneMerge (currently in RateLimiter), rate limiter 
lifecycle bound to OneMerge (MergeScheduler could then use per-merge or global 
accounting, as it pleases).

Just a thought and some initial refactorings for discussion.

  was:
Here is a bit of a background:
- I wanted to implement a custom merging strategy that would have a custom i/o 
flow control (global),
- currently, the CMS is tightly bound with a few classes -- MergeRateLimiter, 
OneMerge, IndexWriter.

Looking at the code it seems to me that everything with respect to I/O control 
could be nicely pulled out into classes that explicitly control the merging 
process, that is only MergePolicy and MergeScheduler. By default, one could 
even run without any additional I/O accounting overhead (which is currently in 
there, even if one doesn't use the CMS's throughput control).

Such refactoring would also give a chance to nicely move things where they 
belong -- job aborting into OneMerge (currently in RateLimiter), rate limiter 
lifecycle bound to OneMerge (MergeScheduler could then use per-merge or global 
accounting, as it pleases).

Just a thought and some initial refactorings for discussion.


> Move throughput control and merge aborting out of IndexWriter's core?
> -
>
> Key: LUCENE-7700
> URL: https://issues.apache.org/jira/browse/LUCENE-7700
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 6.x, master (7.0), 6.5
>
> Attachments: LUCENE-7700.patch, LUCENE-7700.patch, LUCENE-7700.patch
>
>
> Here is a bit of a background:
> - I wanted to implement a custom merging strategy that would have a custom 
> i/o flow control (global),
> - currently, the CMS is tightly coupled with a few classes -- 
> MergeRateLimiter, OneMerge, IndexWriter.
> Looking at the code it seems to me that everything with respect to I/O 
> control could be nicely pulled out into classes that explicitly control the 
> merging process, that is only MergePolicy and MergeScheduler. By default, one 
> could even run without any additional I/O accounting overhead (which is 
> currently in there, even if one doesn't use the CMS's throughput control).
> Such refactoring would also give a chance to nicely move things where they 
> belong -- job aborting into OneMerge (currently in RateLimiter), rate limiter 
> lifecycle bound to OneMerge (MergeScheduler could then use per-merge or 
> global accounting, as it pleases).
> Just a thought and some initial refactorings for discussion.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-7700) Move throughput control and merge aborting out of IndexWriter's core?

2017-03-09 Thread Dawid Weiss (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-7700?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Weiss updated LUCENE-7700:

Attachment: LUCENE-7700.patch

Final patch.

> Move throughput control and merge aborting out of IndexWriter's core?
> -
>
> Key: LUCENE-7700
> URL: https://issues.apache.org/jira/browse/LUCENE-7700
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 6.x, master (7.0), 6.5
>
> Attachments: LUCENE-7700.patch, LUCENE-7700.patch, LUCENE-7700.patch, 
> LUCENE-7700.patch
>
>
> Here is a bit of a background:
> - I wanted to implement a custom merging strategy that would have a custom 
> i/o flow control (global),
> - currently, the CMS is tightly coupled with a few classes -- 
> MergeRateLimiter, OneMerge, IndexWriter.
> Looking at the code it seems to me that everything with respect to I/O 
> control could be nicely pulled out into classes that explicitly control the 
> merging process, that is only MergePolicy and MergeScheduler. By default, one 
> could even run without any additional I/O accounting overhead (which is 
> currently in there, even if one doesn't use the CMS's throughput control).
> Such refactoring would also give a chance to nicely move things where they 
> belong -- job aborting into OneMerge (currently in RateLimiter), rate limiter 
> lifecycle bound to OneMerge (MergeScheduler could then use per-merge or 
> global accounting, as it pleases).
> Just a thought and some initial refactorings for discussion.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10218) The Schema API "replace-field-type" does not generate the SolrParameters for a SimilarityFactory correctly

2017-03-09 Thread Troy Mohl (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10218?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Troy Mohl updated SOLR-10218:
-
Attachment: SOLR-10218.patch

This patch updates the FieldTypeXmlAdapter to parse the similarity options into 
XML elements.

> The Schema API "replace-field-type" does not generate the SolrParameters for 
> a SimilarityFactory correctly
> --
>
> Key: SOLR-10218
> URL: https://issues.apache.org/jira/browse/SOLR-10218
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Schema and Analysis
>Affects Versions: 6.4.1
>Reporter: Benjamin Deininger
>Priority: Minor
> Attachments: SOLR-10218.patch
>
>
> When sending a JSON POST to the Schema API to replace a field type, the 
> following JSON does not pass the SolrParameters properly to the 
> BM25SimilarityFactory.  
> {code:javascript}
> {"replace-field-type":{"name":"tint","class":"solr.TrieIntField","positionIncrementGap":"0","precisionStep":"8","similarity":{"class":"solr.BM25SimilarityFactory","k1":1.25,"b":0.75}}}
> {code}
> The `appendAttrs` function in the FieldTypeXmlAdapter parses k1 and b into 
> attributes instead of children.  
> https://github.com/apache/lucene-solr/blob/master/solr/core/src/java/org/apache/solr/rest/schema/FieldTypeXmlAdapter.java#L155
> {code:xml}
>  class="org.apache.lucene.search.similarities.BM25Similarity" k1="1.25"/>
> {code}
> Based on the XML examples for similarity, this should actually be the 
> following :
> {code:xml}
> 
>  0.1
>  0.1
> 
> {code}
> The similarities block in JSON should be handled differently so that the XML 
> is generated appropriately.
> {code:java}
> protected static Element appendSimilarityAttrs(Document doc, Element elm, 
> Map json) {
> String clazz = (String) json.get("class");
> elm.setAttribute("class", clazz);
> json.remove("class");
> for (Map.Entry entry : json.entrySet()) {
> Object val = entry.getValue();
> if (val != null && !(val instanceof Map)) {
> Element element = 
> doc.createElement(val.getClass().getSimpleName().toLowerCase());
> element.setAttribute("name", entry.getKey());
> element.setTextContent(entry.getValue().toString());
> elm.appendChild(element);
> }
> }
> return elm;
> }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-9838) atomic "inc" when adding doc doesn't respect field "default" from schema

2017-03-09 Thread Ishan Chattopadhyaya (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-9838?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ishan Chattopadhyaya reassigned SOLR-9838:
--

Assignee: Ishan Chattopadhyaya

> atomic "inc" when adding doc doesn't respect field "default" from schema
> 
>
> Key: SOLR-9838
> URL: https://issues.apache.org/jira/browse/SOLR-9838
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Hoss Man
>Assignee: Ishan Chattopadhyaya
> Attachments: SOLR-9838.patch
>
>
> If you do an "atomic update" when adding a document for the first time, then 
> the "inc" operator acts as if the field has a default of 0.
> But if the {{}} has an *actual* default in the schema.xml (example: 
> {{default="42"}}) then that default is ignored by the atomic update code path.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-1105) Use a different stored field for highlighting

2017-03-09 Thread Matthew Caruana Galizia (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903258#comment-15903258
 ] 

Matthew Caruana Galizia commented on SOLR-1105:
---

It also doesn't seem to difficult to implement on the UnifiedHighlighter, at 
least for someone who's familiar with the code.

> Use a different stored field for highlighting
> -
>
> Key: SOLR-1105
> URL: https://issues.apache.org/jira/browse/SOLR-1105
> Project: Solr
>  Issue Type: Improvement
>  Components: highlighter
>Reporter: Dmitry Lihachev
> Attachments: SOLR-1105-1_4_1.patch, 
> SOLR-1105_shared_content_field_1.3.0.patch
>
>
> DefaultSolrHighlighter uses stored field content to highlight. It has some 
> disadvantages, because index grows up fast when using multilingual indexing 
> due to several fields has to be stored with same content. This patch allows 
> DefaultSolrHighlighter to use "contentField" attribute to loockup content in 
> external field.
> Excerpt from old schema:
> {code:xml}
> 
> 
> 
> 
> {code}
> The same after patching, highlighter will now get content stored in "title" 
> field
> {code:xml}
> 
>  contentField="title"/>
>  contentField="title"/>
>  contentField="title"/>
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-10117) Big docs and the DocumentCache; umbrella issue

2017-03-09 Thread Alexandre Rafalovitch (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903226#comment-15903226
 ] 

Alexandre Rafalovitch commented on SOLR-10117:
--

No help with code I am afraid, but a question:

Would this deduplicate large fields replicated in multiple records? We advise 
people to denormalize records, so if big stored fields were actually stored 
once (e.g. repeated description pushed into child records), it would probably 
be a marketing argument, not just a technical one.

> Big docs and the DocumentCache; umbrella issue
> --
>
> Key: SOLR-10117
> URL: https://issues.apache.org/jira/browse/SOLR-10117
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: David Smiley
>Assignee: David Smiley
> Attachments: SOLR_10117_large_fields.patch
>
>
> This is an umbrella issue for improved handling of large documents (large 
> stored fields), generally related to the DocumentCache or SolrIndexSearcher's 
> doc() methods.  Highlighting is affected as it's the primary consumer of this 
> data.  "Large" here is multi-megabyte, especially tens even hundreds of 
> megabytes. We'd like to support such users without forcing them to choose 
> between no DocumentCache (bad performance), or having one but hitting OOM due 
> to massive Strings winding up in there.  I've contemplated this for longer 
> than I'd like to admit and it's a complicated issue with difference concerns 
> to balance.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Closed] (SOLR-5276) highlighter working using stemmed tokens from another field and text from another

2017-03-09 Thread David Smiley (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5276?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Smiley closed SOLR-5276.
--
Resolution: Duplicate

> highlighter working using stemmed tokens from another field and text from 
> another
> -
>
> Key: SOLR-5276
> URL: https://issues.apache.org/jira/browse/SOLR-5276
> Project: Solr
>  Issue Type: Improvement
>  Components: highlighter
>Reporter: Maciej Lizewski
>Priority: Minor
>
> The case is like this:
> I have 'content' field with content extracted with tika and several fields 
> for language versions (like content_pl, content_en, content_es, content_ru, 
> etc). 
> I also use custom langid component which can copy 'content' to serveral 
> content_* fields if it detects more than one language (so those parts are 
> properly stemmed in every language present in text).
> Now to use highlighter in such scenario I need to store all those language 
> fields even if their contents is always same as the one in 'content' field.
> Would be nice if I could configure language specific fields to be not stored, 
> and configure highlighter to take tokens positions from those fields and 
> apply them to text in 'content' field...
> In other words - to say: take tokens from 'content_pl', and prepare highlight 
> based on text in 'content' field.
> It could be administrator responsibility to guarantee that mapped fields have 
> same content.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-9959) SolrInfoMBean-s category and hierarchy cleanup

2017-03-09 Thread Andrzej Bialecki (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-9959?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrzej Bialecki  updated SOLR-9959:

Component/s: metrics

> SolrInfoMBean-s category and hierarchy cleanup
> --
>
> Key: SOLR-9959
> URL: https://issues.apache.org/jira/browse/SOLR-9959
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Affects Versions: master (7.0)
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
>Priority: Blocker
> Fix For: master (7.0)
>
>
> SOLR-9947 changed categories of some of {{SolrInfoMBean-s}}, and it also 
> added an alternative view in JMX, similar to the one produced by 
> {{SolrJmxReporter}}.
> Some changes were left out from that issue because they would break the 
> back-compatibility in 6.x, but they should be done before 7.0:
> * remove the old JMX view of {{SolrInfoMBean}}-s and improve the new one so 
> that it's more readable and useful.
> * in many cases {{SolrInfoMBean.getName()}} just returns a FQCN, but it could 
> be more informative - eg. for highlighter or query plugins this could be the 
> symbolic name of a plugin that users know and use in configuration.
> * top-level categories need more thought. On one hand it's best to minimize 
> their number, on the other hand they need to meaningfully represent the 
> functionality of components that use them. SOLR-9947 made some cosmetic 
> changes, but more discussion is necessary (eg. QUERY vs. SEARCHHANDLER)
> * we should consider removing some of the methods in {{SolrInfoMBean}} that 
> usually don't return any useful information, eg. {{getDocs}}, {{getSource()}} 
> and {{getVersion()}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10247) Support non-numeric metrics

2017-03-09 Thread Andrzej Bialecki (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrzej Bialecki  updated SOLR-10247:
-
Attachment: SOLR-10247.patch

Updated patch.

> Support non-numeric metrics
> ---
>
> Key: SOLR-10247
> URL: https://issues.apache.org/jira/browse/SOLR-10247
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
> Fix For: 6.5, master (7.0)
>
> Attachments: compactFormat.png, currentFormat.png, SOLR-10247.patch, 
> SOLR-10247.patch
>
>
> Metrics API currently supports only numeric values. However, it's useful also 
> to report non-numeric values such as eg. version, disk type, component state, 
> some system properties, etc.
> Codahale {{Gauge}} metric type can be used for this purpose, and 
> convenience methods can be added to {{SolrMetricManager}} to make it easier 
> to use.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-10117) Big docs and the DocumentCache; umbrella issue

2017-03-09 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903287#comment-15903287
 ] 

David Smiley commented on SOLR-10117:
-

bq. Would this deduplicate large fields replicated in multiple records?

No.  If I were tasked to do that, I might implement a customized 
DocValuesFormat that deduplicated per segment (could not dedup at higher tiers) 
by using the length as a crude hash and then verifying the dedup by re-reading 
the original.  Query time would be no overhead; it'd simply share the internal 
offset/length pointer.

> Big docs and the DocumentCache; umbrella issue
> --
>
> Key: SOLR-10117
> URL: https://issues.apache.org/jira/browse/SOLR-10117
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: David Smiley
>Assignee: David Smiley
> Attachments: SOLR_10117_large_fields.patch
>
>
> This is an umbrella issue for improved handling of large documents (large 
> stored fields), generally related to the DocumentCache or SolrIndexSearcher's 
> doc() methods.  Highlighting is affected as it's the primary consumer of this 
> data.  "Large" here is multi-megabyte, especially tens even hundreds of 
> megabytes. We'd like to support such users without forcing them to choose 
> between no DocumentCache (bad performance), or having one but hitting OOM due 
> to massive Strings winding up in there.  I've contemplated this for longer 
> than I'd like to admit and it's a complicated issue with difference concerns 
> to balance.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5872) Eliminate overseer queue

2017-03-09 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5872?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903303#comment-15903303
 ] 

Erick Erickson commented on SOLR-5872:
--

bq: "Also, I do not see why collection A should be aware of collection B state."

What is your evidence of this? Because this was changed quite a while ago. 
Originally there was a single clusterstate.json that held all of the state 
information for all the collections, but that changed to having a state.json in 
each collections' Znode. So either you're misinterpreting something or somehow 
using an older style Zookeeper state. Or something really strange is happening.

See the collections API MIGRATESTATEFORMAT.



> Eliminate overseer queue 
> -
>
> Key: SOLR-5872
> URL: https://issues.apache.org/jira/browse/SOLR-5872
> Project: Solr
>  Issue Type: Improvement
>  Components: SolrCloud
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The overseer queue is one of the busiest points in the entire system. The 
> raison d'être of the queue is
>  * Provide batching of operations for the main clusterstate,json so that 
> state updates are minimized 
> * Avoid race conditions and ensure order
> Now , as we move the individual collection states out of the main 
> clusterstate.json, the batching is not useful anymore.
> Race conditions can easily be solved by using a compare and set in Zookeeper. 
> The proposed solution  is , whenever an operation is required to be performed 
> on the clusterstate, the same thread (and of course the same JVM)
>  # read the fresh state and version of zk node  
>  # construct the new state 
>  # perform a compare and set
>  # if compare and set fails go to step 1
> This should be limited to all operations performed on external collections 
> because batching would be required for others 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-9959) SolrInfoMBean-s category and hierarchy cleanup

2017-03-09 Thread Andrzej Bialecki (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-9959?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903385#comment-15903385
 ] 

Andrzej Bialecki  commented on SOLR-9959:
-

There's also considerable duplication of code responsible for reporting metrics 
in components that implement both {{SolrInfoMBean}} and {{SolrMetricProducer}}. 
We should investigate whether {{SolrInfoMBean.getStatistics()}} is needed at 
all - statistics reported by this method theoretically can be equally well 
reported using the metrics API.

> SolrInfoMBean-s category and hierarchy cleanup
> --
>
> Key: SOLR-9959
> URL: https://issues.apache.org/jira/browse/SOLR-9959
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Affects Versions: master (7.0)
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
>Priority: Blocker
> Fix For: master (7.0)
>
>
> SOLR-9947 changed categories of some of {{SolrInfoMBean-s}}, and it also 
> added an alternative view in JMX, similar to the one produced by 
> {{SolrJmxReporter}}.
> Some changes were left out from that issue because they would break the 
> back-compatibility in 6.x, but they should be done before 7.0:
> * remove the old JMX view of {{SolrInfoMBean}}-s and improve the new one so 
> that it's more readable and useful.
> * in many cases {{SolrInfoMBean.getName()}} just returns a FQCN, but it could 
> be more informative - eg. for highlighter or query plugins this could be the 
> symbolic name of a plugin that users know and use in configuration.
> * top-level categories need more thought. On one hand it's best to minimize 
> their number, on the other hand they need to meaningfully represent the 
> functionality of components that use them. SOLR-9947 made some cosmetic 
> changes, but more discussion is necessary (eg. QUERY vs. SEARCHHANDLER)
> * we should consider removing some of the methods in {{SolrInfoMBean}} that 
> usually don't return any useful information, eg. {{getDocs}}, {{getSource()}} 
> and {{getVersion()}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-9959) SolrInfoMBean-s category and hierarchy cleanup

2017-03-09 Thread Andrzej Bialecki (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-9959?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903382#comment-15903382
 ] 

Andrzej Bialecki  commented on SOLR-9959:
-

There's also considerable duplication of code responsible for reporting metrics 
in components that implement both {{SolrInfoMBean}} and {{SolrMetricProducer}}. 
We should investigate whether {{SolrInfoMBean.getStatistics()}} is needed at 
all - statistics reported by this method theoretically can be equally well 
reported using the metrics API.

> SolrInfoMBean-s category and hierarchy cleanup
> --
>
> Key: SOLR-9959
> URL: https://issues.apache.org/jira/browse/SOLR-9959
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Affects Versions: master (7.0)
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
>Priority: Blocker
> Fix For: master (7.0)
>
>
> SOLR-9947 changed categories of some of {{SolrInfoMBean-s}}, and it also 
> added an alternative view in JMX, similar to the one produced by 
> {{SolrJmxReporter}}.
> Some changes were left out from that issue because they would break the 
> back-compatibility in 6.x, but they should be done before 7.0:
> * remove the old JMX view of {{SolrInfoMBean}}-s and improve the new one so 
> that it's more readable and useful.
> * in many cases {{SolrInfoMBean.getName()}} just returns a FQCN, but it could 
> be more informative - eg. for highlighter or query plugins this could be the 
> symbolic name of a plugin that users know and use in configuration.
> * top-level categories need more thought. On one hand it's best to minimize 
> their number, on the other hand they need to meaningfully represent the 
> functionality of components that use them. SOLR-9947 made some cosmetic 
> changes, but more discussion is necessary (eg. QUERY vs. SEARCHHANDLER)
> * we should consider removing some of the methods in {{SolrInfoMBean}} that 
> usually don't return any useful information, eg. {{getDocs}}, {{getSource()}} 
> and {{getVersion()}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10130) Serious performance degradation in Solr 6.4.1 due to the new metrics collection

2017-03-09 Thread Shawn Heisey (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10130?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shawn Heisey updated SOLR-10130:

Affects Version/s: 6.4.0

> Serious performance degradation in Solr 6.4.1 due to the new metrics 
> collection
> ---
>
> Key: SOLR-10130
> URL: https://issues.apache.org/jira/browse/SOLR-10130
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Affects Versions: 6.4.1, 6.4.0
> Environment: Centos 7, OpenJDK 1.8.0 update 111
>Reporter: Ere Maijala
>Assignee: Andrzej Bialecki 
>Priority: Blocker
>  Labels: perfomance
> Fix For: master (7.0), 6.4.2
>
> Attachments: SOLR-10130.patch, SOLR-10130.patch, 
> solr-8983-console-f1.log
>
>
> We've stumbled on serious performance issues after upgrading to Solr 6.4.1. 
> Looks like the new metrics collection system in MetricsDirectoryFactory is 
> causing a major slowdown. This happens with an index configuration that, as 
> far as I can see, has no metrics specific configuration and uses 
> luceneMatchVersion 5.5.0. In practice a moderate load will completely bog 
> down the server with Solr threads constantly using up all CPU (600% on 6 core 
> machine) capacity with a load that normally  where we normally see an average 
> load of < 50%.
> I took stack traces (I'll attach them) and noticed that the threads are 
> spending time in com.codahale.metrics.Meter.mark. I tested building Solr 
> 6.4.1 with the metrics collection disabled in MetricsDirectoryFactory getByte 
> and getBytes methods and was unable to reproduce the issue.
> As far as I can see there are several issues:
> 1. Collecting metrics on every single byte read is slow.
> 2. Having it enabled by default is not a good idea.
> 3. The comment "enable coarse-grained metrics by default" at 
> https://github.com/apache/lucene-solr/blob/branch_6x/solr/core/src/java/org/apache/solr/update/SolrIndexConfig.java#L104
>  implies that only coarse-grained metrics should be enabled by default, and 
> this contradicts with collecting metrics on every single byte read.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-1105) Use a different stored field for highlighting

2017-03-09 Thread Matthew Caruana Galizia (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-1105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903258#comment-15903258
 ] 

Matthew Caruana Galizia edited comment on SOLR-1105 at 3/9/17 3:57 PM:
---

It also doesn't seem too difficult to implement on the UnifiedHighlighter, at 
least for someone who's familiar with the code.


was (Author: mcaruanagalizia):
It also doesn't seem to difficult to implement on the UnifiedHighlighter, at 
least for someone who's familiar with the code.

> Use a different stored field for highlighting
> -
>
> Key: SOLR-1105
> URL: https://issues.apache.org/jira/browse/SOLR-1105
> Project: Solr
>  Issue Type: Improvement
>  Components: highlighter
>Reporter: Dmitry Lihachev
> Attachments: SOLR-1105-1_4_1.patch, 
> SOLR-1105_shared_content_field_1.3.0.patch
>
>
> DefaultSolrHighlighter uses stored field content to highlight. It has some 
> disadvantages, because index grows up fast when using multilingual indexing 
> due to several fields has to be stored with same content. This patch allows 
> DefaultSolrHighlighter to use "contentField" attribute to loockup content in 
> external field.
> Excerpt from old schema:
> {code:xml}
> 
> 
> 
> 
> {code}
> The same after patching, highlighter will now get content stored in "title" 
> field
> {code:xml}
> 
>  contentField="title"/>
>  contentField="title"/>
>  contentField="title"/>
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-10247) Support non-numeric metrics

2017-03-09 Thread Andrzej Bialecki (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10247?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903323#comment-15903323
 ] 

Andrzej Bialecki  commented on SOLR-10247:
--

1. It doesn't, they are now registered as inline lambdas.
3. Yes, I'll fix this.
4. Good idea, I'll add this.

> Support non-numeric metrics
> ---
>
> Key: SOLR-10247
> URL: https://issues.apache.org/jira/browse/SOLR-10247
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
> Fix For: 6.5, master (7.0)
>
> Attachments: compactFormat.png, currentFormat.png, SOLR-10247.patch
>
>
> Metrics API currently supports only numeric values. However, it's useful also 
> to report non-numeric values such as eg. version, disk type, component state, 
> some system properties, etc.
> Codahale {{Gauge}} metric type can be used for this purpose, and 
> convenience methods can be added to {{SolrMetricManager}} to make it easier 
> to use.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5872) Eliminate overseer queue

2017-03-09 Thread albert vico oton (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5872?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903394#comment-15903394
 ] 

albert vico oton commented on SOLR-5872:


Already did that, but nodes still notify its state change, apparently 
collections need to know about other collections status in order to reroute 
queries to them, this amount of state change msg was killing our io in the ZK 
cluster, causing a lot of cpu wait time and effectively rendering the system 
unusable. But honestly that's as far as we went, we moved away from solrcloud 
and now we are using standalone solr inside each container for each of our 
collections and doing the balancing between replicas through nginx. 

> Eliminate overseer queue 
> -
>
> Key: SOLR-5872
> URL: https://issues.apache.org/jira/browse/SOLR-5872
> Project: Solr
>  Issue Type: Improvement
>  Components: SolrCloud
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The overseer queue is one of the busiest points in the entire system. The 
> raison d'être of the queue is
>  * Provide batching of operations for the main clusterstate,json so that 
> state updates are minimized 
> * Avoid race conditions and ensure order
> Now , as we move the individual collection states out of the main 
> clusterstate.json, the batching is not useful anymore.
> Race conditions can easily be solved by using a compare and set in Zookeeper. 
> The proposed solution  is , whenever an operation is required to be performed 
> on the clusterstate, the same thread (and of course the same JVM)
>  # read the fresh state and version of zk node  
>  # construct the new state 
>  # perform a compare and set
>  # if compare and set fails go to step 1
> This should be limited to all operations performed on external collections 
> because batching would be required for others 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-1105) Use a different stored field for highlighting

2017-03-09 Thread David Smiley (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-1105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Smiley updated SOLR-1105:
---
Fix Version/s: (was: 6.0)
   (was: 4.9)
  Summary: Use a different stored field for highlighting  (was: Using 
external field content for highlighting)

This would be very useful indeed.

> Use a different stored field for highlighting
> -
>
> Key: SOLR-1105
> URL: https://issues.apache.org/jira/browse/SOLR-1105
> Project: Solr
>  Issue Type: Improvement
>  Components: highlighter
>Reporter: Dmitry Lihachev
> Attachments: SOLR-1105-1_4_1.patch, 
> SOLR-1105_shared_content_field_1.3.0.patch
>
>
> DefaultSolrHighlighter uses stored field content to highlight. It has some 
> disadvantages, because index grows up fast when using multilingual indexing 
> due to several fields has to be stored with same content. This patch allows 
> DefaultSolrHighlighter to use "contentField" attribute to loockup content in 
> external field.
> Excerpt from old schema:
> {code:xml}
> 
> 
> 
> 
> {code}
> The same after patching, highlighter will now get content stored in "title" 
> field
> {code:xml}
> 
>  contentField="title"/>
>  contentField="title"/>
>  contentField="title"/>
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-10255) Large psuedo-stored fields via BinaryDocValuesField

2017-03-09 Thread Michael Braun (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10255?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903271#comment-15903271
 ] 

Michael Braun commented on SOLR-10255:
--

We have a use case that would be improved by this. 

> Large psuedo-stored fields via BinaryDocValuesField
> ---
>
> Key: SOLR-10255
> URL: https://issues.apache.org/jira/browse/SOLR-10255
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: David Smiley
>Assignee: David Smiley
> Attachments: SOLR-10255.patch
>
>
> (sub-issue of SOLR-10117)  This is a proposal for a better way for Solr to 
> handle "large" text fields.  Large docs that are in Lucene StoredFields slow 
> requests that don't involve access to such fields.  This is fundamental to 
> the fact that StoredFields are row-stored.  Worse, the Solr documentCache 
> will wind up holding onto massive Strings.  While the latter could be tackled 
> on it's own somehow as it's the most serious issue, nevertheless it seems 
> wrong that such large fields are in row-stored storage to begin with.  After 
> all, relational DBs seemed to have figured this out and put CLOBs/BLOBs in a 
> separate place.  Here, we do similarly by using, Lucene 
> {{BinaryDocValuesField}}.  BDVF isn't well known in the DocValues family as 
> it's not for typical DocValues purposes like sorting/faceting etc.  The 
> default DocValuesFormat doesn't compress these but we could write one that 
> does.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-7700) Move throughput control and merge aborting out of IndexWriter's core?

2017-03-09 Thread Dawid Weiss (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-7700?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Weiss updated LUCENE-7700:

Fix Version/s: master (7.0)
   6.x

> Move throughput control and merge aborting out of IndexWriter's core?
> -
>
> Key: LUCENE-7700
> URL: https://issues.apache.org/jira/browse/LUCENE-7700
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 6.x, master (7.0), 6.5
>
> Attachments: LUCENE-7700.patch, LUCENE-7700.patch, LUCENE-7700.patch
>
>
> Here is a bit of a background:
> - I wanted to implement a custom merging strategy that would have a custom 
> i/o flow control (global),
> - currently, the CMS is tightly bound with a few classes -- MergeRateLimiter, 
> OneMerge, IndexWriter.
> Looking at the code it seems to me that everything with respect to I/O 
> control could be nicely pulled out into classes that explicitly control the 
> merging process, that is only MergePolicy and MergeScheduler. By default, one 
> could even run without any additional I/O accounting overhead (which is 
> currently in there, even if one doesn't use the CMS's throughput control).
> Such refactoring would also give a chance to nicely move things where they 
> belong -- job aborting into OneMerge (currently in RateLimiter), rate limiter 
> lifecycle bound to OneMerge (MergeScheduler could then use per-merge or 
> global accounting, as it pleases).
> Just a thought and some initial refactorings for discussion.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-7700) Move throughput control and merge aborting out of IndexWriter's core?

2017-03-09 Thread Dawid Weiss (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-7700?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Weiss updated LUCENE-7700:

Fix Version/s: 6.5

> Move throughput control and merge aborting out of IndexWriter's core?
> -
>
> Key: LUCENE-7700
> URL: https://issues.apache.org/jira/browse/LUCENE-7700
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 6.x, master (7.0), 6.5
>
> Attachments: LUCENE-7700.patch, LUCENE-7700.patch, LUCENE-7700.patch
>
>
> Here is a bit of a background:
> - I wanted to implement a custom merging strategy that would have a custom 
> i/o flow control (global),
> - currently, the CMS is tightly bound with a few classes -- MergeRateLimiter, 
> OneMerge, IndexWriter.
> Looking at the code it seems to me that everything with respect to I/O 
> control could be nicely pulled out into classes that explicitly control the 
> merging process, that is only MergePolicy and MergeScheduler. By default, one 
> could even run without any additional I/O accounting overhead (which is 
> currently in there, even if one doesn't use the CMS's throughput control).
> Such refactoring would also give a chance to nicely move things where they 
> belong -- job aborting into OneMerge (currently in RateLimiter), rate limiter 
> lifecycle bound to OneMerge (MergeScheduler could then use per-merge or 
> global accounting, as it pleases).
> Just a thought and some initial refactorings for discussion.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Issue Comment Deleted] (SOLR-9959) SolrInfoMBean-s category and hierarchy cleanup

2017-03-09 Thread Andrzej Bialecki (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-9959?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrzej Bialecki  updated SOLR-9959:

Comment: was deleted

(was: There's also considerable duplication of code responsible for reporting 
metrics in components that implement both {{SolrInfoMBean}} and 
{{SolrMetricProducer}}. We should investigate whether 
{{SolrInfoMBean.getStatistics()}} is needed at all - statistics reported by 
this method theoretically can be equally well reported using the metrics API.)

> SolrInfoMBean-s category and hierarchy cleanup
> --
>
> Key: SOLR-9959
> URL: https://issues.apache.org/jira/browse/SOLR-9959
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Affects Versions: master (7.0)
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
>Priority: Blocker
> Fix For: master (7.0)
>
>
> SOLR-9947 changed categories of some of {{SolrInfoMBean-s}}, and it also 
> added an alternative view in JMX, similar to the one produced by 
> {{SolrJmxReporter}}.
> Some changes were left out from that issue because they would break the 
> back-compatibility in 6.x, but they should be done before 7.0:
> * remove the old JMX view of {{SolrInfoMBean}}-s and improve the new one so 
> that it's more readable and useful.
> * in many cases {{SolrInfoMBean.getName()}} just returns a FQCN, but it could 
> be more informative - eg. for highlighter or query plugins this could be the 
> symbolic name of a plugin that users know and use in configuration.
> * top-level categories need more thought. On one hand it's best to minimize 
> their number, on the other hand they need to meaningfully represent the 
> functionality of components that use them. SOLR-9947 made some cosmetic 
> changes, but more discussion is necessary (eg. QUERY vs. SEARCHHANDLER)
> * we should consider removing some of the methods in {{SolrInfoMBean}} that 
> usually don't return any useful information, eg. {{getDocs}}, {{getSource()}} 
> and {{getVersion()}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-7734) FieldType copy constructor should accept IndexableFieldType

2017-03-09 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15904404#comment-15904404
 ] 

ASF subversion and git services commented on LUCENE-7734:
-

Commit d2bf30d58fbfc9279bed663500400153b4d4df44 in lucene-solr's branch 
refs/heads/master from [~dsmiley]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=d2bf30d ]

LUCENE-7734: FieldType copy constructor widened to IndexableFieldType


> FieldType copy constructor should accept IndexableFieldType
> ---
>
> Key: LUCENE-7734
> URL: https://issues.apache.org/jira/browse/LUCENE-7734
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: David Smiley
>Assignee: David Smiley
>Priority: Minor
> Attachments: LUCENE_7734.patch
>
>
> {{FieldType}} is a concrete implementation of {{IndexableFieldType}}.  It has 
> a copy-constructor but it demands a {{FieldType}}.  It should accept  
> {{IndexableFieldType}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-7734) FieldType copy constructor should accept IndexableFieldType

2017-03-09 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15904415#comment-15904415
 ] 

David Smiley commented on LUCENE-7734:
--

So I'm in the middle of a cherry pick from master to branch_6x when I see that, 
surprise surprise, {{IndexableFieldType}} in 6x does not define 
{{numericType()}} or {{numericPrecisionStep()}} (both deprecated things 
associated with legacy numerics).  I could add an {{instanceof FieldType}} and 
cast to this copy constructor on 6x.  Or, just let this be a 7x thing.  I could 
go either way but I'm inclined to just let it be 7x.  If I don't hear any 
opinion further then I'll simply move the changes.txt entry to 7.0 and not 
touch 6x.

> FieldType copy constructor should accept IndexableFieldType
> ---
>
> Key: LUCENE-7734
> URL: https://issues.apache.org/jira/browse/LUCENE-7734
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: David Smiley
>Assignee: David Smiley
>Priority: Minor
> Attachments: LUCENE_7734.patch
>
>
> {{FieldType}} is a concrete implementation of {{IndexableFieldType}}.  It has 
> a copy-constructor but it demands a {{FieldType}}.  It should accept  
> {{IndexableFieldType}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS-EA] Lucene-Solr-master-Linux (64bit/jdk-9-ea+159) - Build # 19135 - Unstable!

2017-03-09 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/19135/
Java: 64bit/jdk-9-ea+159 -XX:+UseCompressedOops -XX:+UseSerialGC

3 tests failed.
FAILED:  junit.framework.TestSuite.org.apache.solr.util.TestSolrCLIRunExample

Error Message:
ObjectTracker found 6 object(s) that were not released!!! [TransactionLog, 
SolrIndexSearcher, MockDirectoryWrapper, MockDirectoryWrapper, 
MDCAwareThreadPoolExecutor, MockDirectoryWrapper] 
org.apache.solr.common.util.ObjectReleaseTracker$ObjectTrackerException: 
org.apache.solr.update.TransactionLog  at 
org.apache.solr.common.util.ObjectReleaseTracker.track(ObjectReleaseTracker.java:42)
  at org.apache.solr.update.TransactionLog.(TransactionLog.java:188)  at 
org.apache.solr.update.UpdateLog.newTransactionLog(UpdateLog.java:443)  at 
org.apache.solr.update.UpdateLog.ensureLog(UpdateLog.java:1102)  at 
org.apache.solr.update.UpdateLog.add(UpdateLog.java:529)  at 
org.apache.solr.update.UpdateLog.add(UpdateLog.java:514)  at 
org.apache.solr.update.DirectUpdateHandler2.doNormalUpdate(DirectUpdateHandler2.java:328)
  at 
org.apache.solr.update.DirectUpdateHandler2.addDoc0(DirectUpdateHandler2.java:247)
  at 
org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:202)
  at 
org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:67)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.DistributedUpdateProcessor.doLocalAdd(DistributedUpdateProcessor.java:987)
  at 
org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:1200)
  at 
org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:749)
  at 
org.apache.solr.update.processor.LogUpdateProcessorFactory$LogUpdateProcessor.processAdd(LogUpdateProcessorFactory.java:103)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.AddSchemaFieldsUpdateProcessorFactory$AddSchemaFieldsUpdateProcessor.processAdd(AddSchemaFieldsUpdateProcessorFactory.java:336)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldMutatingUpdateProcessor.processAdd(FieldMutatingUpdateProcessor.java:118)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldMutatingUpdateProcessor.processAdd(FieldMutatingUpdateProcessor.java:118)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldMutatingUpdateProcessor.processAdd(FieldMutatingUpdateProcessor.java:118)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldMutatingUpdateProcessor.processAdd(FieldMutatingUpdateProcessor.java:118)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldNameMutatingUpdateProcessorFactory$1.processAdd(FieldNameMutatingUpdateProcessorFactory.java:74)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.FieldMutatingUpdateProcessor.processAdd(FieldMutatingUpdateProcessor.java:118)
  at 
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:55)
  at 
org.apache.solr.update.processor.AbstractDefaultValueUpdateProcessorFactory$DefaultValueUpdateProcessor.processAdd(AbstractDefaultValueUpdateProcessorFactory.java:91)
  at 
org.apache.solr.handler.loader.JavabinLoader$1.update(JavabinLoader.java:97)  
at 
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readOuterMostDocIterator(JavaBinUpdateRequestCodec.java:186)
  at 
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readIterator(JavaBinUpdateRequestCodec.java:142)
  at org.apache.solr.common.util.JavaBinCodec.readObject(JavaBinCodec.java:313) 
 at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:258)  at 
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readNamedList(JavaBinUpdateRequestCodec.java:128)
  at org.apache.solr.common.util.JavaBinCodec.readObject(JavaBinCodec.java:278) 
 at org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:258)  at 
org.apache.solr.common.util.JavaBinCodec.unmarshal(JavaBinCodec.java:180)  at 
org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec.unmarshal(JavaBinUpdateRequestCodec.java:193)
  at 
org.apache.solr.handler.loader.JavabinLoader.parseAndLoadDocs(JavabinLoader.java:107)
  at org.apache.solr.handler.loader.JavabinLoader.load(JavabinLoader.java:54)  
at 

[jira] [Updated] (SOLR-10217) Add a query for the background set to the significantTerms streaming expression

2017-03-09 Thread Gethin James (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10217?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gethin James updated SOLR-10217:

Attachment: SOLR-20217.patch

Adding an initial patch for a significant terms background query.  I will work 
next on improving the tests.

> Add a query for the background set to the significantTerms streaming 
> expression
> ---
>
> Key: SOLR-10217
> URL: https://issues.apache.org/jira/browse/SOLR-10217
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Gethin James
> Attachments: SOLR-20217.patch
>
>
> Following the work on SOLR-10156 we now have a significantTerms expression.
> Currently, the background set is always the full index.  It would be great if 
> we could use a query to define the background set.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10076) Hiding keystore and truststore passwords from /admin/info/* outputs

2017-03-09 Thread Mano Kovacs (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mano Kovacs updated SOLR-10076:
---
Attachment: SOLR-10076.patch

[~markrmil...@gmail.com], thank you for the review and comments!

- I added test for case-sensitive property name, it in fact was not properly 
working.
- I changed the redaction text to the one that Greg added. Actually this patch 
is the generalization of his original intent.
- I made the system property redaction configurable with default true. 6.x 
backport only need to vary by the default value of that configuration to have 
it turned off.

> Hiding keystore and truststore passwords from /admin/info/* outputs
> ---
>
> Key: SOLR-10076
> URL: https://issues.apache.org/jira/browse/SOLR-10076
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Mano Kovacs
>Assignee: Mark Miller
> Attachments: SOLR-10076.patch, SOLR-10076.patch
>
>
> Passing keystore and truststore password is done by system properties, via 
> cmd line parameter.
> As result, {{/admin/info/properties}} and {{/admin/info/system}} will print 
> out the received password.
> Proposing solution to automatically redact value of any system property 
> before output, containing the word {{password}}, and replacing its value with 
> {{**}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-10258) Facet functions emit date fields as ticks

2017-03-09 Thread Chris Eldredge (JIRA)
Chris Eldredge created SOLR-10258:
-

 Summary: Facet functions emit date fields as ticks
 Key: SOLR-10258
 URL: https://issues.apache.org/jira/browse/SOLR-10258
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
  Components: faceting
Affects Versions: 5.5.2
Reporter: Chris Eldredge
Priority: Minor


When invoking a facet function in the JSON Facet API, TrieDateField is coerced 
into a numeric value and the result of the function is emitted as numeric 
instead of being converted back into a formatted date.

Example:

curl 
http://localhost:8983/solr/query?q=*:*={most_recent:'max(modified)'}

Produces (in part):

"facets":{
  "count":38304,
  "most_recent":1.489012400831E12}}

The "most_recent" attribute would be more useful if it was converted back to an 
iso8601 formatted date.

There was a thread discussing this issue in 2016: 
http://lucene.472066.n3.nabble.com/min-max-on-date-fields-using-JSON-facets-td4288736.html#a4288781



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-10257) Add logarithm StreamEvaluator

2017-03-09 Thread Joel Bernstein (JIRA)
Joel Bernstein created SOLR-10257:
-

 Summary: Add logarithm StreamEvaluator
 Key: SOLR-10257
 URL: https://issues.apache.org/jira/browse/SOLR-10257
 Project: Solr
  Issue Type: New Feature
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Joel Bernstein


To support custom scoring formulas with StreamEvaluators it would be useful to 
have a *log* evaluator that computes the natural log of a numeric value.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-10247) Support non-numeric metrics

2017-03-09 Thread Shalin Shekhar Mangar (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10247?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903144#comment-15903144
 ] 

Shalin Shekhar Mangar commented on SOLR-10247:
--

# The patch removes the following gauges from UpdateLog: replayLogsCountGauge, 
replayBytesGauge, stateGauge -- I think this is a mistake?
# nit - DirectUpdateHandlerTest has an unused import which will cause precommit 
to fail
# In IndexWriter's constructor, while registering the gauge {{() -> 
runningMajorMerges}}  -- shouldn't that be {{() -> runningMajorMerges.get()}}? 
and so on for all others
# I like the compact output -- we can add a version parameter to the 
MetricsHandler, say version=2, that outputs the compact format -- this can 
become the default in 7.0

> Support non-numeric metrics
> ---
>
> Key: SOLR-10247
> URL: https://issues.apache.org/jira/browse/SOLR-10247
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
> Fix For: 6.5, master (7.0)
>
> Attachments: compactFormat.png, currentFormat.png, SOLR-10247.patch
>
>
> Metrics API currently supports only numeric values. However, it's useful also 
> to report non-numeric values such as eg. version, disk type, component state, 
> some system properties, etc.
> Codahale {{Gauge}} metric type can be used for this purpose, and 
> convenience methods can be added to {{SolrMetricManager}} to make it easier 
> to use.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10217) Add a query for the background set to the significantTerms streaming expression

2017-03-09 Thread Gethin James (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10217?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gethin James updated SOLR-10217:

Attachment: (was: SOLR-20217.patch)

> Add a query for the background set to the significantTerms streaming 
> expression
> ---
>
> Key: SOLR-10217
> URL: https://issues.apache.org/jira/browse/SOLR-10217
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Gethin James
> Attachments: SOLR-20217.patch
>
>
> Following the work on SOLR-10156 we now have a significantTerms expression.
> Currently, the background set is always the full index.  It would be great if 
> we could use a query to define the background set.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10217) Add a query for the background set to the significantTerms streaming expression

2017-03-09 Thread Gethin James (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10217?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gethin James updated SOLR-10217:

Attachment: SOLR-20217.patch

I added a new patch file.  If you prefer a pull request let me know, I have the 
changes on a branch from master 
(https://github.com/covolution/lucene-solr/tree/SOLR-10217).

> Add a query for the background set to the significantTerms streaming 
> expression
> ---
>
> Key: SOLR-10217
> URL: https://issues.apache.org/jira/browse/SOLR-10217
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Gethin James
> Attachments: SOLR-20217.patch
>
>
> Following the work on SOLR-10156 we now have a significantTerms expression.
> Currently, the background set is always the full index.  It would be great if 
> we could use a query to define the background set.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10258) Facet functions emit date fields as ticks

2017-03-09 Thread Yonik Seeley (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonik Seeley updated SOLR-10258:

Component/s: (was: faceting)
 Facet Module

> Facet functions emit date fields as ticks
> -
>
> Key: SOLR-10258
> URL: https://issues.apache.org/jira/browse/SOLR-10258
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: Facet Module
>Affects Versions: 5.5.2
>Reporter: Chris Eldredge
>Priority: Minor
>
> When invoking a facet function in the JSON Facet API, TrieDateField is 
> coerced into a numeric value and the result of the function is emitted as 
> numeric instead of being converted back into a formatted date.
> Example:
> curl 
> http://localhost:8983/solr/query?q=*:*={most_recent:'max(modified)'}
> Produces (in part):
> "facets":{
>   "count":38304,
>   "most_recent":1.489012400831E12}}
> The "most_recent" attribute would be more useful if it was converted back to 
> an iso8601 formatted date.
> There was a thread discussing this issue in 2016: 
> http://lucene.472066.n3.nabble.com/min-max-on-date-fields-using-JSON-facets-td4288736.html#a4288781



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10259) admin/cores?action=STATUS returns 500 when a single core has init failures

2017-03-09 Thread Oliver Bates (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Oliver Bates updated SOLR-10259:

Attachment: SOLR-10259.patch-2.txt
SOLR-10259.patch-1.txt

These are two simple patch ideas.

> admin/cores?action=STATUS returns 500 when a single core has init failures
> --
>
> Key: SOLR-10259
> URL: https://issues.apache.org/jira/browse/SOLR-10259
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>Affects Versions: 5.3
>Reporter: Oliver Bates
>Priority: Trivial
> Attachments: SOLR-10259.patch-1.txt, SOLR-10259.patch-2.txt
>
>
> When I have a healthy core on a node and I call 
> solr/admin/cores?action=STATUS, I get the following healthy response:
> {quote}
> 
>   
> 0
> 1607
>   
>   
>   
> 
>   whoisbanana_shard1_replica1
>   
> /tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/
>   
>   
> /tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/
>   
>   solrconfig.xml
>   schema.xml
>   2017-03-08T15:59:50.18Z
>   380431
>   active
>   0
>   
>   0
>   0
>   0
>   0
>   2
>   0
>   true
>   false
>   
> org.apache.lucene.store.NRTCachingDirectory:NRTCachingDirectory(MMapDirectory@/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index
>  lockFactory=org.apache.lucene.store.NativeFSLockFactory@762404a0; 
> maxCacheMB=48.0 maxMergeSizeMB=4.0)
>   
>   
>   71
>   71 bytes
> 
>   
> 
> 
> {quote}
> If I then corrupt the index file and reload, e.g. like this:
> echo "cheese" >> 
> /tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1
> And then I call the same endpoint (solr/admin/cores?action=STATUS), I get a 
> 500 back:
> {quote}
> 
>   
> 500
> 1508
>   
>   
> Error handling 'status' action
> 
> org.apache.solr.common.SolrException: Error handling 'status' action at 
> org.apache.solr.handler.admin.CoreAdminHandler.handleStatusAction(CoreAdminHandler.java:755)
>  at 
> org.apache.solr.handler.admin.CoreAdminHandler.handleRequestInternal(CoreAdminHandler.java:231)
>  at 
> org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:196)
>  at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:146)
>  at 
> org.apache.solr.servlet.HttpSolrCall.handleAdminRequest(HttpSolrCall.java:676)
>  at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:443) at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:210)
>  at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:179)
>  at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
>  at com.apple.cie.search.auth.TrustFilter.doFilter(TrustFilter.java:44) at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
>  at com.apple.cie.search.id.IdFilter.doFilter(IdFilter.java:38) at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
>  at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) 
> at 
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577) 
> at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223)
>  at 
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1127)
>  at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) 
> at 
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
>  at 
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1061)
>  at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) 
> at 
> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215)
>  at 
> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110)
>  at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97)
>  at org.eclipse.jetty.server.Server.handle(Server.java:499) at 
> org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:310) at 
> org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:257) 
> at org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) 
> at 
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635)
>  at 
> org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555)
>  at 

[jira] [Updated] (SOLR-10259) admin/cores?action=STATUS returns 500 when a single core has init failures

2017-03-09 Thread Oliver Bates (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Oliver Bates updated SOLR-10259:

Description: 
When I have a healthy core on a node and I call solr/admin/cores?action=STATUS, 
I get the following healthy response:

{quote}

  
0
1607
  
  
  

  whoisbanana_shard1_replica1
  
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/
  
  
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/
  
  solrconfig.xml
  schema.xml
  2017-03-08T15:59:50.18Z
  380431
  active
  0
  
  0
  0
  0
  0
  2
  0
  true
  false
  
org.apache.lucene.store.NRTCachingDirectory:NRTCachingDirectory(MMapDirectory@/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index
 lockFactory=org.apache.lucene.store.NativeFSLockFactory@762404a0; 
maxCacheMB=48.0 maxMergeSizeMB=4.0)
  
  
  71
  71 bytes

  


{quote}

If I then corrupt the index file and reload, e.g. like this:
echo "cheese" >> 
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1

And then I call the same endpoint (solr/admin/cores?action=STATUS), I get a 500 
back:

{quote}

  
500
1508
  
  
Error handling 'status' action

org.apache.solr.common.SolrException: Error handling 'status' action at 
org.apache.solr.handler.admin.CoreAdminHandler.handleStatusAction(CoreAdminHandler.java:755)
 at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestInternal(CoreAdminHandler.java:231)
 at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:196)
 at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:146)
 at 
org.apache.solr.servlet.HttpSolrCall.handleAdminRequest(HttpSolrCall.java:676) 
at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:443) at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:210)
 at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:179)
 at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at com.apple.cie.search.auth.TrustFilter.doFilter(TrustFilter.java:44) at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at com.apple.cie.search.id.IdFilter.doFilter(IdFilter.java:38) at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) 
at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) 
at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577) 
at 
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1127)
 at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) 
at 
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1061)
 at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) 
at 
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215)
 at 
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110)
 at 
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) 
at org.eclipse.jetty.server.Server.handle(Server.java:499) at 
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:310) at 
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:257) at 
org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) at 
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635)
 at 
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555) 
at java.lang.Thread.run(Thread.java:745) Caused by: 
org.apache.lucene.index.CorruptIndexException: misplaced codec footer (file 
extended?): remaining=23, expected=16 
(resource=BufferedChecksumIndexInput(MMapIndexInput(path="/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1")))
 at org.apache.lucene.codecs.CodecUtil.validateFooter(CodecUtil.java:411) at 
org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:331) at 
org.apache.lucene.index.SegmentInfos.readCommit(SegmentInfos.java:442) at 
org.apache.lucene.index.SegmentInfos$1.doBody(SegmentInfos.java:493) at 
org.apache.lucene.index.SegmentInfos$1.doBody(SegmentInfos.java:490) at 
org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:731)
 at 
org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:683)
 at 

[jira] [Updated] (SOLR-10259) admin/cores?action=STATUS returns 500 when a single core has init failures

2017-03-09 Thread Oliver Bates (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Oliver Bates updated SOLR-10259:

Description: 
When I have a healthy core on a node and I call solr/admin/cores?action=STATUS, 
I get the following healthy response:


  
0
1607
  
  
  

  whoisbanana_shard1_replica1
  
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/
  
  
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/
  
  solrconfig.xml
  schema.xml
  2017-03-08T15:59:50.18Z
  380431
  active
  0
  
  0
  0
  0
  0
  2
  0
  true
  false
  
org.apache.lucene.store.NRTCachingDirectory:NRTCachingDirectory(MMapDirectory@/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index
 lockFactory=org.apache.lucene.store.NativeFSLockFactory@762404a0; 
maxCacheMB=48.0 maxMergeSizeMB=4.0)
  
  
  71
  71 bytes

  



If I then corrupt the index file and reload, e.g. like this:
echo "cheese" >> 
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1

And then I call the same endpoint (solr/admin/cores?action=STATUS), I get a 500 
back:


  
500
1508
  
  
Error handling 'status' action

org.apache.solr.common.SolrException: Error handling 'status' action at 
org.apache.solr.handler.admin.CoreAdminHandler.handleStatusAction(CoreAdminHandler.java:755)
 at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestInternal(CoreAdminHandler.java:231)
 at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:196)
 at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:146)
 at 
org.apache.solr.servlet.HttpSolrCall.handleAdminRequest(HttpSolrCall.java:676) 
at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:443) at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:210)
 at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:179)
 at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at com.apple.cie.search.auth.TrustFilter.doFilter(TrustFilter.java:44) at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at com.apple.cie.search.id.IdFilter.doFilter(IdFilter.java:38) at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) 
at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) 
at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577) 
at 
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1127)
 at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) 
at 
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1061)
 at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) 
at 
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215)
 at 
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110)
 at 
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) 
at org.eclipse.jetty.server.Server.handle(Server.java:499) at 
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:310) at 
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:257) at 
org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) at 
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635)
 at 
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555) 
at java.lang.Thread.run(Thread.java:745) Caused by: 
org.apache.lucene.index.CorruptIndexException: misplaced codec footer (file 
extended?): remaining=23, expected=16 
(resource=BufferedChecksumIndexInput(MMapIndexInput(path="/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1")))
 at org.apache.lucene.codecs.CodecUtil.validateFooter(CodecUtil.java:411) at 
org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:331) at 
org.apache.lucene.index.SegmentInfos.readCommit(SegmentInfos.java:442) at 
org.apache.lucene.index.SegmentInfos$1.doBody(SegmentInfos.java:493) at 
org.apache.lucene.index.SegmentInfos$1.doBody(SegmentInfos.java:490) at 
org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:731)
 at 
org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:683)
 at 

[jira] [Commented] (SOLR-5872) Eliminate overseer queue

2017-03-09 Thread Shawn Heisey (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5872?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903557#comment-15903557
 ] 

Shawn Heisey commented on SOLR-5872:


bq. We were doing our tests with solr 6.4.1

That version has other problems that may be clouding the issue (pun intended).  
See SOLR-10130.  I advise an immediate upgrade to 6.4.2, to be sure that any 
problems you're encountering are actually caused by SolrCloud, not high CPU 
usage.


> Eliminate overseer queue 
> -
>
> Key: SOLR-5872
> URL: https://issues.apache.org/jira/browse/SOLR-5872
> Project: Solr
>  Issue Type: Improvement
>  Components: SolrCloud
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The overseer queue is one of the busiest points in the entire system. The 
> raison d'être of the queue is
>  * Provide batching of operations for the main clusterstate,json so that 
> state updates are minimized 
> * Avoid race conditions and ensure order
> Now , as we move the individual collection states out of the main 
> clusterstate.json, the batching is not useful anymore.
> Race conditions can easily be solved by using a compare and set in Zookeeper. 
> The proposed solution  is , whenever an operation is required to be performed 
> on the clusterstate, the same thread (and of course the same JVM)
>  # read the fresh state and version of zk node  
>  # construct the new state 
>  # perform a compare and set
>  # if compare and set fails go to step 1
> This should be limited to all operations performed on external collections 
> because batching would be required for others 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-10259) admin/cores?action=STATUS returns 500 when a single core has init failures

2017-03-09 Thread Oliver Bates (JIRA)
Oliver Bates created SOLR-10259:
---

 Summary: admin/cores?action=STATUS returns 500 when a single core 
has init failures
 Key: SOLR-10259
 URL: https://issues.apache.org/jira/browse/SOLR-10259
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Affects Versions: 5.3
Reporter: Oliver Bates
Priority: Trivial


When I have a healthy core on a node and I call solr/admin/cores?action=STATUS, 
I get the following:

# Healthy response 

  
0
1607
  
  
  

  whoisbanana_shard1_replica1
  
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/
  
  
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/
  
  solrconfig.xml
  schema.xml
  2017-03-08T15:59:50.18Z
  380431
  active
  0
  
  0
  0
  0
  0
  2
  0
  true
  false
  

org.apache.lucene.store.NRTCachingDirectory:NRTCachingDirectory(MMapDirectory@/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index
 lockFactory=org.apache.lucene.store.NativeFSLockFactory@762404a0; 
maxCacheMB=48.0 maxMergeSizeMB=4.0)
  
  
  71
  71 bytes

  


###

If I then corrupt the index file and reload, e.g. like this:
echo "cheese" >> 
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1

And then I call the same endpoint (solr/admin/cores?action=STATUS), I get a 500 
back:

 Corrupted response 

  
500
1508
  
  
Error handling 'status' action

org.apache.solr.common.SolrException: Error handling 'status' action at 
org.apache.solr.handler.admin.CoreAdminHandler.handleStatusAction(CoreAdminHandler.java:755)
 at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestInternal(CoreAdminHandler.java:231)
 at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:196)
 at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:146)
 at 
org.apache.solr.servlet.HttpSolrCall.handleAdminRequest(HttpSolrCall.java:676) 
at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:443) at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:210)
 at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:179)
 at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at com.apple.cie.search.auth.TrustFilter.doFilter(TrustFilter.java:44) at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at com.apple.cie.search.id.IdFilter.doFilter(IdFilter.java:38) at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) 
at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) 
at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577) 
at 
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1127)
 at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) 
at 
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1061)
 at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) 
at 
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215)
 at 
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110)
 at 
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) 
at org.eclipse.jetty.server.Server.handle(Server.java:499) at 
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:310) at 
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:257) at 
org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) at 
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635)
 at 
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555) 
at java.lang.Thread.run(Thread.java:745) Caused by: 
org.apache.lucene.index.CorruptIndexException: misplaced codec footer (file 
extended?): remaining=23, expected=16 
(resource=BufferedChecksumIndexInput(MMapIndexInput(path="/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1")))
 at org.apache.lucene.codecs.CodecUtil.validateFooter(CodecUtil.java:411) at 
org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:331) at 
org.apache.lucene.index.SegmentInfos.readCommit(SegmentInfos.java:442) at 

[jira] [Commented] (SOLR-10217) Add a query for the background set to the significantTerms streaming expression

2017-03-09 Thread Dennis Gove (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903656#comment-15903656
 ] 

Dennis Gove commented on SOLR-10217:


What's the reasoning behind having SignificantTermsStream accept a query 
instead of an incoming stream? If it accepted an incoming stream then the 
source data could be the result of any other stream whereas with a query the 
source data can only be the result of a Solr query.

> Add a query for the background set to the significantTerms streaming 
> expression
> ---
>
> Key: SOLR-10217
> URL: https://issues.apache.org/jira/browse/SOLR-10217
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Gethin James
> Attachments: SOLR-20217.patch
>
>
> Following the work on SOLR-10156 we now have a significantTerms expression.
> Currently, the background set is always the full index.  It would be great if 
> we could use a query to define the background set.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10259) admin/cores?action=STATUS returns 500 when a single core has init failures

2017-03-09 Thread Oliver Bates (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Oliver Bates updated SOLR-10259:

Description: 
When I have a healthy core on a node and I call solr/admin/cores?action=STATUS, 
I get the following healthy response:

{quote}

  
0
1607
  
  
  

  whoisbanana_shard1_replica1
  
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/
  
  
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/
  
  solrconfig.xml
  schema.xml
  2017-03-08T15:59:50.18Z
  380431
  active
  0
  
  0
  0
  0
  0
  2
  0
  true
  false
  
org.apache.lucene.store.NRTCachingDirectory:NRTCachingDirectory(MMapDirectory@/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index
 lockFactory=org.apache.lucene.store.NativeFSLockFactory@762404a0; 
maxCacheMB=48.0 maxMergeSizeMB=4.0)
  
  
  71
  71 bytes

  


{quote}

If I then corrupt the index file and reload, e.g. like this:
echo "cheese" >> 
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1

And then I call the same endpoint (solr/admin/cores?action=STATUS), I get a 500 
back:

{quote}

  
500
1508
  
  
Error handling 'status' action

org.apache.solr.common.SolrException: Error handling 'status' action at 
org.apache.solr.handler.admin.CoreAdminHandler.handleStatusAction(CoreAdminHandler.java:755)
 at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestInternal(CoreAdminHandler.java:231)
 at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:196)
 at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:146)
 at 
org.apache.solr.servlet.HttpSolrCall.handleAdminRequest(HttpSolrCall.java:676) 
at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:443) at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:210)
 at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:179)
 at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at com.apple.cie.search.auth.TrustFilter.doFilter(TrustFilter.java:44) at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at com.apple.cie.search.id.IdFilter.doFilter(IdFilter.java:38) at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) 
at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) 
at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577) 
at 
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1127)
 at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) 
at 
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1061)
 at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) 
at 
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215)
 at 
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110)
 at 
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) 
at org.eclipse.jetty.server.Server.handle(Server.java:499) at 
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:310) at 
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:257) at 
org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) at 
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635)
 at 
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555) 
at java.lang.Thread.run(Thread.java:745) Caused by: 
org.apache.lucene.index.CorruptIndexException: misplaced codec footer (file 
extended?): remaining=23, expected=16 
(resource=BufferedChecksumIndexInput(MMapIndexInput(path="/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1")))
 at org.apache.lucene.codecs.CodecUtil.validateFooter(CodecUtil.java:411) at 
org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:331) at 
org.apache.lucene.index.SegmentInfos.readCommit(SegmentInfos.java:442) at 
org.apache.lucene.index.SegmentInfos$1.doBody(SegmentInfos.java:493) at 
org.apache.lucene.index.SegmentInfos$1.doBody(SegmentInfos.java:490) at 
org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:731)
 at 
org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:683)
 at 

[jira] [Updated] (SOLR-10259) admin/cores?action=STATUS returns 500 when a single core has init failures

2017-03-09 Thread Oliver Bates (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Oliver Bates updated SOLR-10259:

Description: 
When I have a healthy core on a node and I call solr/admin/cores?action=STATUS, 
I get the following healthy response:


  
0
1607
  
  
  

  whoisbanana_shard1_replica1
  
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/
  
  
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/
  
  solrconfig.xml
  schema.xml
  2017-03-08T15:59:50.18Z
  380431
  active
  0
  
  0
  0
  0
  0
  2
  0
  true
  false
  
org.apache.lucene.store.NRTCachingDirectory:NRTCachingDirectory(MMapDirectory@/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index
 lockFactory=org.apache.lucene.store.NativeFSLockFactory@762404a0; 
maxCacheMB=48.0 maxMergeSizeMB=4.0)
  
  
  71
  71 bytes

  



If I then corrupt the index file and reload, e.g. like this:
echo "cheese" >> 
/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1

And then I call the same endpoint (solr/admin/cores?action=STATUS), I get a 500 
back:


  
500
1508
  
  
Error handling 'status' action

org.apache.solr.common.SolrException: Error handling 'status' action at 
org.apache.solr.handler.admin.CoreAdminHandler.handleStatusAction(CoreAdminHandler.java:755)
 at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestInternal(CoreAdminHandler.java:231)
 at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:196)
 at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:146)
 at 
org.apache.solr.servlet.HttpSolrCall.handleAdminRequest(HttpSolrCall.java:676) 
at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:443) at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:210)
 at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:179)
 at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at com.apple.cie.search.auth.TrustFilter.doFilter(TrustFilter.java:44) at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at com.apple.cie.search.id.IdFilter.doFilter(IdFilter.java:38) at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) 
at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) 
at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577) 
at 
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1127)
 at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) 
at 
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
 at 
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1061)
 at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) 
at 
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:215)
 at 
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:110)
 at 
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) 
at org.eclipse.jetty.server.Server.handle(Server.java:499) at 
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:310) at 
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:257) at 
org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) at 
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635)
 at 
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555) 
at java.lang.Thread.run(Thread.java:745) Caused by: 
org.apache.lucene.index.CorruptIndexException: misplaced codec footer (file 
extended?): remaining=23, expected=16 
(resource=BufferedChecksumIndexInput(MMapIndexInput(path="/tmp/search_integration_test/solr1/whoisbanana_shard1_replica1/data/index/segments_1")))
 at org.apache.lucene.codecs.CodecUtil.validateFooter(CodecUtil.java:411) at 
org.apache.lucene.codecs.CodecUtil.checkFooter(CodecUtil.java:331) at 
org.apache.lucene.index.SegmentInfos.readCommit(SegmentInfos.java:442) at 
org.apache.lucene.index.SegmentInfos$1.doBody(SegmentInfos.java:493) at 
org.apache.lucene.index.SegmentInfos$1.doBody(SegmentInfos.java:490) at 
org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:731)
 at 
org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:683)
 at 

[JENKINS] Lucene-Solr-6.4-Windows (64bit/jdk1.8.0_121) - Build # 27 - Unstable!

2017-03-09 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-6.4-Windows/27/
Java: 64bit/jdk1.8.0_121 -XX:+UseCompressedOops -XX:+UseParallelGC

1 tests failed.
FAILED:  
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testCollectionsAPI

Error Message:
No live SolrServers available to handle this 
request:[http://127.0.0.1:60801/solr/awhollynewcollection_0]

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: No live SolrServers available 
to handle this request:[http://127.0.0.1:60801/solr/awhollynewcollection_0]
at 
__randomizedtesting.SeedInfo.seed([D229BD309379F302:9A5CC984954ADC97]:0)
at 
org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:414)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1347)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1098)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:1037)
at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:149)
at org.apache.solr.client.solrj.SolrClient.query(SolrClient.java:942)
at 
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testCollectionsAPI(CollectionsAPIDistributedZkTest.java:516)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1713)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:907)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:957)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:367)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:811)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:462)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:916)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:802)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:852)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 

[jira] [Created] (SOLR-10260) Collection not found with count(*) and uppercase name

2017-03-09 Thread Jordan Diehl (JIRA)
Jordan Diehl created SOLR-10260:
---

 Summary: Collection not found with count(*) and uppercase name
 Key: SOLR-10260
 URL: https://issues.apache.org/jira/browse/SOLR-10260
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
  Components: Parallel SQL
Affects Versions: 6.4.1
Reporter: Jordan Diehl
Priority: Minor


I'm on solr 6.4.1 and was testing out using Zeppelin with Solr. I found that 
Parallel SQL queries using aggregate functions do not seem to work with 
collections which have capital letters in the name. 

When searching around for information on this I found [this 
thread|http://lucene.472066.n3.nabble.com/JDBC-Collection-not-found-with-count-and-uppercase-name-td4287590.html]
 where someone else encountered the same issue and was told to open a bug on 
it, but it doesn't look like they ever did.

The thread has an example query and stack trace.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10257) Add logarithm StreamEvaluator

2017-03-09 Thread Joel Bernstein (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10257?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joel Bernstein updated SOLR-10257:
--
Attachment: SOLR-10257.patch

> Add logarithm StreamEvaluator
> -
>
> Key: SOLR-10257
> URL: https://issues.apache.org/jira/browse/SOLR-10257
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Joel Bernstein
> Attachments: SOLR-10257.patch
>
>
> To support custom scoring formulas with StreamEvaluators it would be useful 
> to have a *log* evaluator that computes the natural log of a numeric value.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-7734) FieldType copy constructor should accept IndexableFieldType

2017-03-09 Thread Adrien Grand (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15904063#comment-15904063
 ] 

Adrien Grand commented on LUCENE-7734:
--

True, but this change is API compatible, so if something compiles today, it 
will still compile with your patch. I think the only thing it breaks is ABI 
compatibility so it would not be possible eg. to replace the Lucene JAR without 
recompiling. But this is not something we guarantee anyway, I don't think we 
should add an additional constructor on 6.x.

> FieldType copy constructor should accept IndexableFieldType
> ---
>
> Key: LUCENE-7734
> URL: https://issues.apache.org/jira/browse/LUCENE-7734
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: David Smiley
>Assignee: David Smiley
>Priority: Minor
> Attachments: LUCENE_7734.patch
>
>
> {{FieldType}} is a concrete implementation of {{IndexableFieldType}}.  It has 
> a copy-constructor but it demands a {{FieldType}}.  It should accept  
> {{IndexableFieldType}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5872) Eliminate overseer queue

2017-03-09 Thread albert vico oton (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5872?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903961#comment-15903961
 ] 

albert vico oton commented on SOLR-5872:


thanks for the advice, I'll take a look but the problem we were seeing was in 
Zookeeper cluster not in solr 

> Eliminate overseer queue 
> -
>
> Key: SOLR-5872
> URL: https://issues.apache.org/jira/browse/SOLR-5872
> Project: Solr
>  Issue Type: Improvement
>  Components: SolrCloud
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The overseer queue is one of the busiest points in the entire system. The 
> raison d'être of the queue is
>  * Provide batching of operations for the main clusterstate,json so that 
> state updates are minimized 
> * Avoid race conditions and ensure order
> Now , as we move the individual collection states out of the main 
> clusterstate.json, the batching is not useful anymore.
> Race conditions can easily be solved by using a compare and set in Zookeeper. 
> The proposed solution  is , whenever an operation is required to be performed 
> on the clusterstate, the same thread (and of course the same JVM)
>  # read the fresh state and version of zk node  
>  # construct the new state 
>  # perform a compare and set
>  # if compare and set fails go to step 1
> This should be limited to all operations performed on external collections 
> because batching would be required for others 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-10217) Add a query for the background set to the significantTerms streaming expression

2017-03-09 Thread Joel Bernstein (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15904049#comment-15904049
 ] 

Joel Bernstein commented on SOLR-10217:
---

Gethin and I had discussed using stream for the background query. But it just 
seemed like the right approach in this scenario to make it a query. The 
algorithm is written with an AnalyticsQuery and works directly with the index, 
so collecting a bitset for the background query scales really well. Using a 
stream we would have to feed docID's back into the AnalyticsQuery to get the 
background set, just seemed a lot less scalable.



> Add a query for the background set to the significantTerms streaming 
> expression
> ---
>
> Key: SOLR-10217
> URL: https://issues.apache.org/jira/browse/SOLR-10217
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Gethin James
> Attachments: SOLR-20217.patch
>
>
> Following the work on SOLR-10156 we now have a significantTerms expression.
> Currently, the background set is always the full index.  It would be great if 
> we could use a query to define the background set.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-10217) Add a query for the background set to the significantTerms streaming expression

2017-03-09 Thread Joel Bernstein (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15904049#comment-15904049
 ] 

Joel Bernstein edited comment on SOLR-10217 at 3/9/17 11:19 PM:


Gethin and I had discussed using a stream for the background query. But it just 
seemed like the right approach in this scenario was to make it a query. The 
algorithm is written with an AnalyticsQuery and works directly with the index, 
so collecting a bitset for the background query scales really well. Using a 
stream we would have to feed docID's back into the AnalyticsQuery to get the 
background set, just seemed a lot less scalable.




was (Author: joel.bernstein):
Gethin and I had discussed using stream for the background query. But it just 
seemed like the right approach in this scenario to make it a query. The 
algorithm is written with an AnalyticsQuery and works directly with the index, 
so collecting a bitset for the background query scales really well. Using a 
stream we would have to feed docID's back into the AnalyticsQuery to get the 
background set, just seemed a lot less scalable.



> Add a query for the background set to the significantTerms streaming 
> expression
> ---
>
> Key: SOLR-10217
> URL: https://issues.apache.org/jira/browse/SOLR-10217
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Gethin James
> Attachments: SOLR-20217.patch
>
>
> Following the work on SOLR-10156 we now have a significantTerms expression.
> Currently, the background set is always the full index.  It would be great if 
> we could use a query to define the background set.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (LUCENE-7735) Suffix compression for points

2017-03-09 Thread Adrien Grand (JIRA)
Adrien Grand created LUCENE-7735:


 Summary: Suffix compression for points
 Key: LUCENE-7735
 URL: https://issues.apache.org/jira/browse/LUCENE-7735
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Adrien Grand
Priority: Minor


Points already compress fairly well based on shared prefixes. It would be nice 
if points could also compress suffixes similarly to how doc values perform GCD 
compression.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10217) Add a query for the background set to the significantTerms streaming expression

2017-03-09 Thread Joel Bernstein (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10217?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joel Bernstein updated SOLR-10217:
--
Attachment: SOLR-10257.patch

> Add a query for the background set to the significantTerms streaming 
> expression
> ---
>
> Key: SOLR-10217
> URL: https://issues.apache.org/jira/browse/SOLR-10217
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Gethin James
> Attachments: SOLR-10257.patch, SOLR-20217.patch
>
>
> Following the work on SOLR-10156 we now have a significantTerms expression.
> Currently, the background set is always the full index.  It would be great if 
> we could use a query to define the background set.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10257) Add logarithm StreamEvaluator

2017-03-09 Thread Joel Bernstein (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10257?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joel Bernstein updated SOLR-10257:
--
Attachment: SOLR-10257.patch

> Add logarithm StreamEvaluator
> -
>
> Key: SOLR-10257
> URL: https://issues.apache.org/jira/browse/SOLR-10257
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Joel Bernstein
> Attachments: SOLR-10257.patch, SOLR-10257.patch
>
>
> To support custom scoring formulas with StreamEvaluators it would be useful 
> to have a *log* evaluator that computes the natural log of a numeric value.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10217) Add a query for the background set to the significantTerms streaming expression

2017-03-09 Thread Joel Bernstein (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10217?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joel Bernstein updated SOLR-10217:
--
Attachment: (was: SOLR-10257.patch)

> Add a query for the background set to the significantTerms streaming 
> expression
> ---
>
> Key: SOLR-10217
> URL: https://issues.apache.org/jira/browse/SOLR-10217
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Gethin James
> Attachments: SOLR-20217.patch
>
>
> Following the work on SOLR-10156 we now have a significantTerms expression.
> Currently, the background set is always the full index.  It would be great if 
> we could use a query to define the background set.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10257) Add logarithm StreamEvaluator

2017-03-09 Thread Joel Bernstein (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10257?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joel Bernstein updated SOLR-10257:
--
Attachment: SOLR-10257.patch

> Add logarithm StreamEvaluator
> -
>
> Key: SOLR-10257
> URL: https://issues.apache.org/jira/browse/SOLR-10257
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Joel Bernstein
> Attachments: SOLR-10257.patch, SOLR-10257.patch, SOLR-10257.patch
>
>
> To support custom scoring formulas with StreamEvaluators it would be useful 
> to have a *log* evaluator that computes the natural log of a numeric value.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10184) bin/solr fails to run on java9 due to unrecognized GC options

2017-03-09 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10184?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-10184:

Attachment: SOLR-10184.patch

thanks for the links Uwe, JEP 158 is interesting.

Between the GC logging options that have been removed, and how the unified 
logging now deals with multiple tags (gc, all, etc..) & output specifiers 
(stdout, stderr, for file) in a single '-Xlog' param; it doesn't look viable to 
have an _exact_ java9 equivilent for what the {{bin/solr}} script currently 
does with {{SOLR_LOGS_DIR}} & {{GC_LOG_OPTS}}

The attached patch attempts to support the same ideas as best as (i can see) 
possible under java9...

* if the user doesn't configure a {{GC_LOG_OPTS}}, then we (still) use a 
(verbose) default value for this param via '-Xlog:gc\*'
* if the user configures a blank value for {{GC_LOG_OPTS}}, we (still) leave 
the param blank
* if the effective value of {{GC_LOG_OPTS}} is non blank, then foreach param 
option specified in {{GC_LOG_OPTS}}:
** if it starts with '-Xlog:gc', and does not include an output specifier, then:
*** add a 'file' output specifier based on {{SOLR_LOGS_DIR}} (to mimic the 
'-Xloggc' param we use w/ java8)
*** add {{time,uptime}} decorators (to mimic the -XX params we use w/ java8)
*** add {{filecount=9,filesize=2}} output-options (to mimic the -XX params 
we use w/ java8)

(for simplicity, if the user wants to specify multiple JEP158 tags in the 
'-Xlog' param , then the tag list has to start with "gc" in order for us to add 
a file option -- otherwise we ignore it.  likewise if the user wants to specify 
their own decorators or output-options, then they must also specify their own 
output (file, or stdout, or stderr) specifier as well, since it must come first 
in the '-Xlog' format string)

While fixing this, I also did some small cleanup in how the {{java -version}} 
output is parsed.

Would appreciate detailed review & manual testing from folks since this type of 
script work is hard to write automated tests for.

> bin/solr fails to run on java9 due to unrecognized GC options
> -
>
> Key: SOLR-10184
> URL: https://issues.apache.org/jira/browse/SOLR-10184
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: scripts and tools
>Reporter: Hoss Man
>  Labels: java9
> Attachments: SOLR-10184.patch
>
>
> {noformat}
> hossman@tray:~/lucene/dev/solr [master] $ bin/solr start -f
> Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was 
> deprecated in version 9.0 and will likely be removed in a future release.
> Java HotSpot(TM) 64-Bit Server VM warning: Option UseParNewGC was deprecated 
> in version 9.0 and will likely be removed in a future release.
> Unrecognized VM option 'PrintHeapAtGC'
> Error: Could not create the Java Virtual Machine.
> Error: A fatal exception has occurred. Program will exit.
> {noformat}
> (untested) workaround is to override GC_LOG_OPTS in {{solr.in.sh}}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-10262) Collect request latency metrics for histograms

2017-03-09 Thread Andrzej Bialecki (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10262?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15904588#comment-15904588
 ] 

Andrzej Bialecki  commented on SOLR-10262:
--

Thanks Otis :) Yes, I'm aware of that implementation and the discussion around 
it. It's an easy fix to replace the ExponentiallyDecayingReservoir that metrics 
use by default with HdrHistogram, there's even a pull request against Codahale 
repo with an implementation.

By "making it possible" did you mean that we make it the new default, but we 
also allow people to use a different implementation? This would have to be 
configured early on in {{solr.xml}} or even via system properties, which is a 
bit ugly.

> Collect request latency metrics for histograms
> --
>
> Key: SOLR-10262
> URL: https://issues.apache.org/jira/browse/SOLR-10262
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Reporter: Otis Gospodnetic
>
> Since [~ab] is on a role with metrics...
> There is no way to accurately compute request latency percentiles from 
> metrics exposed by Solr today. We should consider making that possible. c.f. 
> https://github.com/HdrHistogram/HdrHistogram



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-10261) TestStressCloudBlindAtomicUpdates.test_dv() fail

2017-03-09 Thread Cao Manh Dat (JIRA)
Cao Manh Dat created SOLR-10261:
---

 Summary: TestStressCloudBlindAtomicUpdates.test_dv() fail
 Key: SOLR-10261
 URL: https://issues.apache.org/jira/browse/SOLR-10261
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Cao Manh Dat


I found a reproducing seed that cause 
TestStressCloudBlindAtomicUpdates.test_dv() fail

{code}
[junit4]   2> NOTE: reproduce with: ant test  
-Dtestcase=TestStressCloudBlindAtomicUpdates -Dtests.method=test_dv 
-Dtests.seed=AD8E7B56D53B627F -Dtests.nightly=true -Dtests.slow=true 
-Dtests.locale=bg -Dtests.timezone=America/La_Paz -Dtests.asserts=true 
-Dtests.file.encoding=UTF-8
   [junit4] ERROR   1.21s J2 | TestStressCloudBlindAtomicUpdates.test_dv <<<
   [junit4]> Throwable #1: java.util.concurrent.ExecutionException: 
java.lang.RuntimeException: Error from server at 
http://127.0.0.1:49825/solr/test_col: Async exception during distributed 
update: Error from server at 
http://127.0.0.1:49824/solr/test_col_shard2_replica2: Server Error
   [junit4]> request: 
http://127.0.0.1:49824/solr/test_col_shard2_replica2/update?update.distrib=TOLEADER=http%3A%2F%2F127.0.0.1%3A49825%2Fsolr%2Ftest_col_shard5_replica1%2F=javabin=2
   [junit4]> Remote error message: Failed synchronous update on shard 
StdNode: http://127.0.0.1:49836/solr/test_col_shard2_replica1/ update: 
org.apache.solr.client.solrj.request.UpdateRequest@5919dfb3
   [junit4]>at 
__randomizedtesting.SeedInfo.seed([AD8E7B56D53B627F:9B9A19105F66586E]:0)
   [junit4]>at 
java.util.concurrent.FutureTask.report(FutureTask.java:122)
   [junit4]>at 
java.util.concurrent.FutureTask.get(FutureTask.java:192)
   [junit4]>at 
org.apache.solr.cloud.TestStressCloudBlindAtomicUpdates.checkField(TestStressCloudBlindAtomicUpdates.java:281)
   [junit4]>at 
org.apache.solr.cloud.TestStressCloudBlindAtomicUpdates.test_dv(TestStressCloudBlindAtomicUpdates.java:193)
   [junit4]>at java.lang.Thread.run(Thread.java:745)
   [junit4]> Caused by: java.lang.RuntimeException: Error from server at 
http://127.0.0.1:49825/solr/test_col: Async exception during distributed 
update: Error from server at 
http://127.0.0.1:49824/solr/test_col_shard2_replica2: Server Error
   [junit4]> request: 
http://127.0.0.1:49824/solr/test_col_shard2_replica2/update?update.distrib=TOLEADER=http%3A%2F%2F127.0.0.1%3A49825%2Fsolr%2Ftest_col_shard5_replica1%2F=javabin=2
   [junit4]> Remote error message: Failed synchronous update on shard 
StdNode: http://127.0.0.1:49836/solr/test_col_shard2_replica1/ update: 
org.apache.solr.client.solrj.request.UpdateRequest@5919dfb3
   [junit4]>at 
org.apache.solr.cloud.TestStressCloudBlindAtomicUpdates$Worker.run(TestStressCloudBlindAtomicUpdates.java:409)
   [junit4]>at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
   [junit4]>at 
java.util.concurrent.FutureTask.run(FutureTask.java:266)
   [junit4]>at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:229)
   [junit4]>at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
   [junit4]>at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
   [junit4]>... 1 more
   [junit4]> Caused by: 
org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error 
from server at http://127.0.0.1:49825/solr/test_col: Async exception during 
distributed update: Error from server at 
{code}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-10257) Add logarithm StreamEvaluator

2017-03-09 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10257?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15904267#comment-15904267
 ] 

ASF subversion and git services commented on SOLR-10257:


Commit b0a973e083ef9750acc69600480212f391448f6a in lucene-solr's branch 
refs/heads/branch_6x from [~joel.bernstein]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=b0a973e ]

SOLR-10257: Add logarithm StreamEvaluator


> Add logarithm StreamEvaluator
> -
>
> Key: SOLR-10257
> URL: https://issues.apache.org/jira/browse/SOLR-10257
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Joel Bernstein
> Attachments: SOLR-10257.patch, SOLR-10257.patch, SOLR-10257.patch
>
>
> To support custom scoring formulas with StreamEvaluators it would be useful 
> to have a *log* evaluator that computes the natural log of a numeric value.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-10257) Add logarithm StreamEvaluator

2017-03-09 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10257?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15904255#comment-15904255
 ] 

ASF subversion and git services commented on SOLR-10257:


Commit d945a246f6071699790119f07a66fb4c5505cee2 in lucene-solr's branch 
refs/heads/master from [~joel.bernstein]
[ https://git-wip-us.apache.org/repos/asf?p=lucene-solr.git;h=d945a24 ]

SOLR-10257: Add logarithm StreamEvaluator


> Add logarithm StreamEvaluator
> -
>
> Key: SOLR-10257
> URL: https://issues.apache.org/jira/browse/SOLR-10257
> Project: Solr
>  Issue Type: New Feature
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Joel Bernstein
> Attachments: SOLR-10257.patch, SOLR-10257.patch, SOLR-10257.patch
>
>
> To support custom scoring formulas with StreamEvaluators it would be useful 
> to have a *log* evaluator that computes the natural log of a numeric value.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-10262) Collect request latency metrics for histograms

2017-03-09 Thread Otis Gospodnetic (JIRA)
Otis Gospodnetic created SOLR-10262:
---

 Summary: Collect request latency metrics for histograms
 Key: SOLR-10262
 URL: https://issues.apache.org/jira/browse/SOLR-10262
 Project: Solr
  Issue Type: New Feature
  Security Level: Public (Default Security Level. Issues are Public)
  Components: metrics
Reporter: Otis Gospodnetic


Since [~ab] is on a role with metrics...
There is no way to accurately compute request latency percentiles from metrics 
exposed by Solr today. We should consider making that possible. c.f. 
https://github.com/HdrHistogram/HdrHistogram




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-master-Solaris (64bit/jdk1.8.0) - Build # 1183 - Unstable!

2017-03-09 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Solaris/1183/
Java: 64bit/jdk1.8.0 -XX:-UseCompressedOops -XX:+UseParallelGC

1 tests failed.
FAILED:  
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testCollectionsAPI

Error Message:
Error from server at http://127.0.0.1:48825/solr/awhollynewcollection_0: 
Expected mime type application/octet-stream but got text/html.   
 
Error 510HTTP ERROR: 510 Problem 
accessing /solr/awhollynewcollection_0/select. Reason: 
{metadata={error-class=org.apache.solr.common.SolrException,root-error-class=org.apache.solr.common.SolrException},msg={awhollynewcollection_0:6},code=510}
 http://eclipse.org/jetty;>Powered by Jetty:// 
9.3.14.v20161028   

Stack Trace:
org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error 
from server at http://127.0.0.1:48825/solr/awhollynewcollection_0: Expected 
mime type application/octet-stream but got text/html. 


Error 510 


HTTP ERROR: 510
Problem accessing /solr/awhollynewcollection_0/select. Reason:

{metadata={error-class=org.apache.solr.common.SolrException,root-error-class=org.apache.solr.common.SolrException},msg={awhollynewcollection_0:6},code=510}
http://eclipse.org/jetty;>Powered by Jetty:// 
9.3.14.v20161028



at 
__randomizedtesting.SeedInfo.seed([EF11849929B5888A:A764F02D2F86A71F]:0)
at 
org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:595)
at 
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:279)
at 
org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:268)
at 
org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:439)
at 
org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:391)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1361)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1112)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1215)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1215)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1215)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1215)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:1215)
at 
org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:1042)
at 
org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:160)
at org.apache.solr.client.solrj.SolrClient.query(SolrClient.java:942)
at 
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testCollectionsAPI(CollectionsAPIDistributedZkTest.java:523)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1713)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:907)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:957)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:916)
at 

[jira] [Commented] (LUCENE-7700) Move throughput control and merge aborting out of IndexWriter's core?

2017-03-09 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7700?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903178#comment-15903178
 ] 

Michael McCandless commented on LUCENE-7700:


I think it's fine to backport to 6.x.

> Move throughput control and merge aborting out of IndexWriter's core?
> -
>
> Key: LUCENE-7700
> URL: https://issues.apache.org/jira/browse/LUCENE-7700
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Attachments: LUCENE-7700.patch, LUCENE-7700.patch, LUCENE-7700.patch
>
>
> Here is a bit of a background:
> - I wanted to implement a custom merging strategy that would have a custom 
> i/o flow control (global),
> - currently, the CMS is tightly bound with a few classes -- MergeRateLimiter, 
> OneMerge, IndexWriter.
> Looking at the code it seems to me that everything with respect to I/O 
> control could be nicely pulled out into classes that explicitly control the 
> merging process, that is only MergePolicy and MergeScheduler. By default, one 
> could even run without any additional I/O accounting overhead (which is 
> currently in there, even if one doesn't use the CMS's throughput control).
> Such refactoring would also give a chance to nicely move things where they 
> belong -- job aborting into OneMerge (currently in RateLimiter), rate limiter 
> lifecycle bound to OneMerge (MergeScheduler could then use per-merge or 
> global accounting, as it pleases).
> Just a thought and some initial refactorings for discussion.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-6.4-Linux (64bit/jdk1.8.0_121) - Build # 166 - Unstable!

2017-03-09 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-6.4-Linux/166/
Java: 64bit/jdk1.8.0_121 -XX:+UseCompressedOops -XX:+UseSerialGC

2 tests failed.
FAILED:  org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.test

Error Message:
Expected 2 of 3 replicas to be active but only found 1; 
[core_node3:{"core":"c8n_1x3_lf_shard1_replica3","base_url":"http://127.0.0.1:44068","node_name":"127.0.0.1:44068_","state":"active","leader":"true"}];
 clusterState: 
DocCollection(c8n_1x3_lf//collections/c8n_1x3_lf/state.json/17)={   
"replicationFactor":"3",   "shards":{"shard1":{   
"range":"8000-7fff",   "state":"active",   "replicas":{ 
"core_node1":{   "core":"c8n_1x3_lf_shard1_replica2",   
"base_url":"http://127.0.0.1:36373;,   "node_name":"127.0.0.1:36373_",  
 "state":"down"}, "core_node2":{   "state":"down",  
 "base_url":"http://127.0.0.1:38232;,   
"core":"c8n_1x3_lf_shard1_replica1",   "node_name":"127.0.0.1:38232_"}, 
"core_node3":{   "core":"c8n_1x3_lf_shard1_replica3",   
"base_url":"http://127.0.0.1:44068;,   "node_name":"127.0.0.1:44068_",  
 "state":"active",   "leader":"true",   
"router":{"name":"compositeId"},   "maxShardsPerNode":"1",   
"autoAddReplicas":"false"}

Stack Trace:
java.lang.AssertionError: Expected 2 of 3 replicas to be active but only found 
1; 
[core_node3:{"core":"c8n_1x3_lf_shard1_replica3","base_url":"http://127.0.0.1:44068","node_name":"127.0.0.1:44068_","state":"active","leader":"true"}];
 clusterState: DocCollection(c8n_1x3_lf//collections/c8n_1x3_lf/state.json/17)={
  "replicationFactor":"3",
  "shards":{"shard1":{
  "range":"8000-7fff",
  "state":"active",
  "replicas":{
"core_node1":{
  "core":"c8n_1x3_lf_shard1_replica2",
  "base_url":"http://127.0.0.1:36373;,
  "node_name":"127.0.0.1:36373_",
  "state":"down"},
"core_node2":{
  "state":"down",
  "base_url":"http://127.0.0.1:38232;,
  "core":"c8n_1x3_lf_shard1_replica1",
  "node_name":"127.0.0.1:38232_"},
"core_node3":{
  "core":"c8n_1x3_lf_shard1_replica3",
  "base_url":"http://127.0.0.1:44068;,
  "node_name":"127.0.0.1:44068_",
  "state":"active",
  "leader":"true",
  "router":{"name":"compositeId"},
  "maxShardsPerNode":"1",
  "autoAddReplicas":"false"}
at 
__randomizedtesting.SeedInfo.seed([B7A4D9D9CA77C0CF:3FF0E603648BAD37]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.assertTrue(Assert.java:43)
at 
org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.testRf3WithLeaderFailover(LeaderFailoverAfterPartitionTest.java:168)
at 
org.apache.solr.cloud.LeaderFailoverAfterPartitionTest.test(LeaderFailoverAfterPartitionTest.java:55)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1713)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:907)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:957)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:992)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:967)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:367)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:811)
at 

[JENKINS] Lucene-Solr-NightlyTests-master - Build # 1258 - Failure

2017-03-09 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-master/1258/

2 tests failed.
FAILED:  
org.apache.lucene.search.join.TestBlockJoin.testMultiChildQueriesOfDiffParentLevels

Error Message:
Java heap space

Stack Trace:
java.lang.OutOfMemoryError: Java heap space
at 
__randomizedtesting.SeedInfo.seed([49607EEF4FD70C96:98856123AF41F5F0]:0)
at java.util.Arrays.copyOf(Arrays.java:3308)
at 
org.apache.lucene.codecs.compressing.CompressingStoredFieldsIndexReader.(CompressingStoredFieldsIndexReader.java:77)
at 
org.apache.lucene.codecs.compressing.CompressingStoredFieldsReader.(CompressingStoredFieldsReader.java:134)
at 
org.apache.lucene.codecs.compressing.CompressingStoredFieldsFormat.fieldsReader(CompressingStoredFieldsFormat.java:121)
at 
org.apache.lucene.index.SegmentCoreReaders.(SegmentCoreReaders.java:126)
at org.apache.lucene.index.SegmentReader.(SegmentReader.java:77)
at 
org.apache.lucene.index.ReadersAndUpdates.getReader(ReadersAndUpdates.java:143)
at 
org.apache.lucene.index.IndexWriter.mergeMiddle(IndexWriter.java:4486)
at org.apache.lucene.index.IndexWriter.merge(IndexWriter.java:3925)
at 
org.apache.lucene.index.SerialMergeScheduler.merge(SerialMergeScheduler.java:40)
at org.apache.lucene.index.IndexWriter.maybeMerge(IndexWriter.java:2091)
at 
org.apache.lucene.index.IndexWriter.doAfterSegmentFlushed(IndexWriter.java:4971)
at 
org.apache.lucene.index.DocumentsWriter$MergePendingEvent.process(DocumentsWriter.java:731)
at 
org.apache.lucene.index.IndexWriter.processEvents(IndexWriter.java:5009)
at 
org.apache.lucene.index.IndexWriter.processEvents(IndexWriter.java:5000)
at 
org.apache.lucene.index.IndexWriter.updateDocuments(IndexWriter.java:1395)
at 
org.apache.lucene.index.IndexWriter.addDocuments(IndexWriter.java:1368)
at 
org.apache.lucene.index.RandomIndexWriter.addDocuments(RandomIndexWriter.java:203)
at 
org.apache.lucene.search.join.TestBlockJoin.testMultiChildQueriesOfDiffParentLevels(TestBlockJoin.java:1433)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1713)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:907)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:957)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)


FAILED:  org.apache.solr.cloud.TestStressCloudBlindAtomicUpdates.test_dv

Error Message:
java.lang.RuntimeException: Error from server at 
http://127.0.0.1:42500/solr/test_col: Async exception during distributed 
update: Error from server at 
http://127.0.0.1:55012/solr/test_col_shard2_replica2: Server Errorrequest: 
http://127.0.0.1:55012/solr/test_col_shard2_replica2/update?update.distrib=TOLEADER=http%3A%2F%2F127.0.0.1%3A42500%2Fsolr%2Ftest_col_shard3_replica1%2F=javabin=2
 Remote error message: Failed synchronous update on shard StdNode: 
http://127.0.0.1:50853/solr/test_col_shard2_replica1/ update: 
org.apache.solr.client.solrj.request.UpdateRequest@524e0599

Stack Trace:
java.util.concurrent.ExecutionException: java.lang.RuntimeException: Error from 
server at http://127.0.0.1:42500/solr/test_col: Async exception during 
distributed update: Error from server at 
http://127.0.0.1:55012/solr/test_col_shard2_replica2: Server Error



request: 
http://127.0.0.1:55012/solr/test_col_shard2_replica2/update?update.distrib=TOLEADER=http%3A%2F%2F127.0.0.1%3A42500%2Fsolr%2Ftest_col_shard3_replica1%2F=javabin=2
Remote error message: Failed synchronous update on shard StdNode: 
http://127.0.0.1:50853/solr/test_col_shard2_replica1/ update: 
org.apache.solr.client.solrj.request.UpdateRequest@524e0599
at 
__randomizedtesting.SeedInfo.seed([DCDB68FD95F2034:3BD9D4C953021A25]:0)
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.util.concurrent.FutureTask.get(FutureTask.java:192)
at 

[jira] [Commented] (LUCENE-7722) Remove BoostedQuery

2017-03-09 Thread Alan Woodward (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7722?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15902838#comment-15902838
 ] 

Alan Woodward commented on LUCENE-7722:
---

Rewriting is a bit academic for now anyway, let's punt that to another issue 
and just remove the custom scoring queries here.

> Remove BoostedQuery
> ---
>
> Key: LUCENE-7722
> URL: https://issues.apache.org/jira/browse/LUCENE-7722
> Project: Lucene - Core
>  Issue Type: Task
>Reporter: Adrien Grand
>Priority: Minor
>
> We already  have FunctionScoreQuery, which is more flexible than BoostedQuery 
> as it can combine scores in arbitrary ways and only requests scores on the 
> underlying scorer if they are needed. So let's remove BoostedQuery?



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-7700) Move throughput control and merge aborting out of IndexWriter's core?

2017-03-09 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7700?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15902896#comment-15902896
 ] 

Michael McCandless commented on LUCENE-7700:


Thanks [~dweiss] the new patch looks great!  +1 to push, and thank you for 
cleaning this up!

> Move throughput control and merge aborting out of IndexWriter's core?
> -
>
> Key: LUCENE-7700
> URL: https://issues.apache.org/jira/browse/LUCENE-7700
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Attachments: LUCENE-7700.patch, LUCENE-7700.patch, LUCENE-7700.patch
>
>
> Here is a bit of a background:
> - I wanted to implement a custom merging strategy that would have a custom 
> i/o flow control (global),
> - currently, the CMS is tightly bound with a few classes -- MergeRateLimiter, 
> OneMerge, IndexWriter.
> Looking at the code it seems to me that everything with respect to I/O 
> control could be nicely pulled out into classes that explicitly control the 
> merging process, that is only MergePolicy and MergeScheduler. By default, one 
> could even run without any additional I/O accounting overhead (which is 
> currently in there, even if one doesn't use the CMS's throughput control).
> Such refactoring would also give a chance to nicely move things where they 
> belong -- job aborting into OneMerge (currently in RateLimiter), rate limiter 
> lifecycle bound to OneMerge (MergeScheduler could then use per-merge or 
> global accounting, as it pleases).
> Just a thought and some initial refactorings for discussion.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS-EA] Lucene-Solr-6.4-Linux (32bit/jdk-9-ea+159) - Build # 161 - Still Unstable!

2017-03-09 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-6.4-Linux/161/
Java: 32bit/jdk-9-ea+159 -client -XX:+UseParallelGC

1 tests failed.
FAILED:  org.apache.solr.cloud.ShardSplitTest.testSplitAfterFailedSplit

Error Message:
expected:<1> but was:<2>

Stack Trace:
java.lang.AssertionError: expected:<1> but was:<2>
at 
__randomizedtesting.SeedInfo.seed([E7D10676E3E9E46D:1E9C95D9DF9CA9E7]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.failNotEquals(Assert.java:647)
at org.junit.Assert.assertEquals(Assert.java:128)
at org.junit.Assert.assertEquals(Assert.java:472)
at org.junit.Assert.assertEquals(Assert.java:456)
at 
org.apache.solr.cloud.ShardSplitTest.testSplitAfterFailedSplit(ShardSplitTest.java:280)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:547)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1713)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:907)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:957)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:992)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:967)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:367)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:811)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:462)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:916)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:802)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:852)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 

[jira] [Commented] (LUCENE-7700) Move throughput control and merge aborting out of IndexWriter's core?

2017-03-09 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-7700?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15902900#comment-15902900
 ] 

Dawid Weiss commented on LUCENE-7700:
-

Thanks for reviewing, Mike. I'm guessing this should go to 7.x only, right? Or 
can we backport it to 6.x as well, risking some minor API incompatibilities 
(these are internal APIs anyway, but who knows).

> Move throughput control and merge aborting out of IndexWriter's core?
> -
>
> Key: LUCENE-7700
> URL: https://issues.apache.org/jira/browse/LUCENE-7700
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Attachments: LUCENE-7700.patch, LUCENE-7700.patch, LUCENE-7700.patch
>
>
> Here is a bit of a background:
> - I wanted to implement a custom merging strategy that would have a custom 
> i/o flow control (global),
> - currently, the CMS is tightly bound with a few classes -- MergeRateLimiter, 
> OneMerge, IndexWriter.
> Looking at the code it seems to me that everything with respect to I/O 
> control could be nicely pulled out into classes that explicitly control the 
> merging process, that is only MergePolicy and MergeScheduler. By default, one 
> could even run without any additional I/O accounting overhead (which is 
> currently in there, even if one doesn't use the CMS's throughput control).
> Such refactoring would also give a chance to nicely move things where they 
> belong -- job aborting into OneMerge (currently in RateLimiter), rate limiter 
> lifecycle bound to OneMerge (MergeScheduler could then use per-merge or 
> global accounting, as it pleases).
> Just a thought and some initial refactorings for discussion.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5276) highlighter working using stemmed tokens from another field and text from another

2017-03-09 Thread Matthew Caruana Galizia (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5276?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15902687#comment-15902687
 ] 

Matthew Caruana Galizia commented on SOLR-5276:
---

The default highlighter has become the unified highlighter since then, but 
required feature is the same.

> highlighter working using stemmed tokens from another field and text from 
> another
> -
>
> Key: SOLR-5276
> URL: https://issues.apache.org/jira/browse/SOLR-5276
> Project: Solr
>  Issue Type: Improvement
>  Components: highlighter
>Reporter: Maciej Lizewski
>Priority: Minor
>
> The case is like this:
> I have 'content' field with content extracted with tika and several fields 
> for language versions (like content_pl, content_en, content_es, content_ru, 
> etc). 
> I also use custom langid component which can copy 'content' to serveral 
> content_* fields if it detects more than one language (so those parts are 
> properly stemmed in every language present in text).
> Now to use highlighter in such scenario I need to store all those language 
> fields even if their contents is always same as the one in 'content' field.
> Would be nice if I could configure language specific fields to be not stored, 
> and configure highlighter to take tokens positions from those fields and 
> apply them to text in 'content' field...
> In other words - to say: take tokens from 'content_pl', and prepare highlight 
> based on text in 'content' field.
> It could be administrator responsibility to guarantee that mapped fields have 
> same content.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-5276) highlighter working using stemmed tokens from another field and text from another

2017-03-09 Thread Matthew Caruana Galizia (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5276?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15902687#comment-15902687
 ] 

Matthew Caruana Galizia edited comment on SOLR-5276 at 3/9/17 8:44 AM:
---

The default highlighter has become the unified highlighter since then, but 
required feature is the same as SOLR-1105.


was (Author: mcaruanagalizia):
The default highlighter has become the unified highlighter since then, but 
required feature is the same.

> highlighter working using stemmed tokens from another field and text from 
> another
> -
>
> Key: SOLR-5276
> URL: https://issues.apache.org/jira/browse/SOLR-5276
> Project: Solr
>  Issue Type: Improvement
>  Components: highlighter
>Reporter: Maciej Lizewski
>Priority: Minor
>
> The case is like this:
> I have 'content' field with content extracted with tika and several fields 
> for language versions (like content_pl, content_en, content_es, content_ru, 
> etc). 
> I also use custom langid component which can copy 'content' to serveral 
> content_* fields if it detects more than one language (so those parts are 
> properly stemmed in every language present in text).
> Now to use highlighter in such scenario I need to store all those language 
> fields even if their contents is always same as the one in 'content' field.
> Would be nice if I could configure language specific fields to be not stored, 
> and configure highlighter to take tokens positions from those fields and 
> apply them to text in 'content' field...
> In other words - to say: take tokens from 'content_pl', and prepare highlight 
> based on text in 'content' field.
> It could be administrator responsibility to guarantee that mapped fields have 
> same content.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10256) Parentheses in SpellCheckCollator

2017-03-09 Thread Abhishek Kumar Singh (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10256?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Kumar Singh updated SOLR-10256:

Description: 
SpellCheckCollator adds parentheses ( *'('* and *')'* ) around tokens which 
have space between them.  
This should be configurable, because if *_WordBreakSpellCheckComponent_* is 
being used, queries like : *applejuice* will be broken down to *apple juice*. 
Such suggestions are being surrounded by braces by current 
*SpellCheckCollator*. 
And when surrounded by brackets, they represent the same position by 
_EdismaxParser_ , which is not required. 

https://github.com/apache/lucene-solr/blob/master/solr/core/src/java/org/apache/solr/spelling/SpellCheckCollator.java#L227
  

A solution to this will be to have a flag, which can help disable this 
parenthesisation of spell check suggestions.

  was:
SpellCheckCollator adds parentheses ( *'('* and *')'* ) around tokens which 
have space between them.  
This should be configurable, because if *_WordBreakSpellCheckComponent_* is 
being used, queries like : *applejuice* will be broken down to *apple juice*. 
Such suggestions are being surrounded by braces by current 
*SpellCheckCollator*. 
And when surrounded by brackets, they represent the same position, which is not 
required. 

https://github.com/apache/lucene-solr/blob/master/solr/core/src/java/org/apache/solr/spelling/SpellCheckCollator.java#L227
  

A solution to this will be to have a flag, which can help disable this 
parenthesisation of spell check suggestions.


> Parentheses in SpellCheckCollator
> -
>
> Key: SOLR-10256
> URL: https://issues.apache.org/jira/browse/SOLR-10256
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: spellchecker
>Reporter: Abhishek Kumar Singh
>
> SpellCheckCollator adds parentheses ( *'('* and *')'* ) around tokens which 
> have space between them.  
> This should be configurable, because if *_WordBreakSpellCheckComponent_* is 
> being used, queries like : *applejuice* will be broken down to *apple juice*. 
> Such suggestions are being surrounded by braces by current 
> *SpellCheckCollator*. 
> And when surrounded by brackets, they represent the same position by 
> _EdismaxParser_ , which is not required. 
> https://github.com/apache/lucene-solr/blob/master/solr/core/src/java/org/apache/solr/spelling/SpellCheckCollator.java#L227
>   
> A solution to this will be to have a flag, which can help disable this 
> parenthesisation of spell check suggestions.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-5872) Eliminate overseer queue

2017-03-09 Thread albert vico oton (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5872?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15890021#comment-15890021
 ] 

albert vico oton edited comment on SOLR-5872 at 3/9/17 11:51 AM:
-

Hello, we are currently trying to do a deploy of around 200 collections and 
solrcloud can't handle it, it just  dies due update_status messages propagation 
everytime we try to add a new collection, each collection has 3 replicas, and 
sizes are not very large. Also, I do not see why collection A should be aware 
of collection B state.  

But moving to the topic, overseer node dies since he can not handle all the 
stress due the flooding of messages. IMHO we have here a single point of 
failure in a distributed system, which is not very recommended. 

since it would be useful for big fat shards, my suggestion would be to make 
this optional behavior, so people like us, who need to have a more distributed 
approach, can make use of solrcloud. Since right now it is impossible to. and 
I'm not talking about "thousands" of collections actually with as few as 100 we 
are seeing very bad performance.




was (Author: alvico):
Hello, we are currently trying to do a deploy of around 200 collections and 
solrcloud can't handle it, it just  dies due update_status messages propagation 
everytime we try to add a new collection, each collection has 3 replicas, and 
sizes are not very large. Also, I do not see why collection A should be aware 
of collection B state.  

But moving to the topic, overseer node dies since he can not handle all the 
stress due the flooding of messages. IMHO we have here a single point of 
failure in a distributed system, which is not very recommended. 

since it would be useful for big fat shards, my suggestion would be to make 
this optional behavior, so people like use who need to have a more distributed 
approach can make use of solrcloud. Since right now it is impossible to. and 
I'm not talking about "thousands" of collections actually with as few as 100 we 
are seeing very bad performance.



> Eliminate overseer queue 
> -
>
> Key: SOLR-5872
> URL: https://issues.apache.org/jira/browse/SOLR-5872
> Project: Solr
>  Issue Type: Improvement
>  Components: SolrCloud
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The overseer queue is one of the busiest points in the entire system. The 
> raison d'être of the queue is
>  * Provide batching of operations for the main clusterstate,json so that 
> state updates are minimized 
> * Avoid race conditions and ensure order
> Now , as we move the individual collection states out of the main 
> clusterstate.json, the batching is not useful anymore.
> Race conditions can easily be solved by using a compare and set in Zookeeper. 
> The proposed solution  is , whenever an operation is required to be performed 
> on the clusterstate, the same thread (and of course the same JVM)
>  # read the fresh state and version of zk node  
>  # construct the new state 
>  # perform a compare and set
>  # if compare and set fails go to step 1
> This should be limited to all operations performed on external collections 
> because batching would be required for others 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5872) Eliminate overseer queue

2017-03-09 Thread albert vico oton (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5872?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15902928#comment-15902928
 ] 

albert vico oton commented on SOLR-5872:


We were doing our tests with solr 6.4.1

> Eliminate overseer queue 
> -
>
> Key: SOLR-5872
> URL: https://issues.apache.org/jira/browse/SOLR-5872
> Project: Solr
>  Issue Type: Improvement
>  Components: SolrCloud
>Reporter: Noble Paul
>Assignee: Noble Paul
>
> The overseer queue is one of the busiest points in the entire system. The 
> raison d'être of the queue is
>  * Provide batching of operations for the main clusterstate,json so that 
> state updates are minimized 
> * Avoid race conditions and ensure order
> Now , as we move the individual collection states out of the main 
> clusterstate.json, the batching is not useful anymore.
> Race conditions can easily be solved by using a compare and set in Zookeeper. 
> The proposed solution  is , whenever an operation is required to be performed 
> on the clusterstate, the same thread (and of course the same JVM)
>  # read the fresh state and version of zk node  
>  # construct the new state 
>  # perform a compare and set
>  # if compare and set fails go to step 1
> This should be limited to all operations performed on external collections 
> because batching would be required for others 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



org.apache.lucene.index.CorruptIndexException: codec header mismatch

2017-03-09 Thread rosh
Hi,
   lucene search is throwing back this error what does this mean and how can
the indexes be fixed ? Or is rebuilding indexes the only option ?

getIndexReader  
org.apache.lucene.index.CorruptIndexException: codec header mismatch: actual
header=2883712 vs expected header=1071082519 (resource:
MMapIndexInput(path="\\..\_inwu.si"))
at org.apache.lucene.codecs.CodecUtil.checkHeader(CodecUtil.java:128)
~[lucene-core-4.3.0.jar:4.3.0 1477023 - simonw - 2013-04-29 14:55:14]
at
org.apache.lucene.codecs.lucene40.Lucene40SegmentInfoReader.read(Lucene40SegmentInfoReader.java:53)
~[lucene-core-4.3.0.jar:4.3.0 1477023 - simonw - 2013-04-29 14:55:14]
at org.apache.lucene.index.SegmentInfos.read(SegmentInfos.java:301)
~[lucene-core-4.3.0.jar:4.3.0 1477023 - simonw - 2013-04-29 14:55:14]
at
org.apache.lucene.index.StandardDirectoryReader$1.doBody(StandardDirectoryReader.java:56)
~[lucene-core-4.3.0.jar:4.3.0 1477023 - simonw - 2013-04-29 14:55:14]
at
org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:783)
~[lucene-core-4.3.0.jar:4.3.0 1477023 - simonw - 2013-04-29 14:55:14]
at
org.apache.lucene.index.StandardDirectoryReader.open(StandardDirectoryReader.java:52)
~[lucene-core-4.3.0.jar:4.3.0 1477023 - simonw - 2013-04-29 14:55:14]
at org.apache.lucene.index.DirectoryReader.open(DirectoryReader.java:66)
~[lucene-core-4.3.0.jar:4.3.0 1477023 - simonw - 2013-04-29 14:55:14]
at xxx.getIndexReader(FTSearchServiceImpl.java:332) 
[xx-0.0.1.jar:16.2.2]
at xxx.getTextSearch(FTSearchServiceImpl.java:115) [xx-0.0.1.jar:16.2.2]
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
~[na:1.7.0_25]



--
View this message in context: 
http://lucene.472066.n3.nabble.com/org-apache-lucene-index-CorruptIndexException-codec-header-mismatch-tp4324171.html
Sent from the Lucene - Java Developer mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10247) Support non-numeric metrics

2017-03-09 Thread Andrzej Bialecki (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrzej Bialecki  updated SOLR-10247:
-
Attachment: compactFormat.png

> Support non-numeric metrics
> ---
>
> Key: SOLR-10247
> URL: https://issues.apache.org/jira/browse/SOLR-10247
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
> Fix For: 6.5, master (7.0)
>
> Attachments: compactFormat.png, SOLR-10247.patch
>
>
> Metrics API currently supports only numeric values. However, it's useful also 
> to report non-numeric values such as eg. version, disk type, component state, 
> some system properties, etc.
> Codahale {{Gauge}} metric type can be used for this purpose, and 
> convenience methods can be added to {{SolrMetricManager}} to make it easier 
> to use.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10247) Support non-numeric metrics

2017-03-09 Thread Andrzej Bialecki (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrzej Bialecki  updated SOLR-10247:
-
Attachment: currentFormat.png

> Support non-numeric metrics
> ---
>
> Key: SOLR-10247
> URL: https://issues.apache.org/jira/browse/SOLR-10247
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
> Fix For: 6.5, master (7.0)
>
> Attachments: compactFormat.png, currentFormat.png, SOLR-10247.patch
>
>
> Metrics API currently supports only numeric values. However, it's useful also 
> to report non-numeric values such as eg. version, disk type, component state, 
> some system properties, etc.
> Codahale {{Gauge}} metric type can be used for this purpose, and 
> convenience methods can be added to {{SolrMetricManager}} to make it easier 
> to use.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-10247) Support non-numeric metrics

2017-03-09 Thread Andrzej Bialecki (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-10247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrzej Bialecki  updated SOLR-10247:
-
Attachment: SOLR-10247.patch

Initial patch. Some classes have been converted to use the new style for Gauge 
registration eg. {{SolrIndexWriter}}. See {{SolrCore.initializeMetrics}} for an 
example of non-numeric registrations.

Also, I added the ability to produce a much more compact NamedList format in 
{{MetricUtils}}. We should consider using it by default, although it's not 
back-compatible with the format in 6x.

> Support non-numeric metrics
> ---
>
> Key: SOLR-10247
> URL: https://issues.apache.org/jira/browse/SOLR-10247
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
> Fix For: 6.5, master (7.0)
>
> Attachments: SOLR-10247.patch
>
>
> Metrics API currently supports only numeric values. However, it's useful also 
> to report non-numeric values such as eg. version, disk type, component state, 
> some system properties, etc.
> Codahale {{Gauge}} metric type can be used for this purpose, and 
> convenience methods can be added to {{SolrMetricManager}} to make it easier 
> to use.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS-EA] Lucene-Solr-6.x-Linux (32bit/jdk-9-ea+159) - Build # 3022 - Unstable!

2017-03-09 Thread Policeman Jenkins Server
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-6.x-Linux/3022/
Java: 32bit/jdk-9-ea+159 -client -XX:+UseParallelGC

1 tests failed.
FAILED:  org.apache.solr.handler.TestSQLHandler.doTest

Error Message:
--> http://127.0.0.1:41686/pk_/sx/collection1:Failed to execute sqlQuery 
'select str_s, count(*), sum(field_i), min(field_i), max(field_i), cast(avg(1.0 
* field_i) as float) from collection1 where text='' group by str_s order by 
sum(field_i) asc limit 2' against JDBC connection 'jdbc:calcitesolr:'. Error 
while executing SQL "select str_s, count(*), sum(field_i), min(field_i), 
max(field_i), cast(avg(1.0 * field_i) as float) from collection1 where 
text='' group by str_s order by sum(field_i) asc limit 2": From line 1, 
column 39 to line 1, column 50: No match found for function signature 
min()

Stack Trace:
java.io.IOException: --> http://127.0.0.1:41686/pk_/sx/collection1:Failed to 
execute sqlQuery 'select str_s, count(*), sum(field_i), min(field_i), 
max(field_i), cast(avg(1.0 * field_i) as float) from collection1 where 
text='' group by str_s order by sum(field_i) asc limit 2' against JDBC 
connection 'jdbc:calcitesolr:'.
Error while executing SQL "select str_s, count(*), sum(field_i), min(field_i), 
max(field_i), cast(avg(1.0 * field_i) as float) from collection1 where 
text='' group by str_s order by sum(field_i) asc limit 2": From line 1, 
column 39 to line 1, column 50: No match found for function signature 
min()
at 
__randomizedtesting.SeedInfo.seed([7D628A7DB31B3E9A:DA2632D9DEA02D23]:0)
at 
org.apache.solr.client.solrj.io.stream.SolrStream.read(SolrStream.java:235)
at 
org.apache.solr.handler.TestSQLHandler.getTuples(TestSQLHandler.java:2337)
at 
org.apache.solr.handler.TestSQLHandler.testBasicGrouping(TestSQLHandler.java:663)
at org.apache.solr.handler.TestSQLHandler.doTest(TestSQLHandler.java:81)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:547)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1713)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:907)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:943)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:957)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:992)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:967)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:916)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:802)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:852)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 

[jira] [Comment Edited] (SOLR-10247) Support non-numeric metrics

2017-03-09 Thread Andrzej Bialecki (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-10247?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15903064#comment-15903064
 ] 

Andrzej Bialecki  edited comment on SOLR-10247 at 3/9/17 1:48 PM:
--

Initial patch. Some classes have been converted to use the new style for Gauge 
registration eg. {{SolrIndexWriter}}. See {{SolrCore.initializeMetrics}} for an 
example of non-numeric registrations.

Also, I added the ability to produce a much more compact NamedList format in 
{{MetricUtils}}, which is used in the output from {{/admin/metrics}}. We should 
consider using it by default, although it's not back-compatible with the format 
in 6x.


was (Author: ab):
Initial patch. Some classes have been converted to use the new style for Gauge 
registration eg. {{SolrIndexWriter}}. See {{SolrCore.initializeMetrics}} for an 
example of non-numeric registrations.

Also, I added the ability to produce a much more compact NamedList format in 
{{MetricUtils}}. We should consider using it by default, although it's not 
back-compatible with the format in 6x.

> Support non-numeric metrics
> ---
>
> Key: SOLR-10247
> URL: https://issues.apache.org/jira/browse/SOLR-10247
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: metrics
>Reporter: Andrzej Bialecki 
>Assignee: Andrzej Bialecki 
> Fix For: 6.5, master (7.0)
>
> Attachments: compactFormat.png, SOLR-10247.patch
>
>
> Metrics API currently supports only numeric values. However, it's useful also 
> to report non-numeric values such as eg. version, disk type, component state, 
> some system properties, etc.
> Codahale {{Gauge}} metric type can be used for this purpose, and 
> convenience methods can be added to {{SolrMetricManager}} to make it easier 
> to use.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[PATCH] jcc: add python3 support

2017-03-09 Thread Ruediger Meier
Hi,

I did some work to port jcc to python3, see
https://github.com/rudimeier/jcc

There are two interesting branches, py2 and py3
  py2  should still work for python2 >=2.7 without any behavior change
  py3  completes experimental python3 support (but still python2
   incompatible).


regarding py2 branch:
  - fixes many (but not yet all) python3 incompatibilities
  - should still work for python2 >=2.7 without any behavior change
  - succeeds lucene's test-suite completely (pylucene-4.10.1/java-1.7
and (pylucene-6.4.1/java-1.8).
  - removes compatibility for python < 2.7 (though it would not be hard
to keep it)
  

regarding py3 branch:
  - based on py2 it adds a few more patches for python3 support (>3.1)
but still in a way which is python2 incompatible
  - still fails about 25% of pylucene tests.
  - Note I just did pylucene's python3 support for testing trivially
like '2to3  -w  $(find -name "*.py")'


Please comment on this. I'd like to get python3 support upstream. 

Some more notes:

My patches were inspired from the old python-3.1 port
http://svn.apache.org/repos/asf/lucene/pylucene/branches/python_3/jcc/

I've refactored/rebased it and splitted the huge patches into many 
smaller ones inclusive keeping python2 support. Almost all commits in 
my py2 branch are trivially to review and independent of each other. So 
I would be glad if you would merge as many of them as you like.

Since the py3 branch is still not 100% correct (guess still some 
Bytes/Unicode problems) I would be glad if somebody would help to get 
it running.


Cheers,
Rudi