[jira] [Updated] (SOLR-6892) Make update processors toplevel components

2014-12-28 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul updated SOLR-6892:
-
Description: 
The current update processor chain is rather cumbersome and we should be able 
to use the updateprocessors without a chain.

The scope of this ticket is 
* A new tag updateProcessor  becomes a toplevel tag and it will be equivalent 
to the {{processor}} tag inside {{updateRequestProcessorChain}} . The only 
difference is that it should require a {{name}} attribute. The 
{{updateProcessorChain}} tag will continue to exist and it should be possible 
to define processor inside as well . It should also be possible to reference 
a named URP in a chain.
* Any update request will be able  to pass a param {{processor=a,b,c}} , where 
a,b,c are names of update processors. A just in time chain will be created with 
those URPs
* Some in built update processors (wherever possible) will be predefined with 
standard names and can be directly used in requests 
* What happens when I say processor=a,b,c in a request? It will execute the 
default chain after the just-in-time chain {{a-b-c}} . 
* How to execute a different chain other than the default chain? the same old 
mechanism of update.chain=x means that the chain  {{x}} will be applied after 
{{a,b,c}}
* How to avoid the default processor chain from being executed ? There will be 
an implicit URP called {{STOP}} . send your request as processor=a,b,c,STOP. 

  was:
The current update processor chain is rather cumbersome and we should be able 
to use the updateprocessors without a chain.

The scope of this ticket is 
* updateProcessor tag becomes a toplevel tag and it will be equivalent to the 
processor tag inside updateRequestProcessorChain . The only difference is 
that it should require a {{name}} attribute
* Any update request will be able  to pass a param {{processor=a,b,c}} , where 
a,b,c are names of update processors. A just in time chain will be created with 
those update processors
* Some in built update processors (wherever possible) will be predefined with 
standard names and can be directly used in requests 


 Make update processors toplevel components 
 ---

 Key: SOLR-6892
 URL: https://issues.apache.org/jira/browse/SOLR-6892
 Project: Solr
  Issue Type: Bug
Reporter: Noble Paul
Assignee: Noble Paul

 The current update processor chain is rather cumbersome and we should be able 
 to use the updateprocessors without a chain.
 The scope of this ticket is 
 * A new tag updateProcessor  becomes a toplevel tag and it will be 
 equivalent to the {{processor}} tag inside 
 {{updateRequestProcessorChain}} . The only difference is that it should 
 require a {{name}} attribute. The {{updateProcessorChain}} tag will 
 continue to exist and it should be possible to define processor inside as 
 well . It should also be possible to reference a named URP in a chain.
 * Any update request will be able  to pass a param {{processor=a,b,c}} , 
 where a,b,c are names of update processors. A just in time chain will be 
 created with those URPs
 * Some in built update processors (wherever possible) will be predefined with 
 standard names and can be directly used in requests 
 * What happens when I say processor=a,b,c in a request? It will execute the 
 default chain after the just-in-time chain {{a-b-c}} . 
 * How to execute a different chain other than the default chain? the same old 
 mechanism of update.chain=x means that the chain  {{x}} will be applied after 
 {{a,b,c}}
 * How to avoid the default processor chain from being executed ? There will 
 be an implicit URP called {{STOP}} . send your request as 
 processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6892) Make it possible to define update request processors as toplevel components

2014-12-28 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul updated SOLR-6892:
-
Summary: Make it possible to define update request processors as toplevel 
components   (was: Make update processors toplevel components )

 Make it possible to define update request processors as toplevel components 
 

 Key: SOLR-6892
 URL: https://issues.apache.org/jira/browse/SOLR-6892
 Project: Solr
  Issue Type: Bug
Reporter: Noble Paul
Assignee: Noble Paul

 The current update processor chain is rather cumbersome and we should be able 
 to use the updateprocessors without a chain.
 The scope of this ticket is 
 * A new tag updateProcessor  becomes a toplevel tag and it will be 
 equivalent to the {{processor}} tag inside 
 {{updateRequestProcessorChain}} . The only difference is that it should 
 require a {{name}} attribute. The {{updateProcessorChain}} tag will 
 continue to exist and it should be possible to define processor inside as 
 well . It should also be possible to reference a named URP in a chain.
 * Any update request will be able  to pass a param {{processor=a,b,c}} , 
 where a,b,c are names of update processors. A just in time chain will be 
 created with those URPs
 * Some in built update processors (wherever possible) will be predefined with 
 standard names and can be directly used in requests 
 * What happens when I say processor=a,b,c in a request? It will execute the 
 default chain after the just-in-time chain {{a-b-c}} . 
 * How to execute a different chain other than the default chain? the same old 
 mechanism of update.chain=x means that the chain  {{x}} will be applied after 
 {{a,b,c}}
 * How to avoid the default processor chain from being executed ? There will 
 be an implicit URP called {{STOP}} . send your request as 
 processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6892) Make update processors toplevel components

2014-12-28 Thread Noble Paul (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259600#comment-14259600
 ] 

Noble Paul commented on SOLR-6892:
--

I've updated the description with more meat this time. Please comment

 Make update processors toplevel components 
 ---

 Key: SOLR-6892
 URL: https://issues.apache.org/jira/browse/SOLR-6892
 Project: Solr
  Issue Type: Bug
Reporter: Noble Paul
Assignee: Noble Paul

 The current update processor chain is rather cumbersome and we should be able 
 to use the updateprocessors without a chain.
 The scope of this ticket is 
 * A new tag updateProcessor  becomes a toplevel tag and it will be 
 equivalent to the {{processor}} tag inside 
 {{updateRequestProcessorChain}} . The only difference is that it should 
 require a {{name}} attribute. The {{updateProcessorChain}} tag will 
 continue to exist and it should be possible to define processor inside as 
 well . It should also be possible to reference a named URP in a chain.
 * Any update request will be able  to pass a param {{processor=a,b,c}} , 
 where a,b,c are names of update processors. A just in time chain will be 
 created with those URPs
 * Some in built update processors (wherever possible) will be predefined with 
 standard names and can be directly used in requests 
 * What happens when I say processor=a,b,c in a request? It will execute the 
 default chain after the just-in-time chain {{a-b-c}} . 
 * How to execute a different chain other than the default chain? the same old 
 mechanism of update.chain=x means that the chain  {{x}} will be applied after 
 {{a,b,c}}
 * How to avoid the default processor chain from being executed ? There will 
 be an implicit URP called {{STOP}} . send your request as 
 processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6879) Have an option to disable autoAddReplicas temporarily for all collections

2014-12-28 Thread Varun Thacker (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6879?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259603#comment-14259603
 ] 

Varun Thacker commented on SOLR-6879:
-

Thanks Steve.

Couple of ref guide updates that we would need to make - 

1. 
https://cwiki.apache.org/confluence/display/solr/Collections+API#CollectionsAPI-api11
Instead of As of Solr 4.7, only the property urlScheme is supported.
we could have something like - The two supported properties are - 'urlScheme' 
and 'autoAddReplicas' 

2. Under AutoAddReplica Settings on 
https://cwiki.apache.org/confluence/display/solr/Running+Solr+on+HDFS we could 
add the following section- 


Temporarily disable autoAddReplicas for the entire cluster - When doing 
offline maintenance on the cluster and for various other use cases where an 
admin would like to disable autoAddReplicas temporarily, one could use the 
following APIs to disable and enable autoAddReplicas cluster wide.

API to disable autoAddReplicas:

{code}
admin/collections?action=CLUSTERPROPname=autoAddReplicasval=false
{code}

API to enable autoAddReplicas:

{code}
admin/collections?action=CLUSTERPROPname=autoAddReplicas
{code}

 Have an option to disable autoAddReplicas temporarily for all collections
 -

 Key: SOLR-6879
 URL: https://issues.apache.org/jira/browse/SOLR-6879
 Project: Solr
  Issue Type: Improvement
Reporter: Varun Thacker
Assignee: Steve Rowe
 Fix For: 5.0, Trunk

 Attachments: SOLR-6879.patch, SOLR-6879.patch


 When doing offline maintenance on my cluster I would like to disable 
 autoAddReplicas since I don't want replicas being created on nodes during 
 this time.
 It would be useful to have an option to enable/disable autoAddReplicas via an 
 API.
 This API would disable autoAddReplicas:
 {code}
 admin/collections?action=CLUSTERPROPname=autoAddReplicasval=false
 {code}
 Now when I want to enable it again the API call would look like this:
 {code}
 admin/collections?action=CLUSTERPROPname=autoAddReplicas
 {code}
 This uses the CLUSTERPROP semantics of unsetting a property when the {{val}} 
 param is not provided.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6135) re-number fields randomly in tests

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259611#comment-14259611
 ] 

ASF subversion and git services commented on LUCENE-6135:
-

Commit 1648188 from [~rcmuir] in branch 'dev/trunk'
[ https://svn.apache.org/r1648188 ]

LUCENE-6135: renumber fields randomly in tests

 re-number fields randomly in tests
 --

 Key: LUCENE-6135
 URL: https://issues.apache.org/jira/browse/LUCENE-6135
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-6135.patch


 Currently some code (such as stored fields bulk merge) depends upon 
 consistent field number ordering. 
 In the case field numbers do not align, then optimizations are disabled, 
 because the would cause crazy corruption where values are mixed up across 
 different fields. 
 But this is hardly tested today. If i introduce an intentional bug into this 
 logic, then only one lone test fails: TestAddIndexes.testFieldNamesChanged, 
 and only about 10% of the time at best. In general tests pass.
 {code}
 --- 
 lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
 (revision 1647793)
 +++ 
 lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
 (working copy)
 @@ -52,6 +52,10 @@
  for (int i = 0; i  numReaders; i++) {
for (FieldInfo fi : mergeState.fieldInfos[i]) {
  FieldInfo other = mergeState.mergeFieldInfos.fieldInfo(fi.number);
 +// nocommit:
 +if (true) {
 +  break;
 +}
  if (other == null || !other.name.equals(fi.name)) {
continue nextReader;
  }
 {code}
 Don't get me wrong, its a great simple test, but it should not be the only 
 one doing this. And it would be great if it failed more often, i havent 
 looked as to why it only fails rarely if there is a bug in this stuff.
 But in general, we should simulate this more. My current idea is to shuffle 
 up field numbers in MockRandomMergePolicy. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6135) re-number fields randomly in tests

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259621#comment-14259621
 ] 

ASF subversion and git services commented on LUCENE-6135:
-

Commit 1648189 from [~rcmuir] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1648189 ]

LUCENE-6135: renumber fields randomly in tests

 re-number fields randomly in tests
 --

 Key: LUCENE-6135
 URL: https://issues.apache.org/jira/browse/LUCENE-6135
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Attachments: LUCENE-6135.patch


 Currently some code (such as stored fields bulk merge) depends upon 
 consistent field number ordering. 
 In the case field numbers do not align, then optimizations are disabled, 
 because the would cause crazy corruption where values are mixed up across 
 different fields. 
 But this is hardly tested today. If i introduce an intentional bug into this 
 logic, then only one lone test fails: TestAddIndexes.testFieldNamesChanged, 
 and only about 10% of the time at best. In general tests pass.
 {code}
 --- 
 lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
 (revision 1647793)
 +++ 
 lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
 (working copy)
 @@ -52,6 +52,10 @@
  for (int i = 0; i  numReaders; i++) {
for (FieldInfo fi : mergeState.fieldInfos[i]) {
  FieldInfo other = mergeState.mergeFieldInfos.fieldInfo(fi.number);
 +// nocommit:
 +if (true) {
 +  break;
 +}
  if (other == null || !other.name.equals(fi.name)) {
continue nextReader;
  }
 {code}
 Don't get me wrong, its a great simple test, but it should not be the only 
 one doing this. And it would be great if it failed more often, i havent 
 looked as to why it only fails rarely if there is a bug in this stuff.
 But in general, we should simulate this more. My current idea is to shuffle 
 up field numbers in MockRandomMergePolicy. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-6135) re-number fields randomly in tests

2014-12-28 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-6135.
-
   Resolution: Fixed
Fix Version/s: Trunk
   5.0

 re-number fields randomly in tests
 --

 Key: LUCENE-6135
 URL: https://issues.apache.org/jira/browse/LUCENE-6135
 Project: Lucene - Core
  Issue Type: Test
Reporter: Robert Muir
 Fix For: 5.0, Trunk

 Attachments: LUCENE-6135.patch


 Currently some code (such as stored fields bulk merge) depends upon 
 consistent field number ordering. 
 In the case field numbers do not align, then optimizations are disabled, 
 because the would cause crazy corruption where values are mixed up across 
 different fields. 
 But this is hardly tested today. If i introduce an intentional bug into this 
 logic, then only one lone test fails: TestAddIndexes.testFieldNamesChanged, 
 and only about 10% of the time at best. In general tests pass.
 {code}
 --- 
 lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
 (revision 1647793)
 +++ 
 lucene/core/src/java/org/apache/lucene/codecs/compressing/MatchingReaders.java
 (working copy)
 @@ -52,6 +52,10 @@
  for (int i = 0; i  numReaders; i++) {
for (FieldInfo fi : mergeState.fieldInfos[i]) {
  FieldInfo other = mergeState.mergeFieldInfos.fieldInfo(fi.number);
 +// nocommit:
 +if (true) {
 +  break;
 +}
  if (other == null || !other.name.equals(fi.name)) {
continue nextReader;
  }
 {code}
 Don't get me wrong, its a great simple test, but it should not be the only 
 one doing this. And it would be great if it failed more often, i havent 
 looked as to why it only fails rarely if there is a bug in this stuff.
 But in general, we should simulate this more. My current idea is to shuffle 
 up field numbers in MockRandomMergePolicy. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6494) Query filters applied in a wrong order

2014-12-28 Thread Alexander S. (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6494?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259627#comment-14259627
 ] 

Alexander S. commented on SOLR-6494:


As I was told already, Solr does not apply filters incrementally, instead each 
filter runs through the entire data set, then Solr caches the results. In the 
case with filters that contain ranges cache is not effective, especially when 
we need NRT search and commits being triggered multiple times per minute. Then 
big caches make no sense and big autowarming numbers causing Solr to fail. My 
point is that cache is not always efficient and for such cases Solr need to use 
another strategy and apply filters incrementally (read as post filters).

So this:
{quote}
By design, fq clauses like this are calculated for the entire document set and 
the results cached, there is no ordering for that part. Otherwise, how could 
they be re-used for a different query?
{quote}
does not work in all cases.

Something like this:
{code}
fq={!cache=false cost=101}field:value # to run as a post filter
{code}
would definitely solve the problem, but this is not supported.

The frange parser has support for this, but it is not always suitable and fails 
with different errors, like can not use FieldCache on multivalued field: 
type, etc.

Does that look like a missing feature? I mean for me it definitely does, but 
could this be considered as a wish and implemented some day? How can Solr 
community help with missing features?

 Query filters applied in a wrong order
 --

 Key: SOLR-6494
 URL: https://issues.apache.org/jira/browse/SOLR-6494
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.8.1
Reporter: Alexander S.

 This query:
 {code}
 {
   fq: [type:Award::Nomination],
   sort: score desc,
   start: 0,
   rows: 20,
   q: *:*
 }
 {code}
 takes just a few milliseconds, but this one:
 {code}
 {
   fq: [
 type:Award::Nomination,
 created_at_d:[* TO 2014-09-08T23:59:59Z]
   ],
   sort: score desc,
   start: 0,
   rows: 20,
   q: *:*
 }
 {code}
 takes almost 15 seconds.
 I have just ≈12k of documents with type Award::Nomination, but around half 
 a billion with created_at_d field set. And it seems Solr applies the 
 created_at_d filter first going through all documents where this field is 
 set, which is not very smart.
 I think if it can't do anything better than applying filters in the alphabet 
 order it should apply them in the order they were received.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-6494) Query filters applied in a wrong order

2014-12-28 Thread Alexander S. (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6494?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259627#comment-14259627
 ] 

Alexander S. edited comment on SOLR-6494 at 12/28/14 12:50 PM:
---

As I was told already, Solr does not apply filters incrementally, instead each 
filter runs through the entire data set, then Solr caches the results. In the 
case with filters that contain ranges cache is not effective, especially when 
we need NRT search and commits being triggered multiple times per minute. Then 
big caches make no sense and big autowarming numbers causing Solr to fail. My 
point is that cache is not always efficient and for such cases Solr need to use 
another strategy and apply filters incrementally (read as post filters).

So this:
{quote}
By design, fq clauses like this are calculated for the entire document set and 
the results cached, there is no ordering for that part. Otherwise, how could 
they be re-used for a different query?
{quote}
does not work in all cases.

Something like this:
{code}
# cost  100 to run as a post filter, but something like post=true would be 
better I think
fq={!cache=false cost=101}field:value
{code}
would definitely solve the problem, but this is not supported.

The frange parser has support for this, but it is not always suitable and fails 
with different errors, like can not use FieldCache on multivalued field: 
type, etc.

Does that look like a missing feature? I mean for me it definitely does, but 
could this be considered as a wish and implemented some day? How can Solr 
community help with missing features?


was (Author: aheaven):
As I was told already, Solr does not apply filters incrementally, instead each 
filter runs through the entire data set, then Solr caches the results. In the 
case with filters that contain ranges cache is not effective, especially when 
we need NRT search and commits being triggered multiple times per minute. Then 
big caches make no sense and big autowarming numbers causing Solr to fail. My 
point is that cache is not always efficient and for such cases Solr need to use 
another strategy and apply filters incrementally (read as post filters).

So this:
{quote}
By design, fq clauses like this are calculated for the entire document set and 
the results cached, there is no ordering for that part. Otherwise, how could 
they be re-used for a different query?
{quote}
does not work in all cases.

Something like this:
{code}
fq={!cache=false cost=101}field:value # to run as a post filter
{code}
would definitely solve the problem, but this is not supported.

The frange parser has support for this, but it is not always suitable and fails 
with different errors, like can not use FieldCache on multivalued field: 
type, etc.

Does that look like a missing feature? I mean for me it definitely does, but 
could this be considered as a wish and implemented some day? How can Solr 
community help with missing features?

 Query filters applied in a wrong order
 --

 Key: SOLR-6494
 URL: https://issues.apache.org/jira/browse/SOLR-6494
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.8.1
Reporter: Alexander S.

 This query:
 {code}
 {
   fq: [type:Award::Nomination],
   sort: score desc,
   start: 0,
   rows: 20,
   q: *:*
 }
 {code}
 takes just a few milliseconds, but this one:
 {code}
 {
   fq: [
 type:Award::Nomination,
 created_at_d:[* TO 2014-09-08T23:59:59Z]
   ],
   sort: score desc,
   start: 0,
   rows: 20,
   q: *:*
 }
 {code}
 takes almost 15 seconds.
 I have just ≈12k of documents with type Award::Nomination, but around half 
 a billion with created_at_d field set. And it seems Solr applies the 
 created_at_d filter first going through all documents where this field is 
 set, which is not very smart.
 I think if it can't do anything better than applying filters in the alphabet 
 order it should apply them in the order they were received.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5507) Admin UI - Refactoring using AngularJS

2014-12-28 Thread Upayavira (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259635#comment-14259635
 ] 

Upayavira commented on SOLR-5507:
-

[~erickerickson] please feel free to make a dependent issue for your SolrCloud 
enhancement, it is a good idea that shouldn't be lost.

 Admin UI - Refactoring using AngularJS
 --

 Key: SOLR-5507
 URL: https://issues.apache.org/jira/browse/SOLR-5507
 Project: Solr
  Issue Type: Improvement
  Components: web gui
Reporter: Stefan Matheis (steffkes)
Assignee: Stefan Matheis (steffkes)
Priority: Minor
 Attachments: SOLR-5507.patch


 On the LSR in Dublin, i've talked again to [~upayavira] and this time we 
 talked about Refactoring the existing UI - using AngularJS: providing (more, 
 internal) structure and what not ;
 He already started working on the Refactoring, so this is more a 'tracking' 
 issue about the progress he/we do there.
 Will extend this issue with a bit more context  additional information, w/ 
 thoughts about the possible integration in the existing UI and more (:



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6138) ItalianLightStemmer doesn't apply on words shorter then 6 chars in length

2014-12-28 Thread Massimo Pasquini (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259655#comment-14259655
 ] 

Massimo Pasquini commented on LUCENE-6138:
--

The issue you pointed out is related to a different stemmer for Russian 
language. I see no connection to the Italian light stemmer. According to the 
rules of the Italian grammar, I think the bug I found can be fixed (it possibly 
cannot be done in the Russian stemmer according to what I've read on the other 
post).

So I suppose the ItalianLightStemmer can evolve to a better implementation: it 
is possible to find some simple rules on words suffixes in order to make a 
decision about applying the stemming on short words (shorter then 6 characters).

Notice my thoughts are not based on a deep and accurate study of the problem 
though. But I think it could be worth to investigate about it. I may suggest to 
submit this issue to the author of the code and see if he got a better solution 
(I saw he's in the field of computational linguistics). According to the notes 
in the source, the algorithm was written in 2005 as the result of some 
experiments. We don't know if they've put further efforts in investigating the 
problem and they possibly wrote a best algorithm they agree to publish 
according to Lucene's license.

I don't expect the stemmer to be 100% successful, but the issue I pointed out 
affects an important range on words.

 ItalianLightStemmer doesn't apply on words shorter then 6 chars in length
 -

 Key: LUCENE-6138
 URL: https://issues.apache.org/jira/browse/LUCENE-6138
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/analysis
Affects Versions: 4.10.2
Reporter: Massimo Pasquini
Priority: Minor

 I expect a stemmer to transform nouns in their singular and plural forms into 
 a shorter common form. The implementation of the ItalianLightStemmer doesn't 
 apply any stemming to words shorter then 6 characters in length. This leads 
 to some annoying results:
 singular form | plural form
 4|5 chars in length (no stemming)
 alga - alga | alghe - alghe
 fuga - fuga | fughe - fughe
 lega - lega | leghe - leghe
 5|6 chars in length (stemming only on plural form)
 vanga - vanga | vanghe - vang
 verga - verga | verghe - verg
 I suppose that such limitation on words length is to avoid other side effects 
 on shorter words not in the set above, but I think something must be reviewed 
 in the code for better results.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5507) Admin UI - Refactoring using AngularJS

2014-12-28 Thread Jack Krupansky (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259657#comment-14259657
 ] 

Jack Krupansky commented on SOLR-5507:
--

bq.  All I ask, though, is that you forgive the occasional burst of ebullient 
enthusiasm!

No need for it to be forgiven... all ebullient enthusiasm is always welcome and 
encouraged.


 Admin UI - Refactoring using AngularJS
 --

 Key: SOLR-5507
 URL: https://issues.apache.org/jira/browse/SOLR-5507
 Project: Solr
  Issue Type: Improvement
  Components: web gui
Reporter: Stefan Matheis (steffkes)
Assignee: Stefan Matheis (steffkes)
Priority: Minor
 Attachments: SOLR-5507.patch


 On the LSR in Dublin, i've talked again to [~upayavira] and this time we 
 talked about Refactoring the existing UI - using AngularJS: providing (more, 
 internal) structure and what not ;
 He already started working on the Refactoring, so this is more a 'tracking' 
 issue about the progress he/we do there.
 Will extend this issue with a bit more context  additional information, w/ 
 thoughts about the possible integration in the existing UI and more (:



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6141) simplify stored fields bulk merge logic

2014-12-28 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259660#comment-14259660
 ] 

Robert Muir commented on LUCENE-6141:
-

We already have a test: testMixedCompressions. I will improve it with the 
commit though.

 simplify stored fields bulk merge logic
 ---

 Key: LUCENE-6141
 URL: https://issues.apache.org/jira/browse/LUCENE-6141
 Project: Lucene - Core
  Issue Type: Task
Reporter: Robert Muir
 Attachments: LUCENE-6141.patch


 The current logic checks that the same chunksize and compression algorithm 
 were used, but this is obselete as it no longer iterates chunks.
 We only need to check that the format version is the same.
 This also allows for bulk merging across compression algorithms. A good use 
 case is BEST_SPEED - BEST_COMPRESSION for archiving.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6141) simplify stored fields bulk merge logic

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259664#comment-14259664
 ] 

ASF subversion and git services commented on LUCENE-6141:
-

Commit 1648221 from [~rcmuir] in branch 'dev/trunk'
[ https://svn.apache.org/r1648221 ]

LUCENE-6141: simplify stored fields bulk merge logic

 simplify stored fields bulk merge logic
 ---

 Key: LUCENE-6141
 URL: https://issues.apache.org/jira/browse/LUCENE-6141
 Project: Lucene - Core
  Issue Type: Task
Reporter: Robert Muir
 Attachments: LUCENE-6141.patch


 The current logic checks that the same chunksize and compression algorithm 
 were used, but this is obselete as it no longer iterates chunks.
 We only need to check that the format version is the same.
 This also allows for bulk merging across compression algorithms. A good use 
 case is BEST_SPEED - BEST_COMPRESSION for archiving.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6892) Make it possible to define update request processors as toplevel components

2014-12-28 Thread Jack Krupansky (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259665#comment-14259665
 ] 

Jack Krupansky commented on SOLR-6892:
--

Thanks for the description updates. Comments...

1. We need to be explicit about how and when the hard-wired processors are 
invoked. In particular the run update processor. The log update processor 
is somewhat special in that it is not mandatory, but a lot of people are not 
explicitly aware of it, so if they leave it out, they will be wondering why 
they don't get logging of updates.

2. I suggest three parameters: pre.processors to specify processors before 
the default chain, post.processors to specify processors after the default 
chain (before or after run update and log update??), and processors to 
specify a processor list to completely replace the default chain.

3. Make log update be automatically added at the end unless a nolog 
processor is specified.

4. Make run update be automatically added at the end unless a norun 
processor is specified.

5. Discuss processor vs. processors - I prefer the latter since it is 
explicit, but maybe allow both since the singular/plural can be confusing.

6. Consider supporting both a single parameter with a csv list as well as 
multiple parameters each with a single value. I prefer having the choice. 
Having a separate parameter for each processor can be more explicit sometimes.

7. Consider a single-processor parameter with the option to specify the 
parameters for that processor. That would make it possible to invoke the 
various field mutating update processors, which would be especially cool and 
convenient.


 Make it possible to define update request processors as toplevel components 
 

 Key: SOLR-6892
 URL: https://issues.apache.org/jira/browse/SOLR-6892
 Project: Solr
  Issue Type: Bug
Reporter: Noble Paul
Assignee: Noble Paul

 The current update processor chain is rather cumbersome and we should be able 
 to use the updateprocessors without a chain.
 The scope of this ticket is 
 * A new tag updateProcessor  becomes a toplevel tag and it will be 
 equivalent to the {{processor}} tag inside 
 {{updateRequestProcessorChain}} . The only difference is that it should 
 require a {{name}} attribute. The {{updateProcessorChain}} tag will 
 continue to exist and it should be possible to define processor inside as 
 well . It should also be possible to reference a named URP in a chain.
 * Any update request will be able  to pass a param {{processor=a,b,c}} , 
 where a,b,c are names of update processors. A just in time chain will be 
 created with those URPs
 * Some in built update processors (wherever possible) will be predefined with 
 standard names and can be directly used in requests 
 * What happens when I say processor=a,b,c in a request? It will execute the 
 default chain after the just-in-time chain {{a-b-c}} . 
 * How to execute a different chain other than the default chain? the same old 
 mechanism of update.chain=x means that the chain  {{x}} will be applied after 
 {{a,b,c}}
 * How to avoid the default processor chain from being executed ? There will 
 be an implicit URP called {{STOP}} . send your request as 
 processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6141) simplify stored fields bulk merge logic

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259668#comment-14259668
 ] 

ASF subversion and git services commented on LUCENE-6141:
-

Commit 1648222 from [~rcmuir] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1648222 ]

LUCENE-6141: simplify stored fields bulk merge logic

 simplify stored fields bulk merge logic
 ---

 Key: LUCENE-6141
 URL: https://issues.apache.org/jira/browse/LUCENE-6141
 Project: Lucene - Core
  Issue Type: Task
Reporter: Robert Muir
 Fix For: 5.0, Trunk

 Attachments: LUCENE-6141.patch


 The current logic checks that the same chunksize and compression algorithm 
 were used, but this is obselete as it no longer iterates chunks.
 We only need to check that the format version is the same.
 This also allows for bulk merging across compression algorithms. A good use 
 case is BEST_SPEED - BEST_COMPRESSION for archiving.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-6141) simplify stored fields bulk merge logic

2014-12-28 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-6141.
-
   Resolution: Fixed
Fix Version/s: Trunk
   5.0

 simplify stored fields bulk merge logic
 ---

 Key: LUCENE-6141
 URL: https://issues.apache.org/jira/browse/LUCENE-6141
 Project: Lucene - Core
  Issue Type: Task
Reporter: Robert Muir
 Fix For: 5.0, Trunk

 Attachments: LUCENE-6141.patch


 The current logic checks that the same chunksize and compression algorithm 
 were used, but this is obselete as it no longer iterates chunks.
 We only need to check that the format version is the same.
 This also allows for bulk merging across compression algorithms. A good use 
 case is BEST_SPEED - BEST_COMPRESSION for archiving.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6892) Make it possible to define update request processors as toplevel components

2014-12-28 Thread Noble Paul (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259670#comment-14259670
 ] 

Noble Paul commented on SOLR-6892:
--

[~jkrupan] 
A lot of your suggestions are not realy in the scope of this ticket and should 
be addressed separately

bq. Consider supporting both a single parameter with a csv list as well as 
multiple parameters each with a single value. I prefer having the choice.

I think having choices is not really good if it doesn't offer anything better.

bq.Consider a single-processor parameter with the option to specify the 
parameters for that processor. 

This is just too complex to handle with the flat http parameter structure. 
Again beyond the scope

URPs are really powerful but they are a bit hard to configure and use. I wish 
to make it easier and more widely used. That is the  objective of this ticket 

 Make it possible to define update request processors as toplevel components 
 

 Key: SOLR-6892
 URL: https://issues.apache.org/jira/browse/SOLR-6892
 Project: Solr
  Issue Type: Bug
Reporter: Noble Paul
Assignee: Noble Paul

 The current update processor chain is rather cumbersome and we should be able 
 to use the updateprocessors without a chain.
 The scope of this ticket is 
 * A new tag updateProcessor  becomes a toplevel tag and it will be 
 equivalent to the {{processor}} tag inside 
 {{updateRequestProcessorChain}} . The only difference is that it should 
 require a {{name}} attribute. The {{updateProcessorChain}} tag will 
 continue to exist and it should be possible to define processor inside as 
 well . It should also be possible to reference a named URP in a chain.
 * Any update request will be able  to pass a param {{processor=a,b,c}} , 
 where a,b,c are names of update processors. A just in time chain will be 
 created with those URPs
 * Some in built update processors (wherever possible) will be predefined with 
 standard names and can be directly used in requests 
 * What happens when I say processor=a,b,c in a request? It will execute the 
 default chain after the just-in-time chain {{a-b-c}} . 
 * How to execute a different chain other than the default chain? the same old 
 mechanism of update.chain=x means that the chain  {{x}} will be applied after 
 {{a,b,c}}
 * How to avoid the default processor chain from being executed ? There will 
 be an implicit URP called {{STOP}} . send your request as 
 processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6140) simplify inflater usage in deflate CompressionMode

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6140?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259671#comment-14259671
 ] 

ASF subversion and git services commented on LUCENE-6140:
-

Commit 1648224 from [~rcmuir] in branch 'dev/trunk'
[ https://svn.apache.org/r1648224 ]

LUCENE-6140: simplify inflater usage in CompressionMode

 simplify inflater usage in deflate CompressionMode
 --

 Key: LUCENE-6140
 URL: https://issues.apache.org/jira/browse/LUCENE-6140
 Project: Lucene - Core
  Issue Type: Task
Reporter: Robert Muir
 Attachments: LUCENE-6140.patch


 This currently loops-n-grows the output byte[]. But we always decompress the 
 whole block (we dont emit flushes or anything to allow otherwise) and ignore 
 offset/length to the end, and we know how big the uncompressed size is 
 up-front... we can just call inflate one time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-6140) simplify inflater usage in deflate CompressionMode

2014-12-28 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6140?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-6140.
-
   Resolution: Fixed
Fix Version/s: Trunk
   5.0

 simplify inflater usage in deflate CompressionMode
 --

 Key: LUCENE-6140
 URL: https://issues.apache.org/jira/browse/LUCENE-6140
 Project: Lucene - Core
  Issue Type: Task
Reporter: Robert Muir
 Fix For: 5.0, Trunk

 Attachments: LUCENE-6140.patch


 This currently loops-n-grows the output byte[]. But we always decompress the 
 whole block (we dont emit flushes or anything to allow otherwise) and ignore 
 offset/length to the end, and we know how big the uncompressed size is 
 up-front... we can just call inflate one time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6140) simplify inflater usage in deflate CompressionMode

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6140?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259673#comment-14259673
 ] 

ASF subversion and git services commented on LUCENE-6140:
-

Commit 1648226 from [~rcmuir] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1648226 ]

LUCENE-6140: simplify inflater usage in CompressionMode

 simplify inflater usage in deflate CompressionMode
 --

 Key: LUCENE-6140
 URL: https://issues.apache.org/jira/browse/LUCENE-6140
 Project: Lucene - Core
  Issue Type: Task
Reporter: Robert Muir
 Fix For: 5.0, Trunk

 Attachments: LUCENE-6140.patch


 This currently loops-n-grows the output byte[]. But we always decompress the 
 whole block (we dont emit flushes or anything to allow otherwise) and ignore 
 offset/length to the end, and we know how big the uncompressed size is 
 up-front... we can just call inflate one time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6892) Make it possible to define update request processors as toplevel components

2014-12-28 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul updated SOLR-6892:
-
Issue Type: Improvement  (was: Bug)

 Make it possible to define update request processors as toplevel components 
 

 Key: SOLR-6892
 URL: https://issues.apache.org/jira/browse/SOLR-6892
 Project: Solr
  Issue Type: Improvement
Reporter: Noble Paul
Assignee: Noble Paul

 The current update processor chain is rather cumbersome and we should be able 
 to use the updateprocessors without a chain.
 The scope of this ticket is 
 * A new tag updateProcessor  becomes a toplevel tag and it will be 
 equivalent to the {{processor}} tag inside 
 {{updateRequestProcessorChain}} . The only difference is that it should 
 require a {{name}} attribute. The {{updateProcessorChain}} tag will 
 continue to exist and it should be possible to define processor inside as 
 well . It should also be possible to reference a named URP in a chain.
 * Any update request will be able  to pass a param {{processor=a,b,c}} , 
 where a,b,c are names of update processors. A just in time chain will be 
 created with those URPs
 * Some in built update processors (wherever possible) will be predefined with 
 standard names and can be directly used in requests 
 * What happens when I say processor=a,b,c in a request? It will execute the 
 default chain after the just-in-time chain {{a-b-c}} . 
 * How to execute a different chain other than the default chain? the same old 
 mechanism of update.chain=x means that the chain  {{x}} will be applied after 
 {{a,b,c}}
 * How to avoid the default processor chain from being executed ? There will 
 be an implicit URP called {{STOP}} . send your request as 
 processor=a,b,c,STOP. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-6893) DIH doen't work using URL and wget

2014-12-28 Thread Dani (JIRA)
Dani created SOLR-6893:
--

 Summary: DIH doen't work using URL and wget
 Key: SOLR-6893
 URL: https://issues.apache.org/jira/browse/SOLR-6893
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.10.2, 4.8.1, 4.6.1
 Environment: Linux (Ubuntu, CentOS)
Reporter: Dani
Priority: Minor


[ related to #SOLR-2305 ]

I put this URL on browser and import correctly starts:

http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-importentity=fileImportclean=truecommit=true

But if I use wget it doens't work:

wget 
http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-importentity=fileImportclean=truecommit=true;

nor also using escape:

wget 
http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import\entity=fileImport\clean=true\commit=true;



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6138) ItalianLightStemmer doesn't apply on words shorter then 6 chars in length

2014-12-28 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259684#comment-14259684
 ] 

Erick Erickson commented on LUCENE-6138:


Right, the important part of the discussion (I should have pointed it out) was 
that the stemmers are not part of the Solr code base, they're another project 
and that project would be the place to raise possible bugs or submit patches, 

bq: Can you propose your changes to 
http://members.unine.ch/jacques.savoy/clef/index.html?

Sorry for the confusion



 ItalianLightStemmer doesn't apply on words shorter then 6 chars in length
 -

 Key: LUCENE-6138
 URL: https://issues.apache.org/jira/browse/LUCENE-6138
 Project: Lucene - Core
  Issue Type: Bug
  Components: modules/analysis
Affects Versions: 4.10.2
Reporter: Massimo Pasquini
Priority: Minor

 I expect a stemmer to transform nouns in their singular and plural forms into 
 a shorter common form. The implementation of the ItalianLightStemmer doesn't 
 apply any stemming to words shorter then 6 characters in length. This leads 
 to some annoying results:
 singular form | plural form
 4|5 chars in length (no stemming)
 alga - alga | alghe - alghe
 fuga - fuga | fughe - fughe
 lega - lega | leghe - leghe
 5|6 chars in length (stemming only on plural form)
 vanga - vanga | vanghe - vang
 verga - verga | verghe - verg
 I suppose that such limitation on words length is to avoid other side effects 
 on shorter words not in the set above, but I think something must be reviewed 
 in the code for better results.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6893) DIH doen't work using URL and wget

2014-12-28 Thread Dani (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6893?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dani updated SOLR-6893:
---
Description: 
I put this URL on browser and import correctly starts:

http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-importentity=fileImportclean=truecommit=true

But if I use wget it doens't work:

wget 
http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-importentity=fileImportclean=truecommit=true;

nor also using escape:

wget 
http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import\entity=fileImport\clean=true\commit=true;

  was:
[ related to #SOLR-2305 ]

I put this URL on browser and import correctly starts:

http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-importentity=fileImportclean=truecommit=true

But if I use wget it doens't work:

wget 
http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-importentity=fileImportclean=truecommit=true;

nor also using escape:

wget 
http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import\entity=fileImport\clean=true\commit=true;


 DIH doen't work using URL and wget
 --

 Key: SOLR-6893
 URL: https://issues.apache.org/jira/browse/SOLR-6893
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.6.1, 4.8.1, 4.10.2
 Environment: Linux (Ubuntu, CentOS)
Reporter: Dani
Priority: Minor

 I put this URL on browser and import correctly starts:
 http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-importentity=fileImportclean=truecommit=true
 But if I use wget it doens't work:
 wget 
 http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-importentity=fileImportclean=truecommit=true;
 nor also using escape:
 wget 
 http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import\entity=fileImport\clean=true\commit=true;



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5507) Admin UI - Refactoring using AngularJS

2014-12-28 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259692#comment-14259692
 ] 

Erick Erickson commented on SOLR-5507:
--

My desire here is for the really annoying error messages to go away if the 
files are misplaced...

I have no real feel for whether they are useful enough to try to keep around 
anyway, up to the people doing the work of course.

 Admin UI - Refactoring using AngularJS
 --

 Key: SOLR-5507
 URL: https://issues.apache.org/jira/browse/SOLR-5507
 Project: Solr
  Issue Type: Improvement
  Components: web gui
Reporter: Stefan Matheis (steffkes)
Assignee: Stefan Matheis (steffkes)
Priority: Minor
 Attachments: SOLR-5507.patch


 On the LSR in Dublin, i've talked again to [~upayavira] and this time we 
 talked about Refactoring the existing UI - using AngularJS: providing (more, 
 internal) structure and what not ;
 He already started working on the Refactoring, so this is more a 'tracking' 
 issue about the progress he/we do there.
 Will extend this issue with a bit more context  additional information, w/ 
 thoughts about the possible integration in the existing UI and more (:



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6082) Umbrella JIRA for Admin UI and SolrCloud.

2014-12-28 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6082?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259693#comment-14259693
 ] 

Erick Erickson commented on SOLR-6082:
--

[~shalinmangar] Don't know whether this is getting much of your attention, but 
see SOLR-5507. Upayavira is working on moving the admin UI to AngularJS, 
coordinating the two efforts seems like A Good Thing

An open question is how to prioritize this and the work on SOLR-5507. How/when 
should priority be given to reproducing the current admin UI which looks at 
things largely on the basis of individual nodes with SolrCloud bolted on as 
opposed to adding SolrCloud-centric features to the admin UI?

Recording here for posterity, as always the people doing the work are the ones 
who get to choose.

 Umbrella JIRA for Admin UI and SolrCloud.
 -

 Key: SOLR-6082
 URL: https://issues.apache.org/jira/browse/SOLR-6082
 Project: Solr
  Issue Type: Improvement
  Components: web gui
Affects Versions: 4.9, Trunk
Reporter: Erick Erickson
Assignee: Shalin Shekhar Mangar

 It would be very helpful if the admin UI were more cloud friendly. This is 
 an umbrella JIRA so we can collect sub-tasks as necessary. I think there 
 might be scattered JIRAs about this, let's link them in as we find them.
 [~steffkes] - I've taken the liberty of assigning it to you since you 
 expressed some interest. Feel free to assign it back if you want...
 Let's imagine that a user has a cluster with _no_ collections assigned and 
 start from there.
 Here's a simple way to set this up. Basically you follow the reference guide 
 tutorial but _don't_ define a collection.
 1 completely delete the collection1 directory from example
 2 cp -r example example2
 3 in example, execute java -DzkRun -jar start.jar
 4 in example2, execute java -Djetty.port=7574 -DzkHost=localhost:9983 -jar 
 start.jar
 Now the cloud link appears. If you expand the tree view, you see the two 
 live nodes. But, there's nothing in the graph view, no cores are selectable, 
 etc.
 First problem (need to solve before any sub-jiras, so including it here): You 
 have to push a configuration directory to ZK.
 [~thetapi] The _last_ time Stefan and I started allowing files to be written 
 to Solr from the UI it was...unfortunate. I'm assuming that there's something 
 similar here. That is, we shouldn't allow pushing the Solr config _to_ 
 ZooKeeper through the Admin UI, where they'd be distributed to all the solr 
 nodes. Is that true? If this is a security issue, we can keep pushing the 
 config dirs to ZK a manual step for now...
 Once we determine how to get configurations up, we can work on the various 
 sub-jiras.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-6876) Remove unused legacy scripts.conf

2014-12-28 Thread Erick Erickson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erick Erickson updated SOLR-6876:
-
Attachment: SOLR-6876.patch

Patch with entry in CHANGES.txt

 Remove unused legacy scripts.conf
 -

 Key: SOLR-6876
 URL: https://issues.apache.org/jira/browse/SOLR-6876
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.10.2, 5.0, Trunk
Reporter: Alexandre Rafalovitch
Assignee: Erick Erickson
Priority: Minor
 Attachments: SOLR-6876.patch


 Some of the example collections include *scripts.conf* in the *conf* 
 directory. It is not used by anything in the distribution and is somehow left 
 over from the Solr 1.x legacy days.
 It should be possible to safe delete it to avoid confusing users trying to 
 understand what different files actually do.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6876) Remove unused legacy scripts.conf

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259709#comment-14259709
 ] 

ASF subversion and git services commented on SOLR-6876:
---

Commit 1648246 from [~erickoerickson] in branch 'dev/trunk'
[ https://svn.apache.org/r1648246 ]

SOLR-6876: Remove unused legacy scripts.conf

 Remove unused legacy scripts.conf
 -

 Key: SOLR-6876
 URL: https://issues.apache.org/jira/browse/SOLR-6876
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.10.2, 5.0, Trunk
Reporter: Alexandre Rafalovitch
Assignee: Erick Erickson
Priority: Minor
 Attachments: SOLR-6876.patch


 Some of the example collections include *scripts.conf* in the *conf* 
 directory. It is not used by anything in the distribution and is somehow left 
 over from the Solr 1.x legacy days.
 It should be possible to safe delete it to avoid confusing users trying to 
 understand what different files actually do.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6893) DIH doen't work using URL and wget

2014-12-28 Thread Tomoko Uchida (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6893?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259713#comment-14259713
 ] 

Tomoko Uchida commented on SOLR-6893:
-

Hi,
At my environment (Mac OS), DIH works well with wget command.
Maybe you should remove '/#/' from request url.

I have tested two patterns with /#/ and without /#/
It works.
{code}
$ wget -O response 
http://localhost:8983/solr/wikipedia/dataimport?command=full-importindent=true;
$ wget -O response 
http://localhost:8983/solr/wikipedia/dataimport?command=statusindent=true;
$ cat response 
?xml version=1.0 encoding=UTF-8?
response

lst name=responseHeader
  int name=status0/int
  int name=QTime0/int
/lst
lst name=initArgs
  lst name=defaults
str name=configdb-data-config.xml/str
  /lst
/lst
str name=commandstatus/str
str name=statusbusy/str
str name=importResponseA command is still running.../str
lst name=statusMessages
  str name=Time Elapsed0:6:59.792/str
  str name=Total Requests made to DataSource1/str
  str name=Total Rows Fetched121772/str
  str name=Total Documents Processed60885/str
  str name=Total Documents Skipped0/str
  str name=Full Dump Started2014-12-28 19:38:24/str
/lst
/response
{code}

It does not work.
{code}
$ wget -O response 
http://localhost:8983/solr/#/wikipedia/dataimport?command=full-importindent=true;
{code}

Please try again after removing '#' character (this character has special 
meaning in URI specification.)

 DIH doen't work using URL and wget
 --

 Key: SOLR-6893
 URL: https://issues.apache.org/jira/browse/SOLR-6893
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.6.1, 4.8.1, 4.10.2
 Environment: Linux (Ubuntu, CentOS)
Reporter: Dani
Priority: Minor

 I put this URL on browser and import correctly starts:
 http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-importentity=fileImportclean=truecommit=true
 But if I use wget it doens't work:
 wget 
 http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-importentity=fileImportclean=truecommit=true;
 nor also using escape:
 wget 
 http://localhost:8983/solr/#/sintesicontratti/dataimport//importazione?command=full-import\entity=fileImport\clean=true\commit=true;



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-6876) Remove unused legacy scripts.conf

2014-12-28 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6876?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259726#comment-14259726
 ] 

ASF subversion and git services commented on SOLR-6876:
---

Commit 1648252 from [~erickoerickson] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1648252 ]

SOLR-6876: Remove unused legacy scripts.conf

 Remove unused legacy scripts.conf
 -

 Key: SOLR-6876
 URL: https://issues.apache.org/jira/browse/SOLR-6876
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.10.2, 5.0, Trunk
Reporter: Alexandre Rafalovitch
Assignee: Erick Erickson
Priority: Minor
 Fix For: 5.0, Trunk

 Attachments: SOLR-6876.patch


 Some of the example collections include *scripts.conf* in the *conf* 
 directory. It is not used by anything in the distribution and is somehow left 
 over from the Solr 1.x legacy days.
 It should be possible to safe delete it to avoid confusing users trying to 
 understand what different files actually do.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-6876) Remove unused legacy scripts.conf

2014-12-28 Thread Erick Erickson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6876?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Erick Erickson resolved SOLR-6876.
--
   Resolution: Fixed
Fix Version/s: Trunk
   5.0

Thanks Alexandre!

 Remove unused legacy scripts.conf
 -

 Key: SOLR-6876
 URL: https://issues.apache.org/jira/browse/SOLR-6876
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.10.2, 5.0, Trunk
Reporter: Alexandre Rafalovitch
Assignee: Erick Erickson
Priority: Minor
 Fix For: 5.0, Trunk

 Attachments: SOLR-6876.patch


 Some of the example collections include *scripts.conf* in the *conf* 
 directory. It is not used by anything in the distribution and is somehow left 
 over from the Solr 1.x legacy days.
 It should be possible to safe delete it to avoid confusing users trying to 
 understand what different files actually do.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5507) Admin UI - Refactoring using AngularJS

2014-12-28 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259729#comment-14259729
 ] 

Erick Erickson commented on SOLR-5507:
--

[~upayavira] Added a comment to SOLR-6082 about prioritization, I think that's 
enough.



 Admin UI - Refactoring using AngularJS
 --

 Key: SOLR-5507
 URL: https://issues.apache.org/jira/browse/SOLR-5507
 Project: Solr
  Issue Type: Improvement
  Components: web gui
Reporter: Stefan Matheis (steffkes)
Assignee: Stefan Matheis (steffkes)
Priority: Minor
 Attachments: SOLR-5507.patch


 On the LSR in Dublin, i've talked again to [~upayavira] and this time we 
 talked about Refactoring the existing UI - using AngularJS: providing (more, 
 internal) structure and what not ;
 He already started working on the Refactoring, so this is more a 'tracking' 
 issue about the progress he/we do there.
 Will extend this issue with a bit more context  additional information, w/ 
 thoughts about the possible integration in the existing UI and more (:



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-5.x-Windows (32bit/jdk1.8.0_40-ea-b09) - Build # 4412 - Failure!

2014-12-28 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Windows/4412/
Java: 32bit/jdk1.8.0_40-ea-b09 -client -XX:+UseParallelGC (asserts: false)

No tests ran.

Build Log:
[...truncated 10932 lines...]
FATAL: java.io.IOException: Unexpected termination of the channel
hudson.remoting.RequestAbortedException: java.io.IOException: Unexpected 
termination of the channel
at hudson.remoting.Request.abort(Request.java:295)
at hudson.remoting.Channel.terminate(Channel.java:814)
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:69)
at ..remote call to Windows VBOX(Native Method)
at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1356)
at hudson.remoting.Request.call(Request.java:171)
at hudson.remoting.Channel.call(Channel.java:751)
at 
hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:179)
at com.sun.proxy.$Proxy75.join(Unknown Source)
at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:979)
at hudson.Launcher$ProcStarter.join(Launcher.java:388)
at hudson.tasks.Ant.perform(Ant.java:217)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:770)
at hudson.model.Build$BuildExecution.build(Build.java:199)
at hudson.model.Build$BuildExecution.doRun(Build.java:160)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:533)
at hudson.model.Run.execute(Run.java:1759)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
at hudson.model.ResourceController.execute(ResourceController.java:89)
at hudson.model.Executor.run(Executor.java:240)
Caused by: java.io.IOException: Unexpected termination of the channel
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:50)
Caused by: java.io.EOFException
at 
java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2325)
at 
java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:2794)
at 
java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:801)
at java.io.ObjectInputStream.init(ObjectInputStream.java:299)
at 
hudson.remoting.ObjectInputStreamEx.init(ObjectInputStreamEx.java:40)
at 
hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:34)
at 
hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:48)



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[JENKINS-MAVEN] Lucene-Solr-Maven-5.x #801: POMs out of sync

2014-12-28 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Maven-5.x/801/

3 tests failed.
FAILED:  
org.apache.solr.hadoop.MorphlineGoLiveMiniMRTest.org.apache.solr.hadoop.MorphlineGoLiveMiniMRTest

Error Message:
null

Stack Trace:
java.lang.AssertionError: null
at __randomizedtesting.SeedInfo.seed([42DA430FA16CA04]:0)
at 
org.apache.lucene.util.TestRuleTemporaryFilesCleanup.before(TestRuleTemporaryFilesCleanup.java:105)
at 
com.carrotsearch.randomizedtesting.rules.TestRuleAdapter$1.before(TestRuleAdapter.java:26)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:35)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745)


FAILED:  org.apache.solr.hadoop.MorphlineBasicMiniMRTest.testPathParts

Error Message:
Test abandoned because suite timeout was reached.

Stack Trace:
java.lang.Exception: Test abandoned because suite timeout was reached.
at __randomizedtesting.SeedInfo.seed([B98A46E3988AA203]:0)


FAILED:  
org.apache.solr.hadoop.MorphlineBasicMiniMRTest.org.apache.solr.hadoop.MorphlineBasicMiniMRTest

Error Message:
Suite timeout exceeded (= 720 msec).

Stack Trace:
java.lang.Exception: Suite timeout exceeded (= 720 msec).
at __randomizedtesting.SeedInfo.seed([B98A46E3988AA203]:0)




Build Log:
[...truncated 53961 lines...]
BUILD FAILED
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Maven-5.x/build.xml:552: 
The following error occurred while executing this line:
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Maven-5.x/build.xml:204: 
The following error occurred while executing this line:
: Java returned: 1

Total time: 380 minutes 16 seconds
Build step 'Invoke Ant' marked build as failure
Recording test results
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[JENKINS] Lucene-Solr-trunk-Linux (64bit/jdk1.9.0-ea-b34) - Build # 11824 - Failure!

2014-12-28 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Linux/11824/
Java: 64bit/jdk1.9.0-ea-b34 -XX:-UseCompressedOops -XX:+UseG1GC (asserts: true)

2 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.ChaosMonkeySafeLeaderTest

Error Message:
Suite timeout exceeded (= 720 msec).

Stack Trace:
java.lang.Exception: Suite timeout exceeded (= 720 msec).
at __randomizedtesting.SeedInfo.seed([78D62EC33B461E33]:0)


FAILED:  org.apache.solr.cloud.ChaosMonkeySafeLeaderTest.testDistribSearch

Error Message:
Test abandoned because suite timeout was reached.

Stack Trace:
java.lang.Exception: Test abandoned because suite timeout was reached.
at __randomizedtesting.SeedInfo.seed([78D62EC33B461E33]:0)




Build Log:
[...truncated 10414 lines...]
   [junit4] Suite: org.apache.solr.cloud.ChaosMonkeySafeLeaderTest
   [junit4]   2 Creating dataDir: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderTest-78D62EC33B461E33-001/init-core-data-001
   [junit4]   2 1691221 T5088 oas.SolrTestCaseJ4.buildSSLConfig Randomized ssl 
(false) and clientAuth (false)
   [junit4]   2 1691221 T5088 
oas.BaseDistributedSearchTestCase.initHostContext Setting hostContext system 
property: /
   [junit4]   2 1691223 T5088 oas.SolrTestCaseJ4.setUp ###Starting 
testDistribSearch
   [junit4]   2 1691224 T5088 oasc.ZkTestServer.run STARTING ZK TEST SERVER
   [junit4]   1 client port:0.0.0.0/0.0.0.0:0
   [junit4]   2 1691224 T5089 oasc.ZkTestServer$ZKServerMain.runFromConfig 
Starting server
   [junit4]   2 1691324 T5088 oasc.ZkTestServer.run start zk server on 
port:45567
   [junit4]   2 1691325 T5088 
oascc.SolrZkClient.createZkCredentialsToAddAutomatically Using default 
ZkCredentialsProvider
   [junit4]   2 1691325 T5088 oascc.ConnectionManager.waitForConnected Waiting 
for client to connect to ZooKeeper
   [junit4]   2 1691328 T5096 oascc.ConnectionManager.process Watcher 
org.apache.solr.common.cloud.ConnectionManager@733c650b 
name:ZooKeeperConnection Watcher:127.0.0.1:45567 got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
   [junit4]   2 1691328 T5088 oascc.ConnectionManager.waitForConnected Client 
is connected to ZooKeeper
   [junit4]   2 1691329 T5088 oascc.SolrZkClient.createZkACLProvider Using 
default ZkACLProvider
   [junit4]   2 1691329 T5088 oascc.SolrZkClient.makePath makePath: /solr
   [junit4]   2 1691332 T5088 
oascc.SolrZkClient.createZkCredentialsToAddAutomatically Using default 
ZkCredentialsProvider
   [junit4]   2 1691333 T5088 oascc.ConnectionManager.waitForConnected Waiting 
for client to connect to ZooKeeper
   [junit4]   2 1691333 T5099 oascc.ConnectionManager.process Watcher 
org.apache.solr.common.cloud.ConnectionManager@7099daea 
name:ZooKeeperConnection Watcher:127.0.0.1:45567/solr got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
   [junit4]   2 1691334 T5088 oascc.ConnectionManager.waitForConnected Client 
is connected to ZooKeeper
   [junit4]   2 1691334 T5088 oascc.SolrZkClient.createZkACLProvider Using 
default ZkACLProvider
   [junit4]   2 1691334 T5088 oascc.SolrZkClient.makePath makePath: 
/collections/collection1
   [junit4]   2 1691336 T5088 oascc.SolrZkClient.makePath makePath: 
/collections/collection1/shards
   [junit4]   2 1691338 T5088 oascc.SolrZkClient.makePath makePath: 
/collections/control_collection
   [junit4]   2 1691339 T5088 oascc.SolrZkClient.makePath makePath: 
/collections/control_collection/shards
   [junit4]   2 1691340 T5088 oasc.AbstractZkTestCase.putConfig put 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2 1691340 T5088 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/solrconfig.xml
   [junit4]   2 1691342 T5088 oasc.AbstractZkTestCase.putConfig put 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/schema15.xml
 to /configs/conf1/schema.xml
   [junit4]   2 1691342 T5088 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/schema.xml
   [junit4]   2 1691343 T5088 oasc.AbstractZkTestCase.putConfig put 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
 to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2 1691344 T5088 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2 1691345 T5088 oasc.AbstractZkTestCase.putConfig put 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
 to /configs/conf1/stopwords.txt
   [junit4]   2 1691345 T5088 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/stopwords.txt
   [junit4]   2 1691346 T5088 oasc.AbstractZkTestCase.putConfig put 

[jira] [Commented] (SOLR-6888) Decompressing documents on first-pass distributed queries to get docId is inefficient, use indexed values instead?

2014-12-28 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-6888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14259787#comment-14259787
 ] 

David Smiley commented on SOLR-6888:


Erick,
I enjoyed reading what you have here.  I think this issue duplicates SOLR-5478, 
for which I have a patch in fact. I encourage you to review it and kick the 
tires on it!

 Decompressing documents on first-pass distributed queries to get docId is 
 inefficient, use indexed values instead?
 --

 Key: SOLR-6888
 URL: https://issues.apache.org/jira/browse/SOLR-6888
 Project: Solr
  Issue Type: Improvement
Affects Versions: 5.0, Trunk
Reporter: Erick Erickson
Assignee: Erick Erickson
 Attachments: SOLR-6888-hacktiming.patch


 Assigning this to myself to just not lose track of it, but I won't be working 
 on this in the near term; anyone feeling ambitious should feel free to grab 
 it.
 Note, docId used here is whatever is defined for uniqueKey...
 Since Solr 4.1, the compression/decompression process is based on 16K blocks 
 and is automatic, and not configurable. So, to get a single stored value one 
 must decompress an entire 16K block. At least.
 For SolrCloud (and distributed processing in general), we make two trips, one 
 to get the doc id and score (or other sort criteria) and one to return the 
 actual data.
 The first pass here requires that we return the top N docIDs and sort 
 criteria, which means that each and every sub-request has to unpack at least 
 one 16K block (and sometimes more) to get just the doc ID. So if we have 20 
 shards and only want 20 rows, 95% of the decompression cycles will be wasted. 
 Not to mention all the disk reads.
 It seems like we should be able to do better than that. Can we argue that doc 
 ids are 'special' and should be cached somehow? Let's discuss what this would 
 look like. I can think of a couple of approaches:
 1 Since doc IDs are special, can we say that for this purpose returning 
 the indexed version is OK? We'd need to return the actual stored value when 
 the full doc was requested, but for the sub-request only what about returning 
 the indexed value instead of the stored one? On the surface I don't see a 
 problem here, but what do I know? Storing these as DocValues seems useful in 
 this case.
 1a A variant is treating numeric docIds specially since the indexed value 
 and the stored value should be identical. And DocValues here would be useful 
 it seems. But this seems an unnecessary specialization if 1 is implemented 
 well.
 2 We could cache individual doc IDs, although I'm not sure what use that 
 really is. Would maintaining the cache overwhelm the savings of not 
 decompressing? I really don't like this idea, but am throwing it out there. 
 Doing this from stored data up front would essentially mean decompressing 
 every doc so that seems untenable to try up-front.
 3 We could maintain an array[maxDoc] that held document IDs, perhaps lazily 
 initializing it. I'm not particularly a fan of this either, doesn't seem like 
 a Good Thing. I can see lazy loading being almost, but not quite totally, 
 useless, i.e. a hit ratio near 0, especially since it'd be thrown out on 
 every openSearcher.
 Really, the only one of these that seems viable is 1/1a. The others would 
 all involve decompressing the docs anyway to get the ID, and I suspect that 
 caching would be of very limited usefulness. I guess 1's viability hinges 
 on whether, for internal use, the indexed form of DocId is interchangeable 
 with the stored value.
 Or are there other ways to approach this? Or isn't it something to really 
 worry about?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-NightlyTests-5.x - Build # 717 - Failure

2014-12-28 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-5.x/717/

1 tests failed.
REGRESSION:  
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testDistribSearch

Error Message:
Error from server at http://127.0.0.1:54363: Error CREATEing SolrCore 
'halfcollection_shard1_replica1': Unable to create core 
[halfcollection_shard1_replica1] Caused by: Could not get shard id for core: 
halfcollection_shard1_replica1

Stack Trace:
org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException: Error 
from server at http://127.0.0.1:54363: Error CREATEing SolrCore 
'halfcollection_shard1_replica1': Unable to create core 
[halfcollection_shard1_replica1] Caused by: Could not get shard id for core: 
halfcollection_shard1_replica1
at 
org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:566)
at 
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:213)
at 
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:209)
at 
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testErrorHandling(CollectionsAPIDistributedZkTest.java:452)
at 
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.doTest(CollectionsAPIDistributedZkTest.java:201)
at 
org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:869)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 

[JENKINS] Lucene-Solr-Tests-5.x-Java7 - Build # 2394 - Failure

2014-12-28 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-5.x-Java7/2394/

2 tests failed.
REGRESSION:  org.apache.solr.cloud.ChaosMonkeySafeLeaderTest.testDistribSearch

Error Message:
Test abandoned because suite timeout was reached.

Stack Trace:
java.lang.Exception: Test abandoned because suite timeout was reached.
at __randomizedtesting.SeedInfo.seed([2FE41655C33758DD]:0)


FAILED:  
junit.framework.TestSuite.org.apache.solr.cloud.ChaosMonkeySafeLeaderTest

Error Message:
Suite timeout exceeded (= 720 msec).

Stack Trace:
java.lang.Exception: Suite timeout exceeded (= 720 msec).
at __randomizedtesting.SeedInfo.seed([2FE41655C33758DD]:0)




Build Log:
[...truncated 10475 lines...]
   [junit4] Suite: org.apache.solr.cloud.ChaosMonkeySafeLeaderTest
   [junit4]   2 Creating dataDir: 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-5.x-Java7/solr/build/solr-core/test/J2/temp/solr.cloud.ChaosMonkeySafeLeaderTest-2FE41655C33758DD-001/init-core-data-001
   [junit4]   2 1667952 T3808 oas.SolrTestCaseJ4.buildSSLConfig Randomized ssl 
(false) and clientAuth (false)
   [junit4]   2 1667952 T3808 
oas.BaseDistributedSearchTestCase.initHostContext Setting hostContext system 
property: /dt_/lw
   [junit4]   2 1667959 T3808 oas.SolrTestCaseJ4.setUp ###Starting 
testDistribSearch
   [junit4]   2 1667960 T3808 oasc.ZkTestServer.run STARTING ZK TEST SERVER
   [junit4]   1 client port:0.0.0.0/0.0.0.0:0
   [junit4]   2 1667961 T3809 oasc.ZkTestServer$ZKServerMain.runFromConfig 
Starting server
   [junit4]   2 1668060 T3808 oasc.ZkTestServer.run start zk server on 
port:11132
   [junit4]   2 1668061 T3808 
oascc.SolrZkClient.createZkCredentialsToAddAutomatically Using default 
ZkCredentialsProvider
   [junit4]   2 1668062 T3808 oascc.ConnectionManager.waitForConnected Waiting 
for client to connect to ZooKeeper
   [junit4]   2 1668065 T3816 oascc.ConnectionManager.process Watcher 
org.apache.solr.common.cloud.ConnectionManager@418cfca8 
name:ZooKeeperConnection Watcher:127.0.0.1:11132 got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
   [junit4]   2 1668066 T3808 oascc.ConnectionManager.waitForConnected Client 
is connected to ZooKeeper
   [junit4]   2 1668066 T3808 oascc.SolrZkClient.createZkACLProvider Using 
default ZkACLProvider
   [junit4]   2 1668066 T3808 oascc.SolrZkClient.makePath makePath: /solr
   [junit4]   2 1668069 T3808 
oascc.SolrZkClient.createZkCredentialsToAddAutomatically Using default 
ZkCredentialsProvider
   [junit4]   2 1668070 T3808 oascc.ConnectionManager.waitForConnected Waiting 
for client to connect to ZooKeeper
   [junit4]   2 1668072 T3819 oascc.ConnectionManager.process Watcher 
org.apache.solr.common.cloud.ConnectionManager@4b266679 
name:ZooKeeperConnection Watcher:127.0.0.1:11132/solr got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
   [junit4]   2 1668072 T3808 oascc.ConnectionManager.waitForConnected Client 
is connected to ZooKeeper
   [junit4]   2 1668072 T3808 oascc.SolrZkClient.createZkACLProvider Using 
default ZkACLProvider
   [junit4]   2 1668073 T3808 oascc.SolrZkClient.makePath makePath: 
/collections/collection1
   [junit4]   2 1668075 T3808 oascc.SolrZkClient.makePath makePath: 
/collections/collection1/shards
   [junit4]   2 1668076 T3808 oascc.SolrZkClient.makePath makePath: 
/collections/control_collection
   [junit4]   2 1668078 T3808 oascc.SolrZkClient.makePath makePath: 
/collections/control_collection/shards
   [junit4]   2 1668080 T3808 oasc.AbstractZkTestCase.putConfig put 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-5.x-Java7/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2 1668080 T3808 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/solrconfig.xml
   [junit4]   2 1668083 T3808 oasc.AbstractZkTestCase.putConfig put 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-5.x-Java7/solr/core/src/test-files/solr/collection1/conf/schema15.xml
 to /configs/conf1/schema.xml
   [junit4]   2 1668083 T3808 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/schema.xml
   [junit4]   2 1668185 T3808 oasc.AbstractZkTestCase.putConfig put 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-5.x-Java7/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml
 to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2 1668186 T3808 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2 1668188 T3808 oasc.AbstractZkTestCase.putConfig put 
/usr/home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-5.x-Java7/solr/core/src/test-files/solr/collection1/conf/stopwords.txt
 to /configs/conf1/stopwords.txt
   [junit4]   2 1668189 T3808 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/stopwords.txt
   [junit4]   2 1668190 T3808 oasc.AbstractZkTestCase.putConfig put 

Multiple CFS files are generated in lucene 4.10.2

2014-12-28 Thread jaga_india
Hi

I am trying index around 612 record using lucene 4.10.2. It is creating lot
of CF files in index directory.
around 626 CFS file are created. It is taking more time to index.

ENV : java 8, window 7

http://lucene.472066.n3.nabble.com/file/n4176336/index_files.jpg 

Directory dir = FSDirectory.open(file);
IndexWriterConfig config = new
IndexWriterConfig(Version.LUCENE_4_10_2, new ClassicAnalyzer());
if(bufferSizeMB != 0  bufferSizeMB != -1){
config.setRAMBufferSizeMB(bufferSizeMB);
}  else {
config.setRAMBufferSizeMB(DEFAULT_RAM_BUFFER_SIZE_MB);
}  
config.setMaxBufferedDocs(1000);
config.setMaxBufferedDeleteTerms(1000);
config.setMergePolicy(new LogDocMergePolicy());
IndexWriter iwriter = new IndexWriter(dir, config);
iwriter.getConfig().setMaxBufferedDeleteTerms(1000);
iwriter.getConfig().setMaxBufferedDocs(1000);
iwriter.getConfig().setRAMBufferSizeMB(bufferSizeMB)


Thanks and Regards
jaga india



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Multiple-CFS-files-are-generated-in-lucene-4-10-2-tp4176336.html
Sent from the Lucene - Java Developer mailing list archive at Nabble.com.

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org