Re: Problem using TrecTopicsReader()

2013-08-06 Thread Andi Vajda

On Aug 6, 2013, at 10:36, Abhishek Gupta abhi.bansa...@gmail.com wrote:

 Hi,
 I have a problem using TrecTopicsReader which I have already asked on
 #lucene IRC but got no reply till yesterday. I also post it on SO(
 http://stackoverflow.com/questions/18054649/bufferedreader-in-pylucne).
 
 The problem is completely explained on SO.

In the PyLucene Makefile find the jcc invocation and add java.io.BufferedReader 
to the long command line (don't forget the ending \ as needed) and rebuild 
PyLucene.

Andi..

 
 Please Help!!
 
 Thanking You
 Abhishek Gupta,
 9416106204, 9624799165


Re: Problem using TrecTopicsReader()

2013-08-06 Thread Andi Vajda

On Aug 6, 2013, at 21:17, Abhishek Gupta abhi.bansa...@gmail.com wrote:

 Hi,
 Sorry for asking again, but I got another error.
 
 After including the all the java libraries I want in jcc. I wrote code for
 Trec Evaluation. While running it I got the error that `there is no
 bechmark module `. So I modified the Makefile to compile and install
 benchmark jar also.
 
 But, while compiling the compiling pylucene I am getting the following
 error:
 
 -
 BUILD SUCCESSFUL
 Total time: 1 second
 ICU not installed
 /usr/bin/python -m jcc --shared --jar
 lucene-java-4.3.0/lucene/build/core/lucene-core-4.3.0.jar --jar
 lucene-java-4.3.0/lucene/build/analysis/common/lucene-analyzers-common-4.3.0.jar
 --jar lucene-java-4.3.0/lucene/build/benchmark/lucene-benchmark-4.3.0.jar
 --jar lucene-java-4.3.0/lucene/build/memory/lucene-memory-4.3.0.jar --jar
 lucene-java-4.3.0/lucene/build/highlighter/lucene-highlighter-4.3.0.jar
 --jar build/jar/extensions.jar --jar
 lucene-java-4.3.0/lucene/build/queries/lucene-queries-4.3.0.jar --jar
 lucene-java-4.3.0/lucene/build/queryparser/lucene-queryparser-4.3.0.jar
 --jar lucene-java-4.3.0/lucene/build/sandbox/lucene-sandbox-4.3.0.jar --jar
 lucene-java-4.3.0/lucene/build/grouping/lucene-grouping-4.3.0.jar --jar
 lucene-java-4.3.0/lucene/build/join/lucene-join-4.3.0.jar --jar
 lucene-java-4.3.0/lucene/build/facet/lucene-facet-4.3.0.jar --jar
 lucene-java-4.3.0/lucene/build/suggest/lucene-suggest-4.3.0.jar
 --use_full_names --package java.lang java.lang.System java.lang.Runtime
 --package java.util java.util.Arrays java.util.Collections
 java.util.HashMap java.util.HashSet java.util.TreeSet
 java.lang.IllegalStateException java.lang.IndexOutOfBoundsException
 java.util.NoSuchElementException java.text.SimpleDateFormat
 java.text.DecimalFormat java.text.Collator --package java.util.concurrent
 java.util.concurrent.Executors --package java.util.regex --package
 java.iojava.io.StringReader java.io.InputStreamReader
 java.io.FileInputStream
 java.io.BufferedReader java.io.FileReader java.io.PrintWriter --exclude
 org.apache.lucene.sandbox.queries.regex.JakartaRegexpCapabilities --exclude
 org.apache.regexp.RegexpTunnel --python lucene --mapping
 org.apache.lucene.document.Document
 'get:(Ljava/lang/String;)Ljava/lang/String;' --mapping java.util.Properties
 'getProperty:(Ljava/lang/String;)Ljava/lang/String;' --sequence
 java.util.AbstractList 'size:()I' 'get:(I)Ljava/lang/Object;'
 org.apache.lucene.index.IndexWriter:getReader --version 4.3.0 --module
 python/collections.py --module python/ICUNormalizer2Filter.py --module
 python/ICUFoldingFilter.py --module python/ICUTransformFilter.py  --files 8
 --build
 While loading org/apache/lucene/benchmark/byTask/utils/StreamUtils
 Traceback (most recent call last):
  File /usr/lib/python2.7/runpy.py, line 162, in _run_module_as_main
__main__, fname, loader, pkg_name)
  File /usr/lib/python2.7/runpy.py, line 72, in _run_code
exec code in run_globals
  File
 /usr/local/lib/python2.7/dist-packages/JCC-2.16-py2.7-linux-x86_64.egg/jcc/__main__.py,
 line 107, in module
cpp.jcc(sys.argv)
  File
 /usr/local/lib/python2.7/dist-packages/JCC-2.16-py2.7-linux-x86_64.egg/jcc/cpp.py,
 line 583, in jcc
cls = findClass(className.replace('.', '/'))
  File
 /usr/local/lib/python2.7/dist-packages/JCC-2.16-py2.7-linux-x86_64.egg/jcc/cpp.py,
 line 73, in findClass
cls = _findClass(className)
 jcc.cpp.JavaError: java.lang.NoClassDefFoundError:
 org/apache/commons/compress/compressors/CompressorException
 Java stacktrace:
 java.lang.NoClassDefFoundError:
 org/apache/commons/compress/compressors/CompressorException
 at
 org.apache.lucene.benchmark.byTask.utils.StreamUtils.clinit(StreamUtils.java:77)
 Caused by: java.lang.ClassNotFoundException:
 org.apache.commons.compress.compressors.CompressorException
 at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
 at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
 at java.security.AccessController.doPrivileged(Native Method)
 at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
 at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
 ... 1 more
 
 make: *** [compile] Error 1
 -
 
 I think this error is coming because there is no
 `org.apache.commons.compress.compressors.CompressorException`. I don't know
 how to install it for pylucene. Please Help!

If the additional classes you need require org.apache.commons.compressor etcc
.., you need to endure that they're found on the classpath. This is a java 
error.

Andi..

 
 Thanks,
 
 
 On Tue, Aug 6, 2013 at 10:57 PM, Abhishek Gupta 
 abhi.bansa...@gmail.comwrote:
 
 Thanks!! Andi :)
 
 
 On Tue, Aug 6, 2013 

[jira] [Commented] (LUCENE-4542) Make RECURSION_CAP in HunspellStemmer configurable

2013-08-06 Thread JIRA

[ 
https://issues.apache.org/jira/browse/LUCENE-4542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730565#comment-13730565
 ] 

Jan Høydahl commented on LUCENE-4542:
-

I believe this should be closed and set fix-version to 4.4 again?

 Make RECURSION_CAP in HunspellStemmer configurable
 --

 Key: LUCENE-4542
 URL: https://issues.apache.org/jira/browse/LUCENE-4542
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/analysis
Affects Versions: 4.0
Reporter: Piotr
Assignee: Steve Rowe
 Fix For: 5.0, 4.5

 Attachments: Lucene-4542-javadoc.patch, LUCENE-4542.patch, 
 LUCENE-4542-with-solr.patch


 Currently there is 
 private static final int RECURSION_CAP = 2;
 in the code of the class HunspellStemmer. It makes using hunspell with 
 several dictionaries almost unusable, due to bad performance (f.ex. it costs 
 36ms to stem long sentence in latvian for recursion_cap=2 and 5 ms for 
 recursion_cap=1). It would be nice to be able to tune this number as needed.
 AFAIK this number (2) was chosen arbitrary.
 (it's a first issue in my life, so please forgive me any mistakes done).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-4311) HunspellStemFilter returns another values than Hunspell in console / command line with same dictionaries.

2013-08-06 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/LUCENE-4311?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl resolved LUCENE-4311.
-

Resolution: Duplicate

Closing this as duplicate of LUCENE-4542. You can now tune this to your liking 
depending on what dictionary is in use.

 HunspellStemFilter returns another values than Hunspell in console / command 
 line with same dictionaries.
 -

 Key: LUCENE-4311
 URL: https://issues.apache.org/jira/browse/LUCENE-4311
 Project: Lucene - Core
  Issue Type: Bug
  Components: core/other
Affects Versions: 3.5, 4.0-ALPHA, 3.6.1
 Environment: Apache Solr 3.5 - 4.0, Apache Tomcat 7.0
Reporter: Jan Rieger
 Attachments: cs_CZ.aff, cs_CZ.dic


 When I used HunspellStemFilter for stemming the czech language text, it 
 returns me bad results.
 For example word praha returns praha and prahnout, what is not correct.
 So I try the same in my console (Hunspell command line) with exactly same 
 dictionaries and it returns only praha and this is correct.
 Can somebody help me?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-4542) Make RECURSION_CAP in HunspellStemmer configurable

2013-08-06 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/LUCENE-4542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl resolved LUCENE-4542.
-

Resolution: Fixed

Closing for 4.4. Please reopen if this is wrong, but the code and changes entry 
were committed for 4.4 so everything should be fine.

 Make RECURSION_CAP in HunspellStemmer configurable
 --

 Key: LUCENE-4542
 URL: https://issues.apache.org/jira/browse/LUCENE-4542
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/analysis
Affects Versions: 4.0
Reporter: Piotr
Assignee: Steve Rowe
 Fix For: 5.0, 4.4

 Attachments: Lucene-4542-javadoc.patch, LUCENE-4542.patch, 
 LUCENE-4542-with-solr.patch


 Currently there is 
 private static final int RECURSION_CAP = 2;
 in the code of the class HunspellStemmer. It makes using hunspell with 
 several dictionaries almost unusable, due to bad performance (f.ex. it costs 
 36ms to stem long sentence in latvian for recursion_cap=2 and 5 ms for 
 recursion_cap=1). It would be nice to be able to tune this number as needed.
 AFAIK this number (2) was chosen arbitrary.
 (it's a first issue in my life, so please forgive me any mistakes done).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4542) Make RECURSION_CAP in HunspellStemmer configurable

2013-08-06 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/LUCENE-4542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl updated LUCENE-4542:


Fix Version/s: (was: 4.5)
   4.4

 Make RECURSION_CAP in HunspellStemmer configurable
 --

 Key: LUCENE-4542
 URL: https://issues.apache.org/jira/browse/LUCENE-4542
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/analysis
Affects Versions: 4.0
Reporter: Piotr
Assignee: Steve Rowe
 Fix For: 5.0, 4.4

 Attachments: Lucene-4542-javadoc.patch, LUCENE-4542.patch, 
 LUCENE-4542-with-solr.patch


 Currently there is 
 private static final int RECURSION_CAP = 2;
 in the code of the class HunspellStemmer. It makes using hunspell with 
 several dictionaries almost unusable, due to bad performance (f.ex. it costs 
 36ms to stem long sentence in latvian for recursion_cap=2 and 5 ms for 
 recursion_cap=1). It would be nice to be able to tune this number as needed.
 AFAIK this number (2) was chosen arbitrary.
 (it's a first issue in my life, so please forgive me any mistakes done).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5103) Plugin Improvements

2013-08-06 Thread Grant Ingersoll (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5103?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730663#comment-13730663
 ] 

Grant Ingersoll commented on SOLR-5103:
---

Hoss,

A few of my thoughts on it:
# I don't think the config necessarily needs to be in the plugin.  Part of the 
goal here is to be able to easily add a plugin and then easily configure it and 
manage it (from the UI?).
# Plugins should be self-contained and (ideally) class loader isolated
# You shouldn't necessarily have to restart Solr in order to add a plugin


 Plugin Improvements
 ---

 Key: SOLR-5103
 URL: https://issues.apache.org/jira/browse/SOLR-5103
 Project: Solr
  Issue Type: Improvement
Reporter: Grant Ingersoll
Assignee: Grant Ingersoll
 Fix For: 5.0


 I think for 5.0, we should make it easier to add plugins by defining a plugin 
 package, ala a Hadoop Job jar, which is a self--contained archive of a plugin 
 that can be easily installed (even from the UI!) and configured 
 programmatically.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5107) LukeRequestHandler throws NullPointerException when numTerms=0

2013-08-06 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730665#comment-13730665
 ] 

Erick Erickson commented on SOLR-5107:
--

Hoss:

Looks good to me.

 LukeRequestHandler throws NullPointerException when numTerms=0
 --

 Key: SOLR-5107
 URL: https://issues.apache.org/jira/browse/SOLR-5107
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Ahmet Arslan
Assignee: Hoss Man
Priority: Minor
 Attachments: SOLR-5107.patch, SOLR-5107.patch


 Defaults example 
 http://localhost:8983/solr/collection1/admin/luke?fl=catnumTerms=0 yields 
 {code}
 ERROR org.apache.solr.core.SolrCore  – java.lang.NullPointerException
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.getDetailedFieldInfo(LukeRequestHandler.java:610)
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.getIndexedFieldsInfo(LukeRequestHandler.java:378)
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:160)
   at 
 org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
   at org.apache.solr.core.SolrCore.execute(SolrCore.java:1845)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:666)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:369)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:158)
   at 
 org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419)
   at 
 org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455)
   at 
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
   at 
 org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557)
   at 
 org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
   at 
 org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075)
   at 
 org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384)
   at 
 org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
   at 
 org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009)
   at 
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135)
   at 
 org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255)
   at 
 org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154)
   at 
 org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116)
   at org.eclipse.jetty.server.Server.handle(Server.java:368)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489)
   at 
 org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:53)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:942)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1004)
   at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:640)
   at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235)
   at 
 org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:72)
   at 
 org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:264)
   at 
 org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:608)
   at 
 org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:543)
   at java.lang.Thread.run(Thread.java:722)
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5156) CompressingTermVectors termsEnum should probably not support seek-by-ord

2013-08-06 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730673#comment-13730673
 ] 

Michael McCandless commented on LUCENE-5156:


+1

 CompressingTermVectors termsEnum should probably not support seek-by-ord
 

 Key: LUCENE-5156
 URL: https://issues.apache.org/jira/browse/LUCENE-5156
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: LUCENE-5156.patch


 Just like term vectors before it, it has a O(n) seek-by-term. 
 But this one also advertises a seek-by-ord, only this is also O(n).
 This could cause e.g. checkindex to be very slow, because if termsenum 
 supports ord it does a bunch of seeking tests. (Another solution would be to 
 leave it, and add a boolean so checkindex never does seeking tests for term 
 vectors, only real fields).
 However, I think its also kinda a trap, in my opinion if seek-by-ord is 
 supported anywhere, you kinda expect it to be faster than linear time...?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5156) CompressingTermVectors termsEnum should probably not support seek-by-ord

2013-08-06 Thread Adrien Grand (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730688#comment-13730688
 ] 

Adrien Grand commented on LUCENE-5156:
--

Patch looks good, +1

 CompressingTermVectors termsEnum should probably not support seek-by-ord
 

 Key: LUCENE-5156
 URL: https://issues.apache.org/jira/browse/LUCENE-5156
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: LUCENE-5156.patch


 Just like term vectors before it, it has a O(n) seek-by-term. 
 But this one also advertises a seek-by-ord, only this is also O(n).
 This could cause e.g. checkindex to be very slow, because if termsenum 
 supports ord it does a bunch of seeking tests. (Another solution would be to 
 leave it, and add a boolean so checkindex never does seeking tests for term 
 vectors, only real fields).
 However, I think its also kinda a trap, in my opinion if seek-by-ord is 
 supported anywhere, you kinda expect it to be faster than linear time...?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5157) Refactoring MultiDocValues.OrdinalMap to clarify API and internal structure.

2013-08-06 Thread Adrien Grand (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5157?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730727#comment-13730727
 ] 

Adrien Grand commented on LUCENE-5157:
--

bq. on the other hand the fact it uses termsenum at all could be seen as an 
impl detail of merging (sorted)set dictionaries

I tend to see it this way since ordinals are more useful on doc values than on 
an inverted index's terms dictionary?





 Refactoring MultiDocValues.OrdinalMap to clarify API and internal structure.
 

 Key: LUCENE-5157
 URL: https://issues.apache.org/jira/browse/LUCENE-5157
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Boaz Leskes
Priority: Minor
 Attachments: LUCENE-5157.patch


 I refactored MultiDocValues.OrdinalMap, removing one unused parameter and 
 renaming some methods to more clearly communicate what they do. Also I 
 renamed subIndex references to segmentIndex.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5103) Plugin Improvements

2013-08-06 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5103?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730846#comment-13730846
 ] 

Mark Miller commented on SOLR-5103:
---

I think we would still need a configuration option that does not involve Java. 
Other than that, I don't really see too many issues. As far as names, it could 
possibly default to the name of the plugin jar or something else and first one 
wins, second one fails. We could also namespace, kind of how you see with 
eclipse. Honestly though, nothing sounds much better than a simple config file 
with the jar, same name or whatever, that has the plugin name or url. Probably 
that's the default minimal setup - even if you can delete that file and run 
with some default names or whatever for plugins that don't require any config.

I think the path is obviously right - you see it used plenty elsewhere, and it 
resonates with me. But I think Hoss is right in that the common case will still 
need configuration. I don't see how the two points argue with each other though.

We will need to be able to configure with a user friendly format - most plugins 
will have that - if not just to show the user what can be configured. There is 
no reason we have to require it though. Some plugin might simply be there to 
insert some custom code on startup.

 Plugin Improvements
 ---

 Key: SOLR-5103
 URL: https://issues.apache.org/jira/browse/SOLR-5103
 Project: Solr
  Issue Type: Improvement
Reporter: Grant Ingersoll
Assignee: Grant Ingersoll
 Fix For: 5.0


 I think for 5.0, we should make it easier to add plugins by defining a plugin 
 package, ala a Hadoop Job jar, which is a self--contained archive of a plugin 
 that can be easily installed (even from the UI!) and configured 
 programmatically.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5156) CompressingTermVectors termsEnum should probably not support seek-by-ord

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730869#comment-13730869
 ] 

ASF subversion and git services commented on LUCENE-5156:
-

Commit 1511009 from [~rcmuir] in branch 'dev/trunk'
[ https://svn.apache.org/r1511009 ]

LUCENE-5156: remove seek-by-ord from CompressingTermVectors

 CompressingTermVectors termsEnum should probably not support seek-by-ord
 

 Key: LUCENE-5156
 URL: https://issues.apache.org/jira/browse/LUCENE-5156
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: LUCENE-5156.patch


 Just like term vectors before it, it has a O(n) seek-by-term. 
 But this one also advertises a seek-by-ord, only this is also O(n).
 This could cause e.g. checkindex to be very slow, because if termsenum 
 supports ord it does a bunch of seeking tests. (Another solution would be to 
 leave it, and add a boolean so checkindex never does seeking tests for term 
 vectors, only real fields).
 However, I think its also kinda a trap, in my opinion if seek-by-ord is 
 supported anywhere, you kinda expect it to be faster than linear time...?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-5156) CompressingTermVectors termsEnum should probably not support seek-by-ord

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-5156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730880#comment-13730880
 ] 

ASF subversion and git services commented on LUCENE-5156:
-

Commit 1511014 from [~rcmuir] in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1511014 ]

LUCENE-5156: remove seek-by-ord from CompressingTermVectors

 CompressingTermVectors termsEnum should probably not support seek-by-ord
 

 Key: LUCENE-5156
 URL: https://issues.apache.org/jira/browse/LUCENE-5156
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Attachments: LUCENE-5156.patch


 Just like term vectors before it, it has a O(n) seek-by-term. 
 But this one also advertises a seek-by-ord, only this is also O(n).
 This could cause e.g. checkindex to be very slow, because if termsenum 
 supports ord it does a bunch of seeking tests. (Another solution would be to 
 leave it, and add a boolean so checkindex never does seeking tests for term 
 vectors, only real fields).
 However, I think its also kinda a trap, in my opinion if seek-by-ord is 
 supported anywhere, you kinda expect it to be faster than linear time...?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-5156) CompressingTermVectors termsEnum should probably not support seek-by-ord

2013-08-06 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-5156?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-5156.
-

   Resolution: Fixed
Fix Version/s: 4.5
   5.0

 CompressingTermVectors termsEnum should probably not support seek-by-ord
 

 Key: LUCENE-5156
 URL: https://issues.apache.org/jira/browse/LUCENE-5156
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Robert Muir
 Fix For: 5.0, 4.5

 Attachments: LUCENE-5156.patch


 Just like term vectors before it, it has a O(n) seek-by-term. 
 But this one also advertises a seek-by-ord, only this is also O(n).
 This could cause e.g. checkindex to be very slow, because if termsenum 
 supports ord it does a bunch of seeking tests. (Another solution would be to 
 leave it, and add a boolean so checkindex never does seeking tests for term 
 vectors, only real fields).
 However, I think its also kinda a trap, in my opinion if seek-by-ord is 
 supported anywhere, you kinda expect it to be faster than linear time...?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5119) Managed schema problems after adding fields via Schema Rest API

2013-08-06 Thread NilsK (JIRA)
NilsK created SOLR-5119:
---

 Summary: Managed schema problems after adding fields via Schema 
Rest API 
 Key: SOLR-5119
 URL: https://issues.apache.org/jira/browse/SOLR-5119
 Project: Solr
  Issue Type: Bug
  Components: Schema and Analysis
Affects Versions: 4.4
Reporter: NilsK
Priority: Critical


Steps to reproduce:

1. enable managed schema in example/solr/collection1/conf/solrconfig.xml

2. upload that config
{code}sh example/cloud-scripts/zkcli.sh -z localhost:8575 -cmd upconfig -d 
example/solr/collection1/conf/ -n myconfig{code}

3. create a new collection 
{code}curl 
http://localhost:8983/solr/admin/collections?action=CREATEname=mycollectionnumShards=1replicationFactor=1collection.configName=myconfig{code}

4. add some fields
{code}curl http://localhost:8983/solr/mycollection/schema/fields -X POST -H 
'Content-type:application/json' --data-binary '[
{
  name: my_field,
  type: string,
  stored: true,
  indexed: true
},
{
  name: my_field2,
  type: string,
  stored: true,
  indexed: true
}
]'{code}

5. *Problem 1*: http://localhost:8983/solr/#/mycollection_shard1_replica1/schema

{code}
?xml version=1.0 encoding=UTF-8?
response
lst name=responseHeaderint name=status404/intint 
name=QTime2/int/lstlst name=errorstr name=msgCan not find: 
/configs/myconfig/null/strint name=code404/int/lst
/response
{code}

6. *Problem 2*: 
http://localhost:8983/solr/admin/collections?action=RELOADname=mycollection

{code}
response
lst name=responseHeaderint name=status0/intint 
name=QTime845/int/lstlst name=failurestr 
name=10.147.252.2:8983_solrorg.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Server
 at http://10.147.252.2:8983/solr returned non ok status:500, message:Server 
Error/str/lst
/response
{code}

7. when restarting Solr, both 5 and 6 work 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5119) Managed schema problems after adding fields via Schema Rest API

2013-08-06 Thread NilsK (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

NilsK updated SOLR-5119:


Description: 
After adding fields with the Schema API the schema cannot be shown on the Admin 
UI anymore and reloading the Collection/Core throws an NullPointerException. 
The schema itself seems to work.


Steps to reproduce:

1. enable managed schema in example/solr/collection1/conf/solrconfig.xml

2. upload that config
{code}sh example/cloud-scripts/zkcli.sh -z localhost:8575 -cmd upconfig -d 
example/solr/collection1/conf/ -n myconfig{code}

3. create a new collection 
{code}curl 
http://localhost:8983/solr/admin/collections?action=CREATEname=mycollectionnumShards=1replicationFactor=1collection.configName=myconfig{code}

4. add some fields
{code}curl http://localhost:8983/solr/mycollection/schema/fields -X POST -H 
'Content-type:application/json' --data-binary '[
{
  name: my_field,
  type: string,
  stored: true,
  indexed: true
},
{
  name: my_field2,
  type: string,
  stored: true,
  indexed: true
}
]'{code}

5. *Problem 1*: http://localhost:8983/solr/#/mycollection_shard1_replica1/schema

{code}
?xml version=1.0 encoding=UTF-8?
response
lst name=responseHeaderint name=status404/intint 
name=QTime2/int/lstlst name=errorstr name=msgCan not find: 
/configs/myconfig/null/strint name=code404/int/lst
/response
{code}

6. *Problem 2*: 
http://localhost:8983/solr/admin/collections?action=RELOADname=mycollection

{code}
response
lst name=responseHeaderint name=status0/intint 
name=QTime845/int/lstlst name=failurestr 
name=10.147.252.2:8983_solrorg.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Server
 at http://10.147.252.2:8983/solr returned non ok status:500, message:Server 
Error/str/lst
/response
{code}

7. when restarting Solr, both 5 and 6 work 

  was:
Steps to reproduce:

1. enable managed schema in example/solr/collection1/conf/solrconfig.xml

2. upload that config
{code}sh example/cloud-scripts/zkcli.sh -z localhost:8575 -cmd upconfig -d 
example/solr/collection1/conf/ -n myconfig{code}

3. create a new collection 
{code}curl 
http://localhost:8983/solr/admin/collections?action=CREATEname=mycollectionnumShards=1replicationFactor=1collection.configName=myconfig{code}

4. add some fields
{code}curl http://localhost:8983/solr/mycollection/schema/fields -X POST -H 
'Content-type:application/json' --data-binary '[
{
  name: my_field,
  type: string,
  stored: true,
  indexed: true
},
{
  name: my_field2,
  type: string,
  stored: true,
  indexed: true
}
]'{code}

5. *Problem 1*: http://localhost:8983/solr/#/mycollection_shard1_replica1/schema

{code}
?xml version=1.0 encoding=UTF-8?
response
lst name=responseHeaderint name=status404/intint 
name=QTime2/int/lstlst name=errorstr name=msgCan not find: 
/configs/myconfig/null/strint name=code404/int/lst
/response
{code}

6. *Problem 2*: 
http://localhost:8983/solr/admin/collections?action=RELOADname=mycollection

{code}
response
lst name=responseHeaderint name=status0/intint 
name=QTime845/int/lstlst name=failurestr 
name=10.147.252.2:8983_solrorg.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:Server
 at http://10.147.252.2:8983/solr returned non ok status:500, message:Server 
Error/str/lst
/response
{code}

7. when restarting Solr, both 5 and 6 work 


 Managed schema problems after adding fields via Schema Rest API 
 

 Key: SOLR-5119
 URL: https://issues.apache.org/jira/browse/SOLR-5119
 Project: Solr
  Issue Type: Bug
  Components: Schema and Analysis
Affects Versions: 4.4
Reporter: NilsK
Priority: Critical

 After adding fields with the Schema API the schema cannot be shown on the 
 Admin UI anymore and reloading the Collection/Core throws an 
 NullPointerException. The schema itself seems to work.
 Steps to reproduce:
 1. enable managed schema in example/solr/collection1/conf/solrconfig.xml
 2. upload that config
 {code}sh example/cloud-scripts/zkcli.sh -z localhost:8575 -cmd upconfig -d 
 example/solr/collection1/conf/ -n myconfig{code}
 3. create a new collection 
 {code}curl 
 http://localhost:8983/solr/admin/collections?action=CREATEname=mycollectionnumShards=1replicationFactor=1collection.configName=myconfig{code}
 4. add some fields
 {code}curl http://localhost:8983/solr/mycollection/schema/fields -X POST -H 
 'Content-type:application/json' --data-binary '[
 {
   name: my_field,
   type: string,
   stored: true,
   indexed: true
 },
 {
   name: my_field2,
   type: string,
   stored: true,
   indexed: true
 }
 ]'{code}
 5. *Problem 1*: 
 http://localhost:8983/solr/#/mycollection_shard1_replica1/schema
 {code}
 ?xml version=1.0 encoding=UTF-8?
 response
 lst name=responseHeaderint name=status404/intint 
 name=QTime2/int/lstlst name=errorstr name=msgCan not find: 
 

[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730983#comment-13730983
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1511044 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1511044 ]

SOLR-4952: ConvertedLegacyTest shouldn't assuming an absolute score now that we 
are randomizing merge policy settings, term stats may fluctuate

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man

 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test
 * try figure out a way for contrib tests to easily include the same file 
 and/or apply the same rules as above

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Problem using TrecTopicsReader()

2013-08-06 Thread Abhishek Gupta
Thanks!! Andi :)


On Tue, Aug 6, 2013 at 5:17 PM, Andi Vajda va...@apache.org wrote:


 On Aug 6, 2013, at 10:36, Abhishek Gupta abhi.bansa...@gmail.com wrote:

  Hi,
  I have a problem using TrecTopicsReader which I have already asked on
  #lucene IRC but got no reply till yesterday. I also post it on SO(
  http://stackoverflow.com/questions/18054649/bufferedreader-in-pylucne).
 
  The problem is completely explained on SO.

 In the PyLucene Makefile find the jcc invocation and add
 java.io.BufferedReader to the long command line (don't forget the ending \
 as needed) and rebuild PyLucene.

 Andi..

 
  Please Help!!
 
  Thanking You
  Abhishek Gupta,
  9416106204, 9624799165




-- 
Abhishek Gupta,
897876422, 9416106204, 9624799165


[jira] [Commented] (SOLR-5081) Highly parallel document insertion hangs SolrCloud

2013-08-06 Thread Yago Riveiro (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5081?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730995#comment-13730995
 ] 

Yago Riveiro commented on SOLR-5081:


I have this problem too, but in my case, Solr hangs and I can done more 
insertions without restart the nodes.

I do the insertion using a culr post in json format like:

curl http://127.0.0.1:8983/solr/collection/update --data-binary @data -H 
'Content-type:application/json'

 Highly parallel document insertion hangs SolrCloud
 --

 Key: SOLR-5081
 URL: https://issues.apache.org/jira/browse/SOLR-5081
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Affects Versions: 4.3.1
Reporter: Mike Schrag
 Attachments: threads.txt


 If I do a highly parallel document load using a Hadoop cluster into an 18 
 node solrcloud cluster, I can deadlock solr every time.
 The ulimits on the nodes are:
 core file size  (blocks, -c) 0
 data seg size   (kbytes, -d) unlimited
 scheduling priority (-e) 0
 file size   (blocks, -f) unlimited
 pending signals (-i) 1031181
 max locked memory   (kbytes, -l) unlimited
 max memory size (kbytes, -m) unlimited
 open files  (-n) 32768
 pipe size(512 bytes, -p) 8
 POSIX message queues (bytes, -q) 819200
 real-time priority  (-r) 0
 stack size  (kbytes, -s) 10240
 cpu time   (seconds, -t) unlimited
 max user processes  (-u) 515590
 virtual memory  (kbytes, -v) unlimited
 file locks  (-x) unlimited
 The open file count is only around 4000 when this happens.
 If I bounce all the servers, things start working again, which makes me think 
 this is Solr and not ZK.
 I'll attach the stack trace from one of the servers.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730993#comment-13730993
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1511046 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1511046 ]

SOLR-4952: ConvertedLegacyTest shouldn't assuming an absolute score now that we 
are randomizing merge policy settings, term stats may fluctuate (merge r1511044)

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man

 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test
 * try figure out a way for contrib tests to easily include the same file 
 and/or apply the same rules as above

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5107) LukeRequestHandler throws NullPointerException when numTerms=0

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731054#comment-13731054
 ] 

ASF subversion and git services commented on SOLR-5107:
---

Commit 1511064 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1511064 ]

SOLR-5107: Fixed NPE when using numTerms=0 in LukeRequestHandler

 LukeRequestHandler throws NullPointerException when numTerms=0
 --

 Key: SOLR-5107
 URL: https://issues.apache.org/jira/browse/SOLR-5107
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Ahmet Arslan
Assignee: Hoss Man
Priority: Minor
 Attachments: SOLR-5107.patch, SOLR-5107.patch


 Defaults example 
 http://localhost:8983/solr/collection1/admin/luke?fl=catnumTerms=0 yields 
 {code}
 ERROR org.apache.solr.core.SolrCore  – java.lang.NullPointerException
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.getDetailedFieldInfo(LukeRequestHandler.java:610)
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.getIndexedFieldsInfo(LukeRequestHandler.java:378)
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:160)
   at 
 org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
   at org.apache.solr.core.SolrCore.execute(SolrCore.java:1845)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:666)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:369)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:158)
   at 
 org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419)
   at 
 org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455)
   at 
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
   at 
 org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557)
   at 
 org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
   at 
 org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075)
   at 
 org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384)
   at 
 org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
   at 
 org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009)
   at 
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135)
   at 
 org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255)
   at 
 org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154)
   at 
 org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116)
   at org.eclipse.jetty.server.Server.handle(Server.java:368)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489)
   at 
 org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:53)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:942)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1004)
   at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:640)
   at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235)
   at 
 org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:72)
   at 
 org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:264)
   at 
 org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:608)
   at 
 org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:543)
   at java.lang.Thread.run(Thread.java:722)
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-5107) LukeRequestHandler throws NullPointerException when numTerms=0

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731083#comment-13731083
 ] 

ASF subversion and git services commented on SOLR-5107:
---

Commit 1511069 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1511069 ]

SOLR-5107: Fixed NPE when using numTerms=0 in LukeRequestHandler (merge 
r1511064)

 LukeRequestHandler throws NullPointerException when numTerms=0
 --

 Key: SOLR-5107
 URL: https://issues.apache.org/jira/browse/SOLR-5107
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Ahmet Arslan
Assignee: Hoss Man
Priority: Minor
 Attachments: SOLR-5107.patch, SOLR-5107.patch


 Defaults example 
 http://localhost:8983/solr/collection1/admin/luke?fl=catnumTerms=0 yields 
 {code}
 ERROR org.apache.solr.core.SolrCore  – java.lang.NullPointerException
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.getDetailedFieldInfo(LukeRequestHandler.java:610)
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.getIndexedFieldsInfo(LukeRequestHandler.java:378)
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:160)
   at 
 org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
   at org.apache.solr.core.SolrCore.execute(SolrCore.java:1845)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:666)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:369)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:158)
   at 
 org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419)
   at 
 org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455)
   at 
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
   at 
 org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557)
   at 
 org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
   at 
 org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075)
   at 
 org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384)
   at 
 org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
   at 
 org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009)
   at 
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135)
   at 
 org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255)
   at 
 org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154)
   at 
 org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116)
   at org.eclipse.jetty.server.Server.handle(Server.java:368)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489)
   at 
 org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:53)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:942)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1004)
   at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:640)
   at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235)
   at 
 org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:72)
   at 
 org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:264)
   at 
 org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:608)
   at 
 org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:543)
   at java.lang.Thread.run(Thread.java:722)
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-5107) LukeRequestHandler throws NullPointerException when numTerms=0

2013-08-06 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man resolved SOLR-5107.


   Resolution: Fixed
Fix Version/s: 5.0
   4.5

thanks Ahmet and Erick

 LukeRequestHandler throws NullPointerException when numTerms=0
 --

 Key: SOLR-5107
 URL: https://issues.apache.org/jira/browse/SOLR-5107
 Project: Solr
  Issue Type: Bug
Affects Versions: 4.4
Reporter: Ahmet Arslan
Assignee: Hoss Man
Priority: Minor
 Fix For: 4.5, 5.0

 Attachments: SOLR-5107.patch, SOLR-5107.patch


 Defaults example 
 http://localhost:8983/solr/collection1/admin/luke?fl=catnumTerms=0 yields 
 {code}
 ERROR org.apache.solr.core.SolrCore  – java.lang.NullPointerException
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.getDetailedFieldInfo(LukeRequestHandler.java:610)
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.getIndexedFieldsInfo(LukeRequestHandler.java:378)
   at 
 org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:160)
   at 
 org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
   at org.apache.solr.core.SolrCore.execute(SolrCore.java:1845)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:666)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:369)
   at 
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:158)
   at 
 org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419)
   at 
 org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455)
   at 
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
   at 
 org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557)
   at 
 org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
   at 
 org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075)
   at 
 org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384)
   at 
 org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
   at 
 org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009)
   at 
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135)
   at 
 org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255)
   at 
 org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154)
   at 
 org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116)
   at org.eclipse.jetty.server.Server.handle(Server.java:368)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489)
   at 
 org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:53)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:942)
   at 
 org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1004)
   at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:640)
   at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235)
   at 
 org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:72)
   at 
 org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:264)
   at 
 org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:608)
   at 
 org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:543)
   at java.lang.Thread.run(Thread.java:722)
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5113) CollectionsAPIDistributedZkTest fails all the time

2013-08-06 Thread Noble Paul (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5113?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Noble Paul updated SOLR-5113:
-

Attachment: SOLR-5113.patch

all tests pass . can somebody just confirm

 CollectionsAPIDistributedZkTest fails all the time
 --

 Key: SOLR-5113
 URL: https://issues.apache.org/jira/browse/SOLR-5113
 Project: Solr
  Issue Type: Bug
  Components: Tests
Affects Versions: 4.5, 5.0
Reporter: Uwe Schindler
Assignee: Noble Paul
Priority: Blocker
 Attachments: SOLR-5113.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-5120) Solrj Query response error with result number

2013-08-06 Thread JIRA
Łukasz Woźniczka created SOLR-5120:
--

 Summary: Solrj Query response error with result number 
 Key: SOLR-5120
 URL: https://issues.apache.org/jira/browse/SOLR-5120
 Project: Solr
  Issue Type: Bug
 Environment: linux, lubuntu, java version 1.7.0_13.
Reporter: Łukasz Woźniczka
Priority: Critical


This is my simple code : 

 QueryResponse qr;
try {
qr = fs.execute(solrServer);



System.out.println(QUERY RESPONSE :  + qr);

for (EntryString, Object r : qr.getResponse()) {
System.out.println(RESPONSE:  + r.getKey() +  -  + 
r.getValue());
}
SolrDocumentList dl = qr.getResults();
System.out.println(--RESULT SIZE:[  + dl.size() );


} catch (SolrServerException e) {
e.printStackTrace();
}

I am using solrj and solr-core version 4.4.0. And there is a bug probably in 
solrj in query result. I am creating one simple txt doc with content 'anna' and 
then i am restar solr and try to search this phrase. Nothing is found but this 
is my query response system out {numFound=1,start=1,docs=[]}.

So as you can see ther is info that numFound=1 but docs=[] -- is empty. Next i 
add another document with only one word 'anna' and then try search that string 
and this is sysout; 

{numFound=2,start=1,docs=[SolrDocument{file_id=9882, file_name=luk-search2.txt, 
file_create_user=-1, file_department=10, file_mime_type=text/plain, 
file_extension=.txt, file_parents_folder=[5021, 4781, 341, -20, -1], 
_version_=1442647024934584320}]}


So as you can see ther is numFound = 2 but only one document is listed. 




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Closed] (SOLR-5120) Solrj Query response error with result number

2013-08-06 Thread Shawn Heisey (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5120?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shawn Heisey closed SOLR-5120.
--

Resolution: Not A Problem

This should have been brought up on the solr-user mailing list, or perhaps the 
IRC channel.  Even if it were actually a bug, which it's not, it should have 
been discussed first.

http://lucene.apache.org/solr/discussion.html

Your response indicates that your query has a start parameter of 1.  If that 
had been set to zero or omitted entirely, you would have seen both results.

 Solrj Query response error with result number 
 --

 Key: SOLR-5120
 URL: https://issues.apache.org/jira/browse/SOLR-5120
 Project: Solr
  Issue Type: Bug
 Environment: linux, lubuntu, java version 1.7.0_13.
Reporter: Łukasz Woźniczka
Priority: Critical

 This is my simple code : 
  QueryResponse qr;
 try {
 qr = fs.execute(solrServer);
 System.out.println(QUERY RESPONSE :  + qr);
 for (EntryString, Object r : qr.getResponse()) {
 System.out.println(RESPONSE:  + r.getKey() +  -  + 
 r.getValue());
 }
 SolrDocumentList dl = qr.getResults();
 System.out.println(--RESULT SIZE:[  + dl.size() );
 } catch (SolrServerException e) {
 e.printStackTrace();
 }
 I am using solrj and solr-core version 4.4.0. And there is a bug probably in 
 solrj in query result. I am creating one simple txt doc with content 'anna' 
 and then i am restar solr and try to search this phrase. Nothing is found but 
 this is my query response system out {numFound=1,start=1,docs=[]}.
 So as you can see ther is info that numFound=1 but docs=[] -- is empty. Next 
 i add another document with only one word 'anna' and then try search that 
 string and this is sysout; 
 {numFound=2,start=1,docs=[SolrDocument{file_id=9882, 
 file_name=luk-search2.txt, file_create_user=-1, file_department=10, 
 file_mime_type=text/plain, file_extension=.txt, file_parents_folder=[5021, 
 4781, 341, -20, -1], _version_=1442647024934584320}]}
 So as you can see ther is numFound = 2 but only one document is listed. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Interesting failure scenario, SolrCloud and ZK nodes on different times

2013-08-06 Thread Erick Erickson
I've become aware of a situation I thought I'd pass along. A SolrCloud
installation had several ZK nodes that has very significantly offset times.
They were being hit with the ClusterState says we are the leader, but
locally we don't think we are error when nodes were recovering. Of course
whether this problem is now taken care of with recent Solr releases (I
haven't seen this go by the user's list for quite a while) I don't quite
know.

When the times were coordinated, many of the problems with recovery went
away. We're trying to reconstruct the scenario from memory, but it prompted
me to pass the incident in case it sparked any thoughts. Specifically, I
wonder if there's anything that comes to mind if the ZK nodes are
significantly out of synch with each other time-wise.

FWIW,
Erick


[jira] [Commented] (SOLR-5113) CollectionsAPIDistributedZkTest fails all the time

2013-08-06 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731195#comment-13731195
 ] 

Hoss Man commented on SOLR-5113:


Noble:

1) i notice in your patch that you've modified an assertion, and then commented 
it out -- is that a mistake, should that assertion be enabled?

2) 3 failures out of 10 iterations of the test with your patch applied and run 
using -Dtests.dups=10 ... but 3 is also the magic number of JVMs i use to run 
parallel tests, and all 3 failures looked like this...

{noformat}
   [junit4]   2 NOTE: reproduce with: ant test  
-Dtestcase=CollectionsAPIDistributedZkTest -Dtests.method=testDistribSearch 
-Dtests.seed=5E287730E2C75B9B -Dtests.slow=true -Dtests.locale=ar_QA 
-Dtests.timezone=America/Noronha -Dtests.file.encoding=UTF-8
   [junit4] ERROR116s J1 | 
CollectionsAPIDistributedZkTest.testDistribSearch 
   [junit4] Throwable #1: 
org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException: 
collection already exists: implicitcoll0
   [junit4]at 
__randomizedtesting.SeedInfo.seed([5E287730E2C75B9B:DFCEF92895983BA7]:0)
   [junit4]at 
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:424)
   [junit4]at 
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180)
   [junit4]at 
org.apache.solr.client.solrj.impl.LBHttpSolrServer.request(LBHttpSolrServer.java:264)
   [junit4]at 
org.apache.solr.client.solrj.impl.CloudSolrServer.request(CloudSolrServer.java:318)
   [junit4]at 
org.apache.solr.cloud.AbstractFullDistribZkTestBase.createCollection(AbstractFullDistribZkTestBase.java:1537)
   [junit4]at 
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testCustomCollectionsAPI(CollectionsAPIDistributedZkTest.java:678)
   [junit4]at 
org.apache.solr.cloud.CollectionsAPIDistributedZkTest.doTest(CollectionsAPIDistributedZkTest.java:155)
   [junit4]at 
org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:835)
   [junit4]at java.lang.Thread.run(Thread.java:724)
{noformat}

...when i tried with -Dtests.jvms=4 i got 4 failures out of 10



 CollectionsAPIDistributedZkTest fails all the time
 --

 Key: SOLR-5113
 URL: https://issues.apache.org/jira/browse/SOLR-5113
 Project: Solr
  Issue Type: Bug
  Components: Tests
Affects Versions: 4.5, 5.0
Reporter: Uwe Schindler
Assignee: Noble Paul
Priority: Blocker
 Attachments: SOLR-5113.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731298#comment-13731298
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1511123 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1511123 ]

SOLR-4952: missed the repeater config used in TestReplicationHandler

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man

 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test
 * try figure out a way for contrib tests to easily include the same file 
 and/or apply the same rules as above

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731302#comment-13731302
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1511126 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1511126 ]

SOLR-4952: missed the repeater config used in TestReplicationHandler (merge 
r1511123)

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man

 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test
 * try figure out a way for contrib tests to easily include the same file 
 and/or apply the same rules as above

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Interesting failure scenario, SolrCloud and ZK nodes on different times

2013-08-06 Thread Shawn Heisey

On 8/6/2013 1:56 PM, Erick Erickson wrote:

I've become aware of a situation I thought I'd pass along. A SolrCloud
installation had several ZK nodes that has very significantly offset
times. They were being hit with the ClusterState says we are the
leader, but locally we don't think we are error when nodes were
recovering. Of course whether this problem is now taken care of with
recent Solr releases (I haven't seen this go by the user's list for
quite a while) I don't quite know.

When the times were coordinated, many of the problems with recovery went
away. We're trying to reconstruct the scenario from memory, but it
prompted me to pass the incident in case it sparked any thoughts.
Specifically, I wonder if there's anything that comes to mind if the ZK
nodes are significantly out of synch with each other time-wise.


Does this mean that ntp or other strict time synchronization is 
important for SolrCloud?  I strive for this anyway, just to ensure that 
when I'm researching log files between two machines that I can match 
things up properly.


Thanks,
Shawn


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731336#comment-13731336
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1511128 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1511128 ]

SOLR-4952: TestSolrDeletionPolicy1 can't directly use 
solrconfig.snippet.randomindexconfig.xml - but it can leverage the random sys 
props directly

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man

 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test
 * try figure out a way for contrib tests to easily include the same file 
 and/or apply the same rules as above

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Interesting failure scenario, SolrCloud and ZK nodes on different times

2013-08-06 Thread Chris Hostetter

:  When the times were coordinated, many of the problems with recovery went
:  away. We're trying to reconstruct the scenario from memory, but it
:  prompted me to pass the incident in case it sparked any thoughts.
:  Specifically, I wonder if there's anything that comes to mind if the ZK
:  nodes are significantly out of synch with each other time-wise.
: 
: Does this mean that ntp or other strict time synchronization is important for
: SolrCloud?  I strive for this anyway, just to ensure that when I'm researching
: log files between two machines that I can match things up properly.

I don't know if/how Solr/ZK is affected by having machines with clocks out 
of sync, but i do remember seeing discussions a while back about weird 
things happening ot ZK client apps *while* time adjustments are taking 
place to get back in sync. 

IIRC: as the local clock starts accelerating and jumping ahead in 
increments to correct itself with ntp, then those jumps can confuse the 
ZK code into thinking it's been waiting a lot longer then it really 
has for zk heartbeat (or whatever it's called) and it can trigger a 
timeout situation.


-Hoss

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731348#comment-13731348
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1511129 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1511129 ]

SOLR-4952: TestSolrDeletionPolicy1 can't directly use 
solrconfig.snippet.randomindexconfig.xml - but it can leverage the random sys 
props directly (merge r1511128)

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man

 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test
 * try figure out a way for contrib tests to easily include the same file 
 and/or apply the same rules as above

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731366#comment-13731366
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1511134 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1511134 ]

SOLR-4952: get SolrCoreCheckLockOnStartupTest using 
solrconfig.snippet.randomindexconfig.xml

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man

 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test
 * try figure out a way for contrib tests to easily include the same file 
 and/or apply the same rules as above

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731374#comment-13731374
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1511137 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1511137 ]

SOLR-4952: get SolrCoreCheckLockOnStartupTest using 
solrconfig.snippet.randomindexconfig.xml (merge r1511134)

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man

 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test
 * try figure out a way for contrib tests to easily include the same file 
 and/or apply the same rules as above

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4764) When using NRT, just init the reader from IndexWriter

2013-08-06 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4764?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir updated SOLR-4764:
--

Attachment: SOLR-4764_test.patch

here's a test.

 When using NRT, just init the reader from IndexWriter
 -

 Key: SOLR-4764
 URL: https://issues.apache.org/jira/browse/SOLR-4764
 Project: Solr
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: SOLR-4764_test.patch


 Spinoff from SOLR-4761
 Solr first opens a DirectoryReader from the directory, then later will pass 
 this to IW openIfChanged.
 I noticed this when i was confused that mergedsegmentwarmer doesn't appear to 
 work at first until after you've reopened...
 I'm not totally sure what the current behavior causes (does IW's pool reuse 
 segments from this passed-in external reader, or is this causing some 
 horrible doubling-up/inefficient stuff etc?). To some extent i think we 
 should change it even if its actually performant: I think its confusing.
 I think ideally we'd change IndexReaderFactory's method to take writer 
 instead of directory so that custom DirectoryReaders can still work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731386#comment-13731386
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1511138 from hoss...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1511138 ]

SOLR-4952: TestSolrDeletionPolicy2 can't directly use 
solrconfig.snippet.randomindexconfig.xml - but it can leverage the random sys 
props directly

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man

 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test
 * try figure out a way for contrib tests to easily include the same file 
 and/or apply the same rules as above

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4952) audit test configs to use solrconfig.snippet.randomindexconfig.xml in more tests

2013-08-06 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731388#comment-13731388
 ] 

ASF subversion and git services commented on SOLR-4952:
---

Commit 1511139 from hoss...@apache.org in branch 'dev/branches/branch_4x'
[ https://svn.apache.org/r1511139 ]

SOLR-4952: TestSolrDeletionPolicy2 can't directly use 
solrconfig.snippet.randomindexconfig.xml - but it can leverage the random sys 
props directly

 audit test configs to use solrconfig.snippet.randomindexconfig.xml in more 
 tests
 

 Key: SOLR-4952
 URL: https://issues.apache.org/jira/browse/SOLR-4952
 Project: Solr
  Issue Type: Sub-task
Reporter: Hoss Man
Assignee: Hoss Man

 in SOLR-4942 i updated every solrconfig.xml to either...
 * include solrconfig.snippet.randomindexconfig.xml where it was easy to do so
 * use the useCompoundFile sys prop if it already had an {{indexConfig}} 
 section, or if including the snippet wasn't going to be easy (ie: contrib 
 tests)
 As an improvment on this:
 * audit all core configs not already using 
 solrconfig.snippet.randomindexconfig.xml and either:
 ** make them use it, ignoring any previously unimportant explicit 
 incdexConfig settings
 ** make them use it, using explicit sys props to overwrite random values in 
 cases were explicit indexConfig values are important for test
 ** add a comment why it's not using the include snippet in cases where the 
 explicit parsing is part of hte test
 * try figure out a way for contrib tests to easily include the same file 
 and/or apply the same rules as above

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2894) Implement distributed pivot faceting

2013-08-06 Thread Andrew Muldowney (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2894?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731408#comment-13731408
 ] 

Andrew Muldowney commented on SOLR-2894:


I've found a small error which causes largely sharded (30+) data to spiral out 
of control on refinement requests.

I've fixed the error on a previous version of solr and I'll be forward porting 
it to my 4_4 build by tomorrow.

If you are having issues with complex string fields this should help.

 Implement distributed pivot faceting
 

 Key: SOLR-2894
 URL: https://issues.apache.org/jira/browse/SOLR-2894
 Project: Solr
  Issue Type: Improvement
Reporter: Erik Hatcher
 Fix For: 4.5

 Attachments: SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, 
 SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, 
 SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, 
 SOLR-2894.patch, SOLR-2894.patch, SOLR-2894-reworked.patch


 Following up on SOLR-792, pivot faceting currently only supports 
 undistributed mode.  Distributed pivot faceting needs to be implemented.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-2894) Implement distributed pivot faceting

2013-08-06 Thread Otis Gospodnetic (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-2894?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731433#comment-13731433
 ] 

Otis Gospodnetic commented on SOLR-2894:


[~andrew.muldowney] - I didn't have time to link issues, but [~joel.bernstein] 
is working on at least one issue that is, if I recall correctly, an alternative 
implementation of this

 Implement distributed pivot faceting
 

 Key: SOLR-2894
 URL: https://issues.apache.org/jira/browse/SOLR-2894
 Project: Solr
  Issue Type: Improvement
Reporter: Erik Hatcher
 Fix For: 4.5

 Attachments: SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, 
 SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, 
 SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, 
 SOLR-2894.patch, SOLR-2894.patch, SOLR-2894-reworked.patch


 Following up on SOLR-792, pivot faceting currently only supports 
 undistributed mode.  Distributed pivot faceting needs to be implemented.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-5114) stop tracking IndexCommit in replication handler

2013-08-06 Thread Yonik Seeley (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-5114?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yonik Seeley updated SOLR-5114:
---

Attachment: SOLR-5114.patch

Mostly working snapshot patch.
Trying to figure out what the current behavior is supposed to be in many 
different scenarios is difficult.  One test fails: testEmptyCommits...

 stop tracking IndexCommit in replication handler
 

 Key: SOLR-5114
 URL: https://issues.apache.org/jira/browse/SOLR-5114
 Project: Solr
  Issue Type: Improvement
Reporter: Yonik Seeley
Priority: Trivial
 Attachments: SOLR-5114.patch


 It seems like indexCommitPoint tries to track the latest index commit.  This 
 doesn't seem necessary and could lead to bugs if it's not tracked correctly.  
 We should always be able to ask the deletion policy for the latest commit 
 (including a fallback mechanism if no IW has yet been opened)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (SOLR-5114) stop tracking IndexCommit in replication handler

2013-08-06 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731453#comment-13731453
 ] 

Yonik Seeley edited comment on SOLR-5114 at 8/6/13 11:51 PM:
-

Mostly working snapshot patch.
Trying to figure out what the current behavior is supposed to be in many 
different scenarios is difficult.  One test fails: testEmptyCommits...

edit: oh, this patch also removes a race condition between getting an index 
commit and reserving it.

  was (Author: ysee...@gmail.com):
Mostly working snapshot patch.
Trying to figure out what the current behavior is supposed to be in many 
different scenarios is difficult.  One test fails: testEmptyCommits...
  
 stop tracking IndexCommit in replication handler
 

 Key: SOLR-5114
 URL: https://issues.apache.org/jira/browse/SOLR-5114
 Project: Solr
  Issue Type: Improvement
Reporter: Yonik Seeley
Priority: Trivial
 Attachments: SOLR-5114.patch


 It seems like indexCommitPoint tries to track the latest index commit.  This 
 doesn't seem necessary and could lead to bugs if it's not tracked correctly.  
 We should always be able to ask the deletion policy for the latest commit 
 (including a fallback mechanism if no IW has yet been opened)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4764) When using NRT, just init the reader from IndexWriter

2013-08-06 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4764?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir updated SOLR-4764:
--

Attachment: SOLR-4764.patch

Here is a patch. All tests pass.

A few notes:
* the init respects the existing 'reopenReaders' in solrconfig.xml, so if you 
have that set to false, it still does the same thing as before.
* IndexReaderFactory gets a newReader(IndexWriter, ..) so that NRT users can 
still customize the DirectoryReader being used.
* A few tests relied upon the fact that init didnt open an indexwriter: 
MultiCoreExampleTestBase always created a core container and opened cores, so 
this was moved to the embedded subclass which needs it. The ReplicationHandler 
test and the TestArbitraryIndexDir just set reopenReaders=false because they 
rely fundamentally on this 'readonly' case.


 When using NRT, just init the reader from IndexWriter
 -

 Key: SOLR-4764
 URL: https://issues.apache.org/jira/browse/SOLR-4764
 Project: Solr
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: SOLR-4764.patch, SOLR-4764_test.patch


 Spinoff from SOLR-4761
 Solr first opens a DirectoryReader from the directory, then later will pass 
 this to IW openIfChanged.
 I noticed this when i was confused that mergedsegmentwarmer doesn't appear to 
 work at first until after you've reopened...
 I'm not totally sure what the current behavior causes (does IW's pool reuse 
 segments from this passed-in external reader, or is this causing some 
 horrible doubling-up/inefficient stuff etc?). To some extent i think we 
 should change it even if its actually performant: I think its confusing.
 I think ideally we'd change IndexReaderFactory's method to take writer 
 instead of directory so that custom DirectoryReaders can still work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



OutOfMemoryError in TermContext.build

2013-08-06 Thread Gagandeep singh
Hi folks

I am running a mapreduce which generates More Like This for a list of
documents. For every source document, a set of important words is picked by
generating a TermQuery for a word and seeing the score of the query using
searcher.explain()

My index size on disk is 362MB and I have started solr with 4GB of memory.
However, i'm getting this exception which i'm not able to understand:

Caused by: java.lang.OutOfMemoryError: GC overhead limit exceeded
at
org.apache.lucene.codecs.BlockTreeTermsReader$FieldReader$SegmentTermsEnum$Frame.init(BlockTreeTermsReader.java:2266)
at
org.apache.lucene.codecs.BlockTreeTermsReader$FieldReader$SegmentTermsEnum.getFrame(BlockTreeTermsReader.java:1414)
at
org.apache.lucene.codecs.BlockTreeTermsReader$FieldReader$SegmentTermsEnum.pushFrame(BlockTreeTermsReader.java:1445)
at
org.apache.lucene.codecs.BlockTreeTermsReader$FieldReader$SegmentTermsEnum.seekExact(BlockTreeTermsReader.java:1658)
at org.apache.lucene.index.TermContext.build(TermContext.java:95)
at
org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:166)
at
org.apache.lucene.search.ConstantScoreQuery$ConstantWeight.init(ConstantScoreQuery.java:101)
at
org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:270)
at
com.bloomreach.rp.ScoreSquashingQuery.createWeight(ScoreSquashingQuery.java:46)
at
org.apache.lucene.search.BooleanQuery$BooleanWeight.init(BooleanQuery.java:183)
at
org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:384)
at
com.bloomreach.rp.ScoreSquashingQuery.createWeight(ScoreSquashingQuery.java:46)
at
org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:675)
at
org.apache.lucene.search.IndexSearcher.explain(IndexSearcher.java:643)
at
org.apache.solr.search.SolrIndexSearcher.explain(SolrIndexSearcher.java:2058)
at
com.bloomreach.rp.MLTQueryParser.computeScoreForWordOrPhrase(MLTQueryParser.java:321)

I tried digging into the code for TermQuery.createWeight(), the function
says that its caching the term lookup:
 public Weight createWeight(IndexSearcher searcher) throws IOException {
...
// make TermQuery single-pass if we don't have a PRTS or if the
context differs!
termState = TermContext.build(context, term, true); // cache term
lookups!


But how can this cause an OOM, i thought the index is already a map from
term to doc,field,position that is loaded in memory, why is this loading
anything extra in memory?

Thanks
Gagan


[jira] [Commented] (SOLR-5113) CollectionsAPIDistributedZkTest fails all the time

2013-08-06 Thread Noble Paul (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731579#comment-13731579
 ] 

Noble Paul commented on SOLR-5113:
--

Thanks Hoss
bq.should that assertion be enabled?

That assertion should be enabled. But I'm baffled when I see

{noformat}
  [junit4] ERROR   53.3s | CollectionsAPIDistributedZkTest.testDistribSearch 
   [junit4] Throwable #1: 
org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException: Server at 
http://127.0.0.1:55245/iv/y/implicitcoll0 returned non ok status:503, 
message:Service Unavailable
{noformat}

Considering that the previous statement inserted one doc into the same url. 

However I wanted to ensure that others are able to confirm if this much is 
passing

I see the tests failing when the test is the last one to run , and passing when 
I moved to the top. 

I'm wondering , how do the no:of JVMs affect the testitself


 CollectionsAPIDistributedZkTest fails all the time
 --

 Key: SOLR-5113
 URL: https://issues.apache.org/jira/browse/SOLR-5113
 Project: Solr
  Issue Type: Bug
  Components: Tests
Affects Versions: 4.5, 5.0
Reporter: Uwe Schindler
Assignee: Noble Paul
Priority: Blocker
 Attachments: SOLR-5113.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-trunk-Windows (32bit/jdk1.8.0-ea-b99) - Build # 3121 - Failure!

2013-08-06 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/3121/
Java: 32bit/jdk1.8.0-ea-b99 -server -XX:+UseG1GC

1 tests failed.
REGRESSION:  org.apache.lucene.index.TestIndexWriterOutOfFileDescriptors.test

Error Message:
unreferenced files: before delete: 
[_0_TestBloomFilteredLucene41Postings_0.doc, 
_0_TestBloomFilteredLucene41Postings_0.pos, _k.fdt, _k.fdx, _k.fnm, _k.nvd, 
_k.nvm, _k.si, _k.tvd, _k.tvx, _k_Lucene41WithOrds_0.doc, 
_k_Lucene41WithOrds_0.pos, _k_Lucene41WithOrds_0.tib, 
_k_Lucene41WithOrds_0.tii, _k_Lucene42_0.dvd, _k_Lucene42_0.dvm, 
_k_MockFixedIntBlock_0.doc, _k_MockFixedIntBlock_0.frq, 
_k_MockFixedIntBlock_0.pos, _k_MockFixedIntBlock_0.pyl, 
_k_MockFixedIntBlock_0.skp, _k_MockFixedIntBlock_0.tib, 
_k_MockFixedIntBlock_0.tii, _k_MockVariableIntBlock_0.doc, 
_k_MockVariableIntBlock_0.frq, _k_MockVariableIntBlock_0.pos, 
_k_MockVariableIntBlock_0.pyl, _k_MockVariableIntBlock_0.skp, 
_k_MockVariableIntBlock_0.tib, _k_MockVariableIntBlock_0.tii, 
_k_TestBloomFilteredLucene41Postings_0.blm, 
_k_TestBloomFilteredLucene41Postings_0.doc, 
_k_TestBloomFilteredLucene41Postings_0.pos, 
_k_TestBloomFilteredLucene41Postings_0.tim, 
_k_TestBloomFilteredLucene41Postings_0.tip, _l.fdt, _l.fdx, _l.fnm, _l.nvd, 
_l.nvm, _l.si, _l.tvd, _l.tvx, _l_Lucene41WithOrds_0.doc, 
_l_Lucene41WithOrds_0.pos, _l_Lucene41WithOrds_0.tib, 
_l_Lucene41WithOrds_0.tii, _l_Lucene42_0.dvd, _l_Lucene42_0.dvm, 
_l_MockFixedIntBlock_0.doc, _l_MockFixedIntBlock_0.frq, 
_l_MockFixedIntBlock_0.pos, _l_MockFixedIntBlock_0.pyl, 
_l_MockFixedIntBlock_0.skp, _l_MockFixedIntBlock_0.tib, 
_l_MockFixedIntBlock_0.tii, _l_MockVariableIntBlock_0.doc, 
_l_MockVariableIntBlock_0.frq, _l_MockVariableIntBlock_0.pos, 
_l_MockVariableIntBlock_0.pyl, _l_MockVariableIntBlock_0.skp, 
_l_MockVariableIntBlock_0.tib, _l_MockVariableIntBlock_0.tii, 
_l_TestBloomFilteredLucene41Postings_0.blm, 
_l_TestBloomFilteredLucene41Postings_0.doc, 
_l_TestBloomFilteredLucene41Postings_0.pos, 
_l_TestBloomFilteredLucene41Postings_0.tim, 
_l_TestBloomFilteredLucene41Postings_0.tip, _m.cfe, _m.cfs, _m.si, _n.cfe, 
_n.cfs, _n.si, _o.fdt, _o.fdx, _o.fnm, _o.nvd, _o.nvm, _o.si, _o.tvd, _o.tvx, 
_o_Lucene41WithOrds_0.doc, _o_Lucene41WithOrds_0.pos, 
_o_Lucene41WithOrds_0.tib, _o_Lucene41WithOrds_0.tii, _o_Lucene42_0.dvd, 
_o_Lucene42_0.dvm, _o_MockFixedIntBlock_0.doc, _o_MockFixedIntBlock_0.frq, 
_o_MockFixedIntBlock_0.pos, _o_MockFixedIntBlock_0.pyl, 
_o_MockFixedIntBlock_0.skp, _o_MockFixedIntBlock_0.tib, 
_o_MockFixedIntBlock_0.tii, _o_MockVariableIntBlock_0.doc, 
_o_MockVariableIntBlock_0.frq, _o_MockVariableIntBlock_0.pos, 
_o_MockVariableIntBlock_0.pyl, _o_MockVariableIntBlock_0.skp, 
_o_MockVariableIntBlock_0.tib, _o_MockVariableIntBlock_0.tii, 
_o_TestBloomFilteredLucene41Postings_0.blm, 
_o_TestBloomFilteredLucene41Postings_0.doc, 
_o_TestBloomFilteredLucene41Postings_0.pos, 
_o_TestBloomFilteredLucene41Postings_0.tim, 
_o_TestBloomFilteredLucene41Postings_0.tip, _q.fdt, _q.fdx, _q.fnm, _q.nvd, 
_q.nvm, _q.si, _q.tvd, _q.tvx, _q_Lucene41WithOrds_0.doc, 
_q_Lucene41WithOrds_0.pos, _q_Lucene41WithOrds_0.tib, 
_q_Lucene41WithOrds_0.tii, _q_Lucene42_0.dvd, _q_Lucene42_0.dvm, 
_q_MockFixedIntBlock_0.doc, _q_MockFixedIntBlock_0.frq, 
_q_MockFixedIntBlock_0.pos, _q_MockFixedIntBlock_0.pyl, 
_q_MockFixedIntBlock_0.skp, _q_MockFixedIntBlock_0.tib, 
_q_MockFixedIntBlock_0.tii, _q_MockVariableIntBlock_0.doc, 
_q_MockVariableIntBlock_0.frq, _q_MockVariableIntBlock_0.pos, 
_q_MockVariableIntBlock_0.pyl, _q_MockVariableIntBlock_0.skp, 
_q_MockVariableIntBlock_0.tib, _q_MockVariableIntBlock_0.tii, 
_q_TestBloomFilteredLucene41Postings_0.blm, 
_q_TestBloomFilteredLucene41Postings_0.doc, 
_q_TestBloomFilteredLucene41Postings_0.pos, 
_q_TestBloomFilteredLucene41Postings_0.tim, 
_q_TestBloomFilteredLucene41Postings_0.tip, _r.fdt, _r.fdx, _r.fnm, _r.nvd, 
_r.nvm, _r.si, _r.tvd, _r.tvx, _r_Lucene41WithOrds_0.doc, 
_r_Lucene41WithOrds_0.pay, _r_Lucene41WithOrds_0.pos, 
_r_Lucene41WithOrds_0.tib, _r_Lucene41WithOrds_0.tii, _r_Lucene42_0.dvd, 
_r_Lucene42_0.dvm, _r_MockFixedIntBlock_0.doc, _r_MockFixedIntBlock_0.frq, 
_r_MockFixedIntBlock_0.pos, _r_MockFixedIntBlock_0.pyl, 
_r_MockFixedIntBlock_0.skp, _r_MockFixedIntBlock_0.tib, 
_r_MockFixedIntBlock_0.tii, _r_MockVariableIntBlock_0.doc, 
_r_MockVariableIntBlock_0.frq, _r_MockVariableIntBlock_0.pos, 
_r_MockVariableIntBlock_0.pyl, _r_MockVariableIntBlock_0.skp, 
_r_MockVariableIntBlock_0.tib, _r_MockVariableIntBlock_0.tii, 
_r_TestBloomFilteredLucene41Postings_0.blm, 
_r_TestBloomFilteredLucene41Postings_0.doc, 
_r_TestBloomFilteredLucene41Postings_0.pay, 
_r_TestBloomFilteredLucene41Postings_0.pos, 
_r_TestBloomFilteredLucene41Postings_0.tim, 
_r_TestBloomFilteredLucene41Postings_0.tip, _s.fdt, _s.fdx, _s.fnm, _s.nvd, 
_s.nvm, _s.si, _s.tvd, _s.tvx, _s_Lucene41WithOrds_0.doc, 
_s_Lucene41WithOrds_0.pos, _s_Lucene41WithOrds_0.tib, 
_s_Lucene41WithOrds_0.tii, _s_Lucene42_0.dvd, _s_Lucene42_0.dvm,