Re: short-circuit OR operator in lucene/solr
Short answer, no - it has zero sense. But after some thinking, it can make some sense, potentially. DisjunctionSumScorer holds child scorers semi-ordered in a binary heap. Hypothetically inequality can be enforced at that heap, but heap might not work anymore for such alignment. Hence, instead of heap TreeSet can be used for experiment. fwiw, it's a dev list question. On Mon, Jul 22, 2013 at 4:48 AM, Deepak Konidena deepakk...@gmail.comwrote: I understand that lucene's AND (), OR (||) and NOT (!) operators are shorthands for REQUIRED, OPTIONAL and EXCLUDE respectively, which is why one can't treat them as boolean operators (adhering to boolean algebra). I have been trying to construct a simple OR expression, as follows q = +(field1:value1 OR field2:value2) with a match on either field1 or field2. But since the OR is merely an optional, documents where both field1:value1 and field2:value2 are matched, the query returns a score resulting in a match on both the clauses. How do I enforce short-circuiting in this context? In other words, how to implement short-circuiting as in boolean algebra where an expression A || B || C returns true if A is true without even looking into whether B or C could be true. -Deepak -- Sincerely yours Mikhail Khludnev Principal Engineer, Grid Dynamics http://www.griddynamics.com mkhlud...@griddynamics.com
[jira] [Comment Edited] (SOLR-3076) Solr(Cloud) should support block joins
[ https://issues.apache.org/jira/browse/SOLR-3076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13713486#comment-13713486 ] Mikhail Khludnev edited comment on SOLR-3076 at 7/22/13 8:18 AM: - [~ysee...@gmail.com] I thought address this question at SOLR-4836, but it would be great if you cover it altogether. Pls confirm my understanding, if uniqKey is specified, for every block children obtain value for \_root\_ field, from parent's uniqKey field. Hence \_root\_ field is always used for IW.updateDocs call. btw, uber compact codec can be implemented for that \_root\_ field that somehow interfere with Vadim's idea. [~vkirilchuk] -introducing this separate case, blows IndexWriter code. Presumably you can try to introduce a facility to update not by term as now, but for some kind of generic query. fwiw, parent bitset is not really exists in IndexWriter- see below was (Author: mkhludnev): [~ysee...@gmail.com] I thought address this question at SOLR-4836, but it would be great if you cover it altogether. Pls confirm my understanding, if uniqKey is specified, for every block children obtain value for \_root\_ field, from parent's uniqKey field. Hence \_root\_ field is always used for IW.updateDocs call. btw, uber compact codec can be implemented for that \_root\_ field that somehow interfere with Vadim's idea. [~vkirilchuk] introducing this separate case, blows IndexWriter code. Presumably you can try to introduce a facility to update not by term as now, but for some kind of generic query. fwiw, parent bitset is not really exists in IndexWriter Solr(Cloud) should support block joins -- Key: SOLR-3076 URL: https://issues.apache.org/jira/browse/SOLR-3076 Project: Solr Issue Type: New Feature Reporter: Grant Ingersoll Assignee: Yonik Seeley Fix For: 5.0, 4.4 Attachments: 27M-singlesegment-histogram.png, 27M-singlesegment.png, bjq-vs-filters-backward-disi.patch, bjq-vs-filters-illegal-state.patch, child-bjqparser.patch, dih-3076.patch, dih-config.xml, parent-bjq-qparser.patch, parent-bjq-qparser.patch, Screen Shot 2012-07-17 at 1.12.11 AM.png, SOLR-3076-childDocs.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-7036-childDocs-solr-fork-trunk-patched, solrconf-bjq-erschema-snippet.xml, solrconfig.xml.patch, tochild-bjq-filtered-search-fix.patch Lucene has the ability to do block joins, we should add it to Solr. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3076) Solr(Cloud) should support block joins
[ https://issues.apache.org/jira/browse/SOLR-3076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715007#comment-13715007 ] Mikhail Khludnev commented on SOLR-3076: bq. delete everything between two parents in the bitset Indeed! [~ysee...@gmail.com] we don't need \_root\_ if we can submit two queries for deletion: ToChild(parentid:foo) and TQ(parentid:foo)! Second great idea from [~vkirilchuk] and Alexey is that specific _FullBlockQuery_ which matches parent and all its' children by the given parentid. Solr(Cloud) should support block joins -- Key: SOLR-3076 URL: https://issues.apache.org/jira/browse/SOLR-3076 Project: Solr Issue Type: New Feature Reporter: Grant Ingersoll Assignee: Yonik Seeley Fix For: 5.0, 4.4 Attachments: 27M-singlesegment-histogram.png, 27M-singlesegment.png, bjq-vs-filters-backward-disi.patch, bjq-vs-filters-illegal-state.patch, child-bjqparser.patch, dih-3076.patch, dih-config.xml, parent-bjq-qparser.patch, parent-bjq-qparser.patch, Screen Shot 2012-07-17 at 1.12.11 AM.png, SOLR-3076-childDocs.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-7036-childDocs-solr-fork-trunk-patched, solrconf-bjq-erschema-snippet.xml, solrconfig.xml.patch, tochild-bjq-filtered-search-fix.patch Lucene has the ability to do block joins, we should add it to Solr. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-3076) Solr(Cloud) should support block joins
[ https://issues.apache.org/jira/browse/SOLR-3076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715007#comment-13715007 ] Mikhail Khludnev edited comment on SOLR-3076 at 7/22/13 8:40 AM: - bq. delete everything between two parents in the bitset Indeed! [~ysee...@gmail.com] we don't need \_root\_ if we can submit two queries for deletion: ToChild(parentid:foo) and TQ(parentid:foo)! Second great idea from [~vkirilchuk] and [~aaleev] is that specific _FullBlockQuery_ which matches parent and all its' children by the given parentid. was (Author: mkhludnev): bq. delete everything between two parents in the bitset Indeed! [~ysee...@gmail.com] we don't need \_root\_ if we can submit two queries for deletion: ToChild(parentid:foo) and TQ(parentid:foo)! Second great idea from [~vkirilchuk] and Alexey is that specific _FullBlockQuery_ which matches parent and all its' children by the given parentid. Solr(Cloud) should support block joins -- Key: SOLR-3076 URL: https://issues.apache.org/jira/browse/SOLR-3076 Project: Solr Issue Type: New Feature Reporter: Grant Ingersoll Assignee: Yonik Seeley Fix For: 5.0, 4.4 Attachments: 27M-singlesegment-histogram.png, 27M-singlesegment.png, bjq-vs-filters-backward-disi.patch, bjq-vs-filters-illegal-state.patch, child-bjqparser.patch, dih-3076.patch, dih-config.xml, parent-bjq-qparser.patch, parent-bjq-qparser.patch, Screen Shot 2012-07-17 at 1.12.11 AM.png, SOLR-3076-childDocs.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-3076.patch, SOLR-7036-childDocs-solr-fork-trunk-patched, solrconf-bjq-erschema-snippet.xml, solrconfig.xml.patch, tochild-bjq-filtered-search-fix.patch Lucene has the ability to do block joins, we should add it to Solr. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [VOTE] Release 4.4 RC1
+1 Release candidate looks good and smoke tester ran successfully. -- Adrien - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-5054) My private query component does not work in solrcloud
che created SOLR-5054: - Summary: My private query component does not work in solrcloud Key: SOLR-5054 URL: https://issues.apache.org/jira/browse/SOLR-5054 Project: Solr Issue Type: Bug Components: SolrCloud Affects Versions: 4.3.1 Environment: Linux x64 Reporter: che I user my HQ.jar in one solr is ok every query is entry in my HQ. But I make a solrcloud with three shards. The HQ.jar is not work.The log is not print .so I think it is not entry in. Do you know why ,thank you any help. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Contentions observed in lucene execution
Thanks Mike, This is running on Hotspots VM (on unix) the JVM has 24 GB of max heap. Moving to Lucene 4.X is not an easy option because for us Lucene has shipped as part of the Sterling 9.0 product if we upgrade part of the product we would have to give away support. One of my coworkers had the same doubt (that we are running out of heap - although I am not so sure because the warnings appear in the logs after 100's of stuck threads) and so we plan to bump up the heap to a max of 34 and then watch closely. I'll post if that does not work, thanks for taking a look. - Ramesh. -- View this message in context: http://lucene.472066.n3.nabble.com/Contentions-observed-in-lucene-execution-tp4078796p4079410.html Sent from the Lucene - Java Developer mailing list archive at Nabble.com. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
RE: Contentions observed in lucene execution
Hi, you should not use too much heap space for the Java VM because Lucene relies heavily on the file system cache. If your machine has say 32 Gigs of RAM don't use more than 1/4 (8 DB) as heap for the Java application. Ideally only use so much heap + some overhead, that you get no OOMs (this depends on your index size and application). Lucene generally uses only few heap space unless very huge field caches are used. For the concurrency problem, as Mike already said, try using MMapDirectory (which is supported on Lucene 2.4, too - but not yet so efficient and does not support unmapping) instead of FSDirectory. For more information, see http://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html RandomAccessFile (which is used by FSDirectory) is not behaving well in multi-threaded environments, because the underlying file descriptor can only be used single-threaded. Uwe - Uwe Schindler H.-H.-Meier-Allee 63, D-28213 Bremen http://www.thetaphi.de eMail: u...@thetaphi.de -Original Message- From: RameshIyerV [mailto:rameshiy...@hotmail.com] Sent: Monday, July 22, 2013 11:15 AM To: dev@lucene.apache.org Subject: Re: Contentions observed in lucene execution Thanks Mike, This is running on Hotspots VM (on unix) the JVM has 24 GB of max heap. Moving to Lucene 4.X is not an easy option because for us Lucene has shipped as part of the Sterling 9.0 product if we upgrade part of the product we would have to give away support. One of my coworkers had the same doubt (that we are running out of heap - although I am not so sure because the warnings appear in the logs after 100's of stuck threads) and so we plan to bump up the heap to a max of 34 and then watch closely. I'll post if that does not work, thanks for taking a look. - Ramesh. -- View this message in context: http://lucene.472066.n3.nabble.com/Contentions-observed-in-lucene- execution-tp4078796p4079410.html Sent from the Lucene - Java Developer mailing list archive at Nabble.com. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Reopened] (LUCENE-4734) FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight
[ https://issues.apache.org/jira/browse/LUCENE-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Adrien Grand reopened LUCENE-4734: -- FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight Key: LUCENE-4734 URL: https://issues.apache.org/jira/browse/LUCENE-4734 Project: Lucene - Core Issue Type: Bug Components: modules/highlighter Affects Versions: 4.0, 4.1, 5.0 Reporter: Ryan Lauck Assignee: Adrien Grand Labels: fastvectorhighlighter, highlighter Fix For: 4.4 Attachments: LUCENE-4734-2.patch, lucene-4734.patch, LUCENE-4734.patch If a proximity phrase query overlaps with any other query term it will not be highlighted. Example Text: A B C D E F G Example Queries: B E~10 D (D will be highlighted instead of B C D E) B E~10 C F~10 (nothing will be highlighted) This can be traced to the FieldPhraseList constructor's inner while loop. From the first example query, the first TermInfo popped off the stack will be B. The second TermInfo will be D which will not be found in the submap for B E~10 and will trigger a failed match. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-4734) FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight
[ https://issues.apache.org/jira/browse/LUCENE-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Adrien Grand updated LUCENE-4734: - Attachment: LUCENE-4734-2.patch The approach I used can be memory-intensive when there are many positions that have several terms, here is a fix. FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight Key: LUCENE-4734 URL: https://issues.apache.org/jira/browse/LUCENE-4734 Project: Lucene - Core Issue Type: Bug Components: modules/highlighter Affects Versions: 4.0, 4.1, 5.0 Reporter: Ryan Lauck Assignee: Adrien Grand Labels: fastvectorhighlighter, highlighter Fix For: 4.4 Attachments: LUCENE-4734-2.patch, lucene-4734.patch, LUCENE-4734.patch If a proximity phrase query overlaps with any other query term it will not be highlighted. Example Text: A B C D E F G Example Queries: B E~10 D (D will be highlighted instead of B C D E) B E~10 C F~10 (nothing will be highlighted) This can be traced to the FieldPhraseList constructor's inner while loop. From the first example query, the first TermInfo popped off the stack will be B. The second TermInfo will be D which will not be found in the submap for B E~10 and will trigger a failed match. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
RE: Contentions observed in lucene execution
Uwe, we do not have the flexibility to change the implementation to user MMapDirectory, lucene implementation is core to the Sterling 9.0 product (and ships with it). On heap, the box has 250+ GB, our application's (2) is using close to 68. Also the application internally uses ehcache to cache the lucene index files. And we have configured the ehcache to store all the elements in memory, we did this to minimize disk IO. However it still seems to be doing the RandomAccessFile (used by FSDirectory), do you know what it is attempting to ready? -- View this message in context: http://lucene.472066.n3.nabble.com/Contentions-observed-in-lucene-execution-tp4078796p4079428.html Sent from the Lucene - Java Developer mailing list archive at Nabble.com. - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-5055) Using solr.SearchHandler with Apache Tomcat
Christian Lutz created SOLR-5055: Summary: Using solr.SearchHandler with Apache Tomcat Key: SOLR-5055 URL: https://issues.apache.org/jira/browse/SOLR-5055 Project: Solr Issue Type: Bug Components: search Affects Versions: 4.3.1 Environment: RedHat Linux 6.4 x64_86, Ubuntu 10.04.2 LTS i686 Reporter: Christian Lutz Priority: Minor I'm trying to deploy Solr with Tomcat (Version 7.0.27 and 6.0.37) as described in http://wiki.apache.org/solr/SolrTomcat and https://cwiki.apache.org/confluence/display/solr/Running+Solr+on+Tomcat I deployed only one instance (mentioned in Wiki Documentation as Single Solr Instance). In solrconfig.xml the library configuration is changed to absolut paths for the jar-Files. When i try to use the /browse Request Handler it states the following error: ERROR - 2013-07-22 13:04:20.279; org.apache.solr.common.SolrException; null:java.lang.RuntimeException: java.lang.NoClassDefFoundError: org/apache/commons/lang/StringUtils at org.apache.solr.servlet.SolrDispatchFilter.sendError(SolrDispatchFilter.java:670) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:380) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:155) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:243) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:210) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:225) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:169) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:168) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:98) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:407) at org.apache.coyote.ajp.AjpProcessor.process(AjpProcessor.java:200) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:565) at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:307) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:722) Caused by: java.lang.NoClassDefFoundError: org/apache/commons/lang/StringUtils at org.apache.velocity.runtime.resource.ResourceManagerImpl.initialize(ResourceManagerImpl.java:161) at org.apache.velocity.runtime.RuntimeInstance.initializeResourceManager(RuntimeInstance.java:730) at org.apache.velocity.runtime.RuntimeInstance.init(RuntimeInstance.java:263) at org.apache.velocity.app.VelocityEngine.init(VelocityEngine.java:93) at org.apache.solr.response.VelocityResponseWriter.getEngine(VelocityResponseWriter.java:147) at org.apache.solr.response.VelocityResponseWriter.write(VelocityResponseWriter.java:42) at org.apache.solr.core.SolrCore$LazyQueryResponseWriterWrapper.write(SolrCore.java:2278) at org.apache.solr.servlet.SolrDispatchFilter.writeResponse(SolrDispatchFilter.java:644) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:372) ... 15 more Caused by: java.lang.ClassNotFoundException: org.apache.commons.lang.StringUtils at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:423) at java.lang.ClassLoader.loadClass(ClassLoader.java:356) ... 24 more ERROR - 2013-07-22 13:04:20.282; org.apache.solr.common.SolrException; null:java.lang.NoClassDefFoundError: org/apache/commons/lang/StringUtils at org.apache.velocity.runtime.resource.ResourceManagerImpl.initialize(ResourceManagerImpl.java:161) at org.apache.velocity.runtime.RuntimeInstance.initializeResourceManager(RuntimeInstance.java:730) at org.apache.velocity.runtime.RuntimeInstance.init(RuntimeInstance.java:263) at org.apache.velocity.app.VelocityEngine.init(VelocityEngine.java:93) at org.apache.solr.response.VelocityResponseWriter.getEngine(VelocityResponseWriter.java:147) at org.apache.solr.response.VelocityResponseWriter.write(VelocityResponseWriter.java:42) at org.apache.solr.core.SolrCore$LazyQueryResponseWriterWrapper.write(SolrCore.java:2278) at
[jira] [Commented] (LUCENE-5101) make it easier to plugin different bitset implementations to CachingWrapperFilter
[ https://issues.apache.org/jira/browse/LUCENE-5101?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715124#comment-13715124 ] Paul Elschot commented on LUCENE-5101: -- I simplified the benchmark somewhat and ran it on my 32 bit home machine. The relative results for EliasFanoDocIdSet are very similar to the ones above, so this benchmark is a good measuring tool for me. I'll use it to verify performance of an index once it is working, LUCENE-5109. make it easier to plugin different bitset implementations to CachingWrapperFilter - Key: LUCENE-5101 URL: https://issues.apache.org/jira/browse/LUCENE-5101 Project: Lucene - Core Issue Type: Improvement Reporter: Robert Muir Attachments: DocIdSetBenchmark.java, LUCENE-5101.patch Currently this is possible, but its not so friendly: {code} protected DocIdSet docIdSetToCache(DocIdSet docIdSet, AtomicReader reader) throws IOException { if (docIdSet == null) { // this is better than returning null, as the nonnull result can be cached return EMPTY_DOCIDSET; } else if (docIdSet.isCacheable()) { return docIdSet; } else { final DocIdSetIterator it = docIdSet.iterator(); // null is allowed to be returned by iterator(), // in this case we wrap with the sentinel set, // which is cacheable. if (it == null) { return EMPTY_DOCIDSET; } else { /* INTERESTING PART */ final FixedBitSet bits = new FixedBitSet(reader.maxDoc()); bits.or(it); return bits; /* END INTERESTING PART */ } } } {code} Is there any value to having all this other logic in the protected API? It seems like something thats not useful for a subclass... Maybe this stuff can become final, and INTERESTING PART calls a simpler method, something like: {code} protected DocIdSet cacheImpl(DocIdSetIterator iterator, AtomicReader reader) { final FixedBitSet bits = new FixedBitSet(reader.maxDoc()); bits.or(iterator); return bits; } {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5128) Calling IndexSearcher.searchAfter beyond the number of stored documents causes ArrayIndexOutOfBoundsException
[ https://issues.apache.org/jira/browse/LUCENE-5128?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715127#comment-13715127 ] Michael McCandless commented on LUCENE-5128: I think the core change makes sense. If you pass a too-big doc to searchAfter, something is buggy with the app. And we should fix TestShardSearching to not abuse the API like that ... it should do its own tracking. But, it'd be good to see the actual stacktrace here first ... Calling IndexSearcher.searchAfter beyond the number of stored documents causes ArrayIndexOutOfBoundsException - Key: LUCENE-5128 URL: https://issues.apache.org/jira/browse/LUCENE-5128 Project: Lucene - Core Issue Type: Bug Components: core/search Affects Versions: 4.2 Reporter: crocket Attachments: LUCENE-5128.patch ArrayIndexOutOfBoundsException makes it harder to reason about the cause. Is there a better way to notify programmers of the cause? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: short-circuit OR operator in lucene/solr
When calculating a score, you _do_ care about the rest of the OR clauses since they can contribute to the score. It seems that you'd need an syntax to express And I don't care about scoring. Which is really an fq clause, so maybe this would have to go into logic there? On Mon, Jul 22, 2013 at 4:17 AM, Mikhail Khludnev mkhlud...@griddynamics.com wrote: Short answer, no - it has zero sense. But after some thinking, it can make some sense, potentially. DisjunctionSumScorer holds child scorers semi-ordered in a binary heap. Hypothetically inequality can be enforced at that heap, but heap might not work anymore for such alignment. Hence, instead of heap TreeSet can be used for experiment. fwiw, it's a dev list question. On Mon, Jul 22, 2013 at 4:48 AM, Deepak Konidena deepakk...@gmail.com wrote: I understand that lucene's AND (), OR (||) and NOT (!) operators are shorthands for REQUIRED, OPTIONAL and EXCLUDE respectively, which is why one can't treat them as boolean operators (adhering to boolean algebra). I have been trying to construct a simple OR expression, as follows q = +(field1:value1 OR field2:value2) with a match on either field1 or field2. But since the OR is merely an optional, documents where both field1:value1 and field2:value2 are matched, the query returns a score resulting in a match on both the clauses. How do I enforce short-circuiting in this context? In other words, how to implement short-circuiting as in boolean algebra where an expression A || B || C returns true if A is true without even looking into whether B or C could be true. -Deepak -- Sincerely yours Mikhail Khludnev Principal Engineer, Grid Dynamics - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-5054) My private query component does not work in solrcloud
[ https://issues.apache.org/jira/browse/SOLR-5054?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Erick Erickson resolved SOLR-5054. -- Resolution: Invalid Please raise kind of question on the user's or dev list before raising a JIRA, this sounds a lot more like a usage question than a development task. But my first guess is that you haven't put your custom jar files on all your nodes. My private query component does not work in solrcloud -- Key: SOLR-5054 URL: https://issues.apache.org/jira/browse/SOLR-5054 Project: Solr Issue Type: Bug Components: SolrCloud Affects Versions: 4.3.1 Environment: Linux x64 Reporter: che Labels: solr, solrcloud Original Estimate: 5m Remaining Estimate: 5m I user my HQ.jar in one solr is ok every query is entry in my HQ. But I make a solrcloud with three shards. The HQ.jar is not work.The log is not print .so I think it is not entry in. Do you know why ,thank you any help. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-5056) Further clean up of ConfigSolr interface and CoreContainer construction
Alan Woodward created SOLR-5056: --- Summary: Further clean up of ConfigSolr interface and CoreContainer construction Key: SOLR-5056 URL: https://issues.apache.org/jira/browse/SOLR-5056 Project: Solr Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Makes ConfigSolr a bit more typesafe, and pushes a bunch of cloud-specific config into ZkController. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-5056) Further clean up of ConfigSolr interface and CoreContainer construction
[ https://issues.apache.org/jira/browse/SOLR-5056?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alan Woodward updated SOLR-5056: Attachment: SOLR-5056.patch Further clean up of ConfigSolr interface and CoreContainer construction --- Key: SOLR-5056 URL: https://issues.apache.org/jira/browse/SOLR-5056 Project: Solr Issue Type: Improvement Reporter: Alan Woodward Assignee: Alan Woodward Priority: Minor Attachments: SOLR-5056.patch Makes ConfigSolr a bit more typesafe, and pushes a bunch of cloud-specific config into ZkController. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5127) FixedGapTermsIndex should use monotonic compression
[ https://issues.apache.org/jira/browse/LUCENE-5127?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715169#comment-13715169 ] Michael McCandless commented on LUCENE-5127: +1 I think we should just nuke the divisor? FixedGapTermsIndex should use monotonic compression --- Key: LUCENE-5127 URL: https://issues.apache.org/jira/browse/LUCENE-5127 Project: Lucene - Core Issue Type: Improvement Reporter: Robert Muir Attachments: LUCENE-5127.patch for the addresses in the big in-memory byte[] and disk blocks, we could save a good deal of RAM here. I think this codec just never got upgraded when we added these new packed improvements, but it might be interesting to try to use for the terms data of sorted/sortedset DV implementations. patch works, but has nocommits and currently ignores the divisor. The annoying problem there being that we have the shared interface with get(int) for PackedInts.Mutable/Reader, but no equivalent base class for monotonics get(long)... Still its enough that we could benchmark/compare for now. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-5057) queryResultCache should not related with the order of fq's list
huangfeihong created SOLR-5057: -- Summary: queryResultCache should not related with the order of fq's list Key: SOLR-5057 URL: https://issues.apache.org/jira/browse/SOLR-5057 Project: Solr Issue Type: Improvement Components: search Affects Versions: 4.3, 4.2, 4.1, 4.0 Reporter: huangfeihong Priority: Minor There are two case in the same meaning below. But the case2 can't use the queryResultCache when case1 is executed. case1: q=*:*fq=field1:value1fq=field2:value2 case2: q=*:*fq=field2:value2fq=field1:value1 I think queryResultCache should not be related with the order of fq's list. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-4956) make maxBufferedAddsPerServer configurable
[ https://issues.apache.org/jira/browse/SOLR-4956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715210#comment-13715210 ] Yonik Seeley commented on SOLR-4956: Yes, if we're buffering per-thread/request, then the larger the buffer, the larger the practical chances that big update reorderings can happen on replicas. If the reordering is big enough, a recovering node using peer sync could think it was pretty much up to date. It's not just peer sync either - other things like deletes, and deleteByQuery have limited lookback to handle reorders. Buffering isn't ideal for a lot of reasons - update reorderings on the order of nIndexingThreads * bufferSize - increases the intermediate difference between replicas and leaders (wrt queries) - memory size - we don't know how big documents are... they could be huge It would be nice to move in the other direction and do less buffering. make maxBufferedAddsPerServer configurable -- Key: SOLR-4956 URL: https://issues.apache.org/jira/browse/SOLR-4956 Project: Solr Issue Type: Improvement Affects Versions: 4.3, 5.0 Reporter: Erick Erickson Anecdotal user's list evidence indicates that in high-throughput situations, the default of 10 docs/batch for inter-shard batching can generate significant CPU load. See the thread titled Sharding and Replication on June 19th, but the gist is below. I haven't poked around, but it's a little surprising on the surface that Asif is seeing this kind of difference. So I'm wondering if this change indicates some other underlying issue. Regardless, this seems like it would be good to investigate. Here's the gist of Asif's experience from the thread: Its a completely practical problem - we are exploring Solr to build a real time analytics/data solution for a system handling about 1000 qps. We have various metrics that are stored as different collections on the cloud, which means very high amount of writes. The cloud also needs to support about 300-400 qps. We initially tested with a single Solr node on a 16 core / 24 GB box for a single metric. We saw that writes were not a issue at all - Solr was handling it extremely well. We were also able to achieve about 200 qps from a single node. When we set up the cloud ( a ensemble on 6 boxes), we saw very high CPU usage on the replicas. Up to 10 cores were getting used for writes on the replicas. Hence my concern with respect to batch updates for the replicas. BTW, I altered the maxBufferedAddsPerServer to 1000 - and now CPU usage is very similar to single node installation. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[RESULT][VOTE] Release 4.4 RC1
This vote has passed. I'll start the release process now. Steve On Jul 19, 2013, at 10:05 AM, Steve Rowe sar...@gmail.com wrote: Please vote to release Lucene and Solr 4.4, built off revision 1504776 of https://svn.apache.org/repos/asf/lucene/dev/branches/lucene_solr_4_4. RC1 artifacts are available at: http://people.apache.org/~sarowe/staging_area/lucene-solr-4.4.0-RC1-rev1504776 The smoke tester passes for me. Here's my +1. Steve - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5128) Calling IndexSearcher.searchAfter beyond the number of stored documents causes ArrayIndexOutOfBoundsException
[ https://issues.apache.org/jira/browse/LUCENE-5128?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715203#comment-13715203 ] Shai Erera commented on LUCENE-5128: bq. it should do its own tracking That's the part that bothers me. If we don't allow passing such searchAfter, we're basically telling apps that they should keep track of their shards, which means they need to maintain state (not that it's hard to maintain in this case). I'd rather first see the exception because from what I checked both TopScoreDocsCollector and TopFieldsCollector should be fine with searchAfter = reader.maxDoc(), and IndexSearch.searchAfter returns a TopDocs so there aren't many collectors which can be affected. Except maybe if crocket uses DrillSideways... On one hand the core fix makes sense, on the other hand it may unnecessarily complicate apps life. Especially if they need to obtain totalHits ... I wonder if in that case they can run with TotalHitCountCollector instead and construct an empty TopDocs? That's what happens when you currently pass searchAfter=Integer.MAX_VALUE. I'll try to fix the test like that to see if it helps. Calling IndexSearcher.searchAfter beyond the number of stored documents causes ArrayIndexOutOfBoundsException - Key: LUCENE-5128 URL: https://issues.apache.org/jira/browse/LUCENE-5128 Project: Lucene - Core Issue Type: Bug Components: core/search Affects Versions: 4.2 Reporter: crocket Attachments: LUCENE-5128.patch ArrayIndexOutOfBoundsException makes it harder to reason about the cause. Is there a better way to notify programmers of the cause? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-5057) queryResultCache should not related with the order of fq's list
[ https://issues.apache.org/jira/browse/SOLR-5057?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Feihong Huang updated SOLR-5057: Description: There are two case query with the same meaning below. But the case2 can't use the queryResultCache when case1 is executed. case1: q=*:*fq=field1:value1fq=field2:value2 case2: q=*:*fq=field2:value2fq=field1:value1 I think queryResultCache should not be related with the order of fq's list. was: There are two case in the same meaning below. But the case2 can't use the queryResultCache when case1 is executed. case1: q=*:*fq=field1:value1fq=field2:value2 case2: q=*:*fq=field2:value2fq=field1:value1 I think queryResultCache should not be related with the order of fq's list. queryResultCache should not related with the order of fq's list --- Key: SOLR-5057 URL: https://issues.apache.org/jira/browse/SOLR-5057 Project: Solr Issue Type: Improvement Components: search Affects Versions: 4.0, 4.1, 4.2, 4.3 Reporter: Feihong Huang Priority: Minor Original Estimate: 48h Remaining Estimate: 48h There are two case query with the same meaning below. But the case2 can't use the queryResultCache when case1 is executed. case1: q=*:*fq=field1:value1fq=field2:value2 case2: q=*:*fq=field2:value2fq=field1:value1 I think queryResultCache should not be related with the order of fq's list. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [VOTE] Release 4.4 RC1
+1 -Yonik http://lucidworks.com On Fri, Jul 19, 2013 at 10:05 AM, Steve Rowe sar...@gmail.com wrote: Please vote to release Lucene and Solr 4.4, built off revision 1504776 of https://svn.apache.org/repos/asf/lucene/dev/branches/lucene_solr_4_4. RC1 artifacts are available at: http://people.apache.org/~sarowe/staging_area/lucene-solr-4.4.0-RC1-rev1504776 The smoke tester passes for me. Here's my +1. Steve - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-5057) queryResultCache should not related with the order of fq's list
[ https://issues.apache.org/jira/browse/SOLR-5057?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715267#comment-13715267 ] Erick Erickson commented on SOLR-5057: -- Hmmm, my initial reaction was that it'd be hard to figure out what clauses were equivalent. But I suppose a simple lexical ordering of all the relevant clauses would work for this. Of course the work-around is to generate your fq clauses in the same order every time. queryResultCache should not related with the order of fq's list --- Key: SOLR-5057 URL: https://issues.apache.org/jira/browse/SOLR-5057 Project: Solr Issue Type: Improvement Components: search Affects Versions: 4.0, 4.1, 4.2, 4.3 Reporter: Feihong Huang Priority: Minor Original Estimate: 48h Remaining Estimate: 48h There are two case query with the same meaning below. But the case2 can't use the queryResultCache when case1 is executed. case1: q=*:*fq=field1:value1fq=field2:value2 case2: q=*:*fq=field2:value2fq=field1:value1 I think queryResultCache should not be related with the order of fq's list. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-4478) Allow cores to specify a named config set
[ https://issues.apache.org/jira/browse/SOLR-4478?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715250#comment-13715250 ] Noble Paul commented on SOLR-4478: -- handling the ${} substitutions are more work. But a lot of users who need sharing of solrconfig would be happy to sacrifice that feature. if there are ${} is solrconfig.xml and shareConfig=true we can just fail. Longer term we can support that too but will need some more changes the way properties are read Allow cores to specify a named config set - Key: SOLR-4478 URL: https://issues.apache.org/jira/browse/SOLR-4478 Project: Solr Issue Type: Improvement Affects Versions: 4.2, 5.0 Reporter: Erick Erickson Assignee: Erick Erickson Attachments: SOLR-4478.patch, SOLR-4478.patch Part of moving forward to the new way, after SOLR-4196 etc... I propose an additional parameter specified on the core node in solr.xml or as a parameter in the discovery mode core.properties file, call it configSet, where the value provided is a path to a directory, either absolute or relative. Really, this is as though you copied the conf directory somewhere to be used by more than one core. Straw-man: There will be a directory solr_home/configsets which will be the default. If the configSet parameter is, say, myconf, then I'd expect a directory named myconf to exist in solr_home/configsets, which would look something like solr_home/configsets/myconf/schema.xml solrconfig.xml stopwords.txt velocity velocity/query.vm etc. If multiple cores used the same configSet, schema, solrconfig etc. would all be shared (i.e. shareSchema=true would be assumed). I don't see a good use-case for _not_ sharing schemas, so I don't propose to allow this to be turned off. Hmmm, what if shareSchema is explicitly set to false in the solr.xml or properties file? I'd guess it should be honored but maybe log a warning? Mostly I'm putting this up for comments. I know that there are already thoughts about how this all should work floating around, so before I start any work on this I thought I'd at least get an idea of whether this is the way people are thinking about going. Configset can be either a relative or absolute path, if relative it's assumed to be relative to solr_home. Thoughts? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5127) FixedGapTermsIndex should use monotonic compression
[ https://issues.apache.org/jira/browse/LUCENE-5127?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715302#comment-13715302 ] Robert Muir commented on LUCENE-5127: - Maybe, though we could also add a minimal get(long) interface to blockpacked/monotonicblockpacked/appending/monotonicappending. A few notes: * Current patch changes both the disk offsets (termsDictOffsets) and the offsets into the in-ram terms data (termOffsets) * With the current patch as-is, we could remove the interval*2B #terms limitation, as long addressing is used everywhere. * Current patch saves RAM, savings increase as termsindex/termsdict gets larger. With 10M: ||Checkout||TIB||TII|| |Trunk|519329144|19300603| |Patch|519329144|14149524| * Current patch slows down seek-heavy queries a bit: {noformat} Task QPS trunk StdDev QPS patch StdDev Pct diff PKLookup 86.02 (2.9%) 76.17 (2.4%) -11.4% ( -16% - -6%) Respell 39.76 (3.0%) 36.58 (2.5%) -8.0% ( -13% - -2%) Fuzzy2 35.49 (4.1%) 32.88 (2.6%) -7.3% ( -13% -0%) Fuzzy1 31.49 (4.1%) 29.18 (2.6%) -7.3% ( -13% -0%) {noformat} * termOffsets are read twice per seek / binary search iteration: {code} final long offset = fieldIndex.termOffsets.get(idx); final int length = (int) (fieldIndex.termOffsets.get(1+idx) - offset); {code} * termsDictOffsets are only read once... and this is really just an unfortunate consequence of TermsIndexReaderBase's API... ideally they would lazy-decode this until you really needed it, like BlockTree. So I see a few things we could do: # go forward with current patch (maybe add the divisor stuff via a simple get() interface). clean up int-long everywhere. I'm not sure if these perf diffs matter for the use cases where someone needs an ord-enabled terms index? # hybrid patch, where termOffsets stay absolute but termDictOffsets use monotonicpacked. This would still save some space, but restore the seek-heavy perf. But then we wouldnt be able to cleanup int-long and so on. # do nothing, maybe fork the logic of this thing so it can be used in DV. For how DV is used, it'd be the right tradeoff so its no issue there. FixedGapTermsIndex should use monotonic compression --- Key: LUCENE-5127 URL: https://issues.apache.org/jira/browse/LUCENE-5127 Project: Lucene - Core Issue Type: Improvement Reporter: Robert Muir Attachments: LUCENE-5127.patch for the addresses in the big in-memory byte[] and disk blocks, we could save a good deal of RAM here. I think this codec just never got upgraded when we added these new packed improvements, but it might be interesting to try to use for the terms data of sorted/sortedset DV implementations. patch works, but has nocommits and currently ignores the divisor. The annoying problem there being that we have the shared interface with get(int) for PackedInts.Mutable/Reader, but no equivalent base class for monotonics get(long)... Still its enough that we could benchmark/compare for now. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [VOTE] Release 4.4 RC1
+1 - Mark On Jul 19, 2013, at 10:05 AM, Steve Rowe sar...@gmail.com wrote: Please vote to release Lucene and Solr 4.4, built off revision 1504776 of https://svn.apache.org/repos/asf/lucene/dev/branches/lucene_solr_4_4. RC1 artifacts are available at: http://people.apache.org/~sarowe/staging_area/lucene-solr-4.4.0-RC1-rev1504776 The smoke tester passes for me. Here's my +1. Steve - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
RE: [VOTE] Release 4.4 RC1
Sorry a little bit late, here is my +1 - Uwe Schindler H.-H.-Meier-Allee 63, D-28213 Bremen http://www.thetaphi.de eMail: u...@thetaphi.de -Original Message- From: Steve Rowe [mailto:sar...@gmail.com] Sent: Friday, July 19, 2013 4:06 PM To: dev@lucene.apache.org Subject: [VOTE] Release 4.4 RC1 Please vote to release Lucene and Solr 4.4, built off revision 1504776 of https://svn.apache.org/repos/asf/lucene/dev/branches/lucene_solr_4_4 . RC1 artifacts are available at: http://people.apache.org/~sarowe/staging_area/lucene-solr-4.4.0-RC1- rev1504776 The smoke tester passes for me. Here's my +1. Steve - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: svn commit: r870606 - in /websites/production/lucene/content: core/4_4_0/ core/4_4_0/analyzers-common/ core/4_4_0/analyzers-common/org/ core/4_4_0/analyzers-common/org/apache/ core/4_4_0/analyzers
Yeah, you managed it in one commit! sar...@apache.org schrieb: Author: sarowe Date: Mon Jul 22 16:25:46 2013 New Revision: 870606 Log: Lucene/Solr 4.4.0 documentation [This commit notification would consist of 1693 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.] -- Uwe Schindler H.-H.-Meier-Allee 63, 28213 Bremen http://www.thetaphi.de
[jira] [Commented] (LUCENE-4734) FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight
[ https://issues.apache.org/jira/browse/LUCENE-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715361#comment-13715361 ] ASF subversion and git services commented on LUCENE-4734: - Commit 1505731 from [~jpountz] in branch 'dev/trunk' [ https://svn.apache.org/r1505731 ] LUCENE-4734: Better memory efficiency. FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight Key: LUCENE-4734 URL: https://issues.apache.org/jira/browse/LUCENE-4734 Project: Lucene - Core Issue Type: Bug Components: modules/highlighter Affects Versions: 4.0, 4.1, 5.0 Reporter: Ryan Lauck Assignee: Adrien Grand Labels: fastvectorhighlighter, highlighter Fix For: 4.4 Attachments: LUCENE-4734-2.patch, lucene-4734.patch, LUCENE-4734.patch If a proximity phrase query overlaps with any other query term it will not be highlighted. Example Text: A B C D E F G Example Queries: B E~10 D (D will be highlighted instead of B C D E) B E~10 C F~10 (nothing will be highlighted) This can be traced to the FieldPhraseList constructor's inner while loop. From the first example query, the first TermInfo popped off the stack will be B. The second TermInfo will be D which will not be found in the submap for B E~10 and will trigger a failed match. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4734) FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight
[ https://issues.apache.org/jira/browse/LUCENE-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715364#comment-13715364 ] ASF subversion and git services commented on LUCENE-4734: - Commit 1505732 from [~jpountz] in branch 'dev/branches/branch_4x' [ https://svn.apache.org/r1505732 ] LUCENE-4734: Better memory efficiency. FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight Key: LUCENE-4734 URL: https://issues.apache.org/jira/browse/LUCENE-4734 Project: Lucene - Core Issue Type: Bug Components: modules/highlighter Affects Versions: 4.0, 4.1, 5.0 Reporter: Ryan Lauck Assignee: Adrien Grand Labels: fastvectorhighlighter, highlighter Fix For: 4.4 Attachments: LUCENE-4734-2.patch, lucene-4734.patch, LUCENE-4734.patch If a proximity phrase query overlaps with any other query term it will not be highlighted. Example Text: A B C D E F G Example Queries: B E~10 D (D will be highlighted instead of B C D E) B E~10 C F~10 (nothing will be highlighted) This can be traced to the FieldPhraseList constructor's inner while loop. From the first example query, the first TermInfo popped off the stack will be B. The second TermInfo will be D which will not be found in the submap for B E~10 and will trigger a failed match. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: svn commit: r870606 - in /websites/production/lucene/content: core/4_4_0/ core/4_4_0/analyzers-common/ core/4_4_0/analyzers-common/org/ core/4_4_0/analyzers-common/org/apache/ core/4_4_0/analyzers
Yes, only 48 minutes to upload!!! Since Shalin disabled the hourly website update for twitter/svn, checkouts don't get stale, and long commits can succeed. Steve On Jul 22, 2013, at 12:57 PM, Uwe Schindler u...@thetaphi.de wrote: Yeah, you managed it in one commit! sar...@apache.org schrieb: Author: sarowe Date: Mon Jul 22 16:25:46 2013 New Revision: 870606 Log: Lucene/Solr 4.4.0 documentation [This commit notification would consist of 1693 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.] -- Uwe Schindler H.-H.-Meier-Allee 63, 28213 Bremen http://www.thetaphi.de - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [JENKINS] Lucene-Solr-4.x-Linux (64bit/jdk1.8.0-ea-b98) - Build # 6591 - Failure!
I'll look into this. On 7/21/13 1:04 PM, Policeman Jenkins Server jenk...@thetaphi.de wrote: Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/6591/ Java: 64bit/jdk1.8.0-ea-b98 -XX:+UseCompressedOops -XX:+UseParallelGC 1 tests failed. FAILED: org.apache.lucene.spatial.prefix.SpatialOpRecursivePrefixTreeTest.testCont ains {#6 seed=[8C5C2CF3C312F7E6:6F314F7E2AA1141]} Error Message: Stack Trace: java.lang.NullPointerException at __randomizedtesting.SeedInfo.seed([8C5C2CF3C312F7E6:6F314F7E2AA1141]:0) at org.apache.lucene.spatial.prefix.ContainsPrefixTreeFilter$ContainsVisitor. seekExact(ContainsPrefixTreeFilter.java:137) at org.apache.lucene.spatial.prefix.ContainsPrefixTreeFilter$ContainsVisitor. visit(ContainsPrefixTreeFilter.java:108) at org.apache.lucene.spatial.prefix.ContainsPrefixTreeFilter$ContainsVisitor. access$000(ContainsPrefixTreeFilter.java:80) at org.apache.lucene.spatial.prefix.ContainsPrefixTreeFilter.getDocIdSet(Cont ainsPrefixTreeFilter.java:77) at org.apache.lucene.search.ConstantScoreQuery$ConstantWeight.scorer(Constant ScoreQuery.java:131) at org.apache.lucene.search.AssertingWeight.scorer(AssertingWeight.java:66) at org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:621) at org.apache.lucene.search.AssertingIndexSearcher.search(AssertingIndexSearc her.java:93) at org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:488) at org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:444) at org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:281) at org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:269) at org.apache.lucene.spatial.SpatialTestCase.executeQuery(SpatialTestCase.jav a:99) at org.apache.lucene.spatial.prefix.SpatialOpRecursivePrefixTreeTest.doTest(S patialOpRecursivePrefixTreeTest.java:281) at org.apache.lucene.spatial.prefix.SpatialOpRecursivePrefixTreeTest.testCont ains(SpatialOpRecursivePrefixTreeTest.java:103) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java: 57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorIm pl.java:43) at java.lang.reflect.Method.invoke(Method.java:491) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunne r.java:1559) at com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedR unner.java:79) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedR unner.java:737) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedR unner.java:773) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedR unner.java:787) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSet upTeardownChained.java:50) at org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCa cheSanity.java:51) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAf terRule.java:46) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.e valuate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThread AndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleI gnoreAfterMaxFailures.java:70) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure. java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(Stateme ntAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(T hreadLeakControl.java:358) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(Th readLeakControl.java:782) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeak Control.java:442) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(Randomiz edRunner.java:746) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedR unner.java:648) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedR unner.java:682) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedR unner.java:693) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAf terRule.java:46) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClas sName.java:42) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.e valuate(SystemPropertiesInvariantRule.java:55) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRu
[jira] [Commented] (LUCENE-4734) FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight
[ https://issues.apache.org/jira/browse/LUCENE-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715444#comment-13715444 ] Ryan Lauck commented on LUCENE-4734: I'm curious what you think about the comment I had in my original patch before calling addIfNoOverlap. It seems wasteful to traverse the phrase candidate list from the beginning every iteration to search for overlaps, and also prevents gathering the raw highlights as mentioned in my previous comment. What do you think about waiting until all phrase candidates are gathered to optionally filter overlaps? FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight Key: LUCENE-4734 URL: https://issues.apache.org/jira/browse/LUCENE-4734 Project: Lucene - Core Issue Type: Bug Components: modules/highlighter Affects Versions: 4.0, 4.1, 5.0 Reporter: Ryan Lauck Assignee: Adrien Grand Labels: fastvectorhighlighter, highlighter Fix For: 4.4 Attachments: LUCENE-4734-2.patch, lucene-4734.patch, LUCENE-4734.patch If a proximity phrase query overlaps with any other query term it will not be highlighted. Example Text: A B C D E F G Example Queries: B E~10 D (D will be highlighted instead of B C D E) B E~10 C F~10 (nothing will be highlighted) This can be traced to the FieldPhraseList constructor's inner while loop. From the first example query, the first TermInfo popped off the stack will be B. The second TermInfo will be D which will not be found in the submap for B E~10 and will trigger a failed match. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4734) FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight
[ https://issues.apache.org/jira/browse/LUCENE-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715461#comment-13715461 ] Adrien Grand commented on LUCENE-4734: -- I agree this seems wasteful. Maybe we could open an issue about it? FastVectorHighlighter Overlapping Proximity Queries Do Not Highlight Key: LUCENE-4734 URL: https://issues.apache.org/jira/browse/LUCENE-4734 Project: Lucene - Core Issue Type: Bug Components: modules/highlighter Affects Versions: 4.0, 4.1, 5.0 Reporter: Ryan Lauck Assignee: Adrien Grand Labels: fastvectorhighlighter, highlighter Fix For: 4.4 Attachments: LUCENE-4734-2.patch, lucene-4734.patch, LUCENE-4734.patch If a proximity phrase query overlaps with any other query term it will not be highlighted. Example Text: A B C D E F G Example Queries: B E~10 D (D will be highlighted instead of B C D E) B E~10 C F~10 (nothing will be highlighted) This can be traced to the FieldPhraseList constructor's inner while loop. From the first example query, the first TermInfo popped off the stack will be B. The second TermInfo will be D which will not be found in the submap for B E~10 and will trigger a failed match. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-5058) org.apache.solr.update.PeerSync Logging Warning Typo
Thomas Murphy created SOLR-5058: --- Summary: org.apache.solr.update.PeerSync Logging Warning Typo Key: SOLR-5058 URL: https://issues.apache.org/jira/browse/SOLR-5058 Project: Solr Issue Type: Bug Affects Versions: 4.3 Reporter: Thomas Murphy Priority: Trivial Log entry appears on Solr Admin Logging interface: WARNPeerSyncno frame of reference to tell of we've missed updates There is a typo, this looks like it should read to tell if we've PeerSync expands to org.apache.solr.update.PeerSync -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-4677) Improve Solr's use of spec version.
[ https://issues.apache.org/jira/browse/SOLR-4677?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715558#comment-13715558 ] ASF subversion and git services commented on SOLR-4677: --- Commit 1505776 from [~steve_rowe] in branch 'dev/trunk' [ https://svn.apache.org/r1505776 ] SOLR-4677: stop adding obsolete '-Dspecversion=...' to 'ant prepare-release' cmds Improve Solr's use of spec version. --- Key: SOLR-4677 URL: https://issues.apache.org/jira/browse/SOLR-4677 Project: Solr Issue Type: Improvement Components: Build Reporter: Mark Miller Fix For: 4.3, 5.0 Attachments: SOLR-4677.patch, SOLR-4677.patch, SOLR-4677.patch Solr 4.2.1 went out with an impl version of 4.2.1 and a spec version of 4.2.0. This is because you must update the spec version in common-build.xml while the impl is set by the version you pass as a sys prop when doing prepare-release. Do we need this spec version? Does it serve any purpose? I think we should either stop dealing with it or just set it the same way as the impl...or? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-4677) Improve Solr's use of spec version.
[ https://issues.apache.org/jira/browse/SOLR-4677?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715562#comment-13715562 ] ASF subversion and git services commented on SOLR-4677: --- Commit 1505778 from [~steve_rowe] in branch 'dev/branches/branch_4x' [ https://svn.apache.org/r1505778 ] SOLR-4677: stop adding obsolete '-Dspecversion=...' to 'ant prepare-release' cmds (merged trunk r1505776) Improve Solr's use of spec version. --- Key: SOLR-4677 URL: https://issues.apache.org/jira/browse/SOLR-4677 Project: Solr Issue Type: Improvement Components: Build Reporter: Mark Miller Fix For: 4.3, 5.0 Attachments: SOLR-4677.patch, SOLR-4677.patch, SOLR-4677.patch Solr 4.2.1 went out with an impl version of 4.2.1 and a spec version of 4.2.0. This is because you must update the spec version in common-build.xml while the impl is set by the version you pass as a sys prop when doing prepare-release. Do we need this spec version? Does it serve any purpose? I think we should either stop dealing with it or just set it the same way as the impl...or? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-5036) Solr Ref Guide updates for Solr 4.4
[ https://issues.apache.org/jira/browse/SOLR-5036?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13710096#comment-13710096 ] Hoss Man edited comment on SOLR-5036 at 7/22/13 7:51 PM: - I made some more progress on Cassandra's list... * https://cwiki.apache.org/confluence/display/solr/RequestDispatcher+in+SolrConfig ** SOLR-2079 - addHttpRequestToContext verbage * https://cwiki.apache.org/confluence/display/solr/Query+Screen ** SOLR-3838 - fq verbage and updated creenshot ** SOLR-4719 - json default verbage updated screenshot * https://cwiki.apache.org/confluence/display/solr/Core-Specific+Tools ** update order of links to children and actual order of children in navigation * https://cwiki.apache.org/confluence/display/solr/IndexConfig+in+SolrConfig ** SOLR-4941, SOLR-4934 and LUCENE-5038 - already ok, page doesn't go into depth about mergePolicy, example alreayd used correct compound file syntax ** SOLR-4761: added section on mergedSegmentWarmer * https://cwiki.apache.org/confluence/display/solr/Command+Line+Utilities ** SOLR-4972: Add PUT command to ZkCli tool * https://cwiki.apache.org/confluence/display/solr/Other+Parsers ** SOLR-4785: MaxScoreQParser * https://cwiki.apache.org/confluence/display/solr/Upgrading+to+Solr+4.4 ** New page on upgrading to 4.4 ** child of https://cwiki.apache.org/confluence/display/solr/Major+Changes+from+Solr+3+to+Solr+4 ** clone of https://cwiki.apache.org/confluence/display/solr/Upgrading+to+Solr+4.3 ** mention SOLR-4941, SOLR-4934 and LUCENE-5038 ** mention SOLR-4778 and new LogWatcher API * https://cwiki.apache.org/confluence/display/solr/Detecting+Languages+During+Indexing ** SOLR-4412: New parameter langid.lcmap * https://cwiki.apache.org/confluence/display/solr/Spell+Checking ** SOLR-3240: Add spellcheck.collateMaxCollectDocs * https://cwiki.apache.org/confluence/display/solr/Documents+Screen ** Added a basic stub page with a screenshoot ** linked from parent page ** ensured the page order was correct in navigation But there is still a lot more that needs written -- stuff below i either don't understand the new feature enough to document myself, or didn't have the energy to try to tackle because it needs a lot written about it and i don't know it well enough to have a good sense of how to go about it... * https://cwiki.apache.org/confluence/display/solr/Core-Specific+Tools ** SOLR-4921: Admin UI now supports adding documents to Solr (gsingers, steffkes) ** stub page with screenshot exists, but it needs verbage * Completley new docs about the HDFS SolrCloud support ... somewhere ** SOLR-4916: Add support to write and read Solr index files and transaction log files to and from HDFS. (phunt, Mark Miller, Greg Chanan) *** CT: Without studying this more, it's hard to know where this should go. It's not really SolrCloud, and it's not really a client, but depending on why it's being done it could overlap with either...If someone writes up what you'd tell someone about using it, I could give a better idea of where it fits in the existing page organization (if it does). * https://cwiki.apache.org/confluence/display/solr/Documents%2C+Fields%2C+and+Schema+Design ** SOLR-4897: Add solr/example/example-schemaless/, an example config set for schemaless mode. (Steve Rowe) *** CT: Schemaless in general needs to be added. The most likely place today is a new page under https://cwiki.apache.org/confluence/display/solr/Documents%2C+Fields%2C+and+Schema+Design * https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml ** SOLR-4757: Change the example to use the new solr.xml format and core discovery by directory structure. (Mark Miller) *** CT: There is a page on solr.xml: https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml. This should be updated to show the new format and still include information on the old format for anyone with the old format who uses this guide for reference. ** SOLR-4655: Add option to have Overseer assign generic node names so that new addresses can host shards without naming confusion. (Mark Miller, Anshum Gupta) *** CT: I think this only needs to be added to any new content for solr.xml at https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml * https://cwiki.apache.org/confluence/display/solr/Collections+API ** SOLR-4693: A deleteshard collections API that unloads all replicas of a given shard and then removes it from the cluster state. It will remove only those shards which are INACTIVE or have no range (created for custom sharding). (Anshum Gupta, shalin) *** CT: Add to https://cwiki.apache.org/confluence/display/solr/Collections+API * https://cwiki.apache.org/confluence/display/solr/Schema+API ** SOLR-3251: Dynamically add fields to schema. (Steve Rowe, Robert Muir, yonik) *** CT: Add to
[jira] [Created] (SOLR-5059) 4.4 refguide pages on schemaless schema rest api for adding fields
Hoss Man created SOLR-5059: -- Summary: 4.4 refguide pages on schemaless schema rest api for adding fields Key: SOLR-5059 URL: https://issues.apache.org/jira/browse/SOLR-5059 Project: Solr Issue Type: Sub-task Components: documentation Reporter: Hoss Man Assignee: Steve Rowe breaking off from parent... * https://cwiki.apache.org/confluence/display/solr/Documents%2C+Fields%2C+and+Schema+Design ** SOLR-4897: Add solr/example/example-schemaless/, an example config set for schemaless mode. (Steve Rowe) *** CT: Schemaless in general needs to be added. The most likely place today is a new page under https://cwiki.apache.org/confluence/display/solr/Documents%2C+Fields%2C+and+Schema+Design * https://cwiki.apache.org/confluence/display/solr/Schema+API ** SOLR-3251: Dynamically add fields to schema. (Steve Rowe, Robert Muir, yonik) *** CT: Add to https://cwiki.apache.org/confluence/display/solr/Schema+API ** SOLR-5010: Add support for creating copy fields to the Fields REST API (gsingers) *** CT: Add to https://cwiki.apache.org/confluence/display/solr/Schema+API -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-5060) 4.4 refguide pages on hdfs support
Hoss Man created SOLR-5060: -- Summary: 4.4 refguide pages on hdfs support Key: SOLR-5060 URL: https://issues.apache.org/jira/browse/SOLR-5060 Project: Solr Issue Type: Sub-task Components: documentation Reporter: Hoss Man Assignee: Mark Miller breaking off from parent... * Completley new docs about the HDFS SolrCloud support ... somewhere ** SOLR-4916: Add support to write and read Solr index files and transaction log files to and from HDFS. (phunt, Mark Miller, Greg Chanan) *** CT: Without studying this more, it's hard to know where this should go. It's not really SolrCloud, and it's not really a client, but depending on why it's being done it could overlap with either...If someone writes up what you'd tell someone about using it, I could give a better idea of where it fits in the existing page organization (if it does). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-5061) 4.4 refguide pages new solr.xml format
Hoss Man created SOLR-5061: -- Summary: 4.4 refguide pages new solr.xml format Key: SOLR-5061 URL: https://issues.apache.org/jira/browse/SOLR-5061 Project: Solr Issue Type: Sub-task Components: documentation Reporter: Hoss Man Assignee: Erick Erickson breaking off from parent issue... * https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml ** SOLR-4757: Change the example to use the new solr.xml format and core discovery by directory structure. (Mark Miller) *** CT: There is a page on solr.xml: https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml. This should be updated to show the new format and still include information on the old format for anyone with the old format who uses this guide for reference. ** SOLR-4655: Add option to have Overseer assign generic node names so that new addresses can host shards without naming confusion. (Mark Miller, Anshum Gupta) *** CT: I think this only needs to be added to any new content for solr.xml at https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml It should also be noted that cassandra posted some additional deailed suggests in a comment on the existing page in the ref guide... https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml?focusedCommentId=33296160#comment-33296160 -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-5062) 4.4 refguide updates related to shardsplitting and deleteshard
Hoss Man created SOLR-5062: -- Summary: 4.4 refguide updates related to shardsplitting and deleteshard Key: SOLR-5062 URL: https://issues.apache.org/jira/browse/SOLR-5062 Project: Solr Issue Type: Sub-task Components: documentation Reporter: Hoss Man Assignee: Shalin Shekhar Mangar breaking off from parent issue... * https://cwiki.apache.org/confluence/display/solr/Collections+API ** in general, we need to review this in page in lite of all the shardsplitting stuff and make sure everything is up to date. ** SOLR-4693: A deleteshard collections API that unloads all replicas of a given shard and then removes it from the cluster state. It will remove only those shards which are INACTIVE or have no range (created for custom sharding). (Anshum Gupta, shalin) *** CT: Add to https://cwiki.apache.org/confluence/display/solr/Collections+API -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-5063) 4.4 refguide improvements on new doc adding screen in ui
Hoss Man created SOLR-5063: -- Summary: 4.4 refguide improvements on new doc adding screen in ui Key: SOLR-5063 URL: https://issues.apache.org/jira/browse/SOLR-5063 Project: Solr Issue Type: Sub-task Components: documentation Reporter: Hoss Man Assignee: Grant Ingersoll breaking off from parent issue... * https://cwiki.apache.org/confluence/display/solr/Core-Specific+Tools ** SOLR-4921: Admin UI now supports adding documents to Solr (gsingers, steffkes) ** stub page with screenshot exists, but it needs verbage explaining how it works and what the diff options mean -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3688) Bucketing of association value of a certain category
[ https://issues.apache.org/jira/browse/LUCENE-3688?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715581#comment-13715581 ] Sivan Yogev commented on LUCENE-3688: - I want to do this, but not sure when I will get to it. Bucketing of association value of a certain category Key: LUCENE-3688 URL: https://issues.apache.org/jira/browse/LUCENE-3688 Project: Lucene - Core Issue Type: Improvement Components: modules/facet Reporter: Sivan Yogev Priority: Minor Labels: bucket, facet Attachments: LUCENE-association-buckets.patch, LUCENE-association-buckets.r1292224.patch Add facet requests which collect associations of a certain category into buckets, and returns each bucket as a facet result node. Association type is either int or float, and there are two methods to define buckets. The first by providing buckets which contain pre-defined ranges. The other is by declaring the required number of buckets, where the ranges of the different buckets are dynamicly set to create balanced bucket sizes. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-5036) Solr Ref Guide updates for Solr 4.4
[ https://issues.apache.org/jira/browse/SOLR-5036?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715580#comment-13715580 ] Hoss Man commented on SOLR-5036: FYI: I've taken the list of the remaining edits that need to be made, and split them all out into sub-tasks assigned to the individuals who have at one point or another told me they were working on (or planed to work on) these edits. Solr Ref Guide updates for Solr 4.4 --- Key: SOLR-5036 URL: https://issues.apache.org/jira/browse/SOLR-5036 Project: Solr Issue Type: Improvement Components: documentation Reporter: Cassandra Targett Fix For: 4.4 In response to Hoss' email about updating the Solr Ref Guide, I thought it would be helpful if I went through the CHANGES.txt file and noted what I think might need an update (or a new page) and what pages should be reviewed for each change. I hope this helps those who aren't yet fully versed with the way it's organized. I commented on every item in CHANGES.txt, so I'll post my suggestions in a couple of comments. Many items (like bug fixes) don't really need updates, but more eyes on that would be helpful. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-5064) Update dependency org=org.codehaus.woodstox name=wstx-asl rev=3.2.7 transitive=false/ to a more recent version.
Mark Miller created SOLR-5064: - Summary: Update dependency org=org.codehaus.woodstox name=wstx-asl rev=3.2.7 transitive=false/ to a more recent version. Key: SOLR-5064 URL: https://issues.apache.org/jira/browse/SOLR-5064 Project: Solr Issue Type: Improvement Reporter: Mark Miller Assignee: Mark Miller Priority: Minor Fix For: 5.0, 4.5 @whoschek mentioned to me earlier that we were using a fairly old version. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-5062) 4.4 refguide updates related to shardsplitting and deleteshard
[ https://issues.apache.org/jira/browse/SOLR-5062?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715646#comment-13715646 ] Cassandra Targett commented on SOLR-5062: - [~shalinmangar]: I made some suggestions on how to add doc on this as a comment to the page Hoss referenced. Feel free to edit as needed. 4.4 refguide updates related to shardsplitting and deleteshard -- Key: SOLR-5062 URL: https://issues.apache.org/jira/browse/SOLR-5062 Project: Solr Issue Type: Sub-task Components: documentation Reporter: Hoss Man Assignee: Shalin Shekhar Mangar Fix For: 4.4 breaking off from parent issue... * https://cwiki.apache.org/confluence/display/solr/Collections+API ** in general, we need to review this in page in lite of all the shardsplitting stuff and make sure everything is up to date. ** SOLR-4693: A deleteshard collections API that unloads all replicas of a given shard and then removes it from the cluster state. It will remove only those shards which are INACTIVE or have no range (created for custom sharding). (Anshum Gupta, shalin) *** CT: Add to https://cwiki.apache.org/confluence/display/solr/Collections+API -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-5128) Calling IndexSearcher.searchAfter beyond the number of stored documents causes ArrayIndexOutOfBoundsException
[ https://issues.apache.org/jira/browse/LUCENE-5128?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shai Erera updated LUCENE-5128: --- Attachment: LUCENE-5128.patch I had a chat w/ Mike about what's going on in the test and turns out the test is fine, comments needed some improvements. So I fixed IndexSearcher to throw IllegalArgEx if searchAfter.doc = maxDoc + fixed test comments (I think they better explain what's going on there -- Mike, appreciate your review!) and set after.doc to maxDoc-1. I think it's ready. This all makes sense now (to me). Core tests pass. I'd appreciate some review, but I intend to commit this tomorrow. Calling IndexSearcher.searchAfter beyond the number of stored documents causes ArrayIndexOutOfBoundsException - Key: LUCENE-5128 URL: https://issues.apache.org/jira/browse/LUCENE-5128 Project: Lucene - Core Issue Type: Bug Components: core/search Affects Versions: 4.2 Reporter: crocket Attachments: LUCENE-5128.patch, LUCENE-5128.patch ArrayIndexOutOfBoundsException makes it harder to reason about the cause. Is there a better way to notify programmers of the cause? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-5030) determine document specific steps to publish ref guide with each release
[ https://issues.apache.org/jira/browse/SOLR-5030?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man resolved SOLR-5030. Resolution: Fixed I'm calling this done -- the initial version is live, but it's a living doc that will evolve with each release. determine document specific steps to publish ref guide with each release Key: SOLR-5030 URL: https://issues.apache.org/jira/browse/SOLR-5030 Project: Solr Issue Type: Sub-task Components: documentation Reporter: Hoss Man Assignee: Hoss Man need to document the release process for the documentation... * how is the PDF generated (SOLR-4886 * when should things happen (ie: wait until code release vote is final for last minute doc fixes?) * call a vote on the PDF * where should PDF live once vote is done? * where to link to released PDF(s) from site, moinmoin, confluence nav * what content needs changed after being published (ie: 4.3 - 4.4) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-5128) Calling IndexSearcher.searchAfter beyond the number of stored documents causes ArrayIndexOutOfBoundsException
[ https://issues.apache.org/jira/browse/LUCENE-5128?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715669#comment-13715669 ] Michael McCandless commented on LUCENE-5128: +1, patch looks great. Thanks Shai! Calling IndexSearcher.searchAfter beyond the number of stored documents causes ArrayIndexOutOfBoundsException - Key: LUCENE-5128 URL: https://issues.apache.org/jira/browse/LUCENE-5128 Project: Lucene - Core Issue Type: Bug Components: core/search Affects Versions: 4.2 Reporter: crocket Attachments: LUCENE-5128.patch, LUCENE-5128.patch ArrayIndexOutOfBoundsException makes it harder to reason about the cause. Is there a better way to notify programmers of the cause? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-2894) Implement distributed pivot faceting
[ https://issues.apache.org/jira/browse/SOLR-2894?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Muldowney updated SOLR-2894: --- Attachment: SOLR-2894.patch This version fixes the following: 1) Indecisive faceting not being respected on refinement queries 2) Key not being respected 3) Facet.offset not being respected 4) datetimes breaking when trying to refine One point of contention is this: The SolrExampleTests.java (for the SolrJ stuff) had a check that required pivot facet boolean results as strict Boolean.TRUE as opposed to the string true. This came about from the change that was required to fix datetime. I can't find anywhere else where we require a boolean field's value to equal Boolean.True so I think this test was just an artifact of how the original pivot facetting code was written. As it stands now the SolrExampleTests.doPivotFacetTest:1151 has been changed to true instead of Boolean.TRUE Implement distributed pivot faceting Key: SOLR-2894 URL: https://issues.apache.org/jira/browse/SOLR-2894 Project: Solr Issue Type: Improvement Reporter: Erik Hatcher Fix For: 4.5 Attachments: SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894-reworked.patch Following up on SOLR-792, pivot faceting currently only supports undistributed mode. Distributed pivot faceting needs to be implemented. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-2894) Implement distributed pivot faceting
[ https://issues.apache.org/jira/browse/SOLR-2894?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715674#comment-13715674 ] Andrew Muldowney edited comment on SOLR-2894 at 7/22/13 9:22 PM: - Built on 4_4 This version fixes the following: 1) Indecisive faceting not being respected on refinement queries 2) Key not being respected 3) Facet.offset not being respected 4) datetimes breaking when trying to refine One point of contention is this: The SolrExampleTests.java (for the SolrJ stuff) had a check that required pivot facet boolean results as strict Boolean.TRUE as opposed to the string true. This came about from the change that was required to fix datetime. I can't find anywhere else where we require a boolean field's value to equal Boolean.True so I think this test was just an artifact of how the original pivot facetting code was written. As it stands now the SolrExampleTests.doPivotFacetTest:1151 has been changed to true instead of Boolean.TRUE was (Author: andrew.muldowney): This version fixes the following: 1) Indecisive faceting not being respected on refinement queries 2) Key not being respected 3) Facet.offset not being respected 4) datetimes breaking when trying to refine One point of contention is this: The SolrExampleTests.java (for the SolrJ stuff) had a check that required pivot facet boolean results as strict Boolean.TRUE as opposed to the string true. This came about from the change that was required to fix datetime. I can't find anywhere else where we require a boolean field's value to equal Boolean.True so I think this test was just an artifact of how the original pivot facetting code was written. As it stands now the SolrExampleTests.doPivotFacetTest:1151 has been changed to true instead of Boolean.TRUE Implement distributed pivot faceting Key: SOLR-2894 URL: https://issues.apache.org/jira/browse/SOLR-2894 Project: Solr Issue Type: Improvement Reporter: Erik Hatcher Fix For: 4.5 Attachments: SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894.patch, SOLR-2894-reworked.patch Following up on SOLR-792, pivot faceting currently only supports undistributed mode. Distributed pivot faceting needs to be implemented. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
DISCUSS: First official Solr documentation release, plan to vote on wed.
Now that the 4.4 release vote is official and gradually making it's way to the mirrors, I'd like to draw attention to the efforts that have been made to get the Solr Reference Guide into releasable shape, so that we can plan on holding a VOTE to release the 4.4 version ASAP... https://cwiki.apache.org/confluence/display/solr/Apache+Solr+Reference+Guide https://people.apache.org/~hossman/apache-solr-ref-guide-snapshot.pdf TL;DR: I plan to call a vote to formaly release this doc in ~40 hours. Please help out by reviewing/improving the content -- especially if one of the subtasks of SOLR-5036 is asigned to you (because that means you already agreed you would). - - - For those of you who haven't been following along closely, there were a lot of hicups related to the Apache Confluence instance that slowed us down on getting hte RefGuide content donated by LucidWorks up and running and editable. Some of those issues are still outstanding, notably because the ASF is still in the process of trying ot upgrade to Confluence 5.x, and that process is on hold indefinitely (See SOLR-4618 and it's subtasks for history) which means we may have to re-visit a lot of little things once the Confluence upgrade is complete. In the meantime the Ref Guide is in fact up and live and and editable, and progress has been made on two fronts: * Getting the doc up to date with Solr 4.4 (SOLR-5036) * Getting a process in place to release/publish the doc with each minor solr release. There are 4-5 major additions needed to the doc to cover Solr 4.4 adequately, see the children of SOLR-5036 for details. My hope is that the folks who volunteered to help out in writing those new sections of the doc can work on their edits over the next ~40 hours or so, and then on Wed morning (my time) I plan to vall a VOTE to formally release this doc (i'm volunteering to be the Doc RM this time arround)... http://s.apache.org/exs If you are one of the people assigned an outstanding doc issue, please speak up if you don't think you'll have your edits finished by then, or if you would like assistance with structure, wording or formatting... * SOLR-5059 - sarowe - schemaless schema REST * SOLR-5060 - miller - HDFS * SOLR-5061 - erick - new solr.xml format * SOLR-5062 - shalin - shard splitting review new deleteshard cmd * SOLR-5063 - grant - UI screen for adding docs Even if you are not one of the people listed above, there is still a lot you can do to help out... * Reviewing any/all sections of the guide looking for content improvements that can be made (edit or post comments directly on the affected pages as you find them) * Reviewing the snapshot PDF looking for *general* formatting problems that need to be addressed. (reply to this thread or open a Jira) * Reviewing sanity checking the process docs for how we will be maintaining releasing hte guide looking for potential probems (edit or post comments if you see problems)... https://cwiki.apache.org/confluence/display/solr/Internal+-+Maintaining+Documentation https://cwiki.apache.org/confluence/display/solr/Internal+-+How+To+Publish+This+Documentation If anyone has any general comments / concerns, please reply here. -Hoss - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-trunk-MacOSX (64bit/jdk1.7.0) - Build # 667 - Failure!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-MacOSX/667/ Java: 64bit/jdk1.7.0 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC 1 tests failed. REGRESSION: org.apache.solr.client.solrj.TestBatchUpdate.testWithXml Error Message: IOException occured when talking to server at: https://127.0.0.1:51913/solr/collection1 Stack Trace: org.apache.solr.client.solrj.SolrServerException: IOException occured when talking to server at: https://127.0.0.1:51913/solr/collection1 at __randomizedtesting.SeedInfo.seed([C25C53179091FCC5:69BEEAE11297E5A8]:0) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:435) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180) at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:117) at org.apache.solr.client.solrj.SolrServer.commit(SolrServer.java:168) at org.apache.solr.client.solrj.SolrServer.commit(SolrServer.java:146) at org.apache.solr.client.solrj.TestBatchUpdate.doIt(TestBatchUpdate.java:130) at org.apache.solr.client.solrj.TestBatchUpdate.testWithXml(TestBatchUpdate.java:54) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559) at com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:43) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at
[JENKINS-MAVEN] Lucene-Solr-Maven-trunk #916: POMs out of sync
Build: https://builds.apache.org/job/Lucene-Solr-Maven-trunk/916/ 2 tests failed. FAILED: org.apache.solr.cloud.BasicDistributedZkTest.org.apache.solr.cloud.BasicDistributedZkTest Error Message: 1 thread leaked from SUITE scope at org.apache.solr.cloud.BasicDistributedZkTest: 1) Thread[id=6524, name=recoveryCmdExecutor-3585-thread-1, state=RUNNABLE, group=TGRP-BasicDistributedZkTest] at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339) at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200) at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182) at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:391) at java.net.Socket.connect(Socket.java:579) at org.apache.http.conn.scheme.PlainSocketFactory.connectSocket(PlainSocketFactory.java:127) at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:180) at org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:294) at org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:645) at org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:480) at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:906) at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:805) at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:784) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:365) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180) at org.apache.solr.cloud.SyncStrategy$1.run(SyncStrategy.java:291) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:722) Stack Trace: com.carrotsearch.randomizedtesting.ThreadLeakError: 1 thread leaked from SUITE scope at org.apache.solr.cloud.BasicDistributedZkTest: 1) Thread[id=6524, name=recoveryCmdExecutor-3585-thread-1, state=RUNNABLE, group=TGRP-BasicDistributedZkTest] at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339) at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200) at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182) at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:391) at java.net.Socket.connect(Socket.java:579) at org.apache.http.conn.scheme.PlainSocketFactory.connectSocket(PlainSocketFactory.java:127) at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:180) at org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:294) at org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:645) at org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:480) at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:906) at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:805) at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:784) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:365) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180) at org.apache.solr.cloud.SyncStrategy$1.run(SyncStrategy.java:291) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:722) at __randomizedtesting.SeedInfo.seed([8B462EA1B448E33C]:0) FAILED: org.apache.solr.cloud.BasicDistributedZkTest.org.apache.solr.cloud.BasicDistributedZkTest Error Message: There are still zombie threads that couldn't be terminated: 1) Thread[id=6524, name=recoveryCmdExecutor-3585-thread-1, state=RUNNABLE, group=TGRP-BasicDistributedZkTest] at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339) at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200) at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182) at
[jira] [Created] (SOLR-5065) ParseDoubleFieldUpdateProcessorFactory is unable to parse + in exponent
Jack Krupansky created SOLR-5065: Summary: ParseDoubleFieldUpdateProcessorFactory is unable to parse + in exponent Key: SOLR-5065 URL: https://issues.apache.org/jira/browse/SOLR-5065 Project: Solr Issue Type: Bug Components: update Affects Versions: 4.4 Reporter: Jack Krupansky The ParseDoubleFieldUpdateProcessorFactory is unable to parse the full syntax of Java/JSON scientific notation. Parse fails for 4.5E+10, but does succeed for 4.5E10 and 4.5E-10. Using the schema and config from example-schemaless, I added this data: {code} curl http://localhost:8983/solr/update?commit=true; \ -H 'Content-type:application/json' -d ' [{id: doc-1, a1: Hello World, a2: 123, a3: 123.0, a4: 1.23, a5: 4.5E+10, a6: 123, a7: true, a8: false, a9: true, a10: 2013-07-22, a11: 4.5E10, a12: 4.5E-10, a13: 4.5E+10, a14: 4.5E10, a15: 4.5E-10}]' {code} A query returns: {code} doc str name=iddoc-1/str arr name=a1 strHello World/str /arr arr name=a2 long123/long /arr arr name=a3 double123.0/double /arr arr name=a4 double1.23/double /arr arr name=a5 double4.5E10/double /arr arr name=a6 long123/long /arr arr name=a7 booltrue/bool /arr arr name=a8 boolfalse/bool /arr arr name=a9 booltrue/bool /arr arr name=a10 date2013-07-22T00:00:00Z/date /arr arr name=a11 double4.5E10/double /arr arr name=a12 double4.5E-10/double /arr arr name=a13 str4.5E+10/str /arr arr name=a14 double4.5E10/double /arr arr name=a15 double4.5E-10/double /arr long name=_version_1441308941516537856/long/doc {code} The input value of a13 was the same as a5, but was treated as a string, rather than parsed as a double. So, JSON/Java was able to parse 4.5E+10, but this update processor was not. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-4787) Join Contrib
[ https://issues.apache.org/jira/browse/SOLR-4787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715925#comment-13715925 ] Kranti Parisa commented on SOLR-4787: - Joel, I wanted to try implementing pjoin for the following use case. masterCore = 1M keys (id field, long) childCore = 5M documents (with parentid field, long, whose values are equal to the values of id field in the masterCore And syntax looks like http://localhost:8180/solr/masterCore/select?q=title:afq=({!pjoin%20fromIndex=childCore%20from=parentid%20to=id%20v=$childQ})childQ=(fieldOne:somevalue AND fieldTwo:[1 TO 100]) I am getting SyntaxError, but the same syntax works with normal join. any ideas? Also it seems currently pjoin supports only int keys, can you please update pjoin to allow for longs - Kranti Join Contrib Key: SOLR-4787 URL: https://issues.apache.org/jira/browse/SOLR-4787 Project: Solr Issue Type: New Feature Components: search Affects Versions: 4.2.1 Reporter: Joel Bernstein Priority: Minor Fix For: 4.4 Attachments: SOLR-4787-deadlock-fix.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch This contrib provides a place where different join implementations can be contributed to Solr. This contrib currently includes 2 join implementations. The initial patch was generated from the Solr 4.3 tag. Because of changes in the FieldCache API this patch will only build with Solr 4.2 or above. *PostFilterJoinQParserPlugin aka pjoin* The pjoin provides a join implementation that filters results in one core based on the results of a search in another core. This is similar in functionality to the JoinQParserPlugin but the implementation differs in a couple of important ways. The first way is that the pjoin is designed to work with integer join keys only. So, in order to use pjoin, integer join keys must be included in both the to and from core. The second difference is that the pjoin builds memory structures that are used to quickly connect the join keys. It also uses a custom SolrCache named join to hold intermediate DocSets which are needed to build the join memory structures. So, the pjoin will need more memory then the JoinQParserPlugin to perform the join. The main advantage of the pjoin is that it can scale to join millions of keys between cores. Because it's a PostFilter, it only needs to join records that match the main query. The syntax of the pjoin is the same as the JoinQParserPlugin except that the plugin is referenced by the string pjoin rather then join. fq=\{!pjoin fromCore=collection2 from=id_i to=id_i\}user:customer1 The example filter query above will search the fromCore (collection2) for user:customer1. This query will generate a list of values from the from field that will be used to filter the main query. Only records from the main query, where the to field is present in the from list will be included in the results. The solrconfig.xml in the main query core must contain the reference to the pjoin. queryParser name=pjoin class=org.apache.solr.joins.PostFilterJoinQParserPlugin/ And the join contrib jars must be registed in the solrconfig.xml. lib dir=../../../dist/ regex=solr-joins-\d.*\.jar / The solrconfig.xml in the fromcore must have the join SolrCache configured. cache name=join class=solr.LRUCache size=4096 initialSize=1024 / *ValueSourceJoinParserPlugin aka vjoin* The second implementation is the ValueSourceJoinParserPlugin aka vjoin. This implements a ValueSource function query that can return a value from a second core based on join keys and limiting query. The limiting query can be used to select a specific subset of data from the join core. This allows customer specific relevance data to be stored in a separate core and then joined in the main query. The vjoin is called using the vjoin function query. For example: bf=vjoin(fromCore, fromKey, fromVal, toKey, query) This example shows vjoin being called by the edismax boost function parameter. This example will return the fromVal from the fromCore. The fromKey and toKey are used to link the records from the main query to the records in the fromCore. The query is used to select a specific set of records to join with in fromCore. Currently the fromKey and toKey must be longs but this will change in future versions. Like the pjoin, the join SolrCache is used to hold the join memory structures. To configure the vjoin you must register the ValueSource plugin in the solrconfig.xml as follows: valueSourceParser name=vjoin
[jira] [Comment Edited] (SOLR-4787) Join Contrib
[ https://issues.apache.org/jira/browse/SOLR-4787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715925#comment-13715925 ] Kranti Parisa edited comment on SOLR-4787 at 7/23/13 12:39 AM: --- Joel, I wanted to try implementing pjoin for the following use case. masterCore = 1M keys (id field, long) childCore = 5M documents (with parentid field, long, whose values are equal to the values of id field in the masterCore And syntax looks like http://localhost:8180/solr/masterCore/select?q=title:afq=({!pjoin%20fromIndex=childCore%20from=parentid%20to=id%20v=$childQ})childQ=(fieldOne:somevalue AND fieldTwo:[1 TO 100]) I am getting SyntaxError, but the same syntax works with normal join. any ideas? Also it seems currently pjoin supports only int keys, can you please update pjoin to allow long keys - Kranti was (Author: krantiparisa): Joel, I wanted to try implementing pjoin for the following use case. masterCore = 1M keys (id field, long) childCore = 5M documents (with parentid field, long, whose values are equal to the values of id field in the masterCore And syntax looks like http://localhost:8180/solr/masterCore/select?q=title:afq=({!pjoin%20fromIndex=childCore%20from=parentid%20to=id%20v=$childQ})childQ=(fieldOne:somevalue AND fieldTwo:[1 TO 100]) I am getting SyntaxError, but the same syntax works with normal join. any ideas? Also it seems currently pjoin supports only int keys, can you please update pjoin to allow for longs - Kranti Join Contrib Key: SOLR-4787 URL: https://issues.apache.org/jira/browse/SOLR-4787 Project: Solr Issue Type: New Feature Components: search Affects Versions: 4.2.1 Reporter: Joel Bernstein Priority: Minor Fix For: 4.4 Attachments: SOLR-4787-deadlock-fix.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch This contrib provides a place where different join implementations can be contributed to Solr. This contrib currently includes 2 join implementations. The initial patch was generated from the Solr 4.3 tag. Because of changes in the FieldCache API this patch will only build with Solr 4.2 or above. *PostFilterJoinQParserPlugin aka pjoin* The pjoin provides a join implementation that filters results in one core based on the results of a search in another core. This is similar in functionality to the JoinQParserPlugin but the implementation differs in a couple of important ways. The first way is that the pjoin is designed to work with integer join keys only. So, in order to use pjoin, integer join keys must be included in both the to and from core. The second difference is that the pjoin builds memory structures that are used to quickly connect the join keys. It also uses a custom SolrCache named join to hold intermediate DocSets which are needed to build the join memory structures. So, the pjoin will need more memory then the JoinQParserPlugin to perform the join. The main advantage of the pjoin is that it can scale to join millions of keys between cores. Because it's a PostFilter, it only needs to join records that match the main query. The syntax of the pjoin is the same as the JoinQParserPlugin except that the plugin is referenced by the string pjoin rather then join. fq=\{!pjoin fromCore=collection2 from=id_i to=id_i\}user:customer1 The example filter query above will search the fromCore (collection2) for user:customer1. This query will generate a list of values from the from field that will be used to filter the main query. Only records from the main query, where the to field is present in the from list will be included in the results. The solrconfig.xml in the main query core must contain the reference to the pjoin. queryParser name=pjoin class=org.apache.solr.joins.PostFilterJoinQParserPlugin/ And the join contrib jars must be registed in the solrconfig.xml. lib dir=../../../dist/ regex=solr-joins-\d.*\.jar / The solrconfig.xml in the fromcore must have the join SolrCache configured. cache name=join class=solr.LRUCache size=4096 initialSize=1024 / *ValueSourceJoinParserPlugin aka vjoin* The second implementation is the ValueSourceJoinParserPlugin aka vjoin. This implements a ValueSource function query that can return a value from a second core based on join keys and limiting query. The limiting query can be used to select a specific subset of data from the join core. This allows customer specific relevance data to be stored in a separate core and then joined in the main query. The vjoin is called using the vjoin
[jira] [Comment Edited] (SOLR-4787) Join Contrib
[ https://issues.apache.org/jira/browse/SOLR-4787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715944#comment-13715944 ] Kranti Parisa edited comment on SOLR-4787 at 7/23/13 12:59 AM: --- I did review the PostFilterJoinQParserPlugin code and found that it was expecting fromCore instead of fromIndex (like in normal Join) after trying fromCore now getting the expected error related to INT keys, as my keys are Long. - Kranti was (Author: krantiparisa): I did review the PostFilterJoinQParserPlugin code and found that it was expecting fromCore instead of fromIndex (like in normal Join) after trying fromCore now getting the expected error related to INT keys, as my keys are Long. Join Contrib Key: SOLR-4787 URL: https://issues.apache.org/jira/browse/SOLR-4787 Project: Solr Issue Type: New Feature Components: search Affects Versions: 4.2.1 Reporter: Joel Bernstein Priority: Minor Fix For: 4.4 Attachments: SOLR-4787-deadlock-fix.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch This contrib provides a place where different join implementations can be contributed to Solr. This contrib currently includes 2 join implementations. The initial patch was generated from the Solr 4.3 tag. Because of changes in the FieldCache API this patch will only build with Solr 4.2 or above. *PostFilterJoinQParserPlugin aka pjoin* The pjoin provides a join implementation that filters results in one core based on the results of a search in another core. This is similar in functionality to the JoinQParserPlugin but the implementation differs in a couple of important ways. The first way is that the pjoin is designed to work with integer join keys only. So, in order to use pjoin, integer join keys must be included in both the to and from core. The second difference is that the pjoin builds memory structures that are used to quickly connect the join keys. It also uses a custom SolrCache named join to hold intermediate DocSets which are needed to build the join memory structures. So, the pjoin will need more memory then the JoinQParserPlugin to perform the join. The main advantage of the pjoin is that it can scale to join millions of keys between cores. Because it's a PostFilter, it only needs to join records that match the main query. The syntax of the pjoin is the same as the JoinQParserPlugin except that the plugin is referenced by the string pjoin rather then join. fq=\{!pjoin fromCore=collection2 from=id_i to=id_i\}user:customer1 The example filter query above will search the fromCore (collection2) for user:customer1. This query will generate a list of values from the from field that will be used to filter the main query. Only records from the main query, where the to field is present in the from list will be included in the results. The solrconfig.xml in the main query core must contain the reference to the pjoin. queryParser name=pjoin class=org.apache.solr.joins.PostFilterJoinQParserPlugin/ And the join contrib jars must be registed in the solrconfig.xml. lib dir=../../../dist/ regex=solr-joins-\d.*\.jar / The solrconfig.xml in the fromcore must have the join SolrCache configured. cache name=join class=solr.LRUCache size=4096 initialSize=1024 / *ValueSourceJoinParserPlugin aka vjoin* The second implementation is the ValueSourceJoinParserPlugin aka vjoin. This implements a ValueSource function query that can return a value from a second core based on join keys and limiting query. The limiting query can be used to select a specific subset of data from the join core. This allows customer specific relevance data to be stored in a separate core and then joined in the main query. The vjoin is called using the vjoin function query. For example: bf=vjoin(fromCore, fromKey, fromVal, toKey, query) This example shows vjoin being called by the edismax boost function parameter. This example will return the fromVal from the fromCore. The fromKey and toKey are used to link the records from the main query to the records in the fromCore. The query is used to select a specific set of records to join with in fromCore. Currently the fromKey and toKey must be longs but this will change in future versions. Like the pjoin, the join SolrCache is used to hold the join memory structures. To configure the vjoin you must register the ValueSource plugin in the solrconfig.xml as follows: valueSourceParser name=vjoin class=org.apache.solr.joins.ValueSourceJoinParserPlugin / -- This
[jira] [Commented] (SOLR-4787) Join Contrib
[ https://issues.apache.org/jira/browse/SOLR-4787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715944#comment-13715944 ] Kranti Parisa commented on SOLR-4787: - I did review the PostFilterJoinQParserPlugin code and found that it was expecting fromCore instead of fromIndex (like in normal Join) after trying fromCore now getting the expected error related to INT keys, as my keys are Long. Join Contrib Key: SOLR-4787 URL: https://issues.apache.org/jira/browse/SOLR-4787 Project: Solr Issue Type: New Feature Components: search Affects Versions: 4.2.1 Reporter: Joel Bernstein Priority: Minor Fix For: 4.4 Attachments: SOLR-4787-deadlock-fix.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch This contrib provides a place where different join implementations can be contributed to Solr. This contrib currently includes 2 join implementations. The initial patch was generated from the Solr 4.3 tag. Because of changes in the FieldCache API this patch will only build with Solr 4.2 or above. *PostFilterJoinQParserPlugin aka pjoin* The pjoin provides a join implementation that filters results in one core based on the results of a search in another core. This is similar in functionality to the JoinQParserPlugin but the implementation differs in a couple of important ways. The first way is that the pjoin is designed to work with integer join keys only. So, in order to use pjoin, integer join keys must be included in both the to and from core. The second difference is that the pjoin builds memory structures that are used to quickly connect the join keys. It also uses a custom SolrCache named join to hold intermediate DocSets which are needed to build the join memory structures. So, the pjoin will need more memory then the JoinQParserPlugin to perform the join. The main advantage of the pjoin is that it can scale to join millions of keys between cores. Because it's a PostFilter, it only needs to join records that match the main query. The syntax of the pjoin is the same as the JoinQParserPlugin except that the plugin is referenced by the string pjoin rather then join. fq=\{!pjoin fromCore=collection2 from=id_i to=id_i\}user:customer1 The example filter query above will search the fromCore (collection2) for user:customer1. This query will generate a list of values from the from field that will be used to filter the main query. Only records from the main query, where the to field is present in the from list will be included in the results. The solrconfig.xml in the main query core must contain the reference to the pjoin. queryParser name=pjoin class=org.apache.solr.joins.PostFilterJoinQParserPlugin/ And the join contrib jars must be registed in the solrconfig.xml. lib dir=../../../dist/ regex=solr-joins-\d.*\.jar / The solrconfig.xml in the fromcore must have the join SolrCache configured. cache name=join class=solr.LRUCache size=4096 initialSize=1024 / *ValueSourceJoinParserPlugin aka vjoin* The second implementation is the ValueSourceJoinParserPlugin aka vjoin. This implements a ValueSource function query that can return a value from a second core based on join keys and limiting query. The limiting query can be used to select a specific subset of data from the join core. This allows customer specific relevance data to be stored in a separate core and then joined in the main query. The vjoin is called using the vjoin function query. For example: bf=vjoin(fromCore, fromKey, fromVal, toKey, query) This example shows vjoin being called by the edismax boost function parameter. This example will return the fromVal from the fromCore. The fromKey and toKey are used to link the records from the main query to the records in the fromCore. The query is used to select a specific set of records to join with in fromCore. Currently the fromKey and toKey must be longs but this will change in future versions. Like the pjoin, the join SolrCache is used to hold the join memory structures. To configure the vjoin you must register the ValueSource plugin in the solrconfig.xml as follows: valueSourceParser name=vjoin class=org.apache.solr.joins.ValueSourceJoinParserPlugin / -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands,
[jira] [Updated] (SOLR-5043) hostanme lookup in SystemInfoHandler should be refactored to not block core (re)load
[ https://issues.apache.org/jira/browse/SOLR-5043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man updated SOLR-5043: --- Attachment: SOLR-5043.patch Here's a quick and drity patch that spins up background thread to resolve the hostname. in practice, this works fine -- startup and core reload times are short, and an error eventually gets logged by the background thread if/when the JVM gives up on resolving the hostname. the hitch is: if we assume people might have bad DNS when running solr, we should probably also assume they might have bad DNS when running solr tests. As things stand on trunk today, that just means your tests run dog ass slow -- with this patch, the test will finish fast, but then the runner will fail because it treats this background DNS thread as a leak. probably need to figure out a better way of dealing with this. hostanme lookup in SystemInfoHandler should be refactored to not block core (re)load Key: SOLR-5043 URL: https://issues.apache.org/jira/browse/SOLR-5043 Project: Solr Issue Type: Improvement Reporter: Hoss Man Attachments: SOLR-5043.patch SystemInfoHandler currently lookups the hostname of the machine on it's init, and caches for it's lifecycle -- there is a comment to the effect that the reason for this is because on some machines (notably ones with wacky DNS settings) looking up the hostname can take a long ass time in some JVMs... {noformat} // on some platforms, resolving canonical hostname can cause the thread // to block for several seconds if nameservices aren't available // so resolve this once per handler instance //(ie: not static, so core reload will refresh) {noformat} But as we move forward with a lot more multi-core, solr-cloud, dynamically updated instances, even paying this cost per core-reload is expensive. we should refactoring this so that SystemInfoHandler instances init immediately, with some kind of lazy loading of the hostname info in a background thread, (especially since hte only real point of having that info here is for UI use so you cna keep track of what machine you are looking at) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
hello
http://erego.ro/hcnwz/vmwjtgjgjhogxwsitvwganem Arvind Srini 7/23/2013 2:24:19 AM - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-5061) 4.4 refguide pages new solr.xml format
[ https://issues.apache.org/jira/browse/SOLR-5061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715964#comment-13715964 ] Erick Erickson commented on SOLR-5061: -- Ok, first cut done. There are some configuration parameters that I have no real clue how to document. If anyone can take a quick look at these (or a long look and edit or comment what I should do) I'd appreciate it a lot. These pages: https://cwiki.apache.org/confluence/display/solr/Solr+Cores+and+solr.xml https://cwiki.apache.org/confluence/display/solr/Format+of+solr.xml https://cwiki.apache.org/confluence/pages/editpage.action?pageId=33296227 https://cwiki.apache.org/confluence/display/solr/Moving+to+the+New+solr.xml+Format https://cwiki.apache.org/confluence/display/solr/CoreAdminHandler+Parameters+and+Usage 4.4 refguide pages new solr.xml format -- Key: SOLR-5061 URL: https://issues.apache.org/jira/browse/SOLR-5061 Project: Solr Issue Type: Sub-task Components: documentation Reporter: Hoss Man Assignee: Erick Erickson Fix For: 4.4 breaking off from parent issue... * https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml ** SOLR-4757: Change the example to use the new solr.xml format and core discovery by directory structure. (Mark Miller) *** CT: There is a page on solr.xml: https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml. This should be updated to show the new format and still include information on the old format for anyone with the old format who uses this guide for reference. ** SOLR-4655: Add option to have Overseer assign generic node names so that new addresses can host shards without naming confusion. (Mark Miller, Anshum Gupta) *** CT: I think this only needs to be added to any new content for solr.xml at https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml It should also be noted that cassandra posted some additional deailed suggests in a comment on the existing page in the ref guide... https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml?focusedCommentId=33296160#comment-33296160 -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Release documentation for new solr.xml formats needs more eyes.
These pages could use some review. In particular I don't have some definitions for some of the solr.xml options documented, look for ??? particularly in the Format of solr.xml and https://cwiki.apache.org/confluence/display/solr/Solr+Cores+and+solr.xml https://cwiki.apache.org/confluence/display/solr/Format+of+solr.xml https://cwiki.apache.org/confluence/display/solr/Legacy+solr.xml+Configuration https://cwiki.apache.org/confluence/display/solr/Moving+to+the+New+solr.xml+Format https://cwiki.apache.org/confluence/display/solr/CoreAdminHandler+Parameters+and+Usage The Moving to the new Solr.xml Format is particularly weak, but I'm starting to get past the point where I can concentrate so I'll look at it again in the morning. Thanks, Erick - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-5061) 4.4 refguide pages new solr.xml format
[ https://issues.apache.org/jira/browse/SOLR-5061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715964#comment-13715964 ] Erick Erickson edited comment on SOLR-5061 at 7/23/13 1:35 AM: --- Ok, first cut done. There are some configuration parameters that I have no real clue how to document. If anyone can take a quick look at these (or a long look and edit or comment what I should do) I'd appreciate it a lot. These pages: https://cwiki.apache.org/confluence/display/solr/Solr+Cores+and+solr.xml https://cwiki.apache.org/confluence/display/solr/Format+of+solr.xml https://cwiki.apache.org/confluence/display/solr/Legacy+solr.xml+Configuration https://cwiki.apache.org/confluence/display/solr/Moving+to+the+New+solr.xml+Format https://cwiki.apache.org/confluence/display/solr/CoreAdminHandler+Parameters+and+Usage was (Author: erickerickson): Ok, first cut done. There are some configuration parameters that I have no real clue how to document. If anyone can take a quick look at these (or a long look and edit or comment what I should do) I'd appreciate it a lot. These pages: https://cwiki.apache.org/confluence/display/solr/Solr+Cores+and+solr.xml https://cwiki.apache.org/confluence/display/solr/Format+of+solr.xml https://cwiki.apache.org/confluence/pages/editpage.action?pageId=33296227 https://cwiki.apache.org/confluence/display/solr/Moving+to+the+New+solr.xml+Format https://cwiki.apache.org/confluence/display/solr/CoreAdminHandler+Parameters+and+Usage 4.4 refguide pages new solr.xml format -- Key: SOLR-5061 URL: https://issues.apache.org/jira/browse/SOLR-5061 Project: Solr Issue Type: Sub-task Components: documentation Reporter: Hoss Man Assignee: Erick Erickson Fix For: 4.4 breaking off from parent issue... * https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml ** SOLR-4757: Change the example to use the new solr.xml format and core discovery by directory structure. (Mark Miller) *** CT: There is a page on solr.xml: https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml. This should be updated to show the new format and still include information on the old format for anyone with the old format who uses this guide for reference. ** SOLR-4655: Add option to have Overseer assign generic node names so that new addresses can host shards without naming confusion. (Mark Miller, Anshum Gupta) *** CT: I think this only needs to be added to any new content for solr.xml at https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml It should also be noted that cassandra posted some additional deailed suggests in a comment on the existing page in the ref guide... https://cwiki.apache.org/confluence/display/solr/Core+Admin+and+Configuring+solr.xml?focusedCommentId=33296160#comment-33296160 -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
hey.
http://frenchpornographer.com/cilj/mgbgagpdugov Arvind Srini 7/23/2013 3:20:00 AM - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-4787) Join Contrib
[ https://issues.apache.org/jira/browse/SOLR-4787?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kranti Parisa updated SOLR-4787: Attachment: SOLR-4787-pjoin-long-keys.patch Update PostFilterJoinQParserPlugin to support LONG keys Join Contrib Key: SOLR-4787 URL: https://issues.apache.org/jira/browse/SOLR-4787 Project: Solr Issue Type: New Feature Components: search Affects Versions: 4.2.1 Reporter: Joel Bernstein Priority: Minor Fix For: 4.4 Attachments: SOLR-4787-deadlock-fix.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787-pjoin-long-keys.patch This contrib provides a place where different join implementations can be contributed to Solr. This contrib currently includes 2 join implementations. The initial patch was generated from the Solr 4.3 tag. Because of changes in the FieldCache API this patch will only build with Solr 4.2 or above. *PostFilterJoinQParserPlugin aka pjoin* The pjoin provides a join implementation that filters results in one core based on the results of a search in another core. This is similar in functionality to the JoinQParserPlugin but the implementation differs in a couple of important ways. The first way is that the pjoin is designed to work with integer join keys only. So, in order to use pjoin, integer join keys must be included in both the to and from core. The second difference is that the pjoin builds memory structures that are used to quickly connect the join keys. It also uses a custom SolrCache named join to hold intermediate DocSets which are needed to build the join memory structures. So, the pjoin will need more memory then the JoinQParserPlugin to perform the join. The main advantage of the pjoin is that it can scale to join millions of keys between cores. Because it's a PostFilter, it only needs to join records that match the main query. The syntax of the pjoin is the same as the JoinQParserPlugin except that the plugin is referenced by the string pjoin rather then join. fq=\{!pjoin fromCore=collection2 from=id_i to=id_i\}user:customer1 The example filter query above will search the fromCore (collection2) for user:customer1. This query will generate a list of values from the from field that will be used to filter the main query. Only records from the main query, where the to field is present in the from list will be included in the results. The solrconfig.xml in the main query core must contain the reference to the pjoin. queryParser name=pjoin class=org.apache.solr.joins.PostFilterJoinQParserPlugin/ And the join contrib jars must be registed in the solrconfig.xml. lib dir=../../../dist/ regex=solr-joins-\d.*\.jar / The solrconfig.xml in the fromcore must have the join SolrCache configured. cache name=join class=solr.LRUCache size=4096 initialSize=1024 / *ValueSourceJoinParserPlugin aka vjoin* The second implementation is the ValueSourceJoinParserPlugin aka vjoin. This implements a ValueSource function query that can return a value from a second core based on join keys and limiting query. The limiting query can be used to select a specific subset of data from the join core. This allows customer specific relevance data to be stored in a separate core and then joined in the main query. The vjoin is called using the vjoin function query. For example: bf=vjoin(fromCore, fromKey, fromVal, toKey, query) This example shows vjoin being called by the edismax boost function parameter. This example will return the fromVal from the fromCore. The fromKey and toKey are used to link the records from the main query to the records in the fromCore. The query is used to select a specific set of records to join with in fromCore. Currently the fromKey and toKey must be longs but this will change in future versions. Like the pjoin, the join SolrCache is used to hold the join memory structures. To configure the vjoin you must register the ValueSource plugin in the solrconfig.xml as follows: valueSourceParser name=vjoin class=org.apache.solr.joins.ValueSourceJoinParserPlugin / -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-4787) Join Contrib
[ https://issues.apache.org/jira/browse/SOLR-4787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13715987#comment-13715987 ] Kranti Parisa edited comment on SOLR-4787 at 7/23/13 2:53 AM: -- Attached the patch file (SOLR-4787-pjoin-long-keys.patch) for PostFilterJoinQParserPlugin to support LONG keys was (Author: krantiparisa): Update PostFilterJoinQParserPlugin to support LONG keys Join Contrib Key: SOLR-4787 URL: https://issues.apache.org/jira/browse/SOLR-4787 Project: Solr Issue Type: New Feature Components: search Affects Versions: 4.2.1 Reporter: Joel Bernstein Priority: Minor Fix For: 4.4 Attachments: SOLR-4787-deadlock-fix.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787.patch, SOLR-4787-pjoin-long-keys.patch This contrib provides a place where different join implementations can be contributed to Solr. This contrib currently includes 2 join implementations. The initial patch was generated from the Solr 4.3 tag. Because of changes in the FieldCache API this patch will only build with Solr 4.2 or above. *PostFilterJoinQParserPlugin aka pjoin* The pjoin provides a join implementation that filters results in one core based on the results of a search in another core. This is similar in functionality to the JoinQParserPlugin but the implementation differs in a couple of important ways. The first way is that the pjoin is designed to work with integer join keys only. So, in order to use pjoin, integer join keys must be included in both the to and from core. The second difference is that the pjoin builds memory structures that are used to quickly connect the join keys. It also uses a custom SolrCache named join to hold intermediate DocSets which are needed to build the join memory structures. So, the pjoin will need more memory then the JoinQParserPlugin to perform the join. The main advantage of the pjoin is that it can scale to join millions of keys between cores. Because it's a PostFilter, it only needs to join records that match the main query. The syntax of the pjoin is the same as the JoinQParserPlugin except that the plugin is referenced by the string pjoin rather then join. fq=\{!pjoin fromCore=collection2 from=id_i to=id_i\}user:customer1 The example filter query above will search the fromCore (collection2) for user:customer1. This query will generate a list of values from the from field that will be used to filter the main query. Only records from the main query, where the to field is present in the from list will be included in the results. The solrconfig.xml in the main query core must contain the reference to the pjoin. queryParser name=pjoin class=org.apache.solr.joins.PostFilterJoinQParserPlugin/ And the join contrib jars must be registed in the solrconfig.xml. lib dir=../../../dist/ regex=solr-joins-\d.*\.jar / The solrconfig.xml in the fromcore must have the join SolrCache configured. cache name=join class=solr.LRUCache size=4096 initialSize=1024 / *ValueSourceJoinParserPlugin aka vjoin* The second implementation is the ValueSourceJoinParserPlugin aka vjoin. This implements a ValueSource function query that can return a value from a second core based on join keys and limiting query. The limiting query can be used to select a specific subset of data from the join core. This allows customer specific relevance data to be stored in a separate core and then joined in the main query. The vjoin is called using the vjoin function query. For example: bf=vjoin(fromCore, fromKey, fromVal, toKey, query) This example shows vjoin being called by the edismax boost function parameter. This example will return the fromVal from the fromCore. The fromKey and toKey are used to link the records from the main query to the records in the fromCore. The query is used to select a specific set of records to join with in fromCore. Currently the fromKey and toKey must be longs but this will change in future versions. Like the pjoin, the join SolrCache is used to hold the join memory structures. To configure the vjoin you must register the ValueSource plugin in the solrconfig.xml as follows: valueSourceParser name=vjoin class=org.apache.solr.joins.ValueSourceJoinParserPlugin / -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira - To