[jira] [Updated] (LUCENENET-480) Investigate what needs to happen to make both .NET 3.5 and 4.0 builds possible
[ https://issues.apache.org/jira/browse/LUCENENET-480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Prescott Nasser updated LUCENENET-480: -- Attachment: SortedSet.cs Investigate what needs to happen to make both .NET 3.5 and 4.0 builds possible -- Key: LUCENENET-480 URL: https://issues.apache.org/jira/browse/LUCENENET-480 Project: Lucene.Net Issue Type: Task Components: Lucene.Net Contrib, Lucene.Net Core, Lucene.Net Demo, Lucene.Net Test Affects Versions: Lucene.Net 2.9.4, Lucene.Net 2.9.4g, Lucene.Net 3.0.3 Reporter: Christopher Currens Attachments: SortedSet.cs We need to investigate what needs to be done with the source to be able to support both a .NET 3.5 and 4.0 build. There was some concern from at least one member of the community ([see here|http://mail-archives.apache.org/mod_mbox/lucene-lucene-net-dev/201202.mbox/%3C004b01cce111$871f4990$955ddcb0$@fr%3E]) that we've alienated some of our user base by making such a heavy jump from 2.0 to 4.0. There are major benefits to using 4.0 over the 2.0-based runtimes, specifically FAR superior memory management, particularly with the LOH, where Lucene.NET has had major trouble with in the past. Based on what has been done with Lucene.NET 3.0.3, we can't (easily) drop .NET 3.5/3.0 libraries and go back to 2.0. HashSetT and ISetT is used extensively in the code, that would be a major blocker to get done. I suppose it could be possible, but there hasn't been a whole lot of talk about what runtimes we intend to support. The big change between lucene 2.x and 3.x for java was also a change in runtimes, that allowed generics and enums to be used in the base code. We have a similar situation with the API (it's substantially different, with the addition of properties alone) for this next version of Lucene.NET, so I think it's reasonable to at least make a permanent move from 2.0 to 3.5, though that's only my opinion, hasn't been discussed with the committers. It seems that supporting 3.5 and 4.0 should be fairly easy, and is a decent compromise in supporting both a 2.0 and 4.0 runtime. At the very least, we should try our best to support it. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (LUCENENET-480) Investigate what needs to happen to make both .NET 3.5 and 4.0 builds possible
[ https://issues.apache.org/jira/browse/LUCENENET-480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13393683#comment-13393683 ] Prescott Nasser commented on LUCENENET-480: --- I agree, I'd prefer to leave ISet, however, if most of hte ISetT x = new HashSetT, I can't modify hashset to inherit from my own ISet Implementation. So I can either, 1. rewrite hashset to implement ISet (and inherit HashSet) or we can change the code from ISet. Attached is my incomplete SortedSet - i found it from another implementation based on sortedlist. Investigate what needs to happen to make both .NET 3.5 and 4.0 builds possible -- Key: LUCENENET-480 URL: https://issues.apache.org/jira/browse/LUCENENET-480 Project: Lucene.Net Issue Type: Task Components: Lucene.Net Contrib, Lucene.Net Core, Lucene.Net Demo, Lucene.Net Test Affects Versions: Lucene.Net 2.9.4, Lucene.Net 2.9.4g, Lucene.Net 3.0.3 Reporter: Christopher Currens Attachments: SortedSet.cs We need to investigate what needs to be done with the source to be able to support both a .NET 3.5 and 4.0 build. There was some concern from at least one member of the community ([see here|http://mail-archives.apache.org/mod_mbox/lucene-lucene-net-dev/201202.mbox/%3C004b01cce111$871f4990$955ddcb0$@fr%3E]) that we've alienated some of our user base by making such a heavy jump from 2.0 to 4.0. There are major benefits to using 4.0 over the 2.0-based runtimes, specifically FAR superior memory management, particularly with the LOH, where Lucene.NET has had major trouble with in the past. Based on what has been done with Lucene.NET 3.0.3, we can't (easily) drop .NET 3.5/3.0 libraries and go back to 2.0. HashSetT and ISetT is used extensively in the code, that would be a major blocker to get done. I suppose it could be possible, but there hasn't been a whole lot of talk about what runtimes we intend to support. The big change between lucene 2.x and 3.x for java was also a change in runtimes, that allowed generics and enums to be used in the base code. We have a similar situation with the API (it's substantially different, with the addition of properties alone) for this next version of Lucene.NET, so I think it's reasonable to at least make a permanent move from 2.0 to 3.5, though that's only my opinion, hasn't been discussed with the committers. It seems that supporting 3.5 and 4.0 should be fairly easy, and is a decent compromise in supporting both a 2.0 and 4.0 runtime. At the very least, we should try our best to support it. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Comment Edited] (LUCENENET-480) Investigate what needs to happen to make both .NET 3.5 and 4.0 builds possible
[ https://issues.apache.org/jira/browse/LUCENENET-480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13393683#comment-13393683 ] Prescott Nasser edited comment on LUCENENET-480 at 6/18/12 6:30 AM: I agree, I'd prefer to leave ISetT. However, if most of the ISetT x = new HashSetT, I can't modify Hashset to inherit from my own ISet Implementation. So I can either rewrite hashset to implement ISet (and inherit HashSet) or we can change the code from ISet. Attached is my incomplete SortedSet - i found it from another implementation based on sortedlist. was (Author: pnasser): I agree, I'd prefer to leave ISet, however, if most of hte ISetT x = new HashSetT, I can't modify hashset to inherit from my own ISet Implementation. So I can either, 1. rewrite hashset to implement ISet (and inherit HashSet) or we can change the code from ISet. Attached is my incomplete SortedSet - i found it from another implementation based on sortedlist. Investigate what needs to happen to make both .NET 3.5 and 4.0 builds possible -- Key: LUCENENET-480 URL: https://issues.apache.org/jira/browse/LUCENENET-480 Project: Lucene.Net Issue Type: Task Components: Lucene.Net Contrib, Lucene.Net Core, Lucene.Net Demo, Lucene.Net Test Affects Versions: Lucene.Net 2.9.4, Lucene.Net 2.9.4g, Lucene.Net 3.0.3 Reporter: Christopher Currens Attachments: SortedSet.cs We need to investigate what needs to be done with the source to be able to support both a .NET 3.5 and 4.0 build. There was some concern from at least one member of the community ([see here|http://mail-archives.apache.org/mod_mbox/lucene-lucene-net-dev/201202.mbox/%3C004b01cce111$871f4990$955ddcb0$@fr%3E]) that we've alienated some of our user base by making such a heavy jump from 2.0 to 4.0. There are major benefits to using 4.0 over the 2.0-based runtimes, specifically FAR superior memory management, particularly with the LOH, where Lucene.NET has had major trouble with in the past. Based on what has been done with Lucene.NET 3.0.3, we can't (easily) drop .NET 3.5/3.0 libraries and go back to 2.0. HashSetT and ISetT is used extensively in the code, that would be a major blocker to get done. I suppose it could be possible, but there hasn't been a whole lot of talk about what runtimes we intend to support. The big change between lucene 2.x and 3.x for java was also a change in runtimes, that allowed generics and enums to be used in the base code. We have a similar situation with the API (it's substantially different, with the addition of properties alone) for this next version of Lucene.NET, so I think it's reasonable to at least make a permanent move from 2.0 to 3.5, though that's only my opinion, hasn't been discussed with the committers. It seems that supporting 3.5 and 4.0 should be fairly easy, and is a decent compromise in supporting both a 2.0 and 4.0 runtime. At the very least, we should try our best to support it. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (LUCENENET-484) Some possibly major tests intermittently fail
[ https://issues.apache.org/jira/browse/LUCENENET-484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395775#comment-13395775 ] Luc Vanlerberghe commented on LUCENENET-484: TestStressLocks and TestStressLocksNativeFSLockFactory fail intermittently due to an UnauthorizedAccessException in NativeFSLockFactory.Obtain In the code it is actually described that on Windows you can get intermittent Access Denied exceptions and the java version handles them properly. *BUT*: in java, the Access Denied raises an IOException, whereas in .Net this raises an UnauthorizedAccessException So the solution is to add an identical catch block for the UnauthorizedAccessException. This will cause the lock obtain to simply fail and the calling methods to retry. Some possibly major tests intermittently fail -- Key: LUCENENET-484 URL: https://issues.apache.org/jira/browse/LUCENENET-484 Project: Lucene.Net Issue Type: Bug Components: Lucene.Net Core, Lucene.Net Test Affects Versions: Lucene.Net 3.0.3 Environment: All Reporter: Christopher Currens Fix For: Lucene.Net 3.0.3 Attachments: Lucenenet-484-NativeFSLockFactory.patch, Lucenenet-484-WeakDictionary.patch, Lucenenet-484-WeakDictionaryTests.patch These tests will fail intermittently in Debug or Release mode, in the core test suite: # -Lucene.Net.Index:- #- -TestConcurrentMergeScheduler.TestFlushExceptions- # Lucene.Net.Store: #- TestLockFactory.TestStressLocks # Lucene.Net.Search: #- TestSort.TestParallelMultiSort # -Lucene.Net.Util:- #- -TestFieldCacheSanityChecker.TestInsanity1- #- -TestFieldCacheSanityChecker.TestInsanity2- #- -(It's possible all of the insanity tests fail at one point or another)- # -Lucene.Net.Support- #- -TestWeakHashTableMultiThreadAccess.Test- TestWeakHashTableMultiThreadAccess should be fine to remove along with the WeakHashTable in the Support namespace, since it's been replaced with WeakDictionary. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (LUCENENET-484) Some possibly major tests intermittently fail
[ https://issues.apache.org/jira/browse/LUCENENET-484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Luc Vanlerberghe updated LUCENENET-484: --- Attachment: Lucenenet-484-NativeFSLockFactory.patch Fixes the intermittent failures in TestStressLocks and TestStressLocksNativeFSLockFactory Some possibly major tests intermittently fail -- Key: LUCENENET-484 URL: https://issues.apache.org/jira/browse/LUCENENET-484 Project: Lucene.Net Issue Type: Bug Components: Lucene.Net Core, Lucene.Net Test Affects Versions: Lucene.Net 3.0.3 Environment: All Reporter: Christopher Currens Fix For: Lucene.Net 3.0.3 Attachments: Lucenenet-484-NativeFSLockFactory.patch, Lucenenet-484-WeakDictionary.patch, Lucenenet-484-WeakDictionaryTests.patch These tests will fail intermittently in Debug or Release mode, in the core test suite: # -Lucene.Net.Index:- #- -TestConcurrentMergeScheduler.TestFlushExceptions- # Lucene.Net.Store: #- TestLockFactory.TestStressLocks # Lucene.Net.Search: #- TestSort.TestParallelMultiSort # -Lucene.Net.Util:- #- -TestFieldCacheSanityChecker.TestInsanity1- #- -TestFieldCacheSanityChecker.TestInsanity2- #- -(It's possible all of the insanity tests fail at one point or another)- # -Lucene.Net.Support- #- -TestWeakHashTableMultiThreadAccess.Test- TestWeakHashTableMultiThreadAccess should be fine to remove along with the WeakHashTable in the Support namespace, since it's been replaced with WeakDictionary. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (LUCENENET-493) Make lucene.net culture insensitive (like the java version)
[ https://issues.apache.org/jira/browse/LUCENENET-493?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395818#comment-13395818 ] Luc Vanlerberghe commented on LUCENENET-493: Perhaps my comment was not sufficiently clear: I didn't solve a problem in the testcases, I solved a problem in Lucene.Net itself I happened to stumble upon the problem because I executed the tests in a non-US locale. The java version tests if Lucene behaves independently of the culture. Even if we can't duplicate the way this is tested, we should still make sure Lucene.net behaves independently of the culture too. When the patch is applied, as a side-effect, the tests also pass when run in another culture. Make lucene.net culture insensitive (like the java version) --- Key: LUCENENET-493 URL: https://issues.apache.org/jira/browse/LUCENENET-493 Project: Lucene.Net Issue Type: Bug Components: Lucene.Net Core, Lucene.Net Test Affects Versions: Lucene.Net 3.0.3 Reporter: Luc Vanlerberghe Labels: patch Fix For: Lucene.Net 3.0.3 Attachments: Lucenenet-493.patch In Java, conversion of the basic types to and from strings is locale (culture) independent. For localized input/output one needs to use the classes in the java.text package. In .Net, conversion of the basic types to and from strings depends on the default Culture. Otherwise you have to specify CultureInfo.InvariantCulture explicitly. Some of the testcases in lucene.net fail if they are not run on a machine with culture set to US. In the current version of lucene.net there are patches here and there that try to correct for some specific cases by using string replacement (like System.Double.Parse(s.Replace(., CultureInfo.CurrentCulture.NumberFormat.NumberDecimalSeparator)), but that seems really ugly. I submit a patch here that removes the old workarounds and replaces them by calls to classes in the Lucene.Net.Support namespace that try to handle the conversions in a compatible way. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (LUCENENET-495) Use of DateTime.Now causes huge amount of System.Globalization.DaylightTime object allocations
[ https://issues.apache.org/jira/browse/LUCENENET-495?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Christopher Currens resolved LUCENENET-495. --- Resolution: Fixed Use of DateTime.Now causes huge amount of System.Globalization.DaylightTime object allocations -- Key: LUCENENET-495 URL: https://issues.apache.org/jira/browse/LUCENENET-495 Project: Lucene.Net Issue Type: Bug Components: Lucene.Net Core Affects Versions: Lucene.Net 2.9.4, Lucene.Net 3.0.3 Reporter: Christopher Currens Assignee: Christopher Currens Priority: Critical Fix For: Lucene.Net 3.0.3 This issue mostly just affects RAMDirectory. However, RAMFile and RAMOutputStream are used in other (all?) directory implementations, including FSDirectory types. In RAMOutputStream, the file last modified property for the RAMFile is updated when the stream is flushed. It's calculated using {{DateTime.Now.Ticks / TimeSpan.TicksPerMillisecond}}. I've read before that Microsoft has regretted making DateTime.Now a property instead of a method, and after seeing what it's doing, I'm starting to understand why. DateTime.Now is returning local time. In order for it to calculate that, it has to get the utf offset for the machine, which requires the creation of a _class_, System.Globalization.DaylightTime. This is bad for performance. Using code to write 10,000 small documents to an index (4kb sizes), it created 1,570,157 of these DaylightTime classes, a total of 62MB of extra memory...clearly RAMOutputStream.Flush() is called a lot. A fix I'd like to propose is to change the RAMFile from storing the LastModified date to UTC instead of local. DateTime.UtcNow doesn't create any additional objects and is very fast. For this small benchmark, the performance increase is 31%. I've set it to convert to local-time, when {{RAMDirectory.LastModified(string name)}} is called to make sure it has the same behavior (tests fail otherwise). Are there any other side-effects to making this change? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (LUCENENET-493) Make lucene.net culture insensitive (like the java version)
[ https://issues.apache.org/jira/browse/LUCENENET-493?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Christopher Currens updated LUCENENET-493: -- Attachment: UpdatedLocalizedTestCase.patch I've updated LocalizedTestCase, so that it will actually run against all installed cultures. The workaround is unfortunate...there is now only 1 test that does all localization checks. It's done more or less the same way that java does it, however, instead of being able to override the {{runBare()}} method, I've created a test method that will run all methods in all installed cultures. If a method were to fail, it would list the method name, and the culture it failed in. I've attached a patch that shows the solution, so if anyone has a better solution, we can discuss that and possibly use it instead. Interestingly enough, the tests that Java Lucene has set to test, don't actually fail when using the older code that doesn't have localization changes in it. However, when I added {{TestBoost}} to the list in {{TestQueryParser}}, that one did fail before the push that Simon did. So, it concerns me that we don't have enough tests written that actually will cause it to fail, when run as a localized test. So, what I propose we do before we apply Luc's patch, is to write tests that *will fail* when using as a LocalizedTestCase and then make sure his patch makes all of the tests pass. Make lucene.net culture insensitive (like the java version) --- Key: LUCENENET-493 URL: https://issues.apache.org/jira/browse/LUCENENET-493 Project: Lucene.Net Issue Type: Bug Components: Lucene.Net Core, Lucene.Net Test Affects Versions: Lucene.Net 3.0.3 Reporter: Luc Vanlerberghe Labels: patch Fix For: Lucene.Net 3.0.3 Attachments: Lucenenet-493.patch, UpdatedLocalizedTestCase.patch In Java, conversion of the basic types to and from strings is locale (culture) independent. For localized input/output one needs to use the classes in the java.text package. In .Net, conversion of the basic types to and from strings depends on the default Culture. Otherwise you have to specify CultureInfo.InvariantCulture explicitly. Some of the testcases in lucene.net fail if they are not run on a machine with culture set to US. In the current version of lucene.net there are patches here and there that try to correct for some specific cases by using string replacement (like System.Double.Parse(s.Replace(., CultureInfo.CurrentCulture.NumberFormat.NumberDecimalSeparator)), but that seems really ugly. I submit a patch here that removes the old workarounds and replaces them by calls to classes in the Lucene.Net.Support namespace that try to handle the conversions in a compatible way. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
Build failed in Jenkins: Lucene-Core-4x-Beasting #2603
See http://sierranevada.servebeer.com/job/Lucene-Core-4x-Beasting/2603/ -- [...truncated 882 lines...] [junit4] Completed on J2 in 1.20s, 7 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestNRTReaderWithThreads [junit4] Completed on J3 in 1.05s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestOmitPositions [junit4] Completed on J0 in 0.14s, 3 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestDocsAndPositions [junit4] Completed on J3 in 0.15s, 6 tests [junit4] [junit4] Suite: org.apache.lucene.store.TestHugeRamFile [junit4] Completed on J0 in 0.45s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestSameTokenSamePosition [junit4] Completed on J0 in 0.06s, 2 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestCheckIndex [junit4] Completed on J0 in 0.03s, 2 tests [junit4] [junit4] Suite: org.apache.lucene.codecs.lucene3x.TestImpersonation [junit4] Completed on J0 in 0.01s, 1 test [junit4] [junit4] Suite: org.apache.lucene.search.TestSort [junit4] Completed on J1 in 0.91s, 25 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestDisjunctionMaxQuery [junit4] Completed on J1 in 0.19s, 10 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestDirectoryReader [junit4] Completed on J3 in 0.57s, 24 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestIndexWriterCommit [junit4] IGNOR/A 0.00s J2 | TestIndexWriterCommit.testCommitOnCloseDiskUsage [junit4] Assumption #1: This test cannot run with Memory codec [junit4] Completed on J2 in 1.09s, 12 tests, 1 skipped [junit4] [junit4] Suite: org.apache.lucene.search.TestMultiThreadTermVectors [junit4] Completed on J3 in 0.26s, 1 test [junit4] [junit4] Suite: org.apache.lucene.search.TestFieldCacheRewriteMethod [junit4] Completed on J2 in 0.34s, 1 test [junit4] [junit4] Suite: org.apache.lucene.search.TestMultiValuedNumericRangeQuery [junit4] Completed on J2 in 0.21s, 1 test [junit4] [junit4] Suite: org.apache.lucene.search.TestDateSort [junit4] Completed on J0 in 0.01s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestRandomStoredFields [junit4] Completed on J3 in 0.42s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestMaxTermFrequency [junit4] Completed on J2 in 0.13s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestFieldsReader [junit4] Completed on J2 in 0.22s, 4 tests [junit4] [junit4] Suite: org.apache.lucene.util.automaton.TestLevenshteinAutomata [junit4] Completed on J2 in 0.49s, 4 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestIndexWriterNRTIsCurrent [junit4] Completed on J3 in 0.77s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestBackwardsCompatibility [junit4] Completed on J0 in 2.75s, 13 tests [junit4] [junit4] Suite: org.apache.lucene.search.spans.TestPayloadSpans [junit4] Completed on J3 in 0.39s, 10 tests [junit4] [junit4] Suite: org.apache.lucene.util.automaton.TestDeterminism [junit4] Completed on J0 in 0.35s, 2 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestTransactions [junit4] Completed on J2 in 0.56s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestPostingsOffsets [junit4] Completed on J2 in 0.24s, 11 tests [junit4] [junit4] Suite: org.apache.lucene.util.automaton.TestDeterminizeLexicon [junit4] Completed on J3 in 0.40s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestCrash [junit4] Completed on J0 in 0.37s, 5 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestFieldCacheRangeFilter [junit4] Completed on J3 in 0.12s, 9 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestTermVectors [junit4] Completed on J0 in 0.11s, 11 tests [junit4] [junit4] Suite: org.apache.lucene.util.TestWeakIdentityMap [junit4] Completed on J2 in 0.37s, 2 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestAutomatonQuery [junit4] Completed on J3 in 0.03s, 6 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestFilteredQuery [junit4] Completed on J0 in 0.12s, 9 tests [junit4] [junit4] Suite: org.apache.lucene.util.TestIndexableBinaryStringTools [junit4] Completed on J2 in 0.08s, 5 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestMultiLevelSkipList [junit4] Completed on J3 in 0.04s, 1 test [junit4] [junit4] Suite: org.apache.lucene.search.TestSimpleExplanations [junit4] Completed on J1 in 3.78s, 69 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestSimpleSearchEquivalence [junit4] Completed on J0 in 0.10s, 12 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestFuzzyQuery [junit4] Completed on J2 in 0.05s, 5 tests
Jenkins build is back to normal : Lucene-Core-4x-Beasting #2604
See http://sierranevada.servebeer.com/job/Lucene-Core-4x-Beasting/2604/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3406) Support grouped range and query facets.
[ https://issues.apache.org/jira/browse/SOLR-3406?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Martijn van Groningen updated SOLR-3406: Affects Version/s: 4.0 Support grouped range and query facets. --- Key: SOLR-3406 URL: https://issues.apache.org/jira/browse/SOLR-3406 Project: Solr Issue Type: New Feature Affects Versions: 4.0 Reporter: David Assignee: Martijn van Groningen Priority: Critical Fix For: 4.0 Attachments: SOLR-2898-backport.patch, SOLR-3406.patch, SOLR-3406.patch, SOLR-3406.patch Original Estimate: 504h Remaining Estimate: 504h Need the ability to support grouped range and query facets. Grouped facet fields have already been implemented in SOLR-2898 but we still need the ability to compute grouped range and query facets. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3406) Support grouped range and query facets.
[ https://issues.apache.org/jira/browse/SOLR-3406?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Martijn van Groningen updated SOLR-3406: Attachment: SOLR-3406.patch Updated patch to also support facet.range parameter. Support grouped range and query facets. --- Key: SOLR-3406 URL: https://issues.apache.org/jira/browse/SOLR-3406 Project: Solr Issue Type: New Feature Affects Versions: 4.0 Reporter: David Assignee: Martijn van Groningen Priority: Critical Fix For: 4.0 Attachments: SOLR-2898-backport.patch, SOLR-3406.patch, SOLR-3406.patch, SOLR-3406.patch Original Estimate: 504h Remaining Estimate: 504h Need the ability to support grouped range and query facets. Grouped facet fields have already been implemented in SOLR-2898 but we still need the ability to compute grouped range and query facets. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-3406) Support grouped range and query facets.
[ https://issues.apache.org/jira/browse/SOLR-3406?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Martijn van Groningen resolved SOLR-3406. - Resolution: Fixed Fix Version/s: 5.0 Committed to trunk and branch4x. Support grouped range and query facets. --- Key: SOLR-3406 URL: https://issues.apache.org/jira/browse/SOLR-3406 Project: Solr Issue Type: New Feature Affects Versions: 4.0 Reporter: David Assignee: Martijn van Groningen Priority: Critical Fix For: 4.0, 5.0 Attachments: SOLR-2898-backport.patch, SOLR-3406.patch, SOLR-3406.patch, SOLR-3406.patch Original Estimate: 504h Remaining Estimate: 504h Need the ability to support grouped range and query facets. Grouped facet fields have already been implemented in SOLR-2898 but we still need the ability to compute grouped range and query facets. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (LUCENE-4132) IndexWriterConfig live settings
[ https://issues.apache.org/jira/browse/LUCENE-4132?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shai Erera resolved LUCENE-4132. Resolution: Fixed Lucene Fields: New,Patch Available (was: New) Committed revision 1351225 (trunk) and revision 1351229 (4x). Thanks all for your comments ! IndexWriterConfig live settings --- Key: LUCENE-4132 URL: https://issues.apache.org/jira/browse/LUCENE-4132 Project: Lucene - Java Issue Type: Improvement Reporter: Shai Erera Assignee: Shai Erera Priority: Minor Fix For: 4.0, 5.0 Attachments: LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch A while ago there was a discussion about making some IW settings live and I remember that RAM buffer size was one of them. Judging from IW code, I see that RAM buffer can be changed live as IW never caches it. However, I don't remember which other settings were decided to be live and I don't see any documentation in IW nor IWC for that. IW.getConfig mentions: {code} * bNOTE:/b some settings may be changed on the * returned {@link IndexWriterConfig}, and will take * effect in the current IndexWriter instance. See the * javadocs for the specific setters in {@link * IndexWriterConfig} for details. {code} But there's no text on e.g. IWC.setRAMBuffer mentioning that. I think that it'd be good if we make it easier for users to tell which of the settings are live ones. There are few possible ways to do it: * Introduce a custom @live.setting tag on the relevant IWC.set methods, and add special text for them in build.xml ** Or, drop the tag and just document it clearly. * Separate IWC to two interfaces, LiveConfig and OneTimeConfig (name proposals are welcome !), have IWC impl both, and introduce another IW.getLiveConfig which will return that interface, thereby clearly letting the user know which of the settings are live. It'd be good if IWC itself could only expose setXYZ methods for the live settings though. So perhaps, off the top of my head, we can do something like this: * Introduce a Config object, which is essentially what IWC is today, and pass it to IW. * IW will create a different object, IWC from that Config and IW.getConfig will return IWC. * IWC itself will only have setXYZ methods for the live settings. It adds another object, but user code doesn't change - it still creates a Config object when initializing IW, and need to handle a different type if it ever calls IW.getConfig. Maybe that's not such a bad idea? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4148) _TestUtil should be able to generate random longs
[ https://issues.apache.org/jira/browse/LUCENE-4148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395487#comment-13395487 ] Dawid Weiss commented on LUCENE-4148: - I didn't check _TestUtil, to be honest -- I think it did have a bug with range overflow passed to nextInt (which is not a problem in practice, but was my example on how randomized testing can explore unexpected things). Anyway, as for randomLongBetween -- here you'd need to be more careful about double's representation so simple scaling won't work (there are ranges of double that transform to NaNs for example). I am not so strong in numerical analysis to be able to prove something is a correct solution to the problem. My (naive) take at this would be to calculate the range and then if split into a few cases: 1) if it doesn't exceed a positive int, use nextInt() 2) if it doesn't exceed a positive long, use nextLong() 3) if it does exceed positive long, use BigDecimal? :) Maybe this can be done faster without keeping the uniform distribution requirement so strong (i.e. use a modulo from a combination of two concatenated longs as BigDecimal or something like that). _TestUtil should be able to generate random longs - Key: LUCENE-4148 URL: https://issues.apache.org/jira/browse/LUCENE-4148 Project: Lucene - Java Issue Type: Improvement Components: general/test Reporter: Adrien Grand Priority: Trivial It would be helpful in TestPackedInts at least, in order to generate random values (as a workaround, we currently generate a random int between 0 and {{min(Integer.MAX_VALUE, PackedInts.maxValue(bitsPerValue)}}). Moreover, it would help to fix {{nextInt}} for large ranges (calling {{nextInt(random, -10, Integer.MAX_VALUE)}} or even {{nextInt(random, 0, Integer.MAX_VALUE)}} currently fails because the range of values is {{ Integer.MAX_VALUE}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (LUCENE-4149) TestPackedInts test failure
[ https://issues.apache.org/jira/browse/LUCENE-4149?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Adrien Grand resolved LUCENE-4149. -- Resolution: Fixed Fix Version/s: 5.0 4.0 Committed (r1351231 on trunk and r1351239 on brand_4x). TestPackedInts test failure --- Key: LUCENE-4149 URL: https://issues.apache.org/jira/browse/LUCENE-4149 Project: Lucene - Java Issue Type: Bug Affects Versions: 4.0, 5.0 Reporter: Robert Muir Assignee: Adrien Grand Fix For: 4.0, 5.0 Attachments: LUCENE-4149.patch ant test -Dtestcase=TestPackedInts -Dtests.method=testBulkSet -Dtests.seed=ABDA58FE51EFA72B -Dtests.locale=lt -Dtests.timezone=Europe/Athens -Dargs=-Dfile.encoding=UTF-8 -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4069) Segment-level Bloom filters for a 2 x speed up on rare term searches
[ https://issues.apache.org/jira/browse/LUCENE-4069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395773#comment-13395773 ] Mark Harwood commented on LUCENE-4069: -- Interesting results, Mike - thanks for taking the time to run them. bq. BloomFilteredFieldsProducer should just pass through intersect to the delegate? I have tried to make the BloomFilteredFieldsProducer get out of the way of the client app and the delegate PostingsFormat as soon as it is safe to do so i.e. when the user is safely focused on a non-filtered field. While there is a chance the client may end up making a call to TermsEnum.seekExact(..) on a filtered field then I need to have a wrapper object in place which is in a position to intercept this call. In all other method invocations I just end up delegating calls so I wonder if all these extra method calls are the cause of the slowdown you see e.g. when Fuzzy is enumerating over many terms. The only other alternatives to endlessly wrapping in this way are: a) API change - e.g. allow TermsEnum.seekExact to have a pluggable call-out for just this one method. b) Mess around with byte-code manipulation techniques to weave in Bloom filtering(the sort of thing I recall Hibernate resorts to) Neither of these seem particularly appealing options so I think we may have to live with fuzzy+bloom not being as fast as straight fuzzy. For completeness sake - I don't have access to your benchmarking code but I would hope that PostingsFormat.fieldsProducer() isn't called more than once for the same segment as that's where the Bloom filters get loaded from disk so there's inherent cost there too. I can't imagine this is the case. BTW I've just finished a long-running set of tests which mixes up reads and writes here: http://goo.gl/KJmGv This benchmark represents how graph databases such as Neo4j use Lucene for an index when loading (I typically use the Wikipedia links as a test set). I look to get a 3.5 x speed up in Lucene 4 and Lucene 3.6 gets nearly 9 x speedup over the comparatively slower 3.6 codebase. Segment-level Bloom filters for a 2 x speed up on rare term searches Key: LUCENE-4069 URL: https://issues.apache.org/jira/browse/LUCENE-4069 Project: Lucene - Java Issue Type: Improvement Components: core/index Affects Versions: 3.6, 4.0 Reporter: Mark Harwood Priority: Minor Fix For: 4.0, 3.6.1 Attachments: BloomFilterPostingsBranch4x.patch, MHBloomFilterOn3.6Branch.patch, PrimaryKeyPerfTest40.java An addition to each segment which stores a Bloom filter for selected fields in order to give fast-fail to term searches, helping avoid wasted disk access. Best suited for low-frequency fields e.g. primary keys on big indexes with many segments but also speeds up general searching in my tests. Overview slideshow here: http://www.slideshare.net/MarkHarwood/lucene-bloomfilteredsegments Benchmarks based on Wikipedia content here: http://goo.gl/X7QqU Patch based on 3.6 codebase attached. There are no 3.6 API changes currently - to play just add a field with _blm on the end of the name to invoke special indexing/querying capability. Clearly a new Field or schema declaration(!) would need adding to APIs to configure the service properly. Also, a patch for Lucene4.0 codebase introducing a new PostingsFormat -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-2899) Add OpenNLP Analysis capabilities as a module
[ https://issues.apache.org/jira/browse/LUCENE-2899?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395787#comment-13395787 ] Joern Kottmann commented on LUCENE-2899: For NER you should try the perceptron and a cutoff of zero. For NER with a cutoff of 5 you need otherwise much more training data. Add OpenNLP Analysis capabilities as a module - Key: LUCENE-2899 URL: https://issues.apache.org/jira/browse/LUCENE-2899 Project: Lucene - Java Issue Type: New Feature Components: modules/analysis Reporter: Grant Ingersoll Assignee: Grant Ingersoll Priority: Minor Attachments: LUCENE-2899.patch, opennlp_trunk.patch Now that OpenNLP is an ASF project and has a nice license, it would be nice to have a submodule (under analysis) that exposed capabilities for it. Drew Farris, Tom Morton and I have code that does: * Sentence Detection as a Tokenizer (could also be a TokenFilter, although it would have to change slightly to buffer tokens) * NamedEntity recognition as a TokenFilter We are also planning a Tokenizer/TokenFilter that can put parts of speech as either payloads (PartOfSpeechAttribute?) on a token or at the same position. I'd propose it go under: modules/analysis/opennlp -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-3866) Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader
[ https://issues.apache.org/jira/browse/LUCENE-3866?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Uwe Schindler updated LUCENE-3866: -- Attachment: LUCENE-3866-step2.patch This is the final patch, now also moving IndexReaderContext to collections and make them unmodifiable. For safety, IRC.leaves() now throws UnsupportedOperationException, if the context is not top-level (returned null before), this helps to find bugs. All tests pass, JavaDocs are updated. Robert suggested to me yesterday, to maybe make getSequentialSubReaders() protected, as it is no longer used by user code (only tests). All code should use the topReaderContext returned by the IndexReader to get all *atomic* IRC.leaves(), or IRC.children() to get the sub-contexts/readers. For easy use it might be a good idea to add some easy-use methods in ReaderUtil to get a view on all AtomicReader leaves (without docBase, so returns ListAtomicReader). Some code not needing the whole info would get simplier. This is stuff for a new issue. In my opinion we should also move and hide ReaderSlice and BitSlice to index package, those classes are solely privately used from there. I think this is ready to commit to trunk and for backport to 4.x. I will not wait too long (max 24hrs), as the patch may get outdated very quickly. Maybe Mike can do some perf benchmarks with beast to show that it does not affect performance (some parts like creating Multi* should be much more effective now). Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader --- Key: LUCENE-3866 URL: https://issues.apache.org/jira/browse/LUCENE-3866 Project: Lucene - Java Issue Type: Improvement Reporter: Uwe Schindler Assignee: Uwe Schindler Fix For: 4.0 Attachments: LUCENE-3866-step1.patch, LUCENE-3866-step2.patch Since Lucene 2.9 we have the method getSequentialSubReader() returning IndexReader[]. Based on hardly-to-debug errors in user's code, Robert and me realized that returning an array from a public API is an anti-pattern. If the array is intended to be modifiable (like byte[] in terms,...), it is fine to use arrays in public APIs, but not, if the array must be protected from modification. As IndexReaders are 100% unmodifiable in trunk code (no deletions,...), the only possibility to corrumpt the reader is by modifying the array returned by getSequentialSubReaders(). We should prevent this. The same theoretically applies to FieldCache, too - but the party that is afraid of performance problems is too big to fight against that :( For getSequentialSubReaders there is no performance problem at all. The binary search of reader-ids inside BaseCompositeReader can still be done with the internal protected array, but public APIs should expose only a unmodifiable List. The same applies to leaves() and children() in IndexReaderContext. This change to list would also allow to make CompositeReader and CompositeReaderContext IterableR extends IndexReader, so some loops would look nice. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-3866) Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader
[ https://issues.apache.org/jira/browse/LUCENE-3866?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Uwe Schindler updated LUCENE-3866: -- Attachment: LUCENE-3866-step2.patch Samll update with JavaDocs (remove the don't modify returned array @ getSequentialSubReaders) Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader --- Key: LUCENE-3866 URL: https://issues.apache.org/jira/browse/LUCENE-3866 Project: Lucene - Java Issue Type: Improvement Reporter: Uwe Schindler Assignee: Uwe Schindler Fix For: 4.0 Attachments: LUCENE-3866-step1.patch, LUCENE-3866-step2.patch, LUCENE-3866-step2.patch Since Lucene 2.9 we have the method getSequentialSubReader() returning IndexReader[]. Based on hardly-to-debug errors in user's code, Robert and me realized that returning an array from a public API is an anti-pattern. If the array is intended to be modifiable (like byte[] in terms,...), it is fine to use arrays in public APIs, but not, if the array must be protected from modification. As IndexReaders are 100% unmodifiable in trunk code (no deletions,...), the only possibility to corrumpt the reader is by modifying the array returned by getSequentialSubReaders(). We should prevent this. The same theoretically applies to FieldCache, too - but the party that is afraid of performance problems is too big to fight against that :( For getSequentialSubReaders there is no performance problem at all. The binary search of reader-ids inside BaseCompositeReader can still be done with the internal protected array, but public APIs should expose only a unmodifiable List. The same applies to leaves() and children() in IndexReaderContext. This change to list would also allow to make CompositeReader and CompositeReaderContext IterableR extends IndexReader, so some loops would look nice. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Build failed in Jenkins: Lucene-Core-4x-Beasting #2798
See http://sierranevada.servebeer.com/job/Lucene-Core-4x-Beasting/2798/ -- [...truncated 840 lines...] [junit4] Completed on J2 in 0.02s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestRollingUpdates [junit4] Completed on J2 in 1.24s, 2 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestIndexFileDeleter [junit4] Completed on J2 in 0.08s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestIndexWriter [junit4] Completed on J3 in 3.13s, 47 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestFlushByRamOrCountsPolicy [junit4] Completed on J1 in 1.93s, 5 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestSegmentTermDocs [junit4] Completed on J1 in 0.42s, 5 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestBackwardsCompatibility [junit4] Completed on J2 in 2.42s, 13 tests [junit4] [junit4] Suite: org.apache.lucene.util.packed.TestPackedInts [junit4] Completed on J0 in 7.49s, 15 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestNumericRangeQuery32 [junit4] Completed on J3 in 3.69s, 27 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestDuelingCodecs [junit4] Completed on J1 in 3.83s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestDeletionPolicy [junit4] Completed on J0 in 3.52s, 6 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestCompoundFile [junit4] Completed on J0 in 0.93s, 18 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestTimeLimitingCollector [junit4] Completed on J1 in 1.55s, 6 tests [junit4] [junit4] Suite: org.apache.lucene.store.TestBufferedIndexInput [junit4] Completed on J0 in 1.45s, 4 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestNRTReaderWithThreads [junit4] Completed on J1 in 1.44s, 1 test [junit4] [junit4] Suite: org.apache.lucene.store.TestNRTCachingDirectory [junit4] Completed on J0 in 0.48s, 5 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestNorms [junit4] Completed on J2 in 7.87s, 3 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestForceMergeForever [junit4] Completed on J2 in 0.40s, 1 test [junit4] [junit4] Suite: org.apache.lucene.util.automaton.TestMinimize [junit4] Completed on J0 in 1.36s, 3 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestNeverDelete [junit4] Completed on J1 in 1.90s, 1 test [junit4] [junit4] Suite: org.apache.lucene.search.similarities.TestSimilarityBase [junit4] IGNOR/A 0.00s J0 | TestSimilarityBase.testHeartRanking [junit4] Assumption #1: PreFlex codec does not support the stats necessary for this test! [junit4] Completed on J0 in 0.25s, 29 tests, 1 skipped [junit4] [junit4] Suite: org.apache.lucene.search.spans.TestSpans [junit4] Completed on J2 in 0.63s, 25 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestParallelCompositeReader [junit4] Completed on J0 in 0.12s, 8 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestDocsAndPositions [junit4] Completed on J1 in 0.38s, 6 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestMultiThreadTermVectors [junit4] Completed on J0 in 0.20s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestOmitTf [junit4] Completed on J1 in 0.15s, 6 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestNRTThreads [junit4] Completed on J3 in 7.97s, 1 test [junit4] [junit4] Suite: org.apache.lucene.index.TestRandomStoredFields [junit4] Completed on J0 in 0.30s, 1 test [junit4] [junit4] Suite: org.apache.lucene.search.TestDisjunctionMaxQuery [junit4] Completed on J2 in 0.77s, 10 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestQueryWrapperFilter [junit4] Completed on J3 in 0.12s, 3 tests [junit4] [junit4] Suite: org.apache.lucene.search.payloads.TestPayloadNearQuery [junit4] Completed on J2 in 0.09s, 7 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestPrefixRandom [junit4] Completed on J1 in 0.53s, 1 test [junit4] [junit4] Suite: org.apache.lucene.search.TestBooleanOr [junit4] Completed on J3 in 0.05s, 5 tests [junit4] [junit4] Suite: org.apache.lucene.util.automaton.TestLevenshteinAutomata [junit4] Completed on J0 in 0.41s, 4 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestCustomNorms [junit4] Completed on J2 in 0.21s, 3 tests [junit4] [junit4] Suite: org.apache.lucene.search.TestFieldCacheRangeFilter [junit4] Completed on J3 in 0.12s, 9 tests [junit4] [junit4] Suite: org.apache.lucene.index.TestPrefixCodedTerms [junit4] Completed on J2 in 0.08s, 6 tests [junit4] [junit4] Suite: org.apache.lucene.codecs.lucene3x.TestTermInfosReaderIndex [junit4] Completed on J3 in 0.14s, 3 tests
Jenkins build is back to normal : Lucene-Core-4x-Beasting #2799
See http://sierranevada.servebeer.com/job/Lucene-Core-4x-Beasting/2799/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-4147) rollback/preparecommit thread hazard
[ https://issues.apache.org/jira/browse/LUCENE-4147?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Michael McCandless updated LUCENE-4147: --- Attachment: LUCENE-4147.patch Patch, acquiring the commitLock around close and rollback, and adding the ensureOpen before prepareCommit. However, the test still fails after a few hundred beasting iterations, I think because of thread safety issues where one thread calls docWriter.abort while another is flushing ... not sure how to fix that one yet. Simon maybe you can have a look? rollback/preparecommit thread hazard Key: LUCENE-4147 URL: https://issues.apache.org/jira/browse/LUCENE-4147 Project: Lucene - Java Issue Type: Bug Affects Versions: 4.0 Reporter: Robert Muir Attachments: LUCENE-4147.patch, LUCENE-4147.patch found by http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows-Java7-64/70/ rollback should never throw this exception, as it documents it clears any pendingcommits. but it calls closeInternal outside of any sync block, so it looks like there is a race here. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-4152) add one-syllable method to IndexReader enumerate subreaders
Robert Muir created LUCENE-4152: --- Summary: add one-syllable method to IndexReader enumerate subreaders Key: LUCENE-4152 URL: https://issues.apache.org/jira/browse/LUCENE-4152 Project: Lucene - Java Issue Type: Bug Reporter: Robert Muir Priority: Blocker Fix For: 4.0 Description is exactly as written. getSequentialSubReaders/getTopLevelReaderContext, these method names are way too long/unuseable. They also have tricky semantics (e.g. returning null). In lucene 4, people cannot just use any indexreader and get a merged view. So we need to make this stuff easy on them: * single-syllable method name (leaves(), subs(), i will think on this) * supports enhanced for-loop (no returning null or anything like that) * on indexreader (not atomic or composite, plain old indexreader) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4147) rollback/preparecommit thread hazard
[ https://issues.apache.org/jira/browse/LUCENE-4147?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395826#comment-13395826 ] Simon Willnauer commented on LUCENE-4147: - mike, do we really need to acquire the commit lock? from my perspective it would be enough to add an ensure open when we assigne pendingCommit (inside the sync block) so that racing threads hit already close exceptions. regarding the thread safety issue in DocWriter can you paste the trace? rollback/preparecommit thread hazard Key: LUCENE-4147 URL: https://issues.apache.org/jira/browse/LUCENE-4147 Project: Lucene - Java Issue Type: Bug Affects Versions: 4.0 Reporter: Robert Muir Attachments: LUCENE-4147.patch, LUCENE-4147.patch found by http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows-Java7-64/70/ rollback should never throw this exception, as it documents it clears any pendingcommits. but it calls closeInternal outside of any sync block, so it looks like there is a race here. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4069) Segment-level Bloom filters for a 2 x speed up on rare term searches
[ https://issues.apache.org/jira/browse/LUCENE-4069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395830#comment-13395830 ] Michael McCandless commented on LUCENE-4069: bq. Interesting results, Mike - thanks for taking the time to run them. You're welcome! {quote} bq. BloomFilteredFieldsProducer should just pass through intersect to the delegate? I have tried to make the BloomFilteredFieldsProducer get out of the way of the client app and the delegate PostingsFormat as soon as it is safe to do so i.e. when the user is safely focused on a non-filtered field. While there is a chance the client may end up making a call to TermsEnum.seekExact(..) on a filtered field then I need to have a wrapper object in place which is in a position to intercept this call. In all other method invocations I just end up delegating calls so I wonder if all these extra method calls are the cause of the slowdown you see e.g. when Fuzzy is enumerating over many terms. The only other alternatives to endlessly wrapping in this way are: a) API change - e.g. allow TermsEnum.seekExact to have a pluggable call-out for just this one method. b) Mess around with byte-code manipulation techniques to weave in Bloom filtering(the sort of thing I recall Hibernate resorts to) Neither of these seem particularly appealing options so I think we may have to live with fuzzy+bloom not being as fast as straight fuzzy. {quote} I think the fix is simple: you are not overriding Terms.intersect now, in BloomFilteredTerms. I think you should override it and immediately delegate and then FuzzyN/Respell performance should be just as good as Lucene40 codec. bq. For completeness sake - I don't have access to your benchmarking code All the benchmarking code is here: http://code.google.com/a/apache-extras.org/p/luceneutil/ I run it nightly (trunk) and publish the results here: http://people.apache.org/~mikemccand/lucenebench/ bq. but I would hope that PostingsFormat.fieldsProducer() isn't called more than once for the same segment as that's where the Bloom filters get loaded from disk so there's inherent cost there too. I can't imagine this is the case. It's only called once on init'ing the SegmentReader (or at least it better be!). {quote} BTW I've just finished a long-running set of tests which mixes up reads and writes here: http://goo.gl/KJmGv This benchmark represents how graph databases such as Neo4j use Lucene for an index when loading (I typically use the Wikipedia links as a test set). I look to get a 3.5 x speed up in Lucene 4 and Lucene 3.6 gets nearly 9 x speedup over the comparatively slower 3.6 codebase. {quote} Nice results! It looks like bloom(3.6) is faster than bloom(4.0)? Why is that... Also I wonder why you see such sizable (3.5X speedup) gains on PK lookup but in my benchmark I see only ~13% - 24%. My index has 5 segments per level... Segment-level Bloom filters for a 2 x speed up on rare term searches Key: LUCENE-4069 URL: https://issues.apache.org/jira/browse/LUCENE-4069 Project: Lucene - Java Issue Type: Improvement Components: core/index Affects Versions: 3.6, 4.0 Reporter: Mark Harwood Priority: Minor Fix For: 4.0, 3.6.1 Attachments: BloomFilterPostingsBranch4x.patch, MHBloomFilterOn3.6Branch.patch, PrimaryKeyPerfTest40.java An addition to each segment which stores a Bloom filter for selected fields in order to give fast-fail to term searches, helping avoid wasted disk access. Best suited for low-frequency fields e.g. primary keys on big indexes with many segments but also speeds up general searching in my tests. Overview slideshow here: http://www.slideshare.net/MarkHarwood/lucene-bloomfilteredsegments Benchmarks based on Wikipedia content here: http://goo.gl/X7QqU Patch based on 3.6 codebase attached. There are no 3.6 API changes currently - to play just add a field with _blm on the end of the name to invoke special indexing/querying capability. Clearly a new Field or schema declaration(!) would need adding to APIs to configure the service properly. Also, a patch for Lucene4.0 codebase introducing a new PostingsFormat -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4152) add one-syllable method to IndexReader enumerate subreaders
[ https://issues.apache.org/jira/browse/LUCENE-4152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395840#comment-13395840 ] Uwe Schindler commented on LUCENE-4152: --- bq. They also have tricky semantics (e.g. returning null). No longer since January's refactoring, leaves() returns always something != null. And getSeqSubReaders is only available on CompositeReaders, which always have subReaders. See the other issue LUCENE-3866 where I did some more refactoring to make all this easier. In general I agree with that, as leaves() and children()/subs() would then just be a shortcut to getTopReaderContext()'s methods (leaves(), children() - both are Iterable, no need to change anything) (I would rename that one to as[Top]Context()). Returning plain subReaders without Contexts is not really useful, as all of Lucene's Query logic uses AtomicReaderContext, so leaves() on IndexReader returning the same as getTopReaderContext().leaves() is the way to go. This method can be added as final easy-use method. I hope you look at my other patch @ LUCENE-3866, because it shows how simple the code is now without ReaderUtil.Gather. add one-syllable method to IndexReader enumerate subreaders --- Key: LUCENE-4152 URL: https://issues.apache.org/jira/browse/LUCENE-4152 Project: Lucene - Java Issue Type: Bug Reporter: Robert Muir Priority: Blocker Fix For: 4.0 Description is exactly as written. getSequentialSubReaders/getTopLevelReaderContext, these method names are way too long/unuseable. They also have tricky semantics (e.g. returning null). In lucene 4, people cannot just use any indexreader and get a merged view. So we need to make this stuff easy on them: * single-syllable method name (leaves(), subs(), i will think on this) * supports enhanced for-loop (no returning null or anything like that) * on indexreader (not atomic or composite, plain old indexreader) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-tests-only-4.x-java7 - Build # 15 - Failure
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-4.x-java7/15/ No tests ran. Build Log: [...truncated 451 lines...] ivy-availability-check: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/ivy-settings.xml resolve: [echo] Building solr-uima... ivy-availability-check: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/ivy-settings.xml resolve: [echo] Building solr-velocity... ivy-availability-check: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/ivy-settings.xml resolve: validate: [echo] License check under: /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/solr [licenses] Scanned 82 JAR file(s) for licenses (in 0.66s.), 0 error(s). test: test-core: jflex-uptodate-check: jflex-notice: javacc-uptodate-check: javacc-notice: ivy-availability-check: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/ivy-settings.xml resolve: init: clover.setup: clover.info: [echo] [echo] Clover not found. Code coverage reports disabled. [echo] clover: common.compile-core: [mkdir] Created dir: /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/build/core/classes/java [javac] Compiling 649 source files to /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/build/core/classes/java [javac] warning: [options] bootstrap class path not set in conjunction with -source 1.6 [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:506: error: class, interface, or enum expected [javac] package org.apache.lucene.index; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:525: error: class, interface, or enum expected [javac] import org.apache.lucene.analysis.Analyzer; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:526: error: class, interface, or enum expected [javac] import org.apache.lucene.codecs.Codec; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:527: error: class, interface, or enum expected [javac] import org.apache.lucene.index.DocumentsWriterPerThread.IndexingChain; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:528: error: class, interface, or enum expected [javac] import org.apache.lucene.index.IndexWriter.IndexReaderWarmer; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:529: error: class, interface, or enum expected [javac] import org.apache.lucene.index.IndexWriterConfig.OpenMode; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:530: error: class, interface, or enum expected [javac] import org.apache.lucene.search.IndexSearcher; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:531: error: class, interface, or enum expected [javac] import org.apache.lucene.search.similarities.Similarity; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:532: error: class, interface, or enum expected [javac] import org.apache.lucene.util.InfoStream; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x-java7/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:533: error: class, interface, or enum expected [javac] import org.apache.lucene.util.Version; [javac] ^ [javac] 10 errors [...truncated 14 lines...] [...truncated 566 lines...] [...truncated 566 lines...] [...truncated 566 lines...] [...truncated 566 lines...]
[jira] [Comment Edited] (LUCENE-4152) add one-syllable method to IndexReader enumerate subreaders
[ https://issues.apache.org/jira/browse/LUCENE-4152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395840#comment-13395840 ] Uwe Schindler edited comment on LUCENE-4152 at 6/18/12 12:02 PM: - bq. They also have tricky semantics (e.g. returning null). No longer since January's refactoring, leaves() returns always something != null. And getSeqSubReaders is only available on CompositeReaders, which always have subReaders. See the other issue LUCENE-3866 where I did some more refactoring to make all this easier. In general I agree with that, as leaves() and children()/subs() would then just be a shortcut to getTopReaderContext()'s methods (leaves(), children() - both are Iterable, no need to change anything) (I would rename that one to as[Top]Context()). Returning plain subReaders without Contexts is not really useful, as all of Lucene's Query logic uses AtomicReaderContext, so leaves() on IndexReader returning the same as getTopReaderContext().leaves() is the way to go. This method can be added as final easy-use method. I hope you look at my other patch @ LUCENE-3866, because it shows how simple the code is now without ReaderUtil.Gather. Additionally, as noted in LUCENE-3866, getSequentialSubReaders() in CompositeReader should be made protected, user code does not need to use it. It's solely there for building the reader hierarchy, later available using IRC.leaves() and IRC.children(). Currently this method is only used by tests anymore (which can be rewritten). In CompositeReader, getSequentialSubReaders() should just be the protected abstract way for the API on top to get the inner structure, but not for the outside. was (Author: thetaphi): bq. They also have tricky semantics (e.g. returning null). No longer since January's refactoring, leaves() returns always something != null. And getSeqSubReaders is only available on CompositeReaders, which always have subReaders. See the other issue LUCENE-3866 where I did some more refactoring to make all this easier. In general I agree with that, as leaves() and children()/subs() would then just be a shortcut to getTopReaderContext()'s methods (leaves(), children() - both are Iterable, no need to change anything) (I would rename that one to as[Top]Context()). Returning plain subReaders without Contexts is not really useful, as all of Lucene's Query logic uses AtomicReaderContext, so leaves() on IndexReader returning the same as getTopReaderContext().leaves() is the way to go. This method can be added as final easy-use method. I hope you look at my other patch @ LUCENE-3866, because it shows how simple the code is now without ReaderUtil.Gather. add one-syllable method to IndexReader enumerate subreaders --- Key: LUCENE-4152 URL: https://issues.apache.org/jira/browse/LUCENE-4152 Project: Lucene - Java Issue Type: Bug Reporter: Robert Muir Priority: Blocker Fix For: 4.0 Description is exactly as written. getSequentialSubReaders/getTopLevelReaderContext, these method names are way too long/unuseable. They also have tricky semantics (e.g. returning null). In lucene 4, people cannot just use any indexreader and get a merged view. So we need to make this stuff easy on them: * single-syllable method name (leaves(), subs(), i will think on this) * supports enhanced for-loop (no returning null or anything like that) * on indexreader (not atomic or composite, plain old indexreader) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: svn commit: r1351229 - in /lucene/dev/branches/branch_4x: ./ lucene/ lucene/core/ lucene/core/src/java/org/apache/lucene/index/ lucene/core/src/test/org/apache/lucene/index/
Why are the CHANGES totally screwy in this commit? Please put them under the right section (also fix trunk too) On Mon, Jun 18, 2012 at 4:24 AM, sh...@apache.org wrote: Author: shaie Date: Mon Jun 18 08:24:25 2012 New Revision: 1351229 URL: http://svn.apache.org/viewvc?rev=1351229view=rev Log: LUCENE-4132: merge from trunk Added:   lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java    - copied unchanged from r1351225, lucene/dev/trunk/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java Modified:   lucene/dev/branches/branch_4x/  (props changed)   lucene/dev/branches/branch_4x/lucene/  (props changed)   lucene/dev/branches/branch_4x/lucene/CHANGES.txt   lucene/dev/branches/branch_4x/lucene/core/  (props changed)   lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriter.java   lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriterFlushControl.java   lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriterPerThreadPool.java   lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/FlushPolicy.java   lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/IndexWriter.java   lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/IndexWriterConfig.java   lucene/dev/branches/branch_4x/lucene/core/src/test/org/apache/lucene/index/TestFlushByRamOrCountsPolicy.java   lucene/dev/branches/branch_4x/lucene/core/src/test/org/apache/lucene/index/TestIndexWriterConfig.java Modified: lucene/dev/branches/branch_4x/lucene/CHANGES.txt URL: http://svn.apache.org/viewvc/lucene/dev/branches/branch_4x/lucene/CHANGES.txt?rev=1351229r1=1351228r2=1351229view=diff == --- lucene/dev/branches/branch_4x/lucene/CHANGES.txt (original) +++ lucene/dev/branches/branch_4x/lucene/CHANGES.txt Mon Jun 18 08:24:25 2012 @@ -6,6 +6,17 @@ http://s.apache.org/luceneversions  For contrib changes prior to 4.0, please see:  http://svn.apache.org/repos/asf/lucene/dev/tags/lucene_solr_3_6_0/lucene/contrib/CHANGES.txt +New features + +* LUCENE-4108: add replaceTaxonomy to DirectoryTaxonomyWriter, which replaces +  the taxonomy in place with the given one. (Shai Erera) + +API Changes + +* LUCENE-4132: IndexWriter.getConfig() now returns a LiveIndexWriterConfig object +  which can be used to change the IndexWriter's live settings. IndexWriterConfig +  is used only for initializing the IndexWriter. (Shai Erera) +  === Lucene 4.0.0-ALPHA ===  More information about this release, including any errata related to the  release notes, upgrade instructions, or other changes may be found online at: Modified: lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriter.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriter.java?rev=1351229r1=1351228r2=1351229view=diff == --- lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriter.java (original) +++ lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriter.java Mon Jun 18 08:24:25 2012 @@ -133,7 +133,7 @@ final class DocumentsWriter {  final DocumentsWriterFlushControl flushControl;  final Codec codec; -  DocumentsWriter(Codec codec, IndexWriterConfig config, Directory directory, IndexWriter writer, FieldNumbers globalFieldNumbers, +  DocumentsWriter(Codec codec, LiveIndexWriterConfig config, Directory directory, IndexWriter writer, FieldNumbers globalFieldNumbers,    BufferedDeletesStream bufferedDeletesStream) throws IOException {   this.codec = codec;   this.directory = directory; Modified: lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriterFlushControl.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriterFlushControl.java?rev=1351229r1=1351228r2=1351229view=diff == --- lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriterFlushControl.java (original) +++ lucene/dev/branches/branch_4x/lucene/core/src/java/org/apache/lucene/index/DocumentsWriterFlushControl.java Mon Jun 18 08:24:25 2012 @@ -65,10 +65,9 @@ final class DocumentsWriterFlushControl  private final FlushPolicy flushPolicy;  private boolean closed = false;  private final DocumentsWriter documentsWriter; -  private final IndexWriterConfig config; +  private final LiveIndexWriterConfig
[jira] [Reopened] (LUCENE-4108) Add replaceTaxonomy to DirectoryTaxonomyWriter
[ https://issues.apache.org/jira/browse/LUCENE-4108?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Robert Muir reopened LUCENE-4108: - CHANGES.txt is mangled Add replaceTaxonomy to DirectoryTaxonomyWriter -- Key: LUCENE-4108 URL: https://issues.apache.org/jira/browse/LUCENE-4108 Project: Lucene - Java Issue Type: New Feature Components: modules/facet Reporter: Shai Erera Assignee: Shai Erera Priority: Minor Fix For: 4.0, 5.0 Attachments: LUCENE-4108.patch I'd like to add the option to replace a taxonomy with another one, live on DirectoryTaxoWriter. Similar to how one can do so on IndexWriter, by calling deleteAll() and addIndexes(), it is valuable to be able to do so on DirTW as well. That way, if you need to replace a search index and taxonomy index live, you will be able to, given this new API. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Reopened] (LUCENE-4132) IndexWriterConfig live settings
[ https://issues.apache.org/jira/browse/LUCENE-4132?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Robert Muir reopened LUCENE-4132: - CHANGES.txt is mangled IndexWriterConfig live settings --- Key: LUCENE-4132 URL: https://issues.apache.org/jira/browse/LUCENE-4132 Project: Lucene - Java Issue Type: Improvement Reporter: Shai Erera Assignee: Shai Erera Priority: Minor Fix For: 4.0, 5.0 Attachments: LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch A while ago there was a discussion about making some IW settings live and I remember that RAM buffer size was one of them. Judging from IW code, I see that RAM buffer can be changed live as IW never caches it. However, I don't remember which other settings were decided to be live and I don't see any documentation in IW nor IWC for that. IW.getConfig mentions: {code} * bNOTE:/b some settings may be changed on the * returned {@link IndexWriterConfig}, and will take * effect in the current IndexWriter instance. See the * javadocs for the specific setters in {@link * IndexWriterConfig} for details. {code} But there's no text on e.g. IWC.setRAMBuffer mentioning that. I think that it'd be good if we make it easier for users to tell which of the settings are live ones. There are few possible ways to do it: * Introduce a custom @live.setting tag on the relevant IWC.set methods, and add special text for them in build.xml ** Or, drop the tag and just document it clearly. * Separate IWC to two interfaces, LiveConfig and OneTimeConfig (name proposals are welcome !), have IWC impl both, and introduce another IW.getLiveConfig which will return that interface, thereby clearly letting the user know which of the settings are live. It'd be good if IWC itself could only expose setXYZ methods for the live settings though. So perhaps, off the top of my head, we can do something like this: * Introduce a Config object, which is essentially what IWC is today, and pass it to IW. * IW will create a different object, IWC from that Config and IW.getConfig will return IWC. * IWC itself will only have setXYZ methods for the live settings. It adds another object, but user code doesn't change - it still creates a Config object when initializing IW, and need to handle a different type if it ever calls IW.getConfig. Maybe that's not such a bad idea? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4132) IndexWriterConfig live settings
[ https://issues.apache.org/jira/browse/LUCENE-4132?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395846#comment-13395846 ] Uwe Schindler commented on LUCENE-4132: --- The 4.x commit also breaks the builds, looks like duplicate merges to same file. IndexWriterConfig live settings --- Key: LUCENE-4132 URL: https://issues.apache.org/jira/browse/LUCENE-4132 Project: Lucene - Java Issue Type: Improvement Reporter: Shai Erera Assignee: Shai Erera Priority: Minor Fix For: 4.0, 5.0 Attachments: LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch A while ago there was a discussion about making some IW settings live and I remember that RAM buffer size was one of them. Judging from IW code, I see that RAM buffer can be changed live as IW never caches it. However, I don't remember which other settings were decided to be live and I don't see any documentation in IW nor IWC for that. IW.getConfig mentions: {code} * bNOTE:/b some settings may be changed on the * returned {@link IndexWriterConfig}, and will take * effect in the current IndexWriter instance. See the * javadocs for the specific setters in {@link * IndexWriterConfig} for details. {code} But there's no text on e.g. IWC.setRAMBuffer mentioning that. I think that it'd be good if we make it easier for users to tell which of the settings are live ones. There are few possible ways to do it: * Introduce a custom @live.setting tag on the relevant IWC.set methods, and add special text for them in build.xml ** Or, drop the tag and just document it clearly. * Separate IWC to two interfaces, LiveConfig and OneTimeConfig (name proposals are welcome !), have IWC impl both, and introduce another IW.getLiveConfig which will return that interface, thereby clearly letting the user know which of the settings are live. It'd be good if IWC itself could only expose setXYZ methods for the live settings though. So perhaps, off the top of my head, we can do something like this: * Introduce a Config object, which is essentially what IWC is today, and pass it to IW. * IW will create a different object, IWC from that Config and IW.getConfig will return IWC. * IWC itself will only have setXYZ methods for the live settings. It adds another object, but user code doesn't change - it still creates a Config object when initializing IW, and need to handle a different type if it ever calls IW.getConfig. Maybe that's not such a bad idea? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3866) Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader
[ https://issues.apache.org/jira/browse/LUCENE-3866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395847#comment-13395847 ] Uwe Schindler commented on LUCENE-3866: --- I will later also upgrade MIGRATE.txt markdown. Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader --- Key: LUCENE-3866 URL: https://issues.apache.org/jira/browse/LUCENE-3866 Project: Lucene - Java Issue Type: Improvement Reporter: Uwe Schindler Assignee: Uwe Schindler Fix For: 4.0 Attachments: LUCENE-3866-step1.patch, LUCENE-3866-step2.patch, LUCENE-3866-step2.patch Since Lucene 2.9 we have the method getSequentialSubReader() returning IndexReader[]. Based on hardly-to-debug errors in user's code, Robert and me realized that returning an array from a public API is an anti-pattern. If the array is intended to be modifiable (like byte[] in terms,...), it is fine to use arrays in public APIs, but not, if the array must be protected from modification. As IndexReaders are 100% unmodifiable in trunk code (no deletions,...), the only possibility to corrumpt the reader is by modifying the array returned by getSequentialSubReaders(). We should prevent this. The same theoretically applies to FieldCache, too - but the party that is afraid of performance problems is too big to fight against that :( For getSequentialSubReaders there is no performance problem at all. The binary search of reader-ids inside BaseCompositeReader can still be done with the internal protected array, but public APIs should expose only a unmodifiable List. The same applies to leaves() and children() in IndexReaderContext. This change to list would also allow to make CompositeReader and CompositeReaderContext IterableR extends IndexReader, so some loops would look nice. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4147) rollback/preparecommit thread hazard
[ https://issues.apache.org/jira/browse/LUCENE-4147?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395854#comment-13395854 ] Michael McCandless commented on LUCENE-4147: bq. mike, do we really need to acquire the commit lock? The problem is rollback forcefully clears the pendingCommit and then deletes any files it had (alone) referenced, so if a commit is running concurrently the fsyncs will fail since the files were deleted. Also: it doesn't really make sense to allow rollback and commit to proceed concurrently? Why would an app need this? Seems like we can simplify the code by making them exclusive. bq. regarding the thread safety issue in DocWriter can you paste the trace? Will do ... need to re-beast. rollback/preparecommit thread hazard Key: LUCENE-4147 URL: https://issues.apache.org/jira/browse/LUCENE-4147 Project: Lucene - Java Issue Type: Bug Affects Versions: 4.0 Reporter: Robert Muir Attachments: LUCENE-4147.patch, LUCENE-4147.patch found by http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows-Java7-64/70/ rollback should never throw this exception, as it documents it clears any pendingcommits. but it calls closeInternal outside of any sync block, so it looks like there is a race here. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4069) Segment-level Bloom filters for a 2 x speed up on rare term searches
[ https://issues.apache.org/jira/browse/LUCENE-4069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395857#comment-13395857 ] Mark Harwood commented on LUCENE-4069: -- bq. I think the fix is simple: you are not overriding Terms.intersect now, in BloomFilteredTerms Good catch - a quick test indeed shows a speed up on fuzzy queries. I'll prepare a new patch. I'm not sure on why 3.6+Bloom is faster than 4+Bloom in my tests. I'll take a closer look at your benchmark. Segment-level Bloom filters for a 2 x speed up on rare term searches Key: LUCENE-4069 URL: https://issues.apache.org/jira/browse/LUCENE-4069 Project: Lucene - Java Issue Type: Improvement Components: core/index Affects Versions: 3.6, 4.0 Reporter: Mark Harwood Priority: Minor Fix For: 4.0, 3.6.1 Attachments: BloomFilterPostingsBranch4x.patch, MHBloomFilterOn3.6Branch.patch, PrimaryKeyPerfTest40.java An addition to each segment which stores a Bloom filter for selected fields in order to give fast-fail to term searches, helping avoid wasted disk access. Best suited for low-frequency fields e.g. primary keys on big indexes with many segments but also speeds up general searching in my tests. Overview slideshow here: http://www.slideshare.net/MarkHarwood/lucene-bloomfilteredsegments Benchmarks based on Wikipedia content here: http://goo.gl/X7QqU Patch based on 3.6 codebase attached. There are no 3.6 API changes currently - to play just add a field with _blm on the end of the name to invoke special indexing/querying capability. Clearly a new Field or schema declaration(!) would need adding to APIs to configure the service properly. Also, a patch for Lucene4.0 codebase introducing a new PostingsFormat -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-4147) rollback/preparecommit thread hazard
[ https://issues.apache.org/jira/browse/LUCENE-4147?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Michael McCandless updated LUCENE-4147: --- Attachment: fail.log Here's the verbose output from a failure w/ the patch ... rollback/preparecommit thread hazard Key: LUCENE-4147 URL: https://issues.apache.org/jira/browse/LUCENE-4147 Project: Lucene - Java Issue Type: Bug Affects Versions: 4.0 Reporter: Robert Muir Attachments: LUCENE-4147.patch, LUCENE-4147.patch, fail.log found by http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows-Java7-64/70/ rollback should never throw this exception, as it documents it clears any pendingcommits. but it calls closeInternal outside of any sync block, so it looks like there is a race here. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4152) add one-syllable method to IndexReader enumerate subreaders
[ https://issues.apache.org/jira/browse/LUCENE-4152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395866#comment-13395866 ] Robert Muir commented on LUCENE-4152: - ok, so after LUCENE-3866 is resolved maybe all that is needed here is an eclipse rename? add one-syllable method to IndexReader enumerate subreaders --- Key: LUCENE-4152 URL: https://issues.apache.org/jira/browse/LUCENE-4152 Project: Lucene - Java Issue Type: Bug Reporter: Robert Muir Priority: Blocker Fix For: 4.0 Description is exactly as written. getSequentialSubReaders/getTopLevelReaderContext, these method names are way too long/unuseable. They also have tricky semantics (e.g. returning null). In lucene 4, people cannot just use any indexreader and get a merged view. So we need to make this stuff easy on them: * single-syllable method name (leaves(), subs(), i will think on this) * supports enhanced for-loop (no returning null or anything like that) * on indexreader (not atomic or composite, plain old indexreader) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3555) Lucene DirectoryReader can't find *.frq files created by solr
Nik V. Babichev created SOLR-3555: - Summary: Lucene DirectoryReader can't find *.frq files created by solr Key: SOLR-3555 URL: https://issues.apache.org/jira/browse/SOLR-3555 Project: Solr Issue Type: Bug Affects Versions: 4.0 Reporter: Nik V. Babichev Create index with Solr4 In index directory have such files: _0_Lucene40_0.frq _0_Lucene40_0.prx _0_Lucene40_0.tim _0_Lucene40_0.tip When try to open index by Lucene Directory reader got: java.io.FileNotFoundException: /opt/solr4/irotor/data/docs/index/_0_Lucene40.frq (No such file) File idxLocation = new File('path to index'); DirectoryFactory df = new StandardDirectoryFactory(); DirectoryReader.open(df.get(idxLocation.getAbsolutePath(), native)) Exception generated in this method: org.apache.lucene.codecs.lucene40.Lucene40PostingsReader(Directory dir, FieldInfos fieldInfos, SegmentInfo segmentInfo, IOContext ioContext, String segmentSuffix); freqIn = dir.openInput(IndexFileNames.segmentFileName(segmentInfo.name, segmentSuffix, Lucene40PostingsFormat.FREQ_EXTENSION), ioContext); When try to open unexisting _0_Lucene40.frq instead of _0_Lucene40_0.frq -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4152) add one-syllable method to IndexReader enumerate subreaders
[ https://issues.apache.org/jira/browse/LUCENE-4152?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395868#comment-13395868 ] Uwe Schindler commented on LUCENE-4152: --- Yes, and some final methods in IndexReader abstract class for easy usage, so you can do: {code:java} for (AtomicReaderContext ctx : reader.leaves()) ... {code} add one-syllable method to IndexReader enumerate subreaders --- Key: LUCENE-4152 URL: https://issues.apache.org/jira/browse/LUCENE-4152 Project: Lucene - Java Issue Type: Bug Reporter: Robert Muir Priority: Blocker Fix For: 4.0 Description is exactly as written. getSequentialSubReaders/getTopLevelReaderContext, these method names are way too long/unuseable. They also have tricky semantics (e.g. returning null). In lucene 4, people cannot just use any indexreader and get a merged view. So we need to make this stuff easy on them: * single-syllable method name (leaves(), subs(), i will think on this) * supports enhanced for-loop (no returning null or anything like that) * on indexreader (not atomic or composite, plain old indexreader) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-3555) Lucene DirectoryReader can't find *.frq files created by solr
[ https://issues.apache.org/jira/browse/SOLR-3555?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Robert Muir resolved SOLR-3555. --- Resolution: Invalid if you are using trunk, you need to subscribe to dev@ and commits@. when a note says reindex, we really mean it. Lucene DirectoryReader can't find *.frq files created by solr - Key: SOLR-3555 URL: https://issues.apache.org/jira/browse/SOLR-3555 Project: Solr Issue Type: Bug Affects Versions: 4.0 Reporter: Nik V. Babichev Create index with Solr4 In index directory have such files: _0_Lucene40_0.frq _0_Lucene40_0.prx _0_Lucene40_0.tim _0_Lucene40_0.tip When try to open index by Lucene Directory reader got: java.io.FileNotFoundException: /opt/solr4/irotor/data/docs/index/_0_Lucene40.frq (No such file) File idxLocation = new File('path to index'); DirectoryFactory df = new StandardDirectoryFactory(); DirectoryReader.open(df.get(idxLocation.getAbsolutePath(), native)) Exception generated in this method: org.apache.lucene.codecs.lucene40.Lucene40PostingsReader(Directory dir, FieldInfos fieldInfos, SegmentInfo segmentInfo, IOContext ioContext, String segmentSuffix); freqIn = dir.openInput(IndexFileNames.segmentFileName(segmentInfo.name, segmentSuffix, Lucene40PostingsFormat.FREQ_EXTENSION), ioContext); When try to open unexisting _0_Lucene40.frq instead of _0_Lucene40_0.frq -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-4153) Fast vector highlighting should be aware of field boundaries in case of mv fields.
Martijn van Groningen created LUCENE-4153: - Summary: Fast vector highlighting should be aware of field boundaries in case of mv fields. Key: LUCENE-4153 URL: https://issues.apache.org/jira/browse/LUCENE-4153 Project: Lucene - Java Issue Type: Improvement Reporter: Martijn van Groningen Improve fast vector highlighter to respect field boundaries in the case for multivalued fields. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-4153) Fast vector highlighting should be aware of field boundaries in case of mv fields.
[ https://issues.apache.org/jira/browse/LUCENE-4153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Martijn van Groningen updated LUCENE-4153: -- Attachment: LUCENE-4153.patch Initial approach to make fvh respect field boundaries when highlighting on mv fields. Basically in the BaseFragmentsBuilder the fragInfos are checked if not crossing field boundaries. If so a fragInfo might be modified or split. This approach seems good to me. Maybe others have a different opinion? This behavior can be enabled with the discreteMultiValueHighlighting boolean property. Fast vector highlighting should be aware of field boundaries in case of mv fields. -- Key: LUCENE-4153 URL: https://issues.apache.org/jira/browse/LUCENE-4153 Project: Lucene - Java Issue Type: Improvement Reporter: Martijn van Groningen Attachments: LUCENE-4153.patch Improve fast vector highlighter to respect field boundaries in the case for multivalued fields. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-3892) Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.)
[ https://issues.apache.org/jira/browse/LUCENE-3892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Han Jiang updated LUCENE-3892: -- Attachment: LUCENE-3892_for.patch LUCENE-3892_pfor.patch The new 3892_pfor patch fixed some SuppressingCodec stuff since last two weeks. And the 3892_for lazily implements For postingsformat based on current codes. These two patches are temporary separated, in order to prevent performance reduction for the sake of method overriding. Currently, blocksize ranges from 32 to 128 are tested on both two patches. However, for those skipping-intensive queries, there is no significant performance gain when smaller blocksize was applied. Here is a previous result for PFor, with blockSize=64, comparing with 128(in brackets): {noformat} TaskQPS Base StdDev BaseQPS PFor StdDev PFor Pct diff Phrase4.930.363.100.33 -47% - -25% (-47% - -25%) AndHighMed 27.922.26 19.161.72 -42% - -18% (-37% - -15%) SpanNear2.730.161.960.24 -40% - -14% (-36% - -13%) SloppyPhrase4.190.213.200.30 -34% - -12% (-30% - -6%) Wildcard 19.440.87 17.110.94 -20% - -2% (-17% -3%) AndHighHigh7.500.386.610.59 -23% - 1% (-19% -6%) IntNRQ4.060.523.880.35 -22% - 19% (-16% - 24%) Prefix3 31.001.69 30.452.29 -13% - 11% ( -6% - 20%) OrHighHigh4.160.474.110.34 -18% - 20% (-14% - 27%) OrHighMed4.980.594.940.41 -18% - 22% (-14% - 27%) Respell 40.292.11 40.112.13 -10% - 10% (-15% -2%) TermBGroup1M 20.500.32 20.520.80 -5% - 5% ( 1% - 10%) TermGroup1M 13.510.43 13.610.40 -5% - 7% ( 1% -9%) Fuzzy1 43.201.83 44.021.95 -6% - 11% (-11% -1%) PKLookup 87.161.78 89.520.940% - 5% ( -2% -7%) Fuzzy2 16.090.80 16.540.77 -6% - 13% (-11% -6%) Term 43.561.53 45.263.84 -8% - 16% ( 2% - 26%) TermBGroup1M1P 21.330.64 22.241.23 -4% - 13% ( 0% - 14%) {noformat} Also, the For postingsformat shows few performance change. So I suppose the bottleneck isn't in this method: PForUtil.patchException. Here is an example with blockSize=64: {noformat} TaskQPS Base StdDev Base QPS For StdDev For Pct diff Phrase5.030.453.300.43 -47% - -18% AndHighMed 28.052.33 18.831.77 -43% - -19% SpanNear2.690.181.940.25 -40% - -12% SloppyPhrase4.190.203.220.35 -34% - -10% AndHighHigh7.610.466.410.54 -27% - -2% Respell 41.361.65 37.942.42 -17% - 1% Wildcard 19.200.77 17.890.99 -15% - 2% OrHighHigh4.220.373.940.32 -21% - 10% OrHighMed5.060.464.730.39 -21% - 11% Fuzzy1 44.151.31 42.381.74 -10% - 2% Fuzzy2 16.480.59 15.840.76 -11% - 4% TermGroup1M 13.320.35 13.440.53 -5% - 7% PKLookup 87.701.81 88.621.22 -2% - 4% TermBGroup1M 20.140.47 20.400.59 -3% - 6% Prefix3 30.311.49 31.082.26 -9% - 15% TermBGroup1M1P 21.130.46 21.791.42 -5% - 12% IntNRQ3.960.454.140.46 -16% - 31% Term 43.071.51 46.064.50 -6% - 21% {noformat} Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.) - Key: LUCENE-3892 URL: https://issues.apache.org/jira/browse/LUCENE-3892 Project: Lucene - Java Issue Type: Improvement Reporter: Michael McCandless Labels: gsoc2012,
[jira] [Resolved] (LUCENE-4108) Add replaceTaxonomy to DirectoryTaxonomyWriter
[ https://issues.apache.org/jira/browse/LUCENE-4108?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shai Erera resolved LUCENE-4108. Resolution: Fixed I committed a fix for that -- moved the relevant entries to the 4.0.0 section. Hope it's ok now. Add replaceTaxonomy to DirectoryTaxonomyWriter -- Key: LUCENE-4108 URL: https://issues.apache.org/jira/browse/LUCENE-4108 Project: Lucene - Java Issue Type: New Feature Components: modules/facet Reporter: Shai Erera Assignee: Shai Erera Priority: Minor Fix For: 4.0, 5.0 Attachments: LUCENE-4108.patch I'd like to add the option to replace a taxonomy with another one, live on DirectoryTaxoWriter. Similar to how one can do so on IndexWriter, by calling deleteAll() and addIndexes(), it is valuable to be able to do so on DirTW as well. That way, if you need to replace a search index and taxonomy index live, you will be able to, given this new API. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-4148) _TestUtil should be able to generate random longs
[ https://issues.apache.org/jira/browse/LUCENE-4148?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Adrien Grand updated LUCENE-4148: - Attachment: LUCENE-4148.patch Here is a patch proposition: * {{_TestUtil.nextInt}} forwards calls to {{RandomInts.randomIntBetween}}, as Robert suggested, * A new {{_TestUtil.nextLong}} uses {{Random.nextInt(int)}} when the range does not exceed {{Integer.MAX_VALUE}} and scales {{Double.nextDouble}} into the expected range of values otherwise using {{BigDecimal}} arithmetic. I also updated a few tests to use this new method (especially a call to {{Math.abs(random.nextLong())}} in {{TestIndexInput}}, which is buggy since {{random.nextLong()}} might return Long.MIN_VALUE whose absolute value cannot be represented by a long). _TestUtil should be able to generate random longs - Key: LUCENE-4148 URL: https://issues.apache.org/jira/browse/LUCENE-4148 Project: Lucene - Java Issue Type: Improvement Components: general/test Reporter: Adrien Grand Priority: Trivial Attachments: LUCENE-4148.patch It would be helpful in TestPackedInts at least, in order to generate random values (as a workaround, we currently generate a random int between 0 and {{min(Integer.MAX_VALUE, PackedInts.maxValue(bitsPerValue)}}). Moreover, it would help to fix {{nextInt}} for large ranges (calling {{nextInt(random, -10, Integer.MAX_VALUE)}} or even {{nextInt(random, 0, Integer.MAX_VALUE)}} currently fails because the range of values is {{ Integer.MAX_VALUE}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3892) Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.)
[ https://issues.apache.org/jira/browse/LUCENE-3892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395894#comment-13395894 ] Han Jiang commented on LUCENE-3892: --- There's a potential bottleneck during method calling...Here is an example for PFor, with blocksize=128, exception rate = 97%, normal value = 2 bits, exception value = 32 bits: {noformat} Decoding normal values: 4703 ns Patching exceptions: 5797 ns Single call of PForUtil.decompress totally takes: 58318 ns {noformat} In addition, it costs about 4000ns to record the time span. Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.) - Key: LUCENE-3892 URL: https://issues.apache.org/jira/browse/LUCENE-3892 Project: Lucene - Java Issue Type: Improvement Reporter: Michael McCandless Labels: gsoc2012, lucene-gsoc-12 Fix For: 4.1 Attachments: LUCENE-3892_for.patch, LUCENE-3892_pfor.patch, LUCENE-3892_pfor.patch, LUCENE-3892_pfor.patch, LUCENE-3892_settings.patch, LUCENE-3892_settings.patch On the flex branch we explored a number of possible intblock encodings, but for whatever reason never brought them to completion. There are still a number of issues opened with patches in different states. Initial results (based on prototype) were excellent (see http://blog.mikemccandless.com/2010/08/lucene-performance-with-pfordelta-codec.html ). I think this would make a good GSoC project. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-tests-only-4.x - Build # 128 - Still Failing
Build: https://builds.apache.org/job/Lucene-Solr-tests-only-4.x/128/ No tests ran. Build Log: [...truncated 466 lines...] [echo] Building solr-langid... ivy-availability-check: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/ivy-settings.xml resolve: [echo] Building solr-uima... ivy-availability-check: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/ivy-settings.xml resolve: [echo] Building solr-velocity... ivy-availability-check: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/ivy-settings.xml resolve: validate: [echo] License check under: /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/solr [licenses] Scanned 82 JAR file(s) for licenses (in 0.79s.), 0 error(s). test: test-core: jflex-uptodate-check: jflex-notice: javacc-uptodate-check: javacc-notice: ivy-availability-check: ivy-fail: ivy-configure: [ivy:configure] :: loading settings :: file = /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/ivy-settings.xml resolve: init: clover.setup: clover.info: [echo] [echo] Clover not found. Code coverage reports disabled. [echo] clover: common.compile-core: [mkdir] Created dir: /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/build/core/classes/java [javac] Compiling 649 source files to /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/build/core/classes/java [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:506: class, interface, or enum expected [javac] package org.apache.lucene.index; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:525: class, interface, or enum expected [javac] import org.apache.lucene.analysis.Analyzer; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:526: class, interface, or enum expected [javac] import org.apache.lucene.codecs.Codec; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:527: class, interface, or enum expected [javac] import org.apache.lucene.index.DocumentsWriterPerThread.IndexingChain; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:528: class, interface, or enum expected [javac] import org.apache.lucene.index.IndexWriter.IndexReaderWarmer; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:529: class, interface, or enum expected [javac] import org.apache.lucene.index.IndexWriterConfig.OpenMode; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:530: class, interface, or enum expected [javac] import org.apache.lucene.search.IndexSearcher; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:531: class, interface, or enum expected [javac] import org.apache.lucene.search.similarities.Similarity; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:532: class, interface, or enum expected [javac] import org.apache.lucene.util.InfoStream; [javac] ^ [javac] /usr/home/hudson/hudson-slave/workspace/Lucene-Solr-tests-only-4.x/checkout/lucene/core/src/java/org/apache/lucene/index/LiveIndexWriterConfig.java:533: class, interface, or enum expected [javac] import org.apache.lucene.util.Version; [javac] ^ [javac] 10 errors [...truncated 15 lines...] [...truncated 582 lines...] [...truncated 582 lines...] [...truncated 582 lines...] [...truncated 582 lines...] [...truncated 582 lines...] - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3555) Lucene DirectoryReader can't find *.frq files created by solr
[ https://issues.apache.org/jira/browse/SOLR-3555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395904#comment-13395904 ] Nik V. Babichev commented on SOLR-3555: --- Thanks, I resolve my current problem. I build solr.war from SVN branch_4x and use maven artifact from 4.0-SNAPSHOT (https://repository.apache.org/content/groups/snapshots) for lucene requests. As I see 4.0-SNAPSHOT have difference with svn version. Will try trunk and 5.0-SNAPSHOT Lucene DirectoryReader can't find *.frq files created by solr - Key: SOLR-3555 URL: https://issues.apache.org/jira/browse/SOLR-3555 Project: Solr Issue Type: Bug Affects Versions: 4.0 Reporter: Nik V. Babichev Create index with Solr4 In index directory have such files: _0_Lucene40_0.frq _0_Lucene40_0.prx _0_Lucene40_0.tim _0_Lucene40_0.tip When try to open index by Lucene Directory reader got: java.io.FileNotFoundException: /opt/solr4/irotor/data/docs/index/_0_Lucene40.frq (No such file) File idxLocation = new File('path to index'); DirectoryFactory df = new StandardDirectoryFactory(); DirectoryReader.open(df.get(idxLocation.getAbsolutePath(), native)) Exception generated in this method: org.apache.lucene.codecs.lucene40.Lucene40PostingsReader(Directory dir, FieldInfos fieldInfos, SegmentInfo segmentInfo, IOContext ioContext, String segmentSuffix); freqIn = dir.openInput(IndexFileNames.segmentFileName(segmentInfo.name, segmentSuffix, Lucene40PostingsFormat.FREQ_EXTENSION), ioContext); When try to open unexisting _0_Lucene40.frq instead of _0_Lucene40_0.frq -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-3892) Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.)
[ https://issues.apache.org/jira/browse/LUCENE-3892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Han Jiang updated LUCENE-3892: -- Attachment: (was: LUCENE-3892_for.patch) Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.) - Key: LUCENE-3892 URL: https://issues.apache.org/jira/browse/LUCENE-3892 Project: Lucene - Java Issue Type: Improvement Reporter: Michael McCandless Labels: gsoc2012, lucene-gsoc-12 Fix For: 4.1 Attachments: LUCENE-3892_for.patch, LUCENE-3892_pfor.patch, LUCENE-3892_pfor.patch, LUCENE-3892_pfor.patch, LUCENE-3892_settings.patch, LUCENE-3892_settings.patch On the flex branch we explored a number of possible intblock encodings, but for whatever reason never brought them to completion. There are still a number of issues opened with patches in different states. Initial results (based on prototype) were excellent (see http://blog.mikemccandless.com/2010/08/lucene-performance-with-pfordelta-codec.html ). I think this would make a good GSoC project. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-3892) Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.)
[ https://issues.apache.org/jira/browse/LUCENE-3892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Han Jiang updated LUCENE-3892: -- Attachment: LUCENE-3892_for.patch LUCENE-3892_pfor.patch Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.) - Key: LUCENE-3892 URL: https://issues.apache.org/jira/browse/LUCENE-3892 Project: Lucene - Java Issue Type: Improvement Reporter: Michael McCandless Labels: gsoc2012, lucene-gsoc-12 Fix For: 4.1 Attachments: LUCENE-3892_for.patch, LUCENE-3892_pfor.patch, LUCENE-3892_pfor.patch, LUCENE-3892_pfor.patch, LUCENE-3892_settings.patch, LUCENE-3892_settings.patch On the flex branch we explored a number of possible intblock encodings, but for whatever reason never brought them to completion. There are still a number of issues opened with patches in different states. Initial results (based on prototype) were excellent (see http://blog.mikemccandless.com/2010/08/lucene-performance-with-pfordelta-codec.html ). I think this would make a good GSoC project. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-3892) Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.)
[ https://issues.apache.org/jira/browse/LUCENE-3892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Han Jiang updated LUCENE-3892: -- Attachment: (was: LUCENE-3892_pfor.patch) Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.) - Key: LUCENE-3892 URL: https://issues.apache.org/jira/browse/LUCENE-3892 Project: Lucene - Java Issue Type: Improvement Reporter: Michael McCandless Labels: gsoc2012, lucene-gsoc-12 Fix For: 4.1 Attachments: LUCENE-3892_for.patch, LUCENE-3892_pfor.patch, LUCENE-3892_pfor.patch, LUCENE-3892_pfor.patch, LUCENE-3892_settings.patch, LUCENE-3892_settings.patch On the flex branch we explored a number of possible intblock encodings, but for whatever reason never brought them to completion. There are still a number of issues opened with patches in different states. Initial results (based on prototype) were excellent (see http://blog.mikemccandless.com/2010/08/lucene-performance-with-pfordelta-codec.html ). I think this would make a good GSoC project. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-3555) Lucene DirectoryReader can't find *.frq files created by solr
[ https://issues.apache.org/jira/browse/SOLR-3555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395904#comment-13395904 ] Nik V. Babichev edited comment on SOLR-3555 at 6/18/12 2:20 PM: Thanks, I resolve my current problem. I build solr.war from SVN branch_4x and use maven artifact from 4.0-SNAPSHOT (https://repository.apache.org/content/groups/snapshots) for lucene requests. As I see 4.0-SNAPSHOT have difference with svn version. Trunk and 5.0-SNAPSHOT works well was (Author: sharky): Thanks, I resolve my current problem. I build solr.war from SVN branch_4x and use maven artifact from 4.0-SNAPSHOT (https://repository.apache.org/content/groups/snapshots) for lucene requests. As I see 4.0-SNAPSHOT have difference with svn version. Will try trunk and 5.0-SNAPSHOT Lucene DirectoryReader can't find *.frq files created by solr - Key: SOLR-3555 URL: https://issues.apache.org/jira/browse/SOLR-3555 Project: Solr Issue Type: Bug Affects Versions: 4.0 Reporter: Nik V. Babichev Create index with Solr4 In index directory have such files: _0_Lucene40_0.frq _0_Lucene40_0.prx _0_Lucene40_0.tim _0_Lucene40_0.tip When try to open index by Lucene Directory reader got: java.io.FileNotFoundException: /opt/solr4/irotor/data/docs/index/_0_Lucene40.frq (No such file) File idxLocation = new File('path to index'); DirectoryFactory df = new StandardDirectoryFactory(); DirectoryReader.open(df.get(idxLocation.getAbsolutePath(), native)) Exception generated in this method: org.apache.lucene.codecs.lucene40.Lucene40PostingsReader(Directory dir, FieldInfos fieldInfos, SegmentInfo segmentInfo, IOContext ioContext, String segmentSuffix); freqIn = dir.openInput(IndexFileNames.segmentFileName(segmentInfo.name, segmentSuffix, Lucene40PostingsFormat.FREQ_EXTENSION), ioContext); When try to open unexisting _0_Lucene40.frq instead of _0_Lucene40_0.frq -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-4.x-Linux-Java6-64 - Build # 170 - Failure!
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Linux-Java6-64/170/ All tests passed Build Log: [...truncated 4766 lines...] [...truncated 4766 lines...] [...truncated 4766 lines...] [...truncated 4766 lines...] [...truncated 4747 lines...] [junit4] ERROR: Forked JVM execution exception: 0, execution line: /mnt/ssd/jenkins/tools/java/jdk1.6.0_32/jre/bin/java -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/var/lib/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/heapdumps/ -Dtests.prefix=tests -Dtests.seed=4C54E59EBD35BEF0 -Xmx512M -Dtests.iters= -Dtests.verbose=false -Dtests.infostream=false -Dtests.lockdir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build -Dtests.codec=random -Dtests.postingsformat=random -Dtests.locale=random -Dtests.timezone=random -Dtests.directory=random -Dtests.linedocsfile=europarl.lines.txt.gz -Dtests.luceneMatchVersion=4.0 -Dtests.cleanthreads=perMethod -Djava.util.logging.config.file=/dev/null -Dtests.nightly=false -Dtests.weekly=false -Dtests.slow=false -Dtests.asserts.gracious=false -Dtests.multiplier=3 -DtempDir=. -Dlucene.version=4.0-SNAPSHOT -Djetty.testMode=1 -Djetty.insecurerandom=1 -Dsolr.directoryFactory=org.apache.solr.core.MockDirectoryFactory -classpath /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/test-framework/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/core/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/test-framework/lib/junit-4.10.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/test-framework/lib/randomizedtesting-runner-1.5.0.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/analysis/common/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/analysis/common/classes/test:/var/lib/jenkins/tools/ant/supported18/lib/ant-launcher.jar:/var/lib/jenkins/.ant/lib/ivy-2.2.0.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-jsch.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-jai.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-testutil.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-commons-net.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-log4j.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-jmf.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-antlr.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-junit.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-junit4.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-commons-logging.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-swing.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-oro.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-netrexx.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-resolver.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-xalan2.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-regexp.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-jdepend.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-bcel.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-bsf.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-javamail.jar:/mnt/ssd/jenkins/tools/java/jdk1.6.0_32/lib/tools.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/test-framework/lib/junit4-ant-1.5.0.jar -ea:org.apache.lucene... -ea:org.apache.solr... com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe -eventsfile /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/analysis/common/test/junit4-J0-1422064855254770f.events @/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/analysis/common/test/junit4-J0-1422064855254770f.suites [junit4] ERROR: Forked JVM execution exception: 1, execution line: /mnt/ssd/jenkins/tools/java/jdk1.6.0_32/jre/bin/java -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/var/lib/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/heapdumps/ -Dtests.prefix=tests -Dtests.seed=4C54E59EBD35BEF0 -Xmx512M -Dtests.iters= -Dtests.verbose=false -Dtests.infostream=false -Dtests.lockdir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build -Dtests.codec=random -Dtests.postingsformat=random -Dtests.locale=random -Dtests.timezone=random -Dtests.directory=random -Dtests.linedocsfile=europarl.lines.txt.gz -Dtests.luceneMatchVersion=4.0 -Dtests.cleanthreads=perMethod -Djava.util.logging.config.file=/dev/null -Dtests.nightly=false -Dtests.weekly=false -Dtests.slow=false -Dtests.asserts.gracious=false -Dtests.multiplier=3 -DtempDir=. -Dlucene.version=4.0-SNAPSHOT -Djetty.testMode=1 -Djetty.insecurerandom=1 -Dsolr.directoryFactory=org.apache.solr.core.MockDirectoryFactory -classpath
RE: [JENKINS] Lucene-Solr-4.x-Linux-Java6-64 - Build # 170 - Failure!
This was a test by Steven and me! - Uwe Schindler H.-H.-Meier-Allee 63, D-28213 Bremen http://www.thetaphi.de eMail: u...@thetaphi.de -Original Message- From: Jenkins Server with Girls [mailto:jenk...@sd-datasolutions.de] Sent: Monday, June 18, 2012 4:23 PM To: dev@lucene.apache.org; sh...@apache.org Subject: [JENKINS] Lucene-Solr-4.x-Linux-Java6-64 - Build # 170 - Failure! Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Linux-Java6- 64/170/ All tests passed Build Log: [...truncated 4766 lines...] [...truncated 4766 lines...] [...truncated 4766 lines...] [...truncated 4766 lines...] [...truncated 4747 lines...] [junit4] ERROR: Forked JVM execution exception: 0, execution line: /mnt/ssd/jenkins/tools/java/jdk1.6.0_32/jre/bin/java - XX:+HeapDumpOnOutOfMemoryError - XX:HeapDumpPath=/var/lib/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/heapdumps/ -Dtests.prefix=tests -Dtests.seed=4C54E59EBD35BEF0 - Xmx512M -Dtests.iters= -Dtests.verbose=false -Dtests.infostream=false - Dtests.lockdir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build -Dtests.codec=random - Dtests.postingsformat=random -Dtests.locale=random - Dtests.timezone=random -Dtests.directory=random - Dtests.linedocsfile=europarl.lines.txt.gz -Dtests.luceneMatchVersion=4.0 - Dtests.cleanthreads=perMethod -Djava.util.logging.config.file=/dev/null - Dtests.nightly=false -Dtests.weekly=false -Dtests.slow=false - Dtests.asserts.gracious=false -Dtests.multiplier=3 -DtempDir=. - Dlucene.version=4.0-SNAPSHOT -Djetty.testMode=1 -Djetty.insecurerandom=1 -Dsolr.directoryFactory=org.apache.solr.core.MockDirectoryFactory -classpath /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build/test- framework/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux- Java6- 64/checkout/lucene/build/core/classes/java:/mnt/ssd/jenkins/workspace/Lucen e-Solr-4.x-Linux-Java6-64/checkout/lucene/test-framework/lib/junit- 4.10.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/test-framework/lib/randomizedtesting-runner- 1.5.0.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build/analysis/common/classes/java:/mnt/ssd/jenkins/wor kspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build/analysis/common/classes/test:/var/lib/jenkins/tools/ ant/supported18/lib/ant-launcher.jar:/var/lib/jenkins/.ant/lib/ivy- 2.2.0.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- jsch.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- jai.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- testutil.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-commons- net.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- log4j.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- jmf.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- antlr.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant.jar:/var/lib/jenkins/tool s/ant/supported18/lib/ant- junit.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- junit4.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-commons- logging.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- swing.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- oro.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- netrexx.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- resolver.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- xalan2.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- regexp.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- jdepend.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- bcel.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- bsf.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- javamail.jar:/mnt/ssd/jenkins/tools/java/jdk1.6.0_32/lib/tools.jar:/mnt/ssd/jenk ins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/test- framework/lib/junit4-ant-1.5.0.jar -ea:org.apache.lucene... - ea:org.apache.solr... com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe - eventsfile /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build/analysis/common/test/junit4-J0- 1422064855254770f.events @/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x- Linux-Java6-64/checkout/lucene/build/analysis/common/test/junit4-J0- 1422064855254770f.suites [junit4] ERROR: Forked JVM execution exception: 1, execution line: /mnt/ssd/jenkins/tools/java/jdk1.6.0_32/jre/bin/java - XX:+HeapDumpOnOutOfMemoryError - XX:HeapDumpPath=/var/lib/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/heapdumps/ -Dtests.prefix=tests -Dtests.seed=4C54E59EBD35BEF0 - Xmx512M -Dtests.iters= -Dtests.verbose=false -Dtests.infostream=false - Dtests.lockdir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build -Dtests.codec=random - Dtests.postingsformat=random -Dtests.locale=random - Dtests.timezone=random -Dtests.directory=random -
[jira] [Created] (SOLR-3556) Provide LBHttpSolrServer based on Http Components in Solr 3.x
Andreas Hubold created SOLR-3556: Summary: Provide LBHttpSolrServer based on Http Components in Solr 3.x Key: SOLR-3556 URL: https://issues.apache.org/jira/browse/SOLR-3556 Project: Solr Issue Type: Bug Components: clients - java Affects Versions: 3.6 Reporter: Andreas Hubold Fix For: 3.6.1 With SOLR-2020 Solr 3.6 provides a HttpSolrServer based on Apache Http Components. But the LBHttpSolrServer still uses commons-httpclient. It would be cool if you could provide a loadbalancing implementation based on HttpComponents as well. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3556) Provide LBHttpSolrServer based on Http Components in Solr 3.x
[ https://issues.apache.org/jira/browse/SOLR-3556?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andreas Hubold updated SOLR-3556: - Priority: Minor (was: Major) Provide LBHttpSolrServer based on Http Components in Solr 3.x - Key: SOLR-3556 URL: https://issues.apache.org/jira/browse/SOLR-3556 Project: Solr Issue Type: Bug Components: clients - java Affects Versions: 3.6 Reporter: Andreas Hubold Priority: Minor Fix For: 3.6.1 With SOLR-2020 Solr 3.6 provides a HttpSolrServer based on Apache Http Components. But the LBHttpSolrServer still uses commons-httpclient. It would be cool if you could provide a loadbalancing implementation based on HttpComponents as well. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-4069) Segment-level Bloom filters for a 2 x speed up on rare term searches
[ https://issues.apache.org/jira/browse/LUCENE-4069?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mark Harwood updated LUCENE-4069: - Attachment: (was: BloomFilterPostingsBranch4x.patch) Segment-level Bloom filters for a 2 x speed up on rare term searches Key: LUCENE-4069 URL: https://issues.apache.org/jira/browse/LUCENE-4069 Project: Lucene - Java Issue Type: Improvement Components: core/index Affects Versions: 3.6, 4.0 Reporter: Mark Harwood Priority: Minor Fix For: 4.0, 3.6.1 Attachments: BloomFilterPostingsBranch4x.patch, MHBloomFilterOn3.6Branch.patch, PrimaryKeyPerfTest40.java An addition to each segment which stores a Bloom filter for selected fields in order to give fast-fail to term searches, helping avoid wasted disk access. Best suited for low-frequency fields e.g. primary keys on big indexes with many segments but also speeds up general searching in my tests. Overview slideshow here: http://www.slideshare.net/MarkHarwood/lucene-bloomfilteredsegments Benchmarks based on Wikipedia content here: http://goo.gl/X7QqU Patch based on 3.6 codebase attached. There are no 3.6 API changes currently - to play just add a field with _blm on the end of the name to invoke special indexing/querying capability. Clearly a new Field or schema declaration(!) would need adding to APIs to configure the service properly. Also, a patch for Lucene4.0 codebase introducing a new PostingsFormat -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (LUCENE-4069) Segment-level Bloom filters for a 2 x speed up on rare term searches
[ https://issues.apache.org/jira/browse/LUCENE-4069?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mark Harwood updated LUCENE-4069: - Attachment: BloomFilterPostingsBranch4x.patch New patch with Terms.intersect overridden for faster Fuzzy queries. Didn't get too far with running the Wikipedia perf tests due to missing data file (see http://code.google.com/a/apache-extras.org/p/luceneutil/issues/detail?id=7 ) Segment-level Bloom filters for a 2 x speed up on rare term searches Key: LUCENE-4069 URL: https://issues.apache.org/jira/browse/LUCENE-4069 Project: Lucene - Java Issue Type: Improvement Components: core/index Affects Versions: 3.6, 4.0 Reporter: Mark Harwood Priority: Minor Fix For: 4.0, 3.6.1 Attachments: BloomFilterPostingsBranch4x.patch, MHBloomFilterOn3.6Branch.patch, PrimaryKeyPerfTest40.java An addition to each segment which stores a Bloom filter for selected fields in order to give fast-fail to term searches, helping avoid wasted disk access. Best suited for low-frequency fields e.g. primary keys on big indexes with many segments but also speeds up general searching in my tests. Overview slideshow here: http://www.slideshare.net/MarkHarwood/lucene-bloomfilteredsegments Benchmarks based on Wikipedia content here: http://goo.gl/X7QqU Patch based on 3.6 codebase attached. There are no 3.6 API changes currently - to play just add a field with _blm on the end of the name to invoke special indexing/querying capability. Clearly a new Field or schema declaration(!) would need adding to APIs to configure the service properly. Also, a patch for Lucene4.0 codebase introducing a new PostingsFormat -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Solr - ConcurrentHashMap - NullPointerException
I am recalling big noise in JIRA; and Paul Nobble implemented first version of thisŠ what can be wrong now, after few years? Solr 3.6.0 Jun 18, 2012 3:40:43 PM org.apache.solr.common.SolrException log SEVERE: java.lang.NullPointerException at java.util.concurrent.ConcurrentHashMap.get(ConcurrentHashMap.java:768) at org.apache.solr.util.ConcurrentLRUCache.get(ConcurrentLRUCache.java:87) at org.apache.solr.search.FastLRUCache.get(FastLRUCache.java:136) at org.apache.solr.search.SolrIndexSearcher.getPositiveDocSet(SolrIndexSearcher .java:632) at org.apache.solr.search.SolrIndexSearcher.numDocs(SolrIndexSearcher.java:1743 ) at org.apache.solr.request.SimpleFacets.getFacetQueryCounts(SimpleFacets.java:2 34) at org.apache.solr.request.SimpleFacets.getFacetCounts(SimpleFacets.java:194) at org.apache.solr.handler.component.FacetComponent.process(FacetComponent.java :72) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHand ler.java:186) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase. java:129) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1376) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:3 65) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java: 260) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Application FilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterCh ain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.ja va:233) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.ja va:191) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127 ) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102 ) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java :109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:859) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http 11Protocol.java:588) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) at java.lang.Thread.run(Thread.java:680) Fuad Efendi 416-993-2060 http://www.tokenizer.ca
[jira] [Commented] (LUCENE-3866) Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader
[ https://issues.apache.org/jira/browse/LUCENE-3866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395968#comment-13395968 ] Michael McCandless commented on LUCENE-3866: I ran quick perf test: {noformat} TaskQPS base StdDev base QPS patchStdDev patch Pct diff Respell 92.151.40 90.302.70 -6% - 2% PKLookup 130.684.16 128.242.28 -6% - 3% Fuzzy2 41.790.40 41.141.11 -5% - 2% Fuzzy1 108.972.40 107.332.87 -6% - 3% AndHighHigh 16.190.48 15.970.31 -6% - 3% Phrase 12.900.32 12.740.36 -6% - 4% AndHighMed 64.181.81 63.461.83 -6% - 4% SloppyPhrase8.370.298.290.11 -5% - 3% SpanNear5.510.125.460.18 -6% - 4% TermGroup1M 36.190.60 35.890.74 -4% - 2% TermBGroup1M 70.640.49 70.740.57 -1% - 1% Prefix3 61.073.58 61.251.35 -7% - 8% Wildcard 40.842.20 41.000.98 -7% - 8% Term 147.323.85 149.654.87 -4% - 7% TermBGroup1M1P 50.321.58 51.290.88 -2% - 7% IntNRQ9.961.40 10.180.56 -15% - 25% OrHighHigh 10.310.74 10.560.57 -9% - 16% OrHighMed 12.951.01 13.260.79 -10% - 17% {noformat} Basically no real change ... good! Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader --- Key: LUCENE-3866 URL: https://issues.apache.org/jira/browse/LUCENE-3866 Project: Lucene - Java Issue Type: Improvement Reporter: Uwe Schindler Assignee: Uwe Schindler Fix For: 4.0 Attachments: LUCENE-3866-step1.patch, LUCENE-3866-step2.patch, LUCENE-3866-step2.patch Since Lucene 2.9 we have the method getSequentialSubReader() returning IndexReader[]. Based on hardly-to-debug errors in user's code, Robert and me realized that returning an array from a public API is an anti-pattern. If the array is intended to be modifiable (like byte[] in terms,...), it is fine to use arrays in public APIs, but not, if the array must be protected from modification. As IndexReaders are 100% unmodifiable in trunk code (no deletions,...), the only possibility to corrumpt the reader is by modifying the array returned by getSequentialSubReaders(). We should prevent this. The same theoretically applies to FieldCache, too - but the party that is afraid of performance problems is too big to fight against that :( For getSequentialSubReaders there is no performance problem at all. The binary search of reader-ids inside BaseCompositeReader can still be done with the internal protected array, but public APIs should expose only a unmodifiable List. The same applies to leaves() and children() in IndexReaderContext. This change to list would also allow to make CompositeReader and CompositeReaderContext IterableR extends IndexReader, so some loops would look nice. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Solr - ConcurrentHashMap - NullPointerException
Some research, line 768 of ConcurrentHashMap: int hash = hash(key.hashCode()); I believe we have Unit Test with null as a keyŠ very strange for Solr cache implementationŠ /** * Returns the value to which the specified key is mapped, * or {@code null} if this map contains no mapping for the key. * * pMore formally, if this map contains a mapping from a key * {@code k} to a value {@code v} such that {@code key.equals(k)}, * then this method returns {@code v}; otherwise it returns * {@code null}. (There can be at most one such mapping.) * * @throws NullPointerException if the specified key is null */ public V get(Object key) { int hash = hash(key.hashCode()); return segmentFor(hash).get(key, hash); } From: Fuad Efendi f...@efendi.ca Reply-To: lucene-dev dev@lucene.apache.org Date: Monday, 18 June, 2012 11:45 AM To: lucene-dev dev@lucene.apache.org Subject: Solr - ConcurrentHashMap - NullPointerException I am recalling big noise in JIRA; and Paul Nobble implemented first version of thisŠ what can be wrong now, after few years? Solr 3.6.0 Jun 18, 2012 3:40:43 PM org.apache.solr.common.SolrException log SEVERE: java.lang.NullPointerException at java.util.concurrent.ConcurrentHashMap.get(ConcurrentHashMap.java:768) at org.apache.solr.util.ConcurrentLRUCache.get(ConcurrentLRUCache.java:87) at org.apache.solr.search.FastLRUCache.get(FastLRUCache.java:136) at org.apache.solr.search.SolrIndexSearcher.getPositiveDocSet(SolrIndexSearcher .java:632) at org.apache.solr.search.SolrIndexSearcher.numDocs(SolrIndexSearcher.java:1743 ) at org.apache.solr.request.SimpleFacets.getFacetQueryCounts(SimpleFacets.java:2 34) at org.apache.solr.request.SimpleFacets.getFacetCounts(SimpleFacets.java:194) at org.apache.solr.handler.component.FacetComponent.process(FacetComponent.java :72) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHand ler.java:186) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase. java:129) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1376) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:3 65) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java: 260) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Application FilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterCh ain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.ja va:233) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.ja va:191) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127 ) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102 ) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java :109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:859) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http 11Protocol.java:588) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) at java.lang.Thread.run(Thread.java:680) Fuad Efendi 416-993-2060 http://www.tokenizer.ca
[jira] [Commented] (LUCENE-3866) Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader
[ https://issues.apache.org/jira/browse/LUCENE-3866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395975#comment-13395975 ] Uwe Schindler commented on LUCENE-3866: --- Thanks, I was not expecting a change - otherwise all my believes would have been devastated... :-) Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader --- Key: LUCENE-3866 URL: https://issues.apache.org/jira/browse/LUCENE-3866 Project: Lucene - Java Issue Type: Improvement Reporter: Uwe Schindler Assignee: Uwe Schindler Fix For: 4.0 Attachments: LUCENE-3866-step1.patch, LUCENE-3866-step2.patch, LUCENE-3866-step2.patch Since Lucene 2.9 we have the method getSequentialSubReader() returning IndexReader[]. Based on hardly-to-debug errors in user's code, Robert and me realized that returning an array from a public API is an anti-pattern. If the array is intended to be modifiable (like byte[] in terms,...), it is fine to use arrays in public APIs, but not, if the array must be protected from modification. As IndexReaders are 100% unmodifiable in trunk code (no deletions,...), the only possibility to corrumpt the reader is by modifying the array returned by getSequentialSubReaders(). We should prevent this. The same theoretically applies to FieldCache, too - but the party that is afraid of performance problems is too big to fight against that :( For getSequentialSubReaders there is no performance problem at all. The binary search of reader-ids inside BaseCompositeReader can still be done with the internal protected array, but public APIs should expose only a unmodifiable List. The same applies to leaves() and children() in IndexReaderContext. This change to list would also allow to make CompositeReader and CompositeReaderContext IterableR extends IndexReader, so some loops would look nice. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4069) Segment-level Bloom filters for a 2 x speed up on rare term searches
[ https://issues.apache.org/jira/browse/LUCENE-4069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395980#comment-13395980 ] Michael McCandless commented on LUCENE-4069: bq. Didn't get too far with running the Wikipedia perf tests due to missing data file (see http://code.google.com/a/apache-extras.org/p/luceneutil/issues/detail?id=7 ) Woops sorry: I posted a comment there (new, more recent wikipedia export). Segment-level Bloom filters for a 2 x speed up on rare term searches Key: LUCENE-4069 URL: https://issues.apache.org/jira/browse/LUCENE-4069 Project: Lucene - Java Issue Type: Improvement Components: core/index Affects Versions: 3.6, 4.0 Reporter: Mark Harwood Priority: Minor Fix For: 4.0, 3.6.1 Attachments: BloomFilterPostingsBranch4x.patch, MHBloomFilterOn3.6Branch.patch, PrimaryKeyPerfTest40.java An addition to each segment which stores a Bloom filter for selected fields in order to give fast-fail to term searches, helping avoid wasted disk access. Best suited for low-frequency fields e.g. primary keys on big indexes with many segments but also speeds up general searching in my tests. Overview slideshow here: http://www.slideshare.net/MarkHarwood/lucene-bloomfilteredsegments Benchmarks based on Wikipedia content here: http://goo.gl/X7QqU Patch based on 3.6 codebase attached. There are no 3.6 API changes currently - to play just add a field with _blm on the end of the name to invoke special indexing/querying capability. Clearly a new Field or schema declaration(!) would need adding to APIs to configure the service properly. Also, a patch for Lucene4.0 codebase introducing a new PostingsFormat -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4148) _TestUtil should be able to generate random longs
[ https://issues.apache.org/jira/browse/LUCENE-4148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395992#comment-13395992 ] Dawid Weiss commented on LUCENE-4148: - This looks good to me, although I'm wondering how much randomness do we really need -- those BigIntegers are worrying. I think this can be improved later if somebody wishes to dive into it. Btw. an interesting random number generator -- _TestUtil should be able to generate random longs - Key: LUCENE-4148 URL: https://issues.apache.org/jira/browse/LUCENE-4148 Project: Lucene - Java Issue Type: Improvement Components: general/test Reporter: Adrien Grand Priority: Trivial Attachments: LUCENE-4148.patch It would be helpful in TestPackedInts at least, in order to generate random values (as a workaround, we currently generate a random int between 0 and {{min(Integer.MAX_VALUE, PackedInts.maxValue(bitsPerValue)}}). Moreover, it would help to fix {{nextInt}} for large ranges (calling {{nextInt(random, -10, Integer.MAX_VALUE)}} or even {{nextInt(random, 0, Integer.MAX_VALUE)}} currently fails because the range of values is {{ Integer.MAX_VALUE}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (LUCENE-4148) _TestUtil should be able to generate random longs
[ https://issues.apache.org/jira/browse/LUCENE-4148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395992#comment-13395992 ] Dawid Weiss edited comment on LUCENE-4148 at 6/18/12 4:10 PM: -- This looks good to me, although I'm wondering how much randomness do we really need -- those BigIntegers are worrying. I think this can be improved later if somebody wishes to dive into it. Btw. an interesting random number generator -- http://home.southernct.edu/~pasqualonia1/ca/report.html was (Author: dweiss): This looks good to me, although I'm wondering how much randomness do we really need -- those BigIntegers are worrying. I think this can be improved later if somebody wishes to dive into it. Btw. an interesting random number generator -- _TestUtil should be able to generate random longs - Key: LUCENE-4148 URL: https://issues.apache.org/jira/browse/LUCENE-4148 Project: Lucene - Java Issue Type: Improvement Components: general/test Reporter: Adrien Grand Priority: Trivial Attachments: LUCENE-4148.patch It would be helpful in TestPackedInts at least, in order to generate random values (as a workaround, we currently generate a random int between 0 and {{min(Integer.MAX_VALUE, PackedInts.maxValue(bitsPerValue)}}). Moreover, it would help to fix {{nextInt}} for large ranges (calling {{nextInt(random, -10, Integer.MAX_VALUE)}} or even {{nextInt(random, 0, Integer.MAX_VALUE)}} currently fails because the range of values is {{ Integer.MAX_VALUE}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3866) Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader
[ https://issues.apache.org/jira/browse/LUCENE-3866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13395995#comment-13395995 ] Michael McCandless commented on LUCENE-3866: +1, patch looks good! Thanks Uwe. Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader --- Key: LUCENE-3866 URL: https://issues.apache.org/jira/browse/LUCENE-3866 Project: Lucene - Java Issue Type: Improvement Reporter: Uwe Schindler Assignee: Uwe Schindler Fix For: 4.0 Attachments: LUCENE-3866-step1.patch, LUCENE-3866-step2.patch, LUCENE-3866-step2.patch Since Lucene 2.9 we have the method getSequentialSubReader() returning IndexReader[]. Based on hardly-to-debug errors in user's code, Robert and me realized that returning an array from a public API is an anti-pattern. If the array is intended to be modifiable (like byte[] in terms,...), it is fine to use arrays in public APIs, but not, if the array must be protected from modification. As IndexReaders are 100% unmodifiable in trunk code (no deletions,...), the only possibility to corrumpt the reader is by modifying the array returned by getSequentialSubReaders(). We should prevent this. The same theoretically applies to FieldCache, too - but the party that is afraid of performance problems is too big to fight against that :( For getSequentialSubReaders there is no performance problem at all. The binary search of reader-ids inside BaseCompositeReader can still be done with the internal protected array, but public APIs should expose only a unmodifiable List. The same applies to leaves() and children() in IndexReaderContext. This change to list would also allow to make CompositeReader and CompositeReaderContext IterableR extends IndexReader, so some loops would look nice. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4150) Change version properties in branch_4x to 4.0-ALPHA-SNAPSHOT
[ https://issues.apache.org/jira/browse/LUCENE-4150?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13396001#comment-13396001 ] Hoss Man commented on LUCENE-4150: -- bq. About jira, after we release i do think we should make the version in JIRA so people can report bugs against it. But I don't think we should do that now (or people will get confused about what version to terget things). Just to be clear, i think this absolutely MUST be done, and would be total chaos otherwise. If we are going to vote and have official Alpha and Beta releases that we want end users to actually try out and test thoroughly, then we are going to need a way to record when a bug was found/fixed in 4.0-Alpha vs the 4.0-Beta vs the 4.0-final. Until we have a formally voted on Alpha it's not a big deal, but as soon as that moment happens, ever bug that currently says Fixed in: 4.0 should be bulked edited to say Fixed in: 4.0-Alpha, 4.0 -- so that 9 months from now when someone says I'm still running 4.0-Alpha because there was an API change to fix a bug that made it too hard for me to upgrade to 4.0-Beta there is a good clean record of what bug fixes they do/don't have. bq. No. We are targeting 4.0, these are just snapshots of that! Calling the alpha/beta snapshots is extremely missleading and should be avoided. The only usage of the term snapshot in apache (that i know of) are for *unofficial* nightly/jenkins/svn builds. Since the alpha/beta releases will in fact be _official_ (VOTEd) releases we should not ever imply/refer to them as snapshots since that would convey lower status on them (and reduce the likely hood of enn users testing them) then they will deserve. Change version properties in branch_4x to 4.0-ALPHA-SNAPSHOT --- Key: LUCENE-4150 URL: https://issues.apache.org/jira/browse/LUCENE-4150 Project: Lucene - Java Issue Type: Task Components: general/build Affects Versions: 4.0 Reporter: Steven Rowe Priority: Minor The next release off branch_4x will be named 4.0-ALPHA, so the current version string should be 4.0-ALPHA-SNAPSHOT. (Similarly, after 4.0-ALPHA is released, the version string should be changed to 4.0-BETA-SNAPSHOT.) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-4.x-Windows-Java7-64 - Build # 94 - Failure!
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows-Java7-64/94/ No tests ran. Build Log: [...truncated 10392 lines...] [...truncated 10392 lines...] [...truncated 10392 lines...] [...truncated 10392 lines...] [...truncated 10357 lines...] FATAL: hudson.remoting.RequestAbortedException: java.net.SocketException: Connection reset hudson.remoting.RequestAbortedException: hudson.remoting.RequestAbortedException: java.net.SocketException: Connection reset at hudson.remoting.Request.call(Request.java:149) at hudson.remoting.Channel.call(Channel.java:646) at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:158) at $Proxy38.join(Unknown Source) at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:861) at hudson.Launcher$ProcStarter.join(Launcher.java:345) at hudson.tasks.Ant.perform(Ant.java:217) at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:19) at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:717) at hudson.model.Build$BuildExecution.build(Build.java:199) at hudson.model.Build$BuildExecution.doRun(Build.java:160) at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:499) at hudson.model.Run.execute(Run.java:1460) at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46) at hudson.model.ResourceController.execute(ResourceController.java:88) at hudson.model.Executor.run(Executor.java:239) Caused by: hudson.remoting.RequestAbortedException: java.net.SocketException: Connection reset at hudson.remoting.Request.abort(Request.java:273) at hudson.remoting.Channel.terminate(Channel.java:702) at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:69) Caused by: java.net.SocketException: Connection reset at java.net.SocketInputStream.read(SocketInputStream.java:168) at java.io.FilterInputStream.read(FilterInputStream.java:116) at java.io.BufferedInputStream.fill(BufferedInputStream.java:218) at java.io.BufferedInputStream.read(BufferedInputStream.java:237) at java.io.ObjectInputStream$PeekInputStream.peek(ObjectInputStream.java:2248) at java.io.ObjectInputStream$BlockDataInputStream.peek(ObjectInputStream.java:2541) at java.io.ObjectInputStream$BlockDataInputStream.peekByte(ObjectInputStream.java:2551) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1296) at java.io.ObjectInputStream.readObject(ObjectInputStream.java:350) at hudson.remoting.Command.readFrom(Command.java:90) at hudson.remoting.ClassicCommandTransport.read(ClassicCommandTransport.java:59) at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:48) [...truncated 10392 lines...] - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Solr - ConcurrentHashMap - NullPointerException
Rare problemŠ it happened when I provided NULL facet query via solrconfig: str name=facet.query/str We probably should add null check: // only handle positive (non negative) queries DocSet getPositiveDocSet(Query q) throws IOException { DocSet answer; if (filterCache != null) { answer = filterCache.get(q); if (answer!=null) return answer; } answer = getDocSetNC(q,null); if (filterCache != null) filterCache.put( q,answer); return answer; } From: Fuad Efendi f...@efendi.ca Reply-To: lucene-dev dev@lucene.apache.org Date: Monday, 18 June, 2012 11:52 AM To: lucene-dev dev@lucene.apache.org Subject: Re: Solr - ConcurrentHashMap - NullPointerException Some research, line 768 of ConcurrentHashMap: int hash = hash(key.hashCode()); I believe we have Unit Test with null as a keyŠ very strange for Solr cache implementationŠ /** * Returns the value to which the specified key is mapped, * or {@code null} if this map contains no mapping for the key. * * pMore formally, if this map contains a mapping from a key * {@code k} to a value {@code v} such that {@code key.equals(k)}, * then this method returns {@code v}; otherwise it returns * {@code null}. (There can be at most one such mapping.) * * @throws NullPointerException if the specified key is null */ public V get(Object key) { int hash = hash(key.hashCode()); return segmentFor(hash).get(key, hash); } From: Fuad Efendi f...@efendi.ca Reply-To: lucene-dev dev@lucene.apache.org Date: Monday, 18 June, 2012 11:45 AM To: lucene-dev dev@lucene.apache.org Subject: Solr - ConcurrentHashMap - NullPointerException I am recalling big noise in JIRA; and Paul Nobble implemented first version of thisŠ what can be wrong now, after few years? Solr 3.6.0 Jun 18, 2012 3:40:43 PM org.apache.solr.common.SolrException log SEVERE: java.lang.NullPointerException at java.util.concurrent.ConcurrentHashMap.get(ConcurrentHashMap.java:768) at org.apache.solr.util.ConcurrentLRUCache.get(ConcurrentLRUCache.java:87) at org.apache.solr.search.FastLRUCache.get(FastLRUCache.java:136) at org.apache.solr.search.SolrIndexSearcher.getPositiveDocSet(SolrIndexSearcher .java:632) at org.apache.solr.search.SolrIndexSearcher.numDocs(SolrIndexSearcher.java:1743 ) at org.apache.solr.request.SimpleFacets.getFacetQueryCounts(SimpleFacets.java:2 34) at org.apache.solr.request.SimpleFacets.getFacetCounts(SimpleFacets.java:194) at org.apache.solr.handler.component.FacetComponent.process(FacetComponent.java :72) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHand ler.java:186) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase. java:129) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1376) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:3 65) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java: 260) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Application FilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterCh ain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.ja va:233) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.ja va:191) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127 ) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102 ) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java :109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:859) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http 11Protocol.java:588) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) at java.lang.Thread.run(Thread.java:680) Fuad Efendi 416-993-2060 http://www.tokenizer.ca
[JENKINS] Lucene-Solr-4.x-Linux-Java6-64 - Build # 174 - Failure!
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Linux-Java6-64/174/ All tests passed Build Log: [...truncated 5134 lines...] [junit4] ERROR: Forked JVM execution exception: 0, execution line: /mnt/ssd/jenkins/tools/java/jdk1.6.0_32/jre/bin/java -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/var/lib/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/heapdumps/ -Dtests.prefix=tests -Dtests.seed=2A2493D5321058C4 -Xmx512M -Dtests.iters= -Dtests.verbose=false -Dtests.infostream=false -Dtests.lockdir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build -Dtests.codec=random -Dtests.postingsformat=random -Dtests.locale=random -Dtests.timezone=random -Dtests.directory=random -Dtests.linedocsfile=europarl.lines.txt.gz -Dtests.luceneMatchVersion=4.0 -Dtests.cleanthreads=perMethod -Djava.util.logging.config.file=/dev/null -Dtests.nightly=false -Dtests.weekly=false -Dtests.slow=false -Dtests.asserts.gracious=false -Dtests.multiplier=3 -DtempDir=. -Dlucene.version=4.0-SNAPSHOT -Djetty.testMode=1 -Djetty.insecurerandom=1 -Dsolr.directoryFactory=org.apache.solr.core.MockDirectoryFactory -classpath /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/test-framework/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/grouping/lucene-grouping-4.0-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/core/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/test-framework/lib/junit-4.10.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/test-framework/lib/randomizedtesting-runner-1.5.0.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/join/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/join/classes/test:/var/lib/jenkins/tools/ant/supported18/lib/ant-launcher.jar:/var/lib/jenkins/.ant/lib/ivy-2.2.0.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-jsch.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-jai.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-testutil.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-commons-net.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-log4j.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-jmf.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-antlr.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-junit.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-junit4.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-commons-logging.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-swing.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-oro.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-netrexx.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-resolver.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-xalan2.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-regexp.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-jdepend.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-bcel.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache-bsf.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-javamail.jar:/mnt/ssd/jenkins/tools/java/jdk1.6.0_32/lib/tools.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/test-framework/lib/junit4-ant-1.5.0.jar -ea:org.apache.lucene... -ea:org.apache.solr... com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe -eventsfile /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/join/test/junit4-J0-165541892e65c073d.events @/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build/join/test/junit4-J0-165541892e65c073d.suites BUILD FAILED /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/build.xml:29: The following error occurred while executing this line: /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/build.xml:456: The following error occurred while executing this line: /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/common-build.xml:1417: The following error occurred while executing this line: /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/module-build.xml:62: The following error occurred while executing this line: /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/common-build.xml:1019: The following error occurred while executing this line: /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/common-build.xml:745: At least one slave process threw an unexpected exception, first: Forked process exited with an error code: 143 Total time: 18 minutes 16 seconds Build step 'Execute shell' marked build as failure Archiving artifacts Recording test results Email was triggered for: Failure Sending email for trigger: Failure
RE: [JENKINS] Lucene-Solr-4.x-Linux-Java6-64 - Build # 174 - Failure!
This was another test. Steven Rowe finally fixed the bug in the Extended-eMail-Publisher of Jenkins, it was caused not by our regex, it was caused by the parsing of the user input that contained our huge regexp (regexp used to match regexp - LOL). I think we will submit bug report and let the Apache people also update. Thanks for interruption, we are just fixing bugs in all components around Lucene. This time the Continuous Integration Server! Thanks @ Steven, Uwe - Uwe Schindler H.-H.-Meier-Allee 63, D-28213 Bremen http://www.thetaphi.de eMail: u...@thetaphi.de -Original Message- From: Jenkins Server with Girls [mailto:jenk...@sd-datasolutions.de] Sent: Monday, June 18, 2012 7:00 PM To: dev@lucene.apache.org Subject: [JENKINS] Lucene-Solr-4.x-Linux-Java6-64 - Build # 174 - Failure! Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Linux-Java6- 64/174/ All tests passed Build Log: [...truncated 5134 lines...] [junit4] ERROR: Forked JVM execution exception: 0, execution line: /mnt/ssd/jenkins/tools/java/jdk1.6.0_32/jre/bin/java - XX:+HeapDumpOnOutOfMemoryError - XX:HeapDumpPath=/var/lib/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/heapdumps/ -Dtests.prefix=tests -Dtests.seed=2A2493D5321058C4 - Xmx512M -Dtests.iters= -Dtests.verbose=false -Dtests.infostream=false - Dtests.lockdir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build -Dtests.codec=random - Dtests.postingsformat=random -Dtests.locale=random - Dtests.timezone=random -Dtests.directory=random - Dtests.linedocsfile=europarl.lines.txt.gz -Dtests.luceneMatchVersion=4.0 - Dtests.cleanthreads=perMethod -Djava.util.logging.config.file=/dev/null - Dtests.nightly=false -Dtests.weekly=false -Dtests.slow=false - Dtests.asserts.gracious=false -Dtests.multiplier=3 -DtempDir=. - Dlucene.version=4.0-SNAPSHOT -Djetty.testMode=1 -Djetty.insecurerandom=1 -Dsolr.directoryFactory=org.apache.solr.core.MockDirectoryFactory -classpath /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build/test- framework/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux- Java6-64/checkout/lucene/build/grouping/lucene-grouping-4.0- SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build/core/classes/java:/mnt/ssd/jenkins/workspace/Lucen e-Solr-4.x-Linux-Java6-64/checkout/lucene/test-framework/lib/junit- 4.10.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/test-framework/lib/randomizedtesting-runner- 1.5.0.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build/join/classes/java:/mnt/ssd/jenkins/workspace/Lucen e-Solr-4.x-Linux-Java6- 64/checkout/lucene/build/join/classes/test:/var/lib/jenkins/tools/ant/supported 18/lib/ant-launcher.jar:/var/lib/jenkins/.ant/lib/ivy- 2.2.0.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- jsch.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- jai.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- testutil.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-commons- net.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- log4j.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- jmf.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- antlr.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant.jar:/var/lib/jenkins/tool s/ant/supported18/lib/ant- junit.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- junit4.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-commons- logging.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- swing.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- oro.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- netrexx.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- resolver.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- xalan2.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- regexp.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- jdepend.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- bcel.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant-apache- bsf.jar:/var/lib/jenkins/tools/ant/supported18/lib/ant- javamail.jar:/mnt/ssd/jenkins/tools/java/jdk1.6.0_32/lib/tools.jar:/mnt/ssd/jenk ins/workspace/Lucene-Solr-4.x-Linux-Java6-64/checkout/lucene/test- framework/lib/junit4-ant-1.5.0.jar -ea:org.apache.lucene... - ea:org.apache.solr... com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe - eventsfile /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build/join/test/junit4-J0-165541892e65c073d.events @/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build/join/test/junit4-J0-165541892e65c073d.suites BUILD FAILED /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/build.xml:29: The following error occurred while executing this line: /mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux-Java6- 64/checkout/lucene/build.xml:456: The
[jira] [Commented] (LUCENE-3729) Allow using FST to hold terms data in DocValues.BYTES_*_SORTED
[ https://issues.apache.org/jira/browse/LUCENE-3729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13396032#comment-13396032 ] Michael McCandless commented on LUCENE-3729: Very cool! Why did you need a custom subclass of FSTEnum (instead of BytesRefFSTEnum)? Is it for getOutput? Couldn't you getCurrent().output instead? We could probably make a static Util.getCeil (like Util.get, but finds ceiling term if there is no exact match) ... would likely be faster than re-using an FSTEnum ... but sorta hairy to write. I'm curious how DV sort perf compares ... hopefully it's little loss (going on my previous tests here...). Awesome to fix seekCeil to return enum result ... maybe factor out the same enum from TermsEnum so we can share it? And you can remove the two TODOs from FSTEnum.java... Can ToFSTBytesRefConsumer be moved to FSTSortedBytesImpl.java? Allow using FST to hold terms data in DocValues.BYTES_*_SORTED -- Key: LUCENE-3729 URL: https://issues.apache.org/jira/browse/LUCENE-3729 Project: Lucene - Java Issue Type: Improvement Reporter: Michael McCandless Assignee: Michael McCandless Labels: gsoc2012, lucene-gsoc-11 Attachments: LUCENE-3729.patch, LUCENE-3729.patch, LUCENE-3729.patch, LUCENE-3729.patch, LUCENE-3729.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-trunk-Linux-Java7-64 - Build # 323 - Failure!
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-trunk-Linux-Java7-64/323/ 1 tests failed. REGRESSION: org.apache.lucene.analysis.ja.TestJapaneseTokenizer.testRandomHugeStrings Error Message: Java heap space Stack Trace: java.lang.OutOfMemoryError: Java heap space at __randomizedtesting.SeedInfo.seed([1A47DEC61940914D:8264B90547362D05]:0) at org.apache.lucene.util.fst.FST.init(FST.java:331) at org.apache.lucene.codecs.memory.MemoryPostingsFormat$TermsReader.init(MemoryPostingsFormat.java:811) at org.apache.lucene.codecs.memory.MemoryPostingsFormat.fieldsProducer(MemoryPostingsFormat.java:858) at org.apache.lucene.codecs.perfield.PerFieldPostingsFormat$FieldsReader.init(PerFieldPostingsFormat.java:186) at org.apache.lucene.codecs.perfield.PerFieldPostingsFormat.fieldsProducer(PerFieldPostingsFormat.java:250) at org.apache.lucene.index.SegmentCoreReaders.init(SegmentCoreReaders.java:107) at org.apache.lucene.index.SegmentReader.init(SegmentReader.java:55) at org.apache.lucene.index.StandardDirectoryReader$1.doBody(StandardDirectoryReader.java:62) at org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:639) at org.apache.lucene.index.StandardDirectoryReader.open(StandardDirectoryReader.java:52) at org.apache.lucene.index.DirectoryReader.open(DirectoryReader.java:64) at org.apache.lucene.store.MockDirectoryWrapper.close(MockDirectoryWrapper.java:602) at org.apache.lucene.util.IOUtils.close(IOUtils.java:143) at org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:472) at org.apache.lucene.analysis.BaseTokenStreamTestCase.checkRandomData(BaseTokenStreamTestCase.java:378) at org.apache.lucene.analysis.ja.TestJapaneseTokenizer.testRandomHugeStrings(TestJapaneseTokenizer.java:193) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:601) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1969) at com.carrotsearch.randomizedtesting.RandomizedRunner.access$1100(RandomizedRunner.java:132) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:814) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:875) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:889) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:32) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleReportUncaughtExceptions$1.evaluate(TestRuleReportUncaughtExceptions.java:68) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) Build Log: [...truncated 4666 lines...] [junit4] Suite: org.apache.lucene.analysis.ja.TestJapaneseTokenizer [junit4] ERROR 28.6s J0 | TestJapaneseTokenizer.testRandomHugeStrings [junit4] Throwable #1: java.lang.OutOfMemoryError: Java heap space [junit4]at __randomizedtesting.SeedInfo.seed([1A47DEC61940914D:8264B90547362D05]:0) [junit4]at org.apache.lucene.util.fst.FST.init(FST.java:331) [junit4]at org.apache.lucene.codecs.memory.MemoryPostingsFormat$TermsReader.init(MemoryPostingsFormat.java:811) [junit4]at org.apache.lucene.codecs.memory.MemoryPostingsFormat.fieldsProducer(MemoryPostingsFormat.java:858) [junit4]at org.apache.lucene.codecs.perfield.PerFieldPostingsFormat$FieldsReader.init(PerFieldPostingsFormat.java:186) [junit4]at org.apache.lucene.codecs.perfield.PerFieldPostingsFormat.fieldsProducer(PerFieldPostingsFormat.java:250) [junit4]at org.apache.lucene.index.SegmentCoreReaders.init(SegmentCoreReaders.java:107) [junit4]at org.apache.lucene.index.SegmentReader.init(SegmentReader.java:55) [junit4]at org.apache.lucene.index.StandardDirectoryReader$1.doBody(StandardDirectoryReader.java:62) [junit4]at org.apache.lucene.index.SegmentInfos$FindSegmentsFile.run(SegmentInfos.java:639) [junit4]at
[jira] [Commented] (LUCENE-4150) Change version properties in branch_4x to 4.0-ALPHA-SNAPSHOT
[ https://issues.apache.org/jira/browse/LUCENE-4150?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13396116#comment-13396116 ] Steven Rowe commented on LUCENE-4150: - I agree with Hoss. I think 4.0-ALPHA and 4.0-BETA should be full-fledged releases in every way. Change version properties in branch_4x to 4.0-ALPHA-SNAPSHOT --- Key: LUCENE-4150 URL: https://issues.apache.org/jira/browse/LUCENE-4150 Project: Lucene - Java Issue Type: Task Components: general/build Affects Versions: 4.0 Reporter: Steven Rowe Priority: Minor The next release off branch_4x will be named 4.0-ALPHA, so the current version string should be 4.0-ALPHA-SNAPSHOT. (Similarly, after 4.0-ALPHA is released, the version string should be changed to 4.0-BETA-SNAPSHOT.) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4150) Change version properties in branch_4x to 4.0-ALPHA-SNAPSHOT
[ https://issues.apache.org/jira/browse/LUCENE-4150?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13396129#comment-13396129 ] Robert Muir commented on LUCENE-4150: - Everyone has the right to his own opinion: but for these alpha/betas to me they are snapshots with additional guarantees: and its better to get them out sooner than later so we can iteratively incorporate feedback. If some people want these to be typical lucene solr releases each that take months to finish, good for them, but I strongly disagree (and will not be convinced otherwise!). Fortunately its ok for us to disagree here, I only need to convince 2 other PMC members for us to get these things out there: its been way too long for this release. My advice: if you want these alpha/beta releases to be more solid, get your changes in now because next week I will spin RCs every 72 hours until we get something out. I imagine with enough hassling/bribing it won't be long until 2 other people decide its good to just get the alpha release out there and get people testing it instead of dragging things out. Change version properties in branch_4x to 4.0-ALPHA-SNAPSHOT --- Key: LUCENE-4150 URL: https://issues.apache.org/jira/browse/LUCENE-4150 Project: Lucene - Java Issue Type: Task Components: general/build Affects Versions: 4.0 Reporter: Steven Rowe Priority: Minor The next release off branch_4x will be named 4.0-ALPHA, so the current version string should be 4.0-ALPHA-SNAPSHOT. (Similarly, after 4.0-ALPHA is released, the version string should be changed to 4.0-BETA-SNAPSHOT.) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-3557) Avoid NPE for distributed request when shards.tolerant=true
Ryan McKinley created SOLR-3557: --- Summary: Avoid NPE for distributed request when shards.tolerant=true Key: SOLR-3557 URL: https://issues.apache.org/jira/browse/SOLR-3557 Project: Solr Issue Type: Improvement Components: search Reporter: Ryan McKinley Assignee: Ryan McKinley Priority: Minor Fix For: 4.0 If a shard fails to return and shards.tolerant=true, the faceting module will get a null pointer. We should avoid that NPE -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3557) Avoid NPE for distributed request when shards.tolerant=true
[ https://issues.apache.org/jira/browse/SOLR-3557?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ryan McKinley updated SOLR-3557: Attachment: SOLR-3557-tolerant-faceting.patch Avoid NPE for distributed request when shards.tolerant=true --- Key: SOLR-3557 URL: https://issues.apache.org/jira/browse/SOLR-3557 Project: Solr Issue Type: Improvement Components: search Reporter: Ryan McKinley Assignee: Ryan McKinley Priority: Minor Fix For: 4.0 Attachments: SOLR-3557-tolerant-faceting.patch If a shard fails to return and shards.tolerant=true, the faceting module will get a null pointer. We should avoid that NPE -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-3522) literal function can not be parsed
[ https://issues.apache.org/jira/browse/SOLR-3522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man resolved SOLR-3522. Resolution: Fixed Fix Version/s: (was: 3.6.1) 5.0 Committed revision 1351460. - trunk Committed revision 1351462. - 4x deferring the more complicated (and broken) query equalities test to SOLR-3548 literal function can not be parsed Key: SOLR-3522 URL: https://issues.apache.org/jira/browse/SOLR-3522 Project: Solr Issue Type: Bug Reporter: Hoss Man Assignee: Hoss Man Fix For: 4.0, 5.0 Attachments: SOLR-3522.patch attempting to use the literal function in the fl param causes a parse error... Example queries with functions that works fine... {noformat} http://localhost:8983/solr/collection1/select?q=*:*fl=foo:sum%284,5%29 http://localhost:8983/solr/collection1/select?fl=scoreq={!func}strdist%28%22foo%22,%22fo%22,edit%29 {noformat} Examples using literal function that fails... {noformat} http://localhost:8983/solr/collection1/select?q=*:*fl=foo:literal%28%22foo%22%29 http://localhost:8983/solr/collection1/select?fl=scoreq={!func}strdist%28%22foo%22,literal%28%22fo%22%29,edit%29 {noformat} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-4154) migrate luceneutil from apache extras into lucene/benchmark
Michael McCandless created LUCENE-4154: -- Summary: migrate luceneutil from apache extras into lucene/benchmark Key: LUCENE-4154 URL: https://issues.apache.org/jira/browse/LUCENE-4154 Project: Lucene - Java Issue Type: Improvement Reporter: Michael McCandless Assignee: Michael McCandless I think it'd make sense to move all the performance testing scripts from luceneutil at http://code.google.com/a/apache-extras.org/p/luceneutil under lucene/benchmark ... -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4132) IndexWriterConfig live settings
[ https://issues.apache.org/jira/browse/LUCENE-4132?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13396233#comment-13396233 ] Uwe Schindler commented on LUCENE-4132: --- bq. The 4.x commit also breaks the builds, looks like duplicate merges to same file. This was solved by cleaning workspace on Apache's Jenkins. Must have been a SVN problem. IndexWriterConfig live settings --- Key: LUCENE-4132 URL: https://issues.apache.org/jira/browse/LUCENE-4132 Project: Lucene - Java Issue Type: Improvement Reporter: Shai Erera Assignee: Shai Erera Priority: Minor Fix For: 4.0, 5.0 Attachments: LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch, LUCENE-4132.patch A while ago there was a discussion about making some IW settings live and I remember that RAM buffer size was one of them. Judging from IW code, I see that RAM buffer can be changed live as IW never caches it. However, I don't remember which other settings were decided to be live and I don't see any documentation in IW nor IWC for that. IW.getConfig mentions: {code} * bNOTE:/b some settings may be changed on the * returned {@link IndexWriterConfig}, and will take * effect in the current IndexWriter instance. See the * javadocs for the specific setters in {@link * IndexWriterConfig} for details. {code} But there's no text on e.g. IWC.setRAMBuffer mentioning that. I think that it'd be good if we make it easier for users to tell which of the settings are live ones. There are few possible ways to do it: * Introduce a custom @live.setting tag on the relevant IWC.set methods, and add special text for them in build.xml ** Or, drop the tag and just document it clearly. * Separate IWC to two interfaces, LiveConfig and OneTimeConfig (name proposals are welcome !), have IWC impl both, and introduce another IW.getLiveConfig which will return that interface, thereby clearly letting the user know which of the settings are live. It'd be good if IWC itself could only expose setXYZ methods for the live settings though. So perhaps, off the top of my head, we can do something like this: * Introduce a Config object, which is essentially what IWC is today, and pass it to IW. * IW will create a different object, IWC from that Config and IW.getConfig will return IWC. * IWC itself will only have setXYZ methods for the live settings. It adds another object, but user code doesn't change - it still creates a Config object when initializing IW, and need to handle a different type if it ever calls IW.getConfig. Maybe that's not such a bad idea? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3866) Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader
[ https://issues.apache.org/jira/browse/LUCENE-3866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13396237#comment-13396237 ] Uwe Schindler commented on LUCENE-3866: --- I checked CHANGES.txt and MIGRATE.txt, there are no explicit mentions of the datatypes. No need to change anything. But MIGRATE.txt is missing the documentation about how to get atomic leaves from a reader. We should do this in the followup issue, when we simplify some names. I will commit this patch tomorrow morning to trunk and 4.x. Make CompositeReader.getSequentialSubReaders() and the corresponding IndexReaderContext methods return unmodifiable ListR extends IndexReader --- Key: LUCENE-3866 URL: https://issues.apache.org/jira/browse/LUCENE-3866 Project: Lucene - Java Issue Type: Improvement Reporter: Uwe Schindler Assignee: Uwe Schindler Fix For: 4.0 Attachments: LUCENE-3866-step1.patch, LUCENE-3866-step2.patch, LUCENE-3866-step2.patch Since Lucene 2.9 we have the method getSequentialSubReader() returning IndexReader[]. Based on hardly-to-debug errors in user's code, Robert and me realized that returning an array from a public API is an anti-pattern. If the array is intended to be modifiable (like byte[] in terms,...), it is fine to use arrays in public APIs, but not, if the array must be protected from modification. As IndexReaders are 100% unmodifiable in trunk code (no deletions,...), the only possibility to corrumpt the reader is by modifying the array returned by getSequentialSubReaders(). We should prevent this. The same theoretically applies to FieldCache, too - but the party that is afraid of performance problems is too big to fight against that :( For getSequentialSubReaders there is no performance problem at all. The binary search of reader-ids inside BaseCompositeReader can still be done with the internal protected array, but public APIs should expose only a unmodifiable List. The same applies to leaves() and children() in IndexReaderContext. This change to list would also allow to make CompositeReader and CompositeReaderContext IterableR extends IndexReader, so some loops would look nice. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3367) Show Logging Events in Admin UI
[ https://issues.apache.org/jira/browse/SOLR-3367?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13396244#comment-13396244 ] Stefan Matheis (steffkes) commented on SOLR-3367: - Ryan, would you mind to have a look at my last comment? Show Logging Events in Admin UI --- Key: SOLR-3367 URL: https://issues.apache.org/jira/browse/SOLR-3367 Project: Solr Issue Type: New Feature Components: web gui Reporter: Ryan McKinley Assignee: Stefan Matheis (steffkes) Fix For: 4.0 Attachments: SOLR-3367.patch, SOLR-3367.patch, SOLR-3367.patch, SOLR-3367.patch, SOLR-3367.png We can show logging events in the Admin UI. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3529) clarify distnction between index query tables on analysis debug pages
[ https://issues.apache.org/jira/browse/SOLR-3529?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stefan Matheis (steffkes) updated SOLR-3529: Component/s: web gui clarify distnction between index query tables on analysis debug pages --- Key: SOLR-3529 URL: https://issues.apache.org/jira/browse/SOLR-3529 Project: Solr Issue Type: Improvement Components: web gui Reporter: Hoss Man Assignee: Stefan Matheis (steffkes) Fix For: 4.0 Attachments: long-side-by-side-below-fold.png, long-side-by-side.png, med-just-index.png, med-query-only.png, med-side-by-side-below-fold.png, med-side-by-side.png, short-side-by-side.png Working on the tutorial, i noticed that the analysis debug page is a little hard to read when specifying both index and query values * if the inputs are short, it's not entirely obvious that you are looking at two tables (especially compared to how the page looks using only the index textbox) * if the inputs are longer, the query table shifts down below the fold where users may not even be aware of it. Screenshots and ideas for improvement to follow -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-4155) Move hide ReaderSlice and BitSlice classes (and possibly others) to oal.index package; move ReaderUtil to oal.index
Uwe Schindler created LUCENE-4155: - Summary: Move hide ReaderSlice and BitSlice classes (and possibly others) to oal.index package; move ReaderUtil to oal.index Key: LUCENE-4155 URL: https://issues.apache.org/jira/browse/LUCENE-4155 Project: Lucene - Java Issue Type: Bug Components: core/index Reporter: Uwe Schindler Assignee: Uwe Schindler Fix For: 4.0, 5.0 Those are used solely by the index package and are very internal (just helper classes), so they should be hidden from the user. This can be done by adding the pkg-private to index package. ReaderUtil was cleaned up in LUCENE-3866, should stay public, but is in wrong package since Lucene 2.9. We should move it to oal.index package, too. Its name suggests that. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3535) Add block support for XMLLoader
[ https://issues.apache.org/jira/browse/SOLR-3535?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mikhail Khludnev updated SOLR-3535: --- Attachment: SOLR-3535.patch new attach highlights: * UpdReqProc.processAdd() become protected * FlattenerUpdateProcessorFactory has been introduced. * it transforms the given SolrInputDocument with nested subdocs into block of SIDs * i.e. it transforms AddUpdCmd to AddBlockUpdCmd * nested subdocs are placed as collection of SID fields. Hoss, excuse me. It's not really my point - we can later switch to getChildrenDocs(). it just seems to me easier for now. Let's polish it later. * you can see that Flattener is placed between Distributed and Log/Run. i.e. I addressed your guys point - make it compatible with distributed update magic. * forgot to cover tree levels of nesting, my falut. it's trivial transitive closure via recursion. I'll switch to iteration further. * XMLLoader supports field name=skillsdoc../docdoc../docdoc../docdoc../doc/field it's a nested docs. * there are tests, puzzling a little. test harness is harmed a little. Please have a look, and raise your concerns! btw it might be easier https://github.com/m-khl/solr-patches/commits/blockupdate Add block support for XMLLoader --- Key: SOLR-3535 URL: https://issues.apache.org/jira/browse/SOLR-3535 Project: Solr Issue Type: Sub-task Components: update Affects Versions: 4.1, 5.0 Reporter: Mikhail Khludnev Priority: Minor Attachments: SOLR-3535.patch, SOLR-3535.patch I'd like to add the following update xml message: add-block doc/doc doc/doc /add-block out of scope for now: * other update formats * update log support (NRT), should not be a big deal * overwrite feature support for block updates - it's more complicated, I'll tell you why Alt * wdyt about adding attribute to the current tag {pre}add block=true{pre} * or we can establish RunBlockUpdateProcessor which treat every add /add as a block. *Test is included!!* How you'd suggest to improve the patch? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-3367) Show Logging Events in Admin UI
[ https://issues.apache.org/jira/browse/SOLR-3367?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13396271#comment-13396271 ] Ryan McKinley commented on SOLR-3367: - ya, i see the difference. Obviously it is best if we can change the messages so they are reasonably short, but this is not realistic for all error messages. (yes, we should open an issue for this specific error, but it is not an easy fix) Whatever format we pick should be able to support short or long messages (i think) Perhaps overflow:hidden? or an option to toggle overflow:hidden? Show Logging Events in Admin UI --- Key: SOLR-3367 URL: https://issues.apache.org/jira/browse/SOLR-3367 Project: Solr Issue Type: New Feature Components: web gui Reporter: Ryan McKinley Assignee: Stefan Matheis (steffkes) Fix For: 4.0 Attachments: SOLR-3367.patch, SOLR-3367.patch, SOLR-3367.patch, SOLR-3367.patch, SOLR-3367.png We can show logging events in the Admin UI. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3553) Make solr Plugins (FieldType, Analyser etc) be inheritable
[ https://issues.apache.org/jira/browse/SOLR-3553?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Greg Bowyer updated SOLR-3553: -- Attachment: (was: SOLR-3553-Inheritable-schema-elements.patch) Make solr Plugins (FieldType, Analyser etc) be inheritable -- Key: SOLR-3553 URL: https://issues.apache.org/jira/browse/SOLR-3553 Project: Solr Issue Type: New Feature Affects Versions: 4.0 Reporter: Greg Bowyer Attachments: SOLR-3553-Inheritable-schema-elements.patch During defining an experimental schema for trialling all the combinations of scoring models for lucene 4 it struck me how much duplication is involved in the schema xml Using the current schema I am working with as an example, {code:xml} fieldType name=text class=solr.TextField positionIncrementGap=100 autoGeneratePhraseQueries=true analyzer type=index tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.TrimFilterFactory/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=0 catenateNumbers=1 catenateAll=0 splitOnCaseChange=1 splitOnNumerics=0 stemEnglishPossesive=0 preserveOriginal=0 / filter class=solr.ASCIIFoldingFilterFactory/ filter class=solr.StopFilterFactory ignoreCase=true words=stopwords.txt enablePositionIncrements=true / ... And so on {code} What I would like to be able to say is {code:xml} fieldType name=text_bm25 parent=text similarity class=solr.BM25SimilarityFactory float name=k11.2/float float name=b0.76/float /similarity /fieldType {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3553) Make solr Plugins (FieldType, Analyser etc) be inheritable
[ https://issues.apache.org/jira/browse/SOLR-3553?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Greg Bowyer updated SOLR-3553: -- Attachment: SOLR-3553-Inheritable-schema-elements.patch Make solr Plugins (FieldType, Analyser etc) be inheritable -- Key: SOLR-3553 URL: https://issues.apache.org/jira/browse/SOLR-3553 Project: Solr Issue Type: New Feature Affects Versions: 4.0 Reporter: Greg Bowyer Attachments: SOLR-3553-Inheritable-schema-elements.patch During defining an experimental schema for trialling all the combinations of scoring models for lucene 4 it struck me how much duplication is involved in the schema xml Using the current schema I am working with as an example, {code:xml} fieldType name=text class=solr.TextField positionIncrementGap=100 autoGeneratePhraseQueries=true analyzer type=index tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.TrimFilterFactory/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=0 catenateNumbers=1 catenateAll=0 splitOnCaseChange=1 splitOnNumerics=0 stemEnglishPossesive=0 preserveOriginal=0 / filter class=solr.ASCIIFoldingFilterFactory/ filter class=solr.StopFilterFactory ignoreCase=true words=stopwords.txt enablePositionIncrements=true / ... And so on {code} What I would like to be able to say is {code:xml} fieldType name=text_bm25 parent=text similarity class=solr.BM25SimilarityFactory float name=k11.2/float float name=b0.76/float /similarity /fieldType {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3553) Make solr Plugins (FieldType, Analyser etc) be inheritable
[ https://issues.apache.org/jira/browse/SOLR-3553?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Greg Bowyer updated SOLR-3553: -- Attachment: (was: SOLR-3553-Inheritable-schema-elements.patch) Make solr Plugins (FieldType, Analyser etc) be inheritable -- Key: SOLR-3553 URL: https://issues.apache.org/jira/browse/SOLR-3553 Project: Solr Issue Type: New Feature Affects Versions: 4.0 Reporter: Greg Bowyer Attachments: SOLR-3553-Inheritable-schema-elements.patch During defining an experimental schema for trialling all the combinations of scoring models for lucene 4 it struck me how much duplication is involved in the schema xml Using the current schema I am working with as an example, {code:xml} fieldType name=text class=solr.TextField positionIncrementGap=100 autoGeneratePhraseQueries=true analyzer type=index tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.TrimFilterFactory/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=0 catenateNumbers=1 catenateAll=0 splitOnCaseChange=1 splitOnNumerics=0 stemEnglishPossesive=0 preserveOriginal=0 / filter class=solr.ASCIIFoldingFilterFactory/ filter class=solr.StopFilterFactory ignoreCase=true words=stopwords.txt enablePositionIncrements=true / ... And so on {code} What I would like to be able to say is {code:xml} fieldType name=text_bm25 parent=text similarity class=solr.BM25SimilarityFactory float name=k11.2/float float name=b0.76/float /similarity /fieldType {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3553) Make solr Plugins (FieldType, Analyser etc) be inheritable
[ https://issues.apache.org/jira/browse/SOLR-3553?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Greg Bowyer updated SOLR-3553: -- Attachment: (was: SOLR-3553-Inheritable-schema-elements.patch) Make solr Plugins (FieldType, Analyser etc) be inheritable -- Key: SOLR-3553 URL: https://issues.apache.org/jira/browse/SOLR-3553 Project: Solr Issue Type: New Feature Affects Versions: 4.0 Reporter: Greg Bowyer Attachments: SOLR-3553-Inheritable-schema-elements.patch During defining an experimental schema for trialling all the combinations of scoring models for lucene 4 it struck me how much duplication is involved in the schema xml Using the current schema I am working with as an example, {code:xml} fieldType name=text class=solr.TextField positionIncrementGap=100 autoGeneratePhraseQueries=true analyzer type=index tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.TrimFilterFactory/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=0 catenateNumbers=1 catenateAll=0 splitOnCaseChange=1 splitOnNumerics=0 stemEnglishPossesive=0 preserveOriginal=0 / filter class=solr.ASCIIFoldingFilterFactory/ filter class=solr.StopFilterFactory ignoreCase=true words=stopwords.txt enablePositionIncrements=true / ... And so on {code} What I would like to be able to say is {code:xml} fieldType name=text_bm25 parent=text similarity class=solr.BM25SimilarityFactory float name=k11.2/float float name=b0.76/float /similarity /fieldType {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-3892) Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.)
[ https://issues.apache.org/jira/browse/LUCENE-3892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13396325#comment-13396325 ] Michael McCandless commented on LUCENE-3892: On the For patch ... we shouldn't encode/decode numInts right? It's always 128? Up above, in ForFactory, when we readInt() to get numBytes ... it seems like we could stuff the header numBits into that same int and save checking that in FORUtil.decompress I think there are a few possible ideas to explore to get faster PFor/For performance: * Get more direct access to the file as an int[]; eg MMapDir could expose an IntBuffer from its ByteBuffer (saving the initial copy into byte[] that we now do). Or maybe we add IndexInput.readInts(int[]) and dir impl can optimize how that's done (MMapDir could use Unsafe.copyBytes... except for little endian architectures ... we'd probably have to have separate specialized decoder rather than letting Int/ByteBuffer do the byte swapping). This would require the whole file stays aligned w/ int (eg the header must be 0 mod 4). * Copy/share how oal.packed works, i.e. being able to waste a bit to have faster decode (eg storing the 7 bit case as byte[], wasting 1 bit for each value). * Skipping: can we partially decode a block? EG if we are skipping and we know we only want values after the 80th one, then we shouldn't decode those first 80... * Since doc/freq are aligned, when we store pointers to a given spot, eg in the terms dict or in skip data, we should only store the offset once (today we store it twice). * Alternatively, maybe we should only save skip data on doc/freq block boundaries (prox would still need skip-within-block). * Maybe we should store doc frq blocks interleaved in a single file (since they are aligned) and then skip would skip to the start of a doc/frq block pair. Other ideas...? Add a useful intblock postings format (eg, FOR, PFOR, PFORDelta, Simple9/16/64, etc.) - Key: LUCENE-3892 URL: https://issues.apache.org/jira/browse/LUCENE-3892 Project: Lucene - Java Issue Type: Improvement Reporter: Michael McCandless Labels: gsoc2012, lucene-gsoc-12 Fix For: 4.1 Attachments: LUCENE-3892_for.patch, LUCENE-3892_pfor.patch, LUCENE-3892_pfor.patch, LUCENE-3892_pfor.patch, LUCENE-3892_settings.patch, LUCENE-3892_settings.patch On the flex branch we explored a number of possible intblock encodings, but for whatever reason never brought them to completion. There are still a number of issues opened with patches in different states. Initial results (based on prototype) were excellent (see http://blog.mikemccandless.com/2010/08/lucene-performance-with-pfordelta-codec.html ). I think this would make a good GSoC project. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3548) some queries produced by built in QParsers are not .equals() to themselves
[ https://issues.apache.org/jira/browse/SOLR-3548?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man updated SOLR-3548: --- Attachment: SOLR-3548.patch it took a lot longer then i though, but here is a patch that adds QueryUtils checking against (at least one of) the Query objects produced by every default QParser and ValueSourceParser. It includes a future proofing testCoverage that sets a bit informing an AfterClass method to assert that all of the default parsers were tested so we don't risk this probably again the next time someone adds a new parsers. Currently 4 methods are failing, indicating the following problems... * strdist func - identical query strings don't produce equals() queries * join qparser - clone w/diff boost still has equals hashCode * bbox qparser - clone w/diff boost still has equals hashCode * geofilt qparser - clone w/diff boost still has equals hashCode the hashCode equality isn't the end of the world, but it suggests a really poor hashCode impl (that evidently doesn't call super since Query.hashCode already handles the boost) some queries produced by built in QParsers are not .equals() to themselves -- Key: SOLR-3548 URL: https://issues.apache.org/jira/browse/SOLR-3548 Project: Solr Issue Type: Bug Reporter: Hoss Man Assignee: Hoss Man Priority: Blocker Fix For: 4.0, 3.6.1 Attachments: SOLR-3548.patch, SOLR-3548.patch working on SOLR-3522 i discovered that in some cases Solr is producing queries that aren't equals to themselves, so they can't be cached. So far the only observed problem is in func strdist, but i want to make sure we have an exhaustive test of this in case there are others -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3548) some queries produced by builtin QParsers do not satisfy QueryUtils checks
[ https://issues.apache.org/jira/browse/SOLR-3548?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man updated SOLR-3548: --- Summary: some queries produced by builtin QParsers do not satisfy QueryUtils checks (was: some queries produced by built in QParsers are not .equals() to themselves) some queries produced by builtin QParsers do not satisfy QueryUtils checks -- Key: SOLR-3548 URL: https://issues.apache.org/jira/browse/SOLR-3548 Project: Solr Issue Type: Bug Reporter: Hoss Man Assignee: Hoss Man Priority: Blocker Fix For: 4.0, 3.6.1 Attachments: SOLR-3548.patch, SOLR-3548.patch working on SOLR-3522 i discovered that in some cases Solr is producing queries that aren't equals to themselves, so they can't be cached. So far the only observed problem is in func strdist, but i want to make sure we have an exhaustive test of this in case there are others -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-trunk-Windows-Java7-64 - Build # 346 - Failure!
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-trunk-Windows-Java7-64/346/ 1 tests failed. REGRESSION: org.apache.solr.spelling.suggest.SuggesterTSTTest.testReload Error Message: Exception during query Stack Trace: java.lang.RuntimeException: Exception during query at __randomizedtesting.SeedInfo.seed([2C36619596515F9B:EBC619965C12A789]:0) at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:459) at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:426) at org.apache.solr.spelling.suggest.SuggesterTest.testReload(SuggesterTest.java:91) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:601) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1969) at com.carrotsearch.randomizedtesting.RandomizedRunner.access$1100(RandomizedRunner.java:132) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:814) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:875) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:889) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:32) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleReportUncaughtExceptions$1.evaluate(TestRuleReportUncaughtExceptions.java:68) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:821) at com.carrotsearch.randomizedtesting.RandomizedRunner.access$700(RandomizedRunner.java:132) at com.carrotsearch.randomizedtesting.RandomizedRunner$3$1.run(RandomizedRunner.java:669) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:695) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:734) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:745) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleReportUncaughtExceptions$1.evaluate(TestRuleReportUncaughtExceptions.java:68) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at org.apache.lucene.util.TestRuleIcuHack$1.evaluate(TestRuleIcuHack.java:51) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55) at org.apache.lucene.util.TestRuleNoInstanceHooksOverrides$1.evaluate(TestRuleNoInstanceHooksOverrides.java:53) at org.apache.lucene.util.TestRuleNoStaticHooksShadowing$1.evaluate(TestRuleNoStaticHooksShadowing.java:52) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:36) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:56) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSuite(RandomizedRunner.java:605) at com.carrotsearch.randomizedtesting.RandomizedRunner.access$400(RandomizedRunner.java:132) at com.carrotsearch.randomizedtesting.RandomizedRunner$2.run(RandomizedRunner.java:551) Caused by: java.lang.RuntimeException: REQUEST FAILED: xpath=//lst[@name='spellcheck']/lst[@name='suggestions']/lst[@name='ac']/int[@name='numFound'][.='2'] xml response was: ?xml version=1.0 encoding=UTF-8? response lst name=responseHeaderint name=status0/intint name=QTime0/int/lstlst name=spellchecklst name=suggestions//lst /response request was:q=acspellcheck.count=2qt=/suggest_tstspellcheck.onlyMorePopular=true at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:452)
[jira] [Commented] (LUCENE-4145) Unhandled exception from test framework (in json parsing of test output files?)
[ https://issues.apache.org/jira/browse/LUCENE-4145?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13396432#comment-13396432 ] Hoss Man commented on LUCENE-4145: -- Thanks Dawid ... i would totally expect/respect -Dtests.iters=$HUGE to be limited by the amount of Ram in use, it was the odd except on console followed by the aparent infinite hang that concerend me the most (would hate to see that happen in a jenkins build) Unhandled exception from test framework (in json parsing of test output files?) - Key: LUCENE-4145 URL: https://issues.apache.org/jira/browse/LUCENE-4145 Project: Lucene - Java Issue Type: Improvement Reporter: Hoss Man Assignee: Dawid Weiss Working on SOLR-3267 i got a weird exception printed to the junit output... {noformat} [junit4] Unhandled exception in thread: Thread[pumper-events,5,main] [junit4] com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.JsonParseException: No such reference: id#org.apache.solr.search.TestSort[3] ... {noformat} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-3548) some queries produced by builtin QParsers do not satisfy QueryUtils checks
[ https://issues.apache.org/jira/browse/SOLR-3548?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man updated SOLR-3548: --- Attachment: SOLR-3548.patch Fixed JoinQuery and SpatialDistanceQuery classes to consult super.equals() and super.hashCode() in their corrisponding methods, and added completely new equals/hashCode impls to JaroWinklerDistance, NGramDistance, and LevensteinDistance (which never had them before aparently). This gets all the new tests passing some queries produced by builtin QParsers do not satisfy QueryUtils checks -- Key: SOLR-3548 URL: https://issues.apache.org/jira/browse/SOLR-3548 Project: Solr Issue Type: Bug Reporter: Hoss Man Assignee: Hoss Man Priority: Blocker Fix For: 4.0, 3.6.1 Attachments: SOLR-3548.patch, SOLR-3548.patch, SOLR-3548.patch working on SOLR-3522 i discovered that in some cases Solr is producing queries that aren't equals to themselves, so they can't be cached. So far the only observed problem is in func strdist, but i want to make sure we have an exhaustive test of this in case there are others -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Assigned] (LUCENE-4146) -Dtests.iters combined with -Dtestmethod never fails?
[ https://issues.apache.org/jira/browse/LUCENE-4146?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man reassigned LUCENE-4146: Assignee: Hoss Man bq. I realize this is confusing but this really doesn't stem from my code and is a non-fixable JUnit API (because of how many things depend on it I doubt it'll ever be corrected). I totally understand ... the missing piece for me was not realizing that the filtering happens after the paramaterizing so a filter on testFoo won't match testFoo[0] etc... bq. Other than that all your suggestions are fine by me. I'll fix test-help and make the change to our testmethod alias, but i'm not sure about this one... bq. change the code path that reads the system properties used to pick the tests to fail if this combination (tests.iters + a non-globed test method) is used (rather then silently lying and saying (0) tests passed) ...unless i'm missing something, lucene's build xml files test base class don't have anything to do with parsing these system properties correct? should i file this as a feature request in the randomizedtesting tracker? -Dtests.iters combined with -Dtestmethod never fails? - Key: LUCENE-4146 URL: https://issues.apache.org/jira/browse/LUCENE-4146 Project: Lucene - Java Issue Type: Improvement Reporter: Hoss Man Assignee: Hoss Man Attachments: LUCENE-4146.fail.patch, TEST-org.apache.lucene.TestSearch.iters-no-fail.xml, TEST-org.apache.lucene.TestSearch.no-iters-fail.xml a test that is hardcoded to fail will report succes if you run it with -Dtests.iters -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-4.x-Windows-Java6-64 - Build # 115 - Failure!
Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows-Java6-64/115/ All tests passed Build Log: [...truncated 4427 lines...] [junit4] ERROR: Forked JVM execution exception: 0, execution line: C:\Program Files\Java\jdk1.6.0_32\jre\bin\java.exe -Dtests.prefix=tests -Dtests.seed=98EF8DA20B71EC8 -Xmx512M -Dtests.iters= -Dtests.verbose=false -Dtests.infostream=false -Dtests.lockdir=C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build -Dtests.codec=random -Dtests.postingsformat=random -Dtests.locale=random -Dtests.timezone=random -Dtests.directory=random -Dtests.linedocsfile=europarl.lines.txt.gz -Dtests.luceneMatchVersion=4.0 -Dtests.cleanthreads=perMethod -Djava.util.logging.config.file=/dev/null -Dtests.nightly=false -Dtests.weekly=false -Dtests.slow=false -Dtests.asserts.gracious=false -Dtests.multiplier=1 -DtempDir=. -Dlucene.version=4.0-SNAPSHOT -Djetty.testMode=1 -Djetty.insecurerandom=1 -Dsolr.directoryFactory=org.apache.solr.core.MockDirectoryFactory -classpath C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\test-framework\classes\java;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\memory\lucene-memory-4.0-SNAPSHOT.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\highlighter\lucene-highlighter-4.0-SNAPSHOT.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\analysis\common\lucene-analyzers-common-4.0-SNAPSHOT.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\queryparser\lucene-queryparser-4.0-SNAPSHOT.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\facet\lucene-facet-4.0-SNAPSHOT.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\analysis\icu\lib\icu4j-4.8.1.1.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\core\classes\java;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\benchmark\lib\commons-compress-1.2.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\benchmark\lib\xercesImpl-2.9.1.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\test-framework\lib\junit-4.10.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\test-framework\lib\randomizedtesting-runner-1.5.0.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\benchmark\classes\java;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\benchmark\classes\test;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-launcher.jar;C:\Users\JenkinsSlave\.ant\lib\ivy-2.2.0.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-antlr.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache-bcel.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache-bsf.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache-log4j.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache-oro.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache-regexp.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache-resolver.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache-xalan2.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-commons-logging.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-commons-net.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-jai.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-javamail.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-jdepend.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-jmf.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-jsch.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-junit.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-junit4.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-netrexx.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-swing.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-testutil.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant.jar;C:\Program Files\Java\jdk1.6.0_32\lib\tools.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\test-framework\lib\junit4-ant-1.5.0.jar -ea:org.apache.lucene... -ea:org.apache.solr... com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe -flush -eventsfile C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\benchmark\test\junit4-J0-02233930414c9c7bd.events @C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\benchmark\test\junit4-J0-02233930414c9c7bd.suites BUILD FAILED C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\build.xml:29: The following error occurred while executing this line: C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build.xml:456: The following error occurred while executing this line: C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\common-build.xml:1417: The following error occurred while executing this line: C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\module-build.xml:62: The following error occurred while executing this line: C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\common-build.xml:1019: The following error occurred while executing this line: C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\common-build.xml:745:
RE: [JENKINS] Lucene-Solr-4.x-Windows-Java6-64 - Build # 115 - Failure!
Hi, This one was hanging in windows at benchmark tests (1 runner) for several hours with 100% CPU, 2 cores. Unfortunately I was unable to get stack trace. On Windows with Jenkins running as a system service this seems undoable. Although I was logged in as the same user name on the machine's GUI, running jstack pid only produced Access Denied. Uwe - Uwe Schindler H.-H.-Meier-Allee 63, D-28213 Bremen http://www.thetaphi.de eMail: u...@thetaphi.de -Original Message- From: Policeman Jenkins Server [mailto:jenk...@sd-datasolutions.de] Sent: Tuesday, June 19, 2012 7:41 AM To: dev@lucene.apache.org Subject: [JENKINS] Lucene-Solr-4.x-Windows-Java6-64 - Build # 115 - Failure! Build: http://jenkins.sd-datasolutions.de/job/Lucene-Solr-4.x-Windows-Java6- 64/115/ All tests passed Build Log: [...truncated 4427 lines...] [junit4] ERROR: Forked JVM execution exception: 0, execution line: C:\Program Files\Java\jdk1.6.0_32\jre\bin\java.exe -Dtests.prefix=tests - Dtests.seed=98EF8DA20B71EC8 -Xmx512M -Dtests.iters= -Dtests.verbose=false -Dtests.infostream=false -Dtests.lockdir=C:\Jenkins\workspace\Lucene-Solr-4.x- Windows-Java6-64\lucene\build -Dtests.codec=random - Dtests.postingsformat=random -Dtests.locale=random - Dtests.timezone=random -Dtests.directory=random - Dtests.linedocsfile=europarl.lines.txt.gz -Dtests.luceneMatchVersion=4.0 - Dtests.cleanthreads=perMethod -Djava.util.logging.config.file=/dev/null - Dtests.nightly=false -Dtests.weekly=false -Dtests.slow=false - Dtests.asserts.gracious=false -Dtests.multiplier=1 -DtempDir=. - Dlucene.version=4.0-SNAPSHOT -Djetty.testMode=1 -Djetty.insecurerandom=1 -Dsolr.directoryFactory=org.apache.solr.core.MockDirectoryFactory -classpath C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\build\test- framework\classes\java;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6- 64\lucene\build\memory\lucene-memory-4.0- SNAPSHOT.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6- 64\lucene\build\highlighter\lucene-highlighter-4.0- SNAPSHOT.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6- 64\lucene\build\analysis\common\lucene-analyzers-common-4.0- SNAPSHOT.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6- 64\lucene\build\queryparser\lucene-queryparser-4.0- SNAPSHOT.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6- 64\lucene\build\facet\lucene-facet-4.0- SNAPSHOT.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6- 64\lucene\analysis\icu\lib\icu4j-4.8.1.1.jar;C:\Jenkins\workspace\Lucene-Solr- 4.x-Windows-Java6- 64\lucene\build\core\classes\java;C:\Jenkins\workspace\Lucene-Solr-4.x- Windows-Java6-64\lucene\benchmark\lib\commons-compress- 1.2.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6- 64\lucene\benchmark\lib\xercesImpl-2.9.1.jar;C:\Jenkins\workspace\Lucene- Solr-4.x-Windows-Java6-64\lucene\test-framework\lib\junit- 4.10.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-64\lucene\test- framework\lib\randomizedtesting-runner- 1.5.0.jar;C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6- 64\lucene\build\benchmark\classes\java;C:\Jenkins\workspace\Lucene-Solr- 4.x-Windows-Java6- 64\lucene\build\benchmark\classes\test;C:\Jenkins\tools\apache-ant- 1.8.2\lib\ant-launcher.jar;C:\Users\JenkinsSlave\.ant\lib\ivy- 2.2.0.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- antlr.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache- bcel.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache- bsf.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache- log4j.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache- oro.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache- regexp.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache- resolver.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-apache- xalan2.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-commons- logging.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant-commons- net.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- jai.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- javamail.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- jdepend.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- jmf.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- jsch.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- junit.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- junit4.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- netrexx.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- swing.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant- testutil.jar;C:\Jenkins\tools\apache-ant-1.8.2\lib\ant.jar;C:\Program Files\Java\jdk1.6.0_32\lib\tools.jar;C:\Jenkins\workspace\Lucene-Solr-4.x- Windows-Java6-64\lucene\test-framework\lib\junit4-ant-1.5.0.jar - ea:org.apache.lucene... -ea:org.apache.solr... com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe -flush -eventsfile C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6- 64\lucene\build\benchmark\test\junit4-J0-02233930414c9c7bd.events @C:\Jenkins\workspace\Lucene-Solr-4.x-Windows-Java6-