[jira] [Commented] (SOLR-7789) Introduce a ConfigSet management API
[ https://issues.apache.org/jira/browse/SOLR-7789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14716099#comment-14716099 ] ASF subversion and git services commented on SOLR-7789: --- Commit 1698079 from gcha...@apache.org in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1698079 ] SOLR-7789: Introduce a ConfigSet management API > Introduce a ConfigSet management API > > > Key: SOLR-7789 > URL: https://issues.apache.org/jira/browse/SOLR-7789 > Project: Solr > Issue Type: New Feature >Reporter: Gregory Chanan >Assignee: Gregory Chanan > Attachments: SOLR-7789.patch, SOLR-7789.patch, SOLR-7789.patch, > SOLR-7789.patch, SOLR-7789.patch > > > SOLR-5955 describes a feature to automatically create a ConfigSet, based on > another one, from a collection API call (i.e. one step collection creation). > Discussion there yielded SOLR-7742, Immutable ConfigSet support. To close > the loop, we need support for a ConfigSet management API. > The simplest ConfigSet API could have one operation: > create a new config set, based on an existing one, possible modifying the > ConfigSet properties. Note you need to be able to modify the ConfigSet > properties at creation time because otherwise Immutable could not be changed. > Another logical operation to support is ConfigSet deletion; that may be more > complicated to implement than creation because you need to handle the case > where a collection is already using the configuration. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7789) Introduce a ConfigSet management API
[ https://issues.apache.org/jira/browse/SOLR-7789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14716088#comment-14716088 ] ASF subversion and git services commented on SOLR-7789: --- Commit 1698072 from gcha...@apache.org in branch 'dev/trunk' [ https://svn.apache.org/r1698072 ] SOLR-7789: fix jira number in CHANGES.txt > Introduce a ConfigSet management API > > > Key: SOLR-7789 > URL: https://issues.apache.org/jira/browse/SOLR-7789 > Project: Solr > Issue Type: New Feature >Reporter: Gregory Chanan >Assignee: Gregory Chanan > Attachments: SOLR-7789.patch, SOLR-7789.patch, SOLR-7789.patch, > SOLR-7789.patch, SOLR-7789.patch > > > SOLR-5955 describes a feature to automatically create a ConfigSet, based on > another one, from a collection API call (i.e. one step collection creation). > Discussion there yielded SOLR-7742, Immutable ConfigSet support. To close > the loop, we need support for a ConfigSet management API. > The simplest ConfigSet API could have one operation: > create a new config set, based on an existing one, possible modifying the > ConfigSet properties. Note you need to be able to modify the ConfigSet > properties at creation time because otherwise Immutable could not be changed. > Another logical operation to support is ConfigSet deletion; that may be more > complicated to implement than creation because you need to handle the case > where a collection is already using the configuration. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-NightlyTests-5.x - Build # 939 - Still Failing
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-5.x/939/ 1 tests failed. REGRESSION: org.apache.lucene.codecs.lucene46.TestLucene46SegmentInfoFormat.testRandomExceptions Error Message: hit unexpected FileNotFoundException: file=_3g.si Stack Trace: java.lang.AssertionError: hit unexpected FileNotFoundException: file=_3g.si at __randomizedtesting.SeedInfo.seed([1A10EDD0B2491A33:723F83142C474393]:0) at org.apache.lucene.index.IndexFileDeleter.deleteFile(IndexFileDeleter.java:753) at org.apache.lucene.index.IndexFileDeleter.deletePendingFiles(IndexFileDeleter.java:530) at org.apache.lucene.index.IndexFileDeleter.deleteNewFiles(IndexFileDeleter.java:733) at org.apache.lucene.index.IndexWriter.deleteNewFiles(IndexWriter.java:4700) at org.apache.lucene.index.IndexWriter.mergeMiddle(IndexWriter.java:4218) at org.apache.lucene.index.IndexWriter.merge(IndexWriter.java:3664) at org.apache.lucene.index.SerialMergeScheduler.merge(SerialMergeScheduler.java:40) at org.apache.lucene.index.IndexWriter.maybeMerge(IndexWriter.java:1929) at org.apache.lucene.index.IndexWriter.doAfterSegmentFlushed(IndexWriter.java:4731) at org.apache.lucene.index.DocumentsWriter$MergePendingEvent.process(DocumentsWriter.java:695) at org.apache.lucene.index.IndexWriter.processEvents(IndexWriter.java:4757) at org.apache.lucene.index.IndexWriter.processEvents(IndexWriter.java:4748) at org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:1476) at org.apache.lucene.index.IndexWriter.addDocument(IndexWriter.java:1254) at org.apache.lucene.index.BaseIndexFileFormatTestCase.testRandomExceptions(BaseIndexFileFormatTestCase.java:429) at org.apache.lucene.index.BaseSegmentInfoFormatTestCase.testRandomExceptions(BaseSegmentInfoFormatTestCase.java:48) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.
[jira] [Commented] (LUCENE-6762) CheckIndex cannot "fix" indexes that have individual segments with missing or corrupt .si files because sanity checks will fail trying to read the index initially.
[ https://issues.apache.org/jira/browse/LUCENE-6762?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715983#comment-14715983 ] Mike Drob commented on LUCENE-6762: --- I've almost got a patch ready for this, probably will finish it up tomorrow. > CheckIndex cannot "fix" indexes that have individual segments with missing or > corrupt .si files because sanity checks will fail trying to read the index > initially. > --- > > Key: LUCENE-6762 > URL: https://issues.apache.org/jira/browse/LUCENE-6762 > Project: Lucene - Core > Issue Type: Improvement >Reporter: Mark Miller >Priority: Minor > > Seems like we should still be able to partially recover by dropping these > segments with CheckIndex. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7789) Introduce a ConfigSet management API
[ https://issues.apache.org/jira/browse/SOLR-7789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715938#comment-14715938 ] ASF subversion and git services commented on SOLR-7789: --- Commit 1698043 from gcha...@apache.org in branch 'dev/trunk' [ https://svn.apache.org/r1698043 ] SOLR-7789: Introduce a ConfigSet management API > Introduce a ConfigSet management API > > > Key: SOLR-7789 > URL: https://issues.apache.org/jira/browse/SOLR-7789 > Project: Solr > Issue Type: New Feature >Reporter: Gregory Chanan >Assignee: Gregory Chanan > Attachments: SOLR-7789.patch, SOLR-7789.patch, SOLR-7789.patch, > SOLR-7789.patch, SOLR-7789.patch > > > SOLR-5955 describes a feature to automatically create a ConfigSet, based on > another one, from a collection API call (i.e. one step collection creation). > Discussion there yielded SOLR-7742, Immutable ConfigSet support. To close > the loop, we need support for a ConfigSet management API. > The simplest ConfigSet API could have one operation: > create a new config set, based on an existing one, possible modifying the > ConfigSet properties. Note you need to be able to modify the ConfigSet > properties at creation time because otherwise Immutable could not be changed. > Another logical operation to support is ConfigSet deletion; that may be more > complicated to implement than creation because you need to handle the case > where a collection is already using the configuration. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-5.x-Linux (32bit/jdk1.8.0_60) - Build # 13735 - Failure!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Linux/13735/ Java: 32bit/jdk1.8.0_60 -client -XX:+UseParallelGC 1 tests failed. FAILED: junit.framework.TestSuite.org.apache.solr.cloud.HttpPartitionTest Error Message: ObjectTracker found 1 object(s) that were not released!!! [TransactionLog] Stack Trace: java.lang.AssertionError: ObjectTracker found 1 object(s) that were not released!!! [TransactionLog] at __randomizedtesting.SeedInfo.seed([451FCED23C6C9A08]:0) at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.assertTrue(Assert.java:43) at org.junit.Assert.assertNull(Assert.java:551) at org.apache.solr.SolrTestCaseJ4.afterClass(SolrTestCaseJ4.java:236) at sun.reflect.GeneratedMethodAccessor35.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:799) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at java.lang.Thread.run(Thread.java:745) Build Log: [...truncated 10427 lines...] [junit4] Suite: org.apache.solr.cloud.HttpPartitionTest [junit4] 2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-5.x-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.HttpPartitionTest_451FCED23C6C9A08-001/init-core-data-001 [junit4] 2> 717648 INFO (SUITE-HttpPartitionTest-seed#[451FCED23C6C9A08]-worker) [] o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /_lek/qm [junit4] 2> 717650 INFO (TEST-HttpPartitionTest.test-seed#[451FCED23C6C9A08]) [] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER [junit4] 2> 717650 INFO (Thread-2181) [] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0 [junit4] 2> 717651 INFO (Thread-2181) [] o.a.s.c.ZkTestServer Starting server [junit4] 2> 717751 INFO (TEST-HttpPartitionTest.test-seed#[451FCED23C6C9A08]) [] o.a.s.c.ZkTestServer start zk server on port:34917 [junit4] 2> 717751 INFO (TEST-HttpPartitionTest.test-seed#[451FCED23C6C9A08]) [] o.a.s.c.c.SolrZkClient Using default ZkCredentialsProvider [junit4] 2> 717751 INFO (TEST-HttpPartitionTest.test-seed#[451FCED23C6C9A08]) [] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper [junit4] 2> 717753 INFO (zkCallback-1484-thread-1) [] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@2deaf2 name:ZooKeeperConnection Watcher:127.0.0.1:34917 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None [junit4] 2> 717753 INFO (TEST-HttpPartitionTest.test-seed#[451FCED23C6C9A08]) [] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper [junit4] 2> 717754 INFO (TEST-HttpPartitionTest.test-seed#[451FCED23C6C9A08]) [] o.a.s.c.c.SolrZkClient Using default ZkACLProvider [junit4] 2> 717754 INFO (TEST-HttpPartitionTest.test-seed#[451FCED23C6C9A08]) [] o.a.s.c.c.SolrZkClient makePath: /solr [junit4] 2> 717756 INF
[jira] [Commented] (SOLR-7746) Ping requests stopped working with distrib=true in Solr 5.2.1
[ https://issues.apache.org/jira/browse/SOLR-7746?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715882#comment-14715882 ] Gregory Chanan commented on SOLR-7746: -- I compared what happens in 5.1 vs your version of trunk for the commands listed in the description: 5.1: {noformat} curl 'http://localhost:8889/solr/test2/admin/ping?wt=json&distrib=true&indent=true' { "responseHeader":{ "status":0, "QTime":26, "params":{ "df":"text", "echoParams":"all", "indent":"true", "q":"solrpingquery", "distrib":"true", "wt":"json", "preferLocalShards":"false", "rows":"10"}}, "status":"OK"} {noformat} Trunk: {noformat} curl 'http://localhost:8885/solr/test2/admin/ping?wt=json&distrib=true&indent=true' { "responseHeader":{ "status":0, "QTime":28, "params":{ "q":"{!lucene}*:*", "distrib":"true", "df":"text", "preferLocalShards":"false", "indent":"true", "echoParams":"all", "rows":"10", "wt":"json"}}, "status":"OK"} {noformat} Looks good, the change in the q parameter looks like it's that way because the solrconfig.xml being used is different. > Ping requests stopped working with distrib=true in Solr 5.2.1 > - > > Key: SOLR-7746 > URL: https://issues.apache.org/jira/browse/SOLR-7746 > Project: Solr > Issue Type: Bug > Components: SolrCloud >Affects Versions: 5.2.1 >Reporter: Alexey Serba > Attachments: SOLR-7746.patch, SOLR-7746.patch, SOLR-7746.patch, > SOLR-7746.patch > > > {noformat:title="steps to reproduce"} > # start 1 node SolrCloud cluster > sh> ./bin/solr -c -p > # create a test collection (we won’t use it, but I just want to it to load > solr configs to Zk) > ./bin/solr create_collection -c test -d sample_techproducts_configs -p > # create another test collection with 2 shards > curl > 'http://localhost:/solr/admin/collections?action=CREATE&name=test2&numShards=2&replicationFactor=1&maxShardsPerNode=2&collection.configName=test' > # try distrib ping request > curl > 'http://localhost:/solr/test2/admin/ping?wt=json&distrib=true&indent=true' > ... > "error":{ > "msg":"Ping query caused exception: Error from server at > http://192.168.59.3:/solr/test2_shard2_replica1: Cannot execute the > PingRequestHandler recursively" > ... > {noformat} > {noformat:title="Exception"} > 2116962 [qtp599601600-13] ERROR org.apache.solr.core.SolrCore [test2 shard2 > core_node1 test2_shard2_replica1] – org.apache.solr.common.SolrException: > Cannot execute the PingRequestHandler recursively > at > org.apache.solr.handler.PingRequestHandler.handlePing(PingRequestHandler.java:246) > at > org.apache.solr.handler.PingRequestHandler.handleRequestBody(PingRequestHandler.java:211) > at > org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:143) > at org.apache.solr.core.SolrCore.execute(SolrCore.java:2064) > at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:654) > at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:450) > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:227) > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:196) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7795) Fold Interval Faceting into Range Faceting
[ https://issues.apache.org/jira/browse/SOLR-7795?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Tomás Fernández Löbbe updated SOLR-7795: Attachment: SOLR-7795.patch Added unit test > Fold Interval Faceting into Range Faceting > -- > > Key: SOLR-7795 > URL: https://issues.apache.org/jira/browse/SOLR-7795 > Project: Solr > Issue Type: Task >Reporter: Tomás Fernández Löbbe > Fix For: Trunk, 5.4 > > Attachments: SOLR-7795.patch, SOLR-7795.patch, SOLR-7795.patch, > SOLR-7795.patch > > > Now that range faceting supports a "filter" and a "dv" method, and that > interval faceting is supported on fields with {{docValues=false}}, I think we > should make it so that interval faceting is just a different way of > specifying ranges in range faceting, allowing users to indicate specific > ranges. > I propose we use the same syntax for intervals, but under the "range" > parameter family: > {noformat} > facet.range=price& > f.price.facet.range.set=[0,10]& > f.price.facet.range.set=(10,100] > {noformat} > The counts for those ranges would come in the response also inside of the > "range_facets" section. I'm not sure if it's better to include the ranges in > the "counts" section, or in a different section (intervals?sets?buckets?). > I'm open to suggestions. > {code} > "facet_ranges":{ > "price":{ > "counts":[ > "[0,10]",3, > "(10,100]",2] >} > } > {code} > or… > {code} > "facet_ranges":{ > "price":{ > "intervals":[ > "[0,10]",3, > "(10,100]",2] >} > } > {code} > We should support people specifying both things on the same field. > Once this is done, "interval faceting" could be deprecated, as all it's > functionality is now possible through range queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-Tests-trunk-Java8 - Build # 320 - Failure
Build: https://builds.apache.org/job/Lucene-Solr-Tests-trunk-Java8/320/ 1 tests failed. REGRESSION: org.apache.solr.cloud.BasicDistributedZkTest.test Error Message: commitWithin did not work on node: http://127.0.0.1:40118/vx_nn/fj/collection1 expected:<68> but was:<67> Stack Trace: java.lang.AssertionError: commitWithin did not work on node: http://127.0.0.1:40118/vx_nn/fj/collection1 expected:<68> but was:<67> at __randomizedtesting.SeedInfo.seed([19C9AF29BF843D27:919D90F3117850DF]:0) at org.junit.Assert.fail(Assert.java:93) at org.junit.Assert.failNotEquals(Assert.java:647) at org.junit.Assert.assertEquals(Assert.java:128) at org.junit.Assert.assertEquals(Assert.java:472) at org.apache.solr.cloud.BasicDistributedZkTest.test(BasicDistributedZkTest.java:333) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:963) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:938) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evalua
[jira] [Commented] (SOLR-7746) Ping requests stopped working with distrib=true in Solr 5.2.1
[ https://issues.apache.org/jira/browse/SOLR-7746?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715804#comment-14715804 ] Gregory Chanan commented on SOLR-7746: -- 1. You should not modify TestMiniSolrCloudCluster. That's for testing whether the MiniSolrCloudCluster itself works. Write a test that uses the MiniSolrCloudCluster, there should be a number of examples. Or maybe you don't even need it -- can SolrPingTest satisfy what you need? 2. {code} + // Send distributed and non-distributed ping query + cloudSolrClient.setDefaultCollection(collectionName); + SolrPing req = new SolrPing(); + req.setDistrib(true); + SolrPingResponse rsp = req.process(cloudSolrClient, collectionName); + assertEquals(0, rsp.getStatus()); + + cloudSolrClient.setDefaultCollection(collectionName); + req = new SolrPing(); + req.setDistrib(false); + rsp = req.process(cloudSolrClient, collectionName); + assertEquals(0, rsp.getStatus()); {code} Most of this code is unnecessary, you set the default collection multiple times, you pass the collectionName even though it's set, you create a new request when it would just suffice to set distrib. 3. {code} + public SolrPing setDistrib(boolean distrib) { +params.add("distrib", distrib ? "true" : "false"); +return this; + } {code} You shouldn't modify SolrPing just to test it. Just extend getParams on SolrPing for your distrib example. 4. Can you just have a single one of these instead of putting it in each clause? {code} // Send an error or return + if( ex != null ) { +throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, +"Ping query caused exception: "+ex.getMessage(), ex ); + } +} else { {code} > Ping requests stopped working with distrib=true in Solr 5.2.1 > - > > Key: SOLR-7746 > URL: https://issues.apache.org/jira/browse/SOLR-7746 > Project: Solr > Issue Type: Bug > Components: SolrCloud >Affects Versions: 5.2.1 >Reporter: Alexey Serba > Attachments: SOLR-7746.patch, SOLR-7746.patch, SOLR-7746.patch, > SOLR-7746.patch > > > {noformat:title="steps to reproduce"} > # start 1 node SolrCloud cluster > sh> ./bin/solr -c -p > # create a test collection (we won’t use it, but I just want to it to load > solr configs to Zk) > ./bin/solr create_collection -c test -d sample_techproducts_configs -p > # create another test collection with 2 shards > curl > 'http://localhost:/solr/admin/collections?action=CREATE&name=test2&numShards=2&replicationFactor=1&maxShardsPerNode=2&collection.configName=test' > # try distrib ping request > curl > 'http://localhost:/solr/test2/admin/ping?wt=json&distrib=true&indent=true' > ... > "error":{ > "msg":"Ping query caused exception: Error from server at > http://192.168.59.3:/solr/test2_shard2_replica1: Cannot execute the > PingRequestHandler recursively" > ... > {noformat} > {noformat:title="Exception"} > 2116962 [qtp599601600-13] ERROR org.apache.solr.core.SolrCore [test2 shard2 > core_node1 test2_shard2_replica1] – org.apache.solr.common.SolrException: > Cannot execute the PingRequestHandler recursively > at > org.apache.solr.handler.PingRequestHandler.handlePing(PingRequestHandler.java:246) > at > org.apache.solr.handler.PingRequestHandler.handleRequestBody(PingRequestHandler.java:211) > at > org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:143) > at org.apache.solr.core.SolrCore.execute(SolrCore.java:2064) > at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:654) > at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:450) > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:227) > at > org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:196) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7950) Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO)
[ https://issues.apache.org/jira/browse/SOLR-7950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715777#comment-14715777 ] ASF subversion and git services commented on SOLR-7950: --- Commit 1698039 from gcha...@apache.org in branch 'dev/trunk' [ https://svn.apache.org/r1698039 ] SOLR-7950: Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO) > Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO) > - > > Key: SOLR-7950 > URL: https://issues.apache.org/jira/browse/SOLR-7950 > Project: Solr > Issue Type: Bug >Affects Versions: 4.10.3, Trunk >Reporter: Hrishikesh Gadre >Assignee: Gregory Chanan > Fix For: Trunk, 5.4 > > Attachments: solr-7950-v2.patch, solr-7950.patch > > > When using kerberos authentication mechanism (SPNEGO auth scheme), the Apache > Http client is incorrectly configured with *all* auth schemes (e.g. Basic, > Digest, NTLM, Kerberos, Negotiate etc.) instead of just 'Negotiate'. > This issue was identified after configuring Solr with both Basic + Negotiate > authentication schemes simultaneously. The problem in this case is that Http > client is configured with Kerberos credentials and the default (and > incorrect) auth scheme configuration prefers Basic authentication over > Kerberos. Since the basic authentication credentials are missing, the > authentication and as a result the Http request fails. (I ran into this > problem while creating a collection where there is an internal communication > between Solr servers). > The root cause for this issue is that, AbstractHttpClient::getAuthSchemes() > API call prepares an AuthSchemeRegistry instance with all possible > authentication schemes. Hence when we register the SPNEGO auth scheme in Solr > codebase, it overrides the previous configuration for SPNEGO - but doesn't > remove the other auth schemes from the client configuration. Please take a > look at relevant code snippet. > https://github.com/apache/lucene-solr/blob/trunk/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientConfigurer.java#L80 > A trivial fix would be to prepare a new AuthSchemeRegistry instance > configured with just SPENGO mechanism and set it in the HttpClient. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-7950) Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO)
[ https://issues.apache.org/jira/browse/SOLR-7950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gregory Chanan resolved SOLR-7950. -- Resolution: Fixed Fix Version/s: 5.4 Trunk Thanks for the patch Hrishikesh, committed to Trunk and 5x. > Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO) > - > > Key: SOLR-7950 > URL: https://issues.apache.org/jira/browse/SOLR-7950 > Project: Solr > Issue Type: Bug >Affects Versions: 4.10.3, Trunk >Reporter: Hrishikesh Gadre >Assignee: Gregory Chanan > Fix For: Trunk, 5.4 > > Attachments: solr-7950-v2.patch, solr-7950.patch > > > When using kerberos authentication mechanism (SPNEGO auth scheme), the Apache > Http client is incorrectly configured with *all* auth schemes (e.g. Basic, > Digest, NTLM, Kerberos, Negotiate etc.) instead of just 'Negotiate'. > This issue was identified after configuring Solr with both Basic + Negotiate > authentication schemes simultaneously. The problem in this case is that Http > client is configured with Kerberos credentials and the default (and > incorrect) auth scheme configuration prefers Basic authentication over > Kerberos. Since the basic authentication credentials are missing, the > authentication and as a result the Http request fails. (I ran into this > problem while creating a collection where there is an internal communication > between Solr servers). > The root cause for this issue is that, AbstractHttpClient::getAuthSchemes() > API call prepares an AuthSchemeRegistry instance with all possible > authentication schemes. Hence when we register the SPNEGO auth scheme in Solr > codebase, it overrides the previous configuration for SPNEGO - but doesn't > remove the other auth schemes from the client configuration. Please take a > look at relevant code snippet. > https://github.com/apache/lucene-solr/blob/trunk/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientConfigurer.java#L80 > A trivial fix would be to prepare a new AuthSchemeRegistry instance > configured with just SPENGO mechanism and set it in the HttpClient. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7950) Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO)
[ https://issues.apache.org/jira/browse/SOLR-7950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715776#comment-14715776 ] ASF subversion and git services commented on SOLR-7950: --- Commit 1698037 from gcha...@apache.org in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1698037 ] SOLR-7950: Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO) > Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO) > - > > Key: SOLR-7950 > URL: https://issues.apache.org/jira/browse/SOLR-7950 > Project: Solr > Issue Type: Bug >Affects Versions: 4.10.3, Trunk >Reporter: Hrishikesh Gadre >Assignee: Gregory Chanan > Attachments: solr-7950-v2.patch, solr-7950.patch > > > When using kerberos authentication mechanism (SPNEGO auth scheme), the Apache > Http client is incorrectly configured with *all* auth schemes (e.g. Basic, > Digest, NTLM, Kerberos, Negotiate etc.) instead of just 'Negotiate'. > This issue was identified after configuring Solr with both Basic + Negotiate > authentication schemes simultaneously. The problem in this case is that Http > client is configured with Kerberos credentials and the default (and > incorrect) auth scheme configuration prefers Basic authentication over > Kerberos. Since the basic authentication credentials are missing, the > authentication and as a result the Http request fails. (I ran into this > problem while creating a collection where there is an internal communication > between Solr servers). > The root cause for this issue is that, AbstractHttpClient::getAuthSchemes() > API call prepares an AuthSchemeRegistry instance with all possible > authentication schemes. Hence when we register the SPNEGO auth scheme in Solr > codebase, it overrides the previous configuration for SPNEGO - but doesn't > remove the other auth schemes from the client configuration. Please take a > look at relevant code snippet. > https://github.com/apache/lucene-solr/blob/trunk/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientConfigurer.java#L80 > A trivial fix would be to prepare a new AuthSchemeRegistry instance > configured with just SPENGO mechanism and set it in the HttpClient. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
RE: svn commit: r1697131 - in /lucene/dev/trunk/lucene: JRE_VERSION_MIGRATION.txt test-framework/src/java/org/apache/lucene/util/TestUtil.java
: 1) The test framework never randomly disables asserts (it cannot do : this, because asserts are enabled before the VM starts - you cannot do Hmm, ok -- i thought this was happening when the JVM was forked. In any case, my underlying concern still seems valid to me: confusing nonsensical test failures could occur if someone or some jenkins runs with -Dtests.asserts=false (which we should really do in at least one jenkins job to ensure we're not relying on side effects of method called in an assert ... wasn't that the whole point of adding tests.asserts in LUCENE-6019?) : (we also have a test that validates this). In addition, our tests should : not check the JVM for correctness, it is just there to make an : assumption about how WS should behave. So when this assert fails, the : test is still not wrong. i think having sanity checks about our assumptions of the JVM are useful -- just like we (last time i checked) have sanity checks that filesystems behave as we expect. It's one thing to say "we don't need a check that 2 + 2 == 4" but in cases like this where assumptions (about hte definition of whitespace) can evidently change between JVM versions, it's nice to have a sanity check that our tests are testing the right thing (ie: fail fast because the JVM isn't behaving the way we expect, don't fail slow with a weird obscure NPE or AIOBE error because our code expected 2+2==4 and the JVM gave us -42 instead) : 2) Now the answer: I changed it because of performance. The Ah, ok -- totally fair point. Good call. : Based on (1) and (2) this is OK. If you still want to revert to : Assert.assertTrue like before, please use an if-condition instead: : if (not whitespace) Assert.fail(String.format(error message)); Actaully, the more i think about it, the more i want to: a) add an explicit test that loops over this array and fails with a clear "your JVM is weird and doesn't consider thi char whitespace, that's messed up and this test and possible some other code needs updated if this is the new Java/unicode rules" b) leave the plain assert you have in place as a fallback to provide a clera assertion msg in case someone runs a test method/class that uses this utility method but don't run the sanity check test mentioned in (a) because of -Dtestcase or -Dtest.method. (yes, i realize only one test method currently uses it nad the utility has already been refactored up to live in that same test class, but this is all about future proofing in case it ever gets refactred again) -Hoss http://www.lucidworks.com/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7789) Introduce a ConfigSet management API
[ https://issues.apache.org/jira/browse/SOLR-7789?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715775#comment-14715775 ] Gregory Chanan commented on SOLR-7789: -- Oh and forgot to mention, I was able to reproduce with failures with the beasting script and just turning down the number of iterations allowed it to pass with 100 iterations, 8 concurrent. > Introduce a ConfigSet management API > > > Key: SOLR-7789 > URL: https://issues.apache.org/jira/browse/SOLR-7789 > Project: Solr > Issue Type: New Feature >Reporter: Gregory Chanan >Assignee: Gregory Chanan > Attachments: SOLR-7789.patch, SOLR-7789.patch, SOLR-7789.patch, > SOLR-7789.patch, SOLR-7789.patch > > > SOLR-5955 describes a feature to automatically create a ConfigSet, based on > another one, from a collection API call (i.e. one step collection creation). > Discussion there yielded SOLR-7742, Immutable ConfigSet support. To close > the loop, we need support for a ConfigSet management API. > The simplest ConfigSet API could have one operation: > create a new config set, based on an existing one, possible modifying the > ConfigSet properties. Note you need to be able to modify the ConfigSet > properties at creation time because otherwise Immutable could not be changed. > Another logical operation to support is ConfigSet deletion; that may be more > complicated to implement than creation because you need to handle the case > where a collection is already using the configuration. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7789) Introduce a ConfigSet management API
[ https://issues.apache.org/jira/browse/SOLR-7789?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gregory Chanan updated SOLR-7789: - Attachment: SOLR-7789.patch Here's a new version of the patch. Changes: - rebased to lastest trunk - Fixed typo pointed out by Mark - Added comment to OverseerCollectionMessageHandler - Renamed OverseerProcessor -> OverseerTaskProcessor to go along with the OverseerTaskQueue name. I plan on committing this soon if I don't hear any objections and will file follow on jiras for the suggestions above. > Introduce a ConfigSet management API > > > Key: SOLR-7789 > URL: https://issues.apache.org/jira/browse/SOLR-7789 > Project: Solr > Issue Type: New Feature >Reporter: Gregory Chanan >Assignee: Gregory Chanan > Attachments: SOLR-7789.patch, SOLR-7789.patch, SOLR-7789.patch, > SOLR-7789.patch, SOLR-7789.patch > > > SOLR-5955 describes a feature to automatically create a ConfigSet, based on > another one, from a collection API call (i.e. one step collection creation). > Discussion there yielded SOLR-7742, Immutable ConfigSet support. To close > the loop, we need support for a ConfigSet management API. > The simplest ConfigSet API could have one operation: > create a new config set, based on an existing one, possible modifying the > ConfigSet properties. Note you need to be able to modify the ConfigSet > properties at creation time because otherwise Immutable could not be changed. > Another logical operation to support is ConfigSet deletion; that may be more > complicated to implement than creation because you need to handle the case > where a collection is already using the configuration. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6759) Integrate lat/long BKD and spatial 3d, part 2
[ https://issues.apache.org/jira/browse/LUCENE-6759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715759#comment-14715759 ] Karl Wright commented on LUCENE-6759: - I did a repeat run, being sure to "ant clean" first, and it still passed. Hmm. > Integrate lat/long BKD and spatial 3d, part 2 > - > > Key: LUCENE-6759 > URL: https://issues.apache.org/jira/browse/LUCENE-6759 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless > Attachments: LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch > > > This is just a continuation of LUCENE-6699, which became too big. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
RE: svn commit: r1697131 - in /lucene/dev/trunk/lucene: JRE_VERSION_MIGRATION.txt test-framework/src/java/org/apache/lucene/util/TestUtil.java
Hi Hoss, sorry did not see your message! 1) The test framework never randomly disables asserts (it cannot do this, because asserts are enabled before the VM starts - you cannot do that at runtime). So to disable asserts you have to pass this explicitly to the test runner, and the default is asserts enabled. Currently Policeman Jenkins does not disable asserts at the moment, but it could do that. For normal developers running tests, they are always enabled (we also have a test that validates this). In addition, our tests should not check the JVM for correctness, it is just there to make an assumption about how WS should behave. So when this assert fails, the test is still not wrong. 2) Now the answer: I changed it because of performance. The String.format / String concatenation is executed on every single character and if the assert does not fail. If you create a large whitespace string this takes endless (I tried it), because it creates a new string, formats it,... A native Java assert only calls the string part if the assert actually failed. Based on (1) and (2) this is OK. If you still want to revert to Assert.assertTrue like before, please use an if-condition instead: if (not whitespace) Assert.fail(String.format(error message)); Uwe - Uwe Schindler H.-H.-Meier-Allee 63, D-28213 Bremen http://www.thetaphi.de eMail: u...@thetaphi.de > -Original Message- > From: Chris Hostetter [mailto:hossman_luc...@fucit.org] > Sent: Thursday, August 27, 2015 12:15 AM > To: u...@thetaphi.de; Lucene Dev > Subject: Re: svn commit: r1697131 - in /lucene/dev/trunk/lucene: > JRE_VERSION_MIGRATION.txt test- > framework/src/java/org/apache/lucene/util/TestUtil.java > > > Uwe: I'm still concerned about this change and the way it might result in > confusing failure messages in the future (if the whitespace def of other > characters changes) ... can you please explain your choice "Assert.assertTrue > -> assert" ? > > > On Mon, 24 Aug 2015, Chris Hostetter wrote: > > : Uwe: why did you change this from "Assert.assertTrue" to "assert" ? > : > : In the old code the test would fail every time with a clear explanation of > : hte problem -- in your new code, if assertions are randomly disabled by > : the test framework, then the sanity check won't run and instead we'll get > : a strange failure from whatever test called this method. > : > : > : > : : > == > > : : --- lucene/dev/trunk/lucene/test- > framework/src/java/org/apache/lucene/util/TestUtil.java (original) > : : +++ lucene/dev/trunk/lucene/test- > framework/src/java/org/apache/lucene/util/TestUtil.java Sat Aug 22 > 21:33:47 2015 > : : @@ -35,6 +35,7 @@ import java.util.Collections; > : : import java.util.HashMap; > : : import java.util.Iterator; > : : import java.util.List; > : : +import java.util.Locale; > : : import java.util.Map; > : : import java.util.NoSuchElementException; > : : import java.util.Random; > : : @@ -1188,7 +1189,7 @@ public final class TestUtil { > : :int offset = nextInt(r, 0, WHITESPACE_CHARACTERS.length-1); > : :char c = WHITESPACE_CHARACTERS[offset]; > : :// sanity check > : : - Assert.assertTrue("Not really whitespace? (@"+offset+"): " + c, > Character.isWhitespace(c)); > : : + assert Character.isWhitespace(c) : String.format(Locale.ENGLISH, > "Not > really whitespace? WHITESPACE_CHARACTERS[%d] is '\\u%04X'", offset, (int) > c); > : :out.append(c); > : : } > : : return out.toString(); > : : @@ -1307,9 +1308,9 @@ public final class TestUtil { > : : '\u001E', > : : '\u001F', > : : '\u0020', > : : -// '\u0085', faild sanity check? > : : +// '\u0085', failed sanity check? > : : '\u1680', > : : -'\u180E', > : : +// '\u180E', no longer whitespace in Unicode 7.0 (Java 9)! > : : '\u2000', > : : '\u2001', > : : '\u2002', > : : > : : > : : > : > : -Hoss > : http://www.lucidworks.com/ > : > > -Hoss > http://www.lucidworks.com/ > > - > To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional > commands, e-mail: dev-h...@lucene.apache.org - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6759) Integrate lat/long BKD and spatial 3d, part 2
[ https://issues.apache.org/jira/browse/LUCENE-6759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715680#comment-14715680 ] Karl Wright commented on LUCENE-6759: - Huh. Even odder, tried the "repro" line and got this: {code} ant test -Dtestcase=TestGeo3DPointField -Dtests.method=testRandomMedium* -Dtests.seed=E1D51F3E8B12E79D -Dtests.multiplier=10 -Dtests.iters=5 -Dtests.slow=true -Dtests.linedocsfile=/lucenedata/hudson.enwiki.random.lines.txt.fixed -Dtests.locale=pl -Dtests.timezone=America/Inuvik -Dtests.asserts=true -Dtests.file.encoding=UTF-8 ... BUILD SUCCESSFUL Total time: 13 minutes 14 seconds {code} So now I'm very puzzled, [~mikemccand]. > Integrate lat/long BKD and spatial 3d, part 2 > - > > Key: LUCENE-6759 > URL: https://issues.apache.org/jira/browse/LUCENE-6759 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless > Attachments: LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch > > > This is just a continuation of LUCENE-6699, which became too big. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-7981) term based ValueSourceParsers should support an option to run an analyzer for hte specified field on the input
Hoss Man created SOLR-7981: -- Summary: term based ValueSourceParsers should support an option to run an analyzer for hte specified field on the input Key: SOLR-7981 URL: https://issues.apache.org/jira/browse/SOLR-7981 Project: Solr Issue Type: Improvement Reporter: Hoss Man The following functions all take exactly 2 arguments: a field name, and a term value... * idf * termfreq * tf * totaltermfreq ...we should consider adding an optional third argument to indicate if an analyzer for the specified field should be used on the input to find the real "Term" to consider. For example, the following might all result in equivilent numeric values for all docs assuming simple plural stemming and lowercasing... {noformat} termfreq(foo_t,'Bicycles',query) // use the query analyzer for field foo_t on input Bicycles termfreq(foo_t,'Bicycles',index) // use the index analyzer for field foo_t on input Bicycles termfreq(foo_t,'bicycle',none) // no analyzer used to construct Term termfreq(foo_t,'bicycle') // legacy 2 arg syntax, same as 'none' {noformat} (Special error checking needed if analyzer creates more then one term for the given input string) -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6759) Integrate lat/long BKD and spatial 3d, part 2
[ https://issues.apache.org/jira/browse/LUCENE-6759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715658#comment-14715658 ] Karl Wright commented on LUCENE-6759: - So I tried to reproduce this in geo3d-land exclusively, and coded this: {code} c = new GeoCircle(PlanetModel.WGS84,0.006204988457123483,0.003379977917811208,7.780831828380698E-4); p1 = new GeoPoint(PlanetModel.WGS84,0.005231514023315527,0.0034278119211296914); assertTrue(!c.isWithin(p1)); xyzb = new XYZBounds(); c.getBounds(xyzb); area = GeoAreaFactory.makeGeoArea(PlanetModel.WGS84, xyzb.getMinimumX(), xyzb.getMaximumX(), xyzb.getMinimumY(), xyzb.getMaximumY(), xyzb.getMinimumZ(), xyzb.getMaximumZ()); // Doesn't have to be true, but is... assertTrue(!area.isWithin(p1)); {code} The exact point in question shows up as outside of *even* the bounds computed for the circle. So honestly I don't know how it wound up getting included? Unless, perhaps, the descent decisions were made based on the approximation? Looking at the code to see how to delve deeper... > Integrate lat/long BKD and spatial 3d, part 2 > - > > Key: LUCENE-6759 > URL: https://issues.apache.org/jira/browse/LUCENE-6759 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless > Attachments: LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch > > > This is just a continuation of LUCENE-6699, which became too big. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7795) Fold Interval Faceting into Range Faceting
[ https://issues.apache.org/jira/browse/SOLR-7795?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Tomás Fernández Löbbe updated SOLR-7795: Attachment: SOLR-7795.patch Fix pivots use case. Added some tests with pivots + interval facets. I'll add some more. > Fold Interval Faceting into Range Faceting > -- > > Key: SOLR-7795 > URL: https://issues.apache.org/jira/browse/SOLR-7795 > Project: Solr > Issue Type: Task >Reporter: Tomás Fernández Löbbe > Fix For: Trunk, 5.4 > > Attachments: SOLR-7795.patch, SOLR-7795.patch, SOLR-7795.patch > > > Now that range faceting supports a "filter" and a "dv" method, and that > interval faceting is supported on fields with {{docValues=false}}, I think we > should make it so that interval faceting is just a different way of > specifying ranges in range faceting, allowing users to indicate specific > ranges. > I propose we use the same syntax for intervals, but under the "range" > parameter family: > {noformat} > facet.range=price& > f.price.facet.range.set=[0,10]& > f.price.facet.range.set=(10,100] > {noformat} > The counts for those ranges would come in the response also inside of the > "range_facets" section. I'm not sure if it's better to include the ranges in > the "counts" section, or in a different section (intervals?sets?buckets?). > I'm open to suggestions. > {code} > "facet_ranges":{ > "price":{ > "counts":[ > "[0,10]",3, > "(10,100]",2] >} > } > {code} > or… > {code} > "facet_ranges":{ > "price":{ > "intervals":[ > "[0,10]",3, > "(10,100]",2] >} > } > {code} > We should support people specifying both things on the same field. > Once this is done, "interval faceting" could be deprecated, as all it's > functionality is now possible through range queries. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: svn commit: r1697131 - in /lucene/dev/trunk/lucene: JRE_VERSION_MIGRATION.txt test-framework/src/java/org/apache/lucene/util/TestUtil.java
Uwe: I'm still concerned about this change and the way it might result in confusing failure messages in the future (if the whitespace def of other characters changes) ... can you please explain your choice "Assert.assertTrue -> assert" ? On Mon, 24 Aug 2015, Chris Hostetter wrote: : Uwe: why did you change this from "Assert.assertTrue" to "assert" ? : : In the old code the test would fail every time with a clear explanation of : hte problem -- in your new code, if assertions are randomly disabled by : the test framework, then the sanity check won't run and instead we'll get : a strange failure from whatever test called this method. : : : : : == : : --- lucene/dev/trunk/lucene/test-framework/src/java/org/apache/lucene/util/TestUtil.java (original) : : +++ lucene/dev/trunk/lucene/test-framework/src/java/org/apache/lucene/util/TestUtil.java Sat Aug 22 21:33:47 2015 : : @@ -35,6 +35,7 @@ import java.util.Collections; : : import java.util.HashMap; : : import java.util.Iterator; : : import java.util.List; : : +import java.util.Locale; : : import java.util.Map; : : import java.util.NoSuchElementException; : : import java.util.Random; : : @@ -1188,7 +1189,7 @@ public final class TestUtil { : :int offset = nextInt(r, 0, WHITESPACE_CHARACTERS.length-1); : :char c = WHITESPACE_CHARACTERS[offset]; : :// sanity check : : - Assert.assertTrue("Not really whitespace? (@"+offset+"): " + c, Character.isWhitespace(c)); : : + assert Character.isWhitespace(c) : String.format(Locale.ENGLISH, "Not really whitespace? WHITESPACE_CHARACTERS[%d] is '\\u%04X'", offset, (int) c); : :out.append(c); : : } : : return out.toString(); : : @@ -1307,9 +1308,9 @@ public final class TestUtil { : : '\u001E', : : '\u001F', : : '\u0020', : : -// '\u0085', faild sanity check? : : +// '\u0085', failed sanity check? : : '\u1680', : : -'\u180E', : : +// '\u180E', no longer whitespace in Unicode 7.0 (Java 9)! : : '\u2000', : : '\u2001', : : '\u2002', : : : : : : : : -Hoss : http://www.lucidworks.com/ : -Hoss http://www.lucidworks.com/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6759) Integrate lat/long BKD and spatial 3d, part 2
[ https://issues.apache.org/jira/browse/LUCENE-6759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715623#comment-14715623 ] Michael McCandless commented on LUCENE-6759: OK I committed, beasted, hit this failure: {noformat} ant test -Dtestcase=TestGeo3DPointField -Dtests.method=testRandomMedium* -Dtests.seed=E1D51F3E8B12E79D -Dtests.multiplier=10 -Dtests.iters=5 -Dtests.slow=true -Dtests.linedocsfile=/lucenedata/hudson.enwiki.random.lines.txt.fixed -Dtests.locale=pl -Dtests.timezone=America/Inuvik -Dtests.asserts=true -Dtests.file.encoding=UTF-8 [junit4:pickseed] Seed property 'tests.seed' already defined: E1D51F3E8B12E79D [junit4] says 今日は! Master seed: E1D51F3E8B12E79D [junit4] Executing 1 suite with 1 JVM. [junit4] [junit4] Started J0 PID(62227@localhost). [junit4] Suite: org.apache.lucene.bkdtree3d.TestGeo3DPointField [junit4] OK 70.2s | TestGeo3DPointField.testRandomMedium {#0 seed=[E1D51F3E8B12E79D:5C0B2896CA7784FB]} [junit4] 2> sie 26, 2015 4:11:15 PM com.carrotsearch.randomizedtesting.RandomizedRunner$QueueUncaughtExceptionsHandler uncaughtException [junit4] 2> WARNING: Uncaught exception in thread: Thread[T0,5,TGRP-TestGeo3DPointField] [junit4] 2> java.lang.AssertionError: T0: iter=425 id=1237 docID=1237 lat=0.005231514023315527 lon=0.0034278119211296914 expected false but got: true deleted?=false [junit4] 2> point1=[lat=0.005231514023315527, lon=0.0034278119211296914], iswithin=false [junit4] 2> point2=[X=1.0010991445151618, Y=0.003431592678386528, Z=0.00523734247369568], iswithin=false [junit4] 2> query=PointInGeo3DShapeQuery: field=point:PlanetModel: PlanetModel.WGS84 Shape: GeoCircle: {planetmodel=PlanetModel.WGS84, center=[lat=0.006204988457123483, lon=0.003379977917811208], radius=7.780831828380698E-4(0.04458088248672737)} [junit4] 2>at __randomizedtesting.SeedInfo.seed([E1D51F3E8B12E79D]:0) [junit4] 2>at org.junit.Assert.fail(Assert.java:93) [junit4] 2>at org.apache.lucene.bkdtree3d.TestGeo3DPointField$4._run(TestGeo3DPointField.java:632) [junit4] 2>at org.apache.lucene.bkdtree3d.TestGeo3DPointField$4.run(TestGeo3DPointField.java:521) {noformat} > Integrate lat/long BKD and spatial 3d, part 2 > - > > Key: LUCENE-6759 > URL: https://issues.apache.org/jira/browse/LUCENE-6759 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless > Attachments: LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch > > > This is just a continuation of LUCENE-6699, which became too big. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7942) fix error logging when index is locked on startup, add deprecation warning if unlockOnStartup is configured
[ https://issues.apache.org/jira/browse/SOLR-7942?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hoss Man updated SOLR-7942: --- Attachment: SOLR-7942.patch Here's what i had in mind, with the {{true // 'fail' in trunk}} changed to {{false // warn in 5x}} when backporting. Anyone have concerns/feedack about the various warning/err msgs? > fix error logging when index is locked on startup, add deprecation warning if > unlockOnStartup is configured > --- > > Key: SOLR-7942 > URL: https://issues.apache.org/jira/browse/SOLR-7942 > Project: Solr > Issue Type: Bug >Reporter: Hoss Man >Assignee: Hoss Man >Priority: Minor > Attachments: SOLR-7942.patch > > > LUCENE-6508 removed support for unlockOnStartup, but the way the changes are > made are inconsistent with how other config options have been deprecated in > the past, and cause a confusing error message if an index is locked on > startup... > * in 5x, the SolrConfig constructor should log a warning if the > unlockOnStartup option is specified (regardless of whether it's true or > false) so users know they need to cleanup their configs and change their > expectations (even if they never get - or have not yet gotten - a lock > problem) > * in SolrCore, the LockObtainFailedException that is thrown if the index is > locked should _not_ say "{{Solr now longer supports forceful unlocking via > 'unlockOnStartup'}}" > ** besides the no/now typo, this wording missleads users into thinking that > the LockObtainFailedException error is in some way related to that config > option -- creating an implication that using that option lead them to this > error. we shouldn't mention that here. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7844) Zookeeper session expiry during shard leader election can cause multiple leaders.
[ https://issues.apache.org/jira/browse/SOLR-7844?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mike Roberts updated SOLR-7844: --- Attachment: SOLR-7844.patch Initial attempt at a patch. When we reconnect after a zookeeper session expiry we cancel any ongoing leader elections. The current ongoing election threads check status in a few places to ensure that no 'leader specific' operations are taken after the election has been canceled. I don't think this is bulletproof, but it does significantly reduce the size of the window in which this scenario can occur. > Zookeeper session expiry during shard leader election can cause multiple > leaders. > - > > Key: SOLR-7844 > URL: https://issues.apache.org/jira/browse/SOLR-7844 > Project: Solr > Issue Type: Bug >Affects Versions: 4.10.4 >Reporter: Mike Roberts > Attachments: SOLR-7844.patch > > > If the ZooKeeper session expires for a host during shard leader election, the > ephemeral leader_elect nodes are removed. However the threads that were > processing the election are still present (and could believe the host won the > election). They will then incorrectly create leader nodes once a new > ZooKeeper session is established. > This introduces a subtle race condition that could cause two hosts to become > leader. > Scenario: > a three machine cluster, all of the machines are restarting at approximately > the same time. > The first machine starts, writes a leader_elect ephemeral node, it's the only > candidate in the election so it wins and starts the leadership process. As it > knows it has peers, it begins to block waiting for the peers to arrive. > During this period of blocking[1] the ZK connection drops and the session > expires. > A new ZK session is established, and ElectionContext.cancelElection is > called. Then register() is called and a new set of leader_elect ephemeral > nodes are created. > During the period between the ZK session expiring, and new set of > leader_elect nodes being created the second machine starts. > It creates its leader_elect ephemeral nodes, as there are no other nodes it > wins the election and starts the leadership process. As its still missing one > of its peers, it begins to block waiting for the third machine to join. > There is now a race between machine1 & machine2, both of whom think they are > the leader. > So far, this isn't too bad, because the machine that loses the race will fail > when it tries to create the /collection/name/leader/shard1 node (as it > already exists), and will rejoin the election. > While this is happening, machine3 has started and has queued for leadership > behind machine2. > If the loser of the race is machine2, when it rejoins the election it cancels > the current context, deleting it's leader_elect ephemeral nodes. > At this point, machine3 believes it has become leader (the watcher it has on > the leader_elect node fires), and it runs the LeaderElector::checkIfIAmLeader > method. This method DELETES the current /collection/name/leader/shard1 node, > then starts the leadership process (as all three machines are now running, it > does not block to wait). > So, machine1 won the race with machine2 and declared its leadership and > created the nodes. However, machine3 has just deleted them, and recreated > them for itself. So machine1 and machine3 both believe they are the leader. > I am thinking that the fix should be to cancel & close all election contexts > immediately on reconnect (we do cancel them, however it's run serially which > has blocking issues, and just canceling does not cause the wait loop to > exit). That election context logic already has checks on the closed flag, so > they should exit if they see it has been closed. > I'm working on a patch for this. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7950) Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO)
[ https://issues.apache.org/jira/browse/SOLR-7950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715574#comment-14715574 ] Gregory Chanan commented on SOLR-7950: -- +1. Patch looks good and thanks for the explanation. It sounds like from what Ishan is saying there may be some more work integrating the HADOOP-12082 change with Solr's own Basic Auth Plugin, but that can be done in another jira. > Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO) > - > > Key: SOLR-7950 > URL: https://issues.apache.org/jira/browse/SOLR-7950 > Project: Solr > Issue Type: Bug >Affects Versions: 4.10.3, Trunk >Reporter: Hrishikesh Gadre >Assignee: Gregory Chanan > Attachments: solr-7950-v2.patch, solr-7950.patch > > > When using kerberos authentication mechanism (SPNEGO auth scheme), the Apache > Http client is incorrectly configured with *all* auth schemes (e.g. Basic, > Digest, NTLM, Kerberos, Negotiate etc.) instead of just 'Negotiate'. > This issue was identified after configuring Solr with both Basic + Negotiate > authentication schemes simultaneously. The problem in this case is that Http > client is configured with Kerberos credentials and the default (and > incorrect) auth scheme configuration prefers Basic authentication over > Kerberos. Since the basic authentication credentials are missing, the > authentication and as a result the Http request fails. (I ran into this > problem while creating a collection where there is an internal communication > between Solr servers). > The root cause for this issue is that, AbstractHttpClient::getAuthSchemes() > API call prepares an AuthSchemeRegistry instance with all possible > authentication schemes. Hence when we register the SPNEGO auth scheme in Solr > codebase, it overrides the previous configuration for SPNEGO - but doesn't > remove the other auth schemes from the client configuration. Please take a > look at relevant code snippet. > https://github.com/apache/lucene-solr/blob/trunk/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientConfigurer.java#L80 > A trivial fix would be to prepare a new AuthSchemeRegistry instance > configured with just SPENGO mechanism and set it in the HttpClient. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6699) Integrate lat/lon BKD and spatial3d
[ https://issues.apache.org/jira/browse/LUCENE-6699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715538#comment-14715538 ] ASF subversion and git services commented on LUCENE-6699: - Commit 1698004 from [~mikemccand] in branch 'dev/branches/lucene6699' [ https://svn.apache.org/r1698004 ] LUCENE-6699: increase fudge factor; don't check hits if quantization changed the expected result > Integrate lat/lon BKD and spatial3d > --- > > Key: LUCENE-6699 > URL: https://issues.apache.org/jira/browse/LUCENE-6699 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless >Assignee: Michael McCandless > Attachments: Geo3DPacking.java, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch > > > I'm opening this for discussion, because I'm not yet sure how to do > this integration, because of my ignorance about spatial in general and > spatial3d in particular :) > Our BKD tree impl is very fast at doing lat/lon shape intersection > (bbox, polygon, soon distance: LUCENE-6698) against previously indexed > points. > I think to integrate with spatial3d, we would first need to record > lat/lon/z into doc values. Somewhere I saw discussion about how we > could stuff all 3 into a single long value with acceptable precision > loss? Or, we could use BinaryDocValues? We need all 3 dims available > to do the fast per-hit query time filtering. > But, second: what do we index into the BKD tree? Can we "just" index > earth surface lat/lon, and then at query time is spatial3d able to > give me an enclosing "surface lat/lon" bbox for a 3d shape? Or > ... must we index all 3 dimensions into the BKD tree (seems like this > could be somewhat wasteful)? -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7971) Reduce memory allocated by JavaBinCodec to encode large strings
[ https://issues.apache.org/jira/browse/SOLR-7971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715540#comment-14715540 ] Yonik Seeley commented on SOLR-7971: Making 3 copies really feels heavyweight (the current code makes a single copy in the case of large strings). string->scratch, scratch->off-heap-buffer, off-heap-buffer->scratch, scratch->write And this doesn't even use less system memory... just less heap memory. I wonder how it would do against Mikhail's idea of just calculating the UTF8 length first. If we keep with DirectByteBuffer, we should at least eliminate one of the copies by encoding directly to it? > Reduce memory allocated by JavaBinCodec to encode large strings > --- > > Key: SOLR-7971 > URL: https://issues.apache.org/jira/browse/SOLR-7971 > Project: Solr > Issue Type: Sub-task > Components: Response Writers, SolrCloud >Reporter: Shalin Shekhar Mangar >Assignee: Shalin Shekhar Mangar >Priority: Minor > Fix For: Trunk, 5.4 > > Attachments: SOLR-7971-directbuffer.patch, > SOLR-7971-directbuffer.patch, SOLR-7971.patch > > > As discussed in SOLR-7927, we can reduce the buffer memory allocated by > JavaBinCodec while writing large strings. > https://issues.apache.org/jira/browse/SOLR-7927?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14700420#comment-14700420 > {quote} > The maximum Unicode code point (as of Unicode 8 anyway) is U+10 > ([http://www.unicode.org/glossary/#code_point]). This is encoded in UTF-16 > as surrogate pair {{\uDBFF\uDFFF}}, which takes up two Java chars, and is > represented in UTF-8 as the 4-byte sequence {{F4 8F BF BF}}. This is likely > where the mistaken 4-bytes-per-Java-char formulation came from: the maximum > number of UTF-8 bytes required to represent a Unicode *code point* is 4. > The maximum Java char is {{\u}}, which is represented in UTF-8 as the > 3-byte sequence {{EF BF BF}}. > So I think it's safe to switch to using 3 bytes per Java char (the unit of > measurement returned by {{String.length()}}), like > {{CompressingStoredFieldsWriter.writeField()}} does. > {quote} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6759) Integrate lat/long BKD and spatial 3d, part 2
[ https://issues.apache.org/jira/browse/LUCENE-6759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715521#comment-14715521 ] Michael McCandless commented on LUCENE-6759: Ugh sorry I thought I had committed the latest patch ... I'll do that shortly. And I'll also fix the test to skip checking a hit when the quantization changed the expected result... > Integrate lat/long BKD and spatial 3d, part 2 > - > > Key: LUCENE-6759 > URL: https://issues.apache.org/jira/browse/LUCENE-6759 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless > Attachments: LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch > > > This is just a continuation of LUCENE-6699, which became too big. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7950) Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO)
[ https://issues.apache.org/jira/browse/SOLR-7950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715499#comment-14715499 ] Ishan Chattopadhyaya commented on SOLR-7950: bq. So we may have to introduce Basic authentication scheme (in addition to SPNEGO) in Solr by some other way. We already have basic auth, https://cwiki.apache.org/confluence/display/solr/Basic+Authentication+Plugin However, we still don't have it working together as yet. bq. I don't think we use the Hadoop security framework in Solr. The hadoop-auth's kerberos authentication filters are used. The support for delegation tokens still isn't there. > Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO) > - > > Key: SOLR-7950 > URL: https://issues.apache.org/jira/browse/SOLR-7950 > Project: Solr > Issue Type: Bug >Affects Versions: 4.10.3, Trunk >Reporter: Hrishikesh Gadre >Assignee: Gregory Chanan > Attachments: solr-7950-v2.patch, solr-7950.patch > > > When using kerberos authentication mechanism (SPNEGO auth scheme), the Apache > Http client is incorrectly configured with *all* auth schemes (e.g. Basic, > Digest, NTLM, Kerberos, Negotiate etc.) instead of just 'Negotiate'. > This issue was identified after configuring Solr with both Basic + Negotiate > authentication schemes simultaneously. The problem in this case is that Http > client is configured with Kerberos credentials and the default (and > incorrect) auth scheme configuration prefers Basic authentication over > Kerberos. Since the basic authentication credentials are missing, the > authentication and as a result the Http request fails. (I ran into this > problem while creating a collection where there is an internal communication > between Solr servers). > The root cause for this issue is that, AbstractHttpClient::getAuthSchemes() > API call prepares an AuthSchemeRegistry instance with all possible > authentication schemes. Hence when we register the SPNEGO auth scheme in Solr > codebase, it overrides the previous configuration for SPNEGO - but doesn't > remove the other auth schemes from the client configuration. Please take a > look at relevant code snippet. > https://github.com/apache/lucene-solr/blob/trunk/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientConfigurer.java#L80 > A trivial fix would be to prepare a new AuthSchemeRegistry instance > configured with just SPENGO mechanism and set it in the HttpClient. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7950) Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO)
[ https://issues.apache.org/jira/browse/SOLR-7950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hrishikesh Gadre updated SOLR-7950: --- Attachment: solr-7950-v2.patch [~gchanan] >>We don't support Basic + Negotiate now, right? So we need another solr patch >>to expose the underlying problem? Yes that is correct. I identified this issue while working on LDAP integration (Please refer to [HADOOP-12082|https://issues.apache.org/jira/browse/HADOOP-12082]). I don't think we use the Hadoop security framework in Solr. So we may have to introduce Basic authentication scheme (in addition to SPNEGO) in Solr by some other way. >>There's no fall back mechanism? No. When server supports multiple authentication schemes, client needs to pick one scheme to use (based on list of preferences). The default configuration prefers BASIC scheme over SPENGO. Hence the client attempts to use basic auth scheme. But since the username/password credentials are not configured - the authentication fails. With my patch, we explicitly configure client to use SPNEGO. >>Or can you prefer SPNego over basic? Yes. This can be done by adding SPNEGO before BASIC in the preference list. Here is how the default preferences are initialized. http://grepcode.com/file/repo1.maven.org/maven2/org.apache.httpcomponents/httpclient/4.4.1/org/apache/http/impl/client/AbstractHttpClient.java#AbstractHttpClient.createAuthSchemeRegistry%28%29 I think when we configure HttpClientUtil with Krb5HttpClientConfigurer, we intend to use Kerberos authentication mechanism on the client side. Hence I think just configuring one auth scheme (SPNEGO in this case) is preferable than SPNEGO & BASIC I have also updated the patch to fix a unit test failure. > Invalid auth scheme configuration of Http client when using Kerberos (SPNEGO) > - > > Key: SOLR-7950 > URL: https://issues.apache.org/jira/browse/SOLR-7950 > Project: Solr > Issue Type: Bug >Affects Versions: 4.10.3, Trunk >Reporter: Hrishikesh Gadre >Assignee: Gregory Chanan > Attachments: solr-7950-v2.patch, solr-7950.patch > > > When using kerberos authentication mechanism (SPNEGO auth scheme), the Apache > Http client is incorrectly configured with *all* auth schemes (e.g. Basic, > Digest, NTLM, Kerberos, Negotiate etc.) instead of just 'Negotiate'. > This issue was identified after configuring Solr with both Basic + Negotiate > authentication schemes simultaneously. The problem in this case is that Http > client is configured with Kerberos credentials and the default (and > incorrect) auth scheme configuration prefers Basic authentication over > Kerberos. Since the basic authentication credentials are missing, the > authentication and as a result the Http request fails. (I ran into this > problem while creating a collection where there is an internal communication > between Solr servers). > The root cause for this issue is that, AbstractHttpClient::getAuthSchemes() > API call prepares an AuthSchemeRegistry instance with all possible > authentication schemes. Hence when we register the SPNEGO auth scheme in Solr > codebase, it overrides the previous configuration for SPNEGO - but doesn't > remove the other auth schemes from the client configuration. Please take a > look at relevant code snippet. > https://github.com/apache/lucene-solr/blob/trunk/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientConfigurer.java#L80 > A trivial fix would be to prepare a new AuthSchemeRegistry instance > configured with just SPENGO mechanism and set it in the HttpClient. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7980) CorruptIndex exceptions in ChaosMonkeySafeLeaderTest
[ https://issues.apache.org/jira/browse/SOLR-7980?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715446#comment-14715446 ] Yonik Seeley commented on SOLR-7980: Examples: {code} 2> 25342 WARN (qtp1987916571-72) [n:127.0.0.1:32803_ c:collection1 s:shard2 r:core_node1 x:collection1] o.a.s.h.ReplicationHandler Could not read check sum from index file: _0_SimpleText_0.pst 2> org.apache.lucene.index.CorruptIndexException: codec footer mismatch (file truncated?): actual footer=808464432 vs expected footer=-1071082520 (resource=MMapIndexInput(path="/opt/code/lusolr_trunk2/solr/build/solr-core/test/J0/temp/solr.cloud.ChaosMonkeySafeLeaderTest_38F55D766B70D5BD-001/shard-1-001/cores/collection1/data/index/_0_SimpleText_0.pst")) 2>at org.apache.lucene.codecs.CodecUtil.validateFooter(CodecUtil.java:416) 2>at org.apache.lucene.codecs.CodecUtil.retrieveChecksum(CodecUtil.java:401) 2>at org.apache.solr.handler.ReplicationHandler.getFileList(ReplicationHandler.java:558) 2>at org.apache.solr.handler.ReplicationHandler.handleRequestBody(ReplicationHandler.java:247) 2>at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:151) 2>at org.apache.solr.core.SolrCore.execute(SolrCore.java:2079) 2>at org.apache.solr.servlet.HttpSolrCall.execute(HttpSolrCall.java:667) 2>at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:460) 2>at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:210) 2>at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:179) 2>at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652) 2>at org.apache.solr.client.solrj.embedded.JettySolrRunner$DebugFilter.doFilter(JettySolrRunner.java:106) 2>at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652) 2>at org.eclipse.jetty.servlets.UserAgentFilter.doFilter(UserAgentFilter.java:83) 2>at org.eclipse.jetty.servlets.GzipFilter.doFilter(GzipFilter.java:364) 2>at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652) 2>at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585) 2>at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:221) 2>at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1127) 2>at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515) 2>at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185) 2>at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1061) 2>at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) 2>at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97) 2>at org.eclipse.jetty.server.Server.handle(Server.java:499) 2>at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:310) 2>at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:257) 2>at org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540) 2>at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635) 2>at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555) 2>at java.lang.Thread.run(Thread.java:745) {code} {code} 2> 19962 WARN (RecoveryThread-collection1) [n:127.0.0.1:49515_ c:collection1 s:shard1 r:core_node2 x:collection1] o.a.s.h.IndexFetcher Could not retrie ve checksum from file. 2> java.lang.IllegalArgumentException: Seeking to negative position: MMapIndexInput(path="/opt/code/lusolr_trunk2/solr/build/solr-core/test/J0/temp/solr.cloud.ChaosMonkeySafeLeaderTest_1142D018587DF1D8-001/shard-2-001/cores/collection1/data/index/_0.fdx") 2>at org.apache.lucene.store.ByteBufferIndexInput$SingleBufferImpl.seek(ByteBufferIndexInput.java:408) 2>at org.apache.lucene.codecs.CodecUtil.retrieveChecksum(CodecUtil.java:400) 2>at org.apache.solr.handler.IndexFetcher.compareFile(IndexFetcher.java:876) 2>at org.apache.solr.handler.IndexFetcher.downloadIndexFiles(IndexFetcher.java:839) 2>at org.apache.solr.handler.IndexFetcher.fetchLatestIndex(IndexFetcher.java:437) 2>at org.apache.solr.handler.IndexFetcher.fetchLatestIndex(IndexFetcher.java:265) 2>at org.apache.solr.handler.ReplicationHandler.doFetch(ReplicationHandler.java:382) 2>at org.apache.solr.cloud.RecoveryStrategy.replicate(RecoveryStrategy.java:162) 2>at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:437) 2>at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:227) 2> Caused by: java.lang.IllegalArgumentException 2>at java.nio.Buffer.position(Buff
[jira] [Created] (SOLR-7980) CorruptIndex exceptions in ChaosMonkeySafeLeaderTest
Yonik Seeley created SOLR-7980: -- Summary: CorruptIndex exceptions in ChaosMonkeySafeLeaderTest Key: SOLR-7980 URL: https://issues.apache.org/jira/browse/SOLR-7980 Project: Solr Issue Type: Bug Reporter: Yonik Seeley Found during investigation of SOLR-7836 If you loop the test long enough, you'll occasionally see CorruptIndex exceptions, although it won't cause the test to fail. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7956) There are interrupts on shutdown in places that can cause ChannelAlreadyClosed exceptions which prevents proper closing of transaction logs.
[ https://issues.apache.org/jira/browse/SOLR-7956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715413#comment-14715413 ] Scott Blum commented on SOLR-7956: -- fsyncService.shutdownNow() in IndexFetcher.cleanup() seems particularly dangerous, as that executor's tasks dive directly into channel code > There are interrupts on shutdown in places that can cause > ChannelAlreadyClosed exceptions which prevents proper closing of transaction > logs. > > > Key: SOLR-7956 > URL: https://issues.apache.org/jira/browse/SOLR-7956 > Project: Solr > Issue Type: Bug >Reporter: Mark Miller >Assignee: Mark Miller > Fix For: Trunk, 5.4 > > Attachments: SOLR-7956.patch > > > Found this while beast testing HttpPartitionTest. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7956) There are interrupts on shutdown in places that can cause ChannelAlreadyClosed exceptions which prevents proper closing of transaction logs.
[ https://issues.apache.org/jira/browse/SOLR-7956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715407#comment-14715407 ] Scott Blum commented on SOLR-7956: -- I also see a bunch of calls to ExecutorUtil.shutdownNowAndAwaitTermination(), that seems pretty dangerous. It's difficult to audit exactly how those executors are being used, but it seems worrisome. > There are interrupts on shutdown in places that can cause > ChannelAlreadyClosed exceptions which prevents proper closing of transaction > logs. > > > Key: SOLR-7956 > URL: https://issues.apache.org/jira/browse/SOLR-7956 > Project: Solr > Issue Type: Bug >Reporter: Mark Miller >Assignee: Mark Miller > Fix For: Trunk, 5.4 > > Attachments: SOLR-7956.patch > > > Found this while beast testing HttpPartitionTest. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (SOLR-7977) SOLR_HOST in solr.in.sh doesn't apply to Jetty's host property
[ https://issues.apache.org/jira/browse/SOLR-7977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715395#comment-14715395 ] Shawn Heisey edited comment on SOLR-7977 at 8/26/15 7:49 PM: - bq. I can envision situations where the host that the user wants to include in zookeeper is entirely different In any sanely set up network, those two will of course always be identical ... but users have a way of inventing strange networking setups with address translation where not everything matches up the way it would in a straightforward network. Also, a user may want to specify the hostname to SolrCloud, but still listen on all interfaces, not just the one specified by the hostname. was (Author: elyograg): bq. I can envision situations where the host that the user wants to include in zookeeper is entirely different In any sanely set up network, those two will of course always be identical ... but users have a way of inventing strange networking setups with address translation where not everything matches up the way it would in a straightforward network. > SOLR_HOST in solr.in.sh doesn't apply to Jetty's host property > -- > > Key: SOLR-7977 > URL: https://issues.apache.org/jira/browse/SOLR-7977 > Project: Solr > Issue Type: Bug > Components: security, SolrCloud >Reporter: Shalin Shekhar Mangar > Labels: impact-medium > Fix For: Trunk, 5.4 > > > [~sdavids] pointed out that the SOLR_HOST config option in solr.in.sh doesn't > set Jetty's host property (solr.jetty.host) so it still binds to all net > interfaces. Perhaps it should apply to jetty as well because the user > explicitly wants us to bind to specific IP? -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6759) Integrate lat/long BKD and spatial 3d, part 2
[ https://issues.apache.org/jira/browse/LUCENE-6759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715397#comment-14715397 ] Karl Wright commented on LUCENE-6759: - ok, I'm finally done travelling for the moment. [~mikemccand], where do things stand? I notice that my latest patch didn't get committed, FWIW. Also, do you want me to implement your idea of having both isWithin's need to pass before the assert triggers? > Integrate lat/long BKD and spatial 3d, part 2 > - > > Key: LUCENE-6759 > URL: https://issues.apache.org/jira/browse/LUCENE-6759 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless > Attachments: LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch > > > This is just a continuation of LUCENE-6699, which became too big. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7977) SOLR_HOST in solr.in.sh doesn't apply to Jetty's host property
[ https://issues.apache.org/jira/browse/SOLR-7977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715395#comment-14715395 ] Shawn Heisey commented on SOLR-7977: bq. I can envision situations where the host that the user wants to include in zookeeper is entirely different In any sanely set up network, those two will of course always be identical ... but users have a way of inventing strange networking setups with address translation where not everything matches up the way it would in a straightforward network. > SOLR_HOST in solr.in.sh doesn't apply to Jetty's host property > -- > > Key: SOLR-7977 > URL: https://issues.apache.org/jira/browse/SOLR-7977 > Project: Solr > Issue Type: Bug > Components: security, SolrCloud >Reporter: Shalin Shekhar Mangar > Labels: impact-medium > Fix For: Trunk, 5.4 > > > [~sdavids] pointed out that the SOLR_HOST config option in solr.in.sh doesn't > set Jetty's host property (solr.jetty.host) so it still binds to all net > interfaces. Perhaps it should apply to jetty as well because the user > explicitly wants us to bind to specific IP? -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7956) There are interrupts on shutdown in places that can cause ChannelAlreadyClosed exceptions which prevents proper closing of transaction logs.
[ https://issues.apache.org/jira/browse/SOLR-7956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715392#comment-14715392 ] Scott Blum commented on SOLR-7956: -- Supposing I make the following adjustments: pending.cancel(true) => pending.cancel(false) scheduler.shutdownNow() => scheduler.shutdown(); Then, do I need to wait for the scheduler to actually terminate before exiting close()? > There are interrupts on shutdown in places that can cause > ChannelAlreadyClosed exceptions which prevents proper closing of transaction > logs. > > > Key: SOLR-7956 > URL: https://issues.apache.org/jira/browse/SOLR-7956 > Project: Solr > Issue Type: Bug >Reporter: Mark Miller >Assignee: Mark Miller > Fix For: Trunk, 5.4 > > Attachments: SOLR-7956.patch > > > Found this while beast testing HttpPartitionTest. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7956) There are interrupts on shutdown in places that can cause ChannelAlreadyClosed exceptions which prevents proper closing of transaction logs.
[ https://issues.apache.org/jira/browse/SOLR-7956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715371#comment-14715371 ] Yonik Seeley commented on SOLR-7956: bq. The ones in ReplicationHandler and CommitTracker seem suspect to me I think you're right... the one in CommitTracker does look suspect, esp since we now share the IndexWriter across UpdateHandlers on a core reload. > There are interrupts on shutdown in places that can cause > ChannelAlreadyClosed exceptions which prevents proper closing of transaction > logs. > > > Key: SOLR-7956 > URL: https://issues.apache.org/jira/browse/SOLR-7956 > Project: Solr > Issue Type: Bug >Reporter: Mark Miller >Assignee: Mark Miller > Fix For: Trunk, 5.4 > > Attachments: SOLR-7956.patch > > > Found this while beast testing HttpPartitionTest. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7956) There are interrupts on shutdown in places that can cause ChannelAlreadyClosed exceptions which prevents proper closing of transaction logs.
[ https://issues.apache.org/jira/browse/SOLR-7956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715320#comment-14715320 ] Scott Blum commented on SOLR-7956: -- Interrupts permanently breaking index writers is our #1 production issue right now, so I've just started looking at all the code and JIRAs related to this problem. Glad I found this one, I might try a backport to our ~5.2.1 instance. One question, have you also audited the existing Future.cancel(true) calls? I see several of these in Solr. The ones in ReplicationHandler and CommitTracker seem suspect to me, not sure about HttpShardHandler (maybe that one doesn't hit any core stuff directly). > There are interrupts on shutdown in places that can cause > ChannelAlreadyClosed exceptions which prevents proper closing of transaction > logs. > > > Key: SOLR-7956 > URL: https://issues.apache.org/jira/browse/SOLR-7956 > Project: Solr > Issue Type: Bug >Reporter: Mark Miller >Assignee: Mark Miller > Fix For: Trunk, 5.4 > > Attachments: SOLR-7956.patch > > > Found this while beast testing HttpPartitionTest. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-4638) If IndexWriter is interrupted on close and is using a channel (mmap/nio), it can throw a ClosedByInterruptException and prevent you from opening a new IndexWriter in t
[ https://issues.apache.org/jira/browse/LUCENE-4638?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715302#comment-14715302 ] Scott Blum commented on LUCENE-4638: Any update on this? In our mostly-stock 5.2.1 Solr deployment, we are hitting a point where we get cores into a permanently wedged state all the time, and there seems to be no fix except to restart the entire node (JVM). The IndexWriter gets into a broken state with ClosedByInterrupt, and it never gets out of it, and no new IndexWriter (maybe also no new searchers) can be created. This is one of our biggest operational issues right now. > If IndexWriter is interrupted on close and is using a channel (mmap/nio), it > can throw a ClosedByInterruptException and prevent you from opening a new > IndexWriter in the same proceses if you are using Native locks. > -- > > Key: LUCENE-4638 > URL: https://issues.apache.org/jira/browse/LUCENE-4638 > Project: Lucene - Core > Issue Type: Bug >Reporter: Mark Miller > Fix For: 4.9, Trunk > > Attachments: LUCENE-4638.patch > > > The ClosedByInterruptException will prevent the index from being unlocked in > close. If you try and close again, the call will hang. If you are using > native locks and try to open a new IndexWriter, it will fail to get the lock. > If you try IW#forceUnlock, it wont work because the not fully closed IW will > still have the lock. > ideas: > * On ClosedByInterruptException, IW should continue trying to close what it > can and unlock the index? Generally I have see the exception trigger in > commitInternal. > * We should add a non static forceUnlock to IW that lets you remove the lock > and start a new IW? > * We should make the lock protected so IW sub classes could unlock the index > in advanced use cases? > * others? -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7979) Typo in CoreAdminHandler log message
[ https://issues.apache.org/jira/browse/SOLR-7979?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mike Drob updated SOLR-7979: Attachment: SOLR-7979.patch Trivial patch. > Typo in CoreAdminHandler log message > > > Key: SOLR-7979 > URL: https://issues.apache.org/jira/browse/SOLR-7979 > Project: Solr > Issue Type: Bug >Reporter: Mike Drob >Priority: Trivial > Fix For: Trunk > > Attachments: SOLR-7979.patch > > > 2015-08-26 17:55:18,616 ERROR > org.apache.solr.handler.admin.CoreAdminHandler: Cound not find core to call > recovery: -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-7979) Typo in CoreAdminHandler log message
Mike Drob created SOLR-7979: --- Summary: Typo in CoreAdminHandler log message Key: SOLR-7979 URL: https://issues.apache.org/jira/browse/SOLR-7979 Project: Solr Issue Type: Bug Reporter: Mike Drob Priority: Trivial Fix For: Trunk Attachments: SOLR-7979.patch 2015-08-26 17:55:18,616 ERROR org.apache.solr.handler.admin.CoreAdminHandler: Cound not find core to call recovery: -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-7978) Really fix the example/files update-script Java version issues
Erik Hatcher created SOLR-7978: -- Summary: Really fix the example/files update-script Java version issues Key: SOLR-7978 URL: https://issues.apache.org/jira/browse/SOLR-7978 Project: Solr Issue Type: Bug Components: examples Affects Versions: 5.3 Reporter: Erik Hatcher Assignee: Erik Hatcher Fix For: 5.4 SOLR-7652 addressed this issue by having a Java7 version of the script for 5x and a Java8 version on trunk. 5x on Java8 is broken though. I wager that there's got to be some incantations that can make the same script work on Java 7 and 8. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7977) SOLR_HOST in solr.in.sh doesn't apply to Jetty's host property
[ https://issues.apache.org/jira/browse/SOLR-7977?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715243#comment-14715243 ] Shawn Heisey commented on SOLR-7977: I don't think that property was initially intended to be applicable to Jetty. It's how the hostname gets overridden for SolrCloud. I can envision situations where the host that the user wants to include in zookeeper is entirely different from the host they want to use for network binding. > SOLR_HOST in solr.in.sh doesn't apply to Jetty's host property > -- > > Key: SOLR-7977 > URL: https://issues.apache.org/jira/browse/SOLR-7977 > Project: Solr > Issue Type: Bug > Components: security, SolrCloud >Reporter: Shalin Shekhar Mangar > Labels: impact-medium > Fix For: Trunk, 5.4 > > > [~sdavids] pointed out that the SOLR_HOST config option in solr.in.sh doesn't > set Jetty's host property (solr.jetty.host) so it still binds to all net > interfaces. Perhaps it should apply to jetty as well because the user > explicitly wants us to bind to specific IP? -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6759) Integrate lat/long BKD and spatial 3d, part 2
[ https://issues.apache.org/jira/browse/LUCENE-6759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715191#comment-14715191 ] David Smiley commented on LUCENE-6759: -- bq. Maybe, we could just fix the test so that if isWithin differs between the quantized and unquantized x,y,z, we skip checking that hit? Yes; this is also the approach done RandomSpatialOpFuzzyPrefixTreeTest. The "fuzzy" in the name here because of the issue being discussed. > Integrate lat/long BKD and spatial 3d, part 2 > - > > Key: LUCENE-6759 > URL: https://issues.apache.org/jira/browse/LUCENE-6759 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless > Attachments: LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch > > > This is just a continuation of LUCENE-6699, which became too big. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7954) ArrayIndexOutOfBoundsException from distributed HLL serialization logic when using using stats.field={!cardinality=1.0} in a distributed query
[ https://issues.apache.org/jira/browse/SOLR-7954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14715188#comment-14715188 ] ASF subversion and git services commented on SOLR-7954: --- Commit 1697977 from hoss...@apache.org in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1697977 ] SOLR-7954: Fixed an integer overflow bug in the HyperLogLog code used by the 'cardinality' option of stats.field to prevent ArrayIndexOutOfBoundsException in a distributed search when a large precision is selected and a large number of values exist in each shard (merge r1697969) > ArrayIndexOutOfBoundsException from distributed HLL serialization logic when > using using stats.field={!cardinality=1.0} in a distributed query > -- > > Key: SOLR-7954 > URL: https://issues.apache.org/jira/browse/SOLR-7954 > Project: Solr > Issue Type: Bug >Affects Versions: 5.2.1 > Environment: SolrCloud 4 node cluster. > Ubuntu 12.04 > OS Type 64 bit >Reporter: Modassar Ather >Assignee: Hoss Man > Attachments: SOLR-7954.patch, SOLR-7954.patch, SOLR-7954.patch > > > User reports indicate that using {{stats.field=\{!cardinality=1.0\}foo}} on a > field that has extremely high cardinality on a single shard (example: 150K > unique values) can lead to "ArrayIndexOutOfBoundsException: 3" on the shard > during serialization of the HLL values. > using "cardinality=0.9" (or lower) doesn't produce the same symptoms, > suggesting the problem is specific to large log2m and regwidth values. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7971) Reduce memory allocated by JavaBinCodec to encode large strings
[ https://issues.apache.org/jira/browse/SOLR-7971?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shalin Shekhar Mangar updated SOLR-7971: Attachment: SOLR-7971-directbuffer.patch bq. I only glanced at it, but it's probably a little too simplistic. You can't cut UTF16 in random places, encode it as UTF8 and get the same bytes because of 2 char code points. Duh, of course, I used to know that once and yet... bq. We should really consider returning to how v1 of the protocol handled things since it had to do no buffering at all since it simply used String.length(). We just need to consider how to handle back compat of course. Until we go there, how about this patch which has a modified UTF16toUTF8 method that writes to ByteBuffer directly using an intermediate scratch array? > Reduce memory allocated by JavaBinCodec to encode large strings > --- > > Key: SOLR-7971 > URL: https://issues.apache.org/jira/browse/SOLR-7971 > Project: Solr > Issue Type: Sub-task > Components: Response Writers, SolrCloud >Reporter: Shalin Shekhar Mangar >Assignee: Shalin Shekhar Mangar >Priority: Minor > Fix For: Trunk, 5.4 > > Attachments: SOLR-7971-directbuffer.patch, > SOLR-7971-directbuffer.patch, SOLR-7971.patch > > > As discussed in SOLR-7927, we can reduce the buffer memory allocated by > JavaBinCodec while writing large strings. > https://issues.apache.org/jira/browse/SOLR-7927?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14700420#comment-14700420 > {quote} > The maximum Unicode code point (as of Unicode 8 anyway) is U+10 > ([http://www.unicode.org/glossary/#code_point]). This is encoded in UTF-16 > as surrogate pair {{\uDBFF\uDFFF}}, which takes up two Java chars, and is > represented in UTF-8 as the 4-byte sequence {{F4 8F BF BF}}. This is likely > where the mistaken 4-bytes-per-Java-char formulation came from: the maximum > number of UTF-8 bytes required to represent a Unicode *code point* is 4. > The maximum Java char is {{\u}}, which is represented in UTF-8 as the > 3-byte sequence {{EF BF BF}}. > So I think it's safe to switch to using 3 bytes per Java char (the unit of > measurement returned by {{String.length()}}), like > {{CompressingStoredFieldsWriter.writeField()}} does. > {quote} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7954) ArrayIndexOutOfBoundsException from distributed HLL serialization logic when using using stats.field={!cardinality=1.0} in a distributed query
[ https://issues.apache.org/jira/browse/SOLR-7954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14714473#comment-14714473 ] ASF subversion and git services commented on SOLR-7954: --- Commit 1697969 from hoss...@apache.org in branch 'dev/trunk' [ https://svn.apache.org/r1697969 ] SOLR-7954: Fixed an integer overflow bug in the HyperLogLog code used by the 'cardinality' option of stats.field to prevent ArrayIndexOutOfBoundsException in a distributed search when a large precision is selected and a large number of values exist in each shard > ArrayIndexOutOfBoundsException from distributed HLL serialization logic when > using using stats.field={!cardinality=1.0} in a distributed query > -- > > Key: SOLR-7954 > URL: https://issues.apache.org/jira/browse/SOLR-7954 > Project: Solr > Issue Type: Bug >Affects Versions: 5.2.1 > Environment: SolrCloud 4 node cluster. > Ubuntu 12.04 > OS Type 64 bit >Reporter: Modassar Ather >Assignee: Hoss Man > Attachments: SOLR-7954.patch, SOLR-7954.patch, SOLR-7954.patch > > > User reports indicate that using {{stats.field=\{!cardinality=1.0\}foo}} on a > field that has extremely high cardinality on a single shard (example: 150K > unique values) can lead to "ArrayIndexOutOfBoundsException: 3" on the shard > during serialization of the HLL values. > using "cardinality=0.9" (or lower) doesn't produce the same symptoms, > suggesting the problem is specific to large log2m and regwidth values. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7888) Make Lucene's AnalyzingInfixSuggester.lookup() method that takes a BooleanQuery filter parameter available in Solr
[ https://issues.apache.org/jira/browse/SOLR-7888?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Arcadius Ahouansou updated SOLR-7888: - Attachment: SOLR-7888.patch Updated from trunk and added {{testContextFilterIsTrimmed()}} > Make Lucene's AnalyzingInfixSuggester.lookup() method that takes a > BooleanQuery filter parameter available in Solr > -- > > Key: SOLR-7888 > URL: https://issues.apache.org/jira/browse/SOLR-7888 > Project: Solr > Issue Type: New Feature > Components: Suggester >Affects Versions: 5.2.1 >Reporter: Arcadius Ahouansou >Assignee: Jan Høydahl > Fix For: 5.4 > > Attachments: SOLR-7888.patch, SOLR-7888.patch, SOLR-7888.patch, > SOLR-7888.patch, SOLR-7888.patch, SOLR-7888.patch > > > LUCENE-6464 has introduced a very flexible lookup method that takes as > parameter a BooleanQuery that is used for filtering results. > This ticket is to expose that method to Solr. > This would allow user to do: > {code} > /suggest?suggest=true&suggest.build=true&suggest.q=term&suggest.contextFilterQuery=contexts:tennis > /suggest?suggest=true&suggest.build=true&suggest.q=term&suggest.contextFilterQuery=contexts:golf > AND contexts:football > {code} > etc > Given that the context filtering in currently only implemented by the > {code}AnalyzingInfixSuggester{code} and by the > {code}BlendedInfixSuggester{code}, this initial implementation will support > only these 2 lookup implementations. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6759) Integrate lat/long BKD and spatial 3d, part 2
[ https://issues.apache.org/jira/browse/LUCENE-6759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14714399#comment-14714399 ] Karl Wright commented on LUCENE-6759: - [~mikemccand] Yes, that should do it. If you have both, anyways. ;-) > Integrate lat/long BKD and spatial 3d, part 2 > - > > Key: LUCENE-6759 > URL: https://issues.apache.org/jira/browse/LUCENE-6759 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless > Attachments: LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch > > > This is just a continuation of LUCENE-6699, which became too big. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6750) TestMergeSchedulerExternal failure
[ https://issues.apache.org/jira/browse/LUCENE-6750?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14713616#comment-14713616 ] Steve Rowe commented on LUCENE-6750: New fail on Policeman Jenkins [http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-MacOSX/2623/], this time on branch_5x: {noformat} [junit4] Suite: org.apache.lucene.TestMergeSchedulerExternal [junit4] 2> NOTE: reproduce with: ant test -Dtestcase=TestMergeSchedulerExternal -Dtests.method=testSubclassConcurrentMergeScheduler -Dtests.seed=A2D285B04961A340 -Dtests.slow=true -Dtests.locale=it -Dtests.timezone=Pacific/Majuro -Dtests.asserts=true -Dtests.file.encoding=UTF-8 [junit4] FAILURE 0.57s J0 | TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler <<< [junit4]> Throwable #1: java.lang.AssertionError [junit4]>at __randomizedtesting.SeedInfo.seed([A2D285B04961A340:2553381D4D41D944]:0) [junit4]>at org.apache.lucene.TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler(TestMergeSchedulerExternal.java:116) [junit4]>at java.lang.Thread.run(Thread.java:745) [junit4] 2> NOTE: test params are: codec=Asserting(Lucene53): {id=PostingsFormat(name=MockRandom)}, docValues:{}, sim=DefaultSimilarity, locale=it, timezone=Pacific/Majuro [junit4] 2> NOTE: Mac OS X 10.8.5 x86_64/Oracle Corporation 1.8.0_60 (64-bit)/cpus=3,threads=1,free=368533088,total=518979584 [junit4] 2> NOTE: All tests run in this JVM: [TestLogMergePolicy, TestSameTokenSamePosition, TestCustomNorms, TestTerms, Test2BPostings, TestSpansEnum, TestBinaryDocValuesUpdates, TestDocValuesScoring, TestSparseFixedBitSet, TestByteArrayDataInput, FuzzyTermOnShortTermsTest, TestLucene50StoredFieldsFormat, TestIndexWriterDelete, TestTopDocsCollector, TestIndexWriter, TestWildcard, TestSimilarityProvider, TestBytesRefArray, TestQueryRescorer, TestSpans, TestFilterCachingPolicy, TestNot, TestSpanNotQuery, TestTermRangeFilter, TestScorerPerf, TestTermsEnum, TestApproximationSearchEquivalence, TestMinimize, TestSizeBoundedForceMerge, TestReadOnlyIndex, TestConjunctions, TestLongBitSet, TestMinShouldMatch2, TestLucene50SegmentInfoFormat, TestCharTermAttributeImpl, TestRecyclingIntBlockAllocator, TestQueryBuilder, TestPhrasePrefixQuery, TestIndexWriterNRTIsCurrent, TestNeverDelete, TestMatchNoDocsQuery, TestSegmentReader, TestIntsRef, TestBytesStore, TestPackedTokenAttributeImpl, TestTransactions, TestSimpleExplanationsOfNonMatches, TestFieldsReader, TestSearch, TestLucene50DocValuesFormat, Test2BPositions, TestForceMergeForever, TestMultiFields, TestNorms, TestFrequencyTrackingRingBuffer, TestFastCompressionMode, TestNRTReaderCleanup, TestBooleanMinShouldMatch, TestElevationComparator, TestLRUFilterCache, TestPerFieldPostingsFormat2, TestSloppyPhraseQuery2, TestBytesRefHash, TestIndexCommit, TestWindowsMMap, TestLucene50StoredFieldsFormatHighCompression, TestBinaryDocument, TestSleepingLockWrapper, TestNRTReaderWithThreads, TestNamedSPILoader, TestMultiLevelSkipList, TestDirectoryReaderReopen, TestConcurrentMergeScheduler, TestDateSort, TestBoolean2, TestMultiThreadTermVectors, TestRecyclingByteBlockAllocator, Test2BPostingsBytes, TestCachingWrapperFilter, TestMultiPhraseEnum, TestDuelingCodecsAtNight, TestPrefixQuery, TestMergeSchedulerExternal] {noformat} > TestMergeSchedulerExternal failure > -- > > Key: LUCENE-6750 > URL: https://issues.apache.org/jira/browse/LUCENE-6750 > Project: Lucene - Core > Issue Type: Bug >Affects Versions: Trunk, 5.4 >Reporter: Steve Rowe > > Policeman Jenkins found a failure on OS X > [http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-MacOSX/2649/] that I can't > reproduce on OS X 10.10.4 using Oracle Java 1.8.0_20, even after beasting 200 > total suite iterations with the seed: > {noformat} >[junit4] Suite: org.apache.lucene.TestMergeSchedulerExternal >[junit4] 2> NOTE: reproduce with: ant test > -Dtestcase=TestMergeSchedulerExternal > -Dtests.method=testSubclassConcurrentMergeScheduler > -Dtests.seed=3AF868F9E00E5EBA -Dtests.slow=true -Dtests.locale=ru > -Dtests.timezone=Europe/London -Dtests.asserts=true > -Dtests.file.encoding=ISO-8859-1 >[junit4] FAILURE 0.37s J1 | > TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler <<< >[junit4]> Throwable #1: java.lang.AssertionError >[junit4]> at > __randomizedtesting.SeedInfo.seed([3AF868F9E00E5EBA:BD79D554E42E24BE]:0) >[junit4]> at > org.apache.lucene.TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler(TestMergeSchedulerExternal.java:116) >[junit4]> at java.lang.Thread.run(Thread.java:745) >[junit4] 2> NOTE: test params are: codec=Asserting(Lucene53): > {id=PostingsFormat(name=Memory doPackFST= true)}, docValues:{},
[jira] [Updated] (LUCENE-6750) TestMergeSchedulerExternal failure
[ https://issues.apache.org/jira/browse/LUCENE-6750?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Steve Rowe updated LUCENE-6750: --- Affects Version/s: 5.4 Trunk > TestMergeSchedulerExternal failure > -- > > Key: LUCENE-6750 > URL: https://issues.apache.org/jira/browse/LUCENE-6750 > Project: Lucene - Core > Issue Type: Bug >Affects Versions: Trunk, 5.4 >Reporter: Steve Rowe > > Policeman Jenkins found a failure on OS X > [http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-MacOSX/2649/] that I can't > reproduce on OS X 10.10.4 using Oracle Java 1.8.0_20, even after beasting 200 > total suite iterations with the seed: > {noformat} >[junit4] Suite: org.apache.lucene.TestMergeSchedulerExternal >[junit4] 2> NOTE: reproduce with: ant test > -Dtestcase=TestMergeSchedulerExternal > -Dtests.method=testSubclassConcurrentMergeScheduler > -Dtests.seed=3AF868F9E00E5EBA -Dtests.slow=true -Dtests.locale=ru > -Dtests.timezone=Europe/London -Dtests.asserts=true > -Dtests.file.encoding=ISO-8859-1 >[junit4] FAILURE 0.37s J1 | > TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler <<< >[junit4]> Throwable #1: java.lang.AssertionError >[junit4]> at > __randomizedtesting.SeedInfo.seed([3AF868F9E00E5EBA:BD79D554E42E24BE]:0) >[junit4]> at > org.apache.lucene.TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler(TestMergeSchedulerExternal.java:116) >[junit4]> at java.lang.Thread.run(Thread.java:745) >[junit4] 2> NOTE: test params are: codec=Asserting(Lucene53): > {id=PostingsFormat(name=Memory doPackFST= true)}, docValues:{}, > sim=DefaultSimilarity, locale=ru, timezone=Europe/London >[junit4] 2> NOTE: Mac OS X 10.8.5 x86_64/Oracle Corporation 1.8.0_51 > (64-bit)/cpus=3,threads=1,free=16232544,total=54853632 >[junit4] 2> NOTE: All tests run in this JVM: [TestDateSort, > TestWildcardRandom, TestIndexWriterMergePolicy, TestPackedInts, > TestSpansAdvanced, TestBooleanOr, TestParallelReaderEmptyIndex, > TestFixedBitDocIdSet, TestIndexWriterDeleteByQuery, Test4GBStoredFields, > TestMultiThreadTermVectors, TestIndexWriterConfig, TestToken, > TestMergeSchedulerExternal] >[junit4] Completed [21/401] on J1 in 0.39s, 2 tests, 1 failure <<< > FAILURES! > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6759) Integrate lat/long BKD and spatial 3d, part 2
[ https://issues.apache.org/jira/browse/LUCENE-6759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14713598#comment-14713598 ] Michael McCandless commented on LUCENE-6759: Oh wait, we need to do more than simply disable the assert, because the test will still fail, just a bit later when it verifies all hits (the assert was just "early detection"): {noformat} [junit4] Suite: org.apache.lucene.bkdtree3d.TestGeo3DPointField [junit4] 2> Aug 26, 2015 8:48:39 AM com.carrotsearch.randomizedtesting.RandomizedRunner$QueueUncaughtExceptionsHandler uncaughtException [junit4] 2> WARNING: Uncaught exception in thread: Thread[T0,5,TGRP-TestGeo3DPointField] [junit4] 2> java.lang.AssertionError: T0: iter=63 id=71226 docID=71226 lat=-0.004763555725376775 lon=-0.0076479587074575126 expected false but got: true deleted?=false [junit4] 2> point1=[lat=-0.004763555725376775, lon=-0.0076479587074575126], iswithin=true [junit4] 2> point2=[X=1.0010781049211872, Y=-0.007656353133570567, Z=-0.0047688666958216885], iswithin=false [junit4] 2> query=PointInGeo3DShapeQuery: field=point:PlanetModel: PlanetModel.WGS84 Shape: GeoCircle: {planetmodel=PlanetModel.WGS84, center=[lat=-7.573175600018171E-4, lon=-0.001184769535031697], radius=0.007585721238160122(0.4346298115093282)} [junit4] 2>at __randomizedtesting.SeedInfo.seed([D75138C6C25D1BCF]:0) [junit4] 2>at org.junit.Assert.fail(Assert.java:93) [junit4] 2>at org.apache.lucene.bkdtree3d.TestGeo3DPointField$4._run(TestGeo3DPointField.java:625) [junit4] 2>at org.apache.lucene.bkdtree3d.TestGeo3DPointField$4.run(TestGeo3DPointField.java:521) [junit4] 2> [junit4] 2> Aug 26, 2015 8:48:40 AM com.carrotsearch.randomizedtesting.RandomizedRunner$QueueUncaughtExceptionsHandler uncaughtException [junit4] 2> WARNING: Uncaught exception in thread: Thread[T2,5,TGRP-TestGeo3DPointField] [junit4] 2> java.lang.AssertionError: T2: iter=62 id=71226 docID=71226 lat=-0.004763555725376775 lon=-0.0076479587074575126 expected false but got: true deleted?=false [junit4] 2> point1=[lat=-0.004763555725376775, lon=-0.0076479587074575126], iswithin=true [junit4] 2> point2=[X=1.0010781049211872, Y=-0.007656353133570567, Z=-0.0047688666958216885], iswithin=false [junit4] 2> query=PointInGeo3DShapeQuery: field=point:PlanetModel: PlanetModel.WGS84 Shape: GeoCircle: {planetmodel=PlanetModel.WGS84, center=[lat=-7.573175600018171E-4, lon=-0.001184769535031697], radius=0.007585721238160122(0.4346298115093282)} [junit4] 2>at __randomizedtesting.SeedInfo.seed([D75138C6C25D1BCF]:0) [junit4] 2>at org.junit.Assert.fail(Assert.java:93) [junit4] 2>at org.apache.lucene.bkdtree3d.TestGeo3DPointField$4._run(TestGeo3DPointField.java:625) [junit4] 2>at org.apache.lucene.bkdtree3d.TestGeo3DPointField$4.run(TestGeo3DPointField.java:521) [junit4] 2> [junit4] 2> NOTE: reproduce with: ant test -Dtestcase=TestGeo3DPointField -Dtests.method=testRandomMedium -Dtests.seed=D75138C6C25D1BCF -Dtests.multiplier=10 -Dtests.slow=true -Dtests.locale=de_GR -Dtests.timezone=America/Managua -Dtests.asserts=true -Dtests.file.encoding=US-ASCII [junit4] ERROR 62.4s | TestGeo3DPointField.testRandomMedium <<< {noformat} So how to correspondingly fix the test? Right now, in (intentionally) quantizes the double x,y,z of the point to match what the doc values pack/unpack did ... Maybe, we could just fix the test so that if isWithin differs between the quantized and unquantized x,y,z, we skip checking that hit? > Integrate lat/long BKD and spatial 3d, part 2 > - > > Key: LUCENE-6759 > URL: https://issues.apache.org/jira/browse/LUCENE-6759 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless > Attachments: LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch > > > This is just a continuation of LUCENE-6699, which became too big. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-6764) Payloads should be compressed
Adrien Grand created LUCENE-6764: Summary: Payloads should be compressed Key: LUCENE-6764 URL: https://issues.apache.org/jira/browse/LUCENE-6764 Project: Lucene - Core Issue Type: Improvement Reporter: Adrien Grand Priority: Minor I think we should at least try to do something simple, eg. deduplicate or apply simple LZ77 compression. For instance if you use enclosing html tags to give different weights to individual terms, there might be lots of repetitions as there are not that many unique html tags. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7569) Create an API to force a leader election between nodes
[ https://issues.apache.org/jira/browse/SOLR-7569?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ishan Chattopadhyaya updated SOLR-7569: --- Attachment: SOLR-7569.patch Adding a wait for recoveries to finish after the recovery operation in the test. > Create an API to force a leader election between nodes > -- > > Key: SOLR-7569 > URL: https://issues.apache.org/jira/browse/SOLR-7569 > Project: Solr > Issue Type: New Feature > Components: SolrCloud >Reporter: Shalin Shekhar Mangar >Assignee: Shalin Shekhar Mangar > Labels: difficulty-medium, impact-high > Attachments: SOLR-7569.patch, SOLR-7569.patch, SOLR-7569.patch, > SOLR-7569.patch, SOLR-7569_lir_down_state_test.patch > > > There are many reasons why Solr will not elect a leader for a shard e.g. all > replicas' last published state was recovery or due to bugs which cause a > leader to be marked as 'down'. While the best solution is that they never get > into this state, we need a manual way to fix this when it does get into this > state. Right now we can do a series of dance involving bouncing the node > (since recovery paths between bouncing and REQUESTRECOVERY are different), > but that is difficult when running a large cluster. Although it is possible > that such a manual API may lead to some data loss but in some cases, it is > the only possible option to restore availability. > This issue proposes to build a new collection API which can be used to force > replicas into recovering a leader while avoiding data loss on a best effort > basis. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7971) Reduce memory allocated by JavaBinCodec to encode large strings
[ https://issues.apache.org/jira/browse/SOLR-7971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14713470#comment-14713470 ] Yonik Seeley commented on SOLR-7971: bq. limit the intermediate on-heap buffer to 64KB only I only glanced at it, but it's probably a little too simplistic. You can't cut UTF16 in random places, encode it as UTF8 and get the same bytes because of 2 char code points. bq. I've got that buffering is necessary just because we need to calculate length of encoded bytes for starting tag Yeah, the length is really the only reason we need to buffer and copy. We should really consider returning to how v1 of the protocol handled things since it had to do no buffering at all since it simply used String.length(). We just need to consider how to handle back compat of course. > Reduce memory allocated by JavaBinCodec to encode large strings > --- > > Key: SOLR-7971 > URL: https://issues.apache.org/jira/browse/SOLR-7971 > Project: Solr > Issue Type: Sub-task > Components: Response Writers, SolrCloud >Reporter: Shalin Shekhar Mangar >Assignee: Shalin Shekhar Mangar >Priority: Minor > Fix For: Trunk, 5.4 > > Attachments: SOLR-7971-directbuffer.patch, SOLR-7971.patch > > > As discussed in SOLR-7927, we can reduce the buffer memory allocated by > JavaBinCodec while writing large strings. > https://issues.apache.org/jira/browse/SOLR-7927?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14700420#comment-14700420 > {quote} > The maximum Unicode code point (as of Unicode 8 anyway) is U+10 > ([http://www.unicode.org/glossary/#code_point]). This is encoded in UTF-16 > as surrogate pair {{\uDBFF\uDFFF}}, which takes up two Java chars, and is > represented in UTF-8 as the 4-byte sequence {{F4 8F BF BF}}. This is likely > where the mistaken 4-bytes-per-Java-char formulation came from: the maximum > number of UTF-8 bytes required to represent a Unicode *code point* is 4. > The maximum Java char is {{\u}}, which is represented in UTF-8 as the > 3-byte sequence {{EF BF BF}}. > So I think it's safe to switch to using 3 bytes per Java char (the unit of > measurement returned by {{String.length()}}), like > {{CompressingStoredFieldsWriter.writeField()}} does. > {quote} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7885) Add support for loading HTTP resources
[ https://issues.apache.org/jira/browse/SOLR-7885?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14713468#comment-14713468 ] Aaron LaBella commented on SOLR-7885: - Can someone take a look at this patch and commit if approved? Thanks. > Add support for loading HTTP resources > -- > > Key: SOLR-7885 > URL: https://issues.apache.org/jira/browse/SOLR-7885 > Project: Solr > Issue Type: Improvement > Components: contrib - DataImportHandler, SolrJ >Affects Versions: 5.3 >Reporter: Aaron LaBella > Attachments: SOLR-7885-1.patch, SOLR-7885-2.patch > > > I have a need to be able to load data import handler configuration files from > an HTTP server instead of the local file system. So, I modified > {code}org.apache.solr.core.SolrResourceLoader{code} and some of the > respective dataimport files in {code}org.apache.solr.handler.dataimport{code} > to be able to support doing this. > {code}solrconfig.xml{code} now has the option to define a parameter: > *configRemote*, and if defined (and it's an HTTP(s) URL), it'll attempt to > load the resource. If successfully, it'll also persist the resource to the > local file system so that it is available on a solr server restart per chance > that the remote resource is currently unavailable. > Lastly, to be consistent with the pattern that already exists in > SolrResourceLoader, this feature is *disabled* by default, and requires the > setting of an additional JVM property: > {code}-Dsolr.allow.http.resourceloading=true{code}. > Please review and let me know if there is anything else that needs to be done > in order for this patch to make the next release. As far as I can tell, it's > fully tested and ready to go. > Thanks. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7922) JSON API facet doesnt return facet with attribute that equals to 0
[ https://issues.apache.org/jira/browse/SOLR-7922?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14713451#comment-14713451 ] Yaniv Hemi commented on SOLR-7922: -- Hi, this issue wasn't added into Solr 5.3? > JSON API facet doesnt return facet with attribute that equals to 0 > -- > > Key: SOLR-7922 > URL: https://issues.apache.org/jira/browse/SOLR-7922 > Project: Solr > Issue Type: Bug > Components: Facet Module >Affects Versions: 5.3 >Reporter: Yaniv Hemi >Assignee: Yonik Seeley >Priority: Critical > Labels: facet, json, jsonapi > Attachments: SOLR-7922.patch > > > regular facet returns "0",33739, but JSON Facet API returns all values and > counts without "0". > see the example: > {code:json} > { > "responseHeader":{ > "status":0, > "QTime":9, > "params":{ > "q":"*:*", > "json.facet":"{\n\tfacetForMeta_i_interactionSentiment: {\t\t\ntype : > terms, \n\t\tfield : Meta_i_interactionSentiment\n\t}\n}", > "facet.field":"Meta_i_interactionSentiment", > "indent":"true", > "fq":["channel:TelcoDefaultChannel", > "content_type:PARENT"], > "rows":"0", > "wt":"json", > "facet":"true"}}, > "response":{"numFound":167857,"start":0,"maxScore":1.0,"docs":[] > }, > "facet_counts":{ > "facet_queries":{}, > "facet_fields":{ > "Meta_i_interactionSentiment":[ > "-1",33743, > "0",33739, > "-2",33499, > "2",33451, > "1",33425]}, > "facet_dates":{}, > "facet_ranges":{}, > "facet_intervals":{}, > "facet_heatmaps":{}}, > "facets":{ > "count":167857, > "facetForMeta_i_interactionSentiment":{ > "buckets":[{ > "val":-1, > "count":33743}, > { > "val":-2, > "count":33499}, > { > "val":2, > "count":33451}, > { > "val":1, > "count":33425}]}}} > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Should we EOL support for 32 bit systems?
On 8/25/2015 10:23 AM, Erick Erickson wrote: > I have no real skin in this game, but I thought it worth asking after > Uwe's recent e-mail about disabling 32bits with -server tests. > > I guess it boils down to "who is using 32-bit versions?". Are we > spending time/energy supporting a configuration that is not useful to > enough people to merit the effort? > > I'm perfectly content if the response is "That's a really stupid > question to ask, of course we must continue to support 32-bit OSs". > Although some evidence would be nice ;) > > It's just that nobody I work with is running 32 bit OS's. Whether > that's just my limited exposure to people running small systems is > certainly a valid question. As long as Oracle has 32-bit versions of Java available for easy download from the main website, we probably need to keep testing and supporting it. Lucene and Solr won't scale on a 32-bit Java, but I think there are still plenty of people who start there, and some who actually run it on 32-bit. That said, I think we do need to be thinking about dropping support in the future, even though we can't really do so yet. Intel stopped mass-producing 32-bit chips for the server market in 2005, and stopped mass production of 32-bit chips for the consumer market in 2006. For nearly the last decade, it has been very difficult to buy a computer incapable of 64-bit operation. Since Vista and Windows 7 came on the scene, Microsoft has been pushing 64-bit client operating systems. Server 2008R2 and Server 2012 are only available in 64-bit editions. Macs have been 64-bit for a VERY long time. I think the biggest market for 32-bit Java is browsers on Windows. Virtually all installs of Firefox and Chrome for Windows are 32-bit, and require the 32-bit Java. I bet that if the major browser vendors were to all put out 64-bit versions on their main download links, downloads of 32-bit Java would begin to dwindle rapidly. On Windows 10, it looks like Microsoft has finally gone 64-bit by default with the Edge browser. This might force the others to follow suit. When that happens, I think Oracle may strongly consider dropping 32-bit support from the next major Java version ... and even if they don't do it at that time, they probably will do so on the next major version after that. I just went to java.com with Microsoft Edge. It says "In Windows 10, the Edge browser does not support plug-ins and therefore will not run Java. Switch to a different browser (Firefox or Internet Explorer 11) to run the Java plug-in." They are not going to be helpful in pushing 64-bit Java. :) Thanks, Shawn - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7971) Reduce memory allocated by JavaBinCodec to encode large strings
[ https://issues.apache.org/jira/browse/SOLR-7971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14713437#comment-14713437 ] Shalin Shekhar Mangar commented on SOLR-7971: - bq. couldn't it turn that calling frequent call allocateDirect()&clear() takes too much time? in this case isn't it worth to reuse directBuffer across writeStr() calls as JavaBinCodec's field. Yes, allocateDirect() can be slower and we should reuse the buffer as much as possible. This was just an idea as a patch. I don't intend to commit it as it is. bq. I've got that buffering is necessary just because we need to calculate length of encoded bytes for starting tag, is it a big problem if we loop ByteUtils.UTF16toUTF8() twice, the first time to calculate the length and completely dropping the content, then writing content in the second time loop. Hmm, interesting idea. We could also have a method calcUTF16toUTF8Length which avoids all the bitwise operators and just returns the required length. bq. just curious, how much efforts does it take to extend javabin format by http-like chunks? It should be possible. We'll need a new chunked type and an upgrade to the JavaBin version. Or we may be able to get away with modifying only the LogCodec in TransactionLog. > Reduce memory allocated by JavaBinCodec to encode large strings > --- > > Key: SOLR-7971 > URL: https://issues.apache.org/jira/browse/SOLR-7971 > Project: Solr > Issue Type: Sub-task > Components: Response Writers, SolrCloud >Reporter: Shalin Shekhar Mangar >Assignee: Shalin Shekhar Mangar >Priority: Minor > Fix For: Trunk, 5.4 > > Attachments: SOLR-7971-directbuffer.patch, SOLR-7971.patch > > > As discussed in SOLR-7927, we can reduce the buffer memory allocated by > JavaBinCodec while writing large strings. > https://issues.apache.org/jira/browse/SOLR-7927?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14700420#comment-14700420 > {quote} > The maximum Unicode code point (as of Unicode 8 anyway) is U+10 > ([http://www.unicode.org/glossary/#code_point]). This is encoded in UTF-16 > as surrogate pair {{\uDBFF\uDFFF}}, which takes up two Java chars, and is > represented in UTF-8 as the 4-byte sequence {{F4 8F BF BF}}. This is likely > where the mistaken 4-bytes-per-Java-char formulation came from: the maximum > number of UTF-8 bytes required to represent a Unicode *code point* is 4. > The maximum Java char is {{\u}}, which is represented in UTF-8 as the > 3-byte sequence {{EF BF BF}}. > So I think it's safe to switch to using 3 bytes per Java char (the unit of > measurement returned by {{String.length()}}), like > {{CompressingStoredFieldsWriter.writeField()}} does. > {quote} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7971) Reduce memory allocated by JavaBinCodec to encode large strings
[ https://issues.apache.org/jira/browse/SOLR-7971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14713398#comment-14713398 ] Mikhail Khludnev commented on SOLR-7971: Shalin, * couldn't it turn that calling frequent call allocateDirect()&clear() takes too much time? in this case isn't it worth to reuse directBuffer across writeStr() calls as JavaBinCodec's field. * I've got that buffering is necessary just because we need to calculate length of encoded bytes for starting tag, is it a big problem if we loop ByteUtils.UTF16toUTF8() twice, the first time to calculate the length and completely dropping the content, then writing content in the second time loop. * just curious, how much efforts does it take to extend javabin format by http-like chunks? > Reduce memory allocated by JavaBinCodec to encode large strings > --- > > Key: SOLR-7971 > URL: https://issues.apache.org/jira/browse/SOLR-7971 > Project: Solr > Issue Type: Sub-task > Components: Response Writers, SolrCloud >Reporter: Shalin Shekhar Mangar >Assignee: Shalin Shekhar Mangar >Priority: Minor > Fix For: Trunk, 5.4 > > Attachments: SOLR-7971-directbuffer.patch, SOLR-7971.patch > > > As discussed in SOLR-7927, we can reduce the buffer memory allocated by > JavaBinCodec while writing large strings. > https://issues.apache.org/jira/browse/SOLR-7927?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14700420#comment-14700420 > {quote} > The maximum Unicode code point (as of Unicode 8 anyway) is U+10 > ([http://www.unicode.org/glossary/#code_point]). This is encoded in UTF-16 > as surrogate pair {{\uDBFF\uDFFF}}, which takes up two Java chars, and is > represented in UTF-8 as the 4-byte sequence {{F4 8F BF BF}}. This is likely > where the mistaken 4-bytes-per-Java-char formulation came from: the maximum > number of UTF-8 bytes required to represent a Unicode *code point* is 4. > The maximum Java char is {{\u}}, which is represented in UTF-8 as the > 3-byte sequence {{EF BF BF}}. > So I think it's safe to switch to using 3 bytes per Java char (the unit of > measurement returned by {{String.length()}}), like > {{CompressingStoredFieldsWriter.writeField()}} does. > {quote} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [JENKINS] Lucene-Solr-5.x-MacOSX (64bit/jdk1.8.0) - Build # 2623 - Still Failing!
It's a sign. Your license for the Holy Apple has expired... Dawid On Wed, Aug 26, 2015 at 3:08 PM, Michael McCandless wrote: > This is https://issues.apache.org/jira/browse/LUCENE-6750 > > I don't see why the test fails, and it won't fail on beasting for me on Linux. > > It seems to always fail on OS X ... > > Mike McCandless > > http://blog.mikemccandless.com > > > On Wed, Aug 26, 2015 at 4:52 AM, Policeman Jenkins Server > wrote: >> Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-MacOSX/2623/ >> Java: 64bit/jdk1.8.0 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC >> >> 1 tests failed. >> FAILED: >> org.apache.lucene.TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler >> >> Error Message: >> >> >> Stack Trace: >> java.lang.AssertionError >> at >> __randomizedtesting.SeedInfo.seed([A2D285B04961A340:2553381D4D41D944]:0) >> at org.junit.Assert.fail(Assert.java:92) >> at org.junit.Assert.assertTrue(Assert.java:43) >> at org.junit.Assert.assertTrue(Assert.java:54) >> at >> org.apache.lucene.TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler(TestMergeSchedulerExternal.java:116) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:497) >> at >> com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627) >> at >> com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836) >> at >> com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872) >> at >> com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886) >> at >> org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) >> at >> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) >> at >> org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) >> at >> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) >> at >> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) >> at >> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) >> at >> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) >> at >> com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) >> at >> com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) >> at >> com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845) >> at >> com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747) >> at >> com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781) >> at >> com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792) >> at >> org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) >> at >> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) >> at >> org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) >> at >> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) >> at >> com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) >> at >> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) >> at >> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) >> at >> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) >> at >> org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) >> at >> org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) >> at >> org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) >> at >> org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55) >> at >> com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) >> at >> com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunn
Re: [JENKINS] Lucene-Solr-5.x-MacOSX (64bit/jdk1.8.0) - Build # 2623 - Still Failing!
This is https://issues.apache.org/jira/browse/LUCENE-6750 I don't see why the test fails, and it won't fail on beasting for me on Linux. It seems to always fail on OS X ... Mike McCandless http://blog.mikemccandless.com On Wed, Aug 26, 2015 at 4:52 AM, Policeman Jenkins Server wrote: > Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-MacOSX/2623/ > Java: 64bit/jdk1.8.0 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC > > 1 tests failed. > FAILED: > org.apache.lucene.TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler > > Error Message: > > > Stack Trace: > java.lang.AssertionError > at > __randomizedtesting.SeedInfo.seed([A2D285B04961A340:2553381D4D41D944]:0) > at org.junit.Assert.fail(Assert.java:92) > at org.junit.Assert.assertTrue(Assert.java:43) > at org.junit.Assert.assertTrue(Assert.java:54) > at > org.apache.lucene.TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler(TestMergeSchedulerExternal.java:116) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886) > at > org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) > at > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) > at > org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) > at > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) > at > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) > at > com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) > at > com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) > at > com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781) > at > com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792) > at > org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) > at > org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) > at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > at > com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) > at > org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) > at > org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) > at > org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) > at > org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55) > at > com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) > at > com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) > at java.lang.Thread.run(Thread.java:745) > > > > > Build Log: > [...truncated 1086 lines...] >[junit4] Suite: org.apache.lucene.TestMergeSchedulerExternal >[junit4] 2> NOTE: reprod
[jira] [Commented] (SOLR-7888) Make Lucene's AnalyzingInfixSuggester.lookup() method that takes a BooleanQuery filter parameter available in Solr
[ https://issues.apache.org/jira/browse/SOLR-7888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14713078#comment-14713078 ] Jan Høydahl commented on SOLR-7888: --- I think this is close to committable. If there are no objections to moving {{CONTEXTS_FIELD_NAME}} to {{Lookup.java}} by tomorrow, I'll do a round of final reviews & manual testing before committing. > Make Lucene's AnalyzingInfixSuggester.lookup() method that takes a > BooleanQuery filter parameter available in Solr > -- > > Key: SOLR-7888 > URL: https://issues.apache.org/jira/browse/SOLR-7888 > Project: Solr > Issue Type: New Feature > Components: Suggester >Affects Versions: 5.2.1 >Reporter: Arcadius Ahouansou >Assignee: Jan Høydahl > Fix For: 5.4 > > Attachments: SOLR-7888.patch, SOLR-7888.patch, SOLR-7888.patch, > SOLR-7888.patch, SOLR-7888.patch > > > LUCENE-6464 has introduced a very flexible lookup method that takes as > parameter a BooleanQuery that is used for filtering results. > This ticket is to expose that method to Solr. > This would allow user to do: > {code} > /suggest?suggest=true&suggest.build=true&suggest.q=term&suggest.contextFilterQuery=contexts:tennis > /suggest?suggest=true&suggest.build=true&suggest.q=term&suggest.contextFilterQuery=contexts:golf > AND contexts:football > {code} > etc > Given that the context filtering in currently only implemented by the > {code}AnalyzingInfixSuggester{code} and by the > {code}BlendedInfixSuggester{code}, this initial implementation will support > only these 2 lookup implementations. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [JENKINS] Lucene-Solr-SmokeRelease-5.x - Build # 302 - Still Failing
Thanks Shalin. Mike McCandless http://blog.mikemccandless.com On Wed, Aug 26, 2015 at 8:57 AM, Shalin Shekhar Mangar wrote: > Noble hasn't completed the post install steps which add backwards > compatibility tests for the 5.3.0 release. Since he is on vacation for > a few days, I'll add those indexes to avoid this failure. > > On Wed, Aug 26, 2015 at 8:59 AM, Apache Jenkins Server > wrote: >> Build: https://builds.apache.org/job/Lucene-Solr-SmokeRelease-5.x/302/ >> >> No tests ran. >> >> Build Log: >> [...truncated 52844 lines...] >> prepare-release-no-sign: >> [mkdir] Created dir: >> /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/lucene/build/smokeTestRelease/dist >> [copy] Copying 461 files to >> /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/lucene/build/smokeTestRelease/dist/lucene >> [copy] Copying 245 files to >> /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/lucene/build/smokeTestRelease/dist/solr >>[smoker] Java 1.7 >> JAVA_HOME=/home/jenkins/jenkins-slave/tools/hudson.model.JDK/latest1.7 >>[smoker] Java 1.8 >> JAVA_HOME=/home/jenkins/jenkins-slave/tools/hudson.model.JDK/latest1.8 >>[smoker] NOTE: output encoding is UTF-8 >>[smoker] >>[smoker] Load release URL >> "file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/lucene/build/smokeTestRelease/dist/"... >>[smoker] >>[smoker] Test Lucene... >>[smoker] test basics... >>[smoker] get KEYS >>[smoker] 0.1 MB in 0.01 sec (12.1 MB/sec) >>[smoker] check changes HTML... >>[smoker] download lucene-5.4.0-src.tgz... >>[smoker] 28.5 MB in 0.04 sec (708.0 MB/sec) >>[smoker] verify md5/sha1 digests >>[smoker] download lucene-5.4.0.tgz... >>[smoker] 65.8 MB in 0.09 sec (716.3 MB/sec) >>[smoker] verify md5/sha1 digests >>[smoker] download lucene-5.4.0.zip... >>[smoker] 76.0 MB in 0.11 sec (690.8 MB/sec) >>[smoker] verify md5/sha1 digests >>[smoker] unpack lucene-5.4.0.tgz... >>[smoker] verify JAR metadata/identity/no javax.* or java.* classes... >>[smoker] test demo with 1.7... >>[smoker] got 6063 hits for query "lucene" >>[smoker] checkindex with 1.7... >>[smoker] test demo with 1.8... >>[smoker] got 6063 hits for query "lucene" >>[smoker] checkindex with 1.8... >>[smoker] check Lucene's javadoc JAR >>[smoker] unpack lucene-5.4.0.zip... >>[smoker] verify JAR metadata/identity/no javax.* or java.* classes... >>[smoker] test demo with 1.7... >>[smoker] got 6063 hits for query "lucene" >>[smoker] checkindex with 1.7... >>[smoker] test demo with 1.8... >>[smoker] got 6063 hits for query "lucene" >>[smoker] checkindex with 1.8... >>[smoker] check Lucene's javadoc JAR >>[smoker] unpack lucene-5.4.0-src.tgz... >>[smoker] make sure no JARs/WARs in src dist... >>[smoker] run "ant validate" >>[smoker] run tests w/ Java 7 and testArgs='-Dtests.slow=false'... >>[smoker] test demo with 1.7... >>[smoker] got 213 hits for query "lucene" >>[smoker] checkindex with 1.7... >>[smoker] generate javadocs w/ Java 7... >>[smoker] >>[smoker] Crawl/parse... >>[smoker] >>[smoker] Verify... >>[smoker] run tests w/ Java 8 and testArgs='-Dtests.slow=false'... >>[smoker] test demo with 1.8... >>[smoker] got 213 hits for query "lucene" >>[smoker] checkindex with 1.8... >>[smoker] generate javadocs w/ Java 8... >>[smoker] >>[smoker] Crawl/parse... >>[smoker] >>[smoker] Verify... >>[smoker] confirm all releases have coverage in >> TestBackwardsCompatibility >>[smoker] find all past Lucene releases... >>[smoker] run TestBackwardsCompatibility.. >>[smoker] Releases that don't seem to be tested: >>[smoker] 5.3.0 >>[smoker] Traceback (most recent call last): >>[smoker] File >> "/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/dev-tools/scripts/smokeTestRelease.py", >> line 1449, in >>[smoker] main() >>[smoker] File >> "/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/dev-tools/scripts/smokeTestRelease.py", >> line 1394, in main >>[smoker] smokeTest(c.java, c.url, c.revision, c.version, c.tmp_dir, >> c.is_signed, ' '.join(c.test_args)) >>[smoker] File >> "/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/dev-tools/scripts/smokeTestRelease.py", >> line 1432, in smokeTest >>[smoker] unpackAndVerify(java, 'lucene', tmpDir, 'lucene-%s-src.tgz' >> % version, svnRevision, version, testArgs, baseURL) >>[smoker] File >> "/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/dev-tools/scripts/smokeTestRelease.py", >> line 583, in unpackAndVerify >>[smoker]
Re: Solr 5.3: is SOLR-7622 actually in?
Looks like it’s in, but definitely confusing in JIRA. Noble? Ryan? > On Aug 26, 2015, at 8:43 AM, Alexandre Rafalovitch wrote: > > In the release notes, SOLR-7622 is listed as new, but the issue itself > is marked unresolved and not targeted. Nor did it look finished. > > Just a - belated - sanity check. > > Regards, >Alex. > > Solr Analyzers, Tokenizers, Filters, URPs and even a newsletter: > http://www.solr-start.com/ > > - > To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org > For additional commands, e-mail: dev-h...@lucene.apache.org > - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: [JENKINS] Lucene-Solr-SmokeRelease-5.x - Build # 302 - Still Failing
Noble hasn't completed the post install steps which add backwards compatibility tests for the 5.3.0 release. Since he is on vacation for a few days, I'll add those indexes to avoid this failure. On Wed, Aug 26, 2015 at 8:59 AM, Apache Jenkins Server wrote: > Build: https://builds.apache.org/job/Lucene-Solr-SmokeRelease-5.x/302/ > > No tests ran. > > Build Log: > [...truncated 52844 lines...] > prepare-release-no-sign: > [mkdir] Created dir: > /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/lucene/build/smokeTestRelease/dist > [copy] Copying 461 files to > /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/lucene/build/smokeTestRelease/dist/lucene > [copy] Copying 245 files to > /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/lucene/build/smokeTestRelease/dist/solr >[smoker] Java 1.7 > JAVA_HOME=/home/jenkins/jenkins-slave/tools/hudson.model.JDK/latest1.7 >[smoker] Java 1.8 > JAVA_HOME=/home/jenkins/jenkins-slave/tools/hudson.model.JDK/latest1.8 >[smoker] NOTE: output encoding is UTF-8 >[smoker] >[smoker] Load release URL > "file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/lucene/build/smokeTestRelease/dist/"... >[smoker] >[smoker] Test Lucene... >[smoker] test basics... >[smoker] get KEYS >[smoker] 0.1 MB in 0.01 sec (12.1 MB/sec) >[smoker] check changes HTML... >[smoker] download lucene-5.4.0-src.tgz... >[smoker] 28.5 MB in 0.04 sec (708.0 MB/sec) >[smoker] verify md5/sha1 digests >[smoker] download lucene-5.4.0.tgz... >[smoker] 65.8 MB in 0.09 sec (716.3 MB/sec) >[smoker] verify md5/sha1 digests >[smoker] download lucene-5.4.0.zip... >[smoker] 76.0 MB in 0.11 sec (690.8 MB/sec) >[smoker] verify md5/sha1 digests >[smoker] unpack lucene-5.4.0.tgz... >[smoker] verify JAR metadata/identity/no javax.* or java.* classes... >[smoker] test demo with 1.7... >[smoker] got 6063 hits for query "lucene" >[smoker] checkindex with 1.7... >[smoker] test demo with 1.8... >[smoker] got 6063 hits for query "lucene" >[smoker] checkindex with 1.8... >[smoker] check Lucene's javadoc JAR >[smoker] unpack lucene-5.4.0.zip... >[smoker] verify JAR metadata/identity/no javax.* or java.* classes... >[smoker] test demo with 1.7... >[smoker] got 6063 hits for query "lucene" >[smoker] checkindex with 1.7... >[smoker] test demo with 1.8... >[smoker] got 6063 hits for query "lucene" >[smoker] checkindex with 1.8... >[smoker] check Lucene's javadoc JAR >[smoker] unpack lucene-5.4.0-src.tgz... >[smoker] make sure no JARs/WARs in src dist... >[smoker] run "ant validate" >[smoker] run tests w/ Java 7 and testArgs='-Dtests.slow=false'... >[smoker] test demo with 1.7... >[smoker] got 213 hits for query "lucene" >[smoker] checkindex with 1.7... >[smoker] generate javadocs w/ Java 7... >[smoker] >[smoker] Crawl/parse... >[smoker] >[smoker] Verify... >[smoker] run tests w/ Java 8 and testArgs='-Dtests.slow=false'... >[smoker] test demo with 1.8... >[smoker] got 213 hits for query "lucene" >[smoker] checkindex with 1.8... >[smoker] generate javadocs w/ Java 8... >[smoker] >[smoker] Crawl/parse... >[smoker] >[smoker] Verify... >[smoker] confirm all releases have coverage in TestBackwardsCompatibility >[smoker] find all past Lucene releases... >[smoker] run TestBackwardsCompatibility.. >[smoker] Releases that don't seem to be tested: >[smoker] 5.3.0 >[smoker] Traceback (most recent call last): >[smoker] File > "/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/dev-tools/scripts/smokeTestRelease.py", > line 1449, in >[smoker] main() >[smoker] File > "/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/dev-tools/scripts/smokeTestRelease.py", > line 1394, in main >[smoker] smokeTest(c.java, c.url, c.revision, c.version, c.tmp_dir, > c.is_signed, ' '.join(c.test_args)) >[smoker] File > "/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/dev-tools/scripts/smokeTestRelease.py", > line 1432, in smokeTest >[smoker] unpackAndVerify(java, 'lucene', tmpDir, 'lucene-%s-src.tgz' % > version, svnRevision, version, testArgs, baseURL) >[smoker] File > "/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/dev-tools/scripts/smokeTestRelease.py", > line 583, in unpackAndVerify >[smoker] verifyUnpacked(java, project, artifact, unpackPath, > svnRevision, version, testArgs, tmpDir, baseURL) >[smoker] File > "/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-SmokeRelease-5.x/dev-tools/scripts/smokeTestRelease.py", > line 7
Solr 5.3: is SOLR-7622 actually in?
In the release notes, SOLR-7622 is listed as new, but the issue itself is marked unresolved and not targeted. Nor did it look finished. Just a - belated - sanity check. Regards, Alex. Solr Analyzers, Tokenizers, Filters, URPs and even a newsletter: http://www.solr-start.com/ - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7961) Add version command to bin/solr start script
[ https://issues.apache.org/jira/browse/SOLR-7961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14713044#comment-14713044 ] ASF subversion and git services commented on SOLR-7961: --- Commit 1697910 from jan...@apache.org in branch 'dev/branches/branch_5x' [ https://svn.apache.org/r1697910 ] SOLR-7961: Add version command to bin/solr start script. Also adds -h for help (backport) > Add version command to bin/solr start script > > > Key: SOLR-7961 > URL: https://issues.apache.org/jira/browse/SOLR-7961 > Project: Solr > Issue Type: New Feature > Components: scripts and tools >Reporter: Jan Høydahl >Assignee: Jan Høydahl >Priority: Trivial > Fix For: 5.4 > > Attachments: SOLR-7961.patch > > > It would be nice to be able to tell which version of Solr you have. You can > get it with the {{status}} command today, but only if Solr is already > running. Proposal: > {noformat} > $ bin/solr -version > 5.3.0 > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-7961) Add version command to bin/solr start script
[ https://issues.apache.org/jira/browse/SOLR-7961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jan Høydahl resolved SOLR-7961. --- Resolution: Fixed > Add version command to bin/solr start script > > > Key: SOLR-7961 > URL: https://issues.apache.org/jira/browse/SOLR-7961 > Project: Solr > Issue Type: New Feature > Components: scripts and tools >Reporter: Jan Høydahl >Assignee: Jan Høydahl >Priority: Trivial > Fix For: 5.4 > > Attachments: SOLR-7961.patch > > > It would be nice to be able to tell which version of Solr you have. You can > get it with the {{status}} command today, but only if Solr is already > running. Proposal: > {noformat} > $ bin/solr -version > 5.3.0 > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (SOLR-7961) Add version command to bin/solr start script
[ https://issues.apache.org/jira/browse/SOLR-7961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14713032#comment-14713032 ] ASF subversion and git services commented on SOLR-7961: --- Commit 1697904 from jan...@apache.org in branch 'dev/trunk' [ https://svn.apache.org/r1697904 ] SOLR-7961: Add version command to bin/solr start script. Also adds -h for help > Add version command to bin/solr start script > > > Key: SOLR-7961 > URL: https://issues.apache.org/jira/browse/SOLR-7961 > Project: Solr > Issue Type: New Feature > Components: scripts and tools >Reporter: Jan Høydahl >Assignee: Jan Høydahl >Priority: Trivial > Fix For: 5.4 > > Attachments: SOLR-7961.patch > > > It would be nice to be able to tell which version of Solr you have. You can > get it with the {{status}} command today, but only if Solr is already > running. Proposal: > {noformat} > $ bin/solr -version > 5.3.0 > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-trunk-Windows (64bit/jdk1.8.0_60) - Build # 5202 - Failure!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/5202/ Java: 64bit/jdk1.8.0_60 -XX:-UseCompressedOops -XX:+UseParallelGC 2 tests failed. FAILED: org.apache.solr.handler.TestReplicationHandlerBackup.doTestBackup Error Message: Test abandoned because suite timeout was reached. Stack Trace: java.lang.Exception: Test abandoned because suite timeout was reached. at __randomizedtesting.SeedInfo.seed([B466C333BCD6B09]:0) FAILED: junit.framework.TestSuite.org.apache.solr.handler.TestReplicationHandlerBackup Error Message: Suite timeout exceeded (>= 720 msec). Stack Trace: java.lang.Exception: Suite timeout exceeded (>= 720 msec). at __randomizedtesting.SeedInfo.seed([B466C333BCD6B09]:0) Build Log: [...truncated 11052 lines...] [junit4] Suite: org.apache.solr.handler.TestReplicationHandlerBackup [junit4] 2> Creating dataDir: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.handler.TestReplicationHandlerBackup_B466C333BCD6B09-001\init-core-data-001 [junit4] 2> 2667536 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.SolrTestCaseJ4 ###Starting testBackupOnCommit [junit4] 2> 2667537 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.SolrTestCaseJ4 Writing core.properties file to C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.handler.TestReplicationHandlerBackup_B466C333BCD6B09-001\solr-instance-001\collection1 [junit4] 2> 2667546 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.e.j.s.Server jetty-9.2.13.v20150730 [junit4] 2> 2667548 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@52f8bdf8{/solr,null,AVAILABLE} [junit4] 2> 2667550 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.e.j.s.ServerConnector Started ServerConnector@39be83e5{HTTP/1.1}{127.0.0.1:57610} [junit4] 2> 2667550 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.e.j.s.Server Started @2678201ms [junit4] 2> 2667550 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.handler.TestReplicationHandlerBackup_B466C333BCD6B09-001\solr-instance-001\collection1\data, hostContext=/solr, hostPort=57610} [junit4] 2> 2667551 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.s.SolrDispatchFilter SolrDispatchFilter.init(): sun.misc.Launcher$AppClassLoader@4e0e2f2a [junit4] 2> 2667551 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.c.SolrResourceLoader new SolrResourceLoader for directory: 'C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.handler.TestReplicationHandlerBackup_B466C333BCD6B09-001\solr-instance-001\' [junit4] 2> 2667569 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.c.SolrXmlConfig Loading container configuration from C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.handler.TestReplicationHandlerBackup_B466C333BCD6B09-001\solr-instance-001\solr.xml [junit4] 2> 2667577 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.c.CoresLocator Config-defined core root directory: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.handler.TestReplicationHandlerBackup_B466C333BCD6B09-001\solr-instance-001\. [junit4] 2> 2667577 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.c.CoreContainer New CoreContainer 14928761 [junit4] 2> 2667577 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.c.CoreContainer Loading cores into CoreContainer [instanceDir=C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.handler.TestReplicationHandlerBackup_B466C333BCD6B09-001\solr-instance-001\] [junit4] 2> 2667577 INFO (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.c.CoreContainer loading shared library: C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J0\temp\solr.handler.TestReplicationHandlerBackup_B466C333BCD6B09-001\solr-instance-001\lib [junit4] 2> 2667577 WARN (TEST-TestReplicationHandlerBackup.testBackupOnCommit-seed#[B466C333BCD6B09]) [ ] o.a.s.c.So
[JENKINS] Lucene-Solr-NightlyTests-trunk - Build # 776 - Still Failing
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-trunk/776/ 3 tests failed. REGRESSION: org.apache.solr.cloud.TestRebalanceLeaders.test Error Message: No live SolrServers available to handle this request:[http://127.0.0.1:57241, http://127.0.0.1:56384, http://127.0.0.1:40438, http://127.0.0.1:50878, http://127.0.0.1:41922] Stack Trace: org.apache.solr.client.solrj.SolrServerException: No live SolrServers available to handle this request:[http://127.0.0.1:57241, http://127.0.0.1:56384, http://127.0.0.1:40438, http://127.0.0.1:50878, http://127.0.0.1:41922] at __randomizedtesting.SeedInfo.seed([1A55CE4403CDBAD5:9201F19EAD31D72D]:0) at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:352) at org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1086) at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:857) at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:800) at org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1220) at org.apache.solr.cloud.TestRebalanceLeaders.issueCommands(TestRebalanceLeaders.java:281) at org.apache.solr.cloud.TestRebalanceLeaders.rebalanceLeaderTest(TestRebalanceLeaders.java:109) at org.apache.solr.cloud.TestRebalanceLeaders.test(TestRebalanceLeaders.java:75) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:963) at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:938) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(St
[jira] [Created] (SOLR-7977) SOLR_HOST in solr.in.sh doesn't apply to Jetty's host property
Shalin Shekhar Mangar created SOLR-7977: --- Summary: SOLR_HOST in solr.in.sh doesn't apply to Jetty's host property Key: SOLR-7977 URL: https://issues.apache.org/jira/browse/SOLR-7977 Project: Solr Issue Type: Bug Components: security, SolrCloud Reporter: Shalin Shekhar Mangar Fix For: Trunk, 5.4 [~sdavids] pointed out that the SOLR_HOST config option in solr.in.sh doesn't set Jetty's host property (solr.jetty.host) so it still binds to all net interfaces. Perhaps it should apply to jetty as well because the user explicitly wants us to bind to specific IP? -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-7976) Jetty http and https connectors use different property names for host/port
Shalin Shekhar Mangar created SOLR-7976: --- Summary: Jetty http and https connectors use different property names for host/port Key: SOLR-7976 URL: https://issues.apache.org/jira/browse/SOLR-7976 Project: Solr Issue Type: Bug Affects Versions: 4.10.4, 5.3 Reporter: Shalin Shekhar Mangar Assignee: Shalin Shekhar Mangar Priority: Minor Fix For: Trunk, 5.4 Jetty http and https connectors use different property names for host/port i.e. jetty.host vs solr.jetty.host and jetty.port vs solr.jetty.port. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6570) Make BooleanQuery immutable
[ https://issues.apache.org/jira/browse/LUCENE-6570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712958#comment-14712958 ] Greg Huber commented on LUCENE-6570: I guess it will work as its public, not much else I can do! Cheers Greg. :-) > Make BooleanQuery immutable > --- > > Key: LUCENE-6570 > URL: https://issues.apache.org/jira/browse/LUCENE-6570 > Project: Lucene - Core > Issue Type: Task >Reporter: Adrien Grand >Assignee: Adrien Grand >Priority: Minor > Fix For: 5.3, 6.0 > > Attachments: LUCENE-6570.patch > > > In the same spirit as LUCENE-6531 for the PhraseQuery, we should make > BooleanQuery immutable. > The plan is the following: > - create BooleanQuery.Builder with the same setters as BooleanQuery today > (except setBoost) and a build() method that returns a BooleanQuery > - remove setters from BooleanQuery (except setBoost) > I would also like to add some static utility methods for common use-cases of > this query, for instance: > - static BooleanQuery disjunction(Query... queries) to create a disjunction > - static BooleanQuery conjunction(Query... queries) to create a conjunction > - static BooleanQuery filtered(Query query, Query... filters) to create a > filtered query > Hopefully this will help keep tests not too verbose, and the latter will also > help with the FilteredQuery derecation/removal. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (LUCENE-6763) Make MultiPhraseQuery immutable
Adrien Grand created LUCENE-6763: Summary: Make MultiPhraseQuery immutable Key: LUCENE-6763 URL: https://issues.apache.org/jira/browse/LUCENE-6763 Project: Lucene - Core Issue Type: Task Reporter: Adrien Grand Priority: Minor We should make MultiPhraseQuery immutable similarly to PhraseQuery. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6570) Make BooleanQuery immutable
[ https://issues.apache.org/jira/browse/LUCENE-6570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712947#comment-14712947 ] Adrien Grand commented on LUCENE-6570: -- Maybe you can override `addContextToQuery` to override the clause? > Make BooleanQuery immutable > --- > > Key: LUCENE-6570 > URL: https://issues.apache.org/jira/browse/LUCENE-6570 > Project: Lucene - Core > Issue Type: Task >Reporter: Adrien Grand >Assignee: Adrien Grand >Priority: Minor > Fix For: 5.3, 6.0 > > Attachments: LUCENE-6570.patch > > > In the same spirit as LUCENE-6531 for the PhraseQuery, we should make > BooleanQuery immutable. > The plan is the following: > - create BooleanQuery.Builder with the same setters as BooleanQuery today > (except setBoost) and a build() method that returns a BooleanQuery > - remove setters from BooleanQuery (except setBoost) > I would also like to add some static utility methods for common use-cases of > this query, for instance: > - static BooleanQuery disjunction(Query... queries) to create a disjunction > - static BooleanQuery conjunction(Query... queries) to create a conjunction > - static BooleanQuery filtered(Query query, Query... filters) to create a > filtered query > Hopefully this will help keep tests not too verbose, and the latter will also > help with the FilteredQuery derecation/removal. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Updated] (SOLR-7971) Reduce memory allocated by JavaBinCodec to encode large strings
[ https://issues.apache.org/jira/browse/SOLR-7971?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shalin Shekhar Mangar updated SOLR-7971: Attachment: SOLR-7971-directbuffer.patch Here's another idea as a patch to further reduce heap requirement. In this patch I use a direct byte buffer to hold the encoded bytes and limit the intermediate on-heap buffer to 64KB only. This optimization kicks in only if the max bytes required by the string being serialized is greater than 64KB. With this patch I can index the same 100MB JSON document with 1200MB of heap. [~ysee...@gmail.com] - Thoughts? > Reduce memory allocated by JavaBinCodec to encode large strings > --- > > Key: SOLR-7971 > URL: https://issues.apache.org/jira/browse/SOLR-7971 > Project: Solr > Issue Type: Sub-task > Components: Response Writers, SolrCloud >Reporter: Shalin Shekhar Mangar >Assignee: Shalin Shekhar Mangar >Priority: Minor > Fix For: Trunk, 5.4 > > Attachments: SOLR-7971-directbuffer.patch, SOLR-7971.patch > > > As discussed in SOLR-7927, we can reduce the buffer memory allocated by > JavaBinCodec while writing large strings. > https://issues.apache.org/jira/browse/SOLR-7927?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14700420#comment-14700420 > {quote} > The maximum Unicode code point (as of Unicode 8 anyway) is U+10 > ([http://www.unicode.org/glossary/#code_point]). This is encoded in UTF-16 > as surrogate pair {{\uDBFF\uDFFF}}, which takes up two Java chars, and is > represented in UTF-8 as the 4-byte sequence {{F4 8F BF BF}}. This is likely > where the mistaken 4-bytes-per-Java-char formulation came from: the maximum > number of UTF-8 bytes required to represent a Unicode *code point* is 4. > The maximum Java char is {{\u}}, which is represented in UTF-8 as the > 3-byte sequence {{EF BF BF}}. > So I think it's safe to switch to using 3 bytes per Java char (the unit of > measurement returned by {{String.length()}}), like > {{CompressingStoredFieldsWriter.writeField()}} does. > {quote} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6570) Make BooleanQuery immutable
[ https://issues.apache.org/jira/browse/LUCENE-6570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712935#comment-14712935 ] Greg Huber commented on LUCENE-6570: btw, I have multiple contexts so call AnalyzingInfixSuggester.suggester.lookup(term, contexts, nMax, true, true); which will then call AnalyzingInfixSuggester.toQuery(..) eventually, which adds the context with the BooleanClause.Occur.SHOULD. Its a private method so is there a way to override this? private BooleanQuery toQuery(Set contextInfo) { if (contextInfo == null || contextInfo.isEmpty()) { return null; } BooleanQuery.Builder contextFilter = new BooleanQuery.Builder(); for (BytesRef context : contextInfo) { addContextToQuery(contextFilter, context, BooleanClause.Occur.SHOULD); } return contextFilter.build(); } > Make BooleanQuery immutable > --- > > Key: LUCENE-6570 > URL: https://issues.apache.org/jira/browse/LUCENE-6570 > Project: Lucene - Core > Issue Type: Task >Reporter: Adrien Grand >Assignee: Adrien Grand >Priority: Minor > Fix For: 5.3, 6.0 > > Attachments: LUCENE-6570.patch > > > In the same spirit as LUCENE-6531 for the PhraseQuery, we should make > BooleanQuery immutable. > The plan is the following: > - create BooleanQuery.Builder with the same setters as BooleanQuery today > (except setBoost) and a build() method that returns a BooleanQuery > - remove setters from BooleanQuery (except setBoost) > I would also like to add some static utility methods for common use-cases of > this query, for instance: > - static BooleanQuery disjunction(Query... queries) to create a disjunction > - static BooleanQuery conjunction(Query... queries) to create a conjunction > - static BooleanQuery filtered(Query query, Query... filters) to create a > filtered query > Hopefully this will help keep tests not too verbose, and the latter will also > help with the FilteredQuery derecation/removal. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Resolved] (SOLR-7972) VelocityResponseWriter template encoding issue
[ https://issues.apache.org/jira/browse/SOLR-7972?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Erik Hatcher resolved SOLR-7972. Resolution: Fixed > VelocityResponseWriter template encoding issue > -- > > Key: SOLR-7972 > URL: https://issues.apache.org/jira/browse/SOLR-7972 > Project: Solr > Issue Type: Bug >Affects Versions: 5.3 >Reporter: Erik Hatcher >Assignee: Erik Hatcher >Priority: Critical > Fix For: Trunk, 5.4 > > Attachments: SOLR-7972.patch > > > I'm not sure when this got introduced (5.0 maybe?) - the .vm templates are > loaded using ISO-8859-1 rather than UTF-8 as it should be. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Created] (SOLR-7975) Support payloads on primitive types
Jamie Johnson created SOLR-7975: --- Summary: Support payloads on primitive types Key: SOLR-7975 URL: https://issues.apache.org/jira/browse/SOLR-7975 Project: Solr Issue Type: New Feature Components: clients - java, Server Reporter: Jamie Johnson Currently payloads are supported through the use of an analysis chain, this limits the ability to provide payloads on primitive fields like Trie, Bool, etc without copying these classes and adding the ability in custom code. It would be great if payloads could be added to these field types in a pluggable way similar to what is supported for non primitive types, perhaps through extending the base primitive implementations. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6570) Make BooleanQuery immutable
[ https://issues.apache.org/jira/browse/LUCENE-6570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712881#comment-14712881 ] Uwe Schindler commented on LUCENE-6570: --- The boolean clauses have to be created with MUST by from the beginning. > Make BooleanQuery immutable > --- > > Key: LUCENE-6570 > URL: https://issues.apache.org/jira/browse/LUCENE-6570 > Project: Lucene - Core > Issue Type: Task >Reporter: Adrien Grand >Assignee: Adrien Grand >Priority: Minor > Fix For: 5.3, 6.0 > > Attachments: LUCENE-6570.patch > > > In the same spirit as LUCENE-6531 for the PhraseQuery, we should make > BooleanQuery immutable. > The plan is the following: > - create BooleanQuery.Builder with the same setters as BooleanQuery today > (except setBoost) and a build() method that returns a BooleanQuery > - remove setters from BooleanQuery (except setBoost) > I would also like to add some static utility methods for common use-cases of > this query, for instance: > - static BooleanQuery disjunction(Query... queries) to create a disjunction > - static BooleanQuery conjunction(Query... queries) to create a conjunction > - static BooleanQuery filtered(Query query, Query... filters) to create a > filtered query > Hopefully this will help keep tests not too verbose, and the latter will also > help with the FilteredQuery derecation/removal. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (LUCENE-6570) Make BooleanQuery immutable
[ https://issues.apache.org/jira/browse/LUCENE-6570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712881#comment-14712881 ] Uwe Schindler edited comment on LUCENE-6570 at 8/26/15 10:13 AM: - The boolean clauses have to be created with MUST from the beginning. was (Author: thetaphi): The boolean clauses have to be created with MUST by from the beginning. > Make BooleanQuery immutable > --- > > Key: LUCENE-6570 > URL: https://issues.apache.org/jira/browse/LUCENE-6570 > Project: Lucene - Core > Issue Type: Task >Reporter: Adrien Grand >Assignee: Adrien Grand >Priority: Minor > Fix For: 5.3, 6.0 > > Attachments: LUCENE-6570.patch > > > In the same spirit as LUCENE-6531 for the PhraseQuery, we should make > BooleanQuery immutable. > The plan is the following: > - create BooleanQuery.Builder with the same setters as BooleanQuery today > (except setBoost) and a build() method that returns a BooleanQuery > - remove setters from BooleanQuery (except setBoost) > I would also like to add some static utility methods for common use-cases of > this query, for instance: > - static BooleanQuery disjunction(Query... queries) to create a disjunction > - static BooleanQuery conjunction(Query... queries) to create a conjunction > - static BooleanQuery filtered(Query query, Query... filters) to create a > filtered query > Hopefully this will help keep tests not too verbose, and the latter will also > help with the FilteredQuery derecation/removal. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6699) Integrate lat/lon BKD and spatial3d
[ https://issues.apache.org/jira/browse/LUCENE-6699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712798#comment-14712798 ] ASF subversion and git services commented on LUCENE-6699: - Commit 1697865 from [~mikemccand] in branch 'dev/branches/lucene6699' [ https://svn.apache.org/r1697865 ] LUCENE-6699: comment out assert > Integrate lat/lon BKD and spatial3d > --- > > Key: LUCENE-6699 > URL: https://issues.apache.org/jira/browse/LUCENE-6699 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless >Assignee: Michael McCandless > Attachments: Geo3DPacking.java, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch > > > I'm opening this for discussion, because I'm not yet sure how to do > this integration, because of my ignorance about spatial in general and > spatial3d in particular :) > Our BKD tree impl is very fast at doing lat/lon shape intersection > (bbox, polygon, soon distance: LUCENE-6698) against previously indexed > points. > I think to integrate with spatial3d, we would first need to record > lat/lon/z into doc values. Somewhere I saw discussion about how we > could stuff all 3 into a single long value with acceptable precision > loss? Or, we could use BinaryDocValues? We need all 3 dims available > to do the fast per-hit query time filtering. > But, second: what do we index into the BKD tree? Can we "just" index > earth surface lat/lon, and then at query time is spatial3d able to > give me an enclosing "surface lat/lon" bbox for a 3d shape? Or > ... must we index all 3 dimensions into the BKD tree (seems like this > could be somewhat wasteful)? -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6759) Integrate lat/long BKD and spatial 3d, part 2
[ https://issues.apache.org/jira/browse/LUCENE-6759?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712793#comment-14712793 ] Michael McCandless commented on LUCENE-6759: OK I'll disable this assert: it is too anal. > Integrate lat/long BKD and spatial 3d, part 2 > - > > Key: LUCENE-6759 > URL: https://issues.apache.org/jira/browse/LUCENE-6759 > Project: Lucene - Core > Issue Type: New Feature >Reporter: Michael McCandless > Attachments: LUCENE-6699.patch, LUCENE-6699.patch, LUCENE-6699.patch, > LUCENE-6699.patch > > > This is just a continuation of LUCENE-6699, which became too big. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[JENKINS] Lucene-Solr-5.x-MacOSX (64bit/jdk1.8.0) - Build # 2623 - Still Failing!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-MacOSX/2623/ Java: 64bit/jdk1.8.0 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC 1 tests failed. FAILED: org.apache.lucene.TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler Error Message: Stack Trace: java.lang.AssertionError at __randomizedtesting.SeedInfo.seed([A2D285B04961A340:2553381D4D41D944]:0) at org.junit.Assert.fail(Assert.java:92) at org.junit.Assert.assertTrue(Assert.java:43) at org.junit.Assert.assertTrue(Assert.java:54) at org.apache.lucene.TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler(TestMergeSchedulerExternal.java:116) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845) at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747) at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365) at java.lang.Thread.run(Thread.java:745) Build Log: [...truncated 1086 lines...] [junit4] Suite: org.apache.lucene.TestMergeSchedulerExternal [junit4] 2> NOTE: reproduce with: ant test -Dtestcase=TestMergeSchedulerExternal -Dtests.method=testSubclassConcurrentMergeScheduler -Dtests.seed=A2D285B04961A340 -Dtests.slow=true -Dtests.locale=it -Dtests.timezone=Pacific/Majuro -Dtests.asserts=true -Dtests.file.encoding=UTF-8 [junit4] FAILURE 0.57s J0 | TestMergeSchedulerExternal.testSubclassConcurrentMergeScheduler <<< [junit4]> Throwable #1: java.lang.AssertionError [junit4]>at __randomizedtesting.SeedInfo.se
[jira] [Commented] (SOLR-7972) VelocityResponseWriter template encoding issue
[ https://issues.apache.org/jira/browse/SOLR-7972?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712721#comment-14712721 ] Erik Hatcher commented on SOLR-7972: [~anshumg] thanks! I just woke up to the build errors and was fixing it when I noticed you beat me to it. And thanks for removing the 5.3.1 fix version - if we do make such a release this can be merged over there but maybe no need yet. > VelocityResponseWriter template encoding issue > -- > > Key: SOLR-7972 > URL: https://issues.apache.org/jira/browse/SOLR-7972 > Project: Solr > Issue Type: Bug >Affects Versions: 5.3 >Reporter: Erik Hatcher >Assignee: Erik Hatcher >Priority: Critical > Fix For: Trunk, 5.4 > > Attachments: SOLR-7972.patch > > > I'm not sure when this got introduced (5.0 maybe?) - the .vm templates are > loaded using ISO-8859-1 rather than UTF-8 as it should be. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Comment Edited] (LUCENE-6570) Make BooleanQuery immutable
[ https://issues.apache.org/jira/browse/LUCENE-6570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712704#comment-14712704 ] Greg Huber edited comment on LUCENE-6570 at 8/26/15 8:36 AM: - Hello, I am extending AnalyzingInfixSuggester for use with my suggester where I change the query to be a AND rather than an OR in the finishQuery(..) method. ie /** * Subclass can override this to tweak the Query before searching. */ protected Query finishQuery(Builder in, boolean allTermsRequired) { // Update contexts to be ANDs (MUST) rather than ORs (SHOULD) for (BooleanClause booleanClause : in.build().clauses()) { // Change the contexts to be MUST (will be the only BooleanQuery and others will be TermQuery) if (booleanClause.getQuery() instanceof BooleanQuery) { BooleanQuery bq = (BooleanQuery) booleanClause.getQuery(); for (BooleanClause bc : bq) { bc.setOccur(BooleanClause.Occur.MUST); } // We are done break; } } return in.build(); } It says that BooleanClause.setOccur(..) is depreciated and will be immutable in 6.0, how would I then be able to do this? Cheers Greg was (Author: gregh99): Hello, I am extending AnalyzingInfixSuggester for use with my suggester where I change the query to be a AND rather than an OR in the finishQuery(..) method. ie /** * Subclass can override this to tweak the Query before searching. */ protected Query finishQuery(Builder in, boolean allTermsRequired) { // Update contexts to be ANDs (MUST) rather than ORs (SHOULD) for (BooleanClause booleanClause : in.build().clauses()) { // Change the contexts to be MUST (will be the only BooleanQuery and others will be TermQuery) if (booleanClause.getQuery() instanceof BooleanQuery) { BooleanQuery bq = (BooleanQuery) booleanClause.getQuery(); for (BooleanClause bc : bq) { bc.setOccur(BooleanClause.Occur.MUST); } // We are done break; } } return in.build(); } It says that BooleanClause.setOccur(..) is depreciated and will be immutable in 6.0, how would I then be able to do this? > Make BooleanQuery immutable > --- > > Key: LUCENE-6570 > URL: https://issues.apache.org/jira/browse/LUCENE-6570 > Project: Lucene - Core > Issue Type: Task >Reporter: Adrien Grand >Assignee: Adrien Grand >Priority: Minor > Fix For: 5.3, 6.0 > > Attachments: LUCENE-6570.patch > > > In the same spirit as LUCENE-6531 for the PhraseQuery, we should make > BooleanQuery immutable. > The plan is the following: > - create BooleanQuery.Builder with the same setters as BooleanQuery today > (except setBoost) and a build() method that returns a BooleanQuery > - remove setters from BooleanQuery (except setBoost) > I would also like to add some static utility methods for common use-cases of > this query, for instance: > - static BooleanQuery disjunction(Query... queries) to create a disjunction > - static BooleanQuery conjunction(Query... queries) to create a conjunction > - static BooleanQuery filtered(Query query, Query... filters) to create a > filtered query > Hopefully this will help keep tests not too verbose, and the latter will also > help with the FilteredQuery derecation/removal. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
[jira] [Commented] (LUCENE-6570) Make BooleanQuery immutable
[ https://issues.apache.org/jira/browse/LUCENE-6570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712704#comment-14712704 ] Greg Huber commented on LUCENE-6570: Hello, I am extending AnalyzingInfixSuggester for use with my suggester where I change the query to be a AND rather than an OR in the finishQuery(..) method. ie /** * Subclass can override this to tweak the Query before searching. */ protected Query finishQuery(Builder in, boolean allTermsRequired) { // Update contexts to be ANDs (MUST) rather than ORs (SHOULD) for (BooleanClause booleanClause : in.build().clauses()) { // Change the contexts to be MUST (will be the only BooleanQuery and others will be TermQuery) if (booleanClause.getQuery() instanceof BooleanQuery) { BooleanQuery bq = (BooleanQuery) booleanClause.getQuery(); for (BooleanClause bc : bq) { bc.setOccur(BooleanClause.Occur.MUST); } // We are done break; } } return in.build(); } It says that BooleanClause.setOccur(..) is depreciated and will be immutable in 6.0, how would I then be able to do this? > Make BooleanQuery immutable > --- > > Key: LUCENE-6570 > URL: https://issues.apache.org/jira/browse/LUCENE-6570 > Project: Lucene - Core > Issue Type: Task >Reporter: Adrien Grand >Assignee: Adrien Grand >Priority: Minor > Fix For: 5.3, 6.0 > > Attachments: LUCENE-6570.patch > > > In the same spirit as LUCENE-6531 for the PhraseQuery, we should make > BooleanQuery immutable. > The plan is the following: > - create BooleanQuery.Builder with the same setters as BooleanQuery today > (except setBoost) and a build() method that returns a BooleanQuery > - remove setters from BooleanQuery (except setBoost) > I would also like to add some static utility methods for common use-cases of > this query, for instance: > - static BooleanQuery disjunction(Query... queries) to create a disjunction > - static BooleanQuery conjunction(Query... queries) to create a conjunction > - static BooleanQuery filtered(Query query, Query... filters) to create a > filtered query > Hopefully this will help keep tests not too verbose, and the latter will also > help with the FilteredQuery derecation/removal. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
RE: JDK9 b78 Jenkins tests disabled with 32bits/-server compiler (for now)
There is now an issue @ OpenJDK: https://bugs.openjdk.java.net/browse/JDK-8134468 Uwe - Uwe Schindler H.-H.-Meier-Allee 63, D-28213 Bremen http://www.thetaphi.de eMail: u...@thetaphi.de > -Original Message- > From: Uwe Schindler [mailto:u...@thetaphi.de] > Sent: Tuesday, August 25, 2015 2:14 PM > To: dev@lucene.apache.org > Cc: rory.odonn...@oracle.com; Balchandra Vaidya; Dawid Weiss > Subject: JDK9 b78 Jenkins tests disabled with 32bits/-server compiler (for > now) > > Hi, > > after some digging with Dawid and running tests: > - JDK 9b78, 64bits seems fine > - JDK 9b78, 32bits seems fine with -client compiler > - JDK 9b78, 32bits with -server compiler is heavily broken, it fails every > test > run. You see the first test failing ASAP if you enable -Xbatch and tiered > compilation, > > So I disabled the -server variant for now on Jenkins to keep failed runs low. > I > will now open bug report. > > Uwe > > - > Uwe Schindler > H.-H.-Meier-Allee 63, D-28213 Bremen > http://www.thetaphi.de > eMail: u...@thetaphi.de > > > > > - > To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional > commands, e-mail: dev-h...@lucene.apache.org - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Should we EOL support for 32 bit systems?
Plus, it is fun as well. A bit like finding these: https://goo.gl/CQZPYh Dawid On Tue, Aug 25, 2015 at 11:37 PM, Jan Høydahl wrote: > Another reason to keep 32bit support is that people who already have > installed Java JRE from their browser tend to have the 32bit version (even if > on 64bit OS). So they will be able to test-run Solr without re-installing > Java. > > -- > Jan Høydahl, search solution architect > Cominvent AS - www.cominvent.com > >> 25. aug. 2015 kl. 18.47 skrev Erick Erickson : >> >> Fair enough, just wanted to be sure we weren't making extra work for >> ourselves. >> Well, actually extra work for you since you seem to be the one who interacts >> with the compiler folks the most ;). >> >> On Tue, Aug 25, 2015 at 9:43 AM, Uwe Schindler wrote: >>> Hi, >>> >>> From a Java point of view, there is no real difference to not support 32 >>> bit versions. The bug with JDK 9 are just bugs that could also have >>> happened with 64 bit versions. It is just easier to trigger this bug with >>> 32 bits, but I am almost sure the underlying bug also affect 64 bits. >>> >>> So why should we no longer support all platforms Java supports? Bitness >>> does not matter for our code? Should we then also no longer support Sparc, >>> PowerPC, or ARM platform? >>> -1 to add arbitrary restrictions on our runtime environment. If we want >>> this, we should disallow all platforms we don't test on and of course also >>> al processor variants we don't test on! :-) >>> >>> Uwe >>> >>> - >>> Uwe Schindler >>> H.-H.-Meier-Allee 63, D-28213 Bremen >>> http://www.thetaphi.de >>> eMail: u...@thetaphi.de >>> >>> -Original Message- From: Erick Erickson [mailto:erickerick...@gmail.com] Sent: Tuesday, August 25, 2015 6:23 PM To: dev@lucene.apache.org Subject: Should we EOL support for 32 bit systems? I have no real skin in this game, but I thought it worth asking after Uwe's recent e-mail about disabling 32bits with -server tests. I guess it boils down to "who is using 32-bit versions?". Are we spending time/energy supporting a configuration that is not useful to enough people to merit the effort? I'm perfectly content if the response is "That's a really stupid question to ask, of course we must continue to support 32-bit OSs". Although some evidence would be nice ;) It's just that nobody I work with is running 32 bit OS's. Whether that's just my limited exposure to people running small systems is certainly a valid question. If we _do_ decide to drop support for 32 bit systems, what's the right version? 6.0? Random thoughts on a slow Tuesday morning Erick - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org >>> >>> >>> - >>> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org >>> For additional commands, e-mail: dev-h...@lucene.apache.org >>> >> >> - >> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org >> For additional commands, e-mail: dev-h...@lucene.apache.org >> > > > - > To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org > For additional commands, e-mail: dev-h...@lucene.apache.org > - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org